Sunday, December 17, 2017
Friday, December 15, 2017
Microsoft is simplifying IoT through the use of the Azure IoT Suite and the following components:
- Azure IoT Hub – the gateway that allows you to ingest data
- Azure Time Series Insights – this enables you to graphically analyze and explore data points
- Microsoft IoT Central; a fully managed IoT SaaS
- Azure IoT Edge; a virtual representation of your IoT device that can download code to a physical device and have it execute locally.
Friday, September 29, 2017
The idea with Cloud is that each layer is responsible for its own availability and by combining these loosely coupled layers you get higher availability. It should be a consideration before you begin deployment. For example if you start by considering maintenance and how and when you would take a planned outage it informs your design. You should predict the types of failures you can experience so you are mitigating it in your architecture.
There is an emerging concept on Hyper-scale Cloud platforms called grey failures. A grey failure is when your VM, workloads or applications are not down but are not getting the resources they need.
Monitoring and Alerting should be on for any VMs running in IaaS. When you open a ticket in Azure any known issues are surfaced as part of the support request process. This is part of Azure’s automated analytics engine providing support before you input the ticket.
Backup and DR plans should be applied to your VM. Azure allows you to create granular retention policies. When you recover VMs you can restore over the existing or create a new VM. For DR you can leverage ASR to replicate VMs outside another region. ASR is not multi-target however so you could not replicate VMs from the enterprise to an Azure region and then replicate them to a different one. It would be two distinct steps, first replicate and failover the VM to Azure and then you can setup replication between two regions.
Maintenance Microsoft now provides a local endpoint in a region with a simple REST API that provides information on upcoming maintenance events. These can be surfaced within the VM so you can trigger soft shutdowns of your virtual instance. For example, if you have a single VM (outside an availability set) and the host is being patched the VM can complete a graceful shutdown.
Azure uses VM preserving technology when they do underlying host maintenance. For updates that do not require a reboot of the host the VM is frozen for a few seconds while the host files are updated. For most applications this is seamless however if it is impactful you can use the REST API to create a reaction.
Microsoft collects all host reboot requirements so that they are applied at once vs. periodically throughout the year to improve platform availability. You are preemptively notified 30 days out for these events. One notification is sent per subscription to the administrator. The customer can add additional recipients.
Availability Sets is a logical grouping of VMs within a datacenter that allows Azure to understand how your application is built to provide for redundancy and availability. Microsoft recommends that two or more VMs are created within an availability set. To have a 99.95% SLA you need to deploy your VMs in an Availability Sets. Availability Sets provide you fault isolation for your compute.
An Availability Set with Managed Disks is call a Managed Availability Set. With Managed Availability Set you get fault isolation for compute and storage. Essentially it ensures that the managed VM Disks are not on the same storage.
The AzureRM vision is to capture everything you might do or envision in Cloud. This should extend from infrastructure, configuration to governance and security.
Azure is seeing about 200 ARM templates deployed per second. The session will focus on some of the template enhancements and how Microsoft is more closely integrating identity management and delivering new features.
You now have the ability to deploy ARM deployments across subscriptions (service providers pay attention!). You can also deploy across resource groups. The two declaratives within the ARM template are that enable this are:
You may be wondering how you share your templates, increase the reliability and support them after a deployment?
Managed applications is the ability to simplify template sharing. Managed applications can be shared or sold, they are meant to be simple to deploy, they are contained so they cannot be broken and they can be connected. Connected means you define what level of access you need to it after it has been deployed for ongoing management and support.
For additional details on Managed applications please see https://docs.microsoft.com/en-us/azure/azure-resource-manager/managed-application-overview .
Managed applications are available in West US and West US Central but will be global by the end of the year. When you define a managed application through the Azure portal you determine if it is locked or unlocked. If it is locked you need to define who is authorized to write to it.
By default Managed Applications are deployed within your subscription. From within the access pane of the Managed Application you can share it to other users and subscriptions. Delivering Managed Applications to the Azure marketplace is in Public Preview at this moment.
With Managed Identity you can now create virtual machines with a service principal provided by Azure active directory. This allows the VM to get a token to enable service access to avoid having passwords and credentials in code. To learn more have a look here
ARM Templates & Event Grid
You can use Event Grid to collect all ARM events and requests which can be pushed to an end point or listener. To learn more on Event Grid read here
You can use Resource Policies to do Location Ringfencing. Location Ringfencing allows you to define a policy to ensure your data does not leave a certain location.
You can also restrict which VM Classes that people can use. For example to prevent your developers from deploying extremely expensive classes of VMs.
Policies can be used to limit the access to all the marketplace images to just a few. You can find many starting point policies on GitHub
Azure Policies are in Preview and additional information can be found here:
It is important that customers are configured with a single Identity or Tenant. You should look at the Identity as the Control Plane or the single source of truth. Azure Active Director “AD” has grow 30% in Year-over-Year growth to 12.8 million customers. In addition there are now 272,000 Apps in Azure AD. Ironically the most used application in Azure AD is Google Apps. Customers are using Azure AD to authenticate Google Services.
Azure AD is included with O365 so there is no additional cost. Identity in O365 consists of three different types of users:
- Cloud Identity: accounts live in O365 only
- Synchronized Identity: accounts sync with a local AD Server
- Federated Identity: Certificate based authentication based on an on premises deployment of AD Federation Service.
The Identity can be managed using several methods.
Password Hash Sync ensures you have the same password on-premises as in the cloud. The con to Hash sync is that disabled or user edits are not updated until the sync cycle is complete. In hash sync the hashes on-prem are not identical to those in the cloud but the passwords are the same.
Pass-through Authentication You still have the same password but passwords remain on-premises. There is a Pass-through Agent “PTA” agent that gets installed on your enterprise AD server. THE PTA Agent handles the queuing of requests from Azure AD and sends the validations back once authenticated.
Seamless Single Sign-On works with both Password Hash Sync and Pass-through Authentication. This is done with no additional requirement onsite. SSO is enabled during the installation of AD Connect.
You do not need more than on Azure AD if you have more than one AD on premises. One Azure AD can support hundreds of unique domain names. You can also mix cloud only accounts and on prem synchronized accounts. You can use PowerShell Graph API vs. AD Connect to synchronize and manage users and groups but it is much more difficult. AD Connect is necessary for Hybrid Exchange however.
There are six general use cases for Azure AD:
- Employee Secure Access to Applications
- To leverage Dynamic Groups for automated application deployment. Dynamic groups allow move, join and leave workflow processes
- To federate access for Business-to-Business communication and collaboration (included in Azure AD, 1 license enables 5 collaborations)
- Advanced threat and Identity protection. This is enabled through conditional access based on device compliance.
- To Abide by Governance and Compliance industry regulations. Access Review is in public review which identifies accounts that have not accessed the system for a while and prompts the administrator to review them.
- To leverage O365 as an application development platform
With AD Premium you get AD Connect Health, Dynamic Group memberships, Multi-Factor Authentication for all objects that can be applied when needed vs always on. In addition there is a better overall end user experience.
Thursday, September 28, 2017
Microsoft Ignite 2017: Protect Azure IaaS Deployments using Microsoft Security Center with Sarah Gender & Adwait Joshi
Adopting Cloud no longer has a security barrier. It is however a shared responsibility between the cloud provider and tenant. It is important that the tenant understands this principle so they properly secure their resources.
Securing IaaS is not just about just securing VMs but also the networking and services like storage. It is also about securing multiple clouds as many customers have a multi-cloud strategy. While things like malware protection need to be applied in the cloud the application is different in the cloud. Key challenges specific to cloud are:
- Visibility and Control
- Management Complexity (a mix of IaaS, PaaS and SaaS components)
- Rapidly Evolving Threats (you need a solution optimized for cloud as things are more dynamic)
Microsoft ensures that Azure is built on a Secure Foundation by enforcing physical, infrastructure and operational security. Microsoft provides the controls but the customer or tenant is responsible for Identity & Access, Information Protection, Threat Protection and Security Management.
10 Ways Azure Security Center helps protect IaaS deployments
1 Monitor security state of cloud resources
- Security Center Automatically discovers and monitors Azure resources
- You can secure Enterprise and 3rd party clouds like AWS from Security Center
Security Center is built into the Azure Panel so no additional access is required. If you select the Data Collection policy you can automatically push the monitoring agent. This agent is the same as the Operations Management Agent. When you setup Data Collection you can set the level of logging required.
Security Center comes with a policy engine that allows you to tune policies via subscription. For example you can define one policy posture for production and another for dev and test.
2 Ensure VMs are configured in a certain way
- You can see system update status, Antimalware protection (Azure has one that is built in for free), OS and web configuration assessment (e.g. IIS assessment against best practise confirms)
- It will allow you to fix vulnerabilities quickly
3 Encrypt disks and data
4 Control Network Traffic
5 Use NSGs add additional firewalls
6 Collect Security Data
- Analyze and search security logs from many sources
- Security Center allows you to integrate 3rd party products like Qualys scans as well for assessment for other applications and compliance issues. Security Center monitors IaaS VMs and some PaaS components like web apps.
Security Center provides a new dashboard for failed logon attempts on your VMs. The most common attack on cloud VMs are RPD brute force attacks. To avoid this you can use Just-in-Time access so that port 3389 is only open for a window of time from certain IPs. These are all audited and logged.
Another attack vector is Malware. Application whitelisting allows you to track for good behaviour vs. blocking the bad. Unfortunately it has been arduous to apply.
7 Block malware and unwanted applications
Security Center uses the adaptive algorithm to understand what applications are running to develop a set of whitelists. Once you are happy with lists you can move to enforcement.
8 Use advanced analytics to detect threats quickly.
Security Center looks at VMs and network activity and leverages Microsoft’s global threat intelligence to detect threats quickly. This leverages machine learning to understand what is normal activity statistically to identity abnormal behavior.
9 Quickly assess the scopes and impact of the attack
This is a new feature that graphically displays all the related components that were involved in an attack.
10 Automate threat response
Azure uses Logic Apps to automate responses which allows you to trigger workflows form and alert to enable conditional actions. In addition there is a new malicious map that identifies known malicious IPs by region with related threat intelligence.
The basic Policy for Security Center is free so there is no reason to not have more visibility on what is vulnerable in your environment.
For more information check out
Wednesday, September 27, 2017
- Intune for Enterprises
- Intune for Education
- Intune for SMBs (In Public Preview)
Some other features that have been built into conditional access is to restrict access to services based on the location of the user. Microsoft has also enhanced Mobile Threat Protection and extended Geo fencing (In tech preview).
Microsoft Ignite 2017: Azure Security and Management for hybrid environments Jeremy Winter Director, Security and Management
Jeremy Winter is going to do a deep dive on Azure’s Management and Security announcements. Everyone is a different state when it comes to cloud. This digital transformation is having a pretty big ground level impact. Software is everywhere and it is changing the way we think about our business.
Digital transformation requires alignment across teams. This includes Developers, Operational teams and the notion of a Custodian who looks after all the components. It requires cross-team collaboration, automated execution, proactive governance and cost transparency. This is not a perfect science, it is a journey Microsoft is learning, tweaking and adjusting as they go. Operations is going to become more software based as we start to integrate scripting and programmatic languages.
Microsoft’s bet is that management and security should be part of the platform. Microsoft doesn't think you should have to take enterprise tooling and bring it with you to the cloud. You should expect this management and security to be a native capability. Microsoft thinks about this according to a full stack of cloud capabilities
Security is integral, Protect is really about Site Recovery, Monitoring is about visibility to what is going on, Configuration is about automating what you are doing. Microsoft believes they have a good baseline for all these components in Azure and is now focused on Governance.
Many frameworks try to put a layer of abstraction between the developer and the platform. Microsoft’s strategy is different. They want to allow the activity to go direct to the platform but to protect it via policy. This is a different approach and is something that Microsoft is piloting with Microsoft IT. Intercontinental Hotels is used as a reference case for digital transformation.
Strategies for successful management in the cloud:
1. Where to start “Secure and Well-managed”
You need to secure your cloud resources through Azure Security Center, Protect your data via backup and replication and Monitor your cloud health. Ignite announced PowerShell in Azure Cloud Shell as well as Monitoring and Management built right in to the Azure portals.
Jeremy shows an Azure Monitoring a Linux VM. The Linux VM has management native on the Azure Panel. In the Panel you can see the inventory of what is inside the VM. You no longer have to remote in. It is done programmatically.
In the demo the internal version of Java is shown on the VM and we can not look at Change tracking to determine why the version of Java appears to have changed. This is important from an audit perspective as you have to be able to identity changes. You can see the complete activity log of the guest.
Also new is update management. You can see all the missing updates on individual or multiple computers. You can go ahead and schedule patching by defining your own maintenance window. In the future Microsoft will add pre and post activities. You also have the ability to use Azure Management for non-Azure based VMs.
For disaster recovery you are able to replicate from one region to another. You can use this from enterprise to Azure now, but now also region-to-region. For backup it is now exposed as a property of the virtual machine which you simply enable and assign a backup policy to.
With the new Azure Cloud Shell you have the ability to run PowerShell right inline. There are 3449 commands that have been ported over so far.
2. Governance for the cloud
azure.com/policy is in tech preview. Jeremy switches to the demo for Azure policies. You now have the ability to create policy and see non-compliant and compliant VMs. The example shows a sample policy to ensure that all VMs do not have public IPs. With the policy you are able to quickly audit the environment. You can also enforce these policies. It is based on JSON so you can edit the policies directly. Other use cases include things like auditing for unencrypted VMs.
3. Security Management and threat detection
Microsoft is providing unified visibility, adaptive threat and detection and intelligent response. Azure Security Center is fully hybrid, you can apply it to enterprise and Azure workloads. Jeremy switches to Azure Security Center which provides an overview of your entire environments security posture.
Using the portal you can scan and get Microsoft's Security recommendations. Within the portal you can use ‘Just in Time Access’. This allows the developer to request that a port be open but it is enabled for a window of time. Security Center Analysis allows you to whitelist ports and audit what has changed.
Microsoft can track and group Alerts through Security Center. Now you have a new option through continuous investigation to see visually what security incident has been logged. It takes the tracking and pushes it through the analytics engine. It allows you to visually follow the investigation thread to determine the root cause of the security exploit. Azure Log Analytics is the engine that drives these tool sets.
Azure Log Analytics now has an Advanced Analytics component that provides an advanced query language that you can leverage across the entire environment. It will be fully deployed across Azure by December.
4. Get integrated analytics and monitoring
For this you need to start from the app using application insights, bring in network visibility as well as infrastructure perspective. There is a new version of Application Insights. Jeremy shows Azure Monitor which was launched last year at Ignite. Azure Monitor is the place for end-to-end monitoring of your Azure datacenters.
The demo shows the ability to drill in on VM performance and leverage machine learning to pin point the deviations. The demo shows slow response time for the ‘contoso demo site’. It shows that everything is working but slowly. You can see the dependency view of every process on the machine and everything it is talking to. Quickly you are able to see that the problem has nothing to do the website but is actually a problem downstream.
Microsoft is able to do this because they have a team of data scientists baking in the analytics directly into the Azure platform through the Microsoft Smarts Team.
5. Migrate Workloads to the Cloud.
Microsoft announced a new capability for Azure Migration. You can now discover applications on your virtual machines and group them. From directly within the Azure portal you can determine which apps are ready for migration to Azure. In addition it will recommend the types of Migration tools that you can use to complete the migrations. This is in limited preview.
Tuesday, September 26, 2017
Microsoft’s strategy is about addressing storage costs and management complexity through the use of:
- world class infrastructure on commodity hardware
- Finding smarter ways to store data
- using storage active-tiering
- offload the complexity to Microsoft
How does Storage spaces work? You attach the storage directly to your nodes and then connect the nodes together in a cluster. You then create a storage volume across the the cluster.
When you create the volume you select the resiliency. You have the following three options:
- Fast but uses allot of storage
- Slower but uses less storage
- Mirror-accelerated parity
- This allows you to create volumes that use both mirroring and parity. This is fast but conserves space as well
Storage Spaces Direct is a great option for running File Servers as VMs. This allows you to isolate file server VMs by use VMs running on a storage spaces direct volume. In Windows 2016 you also have the ability to introduce Storage QoS on Storage Spaces to deal with noisy neighbors. It allows you to predefine QoS storage policies to prioritize storage performance for some workloads.
You also have the ability to Dedup. Dedup works by taking unique chunks to a dedup chunk store and replacing the original block with a reference to the unique block. Ideal use cases for Microsoft Dedup is general purpose file servers, VDI and backup.
You may apply Dedup to a SQL Server and Hyper-V but it depends on how much demand there is on the system. High Random I/O workloads are not ideal for Dedup. Dedup is only supported on NTFS on Windows Server 2016. It will support ReFS on Windows Server 1709 which is the next release.
Microsoft has introduced Azure File Sync. With Azure File Sync you are able to centralize your File Services. You can use your on-premises File services to cache files for faster local performance. It is a true file share so it is services use standard SMB and NFS.
Shifting your focus from on-premises file services allows you to take advantage of cloud based backup and DR. Azure File Sync has a fast DR recovery option to get you back up and running in minutes.
Azure File Sync requires Windows 2012 or Windows 2016 and enables you to install a service that tracks file usage. It also teams the server with an Azure File Share. Files that are not touched over time are migrated to Azure File services.
To recover you simply deploy a clean server and reinstall the service. The namespace is recovered right away so the service is available quickly. When the users request the files a priority restore is performed from Azure based storage. Azure File Sync allows your branch file server to have a fixed storage profile as older files move to the cloud.
With this technology you can introduce follow the sun scenarios were work on one file server is synced through Azure File Share to a different region so it is available.
On the roadmap is cloud-to-cloud sync which allows the Azure File Shares to sync through the Azure backbone to different regions. When you have cloud-to-cloud sync the moment the branch server cannot connect to its primary Azure File Share it will go to the next closest.
Azure File Sync is now publically available in five “5” Azure Regions.
Monday, September 25, 2017
As part of the session we are supplied an MXChip IoT Developer’s Kit. This provides a physical IoT device enabling us to mock up IoT scenario’s. The device we are leveraging is made by Arduino. The device comes with a myriad of sensors and interfaces including, temperature and a small display screen. The device retails for approx. 30 – 40$ USD and is a great way to get started learning IoT.
When considering IoT one needs to not just connect the device but understand what you want to achieve with the data. Understanding what you want to do with the data allows you to define the backend components for storage, analytics or both. For example is you are ingesting data that you want to analyze you may leverage Azure Stream Analytics. For less complex scenarios you may define an App Service and use functions.
Microsoft’s preferred development suite is Visual Studio Code. Visual Studio Code includes extensions for Arduino. The process to get up and running is a little involved but there are lots of samples to get you started at https://aka.ms/azureiotgetstarted.
One of the more innovative ways to use the device was demonstrated in the session. The speaker created “The Microsoft Cognitive Bot” by leveraging the Arduino physical sensor and leveraging “LUIS” in the Azure Cloud. LUIS is the Language Understanding and Intelligence Service that is the underlying technology that Microsoft Cortana is built on. The speaker talks to the MXChip sensor asking details about the weather and the conference with LUIS responding.
The session starts with an introduction of what A Basic framework of an IoT Solution looks like as shown in the picture below. On the left of the frame are the devices. Devices can connect to the Azure IoT hub directly provided the traffic is secure and they can reach the internet. For devices that either do not connect directly to the internet or cannot communicated securely you can introduce a Field gateway.
Field Gateways can be used for other items as well such as translating data. In cases where you need high responsiveness, you also may analyze the data on a Field Gateway so that there is less latency between the analysis and the response. Often when dealing with IoT there is both Hot and Cold data streams. Hot being defined as data that requires less latency between the analysis and response vs. cold which may not have a time sensitivity.
An ingestion point requires a D2C endpoint or “Device-to-Cloud”. In addition to D2C you have the other traffic flow which is a C2D endpoint or Cloud-to-Device. C2D traffic tends to be queued. There are two other types of endpoints that you can define; a Method endpoint which is instantaneous and is dependent on the device being connected. The other type is a Twin endpoint. With a Twin endpoint’s you can inject a property; wait for the device to report current state and then synchronize it with your desired state.
We then had an opportunity to sit down with IoT experts like Brett from Microsoft. Okay I know he does not look happy in this picture but we had a really great session. We developed an architecture to look at long term traffic patterns as well as analyze abnormal speeding patterns in real-time for more responsive actions.. “Sorry Speeders ; – )”.
The session turned pretty hands on and we had to get our devices communicating with an Azure based IoT hub we created. We then setup communications back to the device to review both ingress and egress traffic. In addition we configured Azure table stores and integrated some cool visualization using Power BI. Okay got to admit I totally geeked out when I first configured a functional IoT service and then started to do some analysis. It is easy to see how IoT will fundamentally change our abilities in the very near future. Great first day as Microsoft Ignite.
Thursday, September 21, 2017
Friday, September 8, 2017
Had a great discussion with Scott Davis the EVP of Product Engineering and CTO of Embotics at VMworld 2017. Scott was kind enough to share some of the future looking innovations they are working hard on at Embotics.
"Clearly the industry is moving from having virtualization or IaaS centric cross-cloud management platforms to more of an application-centric container and microservices focus. We really see customers getting serious about running containers because of the application portability, microservices synergy, development flexibility and seamless production scale-out that is possible. At the same time they are reducing the operational overhead and interacting more programmatically with the environment.
When we look at the work VMware is doing with Pivotal Container Service we believe this is the right direction but we think that the key is really enhanced automation for DevOps. One of the challenges that was pointed out, is that while customers are successfully deploying Kubernetes systems for their container development, production operation can be a struggle. Often the environment gets locked in stasis because the IT team is wary of upgrades in a live environment.
At Embotics we are all about automation. With our vCommander product we have have a lot of intelligence that we can use to build a sophisticated level of iterative automation. So let's take that challenge and let's think about what would be needed to execute a low risk DevOps migration. You would probably want to deploy the new Kubernetes version and test it against your existing set of containers. This should be sandboxed to eliminate the risk to production, validated and then the upgrade should be fully automated.”
Scott proceeds to demonstrate a beta version of Embotics Cloud Management Platform 'CMP 2.0" automating these exact set of steps across a Kubernetes environment and then rolling the changes forward to update the production environment.
“I think fundamentally we can deliver true DevOps, speeding up release cycles, delivering higher quality and providing a better end user experience. In addition we can automatically pull source code out of platforms like Jenkins, spin up instances, regression test and validate. The test instances that are successful can be vaporized, while preserving the ones that are not so that the issues can be remediated.
We are rolling this out In a set of continuous software releases to our product so that as customers are integrating Containers, the Embotics 'CMP' is extended to meet these new use-cases.
We realize as we collect a number of data points spanning user preference, IT specified compliance rules and vCommander environment knowledge across both enterprise and hyper-scale targets like Azure and AWS that we can assist our customers with intelligent placement suggestions.”
Scott switches to a demo in which the recommended cloud target is ranked by the number of stars in the beta interface.
“We are building it in a way that allows the customer to adjust the parameters and their relative importance so if PCI compliance is more important they can adjust a slider in the interface and our ranking system adjusts to the new priority. Things like cost, compliance can be set to be both relative or mandatory to tune the intelligent placement according to what the customer views as important."
Clearly Embotics is making some innovative moves to incorporate a lot of flexibility in their CMP platform. Looking forward to seeing these releases in the product line with cross cloud intelligence for containers and placement.
It is a pleasure to be here with Zerto's, Don Wales the Vice President of Global Cloud Sales at VMworld 2017. Don this is a big show for Zerto, can you tell me about some of the announcements you are showcasing here today?
"Sure Paul, we are extremely excited about our Zerto 5.5 release. With this release we have introduced an exciting number of new capabilities. You know we have many customers looking at Public and Hybrid Cloud Strategies and at Zerto we want them to be able to leverage these new destinations but do so in a way that is simple and straightforward. Our major announcements here are our support for Fail In and Out of Azure, Increase AWS capabilities, Streamlined and Automated Upgradability, significant API enhancements and BI analytics. All these are designed for a much better end-user experience.
One piece of feedback that we are proud of is when customers tell us that Zerto does exactly what they need it to do without a heavy engineering cost for installation. You know Paul when you think about taking DR to a Cloud Platform like Azure it can be very complex. We have made it both simple and bi-directional. You can fail into and out of Azure with all the capabilities that our customers expect from Zerto like live failover, 30-day journal retention and Journal level file restore.
We also recognize that Azure is not the only cloud target our customers want to use. We have increased the recovery times to Amazon Web services. We have improved the performance and it our testing we have seen a 6x improvement in the recovery to AWS. Zerto has also extended out support to AWS regions in Canada, Ohio, London and Mumbai.
All this as well as continuing to enhance the capabilities that our traditional Cloud Service Providers need to make their customers experience simple yet powerful."
Don, with your increased number of supported Cloud targets and regions how do you ensure your customers have visibility on what's going on?
“Paul we have a SaaS product that allows our customers complete visibility on-premise and in public clouds called Zerto Analytics. It does historical reporting across all Zerto DR and recovery sites. It is a significant step forward in providing the kind of Business Intelligence that customers need as their environments grow and expand.”
Don these innovations are great, looks like Zerto is going to have a great show. Let me ask you when Don's not helping customers with their critical problems, what do you do you unwind?
“It’s all about the family Paul. There is nothing I like better than relaxing with the family at home, and being with my wife and twin daughters. One of our favorite things is to spend time at our beach house where our extended family gathers. It’s a great chance to relax and get ready for my next adventure.”
Many thanks for the time Don, it is great to see all innovations released here at VMworld 2017
Friday, September 1, 2017
I had the pleasure of spending a few moments with Crystal Harrison, Pivot3’s dynamic VP of product strategy “@MrsPivot3”.
Crystal, I know Pivot3 from the innovative work you have been doing in security and surveillance. How has the interest been from customers in the move to datacenter and cloud?
“You know with the next generation release of Acuity, the industry’s first priority-aware hyper converged infrastructure “HCI” the demand has been incredible. While we started with 80% of the product being applied to security use cases we are now seeing a distribution of approx. 60% applied to datacenter and cloud with 40% deriving from our security practice. This is not due to any lack of demand on the security side it is just the demand on our cloud and datacenter focus has taken off with Acuity.”
We are pushing the boundaries with our HCI offering as we are leveraging NVM Express “NVMe” to capitalize on the low latency and internal parallelism of flash-based storage devices. All this is wrapped in an intuitive management interface controlled by policy.”
How do you deal with tiering within the storage components of Acuity?
Initially the policies manage where the data or instance lives in the system. We have the ability to dynamically reallocate resources in real-time as needed. Say for example you have a critical business application that is starting to demand additional resources, we can recapture it from lower priority and policy assigned workloads on the fly. This protects your sensitive workloads and ensures they always have what they need.
How has the demand been from Cloud Service providers?
They love it. We have many flexible models including pay-by-the-drip metered and measured cost models. In addition the policy engine gives them the ability to set and charge for a set of performance based tiers for storage and compute. Iron Mountain is one of our big cloud provider customers. What is really unique is because we have lightweight management overhead and patented Erasure coding you can write to just about every terabyte that you buy which is important value to our customers and service providers.
Crystal, it really sounds like Pivot3 has built a high value innovative solution. Getting away from technology, what does Crystal do to relax when she is not helping customers adopt HCI?
My unwind is the gym. After a long day a good workout helps me reset for the next.
Crystal, it has been a pleasure, great to see Pivot3 having a great show here at VMworld 2017.
Thursday, August 31, 2017
Dr. Peter Weinstock is an Intensive Care Unit physician and Director of the Pediatric Simulator Program at Boston Children's Hospital/Harvard Medical School.
Peter wants to talk about game changers in medicine. Peter looks after critically ill children and is interested in any advance that helps his patients. Peter references a few game changes in medicine such as antibiotics. Antibiotics were discovered in the 1800s. With the discovery we were able to save patients that we could not before. Another game changer was anesthetic which allows us to deliver surgeries that were not possible before.
A game changer moves the bar on the outcomes for all patients. Peter’s innovation is Life-like rehearsal through medical simulation.
The problem in pediatrics is that the exception of medical emergencies do not happen often enough to perfect the treatment and approach to them. Medicine is also an apprentice program in which we are often practicing on the patients that we are treating.
In other high stakes industries simulation and practice are foundational. Take for example the airline industry. In the airline industry when they looked at bad outcomes it was often the lack of communication in a crisis. The medical industry is not immune to these freezing or lack of interaction with the whole team. Airline simulators are used to help the cockpit crew to practice interaction and approach to various emergencies.
So how do we take these methods to the medical industry? In Boston they have a full 3D simulator so the team can practice before the actual surgery. Through 3D printing and special effects typically found in the movie industry they are able to recreate surgery simulators using incredible authenticity.
Prior to this one of the real surgical practice techniques involved making an incision on an actual pepper and removing the seeds. By creating a simulation we are able to practice and drill by leveraging techniques common in other high risk industries in the medical field. Pictured below is Peter with one of the medical simulators, notice the realism.
We do not stop at simulation; we also look at execution. Adopting the team approach used in formula one pit crews for quick efficient focused effort and communication we drill the team.This enables our surgical team to reach a level of efficiency not previously possible. This is a game changer in the medical field.
Affectiva’s vision is that one day we will interact with technologies in the way we interact with people. In order to achieve this technologies must become emotionally aware. This is the potential for emotional AI enabling you to change behavior in positive ways. Raina mentions that today we have things like emoticons but that these are a poor way to communicate emotions. They are all very unnatural ways to interact. Even with voice AI, they tend to have allot of smarts but no heart.
Studies have shown that people rage against technologies like Siri because they are emotionally devoid. There is also a risk that interaction with emotionally devoid technology causes a lack of empathy in human beings.
Affectiva’s first foray was to use wearable glasses for autistic people to provide emotional feedback on human interactions. Autistic people struggle to read body queues. They are now partnering with another company to make this commercially available using google glass.
Raina’s demo shows the technology profiling facial expressions in real-time. They do this by using neural networking to feed 1000s of facial expressions to the AI so that it can recognize different emotions. They now have the largest network of facial recognition data. The core engine has been packaged as an sdk to allow developers the ability to add personalize experiences to there applications.
Some of the possible use cases are personalizing movie feeds based on emotional reaction. Another use case is for autonomous cars to recognize if the driver is tired or angry. They have also partnered with educational software companies to develop software that adapts based on the level of engagement of the student.
The careful use of this technology has been why Affectiva has created the Emotion AI Summit. The Summit will explore how Emotion AI can move us to deeper connections with technology, with businesses and with the people we care about. it takes place at the MIT Media Lab on September 13th.
Hugh Herr (@hughherr) takes the stage and mentions that prosthetics have not evolved a great deal over the decades and are passive with little innovation. Hugh Herr mentions that he lost both legs from frostbite iin a mountain climbing accident in 1982. During the postop, Hugh mentioned that he wanted to mountain climb and was told it would never happen by the doctor. The doctor was dead wrong. He did not understand that innovation and technology is not static but is transient and grows over time.
Hugh actually considered himself lucky as because he was a double amputee he could adjust his height by creating prosthetics that were taller. Hugh references the Biomechatronics limbs he is wearing on stage which have three computers and built in actuators. Hugh’s passion is running the Center for Extreme Bionics. Extreme Bionics is anything that is designed or implanted into the body that enhances human capability.
Hugh explains that when limbs are amputated surgeons fold over the muscle so there is no sensory feedback to the patient. Dr. Hugh and team have developed a new way of amputating. The new ways has surgeons create a little joint by ensuring there are two muscles working to expand and contract. With this new method the patient can ‘feel’ a phantom limb. By adding a simple controller you can track sensory movement that can be relayed to a bionic limb.
What they learned is that if you give the spinal cord enough information the body will intrinsically know how to move. But what about paralysis? The approach right now is to add a cumbersome exoskeleton to enable the ability to move. Work is being done to inject stems cell into severed spinal cord with the results being an incredible return of mobility.
Dr. Hugh and team are testing crystals embedded in muscles to relay information along with light emitters to control muscles. In this way they can build a virtual spinal column of sensors enabling mobility that was once considered impossible.
Hugh mentions that they are going to advance from their current foundational science in Biomechatronics to eliminate disabilities and augment physicality. It is important that we need to develop policies that govern the use of this technology so that it is used ethically.
Wednesday, August 30, 2017
Tuesday, August 29, 2017
AWS Native Services Integration with VMware Cloud on AWS with PaulBockelman @boxpaul & Haider Witwit
- VM Automation - provides cloud agnostic blueprint for deployment in any hyper-scale cloud provider.
- VMware NSX Cloud - AWS instances are shown through the interface which provides a consistent view of all the networking
- Wavefront - Wavefront measures the KPIs for the application and infrastructure, thinks VMware Log Insight in the cloud.
- Workspace ONE Mobile Flows - Mobile Flows allows you to automate business processes by using automated workflows