The value of the hyperscaler + hypervisor model

Public cloud offerings for “hyperscalers” such as AWS EC2, Microsoft Azure & Google GCP provide a lot of value when it comes to be able to stand up and run virtual workloads in a timely manner and provide various capabilities to create globally resilient solutions.

All of these offerings also boast a varying/wide range of native services which can compliment or replace services running in traditional virtual machines.

As I’ve previously stated in a post from August 2022, Direct to Cloud Value – Part 1, the hyperscalers have two major advantages customers can benefit from:

  1. A Well understood architecture
  2. Global availability

Designing, deploying and maintaining “on-premises” infrastructure on the other hand is often far less attractive from a time to value perspective and requires significant design efforts by highly qualified, experienced (and paid) individuals in order to get anywhere close to the scalability, reliability and functionality of the hyperscalers.

On-premises infrastructure may not be cost effective for smaller customers/environments who don’t have the quantity of workloads/data to make it cost effective, so “native” public cloud solutions at a high level are often a great choice for customers.

The problem for many customers is they’re established businesses with a wide range of applications from numerous vendors, many of which are not easy to simply migrate to a public cloud provider.

Workload refactoring is often a time consuming and complex task which is not always able to be achieved in a timely manner, and in many cases not at all.

Customers also rarely have the luxury of starting from and/or just building a greenfield environment due to the overall cost and/or the requirement to get a return on investment (ROI) from existing infrastructure.

Customers often have the requirement to burst during peak periods which isn’t something easily achievable on-premises. Customers often need to significantly oversize their on-premises infrastructure just to be able to support end of month, quarter or peak periods such as “Black Friday” for retailers.

This oversizing does help mitigate risks and deliver business outcomes, but it comes at a high cost (CAPEX).

Enter the “Hyperscaler + Hypervisor” model.

The hyperscaler + hypervisor model is where the hyperscaler (AWS/Azure/Goolgle) provides bare metal servers (a.k.a instances) where a hypervisor (in the above example, VMware ESXi) is running along with Virtual SAN (a.k.a “vSAN”) to provide the entire VMware technology stack to run Virtual Machines (VMs).

Nutanix has a similar offering called “Nutanix Cloud Clusters” or “NC2” using their own hypervisor “AHV”.

Both the VMware & Nutanix offerings gives the same look/feel to their customers as they have today on-premises.

The advantages of the hyperscaler + hypervisor model are enormous from both a business and technical perspective, the following are just a few examples.

  • Ease of Migration

A migration of VMware based workloads from an existing on-premises environment can be achieved using a variety of methods including VMware native tools such HCX as well as third party tools from backup vendors such as Commvault without having to refactor workloads.

This is achieved without the cost/complexity and delay of refactoring workloads.

  • Consistent look and feel

The Hyperscaler + hypervisor options provide customers access to the same management tools they’re used to on-premises meaning there is minimal adjustment required for I.T teams.

  • Built-in Cloud exit strategy / No Cloud Vendor “Lock in”

The hypervisor layer allows customers to quickly move from one hyperscaler to another again without refactoring, giving customers real bargaining power when it comes to negotiating commercial arrangements.

It also enables a move off public cloud back to on-premises.

  • Faster Time to value

The ability to stand up net new environments typically within a few hours gives customers the ability to respond to unexpected situations as well as new projects without the time/complexity of procurement and designing/implementing new environments from the ground up.

One very important value here is the ability to respond to critical situations such as ransomware by standing up an entirely isolated net new infrastructure to restore known good data. This is virtually impossible to do on-premises.

  • Lower Risk

In the event of a significant commercial/security/technical issue, a hyperscaler + hypervisor environment can be scaled up, migrated to a new environment/provider or isolated.

This model also mitigates against the delays caused by under-sizing or failure scenarios where new hardware needs to be added as this can occur typically within an hour or so as opposed to days/weeks/months.

As in the next example, workloads can simply be “lifted and shifted” minimising the number of changes/risks involved with a public cloud migration.

In the event of hardware failures, new hardware can be added back to the environment/s straight away without waiting for replacement hardware to be shipped/arrive and be installed. This greatly minimises the chance of double/subsequent failures causing an impact to the environment.

In the case of a disaster such a region failure, a new region can be scaled up to restore production whereas standing/scaling up a new on-prem environment is unlikely to occur in a timely manner.

  • Avoiding the need to “re-factor” workloads

Simply lifting and shifting workloads “as-is” on the same underlying hypervisor ensures the migration can occur with as few dependancies (and risks) as possible.

  • Provides excellent performance

The hardware provided by these offerings varies but often are all NVMe storage with latest or close to latest generation CPU/Memory, ensuring customers are not stuck with older generation hardware.

Having all workloads share a pool of NVMe storage also avoids the issue where some instances (VMs) are assigned to a lower tier of storage due to commercial cost constraints which can have significant downstream effects on other workloads/applications.

The all NVMe option in hyperscalers + hypervisor solutions becomes cost effective due to the economies of scale and elimination of “Cloud waste” which I will discuss next.

In many cases customers will be moving from a multiple year old hardware & storage solutions, simply having an all NVMe storage layer can reduce latency and subsequently make more efficient use of CPU/Memory often resulting in significant performance improvements let alone newer generation CPUs.

  • Economies of scale

In many cases, purchasing on a per instance (VM) basis may be attractive in the beginning, but when you reach a certain level of workloads, it makes more sense to buy in bulk (i.e.: A bare metal instance) and run the workloads on top of a hypervisor.

This gives the customer the benefit of the hypervisors ability to efficiently and effectively oversubscribe CPU and with a hyper-converged (HCI) storage layer (Virtual SAN a.k.a vSAN or Nutanix AOS) customers benefit from the native data reduction capabilities such as Compression, Deduplication and Erasure Coding.

  • Avoids native cloud instance constraints a.k.a “Cloud waste”

Virtual Machine “right-sizing” is to this day one of the most under-rated tasks but this can provide not only lower cost, but significant performance improvements for VMs. Cloud Waste occurs when workloads are forced into pre-defined instance sizes where small amounts of resources such as vCPUs or vRAM are assigned to the VM, but not required/use.

When we have the hypervisor layer, instance sizes can be customised to the exact requirements and eliminate cloud waste which I’ve personally observed in many customer environments to be in the range of 20-40%.

Credit: Steve Kaplan for coining the term “Cloud Waste”.

  • Increased Business Continuity / Disaster Recovery options

The cost/complexity involved with building business continuity and disaster recovery (BC/DR) solutions often lead to customers having to accept and try to mitigate significant risks to their businesses.

The hyperscaler + hypervisor model provides a number of options to have very cost effective BC/DR solutions including across multiple providers to mitigate against large global provider outages.

  • An OPEX commercial model

The ability to commit to a monthly minimum spend to get the most attractive rates while having the flexibility to burst when required (albeit at a less attractive price) means customers don’t have to try and fund large CAPEX projects and have the ability to scale in a “just in time” fashion.

Cost

This sounds to good to be true, what about cost?

On face value, these offerings can appear expensive compared to on-premises equivalents, but from the numerous assessments I’ve conducted I am confident the true cost is closer to or even cheaper than on-premises especially when a proper Total Cost of Ownership (TCO) is performed.

Compared with “native cloud” i.e.: Running workloads without the hypervisor layer, the hyperscaler + hypervisor solution will typically save customers 20-40% while providing equal or better performance and resiliency.

One other area which can make costs higher than necessary is a lack of optimisation with the workloads. I highly recommend for both on-premises and hyperscaler models that customers engage an experienced architect to review their environment thoroughly.

The performance benefits of a right sizing exercise are typically be significant AND it saves valuable IT resources (CPU/RAM). It also means less hardware is required to achieve the same or even a better outcome and therefore lowering costs.

Summary

The hyperscaler + hypervisor model has many advantages both commercially and technically and with the ease of setup, migration to and scaling in public cloud, I expect this model to become extremely popular.

I would strongly recommend anyone looking at replacing their on premises infrastructure in the near future do a thorough assessment of these offerings against their business goals.

End-2-End Enterprise Architecture (@E2EEA) has multiple highly experienced and certified staff at the highest level with both VMware (VCDX) and Nutanix (NPX) technologies and can provide expert level services to help you assess the hyperscaler + hypervisor options as well as design and deliver the solution.

E2EEA can be reached at sales@e2eea.com

Google Wifi Review – 3 Wi-Fi Point Solution

Since moving into a new place earlier this year, I’ve been struggling to get consistent Wi-fi signal/performance especially in the master bedroom. The master bedroom is the furthest room from my home office where I was running my TP-Link Archer D7 (AC1750) Dual-Band Wireless router.

After spending some time playing around trying to get better signal, I purchased the TP-Link AC750 Wi-Fi Range Extender and plugged it in in various positions between the master bedroom and the home office.

I eventually settled on the one location where the Range Extender was reporting maximum signal which was around 7m or 22ft from my master bedroom where I have a Samsung 75″ TV, Apple TV, Nintendo Switch, an iPad and two phones, one iPhone 8 and one Samsung Galaxy S9.

Performance was still inconsistent and I ultimately placed the TP-Link router right in the middle of the apartment which I would say helped a little bit, but ultimately did not solve the problem.

Note: My Sonos wireless speakers are not supported when using Range Extenders which is a real design flaw on Sonos’ part and a pain for customers. Ultimately I’m less than impressed with Sonos so I’ve got their Sound Bar and two wireless speakers and a sub sitting around doing nothing.

One of the many reasons for the Wi-Fi performance issues is likely to be the all to common scenario these days of being surrounded by a ton of Access Points. The below is what I see on my Macbook  when looking for networks, so in my case, Interference is likely a significant factor.

WiFiNoise

 

But long story short, I continued to get dropouts and inconsistent speeds so I bit the bullet and purchased the Google Wi-fi 3 pack (shown below).

20180626_160148.jpg

First impressions?

It’s nice and small, and uses USB-C for power so the cord is also small. You can see the device below beside my oversized Nutanix coffee mug for scale.

20180626_162629.jpg

It’s not a modem, so you’re stuck having multiple devices which is a bit annoying but not the end of the world.

Initial setup was a breeze, step by step instructions after downloading the Google Wifi app, the first device was detected and then verified by scanning a QV code on the base of the device which was cool but also very easy and saved manually entering numbers which saves time and avoids fat-fingering errors.

Now onto the exciting part, getting the “mesh” network setup and tested.

Once you run through the wizard, the app shows you a review of your network including the Wi-Fi name, Password and the Wi-fi points you have and their configuration, in my case, 1 “Primary” and two “Mesh” Wi-Fi points as shown below.

MeshSummary

The app has a cool “Network Check” functionality in the shortcuts menu (shown below).

Screenshot_20180626-164508_Google Wifi.jpg

The network check allows you to test the Internet speed, the connection quality between the access points “mesh” and the one I have found quite useful is the Test Wi-fi to all wireless devices currently connected to the network.

You can run each test individually or start all three as shown below.

Screenshot_20180626-164658_Google Wifi.jpg

Testing the internet speed is a quick and easy way to see how fast your connection is and saves downloading and using another app on your phone which is handy.

Below is how the test results are displayed, and for Australian internet, this is a pretty good result although it would be considered poor in many parts of the world.

Screenshot_20180626-154804_Google Wifi.jpg

Next up we have the “Test Mesh” option which is pretty important so kudos to Google for ensuring this was part of the app as it will avoid not technical people having to bug their I.T friends for help. At this stage I can hear all the I.T professionals all around the world rejoice!

Screenshot_20180626-154811_Google Wifi.jpg

The “Mesh test” is pretty quick and gives you a clear result as shown below.

Screenshot_20180626-154002_Google Wifi.jpg

This was my first “Mesh test” and while the result is not bad, I relocated the Wi-Fi points as it suggested and re-ran the test.

Screenshot_20180626-154845_Google Wifi.jpg

As we can see the result is now “Great” with full green bars which I have to admit I was very happy to see considering how annoying Wi-Fi had been over the past few months.

Next up, the Wi-Fi test for all connected devices. Prior to running the test I went around the apartment and turned on the three TVs, 2 Apple TVs, iPad, Samsung Galaxy Tablet, I also made sure all phones were on Wi-Fi as well as my laptop. 12 Devices in all.

As the test is running it appears to confirm if a device is Idle or not, and then proceeds to drive traffic to it. A couple of things I really like is that it clearly displays what device/s are connected to what Wi-fi point and the speed it was able to achieve.

WifiDevicesTest

One completed the app gives you a summary of the number of devices tested and their network performance as shown below.

12DevicesTested

Back to the main screen of the app, we can see a summary of the network telling us the access point is online and has 12 devices connected as confirms the internet is online.

NetworkSummary

If we go into “Devices” we can also see per device upload and download statistics so it’s quick and easy to identify if one or more devices are hogging the bandwidth.

DevicesRealTimeBandwidth

While I haven’t used the next to features, the app does allow you to setup a Guest Wi-fi network which is handy if you want to keep your devices isolated from guests and/or not give out your password because it matches your internet banking one.. haha!

Screenshot_20180626-161535_Google Wifi.jpg

Google also allows you to “pause” the internet for specific devices, such as your teenage child/ren and pause it on a set schedule if you choose which I think is a good addition for a home network.

Screenshot_20180626-161540_Google Wifi.jpg

Performance when streaming multiple Ultra HD 4k shows on Netflix?

Moving on from the ease of setup and cool app functionality, let’s test how the network performs with 3 TVs streaming Ultra HD 4k (Netflix), my Samsung Tablet streaming YouTube Premium (4k HD) and my laptop streaming HD video (UFC Fight Pass).

Streaming4KHDon3TVsPlusLaptopandTablet

Above we can see some of the per device stats and I am pleased to report I am yet to observe any of the dropouts or buffering which were common on the previous setup.

Summary:

If you have a large house or apartment, and you’re having trouble with Wi-Fi consistency and dropouts, I would recommend the Google Wi-Fi solution for several reasons.

  • Price? $389 AUD or approx $287USD based on the Exchange rate at the time of writing.

Price wise, I think it’s pretty reasonable. If you consider it’s 3 Wi-Fi points, that’s $129 AUD each which isn’t “cheap” but it’s also not expensive. Comparable high end consumer grade Wireless routers are in the $200-300 range.

Leading onto my next point, Because the Google Wi-Fi is scalable, and therefore somewhat future proofed, I believe the price is justified.

  • It’s a scalable Wi-Fi solution 

You can start with one Wi-Fi point and scale out from there. This is important so you don’t need to buy 3 up front, just start with one (or two in a large house) and scale as required after performing Mesh and Device Wi-Fi testing to see how the Mesh is performing.

  • Setup is easy and the app helps you optimise the position of the Wi-Fi points

This is really cool, especially for non-technical people who may not understand how Wireless access points work and the best place to position them etc. It’s easy to run a few tests and re-position, re-test and in my case, get to a scenario where I have “Great” signal/connectivity between all 3 Wi-Fi points.

  • The Google Wi-Fi app has lots of useful features for everyday use

The ability to quickly troubleshoot if required using device, mesh and internet speed testing is great, again especially for non tech savvy folk.

  • Wi-Fi Range/Performance

The difference in Wi-Fi range and performance in my apartment is night and day compared to my previous setup even with the Wi-Fi Range Extender. Performance is now consistent despite the fact I am in a building with a lot of Wi-Fi access points within strong/medium range of the Google Wi-Fi mesh.

Rating:

As for a rating, I’m giving the Google Wi-Fi solution a 9 out of 10.