Environment Infrastructure Design

This chapter is one of a series that make up the VMware Workspace ONE and VMware Horizon Reference Architecture, a framework that provides guidance on the architecture, design considerations, and deployment of Workspace ONE and Horizon solutions. This chapter provides information about the design of the supporting environmental infrastructure.

Introduction

Several environment resources are required to support a VMware Workspace ONE® and VMware Horizon® deployment. In most cases these will already exist. It is important to ensure that minimum version requirements are met and that any specific configuration for Workspace ONE and Horizon is followed. For any supporting infrastructure component that Workspace ONE or Horizon depends on, that component must be designed to be scalable and highly available. Some key items are especially important when the environment is used for a multi-site deployment.

Environment Design

As might be expected, several environment resources might be required to support a Workspace ONE and Horizon deployment, including Active Directory, DNS, DHCP, security certificates, databases, and load balancers. For any external component that Workspace ONE or Horizon depends on, the component must be designed to be scalable and highly available, as described in this section.

The following components should not be treated as an exhaustive list.

Active Directory

Workspace ONE and VMware Horizon require an Active Directory domain structure for user authentication and management. Standard best practices for an Active Directory deployment must be followed to ensure that it is highly available.

Because cross-site traffic should be avoided wherever possible, configure AD sites and services so that each subnet used for desktops and services is associated with the correct site. This guarantees that lookup requests, DNS name resolution, and general use of AD are kept within a site where possible. This is especially important in terms of Microsoft Distributed File System Namespace (DFS-N) to control which specific file server users get referred to.

Table 1: Implementation Strategy for Active Directory Domain Controllers

Decision

Active Directory domain controllers will run in each location.

Core data center locations may have multiple domain controllers.

Justification

This provides domain services close to the consumption.

Multiple domain controllers ensure resilience and redundancy.

For VMware Horizon® specifics, see the Preparing Active Directory for details on supported versions and preparation steps. For Horizon Cloud Service on Microsoft Azure specifics, see Active Directory Domain Configurations, in the Getting Started with VMware Horizon Cloud Service on Microsoft Azure guide for details on supported Active Directory configurations and preparation steps.

Additionally, for Horizon usage, whether Horizon or Horizon Cloud, set up dedicated organizational units (OUs) for the machine accounts for virtual desktops and RDSH servers. Consider blocking inheritance on these OUs to stop any existing GPOs from having an undesired effect.

Group Policy

Group Policy objects (GPOs) can be used in a variety of ways to control and configure both VMware Horizon® Cloud Service on Microsoft Azure components and also standard Windows settings.

These policies are normally applied to the user or the computer Active Directory account, depending on where the objects are located in Active Directory. In a Horizon Cloud Service on Microsoft Azure environment, it is typical to set specific user policy settings for the specific Horizon Cloud Service session only when a user connects to it.

We also want to have user accounts processed separately from computer accounts with GPOs. This is where the loopback policy is widely used in any GPO that also needs to configure user settings. This is particularly important with Dynamic Environment Manager. Dynamic Environment Manager applies only user settings, so if the Dynamic Environment Manager GPOs are applied to computer objects, loopback processing must be enabled.

Group policies can also be associated at a site level. Refer to the Microsoft Web site for details.

DNS

The Domain Name System (DNS) is widely used in a Workspace ONE and a Horizon environment, from server components communication to clients and virtual desktops. Follow standard design principles for DNS, making it highly available. Additionally, ensure that:

  • Forward and reverse zones are working well.
  • Dynamic updates are enabled so that desktops register with DNS correctly.
  • Scavenging is enabled and tuned to cope with the rapid cloning and replacement of virtual desktops.

Table 2: Implementation Strategy for DNS

Decision

DNS were provided by using Active Directory-integrated DNS zones. The DNS service ran on select Windows Servers running the domain controller roles.

Justification

The environment already had DNS servers with Active Directory-integrated DNS zones

DHCP

In a Horizon environment, desktops and RDSH servers rely on DHCP to get IP addressing information. DHCP must be allowed on the VM networks designated for these virtual desktops and RDSH servers.

In Horizon multi-site deployments, the number of desktops a given site is serving usually changes when a failover occurs. Typically, the recommendation is to over-allocate a DHCP range to allow for seamlessly rebuilding pools and avoiding scenarios where IPs are not being released from the DHCP scope for whatever reason.

For example, take a scenario where 500 desktops are deployed in a single subnet in each site. This would normally require, at a minimum, a /23 subnet range for normal production. To ensure additional capacity in a failover scenario, a larger subnet, such as /21, might be used. The /21 subnet range would provide approximately 2,000 IP addresses, meeting the requirements for running all 1,000 desktops in either site during a failover while still leaving enough capacity if reservations are not released in a timely manner.

It is not a requirement that the total number of desktops be run in the same site during a failover, but this scenario was chosen to show the most extreme case of supporting the total number of desktops in each site during a failover.

There are other strategies for addressing this issue. For example, you can use multiple /24 networks across multiple desktop pools or dedicate a subnet size you are comfortable using for a particular desktop pool. The most important consideration is that there be enough IP address leases available.

The environment can be quite fluid, with instant-clone desktops being deleted at logout and recreated when a pool dips below the minimum number. For this reason, make sure the DHCP lease period is set to a relatively short period. The amount of time depends on the frequency of logouts and the lifetime of a clone.

Table 3: Implementation Strategy for DHCP

Decision

DHCP was available on the VM network for desktops and RDSH servers.

DHCP failover was implemented to ensure availability.

The lease duration of the scopes used was set to 4 hours.

Justification

DHCP is required for Horizon environments.

Virtual desktops can be short lived so a shorter lease period ensures that leases are released quicker.

A lease period of 4 hours is based on an average logout after 8 hours.

Microsoft Azure has a built-in DHCP configuration that is a part of every VNet configured in Microsoft Azure. For more information, see IP Configurations in the Microsoft Azure documentation.

Distributed File System 

File shares are critical in delivering a consistent user experience. They store various types of data used to configure or apply settings that contribute to a persistent-desktop experience.

The data can include the following types: 

  • IT configuration data, as specified in Dynamic Environment Manager
  • User settings and configuration data, which are collected by Dynamic Environment Manager
  • Windows mandatory profile 
  • User data (documents, and more) 
  • VMware ThinApp® packages 

The design requirement is to have no single point of failure within a site while replicating the above data types between the two data centers to ensure their availability in a site-failure scenario. This reference architecture uses Microsoft Distributed File System Namespace (DFS-N) with array-level replication.

Table 4: Implementation Strategy for Replicating Data to Multiple Sites

Decision

DFS-Replication (DFS-R) was used to replicate user data from server to server and optionally between sites.

Justification

Replication between servers provides local redundancy.

Replication between sites provides site redundancy.

DFS Namespace 

The namespace is the referred entry point to the distributed file system.

  • A single-entry point is enabled and active for profile-related shares to comply with the Microsoft support statements (for example, Dynamic Environment Manager user settings). 
  • Other entry points can be defined but disabled to stop user referrals to them. They can then be made active in a recovery scenario.
  • Multiple active entry points are possible for shares that contain data that is read-only for end users (for example, Dynamic Environment Manager IT configuration data, Windows mandatory profile, ThinApp packages).

Table 5: Implementation Strategy for Managing Entry Points to the File System

Decision

DFS-Namespace (DFS-N) was used.

Depending on the data type and user access required, either one or multiple referral entry points may be enabled.

Justification

DFS-N provides a common namespace to the multiple referral points of the user data that is replicated by DFS-R.

More detail on how DFS design applies to profile data can be found in Dynamic Environment Manager Architecture.

Certificate Authority

A Microsoft Enterprise Certificate Authority (CA) is often used for certificate-based authentication, SSO, and email protection. A certificate template is created within the Microsoft CA and is used by VMware Workspace ONE® UEM to sign certificate-signing requests (CSRs) that are issued to mobile devices through the Certificate Authority integration capabilities in Workspace ONE UEM and Active Directory Certificate Services.

The Microsoft CA can be used to create CSRs for VMware Unified Access Gateway, Workspace ONE Access, and any other externally facing components. The CSR is then signed by a well-known external CA to ensure that any device connecting to the environment has access to a valid root certificate.

Having a Microsoft Enterprise CA is a prerequisite for Horizon True SSO. A certificate template is created within the Microsoft CA and is used by True SSO to sign CSRs that are generated by the VM. These certificates are short-lived (approximately 1 hour) and are used solely for the purpose of single-signing a user in to a desktop through Workspace ONE Access without prompting for AD credentials.

Details on setting up a Microsoft CA can be found in the Setting Up True SSO section in Horizon Configuration.

Table 6: Implementation Strategy for the Certificate Authority Server

Decision

A Microsoft Enterprise CA was set up.

Justification

This can be used to support certificate authentication for Windows 10 devices and to support the Horizon True SSO capability.

Microsoft RDS Licensing

Applications published with Horizon use Microsoft RDSH servers as a shared server platform to host Windows applications. Microsoft RDSH servers require licensing through a Remote Desktop Licensing service. It is critical to ensure that the Remote Desktop Licensing service is highly available within each site and also redundant across sites.

Table 7: Implementation Strategy for Microsoft RDS Licensing

Decision

Multiple RDS Licensing servers were configured.

At least one was configured per site.

Justification

This provides licensing for Microsoft RDSH servers where used for delivering Horizon published applications.

Microsoft Key Management Service

To activate Windows (and Microsoft Office) licenses in a VDI environment, VMware recommends using Microsoft Key Management Service (KMS) with volume license keys. Because desktops are typically deleted at logout and are recreated frequently, it is important that this service be highly available. See the Microsoft documentation on how best to deploy volume activation. It is critical to ensure that the KMS service is highly available within each site and also redundant across sites. 

Table 8: Implementation Strategy for the KMS Service

Decision

Microsoft KMS was deployed in a highly available manner

Justification

This allows Horizon virtual desktops and RDSH servers to activate their Microsoft licenses.

Load Balancing

To remove a single point of failure from server components, we can deploy more than one instance of the component and use a load balancer. This not only provides redundancy but also allows the load and processing to be spread across multiple instances of the component. To ensure that the load balancer itself does not become a point of failure, most load balancers allow for setup of multiple nodes in an HA or active/passive configuration.

An existing load balancer can be used, or a new one such as the VMware NSX® Advanced Load Balancer (formerly Avi Vantage by Avi Networks) can be deployed. The NSX Advanced Load Balancer is a software-only ADC (analog-to-digital converter), which offers enterprise-grade load balancing, web application firewall, and other application services in multi-cloud environments (with or without NSX networking).

The NSX Advanced Load Balancer offers a distributed architecture with a separate control plane and data plane and uses a software-defined scale-out architecture that is 100 percent based on REST-APIs. The platform provides elastic autoscaling, built-in application and end-user analytics, and full life-cycle automation of load balancers.

The NSX Advanced Load Balancer can be deployed for load balancing Horizon deployed on-premises in vSphere or VMware Cloud on AWS. NSX Advanced Load Balancer also provides full-featured load balancing for Horizon for multiple deployment scenarios, including global server load balancing (GSLB) for multi-site deployments.

Figure 1: NSX Advanced Load Balancer and Horizon Logical Architecture

See VMware NSX Advanced Load Balancer (Avi Networks) for more detail and Load Balancing Horizon View using Avi Vantage for guidance on implementing the NSX Advanced Load Balancer with Horizon.

 

vSphere Design

VMware vSphere® is the foundation that hosts infrastructure and components. All editions of Horizon come bundled with vSphere for Desktops.

This section describes how components of vSphere were used in this reference architecture, including the design for vSAN and VMware NSX-T Data Center (NSX-T).

  • vSAN pools its server-attached HDDs and SSDs to create a distributed shared datastore that abstracts the storage hardware and provides hyper-converged storage optimized for VMs without the need for external SAN or NAS.
  • NSX-T provides network-based services such as security, network virtualization, routing, and switching in a single platform.

This lets us build in a hyper-converged hardware model based on a physical server as the building block. The server provides not only the compute and memory but also the storage in a modular fashion.

Figure 2: vSphere and vSAN High-Level Architecture

Horizon deployments benefit from granular, elastic storage capacity that scales without forklift upgrades. Instead of having to add an entire storage array when more desktops are needed, we can simply add more disks, flash storage, or another vSphere host.

Although this reference architecture utilizes the benefits of an all-flash vSAN, traditional storage (such as SAN or NAS) is of course also still supported.

vSphere

This document does not try to cover vSphere design and installation. That is well documented in other resources, including the VMware vSphere documentation. Best practices around vSphere configuration and vSAN networking should be followed.

For the vSphere clusters hosting management servers such as Horizon Connection Server, or Workspace ONE Access appliances, VMware vSphere® Storage DRS rules should be enabled to prevent the servers that perform identical operations from running on the same vSphere host. This prevents multiple VM failures if a host fails and these VMs exist on the failed physical vSphere host.

For this reference architecture, we recommend the following settings.

Table 9: vSphere Distributed Resource Scheduler Settings

vSphere DRS 

Setting 

vSphere DRS 

Turn on vSphere DRS 

DRS Automation 

Fully Automated 

Power Management 

Off 

Leave all other DRS settings set to the default.

vSAN

vSAN pools its server-attached HDDs and SSDs to create a distributed shared datastore that abstracts the storage hardware and provides hyper-converged storage optimized for VMs without the need for external SAN or NAS.

vSAN uses VM-centric storage policies to automate the storage service levels on a per-VM basis. Horizon integrates into this consumption model and automatically generates the required storage policies as pools are deployed onto a vSAN datastore.

For best-practice recommendations, see the VMware Horizon 7 on VMware vSAN Best Practices technical white paper.

For full design and architectural guidance for deploying vSAN see the various resources on https://storagehub.vmware.com/t/vmware-vsan/

Networking

At a high level, all network components are configured redundantly to operate in either active/passive mode or active/active mode where allowed, and the various traffic types are separated from each other. Quality of service is controlled with network IO control (NIOC) on the configured distributed virtual switch.

Figure 3: vSphere Networking

The configuration uses 10-GbE adapters to simplify the networking infrastructure and remedy other 1-GB drawbacks such as inadequate bandwidth and lower utilization. Even with these 10-GbE advantages, it is still necessary to ensure that traffic flows can access sufficient bandwidth.

NIOC addresses this requirement by enabling diverse workloads to coexist on a single networking pipe, thus taking full advantage of 10 GbE. NIOC revolves around resource pools similarly to those for CPU and memory. The vSphere administrator is given control to ensure predictable network performance when multiple traffic types contend for the same physical network resources.

Table 10: NIOC Configuration

Traffic Type

Shares

Shares Value

Reservation

Limit

Management traffic

Normal

50

0 Mbit/s

Unlimited

Virtual machine traffic

High

100

0 Mbit/s

Unlimited

vSAN traffic

Normal

50

0 Mbit/s

Unlimited

vMotion traffic

Normal

50

0 Mbit/s

Unlimited

Flow control is disabled for VMkernel interfaces tagged for vSAN (vmk2). vSAN networks can use teaming and failover policy to determine how traffic is distributed between physical adapters and how to reroute traffic in the event of adapter failure. NIC teaming is used mainly for high availability for vSAN. However, additional vSphere traffic types sharing the same team still leverage the aggregated bandwidth by distributing different types of traffic to different adapters within the team. Load-based teaming is used because network convergence on these switch ports can happen quickly after the failure due to the port entering the spanning-tree forwarding state immediately, bypassing the listening and learning states.

The following table shows the distributed switch and port group policies. 

Table 11: Distributed Switch Settings

Property 

Setting 

Default

Revised 

General 

Port Binding 

Static 

 

Policies: Security 

Promiscuous mode 

Reject 

 

MAC address changes 

Accept 

Reject 

Forged transmits 

Accept 

Reject 

Policies: Traffic Shaping 

Status 

Disabled 

 

Policies: Teaming and Failover 

Load balancing 

Route based on the originating virtual port ID 

Route based on physical NIC load 

Failover detection 

Caution Link Status only 

 

Notify switches 

Yes 

 

Policies: Resource Allocation 

Network I/O Control 

Disabled 

Enabled 

Advanced 

Maximum MTU 

1500 

9000 

A single vSphere distributed switch was created with two 10-Gb interfaces in a team. Five port groups isolate network traffic: 

  • Virtual machines 
  • VMware ESXi management network 
  • VMware vSphere® vMotion® 
  • iSCSI 1 and iSCSI 2

Note: Two iSCSI port groups are required in order to configure vmknic-based iSCSI multi-pathing.

Quality of service is enforced with network I/O control (NIOC) on the distributed virtual switch, guaranteeing a share of bandwidth to each type of traffic. A vmkernel interface (vmknic) is created on the ESXi management port group, vSphere vMotion port group, and on each iSCSI port group.

Both 10-Gb adapters are configured as active/active for the VM port group, ESXi management network port group, and the vSphere vMotion port group.

  • The iSCSI 1 port group is bound to a single 10-Gb adapter, with only storage traffic permitted on that adapter.
  • The iSCSI 2 port group is bound to the second 10-Gb network adapter, with only storage traffic permitted over that adapter.

For more information, see the vSphere Networking.

vSphere Infrastructure Design for Active-Active and Active-Passive Services

Standard vSphere design and installation should be followed to create a compute and storage platform to run Horizon resources. This section describes the design used for a Horizon multi-site deployment. For a vSAN stretched cluster, within a metro or campus network environment with low network latency between sites, see Horizon Active-Passive Service Using Stretched vSAN Cluster.

The active/active and the active/passive services for a multi-site setup are based on separate sets of vSphere clusters in each site. The storage used was all-flash arrays. Horizon Cloud Pod Architecture was used to give global entitlements across both sites. Details in this section are specific to the environments put in place to validate the designs in this guide. Different choices in some of the configurations, hardware, and components are to be expected.

The vSphere infrastructure is deployed identically in both data centers, following VMware best practices for pod and block design concepts. For further details, see Horizon Architecture.

Storage 

When designing storage for a Horizon environment, consideration should be given to the mixed workloads. There are management servers running in the management block and desktops and RDSH VMs running in the resource blocks. The storage selected should be able to 

  • Handle the overall I/O load and provide sufficient space.
  • Ensure performance for key components so that a noisy neighbor such as a desktop does not affect the performance of the environment.

Depending on the capability of the storage being used, it is generally recommended to separate these mixed workloads.

In the environment built to validate this reference architecture, the underlying physical storage, all-flash, was shared between the management block and the resource blocks running desktop and RDSH workloads. This was possible because all-flash storage can handle mixed workloads while still delivering great performance.

This approach can be seen in the following logical diagram.

Figure 4: All-Flash Storage Usage

Storage Replication 

For some components of the active/passive service, which uses two geographically dispersed sites, storage array data replication is necessary to provide complete business continuity.

If possible, use an asynchronous replication engine that can support bi-directional replication, which facilitates use of DR (disaster recovery) infrastructure for DR and production. VMware recommends using a replication engine that compares the last replicated snapshot to the new one and sends only incremental data between the two snapshots, thus reducing network traffic. Snapshots that can be deduplicated provide space efficiency.

Figure 5: Asynchronous Replication

VMware recommends replicating Dynamic Environment Manager data between sites. In the case of a site failure, a protected volume that is replicated to the DR site can be recovered promptly and presented. In the case of an active/active deployment, VMware Dynamic Environment Manager volumes are replicated in each direction.

For optimal performance, VMware recommends implementing the advanced vSphere configuration changes outlined in the following table. Leave all other HA settings set to the default.

Table 12: vSphere High Availability (HA) Settings

vSphere HA 

Setting 

vSphere HA 

Turn on vSphere HA 

Host Monitoring 

Enabled 

Host Hardware Monitoring – VM Component Protection: “Protect against Storage Connectivity Loss” 

Disabled (default) 

Virtual Machine Monitoring 

Disabled (default) 

 

NSX-T Data Center

VMware NSX-T Data Center (NSX-T) provides network-based services such as security, virtualization networking, routing, and switching in a single platform. These capabilities are delivered for the applications within a data center, regardless of the underlying physical network and without the need to modify the application.

The main use for NSX-T Data Center in a Horizon infrastructure include:

  • Context-aware micro-segmentation
  • Edge and Partner Services
  • Network Virtualization

Figure 6: Main NSX-T Data Center Use Cases for Horizon

As the following figure shows, NSX-T Data Center performs several security functions within a Horizon solution:

  • Protects EUC management infrastructure NSX-T Data Center secures inter-component communication among the management components of a Horizon infrastructure.
  • Protects desktop pool VM communication NSX-T Data Center can apply firewall policy to block or limit desktop to desktop communication, or desktop to enterprise application. Data Security Scanning (PII/PHI/PCI), 3rd party Security Services e.g. Agentless Antivirus, NGFW, IPS
  • Protect user and Enterprise Application NSX-T Data Center allows user-level identity-based micro-segmentation for the Horizon desktops and RDS Hosts used for published applications. This enables fine-grained access control and visibility for each desktop or published application based on the individual user.
  • Virtual desktops contain applications that allow users to connect to various enterprise applications inside the data center. NSX-T Data Center secures and limits access to applications inside the data center from each desktop.

Figure 7: NSX-T Data Center Security Use Cases for Horizon

Micro-segmentation

The concept of micro-segmentation takes network segmentation, typically done with physical devices such as routers, switches, and firewalls at the data center level, and applies the same services at the individual workload (or desktop) level, independent of network topology.

NSX-T Data Center and its Distributed Firewall feature are used to provide a network-least-privilege security model using micro-segmentation for traffic between workloads within the data center. NSX-T Data Center provides firewalling services, within the hypervisor kernel, where every virtual workload gets a stateful firewall at the virtual network card of the workload. This firewall provides the ability to apply extremely granular security policies to isolate and segment workloads regardless of and without changes to the underlying physical network infrastructure. 

Two foundational security needs must be met to provide a network-least-privilege security posture with NSX Data Center micro-segmentation. NSX-T Data Center uses a Distributed Firewall (DFW) and can use network virtualization to deliver the following requirements. 

Isolation

Isolation can be applied for compliance, software life-cycle management, or general containment of workloads. In a virtualized environment, NSX-T can provide isolation by using the DFW to limit which workloads can communicate with each other. In the case of Horizon, NSX-T Data Center can block desktop-to-desktop communications, which are not typically recommended, with one simple firewall rule, regardless of the underlying physical network topology.

Figure 8: Isolation Between Desktops

Another isolation scenario, not configured in this reference architecture, entails isolating the Horizon desktop and application pools. Collections of VDI and RDSH machines that are members of specific Horizon pools and NSX-T Data Center security groups can be isolated at the pool level rather than per machine. Also, identity-based firewalling can be incorporated into the configuration, which builds on the isolation setup by applying firewall rules that depend on users’ Active Directory group membership, for example.

Segmentation

Segmentation can be applied at the VM, application, infrastructure, or network level with NSX-T. Segmentation is accomplished either by segmenting the environment into logical tiers, each on its own separate subnet using virtual networking, or by keeping the infrastructure components on the same subnet and using the NSX-T Data Center DFW to provide segmentation between them. When NSX-T micro-segmentation is coupled with Horizon, NSX-T Data Center can provide logical trust boundaries around the Horizon infrastructure as well as provide segmentation for the infrastructure components, and for the desktop pools.

Figure 9: NSX-T Segmentation

Advanced Services Insertion (Optional)

Although we did not use this option for this reference architecture, it warrants some discussion. As one of its key functionalities, NSX-T Data Center provides stateful DFW services at layers 2–4 for micro-segmentation. Customers who require higher-level inspection for their applications can leverage one or more NSX-T Data Center extensible network frameworks, in this case NetX, in conjunction with third-party next-generation firewall (NGFW) vendors for integration and traffic redirection. Specific traffic types can be sent to the NGFW vendor for deeper inspection or other services. 

The other network extensibility framework for NSX-T Data Center is Endpoint Security, EPSec. NSX-T Data Center uses EPSec for capabilities leveraged to provide enhanced agentless antivirus/malware or endpoint-monitoring functions from third-party security vendors as well. This integration is optional and in Horizon deployments is beneficial for removing the need for antivirus or anti-malware (AV/AM) agents on the guest operating system. This functionality is provided by NSX Data Center through guest introspection, which is also leveraged to provide information for the identity firewall.

To implement guest introspection, a guest introspection VM must be deployed on each ESXi host in the desktop cluster. Also, the VMware Tools driver for guest introspection must be installed on each desktop or RDSH VM. 

NSX-T Components

An NSX-T Data Center architecture consists of the following components. This section is not meant to be an exhaustive guide for covering NSX-T and every component. For a more in-depth look at the NSX-T components and design decisions, reference the VMware NSX-T Reference Design guide and NSX-T Data Center documentation.

Table 13: NSX-T Components

Component

Description

NSX-T Manager Cluster

The NSX-T Manager Cluster is deployed as a virtual appliance in a set of three appliances for redundancy and scalability. The NSX-T Manager Cluster represents the Management and Control Plane of the NSX-T deployment. The NSX-T Manager Cluster operates a separate user-interface for administration.

Transport Node

A Transport Node is a device that is prepared with NSX-T and is participating in traffic forwarding.

Edge Transport Node - The NSX-T Edge Node can be deployed in two form factors, as a virtual appliance in various sizes (Small, Medium, and Large) and also in a bare metal form factor. The NSX-T Edge Nodes provide North/South connectivity as well as centralized services such as NAT, Load Balancing, Gateway Firewall and Routing, DHCP, and VPN capabilities. NSX-T Edge Nodes provide pools of capacity for running these services.

Hypervisor Transport Node - Hypervisor transport nodes are hypervisors prepared and configured to run NSX-T. The N-VDS provides network services to the virtual machines running on those hypervisors. NSX-T currently supports VMware ESXi and KVM hypervisors but also works on Bare Metal workloads as well.

NSX Virtual Distributed Switch (N-VDS)

The N-VDS is a software switch based on the vSphere vSwitch that provides switching functionality for NSX-T and is installed in the hypervisor, in the case of Horizon, VMware vSphere, and managed by NSX-T. The N-VDS provides the capability for building networks using overlay encapsulation/decapsulation. The N-VDS has similarities to the Virtual Distributed Switch that vCenter can build for vSphere hypervisors in concept, but is fully managed by NSX-T and has no dependency on vCenter or vSphere.

Transport Zone

An NSX-T Transport Zone represents the span of an NSX-T Segment and can be either an Overlay or VLAN type. It is associated with an N-VDS which binds it to a Transport Node.

NSX-T Segment

The NSX Segment is virtual Layer 2 broadcast domain. Segments are created and attached to a Transport Zone. The span of a Segment is defined by the Transport Zone. An NSX-T Segment can either be a type Overlay or VLAN and inherits the type from the Transport Zone of which it’s attached. An NSX-T Overlay Segment represents a software construct of a layer 2 broadcast domain. It can be associated with a network subnet and its default gateway is typically the tier-0 or tier-1 logical NSX-T router. A VLAN Segment represents a software extension of a physical layer 2 broadcast domain into the virtualized environment. The subnet ID of the underlying physical network is associated with the VLAN Segment, and the default gateway is typically the physical router that already provides the gateway in the underlay network.

NSX-T Kernel Modules

The NSX-T kernel modules are installed on each hypervisor that NSX-T is managing and build a distributed routing and firewalling model that scales out as a customer adds more hosts to their environment. Deploying NSX-T to the hypervisors installs the software packages that enable in-kernel routing and firewalling for local workloads. NSX-T manages these local firewalls and routing instances through the NSX-T Simplified UI.

NSX-T is broken up into 3 basic planes, Management, Control, and Data. Each of these planes is independent from the other. Loss of function in one plane does not equate loss of function in another plane. The figure below shows the NSX-T logical architecture and the breakdown of the components in each plane.

Figure 10: NSX-T Data Center Logical Architecture

NSX-T Scalability

See the NSX-T Data Center Recommended Configuration Maximums guide. With regard to a Horizon environment, the two most relevant sections are the Distributed Firewall and the Identity Firewall sections.

Design Overview

NSX-T Data Center and Horizon deployments, can accommodate a wide ranges of deployment topologies and options. NSX-T needs to align at logical boundaries with Horizon to scale similarly. Not every customer requires a topology that scales to the Horizon Cloud Pod Architecture limits but guidance is helpful for showcasing the options. The following sections represent the three Horizon and NSX-T topologies that align NSX-T at the Horizon pod boundary as well as providing guidance for customers that fall into the sizing categories. These topologies are not meant to be comprehensive or to replace a services-led engagement recommendation, but provide guidance on what would be required to build each scenario based on each platform’s provided maximums.

Small (Converged Cluster) Topology

The Small topology represents a converged cluster design where the Management, Edge, and Compute Clusters of both the Server and Horizon domains reside on the same logical vSphere compute cluster. This topology would be similar to using hyper-converged hardware to run Horizon, Enterprise Servers, and the Management components where the compute infrastructure consists of a single vSphere cluster. This Horizon and NSX-T topology aligns from an NSX-T design perspective with the Collapsed Management and Edge Cluster design in the NSX-T Reference Architecture.

Figure 11: Example Small Horizon and NSX-T Data Center Topology

Medium Topology

The Medium topology represents the physical splitting of the management, edge, and compute clusters. This Horizon and NSX-T topology aligns, from an NSX-T design perspective, with the collapsed management and edge cluster design in the NSX-T Reference Architecture, into one cluster to minimize the amount of hardware required for deployment.

Figure 12: Medium Horizon and NSX-T Data Center Topology

Large Topology

The large topology represents the scalable Horizon and NSX-T deployment for the large deployments and ability to support multiple Horizon pods. This Horizon and NSX-T topology aligns from an NSX-T design perspective with the enterprise vSphere-based design in the NSX-T Reference Architecture. All clusters are split out completely and separately, from each other, to isolate fault domains for each aspect of the design.

The management clusters for multiple pods could be collapsed and share the management cluster for domains to reduce the overall hardware needed and reduce some complexity of the design. As the scale of an environment increases with more pods, or to spread the fault domain, this could also be split out into separate management clusters.

While this design does require more hardware to accomplish, it provides maximum and independent scalability for large Horizon deployment needs.

 

Figure 13: Large Horizon and NSX-T Data Center Topology

Table 14: Design Strategy for NSX-T

Decision

The medium topology was used.

Justification

Each location was sized for an 8,000 user Horizon environment in a single Horizon pod.

Separate vSphere clusters are already in place for the management components and for the Horizon desktops.

The medium topology can easily be scaled to a large topology if the size of the environment dramatically increases.

References

Reference the VMware NSX-T Reference Design guide and NSX-T Data Center documentation.

The NSX-T Data Center and EUC Design Guide includes guidance on:

  • Network Virtualization for Horizon
  • Micro-segmentation for Horizon
  • Load-balancing for Horizon management components

NSX-T Data Center and Horizon example architectures for small, medium, and large deployments

 

Microsoft Azure Environment Infrastructure Design

In this reference architecture, multiple Azure regional data centers were used to demonstrate multi-site deployments of Horizon Cloud Service on Microsoft Azure. We configured infrastructures in two Microsoft Azure regions (US East, US East 2) to facilitate this example.

Each region was configured with the following components.

Table 15: Microsoft Azure Infrastructure Components

Component

Description

Management VNet

Microsoft Azure Virtual Network (VNet) configured to host shared services for use by the Horizon Cloud deployments.

VNet peer (management to pod)

Unidirectional network connection between two VNets in Microsoft Azure. Both the Allow forwarded traffic and Allow Gateway Transit options were selected in the VNet configuration to ensure proper connectivity between the two VNets.

VNet peer (pod to management)

Unidirectional network connection between two VNets in Microsoft Azure. Both the Allow Forwarded Traffic and the Allow Gateway Transit options were selected in the VNet configuration to ensure proper connectivity between the two VNets.

Microsoft Azure VPN Gateway

VPN gateway resource provided by Microsoft Azure to provide point-to-point private network connectivity to another network.

Two Microsoft Windows Server VMs

Two Windows servers provide redundancy in each Microsoft Azure region for common network services.

Active Directory domain controller

Active Directory was implemented as a service on each Windows server.  Active Directory was configured according to Option 3 in Networking and Active Directory Considerations on Microsoft Azure for use with VMware Horizon Cloud Service.

DNS server

DNS was implemented as a service on each Windows server.

Windows DFS file share

A Windows share with DFS was enabled on each Windows server to contain the Dynamic Environment Manager profile and configuration shares.

Horizon Cloud control pod VNet

Microsoft Azure VNet created for use of the Horizon Cloud pod. This VNet contains all infrastructure and user services components (RDSH servers, VDI desktops) provided by the Horizon Cloud pod.

We also leveraged two separate Microsoft Azure subscriptions to demonstrate that multiple pods could be deployed to different subscriptions and managed from the same Horizon Cloud Service with Microsoft Azure control plane. For more detail on the design decisions that were made for Horizon Cloud pod deployments, see Horizon Cloud on Microsoft Azure Architecture. 

Network Connectivity to Microsoft Azure

You do not need to provide private access to Microsoft Azure as a part of your Horizon Cloud on Microsoft Azure deployments. The Microsoft Azure infrastructure can be provided from the Internet. 

There are several methods for providing private access to infrastructure deployed to any given Microsoft Azure subscription in any given Microsoft Azure region, including by using a VPN or ExpressRoute configurations. 

Table 16: Implementation Strategy for Providing Private Access to Horizon Cloud Service

Decision

VPN connections were leveraged. Connections were made from the on-premises data center to each of the two Microsoft Azure regions used for this design.

Justification

This is the most typical configuration that we have seen in customer environments to date.

See Connecting your on-premises network to Azure for more details on the options available to provide a private network connection to Microsoft Azure.

Microsoft Azure Virtual Network - VNet

In a Horizon Cloud Service on Microsoft Azure deployment, you are required to configure virtual networks (VNets) for use by the Horizon Cloud pod. You must have already created the VNet you want to use in that region in your Microsoft Azure subscription before deploying Horizon Cloud Service.

Note that DHCP is a service that is a part of a VNet configuration. For more information on how to properly configure a VNet for Horizon Cloud Service, see Configure the Required Virtual Network in Microsoft Azure.

Another useful resource is the VMware Horizon Cloud Service on Microsoft Azure Requirements Checklist.

What's Next?

Now that you have come to the end of this chapter, you can return to the landing page and search or scroll to select your next chapter in one of the following sections:

  • Overview chapters provide understanding of business drivers, use cases, and service definitions.
  • Architecture chapters explore the products you are interested in including in your platform, including Workspace ONE UEM, Workspace ONE Access, Workspace ONE Assist, Workspace ONE Intelligence, Horizon, App Volumes Dynamic Environment Manager, and Unified Access Gateway.
  • Integration chapters cover the integration of components and services you need to create the platform capable of delivering what you want.
  • Configuration chapters provide reference for specific tasks as you build your platform, such as installation, deployment, and configuration processes for Horizon, App Volumes, Dynamic Environment Management, and more.

Filter Tags

Document Reference Architecture Advanced Design