Loading...

 

OOPS

Your browser doesn't support required HTML5 capabilities.

This demo works best with the latest version of Firefox, Chrome, or Safari.

OOPS

This demo is incomplete or damaged. Please reload the page, or download again:

For VMware partners:
www.vmware.com/go/partnerdemos

For VMware employees:
www.vmware.com/go/demos

OOPS

Something broke.

Visit the VMware Demo Library
to get more demos!

For VMware partners:
www.vmware.com/go/partnerdemos

For VMware employees:
www.vmware.com/go/demos

This is an interactive demo

Drive it with your mouse... or your finger

Use Learn mode to learn the demo

The orange boxes show you where to click

Use Present mode to hide all that when presenting

Hit the F4 key to show notes in a separate window

 

Left Arrow KeyRight Arrow KeyYou can also use the arrow keys to step forward or backward

Shortcuts jump to different parts of the demo

OK, got it!

Demo

VMware Demo Library

Demos can have multiple paths, so the sequence of clicks may differ from this list.

Please wait...
Unsaved changes! You may continue editing other frames before saving.
Save Changes
0.

This demo showcases the ability to manage health, performance and capacity of a modern Software Defined Datacenter (SDDC) using vRealize Operations Manager. It contains individual demo sections covering the vROps management pack solutions listed below:

  • Management Pack for Storage Devices (MPSD)
  • Management Pack for Network Devices (MPND)
  • NSX-vSphere Management Pack (includes MPND)
  • vCloud Air Management Pack

To access the sections use the menu on the left, or click on the appropriate steps.

[For MPSD, NSX-vSphere, MPND or vCloud Air, click the 'Dashboard List' pulldown]
[For Virtual SAN with MPSD, click the right-side scrollbar down]
10.
[Click from the menu for MPSD, NSX-vSphere, MPND or vCloud Air]
100.

The Management pack for Network Devices (MPND) provides visibility and analysis of the pyhysical network infrastructure in your vSphere environment. Using discovery protocols such as LLDP or CDP, MPND can detect leaf switches connected to vSphere hosts as well as upstream spine switches connected to the leafs.

The topology tree on the left shows parent / child relationships of physical switches, hosts and distributed switches along with their health status. Active alerts are displayed at the top of the middle column. The topology graph on the far right renders the physical map of resources connected to objects selected in the Network Device Overview tree. Scrolling to the bottom of this dashboard reveals additional metrics for the selected object as well as other network traffic information.

[Click the scrollbar down or hover on the first Leaf Switch under 'Network Device Overview']
110.
[Click the scrollbar down]
120.
[Click the scrollbar up or scroll the 'Sparkline Chart' side-to-side by clicking the lower left scrollbar]
130.
[Click the scrollbar up]
140.
Change the context of the charts by selecting a host in the topology tree.

[Click the first 'Host System']
290.
[Click the scrollbar down or hover on the first Host System]
300.
[Click the scrollbar down]
310.
Notice the Sparkline Chart now reflects traffic statistics for the selected Host

[Click the scrollbar up]
320.
[Click the second vSphere Distributed Switch or hover on either of the outer switches in the Topology Graph]
470.
[Click the scrollbar down]
480.
Notice the Sparkline Chart now reflects traffic information for the Distributed Switch

[Click the scrollbar up]
490.
Let's move on to the Network Device Connectivity dashboard.

[Click the 'Network Device Connectivity' tab at the top of the window or hover on the switch in the topology diagram]
600.
Use the Filter field to reduce the number of selectable items in the object list

[Click the 'Filter' text box]
610.
[Hit any key to type '-c3']
670.
[Click the first Host System in the list ('w2-sm-c3b4.mgmt.local')]
700.
[Click the second Host System in the list ('w2-sm-c3b1.mgmt.local')]
800.

The physical network path between the two hosts is rendered in the Path widget. Any displayed item or interface can be hovered over to get property and health information.

MPND is a powerful tool for mapping out the physical network devices in the SDDC and displaying the connectivity chain up through leaf and spine switches. MPND plays a key part in the correlation of logical networks to physical network devices, which we will explore further in the NSX portion of this demo.

(End of this section)

[Hover on the top leaf switch in the topology diagram, or click the Dashboard List to return to the start of the demo]
820.
To skip ahead to the NFS section of the demo, click the checkbox next to 'NFS'
To continue with MPSD, click 'FC-FCoE'

[Click from the menu for MPSD, NSX-vSphere, MPND or vCloud Air]
840.
[Click 'FC-FCoE Troubleshooting']
860.

The Management Pack for Storage Devices (MPSD) provides insight and troubleshooting tools for managing traditional datacenter storage fabrics such as Fibre Channel, ISCS and NFS. MPSD also provides deep analysis into Virtual SAN environments which is covered in a separate section.

In this section we will briefly look at a Fibre Channel fabric and its associated components, then dive deeper into a NFS storage environment. Start by selecting a host object in the topology tree, taking note as the other dashboard widgets display data related to the selected host.

[Click the far-left Host (yellow)]
920.
[Click the right-side scrollbar down]
930.

As we scroll down we see components in the topology tree that make up the Fibre Channel storage fabric. Host w2-mgmtpm-1 is currently selected, so its child components are highlighted in the tree. Storage throughput statistics for the host are displayed in the metric charts on the right.

[Click the scrollbar down (again)]
940.

The bottom section of the topology tree shows the fibre channel swich ports, array target ports and storage array mounts which are child resources of the currently selected host. The tree provides a correlation and relationship chain of the fibre channel fabric components to which this host is directly related.

[Click the scrollbar up]
950.
[Click the 'x' to close the 'FC-FCoE Troubleshooting' tab]
970.
[Click the 'Dashboard List' at the top of the window]
980.
[Click 'MPSD']
1000.

In each of the MPSD dashboard groups, notice the same 3 dashboard types are available for each storage protocol - Component Heatmap, Component Usage, and general Troubleshooting.

[Hover on 'FC-FCoE']
1010.
[Hover on 'iSCSI']
1020.
[Click the checkbox next to 'NFS']
1040.
[Click 'NFS Components Heatmap']
1060.
Using the MPSD NFS dashboards, let's dig deeper into the storage environment highlighting the tools and charts available to manage the NFS storage fabric.

[Hover on the middle of the big red box]
1070.

In the NFS Volume Throughput and Latency heatmap, the box sizes are determined by the amount of current throughput. The bigger the box, the more throughput the volume is experiencing relative to the other volumes in the map. The color red indicates latency is high at the moment, while the sparkline shows the latency spike for this volume began relatively recently.

[Click '(Details)']
1130.

At the summary tab for this NFS volume we see that even though the heatmap indicated high throughput and latency compared to the other volumes, there are no immediate health, risk or efficiency issues with the volume.

[Click the 'Analysis' tab at the top]
1230.

Note that current throughput is just under 75 MBps vs. the historical max of 200 MBps. Even though the heatmap showed this volume was working harder than the other NFS volumes, it is still well under max throughput capabilities.

[Click 'Home' in the upper left corner]
1340.
Let's move over to the Troubleshooting dashboard.

[Click the 'NFS Troubleshooting' tab]
1390.

Similar to what we observed in the FC Troubleshooting dashboard, here we see the familiar topology tree plus other widgets showing active Issues and Alerts for our NFS storage fabric.

[Click the left-most Host (top left yellow badge)]
1450.
Clicking on the host in the topology tree caused the other widgets on this the dashboard to display contextually related information to the selected host.

[Click the scrollbar down]
1460.

As we scroll down the topology tree we see NFS storage fabric components highlighted that are directly related to the selected host. The dark green EsxPnics are child resources of the host, as are the darker IP switchports to which the Pnics are connected. The metric charts on the right reflect storage throughput statistics for the selected host.

[Click the middle scrollbar down]
1470.

At the very bottom of the topology tree we see a highlighted NFS volume that is mounted by this host. Notice there is nothing highlighted at the IPSwitch, NFS Server and Storage Array levels. That is because none of those objects are direct child resouces of a host. Highlighted items in the topology tree only reflect direct parent and child related resources of the selected object.

[Click the right-side scrollbar up]
1480.
[Click the red Datastore]
1580.
Clicking over to the datastore details we can investigate what relationships exist, including which VMs are being hosted on this datastore.

[Click the left panel scrollbar down]
1590.
[Click 'Virtual Machine' in the left panel]
1630.

The virtual machines listed on the left all have VMDKs on this datastore. If the datastore was experiencing critical issues, the VMs listed would potentially be negatively impacted.

Let's dig a little deeper to see if we can tell how long this datastore has been running hot.

[Click the 'Analysis' tab at the top of the window]
1720.
The Anomalies Score graph indicates the anomaly count began to spike a few hours ago.

[Click the right-side scrollbar down]
1730.

In the 'Info' column, we see that several values are above historical maximums or defined thresholds. The time bars show when the anomalies began to appear. The first latency spike was around 10:38am, then a more sustained peak began at 11:08am. This information can assist in correlating other anomalous behavior that showed up in the environment around the same time.

[Click the right-side scrollbar up]
1740.
Let's move to the Time Remaining tab to look at the same volume from a consumption and capacity perspective.

[Click the tab for 'Time Remaining']
1820.
Looking at the datastore from a capacity standpoint, Time Remaining reveals that at current consumption trends, this datastore will run out of space in 86 days.

[Click the right-side scrollbar down]
1830.

Expanding the Disk Space item gives us a more detailed view of the Usage and Allocation details. This provides a better sense of whether disk space usage is truly being exhausted, or rather that we've been overallocating the storage on this datastore. Note this volume has 30% of its capacity currently being consumed, but 40% of the capacity has been allocated.

[Click 'Home' in the upper left corner]
1890.
Let's move to the third NFS dashboard to look at detailed usage analysis of the NFS components themselves.

[Click the 'NFS Components Usage' tab]
1910.

This dashboard provides a detailed look at throughput, latency and errors for each individual NFS storage component including the EsxPnics, IP switchports, and NFS volumes.

[Click the right-side scrollbar down]
1920.

In the lower middle chart, notice that the latency for 'ntp_east_sata_1000_01_nfs' is very high compared to the other NFS volumes. What we can't tell is how long this condition has existed. The default chart setting is to display statistics from the last hour. Let's see how the volumes compare over the last week.

[Click the 'Edit Widget' icon for the 'NFS Volume Read Latency (ms)' widget]
1970.
[Click the pulldown for 'Range: Last hour']
2000.
[Click 'Last 7 days']
2030.
[Click the 'Save' button]
2140.

Here we see a different picture. During the last 7 days, the latency for 'ntp_east_sata_1000_01_nfs' was much lower (8.46 ms compared to 160.17 during the last hour), and the latency relativity amongst the other volumes was more balanced. To confirm, let's go back to look at the stats for the last hour.

[Click the 'Edit Widget' icon for the 'NFS Volume Read Latency (ms)' widget]
2200.
[Click the pulldown for 'Range: Last 7 days']
2230.
[Click 'Last hour']
2260.
[Click the 'Save' button]
2350.

We clearly see there is higher relative latency for this volume compared to the others during the last hour, however it has not been a chronic issue over the last 7 days. The previous dashboards indicated that although the volume is working hard, it is experiencing no immediate health, risk, or efficiency issues that require further investigation. If any of the virtual machines we saw on the datastore relationship list happen to be experiencing performance issues, we can confidently rule out the NFS fabric and volume as the culprit based on the MPSD data.

(End of this section) [Click the Dashboard List to return to the start of the demo]
2360.

This section of the MPSD demo will focus on Virtual SAN storage environments. We will look at troubleshooting VSAN alerts and also analyze VSAN performance and capacity information.

Notice there are several active alerts in the environment. We will drill into the first VSAN alert.

[Click the yellow alert at the top of the list ('One or more VirtualSAN components...')]
2390.
Several virtual machines have VSAN components that are inactive or missing. Let's look at the first one on the list: LiveAuction App2.

[Click the top row of the table]
2480.

The alert indicates there are VSAN components for this VM that are not in an active state on disk. The Device Description indicates the inactive components are on host w2-sm-c3b1. The Recommendations text suggests the host may be unavailable.

Return to the Home screen to look for correlating issues.

[Click the 'Home' button in the upper left]
2650.
Let's activate the VSAN dashboards and go to VSAN Troubleshooting.

[Click the 'Dashboard List' pulldown]
2660.
[Click 'MPSD']
2710.
[Click the checkbox for 'VirtualSAN']
2740.
[Click 'VirtualSAN 6 Troubleshooting']
2810.
The VSAN topology tree shows a host that is disconnected.

[Click the left-most Host (shows '?')]
2890.
Selecting an object in the topology tree reveals related alerts in the Top Issues widget. It appears a host is indeed down. Let's drill into this alert for more information.

[Click the top alert ('Host has lost connection to vCenter Server')]
3110.

Host w2-sm-c3b1 is disconnected from vCenter. This is the same host which held the VSAN missing disk objects for the LiveAuction App2 VM that we were investigating. Reconnect the host to vCenter to restore the missing VSAN objects.

[Click the 'Home' button in the upper left]
3250.
Now let's select a connected host in the topology tree to view the related child objects and check for alerts or issues.

[Click the right-most Host (green)]
3290.
It appears this host had a pNIC link status down at one point, but the host does not currently show degraded health. It is an old alert that has not yet been cancelled.

[Click the right-side scrollbar down]
3310.

Scrolling down allows us to see the other items in the topology tree along with metric charts. The topology items in dark green are direct child resources of the selected host. The metric charts are showing throughput statistics for the host.

Let's change the dashboard context to a SSD.

[Click the right-most Solid State Device]
3380.
Notice there are no highlighted items in the Magnetic Disk or EsxPnic rows because those objects are not directly related to SSDs. Also note the Metric Chart now shows statistics for the selected SSD.

[Click the right-side scrollbar up]
3390.
[Click the right-most Host]
3440.
Re-clicking the host object in the topology tree causes the other dashboard widgets to reflect data for the selected host.

[Click the 'VirtualSAN 6 Heatmap' tab]
3460.
[Hover on the left-most red box in the heatmap]
3470.
[Click 'Show Sparkline']
3480.

This heatmap shows throughput and workload for host storage controllers in the VSAN environment. The boxes are sized by the amount of throughput, while the color is driven by the workload %. Red indicates a high workload. The sparkline for this host controller indicates this is a recent workload spike on host w2-sm-c3b4. Let's get more detail.

[Click 'Details']
3620.
Although the workload % and throughput for this controller were both high, there does not appear to be an immediate Health, Risk or Efficiency issue. Let's analyze further.

[Click the 'Analysis' tab]
3720.

In Workload Breakdown note the highest observed throughput previously seen on this controller was about 278 MBps. The current throughput is running about 271 MBps which is approachng 98% of the max observed capacity. That doesn't mean there is a problem - it simply means the controller is working hard. Most likely the VSAN cluster is attempting to create another copy of the missing VM components caused by one of the VSAN hosts going offline.

[Click 'Home' in the upper left]
3850.
[Click the right-side scrollbar down]
3870.
Scrolling down in the dashboard reveals other heatmaps for Diskgroup and NIC throughput / workload, plus VSAN environment component counts.

[Click right-side scrollbar up]
3890.
Let's have a closer look at VSAN entity usage.

[Click the 'VirtualSAN 6 Entity Usage' tab]
3940.
This dashboard displays throughput and latency for each of the main VSAN objects - host adapters, flash disks and spinning disks. We can observe the workload distribution amongst each entity grouping, plus quickly determine if latency numbers are abnormally high.

[Click the right-side scrollbar down]
3960.
[Click the scrollbar down again]
3970.

At the bottom of the dashboard are the throughput and latency charts for the magnetic spinning disks in the VSAN environment. We can drill into any items in these charts to further analyze workload imbalance or higher than expected latency numbers.

[Click the tab 'VirtualSAN 6 Device Insights']
4000.

The previous dashboard displayed entity usage information from the perspective of throughput and latency. The Device Insights dashboard is focused on capacity and error information. The top row refers to the magnetic disks - how much capacity has been consumed on each disk, and whether a disk has errors or reallocated sectors. The middle row focuses on the SSDs - cache hit rate, wearout indicators and SSD errors.

[Click the right-side scrollbar down]
4150.

The bottom section of the dashboard has CPU and Memory consumption information for the VSAN hosts. There is also a widget for listing IP switchport errors. Use this dashboard to identify VSAN devices that are approaching capacity limits or throwing an abnormal amount of errors compared to other VSAN devices.

[Click the 'VirtualSAN 6 Cluster Insights' tab]
4190.

The remaining VSAN dashboard displays statistics from the perspective of the entire VSAN cluster vs. individual components. These four charts show Disk Group consumption, latency and errors. The default time range for the data being displayed is the previous hour. However we can modity the range to a value of our choosing.

[Click the settings icon (pencil) for the 'Disk Group Usage (MBps)' widget]
4220.
[Click the pulldown for 'Range: Last hour']
4250.
[Click 'Last 24 hours']
4280.
[Click the 'Save' button]
4410.
The chart we edited is now displaying data from the last 24 hours, while the other three charts are still showing data from the last hour.

[Click the right-side scrollbar down]
4460.
[Click the pulldown for 'Range: Last 24 hours']
4490.
[Click 'Last hour']
4520.
[Click the 'Save' button]
4570.
[Click the right-side scrollbar down]
4590.

The four charts at the bottom of the dashboard reflect capacity, throughput and latency from the perpective of the VSAN datastore. There is also a chart indicating VSAN component counts for each host in the cluster. Use this dashboard to analyze performance and capacity of the VSAN disk groups and datastore as opposed to the VSAN devices and individual components shown in the previous dashboards.

This is the final dashboard in the VSAN section of MPSD. [Click the 'Recommendations' tab]
4670.
(End of this section)

[Click the Dashboard List to return to the start of the demo]
4680.
[Click MPSD, NSX-vSphere, MPND or vCloud Air]
4730.

This section of the demo will focus on management of vCloud Air resources and workloads.

Similar to other SDDC management packs, the vCA Troubleshooting dashboard offers a topology tree to display the relationship chain of all the objects in the environment. Clicking on an item in the tree will highlight the corresponding parent and child resources related to the selected object. The Interesting Metrics chart will change to display statistics for the selected object.

[Click the right-side scrollbar down]
4740.
As we scroll down in the dashboard we expose more levels of the topology tree on the left, and a chart showing health, anomalies and events on the right.

[Click the right-side scrollbar up]
4750.

The Troubleshooting dashboard is the place to start if vCA workloads or resources are experiencing problems. At a quick glance we will see items that have degraded health status, or open alerts and events, and can drill into either for further investigation.

Let's move to the next dashboard to see capacity usage information.

[Click the tab 'vCloud Air Data Center Capacity Usage']
4780.
The top widget lists the virtual data centers registered to your vCloud Air account. The Workload section shows current consumption of CPU, Memory and Storage for the selected data center.

[Click the right-side scrollbar down]
4790.
Scrolling to the bottom of the dashboard reveals the overall capacity remaining for the selected data center. Use this dashboard to get a quick glance at cloud data center consumption, and to check remaining capacity.

[Click the right-side scrollbar up]
4800.
[Click the arrow button on the right to scroll the tabs]
4830.
[Click the tab 'vCloud Air Heatmaps']
4870.

The heatmaps dashboard enables us to quickly see which VMs are the highest resource consumers in four separate categories - CPU, Memory, Disk and Network. The CPU and Memory heatmaps have multiple selections for grouping the VMs. Clicking on items in the heatmaps will display the history of their resource consumption in the charts below the heatmaps of each respective resource category.

[Hover on the red rectangle in the 'CPU Load' widget above '8.295']
4890.
[Click the same rectangle]
4910.
[Click the red rectangle on the far right of the 'CPU Load' widget]
4940.

Notice the historical graph of CPU Usage % for the two selected VMs in the CPU Load History chart below the heatmap. We can also change the heatmap to show CPU Ready % for the workload VMs by choosing that category in the Configurations drop down list.

[Click the pulldown at the the top of the 'CPU Load' widget ('VMs by CPU Usage (%)')]
4960.
Let's redraw the map based on CPU Ready percentage.

[Click 'VMs by CPU Ready (%)']
5000.
[Hover on the green rectangle in the 'CPU Load' widget above '5' in the color scale]
5010.
Notice the PAN-VMs that were red in the CPU Usage map are not experiencing any significant CPU Ready contention.

[Move the mouse away from the green rectangle]
5030.
[Click the red rectangle in the 'Memory Load' widget above '19.879' in the color scale]
5060.
[Click the red rectangle in the top right of the 'Memory Load' widget]
5090.
[Click the pulldown at the the top of the 'Memory Load' widget ('VMs by Mem Usage (%)')]
5110.
[Click 'VMs by Mem Swap in Rate (KBps)']
5150.
Notice the VMs that were red in the Memory Usage map are not experiencing any significant memory swapping. These heatmaps are useful for quickly identifying the top resource consumers in the cloud, and whether there is a resource contention issue in play.

[Click the tab 'vCloud Air VM Utilization']
5180.
[Click the double-up arrow in the far-right widget ('Top 25 VMs by Network Usage Rate (KBps)') to collapse it]
5220.
The VM Utilization dashboard has several Top-25 lists to help quickly gauge the balance of resources across your cloud, and which VMs are the top consumers. We can expand or collapse any of the widgets to display just the lists we wish to see.

[Click the tab 'vCloud Air VM Performance']
5250.
The VM Performance dashboard helps identify VMs that could be experiencing resource contention issues. If there are VMs with high levels of CPU Ready or Memory Swapping, the next step would be to check the Alerts dashboard for active warnings.

[Click the tab 'vCloud Air Alerts']
5320.
Notice the active alert regarding the Load Balancer Service. Let's drill in for more detail.

[Click the only row in the list under the 'vCloud Air Alerts' widget]
5410.
The alert indicates the load balancer service has been configured but is not running. A service restart is recommended.

[Click 'Home' in the upper left]
5520.
[Click the 'vCloud Air Edge Services' tab]
5540.
This dashboard groups all of the Edge Gateway services together to quickly view their status and performance.
5550.
[Click the double down-arrows in the top left widget ('Top 25 Edge Gateways by Uplink Rx (KB...') to expand it]
5610.
[Click the double down-arrows in the top right widget ('Top 25 Edge Gateways DHCP by Pool U...') to expand it]
5660.

The Edge Services dashboard provides a single view into the status and performance of all services running on Edge devices in your vCloud environment. The vCloud Air management pack provides a comprehensive view into your cloud performance, capacity and resource consumption. It also helps quickly identify issues and offers recommendations for resolution.

(End of this section) [Click the Dashboard list to return to the start of the demo]
5730.

This section of the SDDC demo will focus on NSX for vSphere components in the SDDC. The NSX-vSphere management pack also contains the Management Pack for Network Devices (MPND) to aid in the correlation of physical to logical network constructs. However, MPND has no requirement for NSX, and can be run independently as a standalone vROps management pack. MPND is covered in a separate section of this demo.

The NSX Main dashboard displays several operational aspects of virtualized networks including overall health, performance and availability. We can select between multiple NSX environments from this single dashboard by choosing the desired NSX implementation in the Environments widget.

[Click 'nsx nsbu']
5840.
The top portion of the dashboard displays alerts, heapmaps and topology trees, while the bottom section reveals top consumers of network resources.

[Click the right-side scrollbar down]
5850.
[Click the right-side scrollbar up]
5860.
Move to the Topology dashboard to display the physical and logical network correlation tools.

[Click the tab for 'NSX-vSphere Topology']
5890.
Let's look at a NSX Logical Switch.

[Click '224-b2-app']
6070.

The dashboard has rendered the logical and physical topology underpinning the selected NSX Logical Switch. On the left we see the logical switch in the center, the VMs connected to the switch, and the upstream NSX Logical Router. On the right we see the physical hosts, network interfaces, switch interfaces, and the leaf and spine switches this Logical Switch is using for communication. At the bottom of the dashboard are alert lists and metric charts for the selected object.

[Click the right-side scrollbar down]
6080.
We can hover over each link in the physical topology chain to view information regarding the health and properties of that particular network object.

[In the top right panel: hover on the leftmost objects, then click the Host System to continue]
6180.
Note the Metrics chart is now showing stats for the selected host.

[Click the logical router on the right side of the left-side panel]
6320.
The Top Issues and Metrics charts now reflect the selected Logical Router.

[Click the tab for 'NSX-vSphere Object Path']
6350.
The Object Path dashboard displays correlating logical and physical network paths between any two entities in the NSX environment.

We will select the desired NSX environment.

[Click 'nsx east']
6380.
We can minimize any dashboard widget to make room on the screen for the other sections.

[Click the double up-arrows for 'NSX-vSphere Environments' to hide the panel]
6400.
Use the filter field to narrow the list of objects to investigate.

[Click in the 'Filter' text box]
6410.
[Hit any key to type 'live']
6470.
We filtered the list to only display objects containing the string 'live'. Now let's investigate the communication path between App1 and NoSql1 in the LiveAuction application.

[Click 'LiveAuctionW1|1App1']
6510.
[Click 'LiveAuctionW1|1NoSql1']
6670.
The logical communication path between App1 and NoSql1 reveals each VM is connected to an upstream Logical Switch, both of which connect to the east-dlr-1 Distributed Logical Router.

Scroll down to fully display the Physical Path widget.

[Click the right-side scrollbar down]
6680.
The communication path between the physical servers hosting these two VMs is revealed all the way through the upstream leaf and spine switches. We can zoom in or out in the path widgets, or drag any object to any location in the widget to unclutter a complex array of devices.

[Click the 'Navigation' icon under 'Logical Path']
6700.
To further investigate network communication between the two selected objects, navigate to the Troubleshooting dashboard directly from the path widgets.

[Click 'NSX-vSphere Troubleshooting']
6760.
The two virtual machines App1 and NoSql1 chosen in the previous Object Path dashboard are still selected although not displayed here. Select the Run Traceflow action to reveal the network hops between these two VMs.

[Click the pulldown for 'Choose Action' in the lower right]
6780.
[Click 'Run traceflow']
6800.
[Click the right-side scrollbar down]
6810.
[Click the blue arrow icon by the action]
6830.
Each step of the communication path between App1 and NoSql1 is revealed, including any firewall rules that were processed in the traceflow path.

[Click the right-side scrollbar up]
6840.
Use the Filter field to narrow the object list to just the Live Auction VMs.

[Click in the 'Filter' text box for 'Objects']
6850.
[Hit any key to type 'live']
6900.
Select the App1 VM to change the list of available actions to be contextually specific to virtual machines.

[Click LiveAuctionW1|1App1]
6940.
[Click the pulldown for 'Choose Action:']
6970.
[Click 'Get Distributed Firewall Statistics']
6980.
[Click the right-side scrollbar down]
6990.
[Click the blue arrow next to the action]
7010.
The Get Distributed Firewall Statistics action reveals the list of firewall rules that are processed when this VM communicates on the network.

[Click the right-side scrollbar up]
7020.
Filter the object list again, this time focusing on NSX components.

[Click in the text box that says 'live']
7030.
[Hit any key to type 'nsx']
7080.
Let's find out what troubleshooting actions are available specific to each NSX component, starting with a logical switch.

[Click '224-b2-app']
7120.
[Click the double up-arrow to close the 'NSX-vSphere Environments' panel]
7140.
With '224-b2-app' selected, view the available actions that can be run against NSX logical switches.

[Click the pulldown for 'Choose Action:']
7160.
Note the Logical Switch troubleshooting actions.

[Click the pulldown for 'Choose Action:' (again)]
7180.
Results from previously run actions can be selected and displayed from the Results List dropdown.

[Click the pulldown for 'Results:']
7210.
Here is the list of results from previous actions run against the switch '224-b2-app'.

[Click the first item in the list]
7220.
[Click the bottom of the right-side scrollbar to view the full results]
7230.
[Click the right-side scrollbar up]
7240.
Now let's view the available actions for a NSX Logical Router.

[Click 'east-dir-1' from the 'Objects' list]
7270.
[Click the pulldown for 'Choose Action:']
7300.
[Click the pulldown for 'Results:']
7330.
[Click the only option ('Show runtime state...')]
7340.
[Click the bottom of the right-side scrollbar to view the full results]
7350.
[Click the right-side scrollbar up]
7360.
Lastly, let's look at the available actions for a NSX Edge device.

[Click 'esg-east-prm-1' on the list of Objects]
7400.
[Click the pulldown for 'Choose Action:']
7430.
[Click the only option ('Check routing configuration')]
7440.
[Click the pulldown for 'Results:']
7470.
[Click the first item in the list]
7480.
[Click the bottom of the right-side scrollbar to view the full results]
7490.
[Click the right-side scrollbar up]
7500.
The Troubleshooting dashboard offers a variety of actions for diagnosing functional operation of any item running in a NSX environment. Actions are available both for NSX devices as well as NSX service consumers such as virtual machines.

Next let's go back to the Main NSX dashboard.

[Click the tab 'NSX-vSphere Main']
7530.
Another key feature of the NSX management pack is Configuration Assurance checking, i.e. comparing the current NSX configuration against VMware Best Practice recommendations. If components of the NSX environment do not align with Best Practice guidelines, alerts will trigger to advise of the recommended settings for items that are out of compliance.

[Click the second open alert ('10.114.221.41 Backups are not using se...')]
7610.
VMware Best Practice recommends using Secure FTP to backup NSX Manager configurations. This environment is using insecure FTP, therefore triggering the warning alert and recommended remediation.

[Click the 'Home' button in the top left]
7750.
Let's look at one more Configuration Assurance check built into the management pack.

[Click the 6th open alert ('univ-dir-1 One or more OSPF areas on...')]
7900.
This rule checks whether OSPF on logical routers has been configured to use secure authentication.

[Click the arrow next to 'univ-dir-1 has symptom OSPF is enabled on the Logical Rounter']
7920.

The logical router 'univ-dlr-1' has not been configured to use the MD5 authentication protocol, hence the triggered alert. Expanding the 'OSPF is enabled...' symptom shows the date and time when OSPF was enabled on the logical router (3/17/16 at 1:53pm). The recommendation is to use MD5 authentication for all OSPF areas.

[Click the 'Home' button in the upper left corner]
8110.
[Click 'nsx nsbu' on the list of NSX-vSphere Environments']
8330.
The heatmaps on the Main dashboard provide a quick glance at the NSX transport layer health and workload.

[Click the arrow button by 'Configurations' on the 'Transport Layer' panel]
8370.
We can look at the transport layer from multiple health and workload angles, and also choose a heatmap for the vSphere Distributed Switches.

[Click 'Transport Node Network Workload']
8450.
This heatmap shows the current network workload of the transport layer is relatively light in the selected NSX environment.

Let's switch to a different NSX environment.

[Click 'nsx east' on the list of NSX-vSphere Environments']
8610.
[Click the bottom of the right-side scrollbar to view the rest of this dashboard]
8620.
The traffic charts display total throughput, egress and ingress statistics for the top consuming logical networks and VMs in the selected NSX environment.

Scroll to the right in the Traffic charts to view the Egress / Ingress stats.

[Click the right side of the bottom scrollbar under the panel 'Top Logical Networks by Traffic (KBps)']
8650.
[Click the left side of the bottom scrollbar under the panel 'Top Logical Networks by Traffic (KBps)']
8660.
[Click the right side of the bottom scrollbar under the panel 'Top VMs by Traffic (KBps)']
8670.
[Click the right-side scrollbar up]
8680.
Let's move to the vROps inventory trees to investigate workload, fault and capacity information for NSX components.

[Click the 'environment' icon in the top left]
8780.
Notice the individual inventory trees for various NSX objects such as Control Plane, Edge Services, Logical Routers and Switches, and Transport Zones.

Let's start with the Control Plane (NSX Manager and Controllers).

[Click 'NSX-vSphere Control Plane']
8930.
Expand the NSX environment to view Control Plane details.

[Click the arrow next to 'nsx east']
8950.
[Click 'nsx-mgr-east.mgmt.local']
8980.

Selecting the NSX Manager East reveals related Health, Risk and Efficiency alerts as well as three associated Controllers as child resources. We could select any of the sub-tabs on the right, i.e. Alerts, Analysis, Troubleshooting, etc., for further information on nsx-mgr-east. Notice the Risk alert indicating less than 3 controllers are active. Clicking on the alert would identify which controller is offline.

Next let's look at the other NSX inventory trees.

[Click the pulldown for 'NSX-vSphere Control Plane' in the top left]
8990.
[Click 'NSX-vSphere Edge Services']
9020.
[Click the arrow icon next to 'nsx nbsu']
9050.
[Click 'ESG-1']
9110.
Selecting an object in the inventory tree displays Health, Risk and Efficiency alerts for the selected objects and its descendants.

[Click 'ESG-1 NAT']
9190.

Selecting the NAT service for this edge device reveals an open alert indicating NAT rules with no destination. We could click that alert to identify the rules in question, but for this demo let's switch to the Analysis tab and look at Capacity information.

[Click the tab for 'Analysis']
9370.
In the Capacity Remaining Breakdown note the value 2,000 as the total rules capacity, with 3 rules consumed as the current peak value.

[Click 'ESG-1 Load Balancer']
9470.
Staying with the Capacity theme, notice the max values of 64 for Pools and Virtual Servers, with current Peak Values of 1 for each on this Edge device.

[Click 'ESG-1 L2 VPN']
9680.

Moving to the Faults tab for the VPN service we quickly see that the L2 VPN tunnel is down.

The inventory trees are useful for viewing performance, alerts and capacity for all NSX components and services.

[Click the pulldown for 'NSX-vSphere Edge Services' in the top left]
9690.

As mentioned at the beginning, the Management Pack for Network Devices (MPND) is included in the NSX-vSphere management pack. MPND discovers and maps the physical network infrastructure underpinning the virtualized environment. It also installs an inventory tree which enables analysis of the physical infrastructure of switches and ports.

[Click 'Physical Network']
9740.
[Click the arrow icon next to 'mpnd snmpv2' to expand]
9780.
The inventory tree shows a list of all the discovered physical network devices connected to the vSphere environments being monitored with vROps.

Let's drill into a Cisco Nexus Leaf Switch.

[Click 'w2m26-ccmm-nexus9372px-1']
9900.
Choosing the Troubleshooting tab and All Metrics, we can create a series of graphs to our liking to show statistics and properties of the selected physical switch.

Create charts for total egress and total overall traffic seen on this switch.

[Click the '+' button for 'Interfaces']
9920.
[Click 'Egress Traffic (Mbps)']
9990.
[Click the bottom of the scrollbar for the bottom center panel]
10000.
[Click the bottom of the scrollbar for the bottom center panel (again)]
10010.
[Click 'Total Traffic (Mbps)']
10120.

In summary, the NSX-vSphere management pack provides comprehensive physical to virtual network correlation, troubleshooting, configuration assurance and capacity modeling for NSX environments. It is a Must Have tool for every NSX implementation!

(End of this section)

[Click the Home icon to return to the start of the demo]