OnCommandInsight6.3Hands

Transcription

OnCommandInsight6.3Hands
OnCommand Insight 6.3
Hands-on Lab Guide
Dave Collins, Technical Marketing Engineer
August 2012
Contents
1
INTRODUCTION .................................................................................................................................... 4
BEGIN LAB .................................................................................................................................................. 5
2
INVENTORY AND ASSURANCE NAVIGATION AND VIEWS ............................................................. 5
2.1
INVENTORY .............................................................................................................................................................. 6
Storage Arrays .................................................................................................................... 6
Grouping ............................................................................................................................. 6
Switches .............................................................................................................................. 7
Paths ................................................................................................................................... 8
Analyze Path Violation......................................................................................................... 8
Virtual Machines and Data store........................................................................................ 10
3
4
5
6
2
ASSURANCE ....................................................................................................................................... 11
3.1
APPLYING POLICIES TO MONITOR CONFIGURATION AND PERFORMANCE .................................................. 11
3.2
FIBRE CHANNEL POLICY SETTINGS ................................................................................................................... 12
3.3
VIOLATIONS BROWSER ........................................................................................................................................ 14
3.4
PORT BALANCE VIOLATIONS .............................................................................................................................. 16
3.5
DISK UTILIZATION VIOLATIONS ........................................................................................................................... 17
PERFORMANCE.................................................................................................................................. 18
4.1
STORAGE PERFORMANCE ................................................................................................................................... 18
4.2
SAN PERFORMANCE (SWITCH PORT PERFORMANCE) .................................................................................... 21
4.3
CANDIDATES FOR HOST VIRTUALIZATION BASED ON ACTUAL PERFORMANCE. ........................................ 23
4.4
STORAGE ARRAY PERFORMANCE BASED ON SAN TRAFFIC.......................................................................... 24
4.5
STORAGE TIERING AND ANALYSIS ..................................................................................................................... 24
4.6
SWITCH ISL TRAFFIC VISIBILITY AND OPTIMIZATION ....................................................................................... 25
4.7
VIRTUAL MACHINE AND DATA STORE PERFORMANCE: TROUBLESHOOTING END TO END
PERFORMANCE ISSUES USING “ANALYZE PERFORMANCE” .......................................................................... 26
4.8
VM PERFORMANCE ............................................................................................................................................... 29
4.9
APPLICATION AND HOST PERFORMANCE ......................................................................................................... 29
PLANNING TOOLS.............................................................................................................................. 30
5.1
TASK AND ACTION PLANNING AND VALIDATION .............................................................................................. 30
5.2
SWITCH MIGRATION TOOL ................................................................................................................................... 33
DATA WAREHOUSE ........................................................................................................................... 35
6.1
INTRODUCTION AND OVERVIEW ......................................................................................................................... 35
6.2
PLAN - CAPACITY FORECAST DASHBOARD ...................................................................................................... 37
6.3
TIER DASHBOARD ................................................................................................................................................. 41
6.4
ACCOUNTABILITY AND COST AWARENESS. ..................................................................................................... 42
Insert Technical Report Title Here
7
6.5
UNCHARGED STORAGE ....................................................................................................................................... 45
6.6
IOPS VS. CAPACITY REPORTING IN THE DATA WAREHOUSE ......................................................................... 45
6.7
DIGGING INTO THE DETAILS ................................................................................................................................ 47
6.8
VM CAPACITY REPORTING................................................................................................................................... 50
CREATE AD-HOC REPORT ............................................................................................................... 53
7.1
HOW TO CREATE A CUSTOM SHOWBACK/CHARGEBACK REPORT USING BUSINESS INSIGHT
ADVANCED. ............................................................................................................................................................ 53
Steps to create this report ................................................................................................. 53
Adding variable costs per VM to your chargeback report. .................................................. 59
Adding Fixed Overhead costs to your chargeback report. ................................................. 63
Formatting and Grouping the report by Application and Tenant ......................................... 65
CLEANING up the report and running it. ........................................................................... 66
7.2
OTHER OPTIONS FOR AD-HOC REPORTS USING QUERY STUDIO .................................................................. 69
8
SCHEDULING REPORTS FOR DISTRIBUTION ................................................................................ 76
9
ENDING COMMENTS AND FEEDBACK ............................................................................................ 78
3
Insert Technical Report Title Here
THIS IS A LAB GUIDE FOR ONCOMMAND INSIGHT 6.3 ASSURE, PERFORM, PLAN AND
DATA WAREHOUSE (DWH).
Content Note: This lab is designed to help you understand the use cases and how to navigate
OnCommand Insight server and DWH to demo them. I did not call out each use case
specifically but provided a smooth flow to learn how to use and demo OnCommand Insight while
picking up the use cases along the way. It’s also designed so you don’t have to show the whole
demo but you can pick out particular areas you may want to show your customer. This is not
all inclusive nor is it designed to show you all features and functions of OnCommand Insight.
However, please note that each portion of this lab builds on the past steps so it is assumed that
you will learn how to navigate as you go through the lab. In other words we stop telling you how
to get somewhere but tell you to go there because you’ve stepped through in previous steps.
This is primarily built to run against the current OnCommand Insight (Demo_V6.3.gz dated
3/21/2012) and DWH demo db (UPdated_dwh_Demo_6.3.0_with_performance.gz dated
06/01/2012). Screenshots shown are accurate to those databases at the time of this publishing.
Your results may vary depending on your database but the functionality is still the same
regardless of the data. Total lab time is about 1-4 hours depending on experience. It’s a large
book but lots of pictures!!!  Your feedback is welcome. My address is at the end of this
document. Have Fun!!
1 INTRODUCTION
Using 100% Agentless technologies, OnCommand Insight automatically discovers all of your
resources and provides you a complete end-to-end view of your entire service path. With
OnCommand Insight you are able to see exactly which resources are being used and who is
using them. You can establish policies based on your best practices, enabling OnCommand
Insight to monitor and alert on violations that fall outside those policies.
OnCommand Insight is a “READ ONLY” tool that inventories and organizes your storage
infrastructure to enable you to easily see all the details related each device and how it relates to
the other devices in the SAN and the entire environment. OnCommand Insight does not
actively manage any component.
It is also a powerful tool for modeling and validating planned changes to minimize impact and
downtime for consolidations and migrations. OnCommand Insight can be used to identify
candidates for virtualization and tiering.
OnCommand Insight correlates discovered resources to business applications, enabling you to
optimize your resources and better align them with business requirements. You can use this
valuable information to help you reclaim orphaned storage and re-tier resources to get the most
out of your current investments.
OnCommand Insight provides trending, forecasting and reporting for capacity management.
OnCommand Insight enables you to apply your business entities for reporting on usage by
business unit, application, data center and tenants. OnCommand Insight provides user
accountability and cost awareness, enabling you generate automated chargeback reporting by
business unit and applications.
You can download a full customer facing demo video from:
OnCommand 6.2 Full demo with Table of Contents for easy viewing. (For best viewing,
download and unzip the zip file and play the HTML doc from your local drive)
4
Insert Technical Report Title Here
https://communities.netapp.com/docs/DOC-14031
BEGIN LAB
2 INVENTORY AND ASSURANCE NAVIGATION AND VIEWS
Let’s take a look at the discovery and inventory of storage, switches, VM systems, etc.
Log onto OnCommand Insight by selecting the OnCommand Insight ICON
the desktop.
Sign-on using Admin/admin123
From the main screen, select the dropdown ICON that looks like a calendar
from
at the top left corner of the screen and select
Inventory>Hosts. You should see hosts from the demo db.
You can use this menu to select Inventory, Assurance, Performance or Planning
categories OR to make it easier to navigate; you can activate the Navigation Pane as
follows: (recommend you use the Navigation Pane as it’s easier when you start.)
o Select Tools>OnCommand Insight Settings from the menu bar
5
o
Check the “Use Navigation Pane” box about half way down the right panel of
the settings box and select OK
o
Notice the full menu bar down the left side of the screen. This will make it easier
for you to navigate.
Insert Technical Report Title Here
2.1
INVENTORY
Expand the Inventory menu by clicking on the Inventory Button on the Navigation Pane
as shown below.
STORAGE ARRAYS
o
o
o
o
o
o
o
Select Storage Arrays from that menu. This opens up the MAIN VIEW of
storage.
Select the NetApp array called Chicago
As you see, Insight provides a full inventory including family model number and
serial number and other elements. USE the scroll bar to see more columns to
the right.
Note the Icons across the bottom of the screen. These are Microviews. You can
toggle them on and off to select more detail about what you have selected in
the main view.
Cycle through the micro views to get an idea of what is there.
Select Internal Volume and Volume Microviews (as shown above)
You can select the Column Management ICON to add and subtract more
columns from the view. Question: Which microview can you add the DeDupe
savings column?
GROUPING
You can group the information in any of the table views in main and micro views by
selecting the dropdown in the upper right corner of the table view and selecting a
grouping.
o Group the storage devices by model number. This enables you to see all of
6
Insert Technical Report Title Here
your arrays grouped by model number.
o Re-group the main view to NO Grouping.
Select the SEARCH icon from the top of the Main View. In the search window below
the main table, type NTAP and see the array selected. You can step through looking for
next or last occurrence of the word NTAP. Close the search window by clicking the X
in the red box on the left.
Search
SWITCHES
Using the same navigation, select switches from the Inventory menu. (no pictures
here) View the switches and get the names, IP addresses, model numbers and firmware
in the main view. View the ports and zone members and all other elements using the
micro views. You can use the Inventory menu to provide you with views of hosts,
storage, switches, VMs, etc. Let’s look closer at paths.
7
Insert Technical Report Title Here
PATHS
OnCommand Insight then correlates the paths. The paths are the logical and physical elements
that make up the service path between the application, VM and or host and its storage. Paths
include Fibre Channel, Cloud, iSCSI, NFS, Block, NAS, etc. Now, let’s set up the views shown
here and detailed steps below.
Select Paths from the Inventory menu.
Group by
Good Path
Bad Path
Topology microview
Group by Host then Storage. As you can see here from the fiber Channel prospective,
when we select a path were looking at the exact paths between the host and it storage
through the redundant fibre connections.
Select Topology micro view icon from the bottom of the Main view
Use Filter to find host kc_ordev7 by mousing over the top of the Host column until you
see a funnel. Then type kc in the Host column top cell.
Expand kc_Ordev7 and the array ports in the main view
Select the RED Violation
port of kc_ordev7
Select Zoning, Masking, Port micro views
In this particular case you can see that you have one green path which is good across
the fiber channel but the other path is blue. See legend on right of topology. WHAT IS
NOT CONFIGURED? (Hint. Blue and yellow make green)
ANALYZE PATH VIOLATION
OK, let’s analyze the violation:
8
Insert Technical Report Title Here
Right Click and select >>>>>
There are a couple ways you can analyze the violations. You can select violation
micro view from bottom icons or you can simply right click and select “Analyze
Violation” from the Red Path for kc_ordev7 in the main view. This opens a root cause
analysis screen below.
o
Expand the violation and change information in the lower pane. What is the
cause of this violation?
The violation tells you that the number ports this change from 2 to 1 which went against the
Redundancy Policy and cause violation. If you look down at the last changes that occurred you
can see that masking was removed which denied access from that host through its HBA
9
Insert Technical Report Title Here
to the FA port on the array and all the way to the volume. To fix the violation, the
administrator needs to reverse those changes and that violation automatically resolves itself.
Close the Analyze Violation screen.
VIRTUAL MACHINES AND DATA STORE
OnCommand Insight gives you a complete inventory of the VM environment. This is discovered
through a Data source configured to talk with Virtual Center. Details include all internal
configurations and elements of the VMs, ESX hosts, the technologies, as well as all the
information to correlate the path from VMs to its storage and details about how that storage is
configured to the disk including performance (if you have OnCommand Insight Perform License
installed). We’ll discuss performance later in this demo.
Select Datastore from the Inventory menu.
Don’t forget to use
the Microview
ICONS
Scroll down the main view and select DS-30 in the Main view.
Display the Topology and Virtual Machine micro views to show which VMs are using
Datastore DS-30.
Toggle through the micro view icons below to show the details of VM, VMDK,
capacities, the storage, the backend storage in the arrays and resources. You can
see full end to end visibility from the Host through a virtualized storage environment to
the backend storage.
Note: You can also select Hosts and Virtual Machines from the Inventory Menu and
cycle through the microviews noting the end to end visibility.
10
Insert Technical Report Title Here
3 ASSURANCE
3.1
APPLYING POLICIES TO MONITOR CONFIGURATION AND PERFORMANCE
Now that we've gathered the inventory and pull all this information into the Database, let's start
to apply policies so we can monitor and get alerts on violations.
We’ll talk about setting global policies, changes and the new violation browser. We’ll show how
we can analyze performance, talk about some port balance violations and disk utilization
violations.
Initially we set global policies within OnCommand Insight so we can monitor the environment
alert us when something falls outside those policies. There are several policies available.
Select Policy from the top menu bar
Select Global Polices
What thresholds can you set from here?
Select Violation Severity from the left menu of the global policy window
11
Insert Technical Report Title Here
What Severities can you set for each threshold?
Select Violation Notification. What are the possible violation notification options?
3.2
FIBRE CHANNEL POLICY SETTINGS
We set fibre channel policies to determine and keep track of path redundancy. We can set
options such as no SPF (single point of failure) or redundant. We can set it for minimum
number of ports on the host the storage and the maximum number switch hops. We can set
exceptions around different volume types that would not necessarily require redundancy like
BCVs, R1, and R2. We can also set policy exceptions around smaller volumes it wouldn't have
redundancy like EMC gatekeepers.
From the Policy menu on the menu bar Select Fibre Channel Policy
12
Insert Technical Report Title Here
What type of redundancy can you set from here?
What is the default number of ports?
Volume Type Exceptions:
What Volume exemptions can you select?
You can set redundancy policies on physical storage behind a virtualizer (Backend Path)
13
Insert Technical Report Title Here
3.3
VIOLATIONS BROWSER
Let’s take a look at the violation browser. The Violations Browser allows us to see the impact of
the violations on your business elements in one place and helps us manage the violations on all
of those global policies you saw above.
From the Assurance Menu on the left, select Violations Browser. (Note: At this point,
you might want to increase your viewing real estate here by closing down the navigation
pane on the left. To close the navigation pane, go to Tools>OnCommand Insight
Setting and uncheck the Navigation Pane box You can use the same process to turn it
back on later)
Back in the Violations browser; expand the All Violations explorer to reveal the violation
categories. This shows violations like Datastore Latency, Disk Utilization, Volume and
internal volume IOPS and response times, Port Balance violations, etc. These should
look familiar to you from the global policies that we just reviewed a few minutes ago.
As you can see, you can look at the violations all pile here as you can see over 12000
violations. (NOTE: Don’t let this scare you. Usually most violations are caused by
events that create multiple violations per event. You fix one event and a bunch of these
go away!) We can see detail on each of these violations by performing the following:
14
Insert Technical Report Title Here
Select the show violations impact
icon to view all the violations in context by
business entity, application, host, virtual machine, datacenter, etc.
Expand the Impacted Business Entity explorer and drill down to Earth Thermal
Tracking.
Expand and select Disk Utilization.
Sort the Description column descending. Now you can see Element, Description,
Severity and Violation type.
Select the top violation element called Disk DISK-14 of Storage Virtualizer.
In the Impact Details microview, toggle the Host, Virtual Machines, Applications and
Business Entities and Storage icons to view the details of the impact of the violations.
Here we see the impact on one application called City Limits, owned by one business
entity called Green Corp>Alternate Energy>Geothermal>Earth Thermal Tracking on
one host. However, 10 virtual machines are affected by this violation from one array.
The chart in the Violation Event microview shows the history and trending of the
utilization on this one disk over time. From here we can analyze the performance details
as we’ll see later in this demo.
REVIEW:
What are the categories that show impact of violations?
What Business Entity is impacted by these violations?
What is the Utilization of this disk?
Which hosts are being affected by this violation?
Which VMs are being affected by the violation?
We’ll do some troubleshooting using these violations and Analyze Performance later.
15
Insert Technical Report Title Here
3.4
PORT BALANCE VIOLATIONS
Let’s take a look at port balance violations. These are violations showing imbalance on SAN
traffic from hosts, Arrays and Switches. These are not performance related violations.
Using either the Navigation Pane or dropdown menu, Open Assurance
Select Port Balance Violations
Group by Type then Device
Expand Hosts and sort the Device column ascending
Select device Host nj_exch002. Note that this host has a balance index of 81. That
means the difference in distribution of traffic (or the load) between the HBA’s on this
host. Any index over 50 indicates significantly unbalanced ports on a device.
Select the Switch Port Performance microview. Note that over 88% of the traffic
distribution is going across one HBA in only 11% of the traffic going across the other. A
failure on the heavily used HBA could choke that application. This could indicate that
port balancing software is not configured on this server, not configured at all or not
installed.
Now collapse the Hosts and expand the Storage devices
Select various storage devices and view the traffic distribution in the switch port
performance microview to understand the balance across the storage ports.
These port balance violations provide valuable data on how your environment is configure
and optimized. It allows you to quickly determine where you need to optimize your
configurations based on actual usage. These are balance violations within each device not
necessarily traffic related performance violations. We’ll look at performance a bit later.
16
Insert Technical Report Title Here
3.5
DISK UTILIZATION VIOLATIONS
We looked Disk Utilization Violations in the Violations Browser a few minutes ago because we
got alerted to a violation. But if you didn't look at the error, you can go directly to disk utilization
violations here and troubleshoot your issues the similar to how we did in the Violations Browser.
The difference is the Violations Browser breaks down the violations by how it impacts your
Business Entities, Applications, Data Centers, etc. so you can troubleshoot by business
priorities, while Disk Utilization Violations lets you easily see your most critical utilization issues
and troubleshoot from the disk utilization violation back to the hosts. You can also add columns
to show applications and Business Entities if you want.
Let’s take a look how you can use Disk Utilization Violations to quickly identify and drilldown to
where your issue is.
Select Disk Utilization Violations from the Assurance Menu
Exceeded Threshold
Host with highest IOPS
Switch Traffic OK
Sort the Utilization column descending to bring your heaviest utilization to the top.
Here you see the utilization of each disk that exceeded the Disk Utilization threshold we
set earlier in our Global Policy. In relation to each violation, you see the disk, array,
hosts that access this disk, the date and time the violation occurred, the percentage of
utilization as well as IOPS and Throughput.
Select the disk with the highest utilization
Now select the Volume Usage of Disk microview to get details on volume usage and
performance.
Sort the Disk IOPS descending and select the top Volume Usage of Disk. Here I see
the volume with the highest usage along with the disk throughput and percentage info by
volume and host.
17
Insert Technical Report Title Here
Select Switch Port Performance microview. I see that my load appears to be balanced
(distribution column) across the storage ports so that’s most likely not a SAN or network
configuration issue.
Since this disk did cause a utilization violation, I can identify possible host candidates that
can cause the high utilization on the disk. OR I can see that the disk may have too many
volumes carved from it and I may need to spread that load out across more disks.
4 PERFORMANCE
OnCommand Insight provides performance information from end-to-end. This is different from
the violations we discussed above in that it provides pure performance information of Volumes,
Internal Volumes (FlexVols), Storage Pools (Aggregates) and Disks. Performance also
provides performance of VMs, Switches, ESX, Hyper-V, VMDK and Datastores. From here we
can use this to troubleshoot congestion, contention and bottlenecks; identify heavily used
storage pools, volumes and disks; SAN ports that are heavily used; possible candidates for
physical to virtual host virtualization and optimizing your storage and tiering.
From the Navigation Pane or the dropdown menu, expand the Performance menu.
Here we see that OnCommand Insight collects and shows you storage performance,
switch performance, data store performance, VM performance and even application
performance as it relates to storage performance.
4.1
STORAGE PERFORMANCE
From the Performance menu select Storage Performance.
Use Scroll Bars to see more
performance info in all windows
Sort Top Volume IOPS column descending (if not already done)
Select Array called Sym-0000500743… in the main view (should be near the top).
18
Insert Technical Report Title Here
Use the Horizontal slide bar in the main view to see the volume response times and
IOPS as well as disk utilization and IOPS columns (far right). Notice there is no Internal
Volume performance information because the EMC Symmetrix does not contain any.
We’ll look at a NetApp Array shortly to see Internal Volume (FlexVol) performance.
Now, in the Main View, notice the column called Volume Partial R/W. This indicates
there are volumes on that array that are misaligned. (we’ll see more detail later)
Select the microviews at the bottom to show you details of disk performance and
volume performance. Which microviews did you open? (hint: view below) Notice this
provides detailed throughput, IOPS, response times at the volume and disk level.
Close the Disk Performance microview.
Select the column customize
icon in the header of the Volume Performance
microview.
Use the vertical scroll bar to view all the columns that can be added or removed from
this report.
Select Partial R/W and the Storage columns and click OK. This adds columns to the
Volume Performance report showing you each Volume on each array that is misaligned.
(Note: you can get a complete list of all your misaligned volumes by selecting all the
arrays in the main view above.) Additionally, you can group the volumes by storage to
make it easier to view all the misaligned volumes across your entire enterprise by array.
(See figure below)
19
Insert Technical Report Title Here
Partial Read/Write indicated
volume mis-alignment
Now Select the Symmetrix-FAST array and toggle the Chart microview on. Here you
see OnCommand Insight showing EMC FAST auto-tiering. You can also see the
NetApp Hybrid Aggregates by selecting the We can also chart this performance over
time.
Notice FAST-T Volume
performance
20
Insert Technical Report Title Here
OnCommand Insight provides complete end to end performance views through virtualize
storage. Let’s take a look.
Select the storage array called Virtualizer from the main view. (Note: This is a V-Series
machine but OCI provides the same visibility through other virtualizers as well.)
Toggle on the Virtual Machine Performance microview.
Toggle the Backend Volume Performance and Datastore Performance microview on.
Use the slide bars at the bottom of the microviews to see more of the performance
columns in each view. (Whoops… there is a red mark in the Latency column! We will
analyze this later.)
I know this is a bit busy, but I wanted to demo to you that you can have deep performance
visibility from the VM, through the Datastore, to the frontend virtualizer array, through the
backend volumes. You can also drill down performance to the disks on that backend array and
you can also select the Switch Port Performance microview to visualize the performance on the
SAN so you can see very deep performance information from end to end. We will “analyze”
these performances from end to end a bit later.
4.2
SAN PERFORMANCE (SWITCH PORT PERFORMANCE)
Switch performance is the actual performance on the SAN at the switch.
Select Switch Port Performance from the Performance Menu. OnCommand Insight
knows whether the switches are connected to Arrays and Hosts so it shows you the
performance in context to the host or Array instead of the switch prospective.
21
Insert Technical Report Title Here
Using the dropdown at the top of the table, group the main view by “Connected Device
Type then Name”
Using the dropdown next to it, set the timeframe to “Last Week” and hit the refresh
icon to the right.
Sort the Distribution column Descending (arrow pointing down)
Expand Hosts
Expand hosts ny_ora1 and exchange_ny1
If you look at the Value and Distribution columns, you can see how HBA’s are balanced
on these hosts. On host ny_ora1, you see three HBA's that are balanced very well. But,
looking at host exchange_ny1, you see that one of your 2 HBA's has over 95% of the
traffic load on it while the other one has less than 5% of traffic. So you can see an
imbalance of the load across your HBA’s. Perhaps the multipath software is not
configured correctly, doesn't work, or installed but not turned on. However, also look at
the fact that one HBA is 4GB the other is 2 GB. The admins may have purposely
configured this host traffic to compensate for the slower HBA…
Select the Port Performance Distribution and Port Performance microviews to view
this analysis over time.
22
Insert Technical Report Title Here
Select exchange_ny1 from the main menu above. View the performance and
distribution of both HBA’s. If you select one or the other the performance and
distribution charts change to show you the details of what you’ve selected. This function
is the same throughout OnCommand Insight GUI.
4.3
CANDIDATES FOR HOST VIRTUALIZATION BASED ON ACTUAL PERFORMANCE.
This performance from the host SAN prospective shows you which are the busiest servers and
which are candidates for virtualization.
Toggle off the 2 performance charts
Collapse the expanded columns using the Collapse All Groups ICON
main view.
23
Insert Technical Report Title Here
on top of the
Expand Hosts again. Notice your busiest servers are at the top of the list.
Use the vertical slide bar to go to the bottom of the host list to see your least busy
hosts. As you see here, there are many hosts down near the bottom that have hardly
any traffic. Note: If you have virtualization project going on you can very quickly isolate
which physical hosts don't have much traffic to the applications and conduct your due
diligence on those applications for possible relocation to VM environment.
You can also use the same information here to choose which ESX hosts are good candidates to
move those applications based on how much traffic they are generating on the SAN.
4.4
STORAGE ARRAY PERFORMANCE BASED ON SAN TRAFFIC
We use the same logic and methods optimizing the traffic across the storage ports of the arrays.
Collapse the Hosts section and expand the Storage section.
You can see the busiest arrays at the top.
Expand storage array XP 1024 to see the traffic flow through the storage ports. In this
case, over 80% of the traffic is going across two of the six ports on the storage arrays.
Not very well-balanced. You can rebalance this traffic OR using this information, you
can select a lesser used storage port to provision your NEXT Tier 1 application too. This
helps you intelligently provision and optimize your environment using real traffic analysis.
4.5
STORAGE TIERING AND ANALYSIS
Similarly, like you did with the hosts, you can see the storage arrays that are NOT so busy.
Scroll to the bottom of the Storage Array list.
24
Insert Technical Report Title Here
There are several expensive tier 1 Symmetrix and other arrays at the bottom of this list
that have very little traffic accessing the arrays. These arrays *may* have lots of data on
them but nobody's using them. Armed with this information, you could take a look at the
application data on these expensive tier 1 arrays and move the applications to less
expensive tier 2 or tier 3 arrays OR archive it. Then you can decommission or
repurpose these expensive arrays. (LOTS of ROI potential here)
4.6
SWITCH ISL TRAFFIC VISIBILITY AND OPTIMIZATION
OnCommand Insight shows you only the ISL’s (Inter Switch Links) under the switches category.
Collapse the Storage category in the main view and expand the Switch category.
Expand Switch 78 and hcis300.
25
Insert Technical Report Title Here
As we saw with Hosts and Arrays, we can see exactly how well balanced the traffic is across the
iSL’s. Switch hcis300 is well balanced, but on Switch 78 we see that 90% of the traffic is going
across one switch link and only 9% of the traffic across the other. If this is a trunk, this is
severely out of balance.
We also see which are the busiest and least busy switches. This allows us to balance out
(optimize) our environment as well as weed out the least busy switches.
4.7
VIRTUAL MACHINE AND DATA STORE PERFORMANCE:
TROUBLESHOOTING END TO END PERFORMANCE ISSUES USING “ANALYZE
PERFORMANCE”
Let’s put all this performance information to good use!
USE CASE: I may have gotten a call from a user that was complaining that the application on
VM 70 is running slow or I may have received the alert from the threshold being breached. Let’s
troubleshoot the problem
Select Virtual Machine Performance
Select Custom from the “Timeframe” dropdown menu next to the grouping menu. Enter
the dates January 1, 2012 through now.
Then hit the Green Recycle button next to the dropdown.
Sort the VM Disk Top Latency column descending to get the longest latency at the top.
Here we are seeing that in fact VM-70 does not appear to have any performance issues but we
do see very high CPU, Memory and Data Store Latency on VM-60 and VM-61.
Look at column 2. The common factor between Vm-70 (the user complaint) and VM-60 is DS30.
Open the Datastore Performance microview to validate the high latency time.
26
Insert Technical Report Title Here
Right click on VM-60 and select Analyze Performance
This opens an analysis of everything associated with VM-60 and DS30.
See the tabs across the top of the window. Each of these tabs provides in-depth visibility into
performance within each category.
Selecting the Disk tab, I see that although I have a few high “top” utilization, overall
utilization and IOPS are relatively low so that can rule out a hot disk issue.
27
Insert Technical Report Title Here
Select Volumes tab and Internal Volumes tab I see there are some relatively high Top
Response times but still very low IOPS which tells me other factors are affecting
response time and the slowness of the application on VM-70.
Select Backend Volumes we see the storage is virtualized and we can see the
performance on the backend volumes here. Here I see some possible higher IOPS but
still no glaring issues in performance.
To make sure I don’t have a SAN problem I select the Switch Performance tab. It
shows an imbalance between the 2GB HBA’s on ESX1 where VM60 and 70 are and a
potential optimization or outage issue but no gridlock.
Select Hosts tab. This tab shows me that host ESX 1 is the same host that holds VM60 and VM-70. VM-60 appears to be causing very high CPU and Memory usage which
is causing contention with time sharing during disk access thus creating a high Disk
Latency. But the Disk IOPS are still very low.
28
Insert Technical Report Title Here
Deduce that VM-60 is probably not sized right for the application that is driving it hard. This is
probably what’s causing the disk latency issue. So chances of a disk issue are slim.
4.8
VM PERFORMANCE
VM Performance helps you troubleshoot the same scenarios. Here you can understand what's
going on the whole environment.
Select Virtual Machine Performance
Sort column Top Disk Latency in descending order so the largest latency rises to the
top. In this case, VM 61 here is chewing up a lot of memory and a lot of CPU time but
using Low Disk IOPS. The VM appears to be causing the latency issues.
Select VM-61. You can open a micro view and see the VMDK performance as well.
Add chart microview.
You can also a break it out by volume performance and data store performance thus
giving you a more holistic picture of the environment and help you troubleshoot to
resolution.
The takeaway is you can troubleshoot performance issues from many different angles and go in
many different directions to quickly narrow down the problem.
4.9
APPLICATION AND HOST PERFORMANCE
You can add to any of these performance views your applications and hosts to help you
understand how your performances affecting your applications. That is important to the
business customer. You can drill down and understand where the performance issue is by
showing you visibility from the application all the way to the disks.
Scroll down to ESX1
Use the horizontal slide bars in the main and microviews to see performance info.
29
Insert Technical Report Title Here
OnCommand Insight shows you performance from the host prospective all the way back to the
storage but remember it does not have agents on the host so it cannot show you the details of
the performance on the host itself.
Review questions:
What is the value of Analyze Performance?
What are the areas we can view performance matrix under Analyze Performance?
5 PLANNING TOOLS
5.1
TASK AND ACTION PLANNING AND VALIDATION
OnCommand Insight as 2 plan tools to help you plan, validate and monitor changes in your
environment. One is a change management tool and the other is a migration tool for switches
only.
The change management tool (or What-IF) helps you to create tasks and actions within those
tasks using a wizard. It helps you logically configure changes that you need to make, test and
validate those changes before you make them, and monitor the progress of changes as you
make them. This significantly reduces your risk when making changes because you can pretest
them before you make any actual changes in your environment.
NOTE: Remember, OnCommand Insight is a READ ONLY tool so it does not perform any
active tasks. Use it in the planning, validating and execution monitoring of your change
management.
30
Insert Technical Report Title Here
Select Planning Menu
Select Plans to access the tool
Select the task ID oadmin 01.08/2007 – Replace HBA Clearcase1
Notice the Actions list for the task. These are generated by you to help you logically and
accurately list out the tasks.
To add more actions, simply right click in the action area and select “Add Action”
In the new action window, scroll down and select the action you want to perform. You
can add description and other parameters then select OK.
31
Insert Technical Report Title Here
Then you can pre-validate the actions to ensure you know the results of each action
BEFORE you actually perform the task. To do this, right click the task and select
Validate task.
As you see below, OnCommand Insight validates each action against the current
configuration in your environment to validate what has been completed correctly
(GREEN CHECKMARK), what is not completed (BLANK BOX) and what is not
completed correctly (RED X)
32
Insert Technical Report Title Here
When you build the action list, OnCommand Insight automatically compares your
planned changes to your existing environment and anticipates any future violations that
*could* occur if you made these changes without correcting planned actions OR
violations that already exist in your environment.
Once you complete creating your list of action items, you can right click and validate the actions
as many times as you want until completed. OnCommand Insight validates every one of these
actions. It will show you whether the actions are complete correctly, done wrong, or not
completed at all. It gives you a preview into potential issues before you make the changes thus
lowering your risk.
5.2
SWITCH MIGRATION TOOL
The migration tool provides you with instantaneous visibility into all of the environment and
business entities that will be affected by a migration to new or updated switches. Say you want
to just update the firmware on a switch. What-if… it goes down in the middle of the upgrade?
What does it affect in your environment? Knowing this ahead of time can reduce your risk by
giving you the complete picture of whom and what will be affected by the interruption.
The Migration tool allows you to tell OnCommand Insight which switches that you want to
upgrade or replace. Because OnCommand Insight knows all hosts, storage arrays, volumes,
business units, applications that are affected by this change, it can provide you with the current
violations as well future violations that will occur when the switches pulled out. This enables
you to validate the total impact of the changes you want to make BEFORE you make them so
you can reduce your risk by fixing issues before they occur.
NOTE: Remember, OnCommand Insight is a READ ONLY tool so it does not perform any
migration tasks. Use it in the planning and execution monitoring of your migration.
Under Planning menu, select Migrations. This shows you the migration tasks already
created and the existing impact on your business entities of existing proposed changes
here.
To add a new task right click on the task area and select Add Task
33
Insert Technical Report Title Here
Complete task details above and click Next to select Switch (s) to migrate
Select the switches to be updated or replaced and click Finish
Select the new task in the main screen and use the microviews to provide you with
affected paths, impact and quality assurance views
34
Insert Technical Report Title Here
Using this information you can speed up the time for you to migrate switches because it cuts the
due diligence time and lowers your risk because you know the impacts before you take any
actions.
NOTE: Remember, OnCommand Insight is a READ ONLY tool so it does not perform any
migration tasks. Use it in the planning and execution monitoring of your migration.
6 DATA WAREHOUSE
6.1
INTRODUCTION AND OVERVIEW
Let’s introduce you to the data warehouse. We’ll talk about the Datamarts and navigation and
then we’ll go into the reports and we’ll finish by showing you how to create Ad-hoc reports using
Query Studio.
The data warehouse is made up of several Datamarts. Datamarts are sets of data that relate to
each other.
Open a browser and go to http://localhost:8080/reporting
Log on using admin/admin123
If you receive this page, uncheck the “show this page…” and select My Home
35
Insert Technical Report Title Here
Data warehouse (DWH) Home Page Public Folders
Public
Folders
DataMart
and
folders
The data warehouse has several built-in Datamarts thoroughout the data warehouse (DWH).
Above, you see the 3 primary Datamarts called Chargeback datamart, Inventory Datamart and
Storage Efficiency Datamart. Additionally, we have two folders which contain other Datamarts
for Capacity and Performance.
Select the Capacity 6.3 folder
As you can see, there are other Datamarts that are Capacity related. Datamarts including
Internal Volume, Volume, Storage and Storage Pool, and VM Capacity Datamarts provide you
with easy to use data elements related to those specific catagories making it easier for you to
use the existing reports but more importantly help you create your own custom reports using
drag and drop technology we’ll show later in this lab.
Select the Storage Capacity Datamart
36
Insert Technical Report Title Here
There are 4 folders located within EVERY Datamart. Most built in reports are in the Reports folder. Any
custom reports you create, you MUST save them in the Customer Report or Customer Dashboard folders
in order to save them during Upgrades.
Select Dashboards. (notice the BREADCRUMBS to help you navigate)
Navigation
Bread crumbs
Folders
Dashboard
Reports
Which dashboards are located in the folder?
6.2
PLAN - CAPACITY FORECAST DASHBOARD
The data warehouse has over 200 built-in dashboards and reports. Let’s take a look at a few.
The capacity forecast dashboard provides a history of how storage is been used as well as
trends and forecast out to the future. It shows by data center and by tier.
Select that Capacity Dashboard. This may take a bit of time to paint so be patient.
The capacity forecast dashboard provides you with trending and forecast of your
capacity across your entire environment. NOTE: your data in the picture may vary
depending on the demo db you are using and the date (because it’s a trending chart)
While we are at it, let’s also stage the tiering dashboard in a new window by holding the
shift key and selecting the Tiering Dashboard so we can discuss it is well in a few
minutes
37
Insert Technical Report Title Here
When it first opens, you see in the upper left, the Capacity Consumption Forecast report by
Datacenter and Tiers. The initial view shows how much storage is left in each datacenter by tier
before it reaches 80% (adjustable by user) of capacity. The graph on the right depicts the
usage trending and forecasting over time. The “Reset Selection” button resets the graphic to
show storage trending across the entire enterprise.
Select Tokyo/Gold-Fast block on the matrix. Notice the graph at the right changes to
reflect the storage consumption, trending and forecasting for that tier at that datacenter.
38
Insert Technical Report Title Here
Reset the Storage Capacity Trend chart by clicking the Resets Selection next to the
Matrix chart. The chart on the right will show the trending and forecasting for the entire
enterprise.
Scroll down the dashboard to view the list of reports on the right side. Each of the
dashboards has a list of related reports on the lower right hand side. You can select
from any number of different reports to provide the detailed information that you need.
The dashboard also contains some dial graphics showing you storage consumption and
capacity in your enterprise and each datacenter.
Continuing down the left side of the Dashboard, these charts show you business level
storage consumption by business entities. Here we can drill down to see usage by
tenant, Line of Business, Business Unit, and Projects
39
Insert Technical Report Title Here
Right click in this graphic and you can drill down to view storage usage by line of
business, drill again to business unit and by project.
As you can see, you get really detailed information on consumption by your business entities
from Tenant, LOB, Business Unit, Project and Application in a very quick form.
40
Insert Technical Report Title Here
6.3
TIER DASHBOARD
Let’s take a look at the tiers dashboard that we opened up a few minutes ago by selecting it
from the tabs at the bottom of your Windows screen.
Note: Your data may vary depending on the database used for this demo.
This dashboard gives us a different perspective on how storage is growing and how is being
used. As you see it looks like gold tier has remained relatively stable over the past few months
while gold-fast storage, which is more expensive storage, has grown considerably over the past
couple months. This tells you how your tiering initiatives are progressing. Bronze, which is
hardly grown at all, could be an indication that we’re spending too much money on storage.
You might want to review your storage usage using OnCommand Insight to see how the storage
is being consumed and by whom.
Scroll down. Let's look a little closer. OnCommand Insight shows storage usage by
business units, applications and by tier. This enables you to understand how storage is
being used. You can also by data center, tier, and business entity.
As we did in the last report you can right click and drill down and look at consumption by
tenant, line of business, business unit, project and application. You can understand how
your data is being consumed at multiple levels and aspects.
Select the “Return” ICON at the top right of the Tier Dashboard to return to the folder.
41
Insert Technical Report Title Here
There is a new Storage Tier Report located in the Storage and Storage Pool Data Mart. Let’s take a
quick look at it.
Use the BreadCrumbs to navigate back to Capacity 6.3 folder.
Then select Storage and Storage Pool Capacity DataMart and Reports folder.
Next select the Storage Capacity By Tier Report to view the report below. This report shows your
capacity by tier and how it trends over time. It also provides a great detail and summary report at the
bottom showing each Array, tiers and how much capacity is used and the percentage. (lots of
information on a single report)
6.4
ACCOUNTABILITY AND COST AWARENESS.
The standard data warehouse chargeback reports are more about accountability than all
chargeback. We’ll show you this now. We’ll also show you how to create your own powerful
“custom” chargeback/showback reports using Business Insight Advanced later in this lab!
Select Public folders in the breadcrumbs on top left of the Data warehouse window.
Select Chargeback Datamart.
In the chargeback Datamart and select reports folder to access various reports that
show capacity and accountability
42
Insert Technical Report Title Here
Chargeback DM
Select Capacity Accountability by Business Entity and Service Level Detail. Here
you have the option to customize this report to your needs by selecting service levels,
resource types, applications, and host and storage names. You also have the option of
selecting the business entity by using the dropdown to select any or all of the business
entities and projects.
Select all in each category to give you a good representation of the in-depth reporting.
Then click finish.
43
Insert Technical Report Title Here
The report provides a very detailed version of the capacity utilization by business entity,
application, and the host it's running on, the storage array, the volume, the actual provisioned
and used storage. This report is grouped by business unit as well as applications; this provides
you with a good representation of who's using what storage.
Note the scroll bar for scrolling on page 1 and they you can also use the Page UP/Page Down
links at the bottom to go to page 2, etc…
Select the Return Icon in the upper right to return to the folder of reports.
44
Insert Technical Report Title Here
6.5
UNCHARGED STORAGE
You can also generate reports that help you understanding what storage is NOT being
accounted for.
Select “Capacity Accountability by Uncharged Capacity per Internal Volume”. This
provides you with a complete listing by array and volumes, and how much storage is not
being charged or accounted for!
You get FULL accountability of which storage is being accounted and which storage is NOT
being accounted for across the entire enterprise regardless of storage vendor.
6.6
IOPS VS. CAPACITY REPORTING IN THE DATA WAREHOUSE
Let's look at performance versus capacity and orphaned storage by last access. This adds
another dimension into how your storage is being used.
Open the Performance Datamart. (hint, use breadcrumbs to select Public Folders and
then select the Performance Datamart)
Select the Internal Volume Daily Performance folder. This provides really good
pictorial view of how your storage being used.
45
Insert Technical Report Title Here
Select reports and select Allocated, used, internal volume Count by IOPS Ranges.
This provides capacity versus IOPS report which is very interesting.
Select Last Year time period
Select All Storage models and Tiers and Click Finish
Select all arrays and all tiers to give you a full view of how your storage is being used
(or not being used…)
Looking at the results. Remember this is storage accessed over the past year. The resulting
report shows you all the storage that has (or has not) been accessed over the past year.
46
Insert Technical Report Title Here
Storage access for a year
Over 7300 Vols
Over 3.4 PB
As you see from the first bar, there are over 7300 volumes that have not been accessed over in
the past year. If we look at it in terms of size, over 3.4 PB has had zero accessed the past year.
Note: this actually real customer but the names have been sanitized.
You can see how impactful this is. There is over 3.4 PB of storage that has had zero use for a
year! This information enables you to start making business decisions on the storage how to
better understand how it’s being used so you can reclaim and re-purpose some of that storage.
(Talk about ROI!!!)
6.7
DIGGING INTO THE DETAILS
These charts are really nice but you need the detail to effectively work on identification and
recovery. OK, let's go look at the underlying details.
Go back to Volume Daily Performance 6.3, folder and drill down to Reports. (hint it’s
in the Performance Datamart )
Select the Array Performance report. This gives you a complete breakdown of the
performance for all storage from the arrays all the way down to the volumes.
Select one year and set the IOPS parameter you want to filter on. (I usually start at
default)
This report starts with the Orphan Summary.
Select Page down to view the Storage array summary.
47
Insert Technical Report Title Here
As you see this is pretty high-level. It shows the total amount of raw and allocated
capacity in each Storage device vs. the total IOPS and the max IOPS actually used over
the past year. This tells a very compelling story but it’s still high level.
Page down a few pages and reach the bottom of this section. You see a Glossary
terms explaining the column headings.
48
Insert Technical Report Title Here
Now continue to page down to the Host tables. This shows you the hostname, the
raw and allocated capacity by host, and the IOPS accessed over the past year. This is
more detail than the Storage tables above.
Page down past the host tables and you look at the orphaned volumes prospective.
Here is a great deal of details that you can use. These are all the volumes that have not
been accessed in a full year. It shows you the Array name, volume, capacities,
hostname as well as the applications and tiers that have not been accessed in the last
year.
Page down to the “Volume by IOPS tables (may be several pages down). This shows
you the storage array, volume, capacity, host, application, tier, Max and total amount of
IOPS. So we can say it's a pretty well-rounded report that shows you actual usage (or
lack thereof) so you can go reclaim the storage that is not used.
49
Insert Technical Report Title Here
6.8
VM CAPACITY REPORTING
There are several different reports in the VM capacity Datamart.
Navigate to the VM Capacity 6.3 Datamart.
As you see we have several reports built-in here already.
Select VM Capacity 6.3 and then navigate into the Reports folder
Select VM Capacity Summary.
50
Insert Technical Report Title Here
Select all so we see the VM Capacity across the entire Enterprise (spanning multiple
vCenters.
The results show all the VMs, their capacity, the data store, the actual capacity, the VM names
in the provisioned storage, and commit ratio of each VM across your entire environment.
NOTE: I paged down to the bottom so you can see the total storage and % commitment across
your whole enterprise plus a glossary of terms.
Select the “return button” in the upper right corner of the report (looks like a left turn
arrow)
51
Insert Technical Report Title Here
Next select Inactive VMs report to show you VMs that have not been accessed in a
defined period of time (default 60 days).
Set this time threshold and click Finish.
This is an excellent report showing you which VMs are powered off and how long they have
been powered off as well as how much capacity each one of those is holding that nobody else
can use. It gives you all the details including the datacenter, VM, OS, ESX host, cluster, VMDK
and how long it’s been powered off. Armed with this information you can go recover these VMs
cover can reclaim storage...
52
Insert Technical Report Title Here
7 CREATE AD-HOC REPORT
Let’s show you how easy it is to create custom reports in the data warehouse.
7.1
HOW TO CREATE A CUSTOM SHOWBACK/CHARGEBACK REPORT USING
BUSINESS INSIGHT ADVANCED.
Below is a great example of the custom Chargeback or Showback Report that you will
create. It shows usage by Business Entity and Application including variable cost of VM
based on configuration, fixed Overhead and Storage usage.
STEPS TO CREATE THIS REPORT
Watch a video on how to create this report Note: You need a user name and password for this
community. To obtain them, click the Become a Member link.
OnCommand Insight Reporting Portal is accessed through http://<reportingserver>:8080/reporting
Enter User name and Password credentials.
From the Welcome page, select My home.
From the Launch menu (at the top right corner of the OnCommand Insight Reporting
portal), select Business Insight Advanced.
From the list of all packages that appears, click on the Capacity <version> folder
and then click on VM Capacity <version>.
Create a new report by selecting New from the dropdown in the upper left corner or
Create New if you are on the Business Insight Advanced landing page.
53
Insert Technical Report Title Here
From the pre-defined report layouts New pop-up, choose List and click OK.
In the lower right pane, select the Source tab and expand Advanced Data Mart
from the VM Capacity package.
From the Advanced Data Mart, expand Business Entity Hierarchy and Business
Entity and drag Tenant and place it on the report work area.
Collapse Advanced Data Mart and expand Simple Data Mart
From Simple Data Mart, drag Application and place it on the report work area to
the right of the Tenant column. (TIP. Make sure you place it on the blinking gray
BAR on the right of the previous column or it will give you an error)
Now we are going to drag multiple columns to the palate to save time building the report.
We will be reporting on the total # of processors (cores) and the memory that is
configured for each VM. So, let’s grab the following elements from the VM
Dimension under the Advanced Data Mart:
From Advanced Data Mart, expand VM Dimension
54
Insert Technical Report Title Here
Select the next columns IN THE FOLLOWING ORDER.
From Advanced Data Mart>VM Dimension, hold the control key and select the
following columns (in order):
o VM Name
o Processors
o Memory
Click and Drag VM Name and place it on the report work area to the right of the
Application column. NOTE: All the columns should follow in the order you selected
them similar to screenshot below. (your data will differ but columns will be the same)
Now let’s bring Capacity information onto the report.
From Simple Data Mart, hold the control key and select the following columns (in
order):
o Tier
o Tier Cost
o Provisioned Capacity (GB)
Click and Drag Tier column and place it on the report work area to the right of the
Application column. NOTE: All the columns should follow in the order you selected
them similar to screenshot below. (your data will differ but columns will be the same)
55
Insert Technical Report Title Here
To create a summary of cost per GB, hold the control key and select Tier Cost
and Provisioned Capacity (GB)
Then Right Click the Provisioned Capacity Column and select Calculate and
select the multiplication calculation.
Business Insight Advanced has created a new column for you, completed the calculations and put it
in the report.
Next let’s format and re-title the column name.
Right click on the new column head and select Show Properties
56
Insert Technical Report Title Here
In the lower Right corner, scroll down to the bottom of the properties box and
select the ellipsis on the Data item name box. Change the name to Storage Cost
and click OK.
Note the column heading is now Storage Cost.
Now select one of the numeric values in that column and select Data Format
ellipsis from the properties box in the lower right corner.
From the Data Format dialog box select currency from the Format type dropdown.
As you see from the Properties dialog box there are lots of options you can set to
format the currency numbers in this column. The default is USD so let’s just click OK
to set the default. You will see the column reformat to USD.
57
Insert Technical Report Title Here
Here is our current report. Let’s filter out storage that is NOT being charged.
Select any BLANK cell in the Tier Cost Column and click on the filter ICON in the
top toolbar.
Select Exclude Null
Here is our current report. Notice all the cells that had NO cost associated with those tiers are
deleted leaving you with only the storage that has charges associated with it. (TIP: in another report,
you can actually reverse the logic and show only storage that is NOT being charged as well…)
You can also format the Tier Cost column with USD currency as well if you want.
58
Insert Technical Report Title Here
OK, that was easy, but not complete. Let’s add other cost factors into your chargeback report for cost of
VM Service Levels by configuration and fixed overhead costs used by each application.
ADDING VARIABLE COSTS PER VM TO YOUR CHARGEBACK REPORT.
Let’s say the customer wants to charge per VM based on the # of CPUs and Memory it’s configured with.
To do that, we need to first create a VM Service Level comprised of the # of CPUs and Memory
configured for each VM. Then allocate cost per Service Level.
To create a VM Service Level, we are going to drop in a small conditional expression to build the Service
Levels per VM. This is an easy example of the flexibility of Business Insight Advanced in creating reports.
(DON’T panic, you can skip the conditional expression and just put a fixed cost on each VM if you want.
See the Overhead example later on… but humor me here in this lab.)
Select the Tier Column to place where we want to insert the new columns.
Select Toolbox tab at lower right corner and double click the Query Calculation
ICON.
In the Create Calculation Dialog box name the column VM Service Level and
select Other Expression and click OK
59
Insert Technical Report Title Here
In the Data Item Expression dialog box, copy and paste the following VM Service
Level conditional expression into the Expression Definition box and select OK
(note: if you are remoted into the OnCommand Insight server, you may have to
create a text document on the OnCommand Server desktop to cut and paste
this into, prior to pasting it into the Expression box!)
Below is an example of the conditional expression that gives you the if-else condition for
VM Service Level
IF([Processors] = 2 AND [Memory] < 2049)
THEN ('Bronze')
ELSE (IF([Processors] = 2 AND [Memory] < 4097)
THEN ('Bronze_Platinum')
ELSE IF([Processors] = 4 AND [Memory] < 8193)
THEN ('Silver')
ELSE IF([Processors] = 4 AND [Memory] > 8193)
THEN ('Silver_Platinum')
ELSE IF([Processors] = 6 AND [Memory] > 8191)
THEN ('Gold')
ELSE IF([Processors] = 8 AND [Memory] > 16383)
THEN ('Gold_Platinum')
ELSE ('tbd'))
60
Insert Technical Report Title Here
Business Insight Advanced will validate the conditional expression (nice to know if
you got it right) and create the column called VM Service Level and populate it
based on the query. (If you get an error, your Conditional expression probably has
a syntax or other error)
You will see a new column added called VM Service Level with the various Service
Levels for each VM based on the # of CPU’s and memory each has. (At this point
there may be duplicates in the list but not to worry. We are not finished formatting or
grouping the report).
Next, let’s add a column that calculates the cost per VM based on Service levels we just established.
Select Toolbox tab at lower right corner and double click the Query Calculation
ICON.
In the Create Calculation Dialog box name the column Cost Per VM and select
Other Expression and click OK
In the Data Item Expression dialog box, cut and paste the conditional expression for Cost of VM
(below) in the Expression Definition box and select OK
61
Insert Technical Report Title Here
Example of conditional expression for Cost per VM
IF([VM Service Level] = 'Bronze')
THEN (10)
ELSE(IF([VM Service Level] = 'Bronze_Platinum')
THEN (15)
ELSE IF([VM Service Level] = 'Silver')
THEN (20)
ELSE IF([VM Service Level] = 'Silver_Platinum')
THEN (25)
ELSE IF([VM Service Level] = 'Gold')
THEN (40)
ELSE IF([VM Service Level] = 'Gold_Platinum')
THEN (55)
ELSE (30))
You will see a new column added called Cost Per VM with variable costs for each
VM based on the Service Level.
Next format the data in the Cost per VM column to USD currency as you did above.
62
Insert Technical Report Title Here
ADDING FIXED OVERHEAD COSTS TO YOUR CHARGEBACK REPORT.
Let’s say the customer has determined that the total cost for Overhead (including items like heat/AC,
Floor space, Power, Rent, Operations personnel, helpdesk, etc.) is $24. Per VM. Let’s create a
column called Cost of Overhead and apply this fixed cost. (note: you can do this for any fixed costs
rather than use SQL as well..)
Select Toolbox tab at lower right corner and double click the Query Calculation
ICON above.
In the Create Calculation Dialog box name the column Cost of Overhead and
select Other Expression and click OK
In the Data Item Expression dialog box, enter a cost of 24 in the Expression
Definition box and select OK
You will see a new column added called Cost of Overhead with 24 for each VM.
(Note: at this point there may be duplicates in the list but not to worry. We are not
finished formatting or grouping the report).
63
Insert Technical Report Title Here
Next format the data in the Cost of Overhead column to USD currency as you did
above. Then drag the column header and drop it to the right of the Storage Cost
column as shown below.
Subtotaling, naming and saving the report
Now that we have a cost per VM, overhead and the cost per storage usage by Tenant, Application
and VM, let’s sum the total costs and finish formatting the report by Tenant and Application.
Hold the control key down and select a numeric cell in the Cost per VM, Storage
Costs and Cost of Overhead columns. Right click one of the numeric cells and
select Calculate and the add function for the three columns.
This will create a new column called “Cost per VM+ Storage Costs + Cost of Overhead” and
calculate each row.
Now format the column for USD currency, and retitle the column to “Total Cost of
Services”
Name the report “Total Storage, VM and Overhead Cost by Tenant and
Application Chargeback (Showback) “by double clicking the title area.
64
Insert Technical Report Title Here
Now save it to the Customer Report folder using the same name.
FORMATTING AND GROUPING THE REPORT BY APPLICATION AND TENANT
We are not done yet. Now we need to format the report by grouping, subtotaling and totaling by
tenant and application.
Hold the control key down and select “Cost per VM, Provisioned Capacity,
Storage Costs, Cost of Overhead and Total Cost of Service” columns
Select the Total ICON from the Summary dropdown ICON
65
Insert Technical Report Title Here
If you Page down to the bottom of the report you will see total columns. We’ll clean
up the summary Rows in a minute.
Let’s Group the report by Tenant and Application
Hold the control key down and select Tenant and Application columns
Select the Grouping ICON from the top toolbar.
CLEANING UP THE REPORT AND RUNNING IT.
To Clean up the report right click and delete the Summary ROWS. (not columns)
66
Insert Technical Report Title Here
Then Go to the bottom of the report, hold control key and select both Summary
Rows and right click and delete them. (Leave the TOTAL rows)
Save the report.
Now let’s run the report to see how it looks:
Select the Run ICON from the toolbar and run the report as HTML. (note the formats
you can run it in if you want…)
The report will show like this in its final format. I’ve paged down in the report below to show you
subtotals and you can page to the bottom and see the totals by Company and Total of all resources
charged.
67
Insert Technical Report Title Here
These reports are extremely flexible to do what you need. Notice the drill down
link in the Tenant column (pictured above in the red circle). If you click on the
LINK you will drill down from Tenant to Line of Business then to Business Unit, etc.
If you Right Click on the link you can drill up as well.
You can now schedule this report to run and distribute in various formats like any other OnCommand
Insight Data Warehouse report
Remember, now that you have created this report, every time you run it will provide the latest in usability
information. You can automate this report by scheduling it to run and email to recipients etc. Lots of
flexibility…
68
Insert Technical Report Title Here
7.2
OTHER OPTIONS FOR AD-HOC REPORTS USING QUERY STUDIO
You can also create simple Ad-hoc reports by using Query Studio. A very simple example is
shown here:
Log onto the data warehouse using Admin/admin123 (must be logged on as Admin to
use Query Studio)
From Public folders, select the Chargeback Datamart.
Select Launch menu in upper right corner of the view and Select Query Studio.
The Datamart is set up in “simple Datamart” and “advanced Datamart”. Simple DM contains
the elements that most users use for reports. The Advanced DM contains all the Facts and
Dimensions for all the elements. At this point we’ll create this report using the simple DM to
show you how easy it is.
69
Insert Technical Report Title Here
Expand the Simple DM and do the following:
Click and drag the Business Unit to the pallet
Click and drag the Application element to the pallet. You see the applications line up
with their proper Business Units.
Click and drag Tier over to the Pallet to organize the storage usage by tier.
Click and drag “Provisioned raw by GB”. (You can select megabyte or terabytes as
well as gigabyte. I’ve selected GB because this is from a volume perspective and
application perspective.
To calculate cost we need to add the “Tier cost” to the report.
Click and drag the “Tier cost” element over place it between the Provisioned Raw and
the Tier column.
70
Insert Technical Report Title Here
To filter out any storage without tier cost associated right click the heading of the Tier
Cost Column and select Filter. (see below for reference)
o Select “Show only following” (default)
o Select “Missing values” to expand it
o Select “Leave out missing values”
o Select OK
71
Insert Technical Report Title Here
See Results below:
72
Insert Technical Report Title Here
Now, let’s calculate the total cost of usage by GB per application.
Hold the control key and highlight the “Provisioned Capacity” and the Tier cost
column until they show yellow.
Select the green Calculation ICON at top of the edit icons above, or right click on the
columns and select “Calculate”
73
Insert Technical Report Title Here
In the calculation window, select multiplication and title the new column “Cost for
Storage and click Insert. It creates a new column, completes the calculation.
To format the column, right click on the new column and select Format Data.
Select currency, number of decimal points (usually 0) and select 1000’s separator
and click OK. See how the column is formatted now.
Double click the “Title” on the report and re-title the report “Chargeback by Application
and BU”.
Now, you don’t really need the tier cost column so you can delete it by right click on
the column and select Delete.
This is good raw report but now let’s make it more useful.
To group storage cost by Business Unit and Application,
Select the Business Unit column (turns yellow) and select the Group by ICON on the top
line.
You see the report reformats itself into cost by application by business unit.
74
Insert Technical Report Title Here
Click “Save As” icon and save the report to the public folders.
Further Editing:
You can go back and further edit the report like this:
Let’s filter out all the N/A and NAs in the BU and Application columns. You have to do this
one column at a time.
Right click the BU column and select filter.
In the filter dialog window, select “Do not show the following (NOT)” from the
“Condition” dropdown
Select NA and click OK
Do the same for the Application column.
Then Save the report again
75
Insert Technical Report Title Here
As you see you now have a better quality report.
To exit Query Studio, click the “Return” icon at the top right corner of the screen.
8 SCHEDULING REPORTS FOR DISTRIBUTION
OK now the report is saved let’s schedule it for running and distribution. You can schedule
all the built in reports in OnCommand.
At the chargeback report we just created, (you should be looking at where you saved
it…)
Select the schedule ICON on the right-hand side where you can set the properties
76
Insert Technical Report Title Here
As you see on the right you can schedule start and finish date
You can just send this report one time by clicking disable
Set the schedule options for weekly, daily, monthly, etc. Schedule to run this report and
send it to yourself at 3pm every Tuesday until Feb 1, 2012. As you can see, you can
schedule biweekly, several times a week, several times a day or you can also set up by
month, year, and even by trigger. As you see lots of options.
There are a lot of options for report format. The default format is HTML but we can
override that default format by clicking and choosing from PDF, Excel, XML, CSV, etc.
For delivery, we can email it, save it, print the report to a specific printer. You can send
the report via e-mail to users, distribution lists, etc. We can include the link of the report
77
Insert Technical Report Title Here
or attached it directly to the email as well. NOTE: They must be able to log into
OnCommand DWH to access the link.
When you are done, click OK and the schedule is set!
9 ENDING COMMENTS AND FEEDBACK
I hope this Lab was of value to you. Your feedback is important to the quality of this lab
document. Please provide feedback to Dave Collins at [email protected]
NetApp provides no representations or warranties regarding the accuracy, reliability or serviceability of any
information or recommendations provided in this publication, or with respect to any results that may be
obtained by the use of the information or observance of any recommendations provided herein. The
information in this document is distributed AS IS, and the use of this information or the implementation of
any recommendations or techniques herein is a customer’s responsibility and depends on the customer’s
ability to evaluate and integrate them into the customer’s operational environment. This document and
the information contained herein may be used solely in connection with the NetApp products discussed
in this document.
© 2012 NetApp, Inc. All rights reserved. No portions of this document may be reproduced without prior written consent of NetApp,
Inc. Specifications are subject to change without notice. NetApp, the NetApp logo, Go further, faster, xxx, and xxx are trademarks or
registered trademarks of NetApp, Inc. in the United States and/or other countries. All other brands or products are trademarks or
registered trademarks of their respective holders and should be treated as such.nTR-XXX-XX
78
Insert Technical Report Title Here