Dashboard Widgets

This page describes each of the dashboard widgets available within ThousandEyes. Dashboard Widget Types This section describes the available dashboard widgets. See Examples of Widgets for the configuration options.
A “widget type” refers to a category of widgets, for example Live Status widgets such as the Alert Grid or Agent Status widgets.

Widget Summary Table

The table below summarizes the widgets available for ThousandEyes dashboards, grouped by widget type.
Widget Type:
Live Status
Alert Grid
A 24-hour, live look at your alerts as they relate to specific tests and agents.
Alert List
A look at the alerts that were active during the configured time period.
Tests
A 12-hour, live display a list of tests configured in your account group for a one stop glance at high level test health. Grayed-out rows show disabled tests.
Agent Status
A live look at the status of your enterprise or endpoint agents to give you an idea of overall agent health.
Widget Type:
Breakdown
Stacked Bar
Provides horizontal histogram bars with multiple values, which are useful for composite metric data (such as HTTP response or fetch time) and for comparing values between multiple tests, or comparing values on a per-country basis. Bars can be oriented horizontally or vertically as columns.
Grouped Bar
Represents multiple values as single bars in a group of bars. Bars can be oriented horizontally or vertically as columns.
Pie
Similar to stacked bar chart widgets, representing multiple values as proportionally sized segments of a pie.
Widget Type:
Data Summary
Table
Allows a breakdown of numbers by rows and columns. Either rows or columns can list by test, country, continent, or data source (agents or BGP monitors) or the aggregate of those ("All").
Multi-Metric Table
Can have columns with different metrics, rather than a single metric for the entire table.
Number
One or more cards, each card displays a single scalar quantity, such as an average of packets lost or page load times, or a number of alerts.
Color Grid
Displays an array of colored cards, where each card's color depends on the configured color scale, with red representing errors or poor target test performance, green signifying good test performance, and shades in between indicating various levels of degradation.
Widget Type:
Time Series
Line
Line plots with time on the horizontal axis, and the selected quantity on the vertical axis.
Stacked Area
Line plots with time on the horizontal axis, and the selected quantities on the vertical axis. They are used similarly to stacked bar charts but showing values over time. Useful when looking at increases in overall response timing, error counts, and other components. Use the stacked area chart in place of where you're using stacked bar charts today, to show a progression of values over time.
Box and Whiskers
Plot data values versus time on the horizontal axis, with the vertical axis displaying the median, minimum and maximum data points per time value, along with a bar for the range of values represented by the second and third quartiles.
Widget Type:
Maps
Map
Displays data on a world map, based on location of the testing systems (agents or BGP monitors). Data can be displayed per country, per continent, or per agent (if a non-BGP metric is displayed).

Widget Configuration Summary Table

Each widget has custom options for data sources, filters, and similar, as shown below.
Widget Type:
Live Status
Alert Grid
Alert type is Agents, BGP, DNS+, or Endpoint.
Alert List
Alert Types are BGP Routing, Cloud & Enterprise Agents, Devices, Endpoint Agents. Beneath these you can select ThousandEyes test types.
Tests
Search your own tests by ThousandEyes test name, or exclude tests by name.
Agent Status
Choose from agents that you own, either endpoint agents or enterprise agents. (No option to see status of cloud agents, as those are managed by ThousandEyes.)
Widget Type:
Breakdown
Stacked Bar
Data sources: Cloud & Enterprise Agents, Endpoint browser sessions, or Endpoint scheduled tests.
Grouped Bar
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Pie
Data sources: Cloud & Enterprise Agents, Endpoint browser sessions, or Endpoint scheduled tests.
Widget Type:
Data Summary
Table
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Multi-Metric Table
You can choose properties, but not data sources.
Number
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Color Grid
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Widget Type:
Time Series
Line
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Stacked Area
Data sources: Cloud & Enterprise Agents, Endpoint browser sessions, or Endpoint scheduled tests.
Box and Whiskers
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
Widget Type:
Maps
Map
Data sources: Alerts, CEA, Devices, Endpoint browser sessions, Endpoint local networks, Endpoint scheduled tests, Endpoint automated session tests, Internet Insights, or Routing.
The remainder of this page contains a more in-depth description of what each widget does, shows an illustrated example of each type of dashboard widget, and explains the configuration options for each widget.
To bring up the widget configuration screen, click the gear icon next to the widget displayed in a dashboard. For adding a new widget, click + Add Widget in the dashboard where you want to add it, select the widget type, and the configuration screen will appear.

Widget Type: Live Status

Live Status widgets include Alert Grid, Alert List, Tests, and Agent Status.

Alert Grid Widget

The Alert Grid widget displays alerts over the past 24 hours using a matrix with tests as rows, and either Enterprise Agents or BGP monitors as columns.
  • An Alert Grid widget can display Cloud, Enterprise, or Endpoint Agent vantage point data, as well as BGP monitors.
  • A test with an active alert is indicated by a red border around an orange dot, under the column of the agent or monitor that generated the alert.
  • Alerts that are no longer active are borderless orange dots.
  • The length of time the alert was active is proportional to the dot size.
  • You can mouse over the dot to see a tooltip with alert data.
For an Alert Grid widget configured with a filter, if no alert matches the filter expression, the widget displays "No active agent alerts" or "No active BGP monitor alerts", along with the number of alerts hidden due to the filter expression:

Alert Grid Widget Configuration

The Alert Grid widget configuration is shown below.
Alert Type: Options are Agents, BGP, DNS+, or Endpoint. Agents displays the account group's agents across the top of the Alert Grid. BGP displays BGP monitors.
Filter all/any: Enter one or more filter criteria. Filter criteria will be displayed when you place focus on the text field.
Exclude all/any: Identical to Filter all/any above, except matches cause exclusion from display in the widget.

Alert Grid Filtering

Alert Grids have two filtering options, Filter all/any: and Exclude all/any:. Each of these options supports AND or OR filtering using “any” or “all”. The filter finds anything containing the specified filter string; it does not have to be a fully exact match.
For Filter all/any:
  • Select "all" to display items that meet all of the criteria in the text field.
  • Select "any" to display items that meet even one of the filter criteria.
For example, suppose you define two Test Name filter criteria: "ThousandEyes" and "Support".
  • Choosing “any” would surface tests named "My ThousandEyes", "ThousandEyes Support site" and "Linux support website".
  • However, choosing "all" would only match to the test named "ThousandEyes Support site".
Use Exclude all/any: to filter out items that match the criteria in a similar manner as above.

Alert List Widget

The Alert List widget displays alerts that are currently active or were recently cleared within the configured period of time:
The Alert List widget includes the following columns:
  • Active alerts display a red dot on the left-hand side.
  • The alert duration displays in orange on the right.
  • Alert Rule references the alert rule, associated with the test, that triggered the alert.
  • Alert Type refers to the ThousandEyes test type. For example “Web - Transaction” refers to transaction tests, which are part of the Web layer or family of tests.
  • If you click the link in the Alert Source column, you see the results view of the test that generated the alert.
  • Click the View All Alerts link in the top right-hand corner of the widget to open the Alerts > Alert List view, where you can see all your alerts.
For further information about ThousandEyes alerting capabilities, see Alerts.

Alert List Widget Configuration

The Alert List widget configuration is shown below.
Alert Type: Configure the types of alerts to show. Various combinations can be configured, from BGP Routing, Cloud and Enterprise Agents, Devices, and Endpoint Agent classes of alert types.
Limit to: Number of alerts to show. If the amount of alerts to show exceeds the configured limit, a Show more alerts link is provided.
Active within: Limit the age of alerts shown. Only alerts that are currently active, or were active during the outlined period, are shown by the widget. The currently active alerts are listed first.
Filters: Additional filters can be configured to narrow down displayed alerts, by specific Tests, Labels, Agents, Alert rules, etc.

Tests Widget

The Tests dashboard widget displays a list of tests as follows:
  • Tests configured in your account group
  • Tests shared to your account group via the Live Share feature
  • Grayed-out rows indicating disabled tests
Each test lists the name, the test type, the test's alert status, and a 12-hour trend graph of key metrics relevant to that test type. The Tests widget only shows test data for cloud and enterprise agents. It doesn’t show data from endpoint agents, devices, or BGP monitors.
  1. 1.
    Header bar – Click a column name to sort the list of results by that column, or to toggle sort order between ascending and descending alphabetical order.
  2. 2.
    Alert status – Shows red for an alert that is currently running, or green for a healthy test. No color indicates either that no alert rules are assigned to the test, or that alerts are disabled on the test.
  3. 3.
    Test name hyperlink – Click to open the results page for that test. For a test type with multiple views, the default view opens.
  4. 4.
    Trending values – Default metrics for each test type, using values collected over the trailing 12 hours. Navigate to any test round by clicking on the desired point on the graph.
  5. 5.
    Current values – Default metrics for each test type, showing the most recent set of values collected when the page was last refreshed.
  6. 6.
    Test configuration – Click to open the test's configuration page. The gear icon appears when you mouse over the row.

Tests Widget Configuration

The Test widget configuration is shown below.
Filter all/any: Enter one or more filter criteria. Filter criteria will be displayed when you place focus on the text field. With "all" the filter displays matches of the criteria that meet all the criteria in the text field. With "any" the filter displays matches if any of the criteria are met. For example, two Test Name criteria of "ThousandEyes" and "Support" would match the tests "My ThousandEyes", "ThousandEyes Support site" and "Linux support website" if the selection were "any". If the selection were "all", then the match would only be "ThousandEyes Support site".
Exclude all/any: Identical to Filter all/any above, except matches cause exclusion from display in the widget.
Data Suppression: Check a box to enable a data exclusion filter.
  • Exclude Disabled tests: Omit tests that have been disabled. This is on by default.
  • Exclude Alert Suppression Window data: Omit data retrieved during an alert suppression window.

Agent Status Widget

The Agent Status widget is a global map that displays enterprise or endpoint agents that belong to your account group or have been shared with your account group, and whether those agents are online, offline, or disabled:
Enterprise Agent Status widgets display Enterprise Agents on a world map, using the agent's IP address with the ThousandEyes geolocation database, or manually configured geographical information to place the agent in the correct location on the map.
Agent locations are determined using:
  • The IP address, using the ThousandEyes geolocation database
  • Manually configured geographical information using the Country and Region settings on an agent (which override geolocation).
If the agent’s exact location cannot be determined:
  • The agent is placed in the correct country if possible
  • If no information is available, the agent is shown in a neutral location on the map
Each circle on the map displays a number if more than one agent is in that location. The color indicates the status of the agent:
  • Green: Indicates an agent with up-to-date data uploads.
  • Red: Indicates an agent which has not uploaded data in the current round.
  • Yellow: Indicates a mix of agents that are and are not current in their data uploads.
  • Clear: Indicates a disabled agent.
A summary legend of online, offline, and disabled agents displays in the upper-left corner. In the upper-right corner of the widget, zoom-in (+), zoom-out (-) and 100% zoom controls are available.
When you mouse over an agent dot, a tooltip displays the names and private and public IP addresses of the Enterprise Agents in that location - or names and host OS information for Endpoint Agents. If you click the name of an agent, the agent's configuration opens in the corresponding Agent Settings page.

Agent Status Widget Configuration

The Agent Status widget configuration is shown below.
Agents: This selector allows the widget to display either Enterprise Agents or Endpoint Agents.
Show: Owned Agents displays only those agents that belong to the user's current account group (i.e., agents installed using the account group's token). All Assigned Agents (Enterprise Agents only) displays owned agents, plus any agents shared to your account group from another account group in your organization.
Filter: Allows for the selection of specific agents or account groups.

Widget Type: Breakdown

Breakdown widgets include Stacked Bar, Grouped Bar, and Pie.

Stacked Bar Widget

The Stacked Bar dashboard widget is useful for composite metric data (such as HTTP response or fetch time) and for comparing values between multiple tests or on a per-country basis. Bars can also be oriented vertically as columns.

Stacked Bar Widget Configuration

The Stacked Bar widget configuration is shown below.
Stacked bar chart widgets have the following settings:
Data Source: The data source indicates which platform service will be referenced within the widget. Current options include:
  • Alerts: Data is obtained from Platform alerts
  • Cloud and Enterprise Agents: Data is obtained from scheduled tests performed by Cloud and Enterprise Agents
  • Devices: Data is obtained from device-layer reporting
  • Routing: Data is obtained from BGP tests
Category: The category choices depend on the Data Source selected. For example, if you select Cloud and Enterprise Agents as the data source, the Category drop-down shows you a subset of the ThousandEyes test type categories, for example Web - Transaction or Web - Page Load.
Metric: Options for metrics depend on the test type chosen in the previous Category selection. For example, in the Stacked Bar dashboard widget configuration, the metrics that you can choose from (Response Time, Total Fetch Time and Total Error Count) are each themselves composites of other metrics found in an HTTP Server test.
To continue with the example, suppose that you want your Stacked Bar widget to display the Response Time metric for all your HTTP server tests. Response Time consists of four distinct sub-metrics that are part of the HTTP server test: DNS resolution time, connect time, SSL time and wait time. In this dashboard widget, all of these four metrics are presented in a single stacked bar.
For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean?
Measure: A statistical measurement. Measure types available depend on the Metric setting, and will be one or more of the following:
  • % positive (Total Error Count only): The percentage of measurements with non-zero error counts.
  • % zero (Total Error Count only): The percentage of measurements with error counts equal to zero.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were between 0 and 500 ms.
  • Maximum: The maximum value in the set of measurements, such as the response times of an HTTP Server test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the response times of a HTTP Server test over the time period of the dashboard.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Response Time" would indicate that the average value of the response time in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Response Time" would indicate that the value of 50 ms is the middle value in the range of all values of the response time in the reporting period.
  • Standard Deviation: The population standard deviation of a set of measurements, such as the response times, over the time period of the dashboard.
  • Total (Total Error Count only): The total number of errors in the reporting period.
X-Axis or Y-Axis: Arrange the data into bars by the following criteria, and orient the label(s) on the axis displayed :
  • All: Produces a single bar with all data.
  • Agent: Produces bars for each agent.
  • Continents: Produces bars for each continent/continental region (North America, Europe/EMEA, Asia/APAC, etc...).
  • Countries: Produces bars for each country.
  • Test: Produces bars for each test.
  • Test Labels: Each test label has its own bar.
  • Sources: Produces dedicated bars for each source.
  • Visited Sites: Each visited site has its own bar.
  • Private Networks: Each private network has its own bar.
  • User: Each user has its own bar.
  • Connection: Each connection gets a dedicated bar.
  • Network: Produces bars for each network.
  • Domain: Each visited domain has its own bar.
  • Location: Each location would have a dedicated bar.
  • Endpoint Agent Labels: A separate bar is produced for each Endpoint Agent label.
  • Endpoint Tests: Each Endpoint test has its own bar.
  • Devices: Each device is dedicated a bar on graph.
  • Interfaces: Device interfaces would have dedicated bars.
  • Device Types: Separate bars produced for each type of device.
  • Interface Types: Each interface type would have dedicated bars.
Either X-Axis or Y-Axis is displayed, based on the Orientation selected.
Orientation: This toggle icon controls the orientation of the bars, either horizontal or vertical. Selecting horizontal or vertical orientation sets the associated pull-down menu label to X-Axis or Y-Axis.
Sort By: For charts containing multiple bars, the Sort By setting will determine in what order the bars appear. That is, the value of the Sort By setting is applied to the value in the X-Axis/Y-Axis setting.
  • Default (Highest): Sort bars optimally, based on the settings for Metric and Show Comparison. The metric that is optimal for your settings is shown in parentheses.
  • Alphabetical: Sort bars by ascending alphabetical order.
  • Alphabetical (Reverse): Sort bars by ascending alphabetical order.
  • Highest: Sort bars by highest value first.
  • Lowest: Sort bars by lowest value first.
Fixed Time Span: When checked, a date selector will appear, allowing selection of a range of days different than the one specified for the entire dashboard. The widget will display data in that range unless the Global Time Overide option is selected.
Limit To: When checked, the chart will display at most the number of groups you set in the selector. When more groups exist than are displayed, the Sort By order determines which rows are displayed.
Sources: Filter on names of tests or agents that will provide the data to the dashboard. To filter by both tests and agents, click the + icon to the right of the pull-down menu.
Drill Down lets you add multiple filters for Agents, Agent Labels, Tests, Test Labels, and Servers.

Example Stacked Bar Chart Widget

Below is an example of a configured stacked bar chart widget and the resulting output.
  1. 1.
    The configured widget visualizes mean Response Time for HTTP Server tests based on Cloud and Enterprise Agents.
  2. 2.
    Graph is configured to be oriented to extend from left to right with Y axis as Servers sorted alphabetically.
  3. 3.
    This widget is visualizing 4 out of 149 available HTTP server tests.

Grouped Bar Widget

The Grouped Bar dashboard widget represents multiple values as single bars in a group of bars.

Grouped Bar Widget Configuration

The Grouped Bar widget configuration is shown below.
Grouped bar chart widgets have the following settings:
Data Source: The data source indicates which platform service will be referenced within the widget. Current options include:
  • Alerts: Data is obtained from Platform alerts
  • Cloud and Enterprise Agents: Data is obtained from scheduled tests performed by Cloud and Enterprise Agents
  • Devices: Data is obtained from device-layer reporting
  • Endpoint Browser Sessions: Data is obtained from endpoint agent browser sessions
  • Endpoint Local Networks: Data is obtained from endpoint agent local networks
  • Endpoint Scheduled Tests: Data is obtained from endpoint agent scheduled tests
  • Endpoint Automated Session Tests: Data is obtained from endpoint automated session tests
  • Internet Insights: Data is obtained from Internet Insights collective intelligence
  • Routing: Data is obtained from BGP tests
Category: The category options depend on the Data Source selected.
Metric: Options for metrics depend on the previous Category selection. All of the ThousandEyes Data Sources have their own metrics. For example, if you choose Cloud and Enterprise Agents as the Data Source, the Category would show test types.
To continue the example, if we choose Cloud and Enterprise Agents as the Data Source, and Agent-to-Server as the Category, the agent-to-server tests have five metrics: Available Bandwidth, Capacity, Jitter, Latency and Packet Loss. In addition to each of the ThousandEyes test types, alerts also have metrics. For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean?
Measure: A statistical measurement. Measure types available depend on the Metric setting, and will be one or more of the following:
  • % positive: The percentage of measurements greater than zero. For example, the percentage of time when a Network alert was active.
  • % zero: The percentage of measurements equal to zero. For example, the percentage of time when a Network alert was active.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were between 0 and 500 ms.
  • Maximum: The maximum value in the set of measurements, such as the response times of an HTTP Server test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the response times of a HTTP Server test over the time period of the dashboard.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the average value of the latency in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the value of 50 ms is the middle value in the range of all values of the latency in the reporting period.
  • Standard Deviation: The population standard deviation of a set of measurements, such as the response times, over the time period of the dashboard.
  • Total: The total number of measurements. For example, a Metric set to "Alerts" and Tests set to a HTTP Server test indicate the total number of HTTP Server alerts received in the reporting period.
Axis: Arrange the data into bars by the following criteria, and orient the label(s) on the axis displayed :
  • All: Produces a single bar with all data.
  • Source: Produces bars for each source of data (test or agent).
  • Continents: Produces bars for each continent/continental region (North America, Europe/EMEA, Asia/APAC, etc...).
  • Countries: Produces bars for each country.
  • Test: Produces bars for each test.
  • Test Labels: Each test label has its own bar.
  • Sources: Produces dedicated bars for each source.
  • Visited Sites: Each visited site has its own bar.
  • Private Networks: Each private network has its own bar.
  • User: Each user has its own bar.
  • Connection: Each connection gets a dedicated bar.
  • Network: Produces bars for each network.
  • Domain: Each visited domain has its own bar.
  • Location: Each location would have a dedicated bar.
  • Endpoint Agent Labels: A separate bar is produced for each Endpoint Agent label.
  • Endpoint Tests: Each Endpoint test has its own bar.
  • Devices: Each device is dedicated a bar on graph.
  • Interfaces: Device interfaces would have dedicated bars.
  • Device Types: Separate bars produced for each type of device.
  • Interface Types: Each interface type would have dedicated bars.
Either X-Axis or Y-Axis is displayed, based on the Orientation selected.
Orientation: The orientation of the bars--either horizontal or vertical. Selecting horizontal or vertical orientation sets the associated pull-down menu label to X-Axis or Y-Axis.
Group By: One or more sets of bars are produced for each of the values in the Axis setting, based on the value of the Group By setting. The available settings will be among the following:
  • All: All data aggregated into one bar
  • Test: Data from all sources and tests are aggregated together into a single line.
  • Source: Data from all tests or agents are given their own bars.
  • Continents: Data are aggregated by continent. A chart line is displayed for each continent that has data from at least one source.
  • Countries: Data are aggregated by country. A chart line is displayed for each country that has data from at least one source.
  • Test Labels: Data from each test label is aggregated into a single line.
  • Sources: Data from each source is aggregated into a single line.
  • Agent Labels: Data from each agent label is aggregated into a single line.
  • Servers: Data aggregated based on servers in a single line.
  • Visited Sites: Endpoint Data from each visited site is aggregated into a single line.
  • Private Networks: Data from each private network is aggregated into a single line.
  • Users: Data is aggregated based on user into a single line.
  • Platforms: Data from each platform Windows/Mac is aggregated into a single line.
  • Connections: Data for each connection is aggregated into a single line.
  • Networks: Data aggregated based on networks into a single line.
  • Domains: Data aggregated based on visited domain into a single line.
  • Endpoint Agent Labels: Data from each Endpoint Agent label is aggregated into a single line.
  • Device: Data from each device is aggregated into a single line.
  • Device Types: Data from each type of device is aggregated into a single line.
  • Interfaces: Data from each device interface is aggregated into a single line.
  • Interface Types: Data from each device interface type is aggregated into a single line.
Sort By: For charts containing multiple bars, the Sort By setting will determine in what order the bars appear. That is, the value of the Sort By setting is applied to the value in the X-Axis/Y-Axis setting.
  • Default: Sort bars optimally, based on the settings for Metric and Show Comparison. The metric that is optimal for your settings is shown in parentheses.
  • Alphabetical: Sort bars by ascending alphabetical order.
  • Alphabetical (Reverse): Sort bars by ascending alphabetical order.
  • Highest: Sort bars by highest value first.
  • Lowest: Sort bars by lowest value first.
Fixed Time Span: When checked, a date selector will appear, allowing selection of a range of days different than the one specified for the entire dashboard. The widget will display data in that range unless the Global Time Overide option is selected.
Limit To: When checked, the chart will display at most the number of groups you set in the selector. When more groups exist than are displayed, the Sort By order determines which rows are displayed.
Sources: Filter on names of tests or agents that will provide the data to the dashboard. To filter by both tests and agents, click the + icon to the right of the pull-down menu.

Example Grouped Bar Chart Widget

Below is an example of a configured grouped bar chart widget and the resulting output.
  1. 1.
    The chart is configured to visualize Mean Receive Time from HTTP Server tests based on Cloud and Enterprise Agents.
  2. 2.
    Graph is configured to extend horizontally with Servers as Y axis grouped by agent labels and sorted in a descending order.
  3. 3.
    Widget is configured to visualize 5 out of 149 tests.

Pie Widget

Pie widgets represent multiple values as proportionally sized segments of a pie.

Pie Chart Widget Configuration

The Pie Chart widget configuration is shown below.
Pie chart widgets have the following settings:
Data Source: The data source indicates which platform service will be referenced within the widget. Current options include:
  • Cloud and Enterprise Agents: Data is obtained from scheduled tests performed by Cloud and Enterprise Agents
  • Endpoint Browser Sessions: Data is obtained from endpoint agent browser sessions
  • Endpoint Local Networks: Data is obtained from endpoint agent local networks
  • Endpoint Scheduled Tests: Data is obtained from endpoint agent scheduled tests
  • Internet Insights: Data is obtained from Internet Insights collective intelligence
Category: The category choices depend on the Data Source selected.
Metric: The metric choices depend on the Category selected. If we select Cloud and Enterprise Agents as the Data Source and Web - HTTP server as the Category, the metrics you can choose from for the Pie Chart widget are (Response Time, Total Error Count and Total Fetch Time) are composites of metrics found in an HTTP server test.
To continue the example, Response Time consists of four distinct metrics from the HTTP server test: DNS resolution time, connect time, SSL time and wait time. Each of the four metrics are presented in a single pie chart. For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean?
Measure: A statistical measurement. Measure types available depend on the Metric setting, and will be one or more of the following:
  • % positive: The percentage of measurements greater than zero. For example, the percentage of time when a Network alert was active.
  • % zero: The percentage of measurements equal to zero. For example, the percentage of time when a Network alert was active.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were between 0 and 500 ms.
  • Maximum: The maximum value in the set of measurements, such as the response times of an HTTP Server test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the response times of a HTTP Server test over the time period of the dashboard.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the average value of the latency in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the value of 50 ms is the middle value in the range of all values of the latency in the reporting period.
  • Standard Deviation: The population standard deviation of a set of measurements, such as the response times, over the time period of the dashboard.
  • Total: The total number of measurements. For example, a Metric set to "Alerts" and Tests set to a HTTP Server test indicate the total number of HTTP Server alerts received in the reporting period.
Group By: One or more pie charts are produced based on the value of the Group By setting. The available settings will be among the following:
  • All: All data aggregated into one pie chart.
  • Agent: Data from all agents are aggregated into a single pie chart.
  • Continents: Data are aggregated by continent. A pie chart is displayed for each continent that has data from at least one test.
  • Countries: Data are aggregated by country. A pie chart is displayed for each country that has data from at least one test.
  • Test: Data from all tests are aggregated into a single pie chart.
  • All: Group all the data as a single chart.
  • Test Labels: Data from all associated test labels are aggregated in a single pie chart.
  • Sources: Data from each source is given their own pie charts.
  • Agent Labels: Data from each agent label is given its own pie chart.
  • Servers: Data aggregated based on servers. Each server would have a dedicated pie chart.
  • Visited Sites: Endpoint Data from each visited site is given its own pie chart.
  • Private Networks: Data from each private network is given its own pie chart.
  • Users: Data is aggregated based on user, each user gets their own pie chart.
  • Platforms: Data from each platform Windows/Mac is given its own pie chart.
  • Connections: Data for each connection is given its own pie chart.
  • Networks: Data aggregated based on networks, each network gets its own pie chart.
  • Domains: Data aggregated based on visited domain, each domain would have a dedicated pie chart.
  • Endpoint Agent Labels: Data from each Endpoint Agent label gets its own pie chart.
  • Device: Data from each device is given its own pie chart.
  • Device Types: Data from each type of device is given its own pie chart.
  • Interfaces: Data from each device interface is given its own pie chart.
  • Interface Types: Data from each device interface type is given its own pie chart.
Filter: Names of tests or agents that will be the source of the data to the dashboard. To filter by both tests and agents, click the + icon to the right of the pull-down menu.

Example Pie Chart Widget

Below is an example of a configured pie chart widget and the resulting output.
  1. 1.
    The widget is configured to visualize Mean Page Load Time of Browser Sessions - Web collected by Endpoint Agents.
  2. 2.
    Group By field is set to Networks and as a result data from each network is visualized as a separate pie chart.
  3. 3.
    The data is not filtered and as a result all available data is visualized.

Widget Type: Data Summary

Data Summary widgets include Table, Multi-Metric Table, Number, and Color Grid.

Table Widget

The Table widget shows a breakdown of numbers by rows and columns.

Table Widget Configuration

The Table widget configuration is shown below.
Could not load image
Table widgets have the following settings:
Data Source: The data source indicates which platform service will be referenced within the widget. Current options include:
  • Alerts: Data is obtained from Platform alerts
  • Cloud and Enterprise Agents: Data is obtained from scheduled tests performed by Cloud and Enterprise Agents
  • Devices: Data is obtained from device-layer reporting
  • Endpoint Browser Sessions: Data is obtained from endpoint agent browser sessions
  • Endpoint Local Networks: Data is obtained from endpoint agent local networks
  • Endpoint Scheduled Tests: Data is obtained from endpoint agent scheduled tests
  • Endpoint Automated Session Tests: Data is obtained from endpoint automated session tests
  • Internet Insights: Data is obtained from Internet Insights collective intelligence
  • Routing: Data is obtained from BGP tests
Category: The category choices depend on the Data Source selected.
*Metric: Options for metrics depend on the test type chosen in the previous Category selection. For example, agent-to-server tests have five metrics: Available Bandwidth, Capacity, Jitter, Latency and Packet Loss. In addition to each of the ThousandEyes test types, alerts also have metrics. For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean? Depending on your choice of Metric, the names of the widget settings may change, or additional settings may appear. For example, selecting any of the alerts metrics will result in the Alert Rules setting appearing in the widget.
Measure: A statistical measurement. Measure types available depend on the Metric setting, and will be one or more of the following:
  • % positive: The percentage of measurements greater than zero. For example, the percentage of time when a Network alert was active.
  • % zero: The percentage of measurements equal to zero. For example, the percentage of time when a Network alert was active.
  • Maximum: The maximum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were between 0 and 500 ms.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the average value of the latency in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the value of 50 ms is the middle value in the range of all values of the latency in the reporting period.
  • Total: The total number of measurements. For example, a Metric set to "Alerts" and Tests set to a HTTP Server test indicate the total number of HTTP Server alerts received in the reporting period.
  • Active Time: The amount of time an alert rule was active. (Alerts metrics only)
  • Inactive time: The amount of time an alert rule was inactive. (Alerts metrics only)
Show Comparison to Previous Timespan: When checked, the dashboard displays the difference between the current data set and a past data set, shown as an additional row of cells below the principle row, with the comparison numbers in red or green along with an up-arrow or down-arrow indicating a numerical increase or decrease. The combination of color and arrow direction changes based on the metric setting. For some settings of metric, a numeric increase is a change for the better and thus is rendered in green. For other metrics a numeric increase is a change for the worse and rendered in red. For some metrics, the situation may be ambiguous; those metrics are rendered in grey font without accompanying arrows.
"N/A" is displayed in place of the comparison numbers when sufficient data is not currently available for a comparison (such as when a test is first configured but hasn't accumulated a full timespan of data). A "-" (single dash) is displayed in place of the comparison numbers when no data is present and will not ever be present, such as when a test error prevented a round of data collection.
By default, the timespan of the past data set is the timespan prior to the current data set's time span, which is specified by the From and To fields of the dashboard's date and time selector. If the Fixed Time Span box is checked, then the timespans of the data sets are specified by the Fixed Time Span setting's date selector.
Fixed Time Span: When checked, a date selector will appear, allowing selection of a range of days different than the one specified for the entire dashboard. The widget will display data in that range unless the Global Time Overide option is selected.
Limit To: When checked, the table rows will display at most the number you set in the selector. When more rows exist than are displayed, the Sort By order determines which rows are displayed.
Rows:
  • Agent: Data from all agents are aggregated by agent. A row is displayed for each agent that has data from at least one test.
  • All: Data from all sources are aggregated into a single row.
  • Continents: Data from all sources are aggregated by continent. A row is displayed for each continent that has data from at least one source.
  • Countries: Data from all sources are aggregated by country. A row is displayed for each country that has data from at least one source.
  • Test: Data from all tests are given their own rows.
  • Test Labels: Data from all test labels are given their own rows.
  • Agent Labels: Data from all agents aggregated based on agent labels.
Columns:
  • Agent: Data from all agents are aggregated by agent. A column is displayed for each agent that has data from at least one test.
  • All: Data from all sources are aggregated into a single column.
  • Continents: Data are aggregated by continent. A column is displayed for each continent that has data from at least one source.
  • Countries: Data are aggregated by country. A column is displayed for each country that has data from at least one source.
  • Test: Data from all sources are aggregated by test. A column is displayed for each test that has data from at least one source.
Sort By: For tables containing multiple rows, the Sort By setting will determine in what order the rows appear. That is, the value of the Sort By setting is applied to the value in the Rows setting.
  • Default: Sort rows optimally, based on the settings for Metric and Show Comparison. The metric that is optimal for your settings is shown in parentheses.
  • Alphabetical: Sort rows by ascending alphabetical order.
  • Alphabetical (Reverse): Sort rows by ascending alphabetical order.
  • Greatest Decrease: Sort rows by greatest decrease between current and past measurements of a row's Metric value. This setting is meaningful only when the Show Comparison box is checked.
  • Greatest Increase: Sort rows by greatest increase between current and past measurements of a row's Metric value. This setting is meaningful only when the Show Comparison box is checked.
  • Highest: Sort rows by highest value first.
  • Lowest: Sort rows by lowest value first.
Tests: Names of tests that will provide the data to the dashboard. Select one or more tests that are appropriate for your selection of Metric setting (e.g. if the Metric is Packet Loss, select one or more Network tests or tests with included Network metrics; if the metric is Availability, select a Web test, etc..)
Sources: Filter the sources from which the data will be taken. This setting's name and the corresponding column heading change once the Metric is set, and will be one of the following:
  • Alert Rules: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of alert rules.
  • Agents: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of Cloud and Enterprise Agents.
  • Agent Labels: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of Cloud and Enterprise Agent labels (Built-in or Custom)
  • Tests: For any Metric, a list of tests.
  • Test Labels: For any Metric, a list of test labels (Built-in or Custom).
  • Monitors: If the Metric is a BGP test or alert, a list of ThousandEyes BGP monitors.

Example Table Widget

Below is an example of a configured table widget and the resulting output.
  1. 1.
    The configuration shows Cloud and Enterprise Agents set as data source, with latency of agent to agent tests set as a reported metric for the table. The widget is set to visualize Standard Deviation of a latency metric in the source-to-target direction.
  2. 2.
    Data is set to visualize with tests forming the Rows and agent forming the columns. Rows are sorted by a reported value, in the ascending order.
  3. 3.
    The filter is set to include 2 of 189 available agent to agent tests. The “+” icon can be used to put on additional filters.

Multi-Metric Table Widget

The Multi-Metric Table can have columns with different metrics, rather than a single metric for the entire table.

Multi-Metric Table Widget Configuration

The Multi-Metric Table widget configuration is shown below.
Rows Tab
The multi-metric table widget's Rows tab has the following settings:
Rows: Select the property that will be used for the rows in the table. Rows will be one of the following:
  • All: Any property that is relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Tests: Tests relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Test Labels: Test labels relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Agents: Agents relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Monitors: Monitors relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Agent Labels: Agent labels relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Servers: Servers relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Countries: Servers relevant to the criteria in the Column(s) tab(s) will be used as a row.
  • Continents: Continents relevant to the criteria in the Column(s) tab(s) will be used as a row.
Sort By: For tables containing multiple rows, the Sort By setting determines in what order the rows appear. That is, the value of the Sort By setting is applied to the value in the Rows setting.
  • Default: Sort rows optimally, based on the settings for Metric and Show Comparison. The Metric that is used for each column shows the column's Measure in parentheses.
  • Alphabetical: Sort rows by ascending alphabetical order.
  • Alphabetical (Reverse): Sort rows by ascending alphabetical order.
  • Greatest Decrease: Sort rows by greatest decrease between current and past measurements of a row's Metric value. This setting is meaningful only when the Show Comparison box is checked.
  • Greatest Increase: Sort rows by greatest increase between current and past measurements of a row's Metric value. This setting is meaningful only when the Show Comparison box is checked.
  • Highest: Sort rows by highest value first.
  • Lowest: Sort rows by lowest value first.
Show Comparison to Previous Timespan: When checked, the dashboard displays the difference between the current data set and a past data set, shown as an additional value in the cells below the principle value, with the comparison numbers in red or green along with an up-arrow or down-arrow indicating a numerical increase or decrease. The combination of color and arrow direction changes based on the metric setting. For some settings of metric, a numeric increase is a change for the better and thus is rendered in green. For other metrics a numeric increase is a change for the worse and rendered in red. For some metrics, the situation may be ambiguous; those metrics are rendered in grey font without accompanying arrows.
"N/A" is displayed in place of the comparison numbers when sufficient data is not currently available for a comparison (such as when a test is first configured but hasn't accumulated a full timespan of data). A "-" (single dash) is displayed in place of the comparison numbers when no data is present and will not ever be present, such as when a test error prevented a round of data collection.
By default, the timespan of the past data set is the timespan prior to the current data set's time span, which is specified by the From and To fields of the dashboard's date and time selector. If the Fixed Time Span box is checked, then the timespans of the data sets are specified by the Fixed Time Span setting's date selector.
Fixed Time Span: When checked, a date selector will appear, allowing selection of a range of days different than the one specified for the entire dashboard. The widget will display data in that range unless the Global Time Overide option is selected.
Limit To: When checked, the table rows will display at most the number you set in the selector. When more rows exist than are displayed, the Sort By order determines which rows are displayed.
Sources: Filter sources from which the data will be taken. This setting's name and the corresponding column heading change once the Metric is set, and will be one of the following:
  • Alert Rules: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of alert rules.
  • Agents: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of Cloud and Enterprise Agents.
  • Agent Labels: If the Metric is a Network, Web, DNS or Voice test or alert metric, a list of Cloud and Enterprise Agent labels (Built-in or Custom)
  • Tests: For any Metric, a list of tests.
  • Test Labels: For any Metric, a list of test labels (Built-in or Custom).
  • Monitors: If the Metric is a BGP test or alert, a list of ThousandEyes BGP monitors.
  • Servers: For any Metric, a list of servers being tested.
  • Location: For any Metric, a list of locations being tested.
Columns Tab
The multi-metric table widget's Columns tabs have the following settings:
Metric: Each type of ThousandEyes test has its own metrics. For example, agent-to-server tests have five metrics: Available Bandwidth, Capacity, Jitter, Latency and Packet Loss. In addition to each of the ThousandEyes test types, alerts also have metrics. For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean? Depending on your choice of Metric, the names of the widget settings may change, or additional settings may appear. For example, selecting any of the alerts metrics will result in the Alert Rules setting appearing in the widget.
Measure: A statistical measurement. Measure types available depend on the Metric setting, and will be one or more of the following:
  • % positive: The percentage of measurements greater than zero. For example, the percentage of time when a Network alert was active.
  • % zero: The percentage of measurements equal to zero. For example, the percentage of time when a Network alert was active.
  • Maximum: The maximum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were between 0 and 500 ms.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the average value of the latency in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the value of 50 ms is the middle value in the range of all values of the latency in the reporting period.
  • Total: The total number of measurements. For example, a Metric set to "Alerts" and Tests set to a HTTP Server test indicate the total number of HTTP Server alerts received in the reporting period.
  • Active Time: The amount of time an alert rule was active. (Alerts metrics only)
  • Inactive time: The amount of time an alert rule was inactive. (Alerts metrics only)
To add a column, click on the Add New Column link. To delete a column, click the tab of the column you wish to delete, then click the X in the column's header.

Example Multi-Metric Table Widget

Below is an example of a configured multi-metric table widget entitled “HTTP Server statistics” and the resulting output.
The Rows setting is set to "Tests", indicating that the table's rows will be individual test names, and the Sources is set to filter on "Tests", where 3 tests of 241 are selected, creating a two-row table with the first column containing those two selected tests. Column 1 has "Availability" selected as the Metric and "Mean" as the Measure. Each test row will display the mean of availability in the first column. Column 2 has "Response Time" selected as the Metric and "Maximum" as the Measure. Each test row will display the largest of all response time values in the round, in the second column.
Sort by is set to "Default (Tests)", meaning that the sorting order of the rows will be the default (the first column that's set by the Rows). The dark switch with the white up-arrow indicates that the value of the Sort by field will be sorted from low to high, alphabetically in this case. The Show Comparison box is checked, producing the second, lower set of numbers along with up or down arrows to indicate whether the value rose or dropped, and a color code to indicate whether that rise or drop is an improving change from the previous period (green), a degrading change (red) or is ambiguous (gray). Fixed Time Span is unchecked, indicating that the dashboard uses the time frame that is set at the top of the Dashboards page. Limit To is set to 2 rows, thereby preventing one of the three tests from being displayed.

Number Widget

The Number widget can show one or more cards, each card displaying a single scalar quantity, such as an average of packets lost or page load times, or the total number of alerts.

Number Widget Configuration

The Number widget configuration is shown below.
Number widgets have the following settings:
Data Source: The data source indicates which platform service will be referenced within the widget. Current options include:
  • Alerts: Data is obtained from platform alerts
  • Cloud and Enterprise Agents: Data is obtained from scheduled tests performed by Cloud and Enterprise Agents
  • Devices: Data is obtained from device-layer reporting
  • Endpoint Agents: Data is obtained from scheduled tests, automated session tests, and monitored domain metrics
  • Internet Insights
  • Routing: Data is obtained from BGP tests
Metric: Each type of ThousandEyes test has its own metrics. For example, agent-to-server tests have five metrics: Available Bandwidth, Capacity, Jitter, Latency and Packet Loss. In addition to each of the ThousandEyes test types, Alerts is also a category of choices for metrics. For more information on ThousandEyes metrics, see ThousandEyes Metrics: What Do Your Results Mean?
Measure: A statistical measurement. Measure choices available depend on the Metric setting, but will come from the following list:
  • % positive: The percentage of measurements greater than zero. For example, the percentage of time when a Network alert was active.
  • % zero: The percentage of measurements equal to zero. For example, the percentage of time when a Network alert was active.
  • Maximum: The maximum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • Minimum: The minimum value in the set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • Standard Deviation: The population standard deviation of a set of measurements, such as the latency times of a Network test over the time period of the dashboard.
  • nth Percentile: The value of the metric that contains n% of the measurements. When "nth Percentile" is selected, a text box will appear, allowing you to specify an integer for the percentile value from 1 to 99. 98% is the default. For example, a 98th Percentile whose value is 500 ms for a Metric of "Response Time" would indicate that 98% of the measurements of the response time were from 0 to 500 ms.
  • Mean: The value of the arithmetic mean (i.e. average) of the measurements. For example, a Mean of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the average value of the latency in the reporting period was 50 ms.
  • Median: The value of the median measurement. For example, a Median of 50 ms for a Metric of "Latency" for one or more Network tests would indicate that the value of 50 ms is the middle value in the range of all values of the latency in the reporting period.
  • Total: The total number of measurements. For example, a Metric set to "Alerts" and Tests set to a HTTP server test indicate the total number of HTTP server alerts received in the reporting period.
  • Active Time: The amount of time an alert rule was active. (Alerts metrics only)