eG Monitoring
 

Measures reported by TermAppTest

This test reports statistics pertaining to the different applications deployed within the Terminal server and their usage by the clients. One set of results is reported for each application.

Note:

This test will report metrics only if the Microsoft RDS server being monitored uses the .Net framework v3.5 (or above).

Measurement Description Measurement Unit Interpretation
Num_procs_running Number of instances of the published application currently executing on the Terminal server Number This value indicates if too many or too few instances corresponding to an application are executing on the host. The detailed diagnosis of this measure, if enabled, displays the complete list of processes executing, the users executing them, and their individual resource utilization. Using these details, you can quickly detect resource-intensive instances and the user executing them.
Cpu_util Percentage of CPU used by the published application Percent A very high value could indicate that the specified application is consuming excessive CPU resources.
Memory_util This value represents the ratio of the resident set size of the memory utilized by the application to the physical memory of the host system, expressed as a percentage. Percent A sudden increase in memory utilization for an application may be indicative of memory leaks in the application.
Handle_count Indicates the number of handles opened by this application. Number A consistent increase in the handle count over a period of time is indicative of malfunctioning of programs. Compare this value across applications to see which application is using a lot of handles.
No_of_threads Indicates the number of threads that are used by this application. Number  
IO_data_rate Indicates the rate at which this application is reading and writing bytes in I/O operations. KB/sec This value counts all I/O activity generated by each instance of the application and includes file, network and device I/Os.
IO_data_oper_rate Indicates the rate at which this application is issuing read and write data to file, network and device I/O operations. Operations/sec This value counts all I/O activity generated by each instance of the application and includes file, network and device I/Os.
IO_read_data_rate Indicates the rate at which this application is reading data from file, network and device I/O operations. KB/sec  
Page_fault_rate Indicates the number of page faults generated by this application per second. Faults/sec

Page faults are generated when an application tries to use memory that is part of its working set, but can't find it. Page faults can be either hard or soft:

  • Hard page faults occur when the page is found in the page file on the hard disk. The hard page faults involve disk I/O and impact performance.

  • Soft page faults happen when the page is found somewhere else in memory. The soft page faults also impact performance, but may not result in heavy performance loss in a physical environment.

  • Compare values across the applications to figure out which application is generating most page faults.

Virtual_memory_used Indicates the total amount memory that is dedicated to this application, and cannot be shared among other applications on the server. MB  
Max_process_input_delay Indicates the maximum amount of time lag detected between the user's input through any input device (e.g., mouse, keyboard) and the time at which this application detected the input. Seconds These measures will be reported only for the Windows 2019 desktops on the target server.

Ideally, the values of these measures should be very low.

High values for these measures can impact the speed of accessing the applications in the environment and badly degrade the overall user experience.  
Avg_process_input_delay Indicates the average amount of time lag detected between the user's input through any input device (e.g., mouse, keyboard) and the time at which this application detected the input. Seconds