Viewing Application Testing Results

Introduction

You can view the results of a single Application Test and compare up to 5 Application Test instances. The Application Test Results page provides a comprehensive overview of measurements and detailed events for each Test. You can also download both single and multiple Test results as PDF reports for easy sharing and analysis.

Single Test

When an Application Test is completed, you can start looking at the results. To do this, in the Login Enterprise sidebar menu > Results, navigate to Application testing.

Frame 1760.png

Single Test Summary

The summary page displays a list of previously executed Application Tests and includes the following:

Time

Name

App. Failures

App. Perf

Comment

Date and time of the Test.

Name and sequence number of the Test.

The number of failed applications relative to the total number of applications tested.

The number of out-of-bound measurements, compared to the configured thresholds. The number on the right shows the total number of configured measurements.

The configured comment for the Test.

Frame 1765.png

To view the Test details, click on the Test you’re interested in:

Frame 1766.png

Tip: View details of events that occurred during the test. Click View details to open a pop-up with additional information.

Frame 1767.png
Frame 1768.png

If the Test includes application errors, you can expand the details, download the log file, or open the screenshot using the camera icon.

Frame 1769.png

Single Test Results

To view the test results for a single test, select a test and click View.

Frame 1770.png

The following page displays the Test summary, Overview (Platform summary and Application summary), Events, and Access.

Platform Summary

Logon performance


Actual

Threshold

Execution

Performance

The actual time it took to log in.

The login time threshold set in the Test.

Green if the login was successful, regardless of the time taken.

  • Green if the actual login time was below the threshold or if no threshold was set.

  • Red if the actual time exceeded the threshold.

Latency


Actual

Threshold

Execution

Performance

The actual latency in milliseconds (ms).

The latency threshold set in the Test, in milliseconds (ms).

  • Green if latency was measured.

  • Red if no latency could be measured.

  • Green if the actual latency was below the threshold or if no threshold was set.

  • Red if the actual latency exceeded the threshold.

Frame 1772.png

Application Summary

Application(s)


Execution

Performance

Screenshot

  • Green if the script was successfully executed.

  • Red if the script had a failure during execution.

  • Green if all measurements/timers in the script were below the threshold or if no threshold was set.

  • Red if one or more measurements exceeded the threshold.

  • If a screenshot was taken, click the camera icon to open a pop-up displaying the screenshot.

  • If multiple screenshots are available, arrows will appear in the pop-up to navigate through them.

  • A download button is also available to save the screenshot on your desktop.

Measurements (Timers)


Actual

Threshold

Status

The actual time recorded for the measurement.

The threshold set in the Test for the specific measurement.

  • Green if the actual time is below the threshold or if no threshold is set.

  • Red if the actual time exceeds the threshold.

By default, the measurements (timers) include the app start time. The app start time is automatically recorded when the START function in the workload scripts is executed. This function is used by default to launch target apps. Additionally, any custom timers defined in the workload scripts will also be displayed in the UI. To learn how to configure custom timers, see the Scripting Functions Overview.

Frame 1787.png

You can choose to show only applications with failed events. This helps you quickly identify failures and troubleshoot issues.

Frame 1774 (1).png

Tip: To scroll horizontally, drag the scrollbar, or place your cursor over the table, hold Shift, and use the mouse wheel.

Generating a Test PDF Report

Automatically

A single Application Test PDF report is automatically generated once the Test is finished. Download the report by clicking the Download button next to the Test:

Frame 1777.png

Note: This is the report for which you can set email notifications. For more information, see the Report Settings.

Manually

Generate a new single Test PDF report manually or using the Public API. To generate the report:

  1. In the Application Testing results, click on the Test you’re interested in.

  2. Click Generate PDF report (A browser pop-up will open).

Frame 1778.png
  1. In the Generate PDF report pop-up, configure the report according to your needs by ticking off the info you want to include. For details and examples, see Configuring a Test PDF Report.

  2. Click Generate, and then Save.

Tip: Alternatively, select the checkbox next to the specific Test, and in the top menu toolbar, click Generate PDF report.

Frame 1779.png

Note: Email notifications for a single Application Test report generated manually aren’t supported.

Compared Tests

To view the test results for the compared tests, select up to 5 tests and click Compare.

Frame 1780.png

The following page displays the summary of the compared Tests, Overview (Platform summary and Application summary), and Events.

Compared Test Summary

This is where you can select a Baseline Test to compare other Application Tests against. Use the checkboxes to add or remove tests from the comparison.

Frame 1781.png

Compared Test Results

Platform Summary

Displays the login time and latency, along with the percentage difference between the compared Tests. For details, see the Platform Summary.

Frame 1784.png

Application Summary

Compares the results of the Application scripts. For details, see the Application Summary.

Frame 1783.png

Note: The results of the Baseline Test are pinned in the first column. The other Test columns are sorted from oldest to newest.

If the results do not fit the screen, scrollbars will appear in the table. The baseline Test column remains pinned and does not move with the scrollbar.

The results of the Baseline Test are the primary reference. All Applications and timers from the baseline are displayed. If a selected Test has additional Applications or timers not present in the Baseline Test, those will only be shown when that Test is designated as the baseline.

The PDF report for the multiple Application Test comparison isn’t available yet.

Events

The Events tab displays a table of key events across selected tests. The table includes the event date, time since test start, Test run name, and event title. Use this tab to identify patterns and troubleshoot issues, including when comparing tests.

Use the event filter on the Test run results page to refine results for single or compared Test runs. You can apply filters from:

  • The filter sidebar

  • Column headers

You can filter by:

  • Event title

  • Event type

  • Account name

  • Launcher name

  • Application name

  • Target host

Note: The Timespan and Occurred after columns always appear in the first and second positions and can't be removed or reordered.

Frame 1712.png

Generating a Test PDF Report

You can generate and download a Test PDF report with the results comparing up to 5 Tests in one of the following ways:

  1. Using the toolbar on the Application Test Results page:

Frame 1785.png
  1. Using the Generate PDF Report button on the Compare page:

Frame 1786.png

Note: An Application Test report comparing multiple Tests is not generated automatically. You can generate the report manually using either the UI or the Public API.

Email notifications for this type of report aren’t supported.

Configuring a Test PDF Report

You can choose what you would like your Application Test PDF report to include. Once you click Generate PDF report, you can configure the following:

Setting

Description

Options / Details

Notes

Baseline test

Select the test run to use as the baseline.

-

  • On the Test Results page, this option isn’t included. The baseline test is determined by the test selected on the page.

  • Available only for multiple test reports.

Introduction text

Include general information about Login Enterprise at the beginning of the report.

  • Include introduction text: Adds the standard intro section.

  • Exclude introduction text: Omits it from the report.

-

Custom text

Add your own text to the PDF report. Useful for describing what was tested, why, or summarizing conclusions.

  • Include custom text: Opens a pop-up to enter your content.

  • Exclude custom text: Skips this section.

  • The text editor has the following options:

    • Font size (heading or normal text)

    • Bold

    • Italic

    • Underlined

    • Bullet points

    • Numeric lists

  • Whatever is entered in the custom text section will be added to the beginning of the report, directly under the introduction text.

  • Use the Enter key to create paragraph breaks. If an uninterrupted block of text exceeds one page, the Application Test PDF will generate a blank page. The custom text section supports up to five pages.

Results

Exclude successful applications from the PDF report to keep it concise and focused on relevant information.

  • Include successful applications: Adds all successful applications to the report.

  • Exclude successful applications: Includes only failed applications in the report.

  • Successful applications are included by default (checkbox unselected).

  • Available for both single multiple test reports.

A PDF report configuration with the results of a single Application Test looks as follows:

Frame 2100.png

A PDF report configuration with the results of the compared Application Tests looks as follows:

Frame 2101.png

Test PDF Report via the Public API

There are two types of Test PDF reports: old and new.

Old Test PDF report

  • Endpoint: /publicApi/v8-preview/reports/{reportId}/pdf

  • Automatically generated after each Test Run.

  • Available for single Application Test Runs only. Comparison reports are not supported.

New Test PDF report

  • Endpoint: /publicApi/v8-preview/reports/application-test-runs/pdf

  • Not saved automatically. You must generate it via the Public API.

  • Supports single and multiple Application Test Run IDs, allowing you to create comparison reports.

For more information, see Accessing the Public API.

Test PDF Report Breakdown

The Test PDF report summarizes the findings obtained from the testing process. It provides an overview of the Test specifics, results, and measurements. The report is divided into the following sections:

  1. Introduction: Provides an overview of Application testing and the metrics used during this type of testing.

  2. Test Specifics: Provides an overview of the key details of the test setup:

    1. Product version: The Virtual Appliance version.

    2. Connector: The specific Connector used for the Test.

    3. Launcher group(s): The Group(s) of Launchers involved in the Test.

    4. Test duration: The total duration of the Test.

    5. Date: The date on which the Test was conducted.

    6. Workload: The type of Workload or Test scenario used.

  3. Test Summary: Highlights key metrics for the Test:

    1. Application failure: Displays the number of failed Applications. For example: script failed to run.

    2. Application performance: Shows the number of Applications where the actual time exceeded the performance threshold.

  4. Platform Summary: Provides insights into the platform's performance based on latency and login performance:

    1. Login performance: Indicates the performance during login operations.

    2. Latency: This shows any latency issues that occurred during the Test.

      1. Login Performance and Latency include the following:

        1. Test name: The name of the Test conducted.

        2. Bar chart: A visual representation of the performance data.

        3. Actual time: The time taken for the specific Test.

        4. Threshold: The predefined threshold, if applicable.

        5. Execution: The result of the Test execution.

        6. Performance: The performance analysis based on the Test outcome.

  5. Application Results: Dives deeper into the results for each Application with the following subsections:

    1. Summary: A high-level result of the Application's performance during the Test.

    2. Measurements: Includes a table for each measurement or timer taken during the Test. Each table will include:

      1. Test name: The name of the Test.

      2. Bar chart: A visual representation of the performance data.

      3. Actual time: The actual time recorded for the measurement.

      4. Threshold: The threshold value (if applicable).

      5. Status: The status of the Test.

      6. Screenshots: This section includes screenshots taken on a particular application under test.

Tips and Tricks

Customizing Table Columns

You can customize the Test Results table by adding, hiding, or sorting columns.

  • Configuration icon: In the top-right corner of the Application Test results table, click the configuration icon to open the column management pop-up.

Frame 1775.png
  • Adding columns: Add additional columns to the table. The following columns are available:

    • App Failures

    • App Performance

    • Comment

    • Test duration

    • Connector

  • Hiding columns: Hide columns you do not want to display using the configuration options.

  • Sorting columns: To sort columns, drag and drop them into your preferred order.

Frame 1776.png

Note: The Time and Test Name columns are fixed and cannot be hidden or moved.

Your configuration settings are saved in your local storage, i.e., your column preferences will persist when you navigate away from the page or log out and back in.

The configuration may reset if you use a different browser or clear your browser history.

Additional Resources