Performance Center Quick Start Guide - ADM Help Centers

Performance Center
Software Version: 12.56
Quick Start
Go to HELP CENTER ONLINE
http://admhelp.microfocus.com/pc/
Document Release Date: February 2018 | Software Release Date: February 2018
Quick Start
Performance Center
Legal Notices
Disclaimer
Certain versions of software and/or documents (“Material”) accessible here may contain branding from Hewlett-Packard Company
(now HP Inc.) and Hewlett Packard Enterprise Company. As of September 1, 2017, the Material is now offered by Micro Focus, a
separately owned and operated company. Any reference to the HP and Hewlett Packard Enterprise/HPE marks is historical in nature,
and the HP and Hewlett Packard Enterprise/HPE marks are the property of their respective owners.
Warranty
The only warranties for Seattle SpinCo, Inc. and its subsidiaries (“Seattle”) products and services are set forth in the express
warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional
warranty. Seattle shall not be liable for technical or editorial errors or omissions contained herein. The information contained herein
is subject to change without notice.
Restricted Rights Legend
Confidential computer software. Except as specifically indicated, valid license from Seattle required for possession, use or copying.
Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for
Commercial Items are licensed to the U.S. Government under vendor's standard commercial license.
Copyright Notice
© Copyright 2001 - 2018 EntIT Software LLC
Trademark Notices
Adobe™ is a trademark of Adobe Systems Incorporated.
Microsoft® and Windows® are U.S. registered trademarks of Microsoft Corporation.
UNIX® is a registered trademark of The Open Group.
Oracle and Java are registered trademarks of Oracle and/or its affiliates.
Performance Center (12.56)
Page 2 of 37
Quick Start
Performance Center
Contents
Performance Center
1
Welcome to This Guide
5
How This Guide is Organized
Chapter 1: Introducing Performance Center
5
6
Performance Center Testing Process
6
What is the testing process?
6
Performance Testing Applications
7
Chapter 2: Creating Vuser Scripts
Recording Vuser Scripts
8
8
How do I start recording user activity?
8
How do I record a script?
9
How do I view my script?
11
How do I save the script?
13
Chapter 3: Creating and Designing Performance Tests
14
Logging in to Performance Center
14
Adding Vuser Scripts to Performance Center
15
How do I upload the Vuser Script?
Monitoring Performance Tests
16
16
What is a monitor profile?
16
How do I create a monitor profile?
16
Creating Performance Tests
How do I create a performance test?
Designing Performance Tests
17
17
17
How do I open the Performance Test Designer?
17
How do I design the performance test?
18
How do I add the monitor profile to the test?
20
Chapter 4: Running Performance Tests
Creating Test Sets
21
21
What is a test set?
21
How do I create a test set?
21
Reserving Timeslots
How do I reserve a timeslot?
Running Performance Tests
How do I run the performance test?
Performance Center (12.56)
22
22
23
23
Page 3 of 37
Quick Start
Performance Center
What does the Performance Test Run page display?
Chapter 5: Post-Run Analysis and Trending
Analyzing Performance Test Run Results
24
26
26
How do I view Analysis information?
26
Did I reach my goals?
28
How can I view the graphs?
28
How can I compare data from different graphs?
30
How can I sort graph data?
31
How can I publish my findings?
32
Viewing Performance Improvements and Regressions
What steps are involved in working with trend reports?
32
32
Creating the Trend Report
32
Viewing Trend Report Data
34
Send Us Feedback
Performance Center (12.56)
36
Page 4 of 37
Welcome to This Guide
Welcome to the Performance Center Quick Start. Performance Center is Micro Focus's Webenabled global performance testing tool, designed to streamline the testing process and increase
the test efficiency for multiple concurrent performance tests across multiple geographic locations.
This Quick Start is a self-paced guide designed to lead you through the process of creating,
running, and analyzing a performance test, and to introduce you to the Performance Center
testing environment.
Note: This guide assumes that your Performance Center environment has been fully
installed and configured, and that a domain and project have been set up. For details,
contact your Performance Center administrator.
How This Guide is Organized
This guide contains the following lessons:
Lesson
Description
"Introducing
Performance Center"
on page 6
Provides an overview of Performance Center and the performance
testing applications.
"Creating Vuser
Scripts" on page 8
Describes the steps involved in recording Vuser scripts using Virtual
User Generator.
"Creating and
Designing
Performance Tests" on
page 14
Describes the steps involved in creating and designing performance
tests.
"Running Performance
Tests" on page 21
Describes the steps that are necessary before running a performance
test, and how to begin test execution.
"Post-Run Analysis
and Trending" on
page 26
Describes how to analyze test run data using LoadRunner Analysis,
and how to view performance improvements and regressions using
the Trend Reports feature.
Performance Center (12.56)
Page 5
Chapter 1: Introducing Performance Center
This lesson provides a brief overview of Performance Center, and of the applications that are involved in
the testing process.
Notes:
l
l
l
The Quick Start describes how to run a performance test with a single host functioning as
both a Controller and a load generator (C+LG). However, due to possible heavy load on the
Controller and load generator hosts during a performance test, it is best practice to assign
these functions to separate host machines.
Ensure that the latest standalone versions of Virtual User Generator (VuGen)
and LoadRunner Analysis are installed on your system.
You can find additional help on developing a Vuser script in VuGen, with a step-by-step
example, in the LoadRunner Tutorial.
In this lesson you will learn about:
l
l
" Performance Center Testing Process" below
"Performance Testing Applications" on the next page
Performance Center Testing Process
Using Performance Center, you create performance tests in which you define the events that occur
during a testing session. During a test, Performance Center replaces human users at physical machines
with virtual users, or Vusers. These Vusers create load on your system by emulating actions of typical
users in a repeatable and predictable manner.
What is the testing process?
The testing process consists of the following basic processes:
l
Creating the script. Capturing typical end-user business processes performed on your application.
l
Designing the performance test. Setting up the test environment by defining events that occur
during the testing session.
l
Preparing to run the performance test. Adding the performance test to a test set and reserving a
timeslot for the test.
l
Running the performance test. Driving, managing, and monitoring the test.
l
Analyzing the results and viewing performance trends. Analyzing the performance data generated
during the test run, and viewing trending information which identifies performance improvements
and regressions over time.
These processes are explained in more detail in the lessons that follow.
Performance Center (12.56)
Page 6 of 37
Quick Start
Chapter 1: Introducing Performance Center
Performance Testing Applications
Each step in the testing process is carried out by one of the Micro Focus performance testing tool
components. The components are as follows:
Application
Description
Virtual User
Generator
(VuGen)
Creates the script and generates virtual users, or Vusers. VuGen does this by
capturing actions that typical end-users would perform on your application, and then
records these actions into automated Vuser scripts. These Vuser scripts form the
foundation of a performance test.
Performance Provides the central console from which you build, manage, and monitor a test.
Center
Analysis
Analyzes the performance test and provides graphs and reports with in-depth
performance analysis information. Using these graphs and reports, you can pinpoint
and identify the bottlenecks in your application and determine what changes need to
be made to your system to improve its performance.
Performance Center (12.56)
Page 7 of 37
Chapter 2: Creating Vuser Scripts
This lesson explains the steps involved in recording Vuser scripts using Virtual User Generator.
In this lesson you will learn about:
l
"Recording Vuser Scripts" below
Recording Vuser Scripts
You record Vuser scripts using Virtual User Generator (VuGen). A Vuser script is a record of a typical
end-user business process. VuGen works on a record-and-playback principle. As you walk through a
business process on your application, VuGen records your actions into an automated script which later
forms the foundation of the performance test.
Tip: You can find additional help on developing a Vuser script in VuGen, with a step-by-step
example, in the LoadRunner Tutorial.
This section includes:
l
"How do I start recording user activity?" below
"How do I record a script?" on the next page
"How do I view my script?" on page 11
l
"How do I save the script?" on page 13
l
l
How do I start recording user activity?
You begin by opening Vugen and creating a blank script.
1. Start VuGen.
Double-click the Virtual User Generator
opens.
shortcut icon on your desktop. The VuGen Start Page
2. Create a blank Web script.
a. In the toolbar above the VuGen Start Page, click the Add New Script button
New Script dialog box opens.
Performance Center (12.56)
. The Create a
Page 8 of 37
Quick Start
Chapter 2: Creating Vuser Scripts
A protocol is the language that your client uses to communicate with the back end of the system.
b. Make sure the Category type is Single Protocol. VuGen displays a list of the protocols that are
available for a single-protocol script.
c. From the list of available protocols, select a protocol (example: for a Web-based application,
select Web - HTTP/HTML) and click Create. VuGen creates a blank Vuser script and displays the
script in the VuGen Editor.
How do I record a script?
The next stage in creating user emulation is recording the events performed by a real user. In the
previous section, you created an empty script. In this section, you will record a script.
Note: The following exercise explains how to record a scripts for a Web-based application (Web -
HTTP/HTML).
Performance Center (12.56)
Page 9 of 37
Quick Start
Chapter 2: Creating Vuser Scripts
1. Start recording.
a. Click the Record button
b.
c.
d.
e.
f.
on the VuGen toolbar. The Start Recording dialog box opens.
In the Record into action box, select Action.
In Record, select Web Browser.
In the URL address box, type the URL address on the Web application under test.
Keep all other default settings.
Click Start Recording. A new web browser opens and displays the Web application under test.
The floating Recording toolbar opens.
2. Record your script.
Navigate through your Web application.
Performance Center (12.56)
Page 10 of 37
Quick Start
Chapter 2: Creating Vuser Scripts
3. Stop recording.
Click the Stop Recording
button on the floating toolbar to stop the recording process.
VuGen generates the required code and inserts the code into the Vuser script.
How do I view my script?
VuGen recorded your steps from the moment you clicked the Start Recording button to the moment
you clicked the Stop Recording button. You can now use VuGen to view the script.
1. Display the Solution Explorer tab.
On the left side pane, make sure the Solution Explorer tab is selected.
This tab provides you structured access to the various parts of a Vuser script, as well as to a number
of files that are associated with the Vuser script.
2. Display the Step Navigator tab.
Click the Step Navigator tab.
Performance Center (12.56)
Page 11 of 37
Quick Start
Chapter 2: Creating Vuser Scripts
The Step Navigator displays an icon-based view of the script that lists the actions of the Vuser as
steps. For each action you performed during recording, VuGen generated a corresponding step in
the Step Navigator.
The Step Navigator displays a snapshot icon to indicate that a specific step contains a snapshot.
3. View a step in the VuGen Editor.
Make sure Tools > Options > Editor> General > Show Class\Function Browser is selected in the
VuGen Editor.
Double-click any step in the Step Navigator to display the corresponding function in the VuGen
Editor. The Editor displays a text-based view of the script.
Performance Center (12.56)
Page 12 of 37
Quick Start
Chapter 2: Creating Vuser Scripts
In the Editor, the actions of the Vuser are listed as API functions. VuGen uses color-coding to show
the functions and their argument values in the script. You can type C or API functions, as well as
control flow statements, directly into the script.
How do I save the script?
Select File > Save Script As and save the script locally on your desktop. For example, type Scripts. The
script file should be zipped. After you have created your performance test, you will upload the script to
Performance Center.
Performance Center (12.56)
Page 13 of 37
Chapter 3: Creating and Designing
Performance Tests
This lesson explains how to manage test assets in Performance Center, and the steps involved in
creating and designing a performance test.
In this lesson you will learn about:
l
l
l
l
l
"Logging in to Performance Center" below
"Adding Vuser Scripts to Performance Center" on the next page
"Monitoring Performance Tests" on page 16
"Creating Performance Tests" on page 17
"Designing Performance Tests" on page 17
Logging in to Performance Center
You log into Performance Center as follows:
1. Open your web browser and type the Performance Center Server URL
http://<PCS server name>/loadtest.
Note: If more than one Performance Center server is installed in your system together with
a load balancer, you should access My Performance Center by entering the load balancer's
URL. Alternatively, you can also enter the server's defined internal URL.
Performance Center (12.56)
Page 14 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
2. The My Performance Center Login window opens.
If Performance Center was configured for external authentication, the Login and Password fields
do not appear in this window. For more details on external authentication, see the ALM External
Authentication Configuration Guide.
3. In the Login box, type your user name.
4. In the Password box, type the password assigned to you by your site administrator.
5. Select the Automatically log in to my last domain and project on this machine check box if you
want Performance Center to automatically log in to the last project in which you were working.
6. Click the Authenticate button. Performance Center verifies your user name and password and
determines which domains and projects you can access. If you specified automatic login,
Performance Center opens.
7. In the Domain list, select a domain. By default, the last domain in which you were working is
displayed.
8. In the Project list, select a project. By default, the last project in which you were working is
displayed.
9. Click Login. Performance Center opens.
Adding Vuser Scripts to Performance Center
The first step in your test design process is to add your Vuser script to Performance Center. Adding the
script involves two steps: Creating a scripts folder, and uploading the script to the folder.
Performance Center (12.56)
Page 15 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
How do I upload the Vuser Script?
You upload the script from within Performance Center, as follows:
1. Make sure the script is saved locally and is zipped.
2. On the My Performance Center navigation bar, select Test Management > Test Plan.
3. Create a script folder.
a. Select Subject.
b. Click the New Folder
button, and specify a name for the folder, for example, Scripts.
c. Click OK to add the folder.
4. On the module toolbar, click the Upload Script
button. The Upload Scripts dialog box opens.
5. In the Select Folder box, select the scripts folder you created above.
6. Click the Select button and navigate to the location of the zipped script file.
7. Click Upload to upload the script.
Monitoring Performance Tests
You monitor performance test execution using the Performance Center online monitors.
For example, you use the System Resource monitors to monitor a machine's system resource usage
during a performance test and identify server performance bottlenecks.
A primary factor in a transaction's response time is its system resource usage. Using the Performance
Center resource monitors, you can monitor the Windows Resource usage on a machine during a test
run, and determine why a bottleneck occurred on a particular machine.
What is a monitor profile?
To monitor server resources during the test, you select the type of monitors to run and the servers
whose resources you want to monitor. You then add the measurements to monitor for each server.
These monitor settings can be saved as a monitor profile which can be used by any performance test in
your project.
The following section will show you how to create a monitor profile and how to add a Windows
Resource Monitor to it.
How do I create a monitor profile?
You create and configure monitor profile as follows:
1. On the My Performance Center navigation bar, select Resources > Test Resources.
2. Create a folder for the monitor profile.
3. Select the folder, and click the New Monitor Profile
Performance Center (12.56)
Page 16 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
button. The Create New Monitor Profile dialog box, enabling you to add a new monitor profile.
4. Enter the required information and click OK to add the monitor profile to the folder.
button. In the Add New Monitor page that
5. In the Monitor Profile page, click the Add Monitor
opens, select the monitor that you want to run.
6. Select Windows Resources. The Edit Monitor dialog box opens, enabling you to select the Windows
Resources counters you want to monitor.
7. Enter the desired information, and click Save. The monitor is added to your monitor profile.
You will add the monitor profile to the test as part of the test design process.
Creating Performance Tests
To create a test, you begin by creating a test folder, and then creating the test and adding it to the
folder. In the next section, you will learn how to design the test in the Test Designer.
How do I create a performance test?
You create the test as follows:
1. On the My Performance Center navigation bar, select Test Management > Test Plan.
2. Select Subject. Click the New Folder button, and specify a name for the folder, for example, Tests.
Click OK to add the folder.
3. Select the folder, and click the New Test button. The Create New Performance Test dialog box
opens.
4. Enter the test name and click OK. The Performance Test Designer opens.
Designing Performance Tests
You design your performance test in the Performance Test Designer.
How do I open the Performance Test Designer?
You open the test designer as follows:
Performance Center (12.56)
Page 17 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
1. On the My Performance Center navigation bar, select Test Management > Test Plan.
2. Select the test you created above, and click the Edit Test button on the toolbar. The Performance
Test Designer opens.
How do I design the performance test?
From the Groups & Workload view in the Performance Test Designer, you perform the basic test design
steps. You design the test as follows:
1. Select a workload type for the test.
When the Performance Test Designer opens, you will be prompted to select a workload type for the
test. Accept the default, Basic Schedule, by test, by number.
2. Assign a Controller to the test.
a. Click the browse button adjacent to the Controller box. The Select Controller dialog box opens.
b. Select Specific.
c. Select from the list the host machine that has been assigned the Controller and Load Generator
(C+LG) purpose.
3. Add Vuser scripts.
a. If the Scripts Tree pane is not displayed on the right, click the Select Scripts button on the
toolbar to open it. The Scripts Tree displays the script that you uploaded previously to
Performance Center.
b. In the Scripts Tree, select the script and click the left arrow to add it to the test. The script
appears in the Groups pane.
4. Add load generators.
a. In the in the Load Generators column of the Groups pane, click the Select load generators link.
Performance Center (12.56)
Page 18 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
The Assign Load Generators to Groups dialog box opens.
b. In the Groups column, select the Vuser groups to use in the test.
c. In the Load Generators column, expand the Specific tab, and from the list that appears, select
the host machine that has been assigned the Controller and load generator (C+LG) purpose.
d. Click Assign and then click OK.
5. Configure the schedule settings.
A schedule defines the actions of the Vusers for the performance test run, for example the behavior
of 10 Vusers emulating 10 users simultaneously logging on to the same Web site.
Because typical users do not perform the same action simultaneously, the Scheduler allows you to
schedule a performance test according to a more realistic portrayal of typical user behavior.
Define the schedule as follows:
In the lower area of the Groups & Workload tab, in the Global Scheduler pane, click in each of the
schedule actions in the Actions grid and define them as follows:
l
Initialize.
l
Start Vusers.
l
Duration.
l
Stop Vusers.
The schedule graph displays a graphical representation of the defined schedule. The lines in the
Performance Center (12.56)
Page 19 of 37
Quick Start
Chapter 3: Creating and Designing Performance Tests
graph correspond to the actions defined in the Actions grid.
How do I add the monitor profile to the test?
You add the monitor profile to the test as follows:
1. In the Performance Test Designer, select the Monitors tab.
2. On the toolbar, click Add Monitor Profile. The Monitor Profiles pane opens on the right, displaying
the available monitor profiles.
3. In the Monitor Profiles Tree, select the monitor profile and click the left arrow to add it to the test.
Performance Center (12.56)
Page 20 of 37
Chapter 4: Running Performance Tests
This lesson explains the steps that are necessary before you run the performance test, and how to begin
test execution.
In this lesson you will learn about:
l
"Creating Test Sets" below
"Reserving Timeslots" on the next page
l
"Running Performance Tests" on page 23
l
Creating Test Sets
The next step in the performance testing process is the creation of a performance test set.
What is a test set?
After you design the performance test in the Test Plan module, you organize the test execution by
creating a test set in the Test Lab module and adding an instance of the test to the test set. The
purpose of a test set is to enable you to group together tests that were created for similar goals.
How do I create a test set?
You create the test set and add the test to it as follows:
1. Create the test set.
a. On the My Performance Center navigation bar, select Test Management > Test Lab.
b. Click the
button, and specify a name for the Test Set folder, for example,
Test_set_folder.
c. Click OK to add the folder.
button. The Create New
d. Select the folder you created above, and click the New Test Set
Performance Test Set dialog box opens.
e. Enter the test set name. Click OK. The test name is added to the tree.
f. Select the test set and click OK.
2. Assign the performance test to the test set.
a. Click the
Performance Center (12.56)
button. The Assign Test to Test Set dialog box opens.
Page 21 of 37
Quick Start
Chapter 4: Running Performance Tests
b. Select your test and click OK.
Reserving Timeslots
Before you run the test, you reserve a timeslot to ensure that the resources needed will be available for
the duration of the test.
How do I reserve a timeslot?
1. On the My Performance Center navigation bar, select Resources > Timeslots.
2. Click the New Timeslot
button. The Add New Timeslot dialog box opens.
3. Define the following information:
l Start. Choose Manually as the method for executing tests. The timeslot reservation reserves
testing resources only.
l Name. Enter a name for the timeslot.
l Select a test instance. Select the test you created and the test set instance to link to the timeslot.
The number of Vusers and hosts that were defined as part of the test design process are
automatically displayed.
l Duration. Set the start time, and enter a duration for the test.
l Post run action. If the administrator did not set an action across the project (the options are
available for selection), select Collate and Analyze.
4. Click Calculate Availability. The availability of the requested resources during the selected timeslot
is calculated. The results of this calculation are displayed in the Timeslot Status tab and graphically
on the time chart.
Performance Center (12.56)
Page 22 of 37
Quick Start
Chapter 4: Running Performance Tests
Note: If the timeslot cannot be reserved, reselect your resources or adjust the start time,
taking into account the reasons displayed in the Timeslot Status tab.
5. When you find a valid timeslot, click Submit to save the timeslot.
Running Performance Tests
Now that you have designed the performance test, added an instance of the test to a test set, and
reserved a timeslot for the test, you are ready to run the test and observe how your application
performs under load.
How do I run the performance test?
You run the performance test as follows:
1. On the My Performance Center navigation bar, select Test Management > Test Lab.
2. Select your test from the Performance Test Set pane.
Performance Center (12.56)
Page 23 of 37
Quick Start
Chapter 4: Running Performance Tests
3. Click the
button. The Run Performance Test dialog box opens.
4. Click the Reserved Timeslot tab.
5. Select the timeslot you reserved, and click Run.
The Initializing Run page displays the performance test initialization steps performed by
Performance Center, and the status for each step. If each step is completed successfully, the
performance test starts running and the Performance Test Run page opens. The Test Run page
enables you to manage and monitor the test as it runs.
What does the Performance Test Run page display?
The Performance Test Run page is the control center from which the test run is managed and
monitored.
The Performance Test Run page displays the following:
Pane/Tab Description
Summary
The Summary pane at the top of the page displays a synopsis of the running
performance test.
Performance Center (12.56)
Page 24 of 37
Quick Start
Chapter 4: Running Performance Tests
Pane/Tab Description
Test Run
Details
Beneath the Summary pane is the Test Run Details area which provides four views:
l
l
Tabs
Graphs. This is the view displayed in the above image. This view displays performance
measurements for those resources being monitored in the test. This enables you to
monitor how the application under test is performing in real time and identify where
potential bottlenecks exist.
Groups. The Groups view displays the statuses of the Vusers in each Vuser group in
the performance test.
l
Transactions. The Transactions view displays how many transactions have been
l
executed successfully or unsuccessfully.
Messages. The Messages view displays error, warning, debug, and output messages
that are sent to the Controller by the Vusers and load generators during a test run.
On the right-side of the page are three tabs.
l
Graphs: Displays a list of the configured online graphs. You can select which graphs
l
to display in the Graphs view.
Topology: Displays any defined topologies for the test.
Diagnostics: Displays diagnostics data collected from the test.
l
Performance Center (12.56)
Page 25 of 37
Chapter 5: Post-Run Analysis and Trending
This lesson explains how to analyze test run data using LoadRunner Analysis, and how to view
performance improvements and regressions using the Trend Report feature.
In this lesson you will learn about:
l
l
"Analyzing Performance Test Run Results" below
"Viewing Performance Improvements and Regressions" on page 32
Analyzing Performance Test Run Results
Now that you have completed running the performance test, you can use LoadRunner Analysis to
analyze the performance data generated during the test run. Analysis gathers the performance data
into detailed graphs and reports. Using these graphs and reports, you can pinpoint and identify the
bottlenecks in the application, and determine what changes need to be made to the system to improve
its performance.
How do I view Analysis information?
l
To view Analysis data for your actual performance test, you perform the following steps:
a. Open Analysis.
Click Start > All Programs > HPE Software > HPE LoadRunner > Analysis.
b. Connect Analysis to Performance Center.
i. In Analysis, select Tools > HPE ALM Connection. The HPE ALM Connection dialog box
opens.
ii. Enter the ALM Platform URL, for example, http://server:8080/qcbin, and click Connect.
iii. Under Step 2: Authenticate User Information, enter your ALM user credentials, and click
Authenticate.
iv. Select your domain and project and click Login. The dialog box should look as follows:
Performance Center (12.56)
Page 26 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
v. Click OK.
c. Open the Analysis session file for your performance test.
i. Select File > Open. The Open Existing Analysis Session dialog box opens.
ii. Click Test Lab. The test sets in the project are displayed.
iii. Navigate to the Results_<run_id>.zip file for the instance of your performance test that you
want to analyze.
Performance Center (12.56)
Page 27 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
l
iv. Double-click the Results_<run_id>.zip file. The Analysis session file is downloaded from
Performance Center and is opened in Analysis.
To provide for more interesting results, a sample Analysis session is provided, which is based on a
performance test similar to the one you ran. To open the sample Analysis session:
a. Double-click the Analysis shortcut icon
on your desktop.
b. Select File > Open. The Open Existing Analysis Session File dialog box opens.
c. Navigate to the <LoadRunner Installation>\Tutorial folder.
d. Select analysis_session and click Open. Analysis opens the session file in the Analysis window.
The data contained in this sample session is examined in more detail in the sections that follow.
Did I reach my goals?
Analysis opens displaying the Summary Report. The Summary Report provides general information
about the performance test run. In the Statistics Summary of the report, you can see how many Vusers
ran in the test and view other statistics such as the total/average throughput, and the total/average
hits. The Transaction Summary of the report lists a summary of the behavior of each transaction.
How can I view the graphs?
The graph tree in the left pane of the Analysis window shows the graphs that are open for viewing.
From the graph tree, you can choose to open new graphs and remove graphs that you no longer want
to view. The graphs are displayed in the graph viewing area in the right pane of the Analysis window.
You can view data from the selected graphs in the graph legend in the lower pane of the window.
As an example, we will now analyze the Average Transaction Response Time graph. Using this graph,
you can view the behavior of the problematic transactions during each second of the performance test
run. In this section, you view the behavior of the check_itinerary transaction.
Performance Center (12.56)
Page 28 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
To open and analyze the Average Transaction Response Time graph:
1. Choose Graph > Add New Graph. The Open a New Graph dialog box opens.
2. Under Transactions, select Average Transaction Response Time.
3. Click Open Graph. The Average Transaction Response Time graph opens in the graph viewing
area.
Click Close to exit the Open a New Graph dialog box.
Performance Center (12.56)
Page 29 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
4. In the legend, click check_itinerary. The check_itinerary transaction is highlighted in the graph and
in the legend below the graph.
Notice how the average response time of the check_itinerary transaction fluctuates greatly
compared to the other transactions at the bottom of the graph, which follow a more or less stable
average response time.
How can I compare data from different graphs?
You can join two graphs together to see the effect of one graph's data upon another graph's data. This
is called correlating two graphs. For example, you can correlate the Running Vusers graph with the
Average Transaction Response Time graph to see the effect of a large number of Vusers on the
average response time of the transactions.
1. Choose Graph > Add New Graph. The Open a New Graph dialog box opens.
2. Under Vusers, select Running Vusers.
Performance Center (12.56)
Page 30 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
3. Click Open Graph. The Running Vusers graph opens in the graph viewing area.
Click Close to exit the Open a New Graph dialog box.
4. Select the Running Vusers graph and click the Merge Graphs button
on the main toolbar.
5. From the Select graph to merge with list, select Average Transaction Response Time.
6. In the Select type of merge area, select Correlate, and click OK.
The Running Vusers and Average Transaction Response Time graphs are now represented by one
graph which opens in the graph viewing area.
Another Analysis tool, Auto Correlate (click ), merges all graphs containing data that could have had
an effect on a given transaction. Correlations of the transaction with each element are displayed,
allowing you to deduce which elements had the greatest effect on the given transaction.
How can I sort graph data?
You can filter graph data to show fewer transactions for a specific segment of the performance test, and
you can sort graph data to show the data in more relevant ways. For example, you can filter the Average
Transaction Response Time graph to display only the check_itinerary transaction.
Performance Center (12.56)
Page 31 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
1. Click Average Transaction Response Time in the graph tree to open the graph.
2. Select the graph and click the Set Filter/Group by button
.
3. In the Transaction Name value box, select check_itinerary and click OK.
The filtered graph displays only the check_itinerary transaction and hides all the other transactions.
How can I publish my findings?
You can publish the findings from your analysis session in an HTML or Microsoft Word report. The
HTML report can be opened and viewed in any browser. The Word report is more comprehensive than
the HTML report, enabling you to include general information about the performance test and to
format the report to include your company's name, logo, and the author's details.
For details about analyzing performance test results, refer to the LoadRunner Analysis User Guide.
Viewing Performance Improvements and
Regressions
A trend report is a Performance Center feature that allows you to view changes in performance from
one performance test to another, or across several performance tests. By analyzing these changes, you
can easily identify improvements or regressions in the measurement's performance.
For example, if you would like to analyze the performance trend of the average transaction response
time measurement of Transaction X, then the trend report would display the changes in the response
time from one test to the next, clearly indicating whether this measurement improved (where the
response time decreased) or regressed (where it increased).
What steps are involved in working with trend reports?
The process of creating and viewing a trend report involves the following steps:
l
Step 1 - Creating the trend report. Creating the basic trend report by defining its name, description
and selecting a template.
l
Step 2 - Adding test runs to the trend report. Extracting test run data from Analysis to the trend
report.
l
Step 3 - Viewing trend report data. Opening the trend view and analyzing the trending information.
Creating the Trend Report
You create the trend report from the main Trend Reports page.
To create the trend report:
1. Open the main Performance Trending page.
On the My Performance Center navigation bar, select Runs & Analysis > Trending.
2. Open the Trend Report Creation page.
Performance Center (12.56)
Page 32 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
On the Performance Trending page, click the
Report page opens.
button. The Create New Trend
3. Define the General Details and Content and Layout settings.
a. In the Create New Trend Report page, define the following information:
o In the General Details pane, enter a name and description for the trend report.
o In the General Contents and Layout pane, select the Transaction Trends template.
b. Click Create to create the trend report. The Select Test Runs to Trend Report dialog box
superimposes on the Trend Overview tab.
4. Add Performance Test Runs to the Trend Report.
From the Project, Test Set and Test lists respectively, select the performance test that you want to
trend. All analyzed instances of the performance test appear in the table.
Performance Center (12.56)
Page 33 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
Select those analyzed instances of the test that you want to add to the trend report, and click Add.
Performance Center uploads the test runs from Analysis and adds them to the trend report. The
trend report opens displaying the Trend Overview tab.
Notes:
l
l
To add test runs to the trend report, there must be a data processor in your project's host
pool. For details, contact the Administrator.
The process whereby Performance Center extracts the performance test data from
Analysis is very labor intensive on the data processor and may take anywhere from
several minutes to over an hour. For best results, use a data processor dedicated for this
purpose.
Viewing Trend Report Data
In the following section, we will look at trending information for the average transaction response time
measurement only.
To view trending information:
On the trend report, click the Performance tab.
In the trend view that opens, you will see a table that displays the transactions that were present in the
test run, as well as the average transaction response times for each selected instance of the test run.
By comparing these average transaction response time figures, you can identify if the performance of
the transaction improved or regressed from one test run to the next.
Performance Center (12.56)
Page 34 of 37
Quick Start
Chapter 5: Post-Run Analysis and Trending
The following example illustrates this point.
In the trend view shown above, four transactions (TRX_01, TRX_02, TRX_03, and TRX_04) are
displayed and their average Transaction Response Time measurement is being trended from four
performance test runs: 3, 4, 5, and 6.
Test run 3 has been automatically defined as the baseline run, (as indicated by the word Base in
brackets). This means that the average transaction response times contained in the other tests are
compared to test run 3 only.
In test run 3, the average transaction response time for TRX_01 was 2.045. The average transaction
response time for the same transaction in test run 4 was 4.073, which represents a slower response time
and therefore a regression in the performance of this measurement. The percentage difference
between the two figures is displayed in parenthesis, in this case +99.17%.
In test run 6, the average transaction response time for TRX_01 was 1.05, which represents a faster
response time that test run 3, and therefore a performance improvement. The percentage difference
between the two figures is displayed in parenthesis, in this case -48.66%.
Performance Center (12.56)
Page 35 of 37
Send Us Feedback
Let us know how we can improve your experience with the Quick Start.
Send your email to: docteam@microfocus.com
Performance Center (12.56)
Page 36 of 37