A Complete Performance Testing Guide With Examples

What is Performance Testing?

Performance Testing also knows as ‘Perf Testing’, is a type of testing performed to check how application or software performs under workload in terms of responsiveness and stability. The Performance Test goal is to identify and remove performance bottlenecks from an application.

This test is mainly performed to check whether the software meets the expected requirements for application speed, scalability, and stability. 

In this tutorial series, we will cover complete details like – Perf Testing Types, Process, and Writing Performance Test Strategy document from scratch.

This is a detailed tutorial series you may want to bookmark!

Let’s explore!

A Complete Performance Testing Guide

List of ALL the Performance Testing Tutorials in this Series:

Tutorial #1: Performance Testing Complete Guide (This Tutorial)
Tutorial #2: Difference Between Performance, Load and Stress Testing
Tutorial #3: Functional Testing Vs Performance Testing
Tutorial #4: Performance Test Plan and Test Strategy
Tutorial #5: Ways to Supercharge your Performance Testing
Tutorial #6: Cloud Performance Testing Guide
Tutorial #7: Mobile App Performance Testing Guide
Tutorial #8: How to Perform Manual Performance Testing
Tutorial #9: Website Performance Testing Tutorial
Tutorial #10: Performance Testing Companies
Tutorial #11: Performance Testing with LoadRunner (Series)


Tutorial #12: Top Performance Testing Tools
Tutorial #13: Neoload Performance Test Tutorial
Tutorial #14: BlazeMeter Mobile Performance Test Tutorial
Tutorial #15: WAPT Load, Stress and Performance Test Tutorial
Tutorial #16: SmartMeter.io Website Performance Test Tutorial

Types Of Performance Testing

Performance testing types

Load Testing

Load Testing is a type of performance test where the application is tested for its performance on normal and peak usage. The performance of an application is checked with respect to its response to the user request and its ability to respond consistently within an accepted tolerance on different user loads.

The key considerations are:

  1. What is the maximum load the application is able to hold before the application starts behaving unexpectedly?
  2. How much data the Database able to handle before the system slows or the crash is observed?
  3. Are there any network related issues to be addressed?

Stress Testing

Stress Testing is used to find ways to break the system. The test also provides the range of maximum load the system can hold.

Generally, Stress Testing has an incremental approach where the load is increased gradually. The test is started with a load for which the application has already been tested. Then, more load is added slowly to stress the system. The point at which we start seeing servers not responding to the requests is considered the breaking point.

The following questions are to be addressed:

  • What is the maximum load a system can sustain before it breaks down?
  • How is the system break down?
  • Is the system able to recover once it’s crashed?
  • In how many ways a system can break and which are the weak node while handling the unexpected load?

Volume Testing

Volume Testing is to verify that the performance of the application is not affected by the volume of data that is being handled by the application. In order to execute a Volume Test, a huge volume of data is entered into the database. This test can be an incremental or steady test. In the incremental test, the volume of data is increased gradually.

Generally, with the application usage, the database size grows, and it is necessary to test the application against a heavy database.  A good example of this could be a website of a new school or a college having small amounts of data to store initially, but after 5-10 years, the data stores in the database of the website is much more.

Capacity Testing

=> Is the application capable of meeting business volume under both normal and peak load conditions?

Capacity Testing is generally done for future prospects.  Capacity Testing addresses the following:

  1. Will the application be able to support the future load?
  2. Is the environment capable of standing for the upcoming increased load?
  3. What are the additional resources required to make the environment capable enough?

Capacity Testing is used to determine how many users and/or transactions a given web application will support and still meet performance. During this testing, resources such as processor capacity, network bandwidth, memory usage, disk capacity, etc. are considered and altered to meet the goal.

Online Banking is a perfect example of where capacity testing could play a major role.

Reliability/Recovery Testing

Reliability Testing or Recovery Testing – is to verify whether or not the application is able to return back to its normal state after a failure or abnormal behavior and how long does it take for it to do so (in other words, time estimation).

If an online trading site experiences a failure where the users are not able to buy/sell shares at a certain point of the day (peak hours) but are able to do so after an hour or two, we can say the application is reliable or recovered from the abnormal behavior.

Performance Test Process

Here are all the activities performed in this testing:

Performance Test workflow

#1) Requirement Analysis/Gathering

The performance team interacts with the client for identification and gathering of requirements – technical and business. This includes getting information on the application’s architecture, technologies, and database used, intended users, functionality, application usage, test requirement, hardware & software requirements, etc.

#2) POC/Tool selection

Once the key functionality is identified, POC (Proof Of Concept – which is a sort of demonstration of the real-time activity but in a limited sense) is done with the available tools.

The list of available tools depends on the cost of the tool, protocol that application is using, the technologies used to build the application, the number of users we are simulating for the test, etc. During POC, scripts are created for the identified key functionality and executed with 10-15 virtual users.

#3) Performance Test Plan & Design

Depending on the information collected in the preceding stages, test planning and designing are conducted.

Test Planning involves information on how the performance test is going to take place – test environment, workload, hardware, etc.

More on the Test Strategy document below.

#4) Performance Test Development

  • Use cases are created for the functionality identified in the test plan as the scope of PT.
  • These use cases are shared with the client for their approval. This is to make sure the script will be recorded with the correct steps.
  • Once approved, script development starts with a recording of the steps in use cases with the performance test tool selected during the POC (Proof of Concepts) and enhanced by performing Correlation (for handling dynamic value), Parameterization (value substitution) and custom functions as per the situation or need. More on these techniques in our video tutorials.
  • The Scripts are then validated against different users.
  • Parallel to script creation, the performance team also keeps working on setting up the test environment (Software and hardware).
  • The performance team will also take care of Metadata (back-end) through scripts if this activity is not taken up by the client.

#5) Performance Test Modeling

Performance Load Model is created for the test execution. The main aim of this step is to validate whether the given Performance metrics (provided by clients) are achieved during the test or not. There are different approaches to create a Load model. “Little’s Law” is used in most cases.

#6) Test Execution

The scenario is designed according to the Load Model in Controller or Performance Center but the initial tests are not executed with maximum users that are in the Load model.

Test Execution is done incrementally. For Example, If the maximum number of users is 100, the scenarios are first run with 10, 25, 50 users and so on, eventually moving on to 100 users.

#7) Test Results Analysis

Test results are the most important deliverable for the performance tester. This is where we can prove the ROI (Return on Investment) and productivity that a performance testing effort can provide.

Some of the best practices that help the Result Analysis process:

  1. A unique and meaningful name to every test result – this helps in understanding the purpose of the test.
  2. Include the following information in the test result summary:
  • Reason for the failure/s
  • Change in the performance of the application compared to the previous test run
  • Changes made in the test from the point of application build or test environment.
  • It’s a good practice to make a result summary after each test run so that analysis results are not compiled every time test results are referred.
  • PT generally requires many test runs to reach the correct conclusion.
  • It is good to have the following points in result summary:
    • Purpose of test
    • Number of virtual users
    • Scenario summary
    • Duration of test
    • Throughput
    • Graphs
    • Graphs comparison
    • Response Time
    • Error occurred
    • Recommendations

#8) Report

Test results should be simplified so the conclusion is clearer and should not need any derivation. Development Team needs more information on analysis, comparison of results, and details of how the results were obtained.

The test report is considered to be good if it is brief, descriptive and to the point.

How To Write Performance Test Strategy Document?

This tutorial will explain how to write a sample Performance Test Strategy for a Messaging Application.

Remember, that this is just an example and the requirements will differ from one client to another, we will also get to know the best practices for Performance Testing in this tutorial.

Performance Test Strategy

Sample Performance Test Strategy Template

About ABC chat Application – Let’s assume that this is a chat workbench that is used in a company by their customer support agent, this chat application uses XMPP protocol i.e, Extensible Messaging and Presence Protocol and Open fire server for sending and receiving Instant messages.

Some enhancements have been made to this existing chat client like Remote PC control, PC diagnosis, Repair tools, Online chat, etc., so this performance Test strategy is a sample of such applications.

For this application let’s assume that the project team has decided to use JMeter for Performance Testing and JIRA for defect tracking.

The first page of the Performance Test Strategy document should contain Title of the Document and the Copyrights of the Company.

The second page should contain Document Control which includes, Document Version history, Reviewers & Approvers list and Contributors list.

The third page should contain the Table of contents, followed by the below topics.

#1) Introduction

The purpose of this document is to define/explain how Performance Testing will be performed on the ABC chat application for the current and future state.

ABC chat application is an in-house remote support Agent workbench. This workbench will be used to fulfill customer requests. This Workbench has capabilities such as Online chat, Customer Identification, Remote PC control, PC diagnosis, and repair tools.


The key objectives of Performance Testing are as follows:

  • To gain the confidence that the changes to the existing chat application are in line with the defined Service Level Agreement.
  • To ensure that the application performance, service availability, and the stability of the application are not impacted as a result of the new enhancements.
  • Transaction Response Times remain within the acceptable tolerance over the increasing Load profile.
  • JVMs show stable memory usage over the increasing load profiles.

The below picture clearly explains Performance Testing & Optimization process:

Performance Testing & Optimization process


You need to incorporate the architecture diagram of your project in this session.

#2) Scope

In Scope

Below is the Performance Testing scope for ABC chat workbench:

  • Knowledge acquisition of the key business transactions and build load distribution after a detailed study of the system.
  • Identify the critical scenarios for performance testing with assistance from different project tracks.
  • Use the previous release results as a baseline for future releases.
  • Verify and validate the performance test environment and the Performance/Load test tool infrastructure for any additional Agent Machines.
  • Preparation of performance test scripts using JMeter for the identified scenarios that mimic the identified peak load.
  • Setup performance monitoring on the servers for monitoring the test in order to identify the bottlenecks during the test execution phase.
  • Publish Performance test results.
  • Coordinate with various stakeholders to resolve the identified performance issues.
  • Baseline the performance level for future releases.

Out of Scope

  • Functional Testing, UAT, System Testing & Security Testing.
  • Performance testing/monitoring of any third-party interfaces.
  • Performance Tuning. (Most of the times Tuning is done by a different team, if in case you have performance engineers to tune the system then you can add this in the Inscope).
  • Code profiling / Hardware sizing / Capacity planning.
  • Security / Vulnerability testing / UAT/ White box testing.
  • Data generation for Performance Testing.
  • Non-functional tests (For Example, failover, disaster recovery, back-up, usability) other than the performance tests.
  • Testing of any mobile solution.
  • Third-Party Application Performance Testing & Tuning.
  • Realization of performance recommendations, Application code changes and the vendor-supported products/server configuration changes will be out of scope from the Performance Team perspective.
  • Infrastructure Support / Build Deployment/ Environment Readiness/ Database Restore/ Network Support etc.

#3) Approach

Performance testing for ABC chat will be conducted using Jmeter by writing custom XMPP plugins that use a smack library for XMPP connections. These libraries are used to set up connections, login and send chat messages to the XMPP server.

These libraries are bundled into a jar file that is deployed into the Jmeter and is designed based on the scenarios to be tested. The Jmeter Work Bench is installed in the local machine which connects to the JMeter server which has the Load Generators to generate the required load on the Chat server system to monitor the system behavior.

The test scenario will be scripted using the JMeter tool. The scripts would be customized as required. The schedule will be created with the required ramp-up to simulate the real-world scenarios.

The Test Scenario would be broken up and measured in the below aspects:

a) Baseline Test: To run each scenario with 1 Vuser and multiple iterations in order to identify whether the application performance meets the business Service Level Agreement or not.

b) Base Load Test: To meet the Business Benchmark under load test, the Performance Testing team will perform a baseload test which will help to identify any system performance issues with increasing load and creates the baseline for the next level of performance testing.

c) Peak Load / Scalability Test: Performance Testing team will perform multiple tests with increasing Vusers to meet the expected load and also to measure the application performance to establish the performance curve and identify whether the deployment can support the service level agreements under the peak user load.

It helps in tuning or capacity planning of the individual Java virtual machines (JVM), the total number of required JVMs, and the processors. This will be achieved by increasing the no of Vusers to 50%, 75%, 100% and 125% of peak capacity.

d) Endurance Test: Performance Testing team will run this test for a period of 8 Hours / 16 Hours /24 hours to identify memory leaks, performance issues over time, and overall system stability. During endurance tests, the Performance Testing team monitors the key performance indicators, such as transaction response times and the stability of memory usage.

System resources like CPU, Memory, and IO need to be monitored with the help of the project team.

The Performance test environment is assumed to be a replica of the production environment. The tests will be run with an incremental load to identify where the application fails.

Performance Test Scenarios

Include the excel with the set of scenarios.

For Example,

Scenario 1: To validate the Agent and customer chat for X no. of concurrent sessions.

Performance Test Types

The table given below explains the various types of Performance Tests along with their objectives.

Test TypeObjective
Baseline TestEstablish the best performance under specific volumes which will be used as a reference for subsequent measurements.
Load TestMeasure the system performance under anticipated peak production load.
Endurance TestMeasuring the system stability under high volume for an extended period.
Stress TestMeasure the system performance under unfavorable conditions.

Performance Metrics

  • Client-side Metrics
1Transaction Response TimeResponse time of pages during the steady state of the performance testGraph
2ThroughputThe amount of data that the VUsers received from the server over timeGraph
3Hits/secondThe number of HTTP requests made by VUsers to the Web server during the scenario runGraph
4Number of Passed/Failed Transactions

Total number of transactions that Passed and Failed during the test executionExcel
5Transaction Error RateThe Percentage of transactions that failed during the test executionGraph
  • System & Network Performance Metrics

System & Network Performance Metrics

Performance Testing Activities & Deliverables

Performance Testing Activities

#4) Test Data

It is being assumed that the Performance environment data will be a copy of the production data and the required test data will be provided by the project team.

#5) Entry & Exit Criteria

  • Access to all the applications in the environment.
  • Environment readiness complete.
  • Performance Test Data readiness.

Entry & Exit Criteria

#6) Defect Management

  • The Defect Management module in JIRA will be used in the project for defect logging and for tracking to closure.
  • Identification of defects that are found during the test execution phase will be captured in JIRA and these defects will be resolved by the development team according to the below severities.
  • Defect review meetings would be held on a daily basis with the participation from the testing, development, Quality Analysts, and business teams.
  • The criteria to fix defects would get stringent as the project approaches the Go Live date. Guidelines for defect fix criteria to be published in defect review meetings.

Defect Severity Definition

The definitions of severity codes are as follows:

SeverityDescription for Development and Enhancement Problems
BlockerSystem error, show stopper, Network issues
CriticalSystem errors, no clear workaround, interruption or missing business functionality

MajorA serious problem was detected for which the workaround exists that might not be clear to all the users, however, product should not be released without fixing
MediumProblem exists with easy/simple work around but this type of defect may be released upon approval by Business and/or Project Manager
LowCosmetic issues that do not interfere with business functionality or other intermittent problems that are not reproducible every time

#7) Testing Tools & Techniques

JmeterTo verify the Load and Performance of the ABC Chat application.

#8) Suspension and Resumption Criteria

Given below are the Critical Suspension and Resumption criteria which will impact the testing activities:

Environment not set upTesting cannot proceed

Environment readiness.
Application found to be unstableTesting cannot proceed.

Issue resolved
Test Data not availableTesting cannot proceed.

Test Data ready

#9) Test Deliverables

The Performance Test Deliverables include:

  • Performance Testing Strategy
  • Performance Requirements Document
  • Performance Test Scenario Document
  • Performance Test Scripts
  • Performance Test Results

#10) Roles & Responsibilities

Roles & Responsibilities are clearly explained in the table given below.

Roles & Responsibilities

#11) Potential Risks & Mitigation Plan

S.NoRiskProbabilityImpactMitigation PlanOwner
1Test Data unavailability for performance load test executionsHHEstimated dates for the performance test executions should be reviewed and updated. Functional/Dev team support required for data gathering.--
2Environmental IssuesLMRe prioritize Deliverables--
3Change in Functionality/design during performance test execution

MHThis requires rework on the performance test scenarios--
4Extra performance runs to troubleshoot performance issuesMHPerformance testing schedules would be modified and updated to the product team.--
5Estimations are prepared based on 1 bug fix build for performance. Multiple bug fix builds will delay test cycles and eventually it depends on when the next build will be available for rerun.HHRe prioritize the performance test execution cycles.--
6Hardware AvailabilityMHSchedule start date would be moved accordingly.--

#12) Assumptions

  • Performance Test Environment will be a replica of the product architecture landscape. (i.e. correct Hardware, Software, Interfaces, Integration Layers, etc).
  • Performance scripts will be designed based on the critical flows for which the usage is high.
  • All Infrastructure Issues should be resolved before the beginning of Performance testing. Any system configuration changes made later will invalidate the test results.
  • An application is stable and ready to use in the Performance test environment.
  • Necessary hardware and software resources (like load generator machines/software, controller/agent machines) are made available.
  • Any changes to the scope will go through a change control process and the performance testing team will assess the impact of timelines and resources.
  • Respective Servers are expected to handle the load.
  • Application trace logs have to be enabled for the supporting systems for monitoring purposes.

#13) Dependencies

  • Availability of the Performance test environment which is a replica of the product architecture landscape.
  • Support required from various Functional, Development, Database and Infrastructure teams during the test preparation and execution stages.
  • No code changes are implemented during the entire Performance testing phase as time is very limited.
  • In the event of unforeseen issues that lead to restrictions within the timelines, if timelines do not allow for all the test scopes to be met within the original milestone dates support is available from the Release Managers, to provide a scoping and prioritization decision.
  • Application Business Users / Subject Matter Experts will be made available for functional clarifications, and business transactions sign-off.
  • ABC chat Program Manager will review and sign-off.

#14) Abbreviations

HttpHyper Text Transfer Protocol
JDBCJava Database Connectivity
QAQuality Assurance
SLAService Level Agreement
SMESubject Matter Expert
UATUser Acceptance Testing

By now you must have clearly understood how to write an effective Performance Test Strategy for a Messaging application.

Best Practices For Realistic Performance Testing

In order to complete a Performance Test project successfully, we need to ensure that we are doing it in the right way from the planning stage ie planning, development, execution and analysis.

Let’s take a look at each stage in detail in order to conduct Performance Testing effectively.

#1) Planning

  • Try to identify the most common workflows i.e the business scenarios which have to be tested. If the application is an existing one, then check the server logs to understand the most frequently accessed scenarios. If the application is new than talk to the project management team to understand the major business flow.
  • Plan the Load test in such a way that you cover a broad range of workflows like light usage, medium usage and peak loads.
  •  You need to perform many cycles of the Load Test, so try to create a framework so that you can use the same scripts again and again. Also, try to have a backup of the scripts.
  • Try to analyze how long a test has to run, is it one hour? 8 hours? A day or a week? Usually, long-duration tests will uncover many major defects such as OS bugs, Memory leaks, etc.
  • If your organization is using any APM (Application Monitoring Tool), then you can include it during the test runs so that you can easily identify the performance issues and identify the root cause more easily.

#2) Development

  • While developing the scripts i.e recording, try to give a more meaningful transaction name based on the business flow names that are mentioned in the plan.
  • Don’t record any third-party applications and if it gets recorded, try to filter it out while enhancing the scripts.
  • Not all the dynamic values can be correlated using Autocorrelation feature in the tool, so try to do a manual correlation to avoid errors.
  • Try to design your performance tests in such a way that you are hitting the backend of the application and not just the cache server.

#3) Execution

  • Make sure to run the tests in a production-like environment, including factors like SSL, Load Balancer, and Firewalls. This is necessary to simulate a realistic load on the system.
  • Try to create a workload which is very realistic, you can get this by checking the server logs if it is an existing application and if it is a new application you need to get this info from the business team. Remember that workload is very important for conducting successful performance tests.
  • Never come to a conclusion by running tests with half the production size environment, it is always advised to conduct tests in an environment which is just the same as production.
  • While executing long-run tests try to watch the run at frequent intervals in order to make sure that the test is running smoothly.

#4) Analysis

  • Try to analyze the application by adding a few important counters first, when a bottleneck is found then try to add additional counters with respect to the bottleneck. This, in turn, will help in finding the issue more easily.
  • An application can fail for many reasons like it can fail to respond to a request, respond with an error code, fail your validation logic or responding too slowly. So try to look into all these before coming to a conclusion.


I’m sure that this tutorial would have given you immense knowledge on Performance tests and how to write a Performance Test Strategy Document with detailed examples.

In our upcoming tutorial, we will learn the differences between Performance, Load and Stress Testing in detail.

Also, Check => Free LoadRunner In-depth Training Series

NEXT Tutorial

Recommended Reading

78 thoughts on “A Complete Performance Testing Guide With Examples”

  1. Thank you so much for this great article. It really informative. Keep up the good work :)

  2. Can you please add tutorial of Jmeter.

  3. Nice article to know the funda of performance testing

  4. Many Thanks for sharing this articles.

  5. Can you please add tutorial for selenium & Jmeter.

  6. can u pls add tutorials for selemium & jmeter.

  7. Thanks for this article. Good information shred for performance testing basics.

    Along with Load runner can you please share tutorial for performance testing tool – Cloud test.

  8. when we will grt its second tutorial ? After 2 3 days ihave to work on load runner.so pls post it soon :(

  9. Great Documentation on the Performance testing terms (statements) .. :)

    Now, the question which always arrives in my mind about performance testing is –

    What is the actually time to do (start) performance testing ?

    As we know the performance testing deals with load, stress, Volume testing…., so is that mean we should do performance testing at the end of development process, or if not……. then what is the relevant way to begin with performance testing in Software Project ?

    Kindly elaborate your points on above issue.

  10. Thanks for such a informative articles on performance testing.

  11. How to decide the application performance is good or poor?
    Suppose there is application take 1 hr to send the 1GB file to database,So how to decide its performance?

  12. I would like to pursue my software testing career ahead into performance testing. what are the job opportunities available in this respect to performance testing.

    Please suggest

  13. Thanx Vijay for this greate intresting article……..thanx again…

  14. Thanks everybody for appreciating the article and work.

  15. @Kiran, @Udara- We will definitely consider your request. Thanks

  16. @Sachin-Previously when Performance testing was not given so much importance so it was done at the end and whatever were the minimum changes required to tune the application after the performance testing were considered . But now industry is very much aware of need of Performance testing. So the culture is now to have “Built to Performance” application where performance of application is checked from time to time.

    Developers check the performance of their code in unit testing through profilers. And when a functionality is ready to test, Performance team starts their performance testing and give feedback to Dev team so that they can fix the performance issues along with the release of next functionality. So with each release Dev team keep adding functionality and tuning the application in their integration phase.

  17. You have tremendously educate me about performance testing. Thanks.

  18. @ Ankit – Video tutorials will be on Monday.

  19. when can we get the next part of tutorial and practical session?

  20. Thank you and please also help us on Selenium…:)

    Thanks In Advance

  21. It’s great article for the Beginner’s.
    I would be expecting more practical example’s, apporach , and solution to the error’s which we encountered while creating Load Script’s and Performance Test.

    Hope you full fill my expectation.

    All the Best.

  22. Great article!! Can u please provide tutorials for JMeter…

  23. @ Chetan.

    Thanks a lot Sir !

    Firstly you understan dmy Question very much exact.
    and the answer is great along with the Software Development Track.

    Really Nice :)


  24. @ Yogesh.

    Below are two bullet points i would like to mention (for answering you) –

    1. PM (Product Manger) should decide the Performance limits. and if your Application (software) full fill the limits than the Performance is as Expected.

    In your example first you should know what is the expected behaviour (expected speed) and then you can say it is Good or Poor.

    2. Performance Testing has Load testing, Stress Testing, Volume Testing… (you can see in tutorial above).
    So if all this Testing Results are Good (Actual Result = Expected Result) then the performance is Good and if not you can guess it.

    Your nots are Welcome !

  25. Excellent article to understand the need of Performance Testing.

  26. Thanks. Excellent material.
    Looking forward to the tutorials which I expect to cover not merely basics.

  27. Thanks for your efforts Chetan Kaushal. i am the old reader of this blog can we start same tutorial for Jmeter I am interested in Jmeter because of open source, any help from your side will be helpful.

  28. Hello Sir,

    I am very thank full to you that u understand me, what is performance testing exactly…thanks alot my blassings are with you….

    Keep it up sir……

  29. Thankq so much for providing very useful information :-)

  30. Thanks for this Article. This is a nice Documentation on the Performance testing terms and statements.

    Looking forward to your next article.

  31. Thanks! Amazing artical on Performance Testing.
    Good Work Thank you so much.

  32. I just want to know that theoritically the article was best and can we have practise of this at home…I mean to say is it essential that we should work in a corporate for getting full hands on experience of load runner.

    We can learn it from here but where can we practise these steps any link from where we can download.

  33. Can some please tell me what problems they have faced doing perf testing and how it was resolved..?

  34. hello sir,actually i am a begginer for load runner and i know all the basics.but i want to learn loadrunner scripting using c.and i search everywhere.so can anyone please help me to how to start the scripting part.

  35. Hello Sir

    It was a wonderful article about Performance Testing.
    Was very interesting to go through …and was helpful as well.
    Thank you so much for taking the interest in creating such pages..

  36. Hi Vijay !

    I’ve one query about automation testing. Is there any open source automation testing tool available that support flex based applications?

    Need your thoughts on it.



  37. Hi Chetan,

    I am a manual testing professional with 5 yrs.of experience. I want to switch to automation industry now. Could I please request you to help me which automation tool I should go into. I have knowledge of qtp and have done 1 small project of mobile application as well. I have done functional testing of web services using SOAP UI but not fully into the tool. I am befazzled about which one I should go in? Which tool/technology will help me in career growth? I thought to prepare loadrunner by myself, hence explored your site and found it very useful n informative. Could you pl.suggest will these tutorials of loadrunner be enough to enter in performance testing automation or to clear interview? Will you cover any project to make us understand practically how to start performance testing using loadrunner? Is there any programming language or scripting lang.required to learn loadrunner? I am not able to take any decision in regardence of qtp, selenium, loadrunner, mobile testing, soap ui…please help me.Looking forward for your suggestions, I hope you won’t disappoint me. Thanks n Best Regards!

  38. I wanted to learn testingtools. .can you pls help me out for complete course soft copy. ..&materials which u have…
    I am middle class family.i can’t pay money…I am providing my contact& e mail ids…pls share d video tutorials to my mail id..

    Mob:8099174104 email:kvrsrikanth6@gmail.com. ..waiting

  39. Highly informative. Thanks for sharing.

  40. Good and basic required information provided.
    Good for beginers.

  41. hello sir,
    This article is very helpfull to me. I have two year experiance on PHP development, But now I want to switch into testing field.Currently I am not working anywhere. I dont have any idea about automation testing. Will you please help me to make my carrer in software testing.

  42. I need to test a software for performance which will have lot of servers in it with proper IP Address, does any of the performance testing tools support creating virtual servers as many as needed for testing.

  43. Good 1 to a fresher whoz going 2 start the career of performance testing

  44. I want to thank for the information, it really helpful to know a bit of what is performance testing? . I have one year experience in Manual Testing with background of B.E (ECE), As I am planning to learn Loadrunner,but i don’t have any programming or scripting language so far..Is that create any conflict to learn loadrunner..or do i need to know any programming language to learn..please advise i will decide my career based on your kind reply

  45. It is good article for beginner in performance Testing

  46. I want to learn how the performance is tested using load runner or any other testing tool.
    If any one is having notes or any study material please forward it to my mail id.

  47. It is really great one for beginner


Leave a Comment