Sei sulla pagina 1di 65

1. What is the MAIN benefit of designing tests early in the life cycle?

It helps prevent defects from being introduced into the code.

2. What is risk-based testing?

Risk-based testing is the term used for an approach to creating a test strategy that is based
on prioritizing tests by risk. The basis of the approach is a detailed risk analysis and
prioritizing of risks by risk level. Tests to address each risk are then specified, starting with
the highest risk first.

3. A wholesaler sells printer cartridges. The minimum order quantity is 5. There is a


20% discount for orders of 100 or more printer cartridges. You have been asked to
prepare test cases using various values for the number of printer cartridges ordered.
Which of the following groups contain three test inputs that would be generated
using Boundary Value Analysis?

4, 5, 99

4. What is the KEY difference between preventative and reactive approaches to


testing?

Preventative tests are designed early; reactive tests are designed after the software has
been produced.

5. What is the purpose of exit criteria?

To define when a test level is complete.

6. What determines the level of risk?

The likelihood of an adverse event and the impact of the event

7. When is used Decision table testing?

Decision table testing is used for testing systems for which the specification takes the
form of rules or cause-effect combinations. In a decision table the inputs are listed in a
column, with the outputs in the same column but below the inputs. The remainder of the
table explores combinations of inputs to define the outputs produced.
Learn More About Decision Table Testing Technique in the Video Tutorial here

8. What is the MAIN objective when reviewing a software deliverable?

To identify defects in any software work product.

9. Which of the following defines the expected results of a test? Test case
specification or test design specification.

Test case specification.

10. Which is a benefit of test independence?

It avoids author bias in defining effective tests.

11. As part of which test process do you determine the exit criteria?

Test planning.

12. What is beta testing?

Testing performed by potential customers at their own locations.

13. Given the following fragment of code, how many tests are required for 100%
decision coverage?

if width > length

then biggest_dimension = width

if height > width

then biggest_dimension = height

end_if

else biggest_dimension = length

if height > length

then biggest_dimension = height

end_if
end_if

14. You have designed test cases to provide 100% statement and 100% decision
coverage for the following fragment of code. if width > length then
biggest_dimension = width else biggest_dimension = length end_if The following has
been added to the bottom of the code fragment above. print "Biggest dimension is "
& biggest_dimension print "Width: " & width print "Length: " & length How
many more test cases are required?

None, existing test cases can be used.

15. Rapid Application Development ?

Rapid Application Development (RAD) is formally a parallel development of functions


and subsequent integration. Components/functions are developed in parallel as if they
were mini projects, the developments are time-boxed, delivered, and then assembled into
a working prototype. This can very quickly give the customer something to see and use
and to provide feedback regarding the delivery and their requirements. Rapid change and
development of the product is possible using this methodology. However the product
specification will need to be developed for the product at some point, and the project will
need to be placed under more formal controls prior to going into production.

16. What is the difference between Testing Techniques and Testing Tools?

Testing technique: – Is a process for ensuring that some aspects of the application system
or unit functions properly there may be few techniques but many tools.

Testing Tools: – Is a vehicle for performing a test process. The tool is a resource to the
tester, but itself is insufficient to conduct testing

Learn More About Testing Tools here

17. We use the output of the requirement analysis, the requirement specification as
the input for writing …

User Acceptance Test Cases

18. Repeated Testing of an already tested program, after modification, to discover


any defects introduced or uncovered as a result of the changes in the software being
tested or in another related or unrelated software component:

Regression Testing
19. What is component testing ?

Component testing, also known as unit, module and program testing, searches for defects
in, and verifies the functioning of software (e.g. modules, programs, objects, classes, etc.)
that are separately testable. Component testing may be done in isolation from the rest of
the system depend-ing on the context of the development life cycle and the system. Most
often stubs and drivers are used to replace the missing software and simulate the interface
between the software components in a simple manner. A stub is called from the software
component to be tested; a driver calls a component to be tested.

Here is an aswesome video on Unit Testing

20. What is functional system testing ?

Testing the end to end functionality of the system as a whole.

21. What is the benefits of Independent Testing

Independent testers see other and different defects and are unbiased.

22. In a REACTIVE approach to testing when would you expect the bulk of the test
design work to be begun?

After the software or system has been produced.

23. What are the different Methodologies in Agile Development Model?

There are currently seven different Agile methodologies that I am aware of:

1. Extreme Programming (XP)


2. Scrum
3. Lean Software Development
4. Feature-Driven Development
5. Agile Unified Process
6. Crystal
7. Dynamic Systems Development Model (DSDM)

24. Which activity in the fundamental test process includes evaluation of the
testability of the requirements and system?

A Test analysis and design.

25. What is typically the MOST important reason to use risk to drive testing efforts?

Because testing everything is not feasible.


26. Which is the MOST important advantage of independence in testing?

An independent tester may be more effective at finding defects missed by the person who
wrote the software.

27. Which of the following are valid objectives for incident reports?

i. Provide developers and other parties with feedback about the problem to enable
identification, isolation and correction as necessary.

ii. Provide ideas for test process improvement.

iii. Provide a vehicle for assessing tester competence.

iv. Provide testers with a means of tracking the quality of the system under test.

i. Provide developers and other parties with feedback about the problem to enable
identification, isolation and correction as necessary,

ii.Provide ideas for test process improvement,

iv.Provide testers with a means of tracking the quality of the system under test

28. Consider the following techniques. Which are static and which are dynamic
techniques?

i. Equivalence Partitioning.

ii. Use Case Testing.

iii.Data Flow Analysis.

iv.Exploratory Testing.

v. Decision Testing.

vi. Inspections.

Data Flow Analysis and Inspections are static, Equivalence Partitioning, Use Case
Testing, Exploratory Testing and Decision Testing are dynamic.

29. Why are static testing and dynamic testing described as complementary?
Because they share the aim of identifying defects but differ in the types of defect they
find.

30. What are the phases of a formal review ?

In contrast to informal reviews, formal reviews follow a formal process. A typical formal
review process consists of six main steps:

1. Planning
2. Kick-off
3. Preparation
4. Review meeting
5. Rework
6. Follow-up.

31. What is the role of moderator in review process?

The moderator (or review leader) leads the review process. He or she deter-mines, in co-
operation with the author, the type of review, approach and the composition of the review
team. The moderator performs the entry check and the follow-up on the rework, in order
to control the quality of the input and output of the review process. The moderator also
schedules the meeting, disseminates documents before the meeting, coaches other team
members, paces the meeting, leads possible discussions and stores the data that is
collected.

Learn More About Review process in Video Tutorial here

32. What is an equivalence partition (also known as an equivalence class)?

An input or output range of values such that only one value in the range becomes a test
case.

33. When should configuration management procedures be implemented?

During test planning.

34. A Type of functional Testing, which investigates the functions relating to


detection of threats, such as virus from malicious outsiders.

Security Testing
35. Testing where in we subject the target of the test , to varying workloads to
measure and evaluate the performance behaviors and ability of the target and of the
test to continue to function properly under these different workloads. Load Testing

36. Testing activity which is performed to expose defects in the interfaces and in the
interaction between integrated components is:

Integration Level Testing

37. What are the Structure-based (white-box) testing techniques ?

Structure-based testing techniques (which are also dynamic rather than static) use the
internal structure of the software to derive test cases. They are com-monly called 'white-
box' or 'glass-box' techniques (implying you can see into the system) since they require
knowledge of how the software is implemented, that is, how it works. For example, a
structural technique may be concerned with exercising loops in the software. Different
test cases may be derived to exercise the loop once, twice, and many times. This may be
done regardless of the func-tionality of the software.

38. When should be performed Regression testing ?

After the software has changed or when the environment has changed

39. When should testing be stopped?

It depends on the risks for the system being tested

40. What is the purpose of a test completion criterion?

To determine when to stop testing

41. What can static analysis NOT find?

For example memory leaks

42. What is the difference between re-testing and regression testing?

Re-testing ensures the original fault has been removed; regression testing looks for
unexpected sideeffects

43. What are the Experience-based testing techniques ?

In experience-based techniques, people's knowledge, skills and background are a prime


contributor to the test conditions and test cases. The experience of both technical and
business people is important, as they bring different perspectives to the test analysis and
design process. Due to previous experience with similar systems, they may have insights
into what could go wrong, which is very useful for testing.

44. What type of review requires formal entry and exit criteria, including metrics?
Inspection 45. Could reviews or inspections be considered part of testing?

Yes, because both help detect faults and improve quality

46. An input field takes the year of birth between 1900 and 2004 What are the
boundary values for testing this field ? 1899,1900,2004,2005

47. Which of the following tools would be involved in the automation of regression
test? a. Data tester b. Boundary tester c. Capture/Playback d. Output comparator.

d. Output comparator

48. To test a function,what has to write a programmer, which calls the function to be
tested and passes it test data.

Driver

49. What is the one Key reason why developers have difficulty testing their own
work?

Lack of Objectivity

50.“How much testing is enough?”

The answer depends on the risk for your industry, contract and special requirements. 51.
When should testing be stopped? It depends on the risks for the system being tested.

52. Which of the following is the main purpose of the integration strategy for
integration testing in the small?

To specify which modules to combine when, and how many at once.

53. What is the purpose of a test completion criterion?

To determine when to stop testing

54. Given the following code, which statement is true about the minimum number of
test cases required for full statement and branch coverage?

Read p

Read q
IF p+q> 100

THEN Print "Large"

ENDIF

IF p > 50

THEN Print "p Large"

ENDIF

1 test for statement coverage, 2 for branch coverage

55. What is the difference between re-testing and regression testing?

Re-testing ensures the original fault has been removed; regression testing looks for
unexpected side-effects.

56. Which review is normally used to evaluate a product to determine its suitability
for intended use and to identify discrepancies?

Technical Review.

57. Why we use decision tables?.

The techniques of equivalence partitioning and boundary value analysis are often applied
to specific situations or inputs. However, if different combinations of inputs result in
different actions being taken, this can be more difficult to show using equivalence
partitioning and boundary value analysis, which tend to be more focused on the user
interface. The other two specification-based tech-niques, decision tables and state
transition testing are more focused on business logic or business rules. A decision table is
a good way to deal with combinations of things (e.g. inputs). This technique is sometimes
also referred to as a 'cause-effect' table. The reason for this is that there is an associated
logic diagramming technique called 'cause-effect graphing' which was sometimes used to
help derive the decision table

58. Faults found should be originally documented by who?

By testers.

59. Which is the current formal world-wide recognized documentation standard?

There isn’t one.


60. Which of the following is the review participant who has created the item to be
reviewed?

Author

61. A number of critical bugs are fixed in software. All the bugs are in one module,
related to reports. The test manager decides to do regression testing only on the
reports module.

Regression testing should be done on other modules as well because fixing one module
may affect other modules.

62. Why does the boundary value analysis provide good test cases?

Because errors are frequently made during programming of the different cases near the
‘edges’ of the range of values.

63. What makes an inspection different from other review types?

It is led by a trained leader, uses formal entry and exit criteria and checklists.

64. Why can be tester dependent on configuration management?

Because configuration management assures that we know the exact version of the
testware and the test object.

65. What is a V-Model ?

A software development model that illustrates how testing activities integrate with
software development phases

66. What is maintenance testing?

Triggered by modifications, migration or retirement of existing software

67. What is test coverage?

Test coverage measures in some specific way the amount of testing performed by a set of
tests (derived in some other way, e.g. using specification-based techniques). Wherever we
can count things and can tell whether or not each of those things has been tested by some
test, then we can measure coverage.

68. Why is incremental integration preferred over “big bang” integration?

Because incremental integration has better early defects screening and isolation ability
69. When do we prepare RTM (Requirement traceability matrix), is it before test
case designing or after test case designing?

The would be before. Requirements should already be traceable from Review activities
since you should have traceability in the Test Plan already. This question also would
depend on the organisation. If the organisation do test after development started then
requirements must be already traceable to their source. To make life simpler use a tool to
manage requirements.

70. What is called the process starting with the terminal modules ?

Bottom-up integration

71. During which test activity could faults be found most cost effectively?

During test planning

72. The purpose of requirement phase is

To freeze requirements, to understand user needs, to define the scope of testing

73. How much testing is enough?

The answer depends on the risks for your industry, contract and special requirements 74.
Why we split testing into distinct stages? Each test stage has a different purpose.

75. Which of the following is likely to benefit most from the use of test tools
providing test capture and replay facilities? a) Regression testing b) Integration
testing c) System testing d) User acceptance testing

Regression testing

76. How would you estimate the amount of re-testing likely to be required?

Metrics from previous similar projects and discussions with the development team

77. What studies data flow analysis ?

The use of data on paths through the code.

78. What is Alpha testing?

Pre-release testing by end user representatives at the developer’s site.

79. What is a failure?


Failure is a departure from specified behaviour.

80. What are Test comparators ?

Is it really a test if you put some inputs into some software, but never look to see whether
the software produces the correct result? The essence of testing is to check whether the
software produces the correct result, and to do that, we must compare what the software
produces to what it should produce. A test comparator helps to automate aspects of that
comparison.

81. Who is responsible for document all the issues, problems and open point that
were identified during the review meeting Scribe

82. What is the main purpose of Informal review

Inexpensive way to get some benefit

83. What is the purpose of test design technique?

Identifying test conditions and Identifying test cases

84. When testing a grade calculation system, a tester determines that all scores from
90 to 100 will yield a grade of A, but scores below 90 will not. This analysis is known
as:

Equivalence partitioning

85. A test manager wants to use the resources available for the automated testing of
a web application. The best choice is Tester, test automater, web specialist, DBA

86. During the testing of a module tester ‘X’ finds a bug and assigned it to developer.
But developer rejects the same, saying that it’s not a bug. What ‘X’ should do?

Send to the detailed information of the bug encountered and check the reproducibility

87. A type of integration testing in which software elements, hardware elements, or


both are combined all at once into a component or an overall system, rather than in
stages.

Big-Bang Testing

88. In practice, which Life Cycle model may have more, fewer or different levels of
development and testing, depending on the project and the software product. For
example, there may be component integration testing after component testing, and
system integration testing after system testing.
V-Model

89. Which technique can be used to achieve input and output coverage? It can be
applied to human input, input via interfaces to a system, or interface parameters in
integration testing.

Equivalence partitioning

90. “This life cycle model is basically driven by schedule and budget risks” This
statement is best suited for…

V-Model

91. In which order should tests be run?

The most important tests first

92. The later in the development life cycle a fault is discovered, the more expensive it
is to fix. why?

The fault has been built into more documentation,code,tests, etc

93. What is Coverage measurement?

It is a partial measure of test thoroughness.

94. What is Boundary value testing?

Test boundary conditions on, below and above the edges of input and output equivalence
classes.

95. What is Fault Masking ?

Error condition hiding another error condition.

96. What does COTS represent?

Commercial Off The Shelf.

97.The purpose of wich is allow specific tests to be carried out on a system or


network that resembles as closely as possible the environment where the item under
test will be used upon release?

Test Environment
98. What can be though of as being based on the project plan, but with greater
amounts of detail?

Phase Test Plan

99. What is exploratory testing?

Exploratory testing is a hands-on approach in which testers are involved in minimum


planning and maximum test execution. The planning involves the cre-ation of a test
charter, a short declaration of the scope of a short (1 to 2 hour) time-boxed test effort, the
objectives and possible approaches to be used. The test design and test execution
activities are performed in parallel typi-cally without formally documenting the test
conditions, test cases or test scripts. This does not mean that other, more formal testing
techniques will not be used. For example, the tester may decide to use boundary value
analysis but will think through and test the most important boundary values without
necessarily writing them down. Some notes will be written during the exploratory-testing
session, so that a report can be produced afterwards.

100. What is failure?

Deviation from expected result to actual result

Why Error Guessing is significant in testing?

Error Guessing is not a testing type nor testing method, does Error
Guessing always come with the experience. Usually error guessing
scope rely on the software tester involved which means who uses past
experience and intuition to determine what situations commonly cause
software or application failure, or which functionality leads to break or
release blocks.

There are no particular rules to define error guessing but yes always
Error Guessing will conduct on Ad-hoc basis. Whoever evolving the
Error Guessing testing should have complete knowledge on application
or Software functionality.

Now let’s discuss why Error Guessing is significant in Testing, Not


only testing Error Guessing significant for Development also. Always it
helps to find more valuable defects or issues can find prior end users
or customers that help to the Organization to get good reputation from
customers and also it reduces huge cost of maintenance.

0 comments

What is requirement traceability matrix and explain Importance of


Traceability matrix?

Requirement Trace-ability matrix is matrix to map between


requirements and test cases or test scenarios. Normally testing team
should involve rite from the requirement phase for any project if
requirements are clear to develop application or product. If testing
team gets involved right from the requirement phase then requirement
traceability is one of the important documents that can detail what
kind of test coverage the test cases or scenarios have.

Below are elements should present in Traceability Matrix:

Project Name: For Which project Traceability matrix preparing

Project Manager Name:

Project Description: Description of the Project

For each issue identified, complete the following:

ID: A unique ID number used to identify the traceability item in the


requirements traceability matrix.

Associated ID(s): This column should contain the ID of any


associated utilities used for requirements tracking such as a
repository, pipeline document, etc

Technical Assumption(s) or Customer Need(s): This column


should be populated with a description of the technical assumption or
customer need linked to the functional requirement.

Functional Requirement: This column should be populated with a


description of the functional requirement.

Status: This column should be populated with the current status of


the functional requirement.
Architectural/Design Document: This column should be populated
with a description of the architectural/design document linked to the
functional requirement.

Technical Specification: This column should be populated with a


description of the technical specification linked to the functional
requirement.

System Component(s): This column should be populated with a


description of the system component(s) linked to the functional
requirement.

Software Module(s): This column should be populated with a


description of the software module(s) linked to the functional
requirement.

Test Case Number: This column should be populated with the test
case number linked to the functional requirement.

Tested In: This column should be populated with the module that the
functional requirement has been tested in.

Implemented In: This column should be populated with the module


that the functional requirement has been implemented in.

Verification: This column should be populated with a description of


the verification document linked to the functional requirement.

Additional Comments: This column should be populated with any


additional comments

3 comments

Explain Work bench concept Phases?

Below are the phases of Bench mark concept:

workbench: phase Requirement For this phase the input will be


the customer’s or end users requirements and team execute the task
of writing a requirement documents.
workbench: phase Design In this phase the input will be the
requirement documents and team execute the task of preparing a
technical document.

workbench: phase Execution In this phase actual execution starts.


The input will be the technical document.

workbench: phase Testing This phase is the actual testing phase of


the project. The input is the source code which needs to be tested.

workbench: phase Deployment This is the deployment phase.


There are two inputs for this phase: one is the source code which
needs to be deployed and that is dependent on the test results. The
output of this project is that the customer gets the product which he
can now start using.

workbench: phase Maintenance The input to this phase is the


deployment results, execution is implementing change requests from
the end customer, the check part is nothing but running regression.
3 comments

Explain how to define a testing policy?

Explain how to define a testing policy?

During the Interview process this question asks to see whether


interviewee can able handle the team and follow the life cycle or not.
In the market still there are so many companies think testing is
secondary than development so testing managers should show the
importance of testing in terms of Test approaches, testing policies and
testing processes.

Below are the few steps to define a testing policy in general but these
are not a standard steps again it will vary company to company.

Definition: Any company should start with define one unique


definition for testing within the company because everybody needs to
be with the same mind set.

achieve: to How In this step needs to define how are we going to


achieve our objective? Is there going to be a testing committee, will
there be compulsory test plans which need to be executed, etc...
Evaluate: And in the evaluate step after testing is implemented for a
particular application or project how do we evaluate it? Are we going
to derive metrics of defects per phase, per programmer, etc.? Finally,
it’s important to let everyone know how testing has added value to the
project.

Standards: And in this step what are the standards we want to


achieve by testing the application for instance, we can say that more
than 20 defects per KLOC will be considered below standard and code
review should be done for it.
1 comments

What is Security testing and what are the main things to test in Security
Testing?

Whenever we develop any web applications security testing should be


on top priority basically for Finance domain and banking applications.
Commonly in security testing below terms uses most of the times.

- Password cracking
- Vulnerability
- URL manipulation
- SQL injection
- Cross Site Scripting
- Spoofing

Below are few things needed to concentrate while doing security


testing:

- Authentication validations and Password protection


- Direct URL’s should not work after logging to the application
- HTTP and HTTP’s validations
- Protocols and IP config validations
- Memory leeks
- Configuration of the application in servers
1 comments

Explain Software QA Roles and Responsibilities

Trainee QA Responsibilities:

 Read and understand the Business requirements


 Read and understand QA test plan and approach
 Prepare Test Scenarios
 Prepare Test cases
 Test Execution
 Defect raising

Senior QA Responsibilities:
 Read and understand the requirements
 Prepare module level or iteration level QA Test plan and
approach
 Prepare test scenarios and Test cases
 Prepare Automation test scripts
 Review Test scenarios and Test cases prepared by team
members
 Review Defects raised by team members and Defect analysis
 daily status reports
 Schedule KT session to team members

QA Lead Responsibilities:
 Under stand the Business by attending meetings with the client
and business analysts
 Prepare whole project Test plan and Approach
 Prepare Daily and weekly status reports formats
 Send weekly execution status reports to the shake holders
 Resource planning
 Have calls with customers weekly to update the project status
and test execution status
 Automation frame works
 Review the test design and automation scripts very high level

3 comments

Explain the concept of defect cascading?

Defect cascading explains how one defect leads to other defects.


Example: In the above flow chart for an employee application salary
calculations are not correct when totaling Basic, Tax and Gross, so for
this issue defect raised. In this case other modules will block and also
same defect will exists in all those modules due to this defect. This is
called Defect Cascading.

1 comments

What are the different defect categories and explain?

We have 3 types of Defect Categories:


 Wrong
 Missing
 Extra

Wrong: This defect category is a variance from the given requirement


specification. This means if there are any discrepancies in requirement
documents it will fall under this category.

Missing: During the development or design if we miss any


requirement which is given by the customer those type of defect
categories fall under this. This type of missing causes by when
requirement explain by the customer was not noted properly and
customer requirement not explained to the team properly when they
were developed.

Extra: During the development if a new requirement added to the


existing requirements which is not given by the customer those
changes or defects fall under this category.
0 comments

What is PDCA cycle?

PDCA cycle depends on Software development life cycle, in any


Software Development process Software testing plays key role always.
In software development process there should be four important steps,
they are Plan, Do, Check and Act which refers PDCA cycle.

In PDCA cycle actual test execution will start during the Check Stage,
in Plan and Do phases management and leads will come up with the
plan and strategy and in Check stage testers will start execution and in
Act stage leads and management take decisions based on the issues
testers facing in Check stage.

Detail explanation of above four steps:

Plan: Defines what to test, where to test and who to test and explains
time lines and testing cycles.

Do:Executing the above plan with the time lines specified  

Check: Verifying and cross checking whether everything going as


planned and within the project time lines

Act: Taking decision and supporting the team if any issues facing in
above steps
1 comments

Explain few Software Development Models?

Spiral Model: The spiral model is a Software Development model in


this model there are number of cycles (or spirals) of all the sequential
steps of the waterfall model. Whenever first iteration or initial cycle
complete then will go for the reviews and demos of the developed
model if it is not as per the specified requirements or expected
standards, a second cycle follows, and so on. Always Spiral model or
method follows an iterative approach and mainly it is useful for very
large projects having complex and constantly changing requirements.

Rational Unified Process (RUP): In Rational Unified Process each


cycle should have four phases those are inception, elaboration,
construction and transition. At the end of each cycle conduct reviews
of the product or the application and also further cycle follows if
necessary. This model also similar to the spiral model because entire
cycle broken up into multiple cycles or processes.

Agile Model: Now a days most of the Organizations adapting Agile


Methodology because Agile methodology is more flexible to continue
Product or application development. Agile methodology is a selective
mix of both Spiral and RUP approaches in addition to quite a few new
developmental methods. Fast and incremental development is one of
the key principles of this methodology. The focus is on obtaining quick,
practical and visible outputs and results, rather than merely following
theoretical processes.

Rapid Application Development (RAD): It is very rare approach to


follow; very few Organizations are following this method.
3 comments

How to control warnings from the “Test results page” in QTP?

To Suppress warnings from the Test results Viewer go to “Tools >


Filters > and un-check Warnings” check box.

Explain QuickTest Window in detail?

The QTP window contains the following key elements:

Title bar: Displays the title of the currently open test.


Testing toolbar: Contains buttons to assist the testing process.
Menu bar: Displays menus of QTP commands.
Test pane: Contains the Keyword View and Expert View tabs.
Active Screen: Provides a snapshot of your application as it appeared
when you performed a certain step during the recording session.
File toolbar: Contains buttons to assist you in managing your test.
Debug toolbar: Contains buttons to assist you in debugging tests.
Data Table: Assists you in parameterization of your test.
Status bar: Displays the status of the QuickTest application.
Debug Viewer pane: Assists in debugging tests or scripts. The Debug
Viewer pane contains the Watch Expressions, Variables, and Command
tabs.
0 comments

What are the extension of script and object repository files in QTP?

QTP Object Repository extension: .tsr


QTP Script extension: .mts

How to identify Object while Recording script in QTP?

While recording any script it stores Object as Test Object. For each and
every test object class checks whether this description is enough to
uniquely identify the object. If it is not, QTP adds assistive properties
to the test objects one by one to the description until it has a unique
description.

How to identify Object during test run?

When ever test scripts runs it searches for a run-time object that
exactly matches the description of the test object. Always it will
expects to find a perfect match for both the mandatory and any
Assistive properties of test objectuses Smart Identification mechanism
to identify an object, even when the recorded description is no longer
accurate.

0 comments

How to do Load testing for web based Application by using QTP?

First of all QTP is not a right option for the Loadtesting,the best option
will be Loadrunner.

1. Open web application and record script of particular scenario in


QTP
2. Make 100 copies of that script of particular scenario and run the test
3. In that case, do the load of application on server.
4. The basic logic of running the copy 100 times is to create same
scenario as if 100 users were working.
0 comments

How to open a new script using QTP?


To open any script in QTP follow below steps:

First Method: If QTP not yet open then go to Start > Programs >
QuickTest Professional > QuickTest Professional then welcome window
opens now click Blank Test.

a. The other way is go to File menu > and click New then a blank test
opens.

b. Another way is or click the New button on the tool bar then a blank
test opens.

Second Method: If QTP is already open then check which add-ins are
loaded by selecting Help > About QuickTest Professional. If the Web
Add-in is not loaded, you must exit and restart QuickTest. When the
Add-in Manager opens, select the Web Add-in, and clear all other add-
ins. Go to File menu > click on New then blank test opens. Another
way is click the New button on the tool bar.

Explain Test Fusion Report of QTP ?

After Test run comeplete Test Fusion report displays all aspects of
the test run. a high-level results overview, an expandable Tree View of
the test specifying exactly where application failures occurred, the test
data used, application screen shots for every step that highlight any
discrepancies, and detailed explanations of each checkpoint pass and
failure. By combining TestFusion reports with QuickTest Professional,
you can share reports across an entire QA and development team.

1 comments

How to Run a Test Scripts using QTP?

We can run Test Scripts in 2 different ways:

1. Start QuickTest and open your test.

2. If QuickTest is not already open, choose Start > Programs >


QuickTest Professional then welcome window opens, click on Open
Existing.

Now Click Run button or go to Test >and click Run then Run dialog box
will open. Then select New run results folder and accept the default
results folder name and click OK button then Run dialog box will close.
How to confirm that all images are saved to the test results or
not?

QTP allows you to determine when to save images to the test results.
Choose Tools > Options and select the Run tab. In the Save step
screen capture to test results option, select Always then click OK
button to close the Options dialog box. Then automatically all the
images will save in the test results.

0 comments

How to Save recorded or written script using QTP?

Select File > and click Save or the another way is click the Save button
on tool bar then Save dialog box opens to the Tests folder.Now create
a folder to save the script then type your test name in the File name
field and click Save. After successfully script saved test name should
display in the title bar of the main QuickTest window.

How many types of Object Repository's available in QTP?

QTP have two types of Object Repository's


1. Shared Object Repository
2. Action Object Repository

The object repository per-action mode is the default setting. and


during the test script creation we can choose eassly which object
repository is the default one. If we modify values in an action object
repository, the changes do not have any effect on other actions.But , if
the same test object exists in more than one action and we modify an
object’s property values in one action then we may need to make the
same change in every action containing the object.

0 comments

How to insert a Image Check point to check enable property in QTP?

All images are as push button than we can check the property enabled
or disabled. After verifying not able to find the image property then
have to go object repository for specific image object and click on
add/Remove button to add the available properties to the image
object.
And if we take it as image then we have to check visible or invisible
property that will help to know enable or disable properties for the
image object.

Another way:

The Image Checkpoint does not have any property to verify the
enable/disable property.Here one thing we have to check that is first
find out from the Developer if developer is showing different images
for activating/deactivating for example grayed out image.

That is the only way a developer can show deactivate/activate if


developer is using an “image” else developer might be using a button
having a heads up with an image.If it is a button used to display with
the heads up as an image have to use the object Properties as a
checkpoint.
0 comments

How to Start recording using QTP?

Below are the few ways to record script in QTP:

To Open QTP : choose Start > Programs > QuickTest Professional >
QuickTest Professional. or click QuickTest short cut key on Desktop.
After following the above step Add-in Manager dialog box opens, to
confirm that select Web Add-in and clear all other add-ins then click
OK to close the Add-in Manager and open QuickTest.

If QuickTest is already open, check which add-ins are loaded by


selecting Help > About QuickTest Professional. If the Web Add-in is not
loaded, you must exit and restart QuickTest. When the Add-in
Manager opens, select the Web Add-in, and clear all other add-ins.
Now choose File > New, or click the New button.A blank test opens.

Now Start recording:

1. Choose Test > Record or click the Record button The Record and
Run Settings dialog box opens.
2. In the Web tab, select Open the following browser when a record or
run session begins.
3. Now select browser from the Type list and confirm that the
URL(application URL) in the Address box.
4. Next in the Windows Applications tab, confirm that Record and run
on those applications is selected
5. And finally Click OK.
6. Now QTP begins recording, and browser opens the specified web
application

0 comments

What is the QTP testing process?

QTP testing process have seven main steps:

1. Preparing to recording
2. Recording
3. Enhancing your script
4. Debugging
5. Run
6. Analyze
7. Report Defects (more…)

How QTP recognizes Objects in AUT?

QTP stores the definitions for application objects in a file called the
Object Repository. QTP will add an entry for each item what we
interact with which mean all the objects of an application. Each Object
Repository entry will be identified by a logical name, and will contain a
set of properties that uniquely identify each object.

Each line in the script will contain a reference to the object that we
interacted with, a call to the appropriate method and any parameters
for that method. The references to objects in the script will all be
identified by the logical name, rather than any physical, descriptive
properties.

1 comments

How many types of Parameters are available in QTP?

We have 3 types of Parameters available in QTP:

1. Method Argument
2. Data Driven
3. Dynamic

What is keyword view and Expert view in QTP?

QTP have two views those are Expert view and Keyword View, The
Expert View provides an alternative to the Tree View for testers who
are familiar with VBScript. The Expert View shows the recorded scripts
in VBScript and enhance it with programming. And The Expert View
displays the steps which are executed while recording in VBScript.
After recording any test, can increase its power and flexibility by
adding recordable and non-recordable VBScript statements.

Advanced testers who are know the VB script can view and edit their
tests in the Expert View, which reveals the underlying industry-
standard VBScript that QuickTest Professional automatically generates.
Any changes made in the Expert View are automatically synchronized
with the Keyword View.

Verification: Verification is finding defects or faults by reading out the documents.


Example-review of any document

Validation: Validation is finding defects by executing test cases or testing the product.

1. What is Acceptance Testing?


Testing conducted to enable a user/customer to determine whether to
accept a software product. Normally performed to validate the
software meets a set of agreed acceptance criteria.

2. What is Accessibility Testing?


Verifying a product is accessible to the people having disabilities (deaf,
blind, mentally disabled etc.).

3. What is Ad Hoc Testing?


A testing phase where the tester tries to 'break' the system by
randomly trying the system's functionality. Can include negative
testing as well. See also Monkey Testing.

4. What is Agile Testing?


Testing practice for projects using agile methodologies, treating
development as the customer of testing and emphasizing a test-first
design paradigm. See also Test Driven Development.

5. What is Application Binary Interface (ABI)?


A specification defining requirements for portability of applications in
binary forms across different system platforms and environments.

6. What is Application Programming Interface (API)?


A formalized set of software calls and routines that can be referenced
by an application program in order to access supporting system or
network services.

7. What is Automated Software Quality (ASQ)?


The use of software tools, such as automated testing tools, to improve
software quality.

8. What is Automated Testing?


Testing employing software tools which execute tests without manual
intervention. Can be applied in GUI, performance, API, etc. testing.
The use of software to control the execution of tests, the comparison
of actual outcomes to predicted outcomes, the setting up of test
preconditions, and other test control and test reporting functions.

9. What is Backus-Naur Form?


A metalanguage used to formally describe the syntax of a language.

10. What is Basic Block?


A sequence of one or more consecutive, executable statements
containing no branches.

11. What is Basis Path Testing?


A white box test case design technique that uses the algorithmic flow
of the program to design tests.

12. What is Basis Set?


The set of tests derived using basis path testing.

13. What is Baseline?


The point at which some deliverable produced during the software
engineering process is put under formal change control.

14. What you will do during the first day of job?


What would you like to do five years from now?

15. What is Beta Testing?


Testing of a release of a software product conducted by customers.

16. What is Binary Portability Testing?


Testing an executable application for portability across system
platforms and environments, usually for conformation to an ABI
specification.

17. What is Black Box Testing?


Testing based on an analysis of the specification of a piece of software
without reference to its internal workings. The goal is to test how well
the component conforms to the published requirements for the
component.

18. What is Bottom Up Testing?


An approach to integration testing where the lowest level components
are tested first, then used to facilitate the testing of higher level
components. The process is repeated until the component at the top of
the hierarchy is tested.

19. What is Boundary Testing?


Test which focus on the boundary or limit conditions of the software
being tested. (Some of these tests are stress tests).

20. What is Bug?


A fault in a program which causes the program to perform in an
unintended or unanticipated manner.

20. What is Defect?


If software misses some feature or function from what is there in
requirement it is called as defect.

21. What is Boundary Value Analysis?


BVA is similar to Equivalence Partitioning but focuses on "corner cases"
or values that are usually out of range as defined by the specification.
his means that if a function expects all values in range of negative 100
to positive 1000, test inputs would include negative 101 and positive
1001.

22. What is Branch Testing?


Testing in which all branches in the program source code are tested at
least once.

23. What is Breadth Testing?


A test suite that exercises the full functionality of a product but does
not test features in detail.

24. What is CAST?


Computer Aided Software Testing.

25. What is Capture/Replay Tool?


A test tool that records test input as it is sent to the software under
test. The input cases stored can then be used to reproduce the test at
a later time. Most commonly applied to GUI test tools.

26. What is CMM?


The Capability Maturity Model for Software (CMM or SW-CMM) is a
model for judging the maturity of the software processes of an
organization and for identifying the key practices that are required to
increase the maturity of these processes.

27. What is Cause Effect Graph?


A graphical representation of inputs and the associated outputs effects
which can be used to design test cases.

28. What is Code Complete?


Phase of development where functionality is implemented in entirety;
bug fixes are all that are left. All functions found in the Functional
Specifications have been implemented.

29. What is Code Coverage?


An analysis method that determines which parts of the software have
been executed (covered) by the test case suite and which parts have
not been executed and therefore may require additional attention.

30. What is Code Inspection?


A formal testing technique where the programmer reviews source code
with a group who ask questions analyzing the program logic, analyzing
the code with respect to a checklist of historically common
programming errors, and analyzing its compliance with coding
standards.

31. What is Code Walkthrough?


A formal testing technique where source code is traced by a group with
a small set of test cases, while the state of program variables is
manually monitored, to analyze the programmer's logic and
assumptions.

32. What is Coding?


The generation of source code.

33. What is Compatibility Testing?


Testing whether software is compatible with other elements of a
system with which it should operate, e.g. browsers, Operating
Systems, or hardware.

34. What is Component?


A minimal software item for which a separate specification is available.

35. What is Component Testing?


Testing of individual software components (Unit Testing).

36. What is Concurrency Testing?


Multi-user testing geared towards determining the effects of accessing
the same application code, module or database records. Identifies and
measures the level of locking, deadlocking and use of single-threaded
code and locking semaphores.

37. What is Conformance Testing?


The process of testing that an implementation conforms to the
specification on which it is based. Usually applied to testing
conformance to a formal standard.

38. What is Context Driven Testing?


The context-driven school of software testing is flavor of Agile Testing
that advocates continuous and creative evaluation of testing
opportunities in light of the potential information revealed and the
value of that information to the organization right now.

39. What is Conversion Testing?


Testing of programs or procedures used to convert data from existing
systems for use in replacement systems.
40. What is Cyclomatic Complexity?
A measure of the logical complexity of an algorithm, used in white-box
testing.

41. What is Data Dictionary?


A database that contains definitions of all data items defined during
analysis.

42. What is Data Flow Diagram?


A modeling notation that represents a functional decomposition of a
system.

43. What is Data Driven Testing?


Testing in which the action of a test case is parameterized by
externally defined data values, maintained as a file or spreadsheet. A
common technique in Automated Testing.

44. What is Debugging?


The process of finding and removing the causes of software failures.

45. What is Defect?


Nonconformance to requirements or functional / program specification

46. What is Dependency Testing?


Examines an application's requirements for pre-existing software,
initial states and configuration in order to maintain proper
functionality.

47. What is Depth Testing?


A test that exercises a feature of a product in full detail.

48. What is Dynamic Testing?


Testing software through executing it. See also Static Testing.

49. What is Emulator?


A device, computer program, or system that accepts the same inputs
and produces the same outputs as a given system.

50. What is Endurance Testing?


Checks for memory leaks or other problems that may occur with
prolonged execution.

51. What is End-to-End testing?


Testing a complete application environment in a situation that mimics
real-world use, such as interacting with a database, using network
communications, or interacting with other hardware, applications, or
systems if appropriate.

52. What is Equivalence Class?


A portion of a component's input or output domains for which the
component's behaviour is assumed to be the same from the
component's specification.

53. What is Equivalence Partitioning?


A test case design technique for a component in which test cases are
designed to execute representatives from equivalence classes.

54. What is Exhaustive Testing?


Testing which covers all combinations of input values and
preconditions for an element of the software under test.

55. What is Functional Decomposition?


A technique used during planning, analysis and design; creates a
functional hierarchy for the software.

54. What is Functional Specification?


A document that describes in detail the characteristics of the product
with regard to its intended features.

55. What is Functional Testing?


Testing the features and operational behavior of a product to ensure
they correspond to its specifications. Testing that ignores the internal
mechanism of a system or component and focuses solely on the
outputs generated in response to selected inputs and execution
conditions. or Black Box Testing.

56. What is Glass Box Testing?


A synonym for White Box Testing.

57. What is Gorilla Testing?


Testing one particular module, functionality heavily.

58. What is Gray Box Testing?


A combination of Black Box and White Box testing methodologies?
testing a piece of software against its specification but using some
knowledge of its internal workings.

59. What is High Order Tests?


Black-box tests conducted once the software has been integrated.

60. What is Independent Test Group (ITG)?


A group of people whose primary responsibility is software testing,

61. What is Inspection?


A group review quality improvement process for written material. It
consists of two aspects; product (document itself) improvement and
process improvement (of both document production and inspection).

62. What is Integration Testing?


Testing of combined parts of an application to determine if they
function together correctly. Usually performed after unit and functional
testing. This type of testing is especially relevant to client/server and
distributed systems.

63. What is Installation Testing?


Confirms that the application under test recovers from expected or
unexpected events without loss of data or functionality. Events can
include shortage of disk space, unexpected loss of communication, or
power out conditions.

64. What is Load Testing?


See Performance Testing.

65. What is Localization Testing?


This term refers to making software specifically designed for a specific
locality.

66. What is Loop Testing?


A white box testing technique that exercises program loops.

67. What is Metric?


A standard of measurement. Software metrics are the statistics
describing the structure or content of a program. A metric should be a
real objective measurement of something such as number of bugs per
lines of code.

68. What is Monkey Testing?


Testing a system or an Application on the fly, i.e just few tests here
and there to ensure the system or an application does not crash out.

69. What is Negative Testing?


Testing aimed at showing software does not work. Also known as "test
to fail". See also Positive Testing.

70. What is Path Testing?


Testing in which all paths in the program source code are tested at
least once.

71. What is Performance Testing?


Testing conducted to evaluate the compliance of a system or
component with specified performance requirements. Often this is
performed using an automated test tool to simulate large number of
users. Also know as "Load Testing".

72. What is Positive Testing?


Testing aimed at showing software works. Also known as "test to
pass". See also Negative Testing.

73. What is Quality Assurance?


All those planned or systematic actions necessary to provide adequate
confidence that a product or service is of the type and quality needed
and expected by the customer.

74. What is Quality Audit?


A systematic and independent examination to determine whether
quality activities and related results comply with planned
arrangements and whether these arrangements are implemented
effectively and are suitable to achieve objectives.

75. What is Quality Circle?


A group of individuals with related interests that meet at regular
intervals to consider problems or other matters related to the quality
of outputs of a process and to the correction of problems or to the
improvement of quality.

76. What is Quality Control?


The operational techniques and the activities used to fulfill and verify
requirements of quality.

77. What is Quality Management?


That aspect of the overall management function that determines and
implements the quality policy.

78. What is Quality Policy?


The overall intentions and direction of an organization as regards
quality as formally expressed by top management.
79. What is Quality System?
The organizational structure, responsibilities, procedures, processes,
and resources for implementing quality management.

80. What is Race Condition?


A cause of concurrency problems. Multiple accesses to a shared
resource, at least one of which is a write, with no mechanism used by
either to moderate simultaneous access.

81. What is Ramp Testing?


Continuously raising an input signal until the system breaks down.

82. What is Recovery Testing?


Confirms that the program recovers from expected or unexpected
events without loss of data or functionality. Events can include
shortage of disk space, unexpected loss of communication, or power
out conditions.

83. What is Regression Testing?


Retesting a previously tested program following modification to ensure
that faults have not been introduced or uncovered as a result of the
changes made.

84. What is Release Candidate?


A pre-release version, which contains the desired functionality of the
final version, but which needs to be tested for bugs (which ideally
should be removed before the final version is released).

85. What is Sanity Testing?


Brief test of major functional elements of a piece of software to
determine if its basically operational. See also Smoke Testing.

86. What is Scalability Testing?


Performance testing focused on ensuring the application under test
gracefully handles increases in work load.

87. What is Security Testing?


Testing which confirms that the program can restrict access to
authorized personnel and that the authorized personnel can access the
functions available to their security level.

88. What is Smoke Testing?


A quick-and-dirty test that the major functions of a piece of software
work. Originated in the hardware testing practice of turning on a new
piece of hardware for the first time and considering it a success if it
does not catch on fire.

89. What is Soak Testing?


Running a system at high load for a prolonged period of time. For
example, running several times more transactions in an entire day (or
night) than would be expected in a busy day, to identify and
performance problems that appear after a large number of
transactions have been executed.

90. What is Software Requirements Specification?


A deliverable that describes all data, functional and behavioral
requirements, all constraints, and all validation requirements for
software/

91. What is Software Testing?


A set of activities conducted with the intent of finding errors in
software.

92. What is Static Analysis?


Analysis of a program carried out without executing the program.

93. What is Static Analyzer?


A tool that carries out static analysis.

94. What is Static Testing?


Analysis of a program carried out without executing the program.

95. What is Storage Testing?


Testing that verifies the program under test stores data files in the
correct directories and that it reserves sufficient space to prevent
unexpected termination resulting from lack of space. This is external
storage as opposed to internal storage.

96. What is Stress Testing?


Testing conducted to evaluate a system or component at or beyond
the limits of its specified requirements to determine the load under
which it fails and how. Often this is performance testing using a very
high level of simulated load.

97. What is Structural Testing?


Testing based on an analysis of internal workings and structure of a
piece of software. See also White Box Testing.
98. What is System Testing?
Testing that attempts to discover defects that are properties of the
entire system rather than of its individual components.

99. What is Testability?


The degree to which a system or component facilitates the
establishment of test criteria and the performance of tests to
determine whether those criteria have been met.

100. What is Testing?


The process of exercising software to verify that it satisfies specified
requirements and to detect errors. The process of analyzing a software
item to detect the differences between existing and required conditions
(that is, bugs), and to evaluate the features of the software item (Ref.
IEEE Std 829). The process of operating a system or component under
specified conditions, observing or recording the results, and making an
evaluation of some aspect of the system or component. What is Test
Automation? It is the same as Automated Testing.

101. What is Test Bed?


An execution environment configured for testing. May consist of
specific hardware, OS, network topology, configuration of the product
under test, other application or system software, etc. The Test Plan for
a project should enumerated the test beds(s) to be used.

102. What is Test Case?


Test Case is a commonly used term for a specific test. This is usually
the smallest unit of testing. A Test Case will consist of information
such as requirements testing, test steps, verification steps,
prerequisites, outputs, test environment, etc. A set of inputs,
execution preconditions, and expected outcomes developed for a
particular objective, such as to exercise a particular program path or
to verify compliance with a specific requirement. Test Driven
Development? Testing methodology associated with Agile
Programming in which every chunk of code is covered by unit tests,
which must all pass all the time, in an effort to eliminate unit-level and
regression bugs during development. Practitioners of TDD write a lot of
tests, i.e. an equal number of lines of test code to the size of the
production code.

103. What is Test Driver?


A program or test tool used to execute a tests. Also known as a Test
Harness.
104. What is Test Environment?
The hardware and software environment in which tests will be run, and
any other software with which the software under test interacts when
under test including stubs and test drivers.

105. What is Test First Design?


Test-first design is one of the mandatory practices of Extreme
Programming (XP).It requires that programmers do not write any
production code until they have first written a unit test.

106. What is Test Harness?


A program or test tool used to execute a tests. Also known as a Test
Driver.

107. What is Test Plan?


A document describing the scope, approach, resources, and schedule
of intended testing activities. It identifies test items, the features to be
tested, the testing tasks, who will do each task, and any risks
requiring contingency planning.

108. What is Test Procedure?


A document providing detailed instructions for the execution of one or
more test cases.

109. What is Test Script?


Commonly used to refer to the instructions for a particular test that
will be carried out by an automated test tool.

110. What is Test Specification?


A document specifying the test approach for a software feature or
combination or features and the inputs, predicted results and
execution conditions for the associated tests.

111. What is Test Suite?


A collection of tests used to validate the behavior of a product. The
scope of a Test Suite varies from organization to organization. There
may be several Test Suites for a particular product for example. In
most cases however a Test Suite is a high level concept, grouping
together hundreds or thousands of tests related by what they are
intended to test.

112. What is Test Tools?


Computer programs used in the testing of a system, a component of
the system, or its documentation.

113. What is Thread Testing?


A variation of top-down testing where the progressive integration of
components follows the implementation of subsets of the
requirements, as opposed to the integration of components by
successively lower levels.

114. What is Top Down Testing?


An approach to integration testing where the component at the top of
the component hierarchy is tested first, with lower level components
being simulated by stubs. Tested components are then used to test
lower level components. The process is repeated until the lowest level
components have been tested.

115. What is Total Quality Management?


A company commitment to develop a process that achieves high
quality product and customer satisfaction.

116. What is Traceability Matrix?


A document showing the relationship between Test Requirements and
Test Cases.

117. What is Usability Testing?


Testing the ease with which users can learn and use a product.

118. What is Use Case?


The specification of tests that are conducted from the end-user
perspective. Use cases tend to focus on operating software as an end-
user would conduct their day-to-day activities.

119. What is Unit Testing?


Testing of individual software components.

120. What is Validation?


The process of evaluating software at the end of the software
development process to ensure compliance with software
requirements. The techniques for validation is testing, inspection and
reviewing.

121. What is Verification?


The process of determining whether of not the products of a given
phase of the software development cycle meet the implementation
steps and can be traced to the incoming objectives established during
the previous phase. The techniques for verification are testing,
inspection and reviewing.

122. What is White Box Testing?


Testing based on an analysis of internal workings and structure of a
piece of software. Includes techniques such as Branch Testing and
Path Testing. Also known as Structural Testing and Glass Box Testing.
Contrast with Black Box Testing. White box testing is used to test the
internal logic of the code for ex checking whether the path has been
executed once, checking whether the branches has been executed
atleast once .....Used to check the structure of the code.

123. What is Workflow Testing?


Scripted end-to-end testing which duplicates specific workflows which
are expected to be utilized by the end-user.

124. What's the difference between load and stress testing ?


One of the most common, but unfortunate misuse of terminology is
treating “load testing” and “stress testing” as synonymous. The
consequence of this ignorant semantic abuse is usually that the
system is neither properly “load tested” nor subjected to a meaningful
stress test. Stress testing is subjecting a system to an unreasonable
load while denying it the resources (e.g., RAM, disc, mips, interrupts,
etc.) needed to process that load. The idea is to stress a system to the
breaking point in order to find bugs that will make that break
potentially harmful. The system is not expected to process the
overload without adequate resources, but to behave (e.g., fail) in a
decent manner (e.g., not corrupting or losing data). Bugs and failure
modes discovered under stress testing may or may not be repaired
depending on the application, the failure mode, consequences, etc.
The load (incoming transaction stream) in stress testing is often
deliberately distorted so as to force the system into resource
depletion. Load testing is subjecting a system to a statistically
representative (usually) load. The two main reasons for using such
loads is in support of software reliability testing and in performance
testing. The term 'load testing' by itself is too vague and imprecise to
warrant use. For example, do you mean representative load,'
'overload,' 'high load,' etc. In performance testing, load is varied from
a minimum (zero) to the maximum level the system can sustain
without running out of resources or having, transactions >suffer
(application-specific) excessive delay. A third use of the term is as a
test whose objective is to determine the maximum sustainable load
the system can handle. In this usage, 'load testing' is merely testing at
the highest transaction arrival rate in performance testing.
125. What's the difference between QA and testing?
QA is more a preventive thing, ensuring quality in the company and
therefore the product rather than just testing the product for software
bugs? TESTING means 'quality control' QUALITY CONTROL measures
the quality of a product QUALITY ASSURANCE measures the quality of
processes used to create a quality product.

126. What is the best tester to developer ratio?


Reported tester: developer ratios range from 10:1 to 1:10. There's no
simple answer. It depends on so many things, Amount of reused code,
number and type of interfaces, platform, quality goals, etc. It also can
depend on the development model. The more specs, the less testers.
The roles can play a big part also. Does QA own beta? Do you include
process auditors or planning activities? These figures can all vary very
widely depending on how you define 'tester' and 'developer'. In some
organizations, a 'tester' is anyone who happens to be testing software
at the time -- such as their own. In other organizations, a 'tester' is
only a member of an independent test group. It is better to ask about
the test labor content than it is to ask about the tester/developer ratio.
The test labor content, across most applications is generally accepted
as 50%, when people do honest accounting. For life-critical software,
this can go up to 80%.

127. How can new Software QA processes be introduced in an existing


organization?
- A lot depends on the size of the organization and the risks involved.
For large organizations with high-risk (in terms of lives or property)
projects, serious management buy-in is required and a formalized QA
process is necessary. - Where the risk is lower, management and
organizational buy-in and QA implementation may be a slower, step-
at-a-time process. QA processes should be balanced with productivity
so as to keep bureaucracy from getting out of hand. - For small groups
or projects, a more ad-hoc process may be appropriate, depending on
the type of customers and projects. A lot will depend on team leads or
managers, feedback to developers, and ensuring adequate
communications among customers, managers, developers, and
testers. - In all cases the most value for effort will be in requirements
management processes, with a goal of clear, complete, testable
requirement specifications or expectations.

128. What are 5 common problems in the software development


process?
1. poor requirements - if requirements are unclear, incomplete, too
general, or not testable, there will be problems. 2. unrealistic schedule
- if too much work is crammed in too little time, problems are
inevitable. 3. inadequate testing - no one will know whether or not the
program is any good until the customer complains or systems crash. 4.
features - requests to pile on new features after development is
underway; extremely common. 5. miscommunication - if developers
don't know what's needed or customer's have erroneous expectations,
problems are guaranteed.

129. What are 5 common solutions to software development


problems?
1. solid requirements - clear, complete, detailed, cohesive, attainable,
testable requirements that are agreed to by all players. Use prototypes
to help nail down requirements. 2. realistic schedules - allow adequate
time for planning, design, testing, bug fixing, re-testing, changes, and
documentation; personnel should be able to complete the project
without burning out. 3. adequate testing - start testing early on, re-
test after fixes or changes, plan for adequate time for testing and bug-
fixing. 4. stick to initial requirements as much as possible - be
prepared to defend against changes and additions once development
has begun, and be prepared to explain consequences. If changes are
necessary, they should be adequately reflected in related schedule
changes. If possible, use rapid prototyping during the design phase so
that customers can see what to expect. This will provide them a higher
comfort level with their requirements decisions and minimize changes
later on. 5. communication - require walkthroughs and inspections
when appropriate; make extensive use of group communication tools -
e-mail, groupware, networked bug-tracking tools and change
management tools, intranet capabilities, etc.; insure that
documentation is available and up-to-date - preferably electronic, not
paper; promote teamwork and cooperation; use prototypes early on so
that customers' expectations are clarified.

130. What is 'good code'?


'Good code' is code that works, is bug free, and is readable and
maintainable. Some organizations have coding 'standards' that all
developers are supposed to adhere to, but everyone has different
ideas about what's best, or what is too many or too few rules. There
are also various theories and metrics, such as McCabe Complexity
metrics. It should be kept in mind that excessive use of standards and
rules can stifle productivity and creativity. 'Peer reviews', 'buddy
checks' code analysis tools, etc. can be used to check for problems and
enforce standards. For C and C++ coding, here are some typical ideas
to consider in setting rules/standards; these may or may not apply to
a particular situation: - minimize or eliminate use of global variables. -
use descriptive function and method names - use both upper and
lower case, avoid abbreviations, use as many characters as necessary
to be adequately descriptive (use of more than 20 characters is not
out of line); be consistent in naming conventions. - use descriptive
variable names - use both upper and lower case, avoid abbreviations,
use as many characters as necessary to be adequately descriptive (use
of more than 20 characters is not out of line); be consistent in naming
conventions. - function and method sizes should be minimized; less
than 100 lines of code is good, less than 50 lines is preferable. -
function descriptions should be clearly spelled out in comments
preceding a function's code.- organize code for readability. - use
whitespace generously - vertically and horizontally - each line of code
should contain 70 characters max. - one code statement per line. -
coding style should be consistent throughout a program (eg, use of
brackets, indentations, naming conventions, etc.) - in adding
comments, err on the side of too many rather than too few comments;
a common rule of thumb is that there should be at least as many lines
of comments (including header blocks) as lines of code. - no matter
how small, an application should include documentation of the overall
program function and flow (even a few paragraphs is better than
nothing); or if possible a separate flow chart and detailed program
documentation. - make extensive use of error handling procedures and
status and error logging. - for C++, to minimize complexity and
increase maintainability, avoid too many levels of inheritance in class
hierarchies (relative to the size and complexity of the application).
Minimize use of multiple inheritance, and minimize use of operator
overloading (note that the Java programming language eliminates
multiple inheritance and operator overloading.) - for C++, keep class
methods small, less than 50 lines of code per method is preferable. -
for C++, make liberal use of exception handlers

131. What is 'good design'?


'Design' could refer to many things, but often refers to 'functional
design' or 'internal design'. Good internal design is indicated by
software code whose overall structure is clear, understandable, easily
modifiable, and maintainable; is robust with sufficient error-handling
and status logging capability; and works correctly when implemented.
Good functional design is indicated by an application whose
functionality can be traced back to customer and end-user
requirements. For programs that have a user interface, it's often a
good idea to assume that the end user will have little computer
knowledge and may not read a user manual or even the on-line help;
some common rules-of-thumb include: - the program should act in a
way that least surprises the user - it should always be evident to the
user what can be done next and how to exit - the program shouldn't
let the users do something stupid without warning them.

132. What makes a good test engineer?


A good test engineer has a 'test to break' attitude, an ability to take
the point of view of the customer, a strong desire for quality, and an
attention to detail. Tact and diplomacy are useful in maintaining a
cooperative relationship with developers, and an ability to
communicate with both technical (developers) and non-technical
(customers, management) people is useful. Previous software
development experience can be helpful as it provides a deeper
understanding of the software development process, gives the tester
an appreciation for the developers' point of view, and reduce the
learning curve in automated test tool programming. Judgment skills
are needed to assess high-risk areas of an application on which to
focus testing efforts when time is limited.

133. What makes a good Software QA engineer?


The same qualities a good tester has are useful for a QA engineer.
Additionally, they must be able to understand the entire software
development process and how it can fit into the business approach and
goals of the organization. Communication skills and the ability to
understand various sides of issues are important. In organizations in
the early stages of implementing QA processes, patience and
diplomacy are especially needed. An ability to find problems as well as
to see 'what's missing' is important for inspections and reviews.

134. What makes a good QA or Test manager?


A good QA, test, or QA/Test(combined) manager should: - be familiar
with the software development process - be able to maintain
enthusiasm of their team and promote a positive atmosphere, despite
what is a somewhat 'negative' process (e.g., looking for or preventing
problems) - be able to promote teamwork to increase productivity - be
able to promote cooperation between software, test, and QA engineers
- have the diplomatic skills needed to promote improvements in QA
processes -have the ability to withstand pressures and say 'no' to
other managers when quality is insufficient or QA processes are not
being adhered to - have people judgement skills for hiring and keeping
skilled personnel- be able to communicate with technical and non-
technical people, engineers, managers, and customers. - be able to
run meetings and keep them focused

135. What's the role of documentation in QA?


Critical. (Note that documentation can be electronic, not necessarily
paper.) QA practices should be documented such that they are
repeatable. Specifications, designs, business rules, inspection reports,
configurations, code changes, test plans, test cases, bug reports, user
manuals, etc. should all be documented. There should ideally be a
system for easily finding and obtaining documents and determining
what documentation will have a particular piece of information.
Change management for documentation should be used if possible.

136. What's the big deal about 'requirements'?


One of the most reliable methods of insuring problems, or failure, in a
complex software project is to have poorly documented requirements
specifications. Requirements are the details describing an application's
externally-perceived functionality and properties. Requirements should
be clear, complete, reasonably detailed, cohesive, attainable, and
testable. A non-testable requirement would be, for example, 'user-
friendly' (too subjective). A testable requirement would be something
like 'the user must enter their previously-assigned password to access
the application'. Determining and organizing requirements details in a
useful and efficient way can be a difficult effort; different methods are
available depending on the particular project. Many books are
available that describe various approaches to this task. Care should be
taken to involve ALL of a project's significant 'customers' in the
requirements process. 'Customers' could be in-house personnel or out,
and could include end-users, customer acceptance testers, customer
contract officers, customer management, future software maintenance
engineers, salespeople, etc. Anyone who could later derail the project
if their expectations aren't met should be included if possible.
Organizations vary considerably in their handling of requirements
specifications. Ideally, the requirements are spelled out in a document
with statements such as 'The product shall.....'. 'Design' specifications
should not be confused with 'requirements'; design specifications
should be traceable back to the requirements. In some organizations
requirements may end up in high level project plans, functional
specification documents, in design documents, or in other documents
at various levels of detail. No matter what they are called, some type
of documentation with detailed requirements will be needed by testers
in order to properly plan and execute tests. Without such
documentation, there will be no clear-cut way to determine if a
software application is performing correctly.

137. What steps are needed to develop and run software tests?
The following are some of the steps to consider: - Obtain
requirements, functional design, and internal design specifications and
other necessary documents - Obtain budget and schedule
requirements - Determine project-related personnel and their
responsibilities, reporting requirements, required standards and
processes (such as release processes, change processes, etc.) -
Identify application's higher-risk aspects, set priorities, and determine
scope and limitations of tests - Determine test approaches and
methods - unit, integration, functional, system, load, usability tests,
etc. - Determine test environment requirements (hardware, software,
communications, etc.) -Determine testware requirements
(record/playback tools, coverage analyzers, test tracking, problem/bug
tracking, etc.) - Determine test input data requirements - Identify
tasks, those responsible for tasks, and labor requirements - Set
schedule estimates, timelines, milestones - Determine input
equivalence classes, boundary value analyses, error classes - Prepare
test plan document and have needed reviews/approvals - Write test
cases - Have needed reviews/inspections/approvals of test cases -
Prepare test environment and testware, obtain needed user
manuals/reference documents/configuration guides/installation guides,
set up test tracking processes, set up logging and archiving processes,
set up or obtain test input data - Obtain and install software releases -
Perform tests - Evaluate and report results - Track problems/bugs and
fixes - Retest as needed - Maintain and update test plans, test cases,
test environment, and testware through life cycle

138. What is 'configuration management'?


Configuration management covers the processes used to control,
coordinate, and track: code, requirements, documentation, problems,
change requests, designs, tools/compilers/libraries/patches, changes
made to them, and who makes the changes.

139. What if the software is so buggy it can't really be tested at all?


The best bet in this situation is for the testers to go through the
process of reporting whatever bugs or blocking-type problems initially
show up, with the focus being on critical bugs. Since this type of
problem can severely affect schedules, and indicates deeper problems
in the software development process (such as insufficient unit testing
or insufficient integration testing, poor design, improper build or
release procedures, etc.) managers should be notified, and provided
with some documentation as evidence of the problem.

140. How can it be known when to stop testing?


This can be difficult to determine. Many modern software applications
are so complex, and run in such an interdependent environment, that
complete testing can never be done. Common factors in deciding when
to stop are: - Deadlines (release deadlines, testing deadlines, etc.)-
Test cases completed with certain percentage passed - Test budget
depleted - Coverage of code/functionality/requirements reaches a
specified point - Bug rate falls below a certain level - Beta or alpha
testing period ends

141. What if there isn't enough time for thorough testing?


Use risk analysis to determine where testing should be focused. Since
it's rarely possible to test every possible aspect of an application,
every possible combination of events, every dependency, or
everything that could go wrong, risk analysis is appropriate to most
software development projects. This requires judgement skills,
common sense, and experience. (If warranted, formal methods are
also available.) Considerations can include: - Which functionality is
most important to the project's intended purpose? - Which
functionality is most visible to the user? - Which functionality has the
largest safety impact? - Which functionality has the largest financial
impact on users? - Which aspects of the application are most
important to the customer? - Which aspects of the application can be
tested early in the development cycle? - Which parts of the code are
most complex, and thus most subject to errors? - Which parts of the
application were developed in rush or panic mode? - Which aspects of
similar/related previous projects caused problems? - Which aspects of
similar/related previous projects had large maintenance expenses? -
Which parts of the requirements and design are unclear or poorly
thought out? - What do the developers think are the highest-risk
aspects of the application? - What kinds of problems would cause the
worst publicity? - What kinds of problems would cause the most
customer service complaints?- What kinds of tests could easily cover
multiple functionalities? - Which tests will have the best high-risk-
coverage to time-required ratio?

142. What can be done if requirements are changing continuously?


A common problem and a major headache. - Work with the project's
stakeholders early on to understand how requirements might change
so that alternate test plans and strategies can be worked out in
advance, if possible. - It's helpful if the application's initial design
allows for some adaptability so that later changes do not require
redoing the application from scratch. - If the code is well-commented
and well-documented this makes changes easier for the developers.-
Use rapid prototyping whenever possible to help customers feel sure of
their requirements and minimize changes. - The project's initial
schedule should allow for some extra time commensurate with the
possibility of changes.- Try to move new requirements to a 'Phase 2'
version of an application, while using the original requirements for the
'Phase 1' version. - Negotiate to allow only easily-implemented new
requirements into the project, while moving more difficult new
requirements into future versions of the application. - Be sure that
customers and management understand the scheduling impacts,
inherent risks, and costs of significant requirements changes. Then let
management or the customers (not the developers or testers) decide if
the changes are warranted - after all, that's their job. - Balance the
effort put into setting up automated testing with the expected effort
required to re-do them to deal with changes. - Try to design some
flexibility into automated test scripts. - Focus initial automated testing
on application aspects that are most likely to remain unchanged. -
Devote appropriate effort to risk analysis of changes to minimize
regression testing needs. - Design some flexibility into test cases (this
is not easily done; the best bet might be to minimize the detail in the
test cases, or set up only higher-level generic-type test plans) - Focus
less on detailed test plans and test cases and more on ad hoc testing
(with an understanding of the added risk that this entails).

143. What if the project isn't big enough to justify extensive testing?
Consider the impact of project errors, not the size of the project.
However, if extensive testing is still not justified, risk analysis is again
needed and the same considerations as described previously in 'What
if there isn't enough time for thorough testing?' apply. The tester
might then do ad hoc testing, or write up a limited test plan based on
the risk analysis.

144. What if the application has functionality that wasn't in the


requirements?
It may take serious effort to determine if an application has significant
unexpected or hidden functionality, and it would indicate deeper
problems in the software development process. If the functionality isn't
necessary to the purpose of the application, it should be removed, as it
may have unknown impacts or dependencies that were not taken into
account by the designer or the customer. If not removed, design
information will be needed to determine added testing needs or
regression testing needs. Management should be made aware of any
significant added risks as a result of the unexpected functionality. If
the functionality only effects areas such as minor improvements in the
user interface, for example, it may not be a significant risk.

145. How can Software QA processes be implemented without stifling


productivity?
By implementing QA processes slowly over time, using consensus to
reach agreement on processes, and adjusting and experimenting as an
organization grows and matures, productivity will be improved instead
of stifled. Problem prevention will lessen the need for problem
detection, panics and burn-out will decrease, and there will be
improved focus and less wasted effort. At the same time, attempts
should be made to keep processes simple and efficient, minimize
paperwork, promote computer-based processes and automated
tracking and reporting, minimize time required in meetings, and
promote training as part of the QA process. However, no one -
especially talented technical types - likes rules or bureaucracy, and in
the short run things may slow down a bit. A typical scenario would be
that more days of planning and development will be needed, but less
time will be required for late-night bug-fixing and calming of irate
customers.

146. What if an organization is growing so fast that fixed QA processes


are impossible?
This is a common problem in the software industry, especially in new
technology areas. There is no easy solution in this situation, other
than: - Hire good people - Management should 'ruthlessly prioritize'
quality issues and maintain focus on the customer - Everyone in the
organization should be clear on what 'quality' means to the customer

147. How does a client/server environment affect testing?


Client/server applications can be quite complex due to the multiple
dependencies among clients, data communications, hardware, and
servers. Thus testing requirements can be extensive. When time is
limited (as it usually is) the focus should be on integration and system
testing. Additionally, load/stress/performance testing may be useful in
determining client/server application limitations and capabilities. There
are commercial tools to assist with such testing.

148.How can World Wide Web sites be tested?


Web sites are essentially client/server applications - with web servers
and 'browser' clients. Consideration should be given to the interactions
between html pages, TCP/IP communications, Internet connections,
firewalls, applications that run in web pages (such as applets,
javascript, plug-in applications), and applications that run on the
server side (such as cgi scripts, database interfaces, logging
applications, dynamic page generators, asp, etc.). Additionally, there
are a wide variety of servers and browsers, various versions of each,
small but sometimes significant differences between them, variations
in connection speeds, rapidly changing technologies, and multiple
standards and protocols. The end result is that testing for web sites
can become a major ongoing effort. Other considerations might
include: - What are the expected loads on the server (e.g., number of
hits per unit time?), and what kind of performance is required under
such loads (such as web server response time, database query
response times). What kinds of tools will be needed for performance
testing (such as web load testing tools, other tools already in house
that can be adapted, web robot downloading tools, etc.)? - Who is the
target audience? What kind of browsers will they be using? What kind
of connection speeds will they by using? Are they intra- organization
(thus with likely high connection speeds and similar browsers) or
Internet-wide (thus with a wide variety of connection speeds and
browser types)? - What kind of performance is expected on the client
side (e.g., how fast should pages appear, how fast should animations,
applets, etc. load and run)? - Will down time for server and content
maintenance/upgrades be allowed? how much? - What kinds of
security (firewalls, encryptions, passwords, etc.) will be required and
what is it expected to do? How can it be tested? - How reliable are the
site's Internet connections required to be? And how does that affect
backup system or redundant connection requirements and testing? -
What processes will be required to manage updates to the web site's
content, and what are the requirements for maintaining, tracking, and
controlling page content, graphics, links, etc.? - Which HTML
specification will be adhered to? How strictly? What variations will be
allowed for targeted browsers? - Will there be any standards or
requirements for page appearance and/or graphics throughout a site
or parts of a site?? - How will internal and external links be validated
and updated? how often? - Can testing be done on the production
system, or will a separate test system be required? How are browser
caching, variations in browser option settings, dial-up connection
variabilities, and real-world internet 'traffic congestion' problems to be
accounted for in testing?- How extensive or customized are the server
logging and reporting requirements; are they considered an integral
part of the system and do they require testing?- How are cgi
programs, applets, javascripts, ActiveX components, etc. to be
maintained, tracked, controlled, and tested? - Pages should be 3-5
screens max unless content is tightly focused on a single topic. If
larger, provide internal links within the page. - The page layouts and
design elements should be consistent throughout a site, so that it's
clear to the user that they're still within a site. - Pages should be as
browser-independent as possible, or pages should be provided or
generated based on the browser-type. - All pages should have links
external to the page; there should be no dead-end pages. - The page
owner, revision date, and a link to a contact person or organization
should be included on each page.
149. How is testing affected by object-oriented designs?
Well-engineered object-oriented design can make it easier to trace
from code to internal design to functional design to requirements.
While there will be little affect on black box testing (where an
understanding of the internal design of the application is unnecessary),
white-box testing can be oriented to the application's objects. If the
application was well-designed this can simplify test design.

150. What is Extreme Programming and what's it got to do with


testing?
Extreme Programming (XP) is a software development approach for
small teams on risk-prone projects with unstable requirements. It was
created by Kent Beck who described the approach in his book 'Extreme
Programming Explained'. Testing ('extreme testing') is a core aspect of
Extreme Programming. Programmers are expected to write unit and
functional test code first - before the application is developed. Test
code is under source control along with the rest of the code.
Customers are expected to be an integral part of the project team and
to help develop scenarios for acceptance/black box testing. Acceptance
tests are preferably automated, and are modified and rerun for each of
the frequent development iterations. QA and test personnel are also
required to be an integral part of the project team. Detailed
requirements documentation is not used, and frequent re-scheduling,
re-estimating, and re-prioritizing is expected.

151. Will automated testing tools make testing easier?


- Possibly. For small projects, the time needed to learn and implement
them may not be worth it. For larger projects, or on-going long-term
projects they can be valuable. - A common type of automated tool is
the 'record/playback' type. For example, a tester could click through
all combinations of menu choices, dialog box choices, buttons, etc. in
an application GUI and have them 'recorded' and the results logged by
a tool. The 'recording' is typically in the form of text based on a
scripting language that is interpretable by the testing tool. If new
buttons are added, or some underlying code in the application is
changed, etc. the application can then be retested by just 'playing
back' the 'recorded' actions, and comparing the logging results to
check effects of the changes. The problem with such tools is that if
there are continual changes to the system being tested, the
'recordings' may have to be changed so much that it becomes very
time-consuming to continuously update the scripts. Additionally,
interpretation of results (screens, data, logs, etc.) can be a difficult
task. Note that there are record/playback tools for text-based
interfaces also, and for all types of platforms.- Other automated tools
can include: code analyzers - monitor code complexity, adherence to
standards, etc. coverage analyzers - these tools check which parts of
the code have been exercised by a test, and may be oriented to code
statement coverage, condition coverage, path coverage, etc. memory
analyzers - such as bounds-checkers and leak detectors.
load/performance test tools - for testing client/server and web
applications under various load levels. web test tools - to check that
links are valid, HTML code usage is correct, client-side and server-side
programs work, a web site's interactions are secure. other tools - for
test case management, documentation management, bug reporting,
and configuration management.

152. What's the difference between black box and white box testing?
Black-box and white-box are test design methods. Black-box test
design treats the system as a “black-box”, so it doesn't explicitly use
knowledge of the internal structure. Black-box test design is usually
described as focusing on testing functional requirements. Synonyms
for black-box include: behavioral, functional, opaque-box, and closed-
box. White-box test design allows one to peek inside the “box”, and it
focuses specifically on using internal knowledge of the software to
guide the selection of test data. Synonyms for white-box include:
structural, glass-box and clear-box. While black-box and white-box are
terms that are still in popular use, many people prefer the terms
'behavioral' and 'structural'. Behavioral test design is slightly different
from black-box test design because the use of internal knowledge isn't
strictly forbidden, but it's still discouraged. In practice, it hasn't proven
useful to use a single test design method. One has to use a mixture of
different methods so that they aren't hindered by the limitations of a
particular one. Some call this 'gray-box' or 'translucent-box' test
design, but others wish we'd stop talking about boxes altogether.It is
important to understand that these methods are used during the test
design phase, and their influence is hard to see in the tests once
they're implemented. Note that any level of testing (unit testing,
system testing, etc.) can use any test design methods. Unit testing is
usually associated with structural test design, but this is because
testers usually don't have well-defined requirements at the unit level
to validate.

153. What kinds of testing should be considered?


Black box testing - not based on any knowledge of internal design or
code. Tests are based on requirements and functionality. White box
testing - based on knowledge of the internal logic of an application's
code. Tests are based on coverage of code statements, branches,
paths, conditions.unit testing - the most 'micro' scale of testing; to test
particular functions or code modules. Typically done by the
programmer and not by testers, as it requires detailed knowledge of
the internal program design and code. Not always easily done unless
the application has a well-designed architecture with tight code; may
require developing test driver modules or test harnesses. incremental
integration testing - continuous testing of an application as new
functionality is added; requires that various aspects of an application's
functionality be independent enough to work separately before all
parts of the program are completed, or that test drivers be developed
as needed; done by programmers or by testers. integration testing -
testing of combined parts of an application to determine if they
function together correctly. The 'parts' can be code modules, individual
applications, client and server applications on a network, etc. This type
of testing is especially relevant to client/server and distributed
systems. functional testing - black-box type testing geared to
functional requirements of an application; this type of testing should
be done by testers. This doesn't mean that the programmers shouldn't
check that their code works before releasing it (which of course applies
to any stage of testing.) system testing - black-box type testing that is
based on overall requirements specifications; covers all combined
parts of a system. end-to-end testing - similar to system testing; the
'macro' end of the test scale; involves testing of a complete application
environment in a situation that mimics real-world use, such as
interacting with a database, using network communications, or
interacting with other hardware, applications, or systems if
appropriate. sanity testing or smoke testing - typically an initial testing
effort to determine if a new software version is performing well enough
to accept it for a major testing effort. For example, if the new software
is crashing systems every 5 minutes, bogging down systems to a
crawl, or corrupting databases, the software may not be in a 'sane'
enough condition to warrant further testing in its current state.
regression testing - re-testing after fixes or modifications of the
software or its environment. It can be difficult to determine how much
re-testing is needed, especially near the end of the development cycle.
Automated testing tools can be especially useful for this type of
testing. acceptance testing - final testing based on specifications of the
end-user or customer, or based on use by end-users/customers over
some limited period of time. load testing - testing an application under
heavy loads, such as testing of a web site under a range of loads to
determine at what point the system's response time degrades or
fails.stress testing - term often used interchangeably with 'load' and
'performance' testing. Also used to describe such tests as system
functional testing while under unusually heavy loads, heavy repetition
of certain actions or inputs, input of large numerical values, large
complex queries to a database system, etc. performance testing - term
often used interchangeably with 'stress' and 'load' testing. Ideally
'performance' testing (and any other 'type' of testing) is defined in
requirements documentation or QA or Test Plans.usability testing -
testing for 'user-friendliness'. Clearly this is subjective, and will depend
on the targeted end-user or customer. User interviews, surveys, video
recording of user sessions, and other techniques can be used.
Programmers and testers are usually not appropriate as usability
testers. install/uninstall testing - testing of full, partial, or upgrade
install/uninstall processes. recovery testing - testing how well a
system recovers from crashes, hardware failures, or other catastrophic
problems. failover testing - typically used interchangeably with
'recovery testing'security testing - testing how well the system
protects against unauthorized internal or external access, willful
damage, etc; may require sophisticated testing techniques.
compatibility testing - testing how well software performs in a
particular hardware/software/operating system/network/etc.
environment. exploratory testing - often taken to mean a creative,
informal software test that is not based on formal test plans or test
cases; testers may be learning the software as they test it. ad-hoc
testing - similar to exploratory testing, but often taken to mean that
the testers have significant understanding of the software before
testing it. context-driven testing - testing driven by an understanding
of the environment, culture, and intended use of software. For
example, the testing approach for life-critical medical equipment
software would be completely different than that for a low-cost
computer game. user acceptance testing - determining if software is
satisfactory to an end-user or customer. comparison testing -
comparing software weaknesses and strengths to competing products.
alpha testing - testing of an application when development is nearing
completion; minor design changes may still be made as a result of
such testing. Typically done by end-users or others, not by
programmers or testers. beta testing - testing when development and
testing are essentially completed and final bugs and problems need to
be found before final release. Typically done by end-users or others,
not by programmers or testers. mutation testing - a method for
determining if a set of test data or test cases is useful, by deliberately
introducing various code changes ('bugs') and retesting with the
original test data/cases to determine if the 'bugs' are detected. Proper
implementation requires large computational resources.

154. Why is it often hard for management to get serious about quality
assurance?
Solving problems is a high-visibility process; preventing problems is
low-visibility.This is illustrated by an old parable:In ancient China there
was a family of healers, one of whom was known throughout the land
and employed as a physician to a great lord. The physician was asked
which of his family was the most skillful healer. He replied, "I tend to
the sick and dying with drastic and dramatic treatments, and on
occasion someone is cured and my name gets out among the
lords.""My elder brother cures sickness when it just begins to take
root, and his skills are known among the local peasants and
neighbors." "My eldest brother is able to sense the spirit of sickness
and eradicate it before it takes form. His name is unknown outside our
home."

155. Why does software have bugs?


1. Miscommunication or no communication - as to specifics of what an
application should or shouldn't do (the application's requirements). 2.
Software complexity - the complexity of current software applications
can be difficult to comprehend for anyone without experience in
modern-day software development. Multi-tiered applications, client-
server and distributed applications, data communications, enormous
relational databases, and sheer size of applications have all
contributed to the exponential growth in software/system complexity.
programming errors - programmers, like anyone else, can make
mistakes. 3. Changing requirements (whether documented or
undocumented) - the end-user may not understand the effects of
changes, or may understand and request them anyway - redesign,
rescheduling of engineers, effects on other projects, work already
completed that may have to be redone or thrown out, hardware
requirements that may be affected, etc. If there are many minor
changes or any major changes, known and unknown dependencies
among parts of the project are likely to interact and cause problems,
and the complexity of coordinating changes may result in errors.
Enthusiasm of engineering staff may be affected. In some fast-
changing business environments, continuously modified requirements
may be a fact of life. In this case, management must understand the
resulting risks, and QA and test engineers must adapt and plan for
continuous extensive testing to keep the inevitable bugs from running
out of control. 3. Poorly documented code - it's tough to maintain and
modify code that is badly written or poorly documented; the result is
bugs. In many organizations management provides no incentive for
programmers to document their code or write clear, understandable,
maintainable code. In fact, it's usually the opposite: they get points
mostly for quickly turning out code, and there's job security if nobody
else can understand it ('if it was hard to write, it should be hard to
read'). 4. software development tools - visual tools, class libraries,
compilers, scripting tools, etc. often introduce their own bugs or are
poorly documented, resulting in added bugs.

156. How can new Software QA processes be introduced in an existing


organization?
A lot depends on the size of the organization and the risks involved.
For large organizations with high-risk (in terms of lives or property)
projects, serious management buy-in is required and a formalized QA
process is necessary. Where the risk is lower, management and
organizational buy-in and QA implementation may be a slower, step-
at-a-time process. QA processes should be balanced with productivity
so as to keep bureaucracy from getting out of hand. For small groups
or projects, a more ad-hoc process may be appropriate, depending on
the type of customers and projects. A lot will depend on team leads or
managers, feedback to developers, and ensuring adequate
communications among customers, managers, developers, and
testers. The most value for effort will often be in (a) requirements
management processes, with a goal of clear, complete, testable
requirement specifications embodied in requirements or design
documentation, or in 'agile'-type environments extensive continuous
coordination with end-users, (b) design inspections and code
inspections, and (c) post-mortems/retrospectives.

157. how do the companies expect the defect reporting to be


communicated by the tester to the development team. Can the excel
sheet template be used for defect reporting. If so what are the
common fields that are to be included ? who assigns the priority and
severity of the defect
To report bugs in excel: Sno. Module Screen/ Section Issue detail
SeverityPriority Issuestatusthis is how to report bugs in excel sheet
and also set filters on the Columns attributes.But most of the
companies use the share point process of reporting bugs In this when
the project came for testing a module wise detail of project is inserted
to the defect management system they are using. It contains following
field1. Date2. Issue brief3. Issue description (used for developer to
regenerate the issue)4. Issue status ( active, resolved, onhold,
suspend and not able to regenerate)5. Assign to (Names of members
allocated to project)6. Priority (High, medium and low)7. severity
(Major, medium and low)

158. What are the tables in testplans and testcases?


Test plan is a document that contains the scope, approach, test design
and test strategies. It includes the following:-1. Test case identifier2.
Scope3.Features to be tested4. Features not to be tested.5. Test
strategy.6. Test Approach7. Test Deliverables8. Responsibilities.9
Staffing and Training10.Risk and Contingencies11. ApprovalWhile A
test case is a noted/documented set of steps/activities that are carried
out or executed on the software in order to confirm its
functionality/behavior to certain set of inputs.

159. What are the table contents in testplans and test cases?
Test Plan is a document which is prepared with the details of the
testing priority. A test Plan generally includes: 1. Objective of
Testing2. Scope of Testing3. Reason for testing4. Timeframe5.
Environment6. Entrance and exit criteria7. Risk factors involved8.
Deliverables

160. What automating testing tools are you familiar with?


Win Runner , Load runner, QTP , Silk Performer, Test director, Rational
robot, QA run.

161. How did you use automating testing tools in your job?
1. For regression testing2. Criteria to decide the condition of a
particular build3. Describe some problem that you had with
automating testing tool.The problem of winrunner identifying the third
party controls like infragistics control.

162. How do you plan test automation?


1. Prepare the automation Test plan2. Identify the scenario3. Record
the scenario4. Enhance the scripts by inserting check points and
Conditional Loops5. Incorporated Error Handler6. Debug the script7.
Fix the issue8. Rerun the script and report the result.

163. Can test automation improve test effectiveness?


Yes, Automating a test makes the test process:1.Fast2.Reliable3.
Repeatable4.Programmable5.Reusable6.Comprehensive6. What is data
- driven automation?Testing the functionality with more test cases
becomes laborious as the functionality grows. For multiple sets of data
(test cases), you can execute the test once in which you can figure out
for which data it has failed and for which data, the test has passed.
This feature is available in the WinRunner with the data driven test
where the data can be taken from an excel sheet or notepad.

164. What are the main attributes of test automation?


software test automation attributes :Maintainability - the effort needed
to update the test automation suites for each new releaseReliability -
the accuracy and repeatability of the test automationFlexibility - the
ease of working with all the different kinds of automation test
wareEfficiency - the total cost related to the effort needed for the
automationPortability - the ability of the automated test to run on
different environmentsRobustness - the effectiveness of automation on
an unstable or rapidly changing systemUsability - the extent to which
automation can be used by different types of users

165. Does automation replace manual testing?


There can be some functionality which cannot be tested in an
automated tool so we may have to do it manually. therefore manual
testing can never be replaced. (We can write the scripts for negative
testing also but it is hectic task).When we talk about real environment
we do negative testing manually.

166. How will you choose a tool for test automation?


choosing of a tool depends on many things ...1. Application to be
tested2. Test environment3. Scope and limitation of the tool.4. Feature
of the tool.5. Cost of the tool.6. Whether the tool is compatible with
your application which means tool should be able to interact with your
application7. Ease of use

167. How you will evaluate the tool for test automation?
We need to concentrate on the features of the tools and how this could
be beneficial for our project. The additional new features and the
enhancements of the features will also help.

168. What are main benefits of test automation?


FAST ,RELIABLE,COMPREHENSIVE,REUSABLE

169. What could go wrong with test automation?


1. The choice of automation tool for certain technologies 2. Wrong set
of test automated

170. How you will describe testing activities?


Testing activities start from the elaboration phase. The various testing
activities are preparing the test plan, Preparing test cases, Execute the
test case, Log the bug, validate the bug & take appropriate action for
the bug, Automate the test cases.

171. What testing activities you may want to automate?


Automate all the high priority test cases which needs to be executed
as a part of regression testing for each build cycle.

172. Describe common problems of test automation.


The common problems are:1. Maintenance of the old script when there
is a feature change or enhancement2. The change in technology of the
application will affect the old scripts

173. What types of scripting techniques for test automation do you


know?
5 types of scripting techniques:LinearStructuredSharedData DrivenKey
Driven

174. What are principles of good testing scripts for automation?


1. Proper code guiding standards2. Standard format for defining
functions, exception handler etc3. Comments for functions4. Proper
errorhandling mechanisms5. The appropriate synchronisation
techniques18. What tools are available for support of testing during
software development life cycle?Testing tools for regression and
load/stress testing for regression testing like, QTP, load runner,
rational robot, winrunner, silk, testcomplete, Astra are available in the
market. -For defect tracking BugZilla, Test Runner are available.

175. Can the activities of test case design be automated?


As I know it, test case design is about formulating the steps to be
carried out to verify something about the application under test. And
this cannot be automated. However, I agree that the process of
putting the test results into the excel sheet.

176. What are the limitations of automating software testing?


Hard-to-create environments like “out of memory”, “invalid
input/reply”, and “corrupt registry entries” make applications behave
poorly and existing automated tools can’t force these condition - they
simply test your application in “normal” environment.

177. What skills needed to be a good test automator?


1.Good Logic for programming.2. Analytical skills.3.Pessimestic in
Nature.

178. How to find that tools work well with your existing system?
1. Discuss with the support officials2. Download the trial version of the
tool and evaluate3. Get suggestions from people who are working on
the tool

179. Describe some problem that you had with automating testing tool
1. The inability of winrunner to identify the third party control like
infragistics controls2. The change of the location of the table object will
cause object not found error.3. The inability of the winrunner to
execute the script against multiple languages

180. What are the main attributes of test automation?


Maintainability, Reliability, Flexibility, Efficiency, Portability,
Robustness, and Usability - these are the main attributes in test
automation.

181. What testing activities you may want to automate in a project?


Testing tools can be used for :* Sanity tests(which is repeated on
every build),* stress/Load tests(U simulate a large no of users, which
is manually impossible) &* Regression tests(which are done after
every code change)

182. How to find that tools work well with your existing system?
To find this, select the suite of tests which are most important for your
application. First run them with automated tool. Next subject the same
tests to careful manual testing. If the results are coinciding you can
say your testing tool has been performing.

183. How will you test the field that generates auto numbers of AUT
when we click the button 'NEW" in the application?
We can create a textfile in a certain location, and update the auto
generated value each time we run the test and compare the currently
generated value with the previous one will be one solution.

184. How will you evaluate the fields in the application under test
using automation tool?
We can use Verification points(rational Robot) to validate the fields
.Ex.Using objectdata, objectdata properties VP we can validate fields.

185. Can we perform the test of single application at the same time
using different tools on the same machine?
No. The Testing Tools will be in the ambiguity to determine which
browser is opened by which tool.

186. Difference between Web application Testing and Client Server


Testing. State the different types for Web application Testing and
Client Server Testing types?
which winrunner 7.2 version compatible with internet explorer, firefox

187. What is 'configuration management'?


Configuration management is a process to control and document any
changes made during the life of a project. Revision control, Change
Control, and Release Control are important aspects of Configuration
Management.

188. How to test the Web applications?


The basic difference in webtesting is here we have to test for URL's
coverage and links coverage. Using WinRunner we can conduct
webtesting. But we have to make sure that Webtest option is selected
in "Add in Manager". Using WR we cannot test XML objects.

189. What are the problems encountered during the testing the
application compatibility on different browsers and on different
operating systems
Font issues, alignment issues

190. How exactly the testing the application compatibility on different


browsers and on different operating systems is done
Please submit Your Suggestion to our forum at
http://www.fyicenter.com

191. How testing is proceeded when SRS or any other document is not
given?
If SRS is not there we can perform Exploratory testing. In Exploratory
testing the basic module is executed and depending on its results, the
next plan is executed.

192. How do we test for severe memory leakages ?


By using Endurance Testing . Endurance Testing means checking for
memory leaks or other problems that may occur with prolonged
execution.

193. What is the difference between quality assurance and testing?


Quality assurance involves the entire software development process
and testing involves operation of a system or application to evaluate
the results under certain conditions. QA is oriented to prevention and
Testing is oriented to detection.

194. Why does software have bugs?


1.miscommunication2.programming errors3.time pressures.4.changing
requirements5.software complexity

195. How do u do usability testing, security testing, installation


testing, ADHOC, safety and smoke testing?
Please submit Your Suggestion to our forum at
http://www.fyicenter.com
196. What is memory leaks and buffer overflows ?
Memory leaks means incomplete deallocation - are bugs that happen
very often. Buffer overflow means data sent as input to the server that
overflows the boundaries of the input area, thus causing the server to
misbehave. Buffer overflows can be used.

197. What are the major differences between stress testing,load


testing,Volume testing?
Stress testing means increasing the load ,and checking the
performance at each level. Load testing means at a time giving more
load by the expectation and checking the performance at that level.
Volume testing means first we have to apply initial.

198. What is Exhaustive Testing?


Testing which covers all combinations of input values and
preconditions for an element of the software under test.

199. What is Functional Decomposition?


A technique used during planning, analysis and design; creates a
functional hierarchy for the software.

200. What is Functional Specification?


A document that describes in detail the characteristics of the product
with regard to its intended features.

1.What is 'Software Quality Assurance'?

Software QA involves the entire software development Process - monitoring and


improving the process, making sure that any agreed-upon standards and procedures are
followed, and ensuring that problems are found and dealt with.

It is oriented to 'prevention'. (See the Books section for a list of useful books on Software
Quality Assurance.)

2.What is 'Software Testing'?


Testing involves operation of a system or application under controlled conditions and
evaluating the results (eg, 'if the user is in interface A of the application while using
hardware B, and does C, then D should happen'). The controlled conditions should
include both normal and abnormal conditions. Testing should intentionally attempt to
make things go wrong to determine if things happen when they shouldn't or things don't
happen when they should. It is oriented to 'detection'.

Organizations vary considerably in how they assign responsibility for QA and testing.
Sometimes they're the combined responsibility of one group or individual. Also common
are project teams that include a mix of testers and developers who work closely together,
with overall QA processes monitored by project managers. It will depend on what best
fits an organization's size and business structure.

7.What is verification? validation?

* Verification typically involves reviews and meetings to evaluate documents, plans,


code, requirements, and specifications. This can be done with checklists, issues lists,
walkthroughs, and inspection meetings. Validation typically involves actual testing and
takes place after verifications are completed. The term 'IV & V' refers to Independent
Verification and Validation.

8.What is a 'walkthrough'?

* A 'walkthrough' is an informal meeting for evaluation or informational purposes. Little


or no preparation is usually required.

Potrebbero piacerti anche