Important Manual Testing Interview Questions and Answers
We have collected some really important manual testing interview questions and answers for our students. These unlike other blogs are not just one liners but we have explained each and every asnwer for every question in brief so that our students get the best and are all prepared for their interview.
Below are some of the most asked manual testing interview questions:
Explain me about Software Testing Life Cycle (STLC)?
Software Testing Life Cycle will have below phases: Every tester in software testing will go through below phases.
- Understanding Requirements / User Stories: Understanding Use Stories (if you are in Agile model), OR understanding requirements by going through Software Requirement Specifications(if you are in Non-Agile model) which we generally call as SRS
- Clarify your doubts: Raise your doubts / questions after going through above user stories or requirements and get them clarified
- Writing the Test Cases: Here you guys start writing the test cases based on understandings from above user stories / requirements. We generally use test management tools to write test cases. We have so many tools available in the market to use like TFS which is from Microsoft and Test Link etc. Most of the tools are similar to use.
- Review of Test Cases: Once test cases are prepared they will be reviewed by internal testing team. We call it as peer review. I mean let’s say we have 4 testers in the project and one tester test cases will be reviewed by other tester and provide review comments. We also send these test cases to client for review but not all the time. It differs to client to client. If client request we have to send for review and get them approved
- Execute Test Cases: Once system is given for testing team to do testing we execute the test cases which are prepared. IF any test case gets failed we go ahead and raise defect in defect management tool for tracking.
- Defect Logging: While doing the test execution we raise defect when we see actual result is not matching with expected result. Once we raise defect it will be discussed in triage calls and make active OR reject based on discussion with stake holders.
- Defect Re-testing: Once defect is activated after discussion development team will start working on these all defects which are agreed as defect.
- Automation: Now a days automation has become mandatory in software testing industry as most of the projects are running in Agile. Once one release / sprint is completed we do automation of manual test cases which generally cover functional flow.
What is black box testing?
Black box testing is a kind of testing where tester performs the testing without knowing internal structure of code OR Architecture of the system. Here we perform testing by giving input and validate output is correct as per the input given. Let’s take below two examples to understand it better:
- Example 1: Let’s take a calculator where it allows to perform some arithmetic operations. Let’s say I would like to add two 2 numbers, so just input number say 20 then press “+” then enter 40 then press “=” we get output as 60. That’s it. We don’t know internally what is the code written to do this OR we don’t know what is the logic that is written in the code OR Architecture of system how it is defined. So out input is providing two numbers and validating the output is correct or not in this example we have to validate whether out is 60 or not.
- Example 2: Let’s take an ATM machine where we withdraw some amount. Here we insert card input your pin and select Withdraw option from menu and enter amount. Let’s say we have given input as $1000. We just have to validate to check whether amount what is given as output from ATM machine is $1000 or not. Here also we don’t what is technology in which ATM is implemented, we don’t know Architecture of the ATM, we don’t know what is the logic written in the code.
So in black box testing we just provide input and check output without knowing below
- What is the logic that is written in code to perform specific action
- How is the architecture of system is defined
- What is the technology in which system is developed
What are different techniques in this black box testing
We have below techniques which are used in black box testing
- Boundary value analysis (BVA)
- Equivalence Partitioning
- Error Guessing
Explain Boundary Value Analysis Technique?
We use Boundary Value Analysis testing in black box testing. We generally use this technique to test if any field is taking range of data. As the name says it always checks with the boundaries of the range. It checks for both positive and negative cases.
Let’s take an example: I have an application in which we have to enter employ number in order to create new employ in the system. Let’s say requirement given for this filed is it should be minimum of 1 and maximum of 1000 as minimum and maximum range. For this example if we use boundary value analysis we have to test with by entering following numbers
Valid / Positive cases:
- Enter 1 as input(this is minimum value)
- Enter 2 as input(this is minimum+1)
- Enter 1000 as input(this is maximum value)
- Enter 999 as input(this is maximum-1)
Invalid / Negative cases:
- Enter Zero as input(this is minium-1)
- Enter 1001 as input(this is maximum+1)
So here we are always testing application with boundaries of the minimum & maximum of the range that is given as part of requirement for a specific field. That’s it how we use this technique for testing.
What is Equivalence Partitioning?
Equivalence Partitioning is another block box testing technique which is used. In case of Boundary Value Analysis we try to input data with boundaries of minimum and maximum
In equivalence Partitioning we basically have two classes one is valid and other tow are invalid.
Example: Let’s take same example which understood from above question for BVA and see how we input in case of Equivalence Partitioning I have an application in which we have to enter employ number in order to create new employ in the system. Let’s say requirement given for this filed is it should be minimum of 1 and maximum of 1000 as minimum and maximum range. For this example if
We test with below one valid and two invalid classes using Equivalence Partitioning
One Valid Class:
Enter any number between the range which is nothing but 1 to 1000 so you can enter any number like say 90
Two In-Valid Classes:
Enter any number above the range which is nothing but more than maximum value which is 1000 in this case. so you can enter any number like say 2000
Enter any number below the range which is nothing but less than minimum value which is 1 in this case. so you can enter any number below range like say -10
What is white box testing?
In block box testing we understand that we do testing by just giving input and validate output without knowing any internal structure of the code, logic what is written, not knowing the technology and also without know the technology as well. But when it comes to white box testing it is reverse to black box which means you guys have to look at the code and verify each and every line of the code and conditions for both true and false outcomes. This testing makes sure whatever code is written is correct and validates at code level.
What are the white box testing techniques available
- Statement Coverage Testing
- Condition Coverage Testing
- Branch coverage also known as Decision coverage
Difference between Verification & Validation
Verification: Verification happens in every phase of SDLC to make sure what we are building is correct or not. It is continues process that is performed by individual resource whoever is responsible for every phase of SDLC. Verification is performed using below
Validation: We do validation to find out whether we build system right or not. Typically validation is performed by testing team.
In simple term to say “Are we building system right are known as verification” & “Did we build system is known as Validation”
9. What is Static Testing and Dynamic Testing
We do static testing before actual system is ready for testing. It means that we do reviews walkthroughs & inspections. Like we do Test Plan review, Test Case review etc before system is actually implemented. Dynamic testing we do once system is implemented and given for testing team to testing.
In simple words to say static testing falls under “Verification” & dynamic testing falls on “Validation”.
10. Explain me about traceability Matrix
Requirement Traceability Matrix or RTM captures all requirements OR user stories given by the client and it helps in tracing test cases which are written against these requirements.
In simple words to say, it is a document that maps and traces client requirement with test cases. The main purpose of Requirement Traceability Matrix is to make sure that testing team written test cases to cover client requirements so that no functionality will miss from testing team under testing.
11. Can you write template for Traceability Matrix
Let’s take an example we have requirements / use stories from client show below, as a tester we write the test cases for every requirement / user story. In below Traceability Matrix template we can see every requirement / user story is mapped with Test Cases that are written by testing team. This way Traceability Matrix allows to make sure we cover all the requirements / user stories from testing team.
12. Explain me different Test Levels we have
We have below different test levels
- Unit Testing
- Integration Testing
- System Testing
- User Acceptance Testing (UAT)
13. What is Integration Testing?
Integration testing mean we do integration of each developer code and do this testing to make sure each developers code is working after integration of each developer’s code.
This testing is mainly performed to makes sure each component is interacts well without any issues so that we can prevent defects in system testing before it is giving to testing team..
14. What is UAT testing?
User Acceptance Testing is performed by client once testing team completed system testing. We get approval from client after doing this testing to make it available for end users.
15. Explain difference between Test Scenario & Test Case
Test Scenario: As a testing team we have to always understand requirements / user stories from client. So for every requirement we always write test scenarios considering what are the scenarios we have to test. For every requirement / user story we write one test scenario. Test scenario is mostly a one line statement which tells what we are going to test as part of this requirement / user story.
Test Case: Test cases will be in detail not like one statement when compared to test scenario. Test case will cover in detail like what are the steps to be covered, what data we have to provide & what is expected result for each step. We generally write multiple test cases for one test scenario.
Example: Let’s take an example to understand it better
Let’s say that we have below requirement / user story from client
“User should be allowed to register an account with FaceBook and duplication of user should not be allowed to create”
For above requirement / user story we write below scenario and test cases
Test Scenario: Verify user allowed to register an account
Test Cases: We write multiple test cases for the above scenario:
- Create an account and check account is created
- Check system is not allowed to create duplication of user
- Verify GUI (Graphical User Interface) of Register Account page.
16. What are the different types of test cases you write?
We always have to consider below types of test cases for each and every requirement / user story
- GUI: Graphical user interface test cases cover UI of application. Here we write test to check whether application screen showing as per the expectations which are given by client or not. Also we check look and feel of the screen like alignment issues, overlapping of the fields and spell check etc.. But for everything we need not to write a test case but we write at least one test case for checking UI.
- Functional:Here we write test cases to check the functionality of application is working as expected or not
- Filed Level Validation: Here we write test cases to check for each and every field for below
- Mandatory Data Validation: If field is mandatory then leave that filed blank click on submit and see system throws valid error message
- Max size data: Enter data more than maximum size specified. Let’s say FirstName filed supposed to take maximum of 100 letters then enter more than 100 letters and check proper error message is displayed or not.
- Minimum size data: Enter data less than minimum size specified. Let’s say FirstName filed supposed to take minimum of 5 letters then enter less than 5 letters and check proper error message is displayed or not.
- Invalid data: Let’s say FirstName field should take only letter then enter some invalid data like #%#%%%^1344 and check proper error message is displayed or not.
17. Difference between Error, defect & Bug?
Error: let’s say developer written code to implement addition of two numbers and where by mistake instead of “+” he added as “-“. So there is an error in the code instead of addition he is doing subtraction. This is called error in the code
Defect: Tester founds this error while doing the testing is called as a defect
Bug: Once defect is agreed as valid from development team then it is called as bug.
18. Explain the Defect Life Cycle.
Once tester finds a defect while doing the testing it will go with below life cycles
- When defect is found it will be raised in one of the defect management tool, it will be with status as “New” (status varies between the defect management tools)
- This defect will be taken with development team / stakeholders in “Triage” status to discuss on this defect
- If this defect is valid and agrees by development team then it will go to status as “Active” and will be assigned to develper. All the defects agreed as valid defects will go as Active.
- If this defect is not valid then it will go as “Rejected”
- Once developer starts working on it will go status as “In-Progress”
- Once developer fixes the defect it will go as “Resolved” and assigned back to tester for re-testing
- Once tester tests the defect and confirms it is working fine then tester will change status of defect to “Closed”
- If defect is still exist even after fix then tester will change status to “Re-Open” as it is not working
- Once defect is in “Re-Open” status by tester, it will go above process again like active, in-progress, “Resolved”
19. What are the different status of defect?
Status of defect varies from tool to tool but below are the common status in most of the tool
20. What is defect severity & priority?
Severity: Tells how severely a defect is impacting to business / system
Priority: Tells urgency of fix of defect how soon it should be fixed
21. Explain me different severities & priorities?
Every defect management tool will have its own severities and priorities. But most the tools will have below
22. As a tester you raised defect and developer is not accepting it as defect. What you do in this situation?
It is common case in most of the projects as understanding of requirement differs from person to person. So have to explain with requirement reference number / user story number which you are referring as it is valid defect. We need to explain that we have not raised defect in assumptions and it is raised as it is not working as per the expected.
23. What is Test Plan and explain me contents of Test Plan?
Test Plan is a document which will have below contents
- What is this document about
- Out Of scope
- Entry criteria
- Exit criteria
- Test Deliverables
- Automation and tolls planned to use
24. When to stop Testing?
When we met the Exit Criteria that is defined in Test Plan. In most the projects below will be exist criteria in common
- All the test cases are executed and no test case is not in run state
- 95% of the test cases are passed
- No test case is in blocked status
- No critical, high & medium defects are in open / active status
25. What you do if no sufficient time is given for testing?
- We plan to test core functionalities of the system
- We try to cover complete end to end flow of testing
- We may not concentrate on field level validation.