ISTQB Certification
ISTQB Certification
1. Summarize, read, learn and memmorize Certified Tester Fundation Level of ISTQB
certification.
2. Look for some common interview question and preparate structured responses to them
(including current work experience).
3. Extract some basic concepts on SQL and Other software tools listed in Apymsa work
request.
4. Repasar Agile.
5. Hacer algunos ejercicios sobre extraer pruebas a un software en particular.
ISTQB (R)
What is testing?
Software testing is a set of activities to discover defects and evaluate the quality of software
artifacts. These artifacts, when being tested, are known as test objects.
Testing do not only consists of executing tests, it also includes other activities and must be aligned
with the software development lifecycle.
Testing does not only focuses entirely on verifying the test object. Testing involves verification like
checking whether the system meets specified requirements, but it also involves validation, which
means checking whether the system meets users' and other stakeholders' needs in its operational
environment.
Testing may be dynamic or static. Dynamic testing involves the execution of software, Static testing
does not. Static testing includes reviews and static analysis. Dynamic testing uses different types of
test techniques and test approaches to derive test cases.
Test Objectives:
Evaluating work products such as requirements, user stories, designs and code.
Triggering failures and finding defects.
Ensuring required coverage of a test object.
Reducing the level of risk of inadequate software quality.
Verifying whether specified requirements have been fuldilled.
Verifying that a test object complies with contractual, legal, and regulatory requirements.
Provide information to stakeholders to allow them to make informed decisions.
Building confidence in the quality of the test object.
Validating whether the test object is complete and works as expected by the stakeholders.
Testing and Debugging are separate activities. Testing can trigger failures that are caused by defects
in the software (dynamic testing) or can directly find defects in the test object (static testing). When
dynamic testing triggers a failure, debugging is concerned with finding causes of this failure
(defects), analyzing these causes, and eliminating them. The typical debugging process in this cases
involves:
Reproduction of a failure
Diagnosis (finding the root cause)
Fixing the cause
Subsequent confirmation testing checks whether the fixes resolved the problem. Preferably,
confirmation testing is done by the same person who performed the initial test. Subsequent
regression testing can also be performed, to check whether the fixes are causing failures in other
parts of the test object.
When static testing identifies a defect, debugging is concerned with removing it. There is no need
for reproduction or diagnosis, since static testing directly finds defects, an cannot cause failures.
Why is testing necessary?
Testing provides cost-effective means of detecting defects. These defects can then be removed (by
debugging - a non-testing activity), so testing indirectly contributes to higher quality test objects.
Testing provides a means of directly evaluating the quality of a test object at various stages in the
SDLC. These measures are used as part of a larger project management activity, contributing to
decisions to move to the next stage of the SDLC, such as release decision. Testers ensure that their
understanding of users' needs are considered throughout the development lifecycle. The alternative
is to involve a representative set of users as part of the development project, which is not usually
possible due to the high costs and lack of availability of suitable users.
While people often use the terms "testing" and "quality assurance (QA)" interchangeably, testing
and QA are not the same. Testing is a form of quality control (QC). QC is a product-oriented,
corrective approach that focuses on those activities supporting the achievement of appropriate levels
of quality. Testing is a major form of quality control, while others include formal methods (model
checking and proof of correctness), simulation and prototyping.
QA is a process-oriented, preventive approach that focuses on the implementation and improvement
of processes. It works on the basis that if a good process is followed correctly, then it will generate a
good product. QA applies to both the development and testing processes, and is the responsibility of
everyone on a project. Test results are used by QA and QC. In QC they are used to fix defects, while
in QA they provide feedback on how well the development and testing processes are performing.
Errors, Defects, Failures and Root Causes
Human beings make errors (mistakes), which produce defects (faults, bugs), which in turn may
result in failures. Humans make errors for various reasons, such as time pressure, complexity of
work products, processes, infraestructure or interactions, or simply because they are tired or lack of
adequate training.
Defects can be found in documentation, such as requirements specification or a test script, in source
code, or in a supporting artifact such as a build file. Defects in artifacts produced earlier in the
SDLC, if undetected, often lead to defective artifacts later in the lifecycle. If a defect in code is
executed, the system may fail to do what it should do, or do something it shouldn't, causing a
failure. Some defects will always result in a failure if executed, while others will only result in a
failure in specific circumstances, and some may never result in a failure.
A root cause is a fundamental reason for the occurrence of a problem (a situation that leads to an
error). Root causes are identified through root cause analysis, which is typically performed when a
failure occurs or a defect is identified. It is believed that further similar failures or defects can be
prevented or their frequency reduced by addressing the root cause, such as by removing it.
The seven testing principles:
1. Testing shows the presence, not the absence of defects: Testing can show that defects are
present in the test object, but cannot prove that there are no defects. Testing reduces the
probability of defects remaining undiscovered in the test object, but even if no defects are
found, testing cannot prove test object correctness.
2. Exhaustive testing is impossible: Testing everything is not feasible except in trivial cases.
Rather than attempting to test exhaustively, test techniques, test case prioritization, and risk-
based testing, should be used to focus testing efforts.
3. Early testing saves time and money: Defects that are removed early in the process will
not cause subsequent defects in derived work products. The cost of quality will be reduced
since fewer failures will occur later in the SDLC. To find defects early, both static testing
and dynamic testing should be started as early as possible.
4. Defects cluster together: A small number of system components usually contain most of
the defects discovered or are responsible for most of the operational failures. This
phenomenon is an ilustration of the pareto principle. Predicted defect clusters, and actual
defect clusters observed during testing or in operation, are an important input for risk-based
testing.
5. Tests wear out: If same tests are repeated many times, they become increasingly
ineffective in detecting new defects. To overcome this effect, existing tests and test data
may need to be modified, and new tests may need to be written. However, in some cases,
repeating the same tests can have a beneficial outcome, like in automated regression testing.
6. Testing is context dependent: There is no single universally applicable approach to
testing. Testing is done differently in different contexts.
7. Absence-of-defects fallacy: It is a fallacy to expect that software verification will ensure
the success of a system. Thoroughly testing all the specified requirements and fixing all the
defects found could still produce a system that does not fulfill the users' needs and
expectations, that does not help in achieving the customer's business goals, and that is
inferior compared to other competing systems. In addition to verification, validation should
also be carried out.
Test activities, Testware and Test Roles.
Testing is context dependent, but there are some common sets of test activities at a high level
without which testing is less likely to achieve test objectives. These sets of test activities form a test
process. The test process can be tailored to a given situation based on various factors. Which test
activities are included in this test process, how they are implemented, and when they occur is
normally decided as part of the test planning for the specific situation.
A test process usually consists of the main groups of activities described below. Although many of
these activities may appear to follow a logical sequence, they are often implemented iteratively or in
parallel. These testing activities usually need to be tailored to the system and the project.
Test planning: Defining the test objectives and then selecting an approach that best
achieves the objectives within the constraints imposed by the overall context.
Test monitoring and control: Test monitoring involves the ongoing checking of all test
activities and the comparison of actual progress against the plan. Test control involves
taking the actions necessary to meet the objectives of testing.
Test analysis: Includes analyzing the test basis to identify testable features and to define
and prioritize associated test conditions, together with the related risks and risk levels. The
test basis and the test objects are also evaluated to identify defects that may contain and to
assess their testability. Test analysis is often supported using test techniques. Test analysis
answers the question "what to test?" in terms of measurable coverage criteria.
Test design: Includes elaborating the test conditions into test cases and other testware (like
test charters). This activity often involves the identification of coverage items, which sever
as a guide to specify test case inputs. Test techniques can be used to support this activity.
Test design also includes defining the test data requirements, designing the test environment
and identifying any other required infrastructure and tools. Test design answers the question
"how to test?".
Test implementation: It includes creating or acquiring the testware necessary for test
execution (like test data). Test cases can be organized into test procedures and are often
assembled into test suites. Manual and automated test scripts are created. Test procedures
are prioritized and arranged within a test execution schedule for efficient test execution.
The test environment is built and verified to be set up correctly.
Test execution: Includes running the tests in accordance with the test execution schedule
(test runs). Test execution may be manual or automated. Test execution can take many
forms, including continuous testing or pair testing sessions. Actual test results are compared
with the expected results. The test results are logged. Anomalies are analyzed to identify
their likely causes. This analysis allows us to report the anomalies based on the failures
observed.
Test completion: these activities usually occur at project milestones (release, end of
iteration, test level completion) for any unresolved defects, change requests or product
backlog items created. Any testware that may be useful in the future is identified and
archived or handed over to the appropriate teams. The test environment is shut down to an
agreed state. The test activities are analyzed to identify lessons learned and improvements
for future iterations, releases, or projects. A test completion report is created and
communicated to the stakeholders.
Testing is not performed in isolation, test activities are an integral part of the development processes
carried out within an organization. Testing is also funded by stakeholders and its final goal is to help
fulfill the stakeholders’ business needs. Therefore, the way the testing is carried out will depend on
a number of contextual factors including:
Traceability of test cases to requirements can verify that the requirements are covered by
test cases.
Traceability of test results to risks can be used to evaluate the level of residual risk in a
test object.
Good traceability makes it possible to determine the impact of changes, facilitates test audits, and
helps meet IT governance criteria. It makes test progress and completion reports more easily
understandable by including the status of test basis elements. It can help to communicate the
technical aspects of testing to stakeholders in an understandable manner.
Roles in testing.
Two principal roles: A test manager role and a testing role. The activities assigned to these roles
depend on several factors such as the project and product context, skills or the organization.
Test management role takes overall responsibility for the test process, test team and leadership of
the test activities. It is mainly focused on the activities of test planning, test monitoring and control,
and test completion. In Agile software development some of the test management tasks may be
handled by the Agile team, while tasks that are covered by multiple teams or the entire organization
may be performed by test managers outside of the development team.
Testing role takes overall responsibility for the engineering (techincal) aspect of testing. The testing
role is mainly focused on the activities of test analysis, test design, test implementation and test
execution.
Different organizations may take these roles at different ways. It is also possible for one person to
take on the roles of testing and test management at the same time.
Essential skills and good practices in testing.
Scope and timing of the test activities: Test levels and test types.
Level of detail of test documentation.
Choice of test techniques and test approach.
Extent of test automation.
Role and responsibilities of a tester.
In sequential development models, in the initial phases testers typically participate in requirement
reviews, test analysis, and test design. The executable code is usually created in the later phases, so
typically dynamic testing cannot be performed early in the SDLC.
In some iterative and incremental development models, it is assumed that each iteration delivers a
working prototype or product increment. This implies that in each iteration both static and dynamic
testing may be performed at all test levels. Frequent delivery of increments requires fast feedback
and extensive regression testing.
Agile software development assumes that change may occur throughout the project. Therefore,
lightweight work product documentation and extensive test automation to make regression testing
easier are favored in agile projects. Also, most of the manual testing tends to be done using
experience-based test techniques that do not require extensive prior test analysis and design.
SDLC and good Testing Practices.
For every software development activity, there is a corresponding test activity, so that all
development activities are subject to quality control.
Different test levels have specific and different test objectives, which allows for testing to
be appropriately comprehensive while avoiding redundancy.
Test analysis and design for a given test level begins during the corresponding development
phase of the SDLC, so that testing can adhere to the principle of early testing.
Testers are involved in reviewing work products as soon as drafts of this documentation are
available, so that this earlier testing and defect detection can support the shift-left strategy.
TDD, ATDD and BDD are similar development approaches, where tests are defined as a means of
direct development. Each of these approaches implements the principle of early testing and follows
a shift-left approach. Since the tests are defined before the code is written. They should support an
iterative development model. These approaches are characterized as follows:
TDD:
Directs the coding through test cases (instead of extensive software design)
Test are written first, then the code is written to satisfy the tests, and then the tests and code
are refactored.
Acceptance Test-Driven Development (ATDD):
Derives tests from acceptance criteria as part of the system design process
Tests are written before the part of the application is developed to satisfy the tests
Behaviour-Driven Development (BDD):
Expresses the desired behaviour of an application with test cases written in a simple form of
natural language, which is easy to understand by stakeholders – Usually using the
Given/When/Then format.
Test cases are then automatically translated into executable tests.
For all the above approaches, tests may persist as automated tests to ensure the code quality in
future adaptions / refactoring.
DevOps and Testing
DevOps is an organizational approach aiming to create synergy by getting development (including
testing) and operations to work together to achieve a set of common goals. DevOps requires a
cultural shift within an organization to bridge the gaps between development and operations while
treating their functions with equal value. DevOps promotes team autonomy, fast feedback,
integrated toolchains, and technical practices like continuous integration and continuous delivery.
This enables the teams to build, test and release high-quality code faster through a DevOps delivery
pipeline.
Benefits of DevOps from Testing perspective:
Fast feedback on the code quality, and whether changes adversely affect existing code
CI promotes a shift-left approach in testing by encouraging developers to submit high
quality code accompanied by component tests and static analysis.
Promotes automated processes like CI/CD that facilitate establishing stable test
environments.
Increases the view on non-functional quality characteristics: performance, reliability.
Automation through a delivery pipeline reduces the need for repetitive manual testing
The risk in regression is minimized due to the scale and range of automated regression tests.
Risks and challenges of DevOps:
Reviewing the specification from the perspective of testing. These review activities on
specifications often find potential defects, such as ambiguities, incompleteness, and
inconsistencies.
Writing test cases before the code is written and have the code run in a test harness during
code implementation.
Using CI and even better CD as it comes with fast feedback and automated component tests
to accompany source code when it is submitted to the code repository.
Completing static analysis of source code prior to dynamic testing, or as part of an
automated process.
Performing non-functional testing starting at the component test level, where possible. This
is a form of shift-left as these non-functional test types tend to be performed later in the
SDLC when a complete system and a representative test environment are available.
A shift-left approach might result in extra training, effort and/or costs earlier in the process but is
expected to save efforts and/or costs later in the process. For this approach it is important that
stakeholders are convinced and bought into this concept.
Project retrospectives.
Are often held at the end of a project or an iteration, at a release milestone, or can be held when
needed. The timing depends on the particular SDLC model being followed. In these meetings
participate testers, developers, architects, product owner, business analysts and they discuss:
Functional Testing: Evaluates the functions that a component or a system should perform.
The functions are “what” the test object should do. The main objective of functional testing
is checking the functional completeness, functional correctness and functional
appropriateness.
Non-Functional testing: Evaluates attributes other than functional characteristics of a
component or system. Non-functional testing is the testing of “how well the system
behaves”. The main objective of non-functional testing is checking the non-functional
software quality characteristics.
o Performance efficiency
o Compatibility
o Usability
o Reliability
o Security
o Maintainability
o Portability
Many non-functional tests are derived from functional tests as they use the same functional tests,
but check that wile performing the function, a non-functional constraint is satisfied (checking that a
function performs within an specified time, or a function can be ported to a new platform). The late
discovery of non-functional defects can pose a serious threat to the success of a project. Non-
functional testing sometimes needs a very specific test environment, such as a usability lab for
usability testing.
Confirmation Testing: Confirms that an original defect has been successfully fixed.
Depending on the risk, one can test the fixed version of the software in several ways,
including:
o Executing all test cases that previously have failed due to the defect
o Adding new tests to cover any changes that were needed to fix the defect
o However when time or money is short when fixing defects, confirmation testing
might be restricted to simply exercising the steps that should reproduce the failure
caused by the defect and checking that the failure does not occur.
Regression Testing: Confirms that no adverse consequences have been caused by a change,
including a fix that has already been confirmation tested. These adverse consequences could
affect the same component where the change was made, other components in the same
system, or even other connected systems. Regression testing may not be restricted to the
test object itself but can also be related to the environment. It is advisable first to perform
an impact analysis to optimize the extent of the regression testing. Impact analysis shows
which parts of the software could be affected.
Regression test suites are run many times and generally the number of regression test cases
will increase with each iteration or release, so regression testing is a strong candidate for
automation. Automation of these tests should start early in the project. Where CI is used,
such as in DevOps, it is good practices to also include automated regression tests.
Depending on the situation, this may include regression tests on different levels.
Static Testing:
In static testing the software under test does not need to be executed. Code, process specification,
system architecture specification or other work products are evaluated through manual examination
or with the help of a tool. Test objectives include improving quality, detecting defects and assessing
characteristics like readability, completeness, correctness, testability and consistency. Static testing
can be applied for both verification and validation.
Testers, business representatives and developers work together during example mappings,
collaborative user story writing and backlog refinement sessions to ensure that user stories and
related work products meet defined criteria. The definition of ready review techniques can be
applied to ensure user stories are complete and understandable and include testable acceptance
criteria. By asking the right questions, testers explore, challenge and help improve the proposed
user stories.
Static analysis can identify problems prior to dynamic testing with less effort since no test cases are
required, and tools are typically used. Static analysis is often incorporated into CI frameworks.
While largely used to detect specific code defects, is also used to evaluate maintainability and
security. Spelling checkers and readability tools are other examples of static analysis tools.
Work Products Examinable by Static Testing
Requirement specification documents, source code, test plans, test cases, product backlog items, test
charters, project documentation, contracts and models.
Any product that can be read and understood can be subject of a review. However for static testing,
work products need a structure against which they can be checked.
Value of Static Testing
It can detect defects in the earliest phases of the SDLC. It can also identify defects which cannot be
detected by dynamic testing like unreachable code, design patterns not implemented as desired,
defects in non-executable work products.
It is recommended to involve a wide variety of stakeholders in static testing. Even though reviews
can be costly to implement, the overall project costs are usually much lower than when no reviews
are performed because less time and effort needs to be spent on fixing defects later in the project.
Code defects can be detected using static analysis more efficiently than in dynamic testing, usually
resulting in both fewer code defects and a lower overall development effort.
Static and dynamic testing can both lead to the detection of defects, however there are some
defect types that can only be found by either static or dynamic testing.
Static testing finds defects directly, while dynamic testing causes failures from which the
associated defects are determined through subsequent analysis.
Static testing may more easily detect defects that lay on paths through the code that are
rarely executed or hard to reach using dynamic testing.
Static testing can be applied to non-executable work products, while dynamic testing can
only be applied to executable work products.
Static testing can be used to measure quality characteristics that are not dependent on
executing code like maintainability, while dynamic testing can be used to measure quality
characteristics that are dependent on executing code like performance efficiency.
Defects easier or cheaper to find with static testing:
Informal review: Do not follow a defined process and do not require a formal documented
output. The main objective is detecting anomalies.
Walkthrough: It is lead by the author and can serve many objectives, such as evaluating
quality and building confidence in the work product, educating reviewers, gaining
consensus, generating new ideas.
Technical Review: Is performed by technically qualified reviewers and led by a moderator.
The objectives of a technical review are to gain consensus and make decisions regarding a
technical problem, but also detect anomalies, evaluate quality and build confidence in the
work product.
Inspection: Are the most formal type of review, the follow a complete generic process. The
main objective is to find the maximum number of anomalies.
Test techniques.
Test techniques support the tester in test analysis (what to test) and in test design (how to test). Test
techniques help to develop a relatively small, but sufficient, set of test cases in a systematic way.
Test techniques also help the tester to define test conditions, identify coverage items, and identify
test data during the test analysis and design.
In iterative SDLCs, typically two kinds of planning occur: release planning and iteration planning.
Release planning looks ahead to the release of a product, defines and re-defines the product
backlog, and may involve refining larger user stories into a set of smaller user stories. It also serves
as the basis for the test approach and test plan across all iterations. Testers involved in release
planning participate in writing testable user stories and acceptance criteria, participate in project and
quality risk analyses, estimate test effort associated with user stories, determine the test approach,
and plan the testing for the release.
Iteration planning looks ahead to the end of a single iteration and is concerned with the iteration
backlog. Testers involved in iteration planning participate in the detailed risk analysis of user
stories, determine the testability of user stories, break down user stories into tasks, estimate test
effort for all testing tasks, and identify and refine functional and non-functional aspects of the test
object.
Entry Criteria and Exit Criteria.
Entry criteria define the preconditions for undertaking a given activity. If entry criteria are not met,
it is likely that the activity will prove to be more difficult, time-consuming, costly, and riskier. Exit
criteria define what must be achieved in order to declare an activity completed. Entry criteria and
exit criteria should be defined for each test level, and will differ based on the test objectives.
Typical entry criteria include: availability of resources (people, tools, environment, test data,
budget, time) availability of testware (test basis, testable requirements, user stories, test cases), and
initial quality level of a test object (all smoke tests have passed).
Typical exit criteria include: measure of thoroughness (achived level of coverage, number of
unresolved defects, defect density, number of failed test cases), and completion criteria (planned
tests have been executed, static testing has been performed, all defects found are reported, all
regression tests are automated).
Running out of time or budget can also be viewed as valid exit criteria. Even without other exit
criteria being satisfied, it can be acceptable to end testing under such circumstances, if the
stakeholders have reviewed and accepted the risk to go live without further testing.
In Agile software development, exit criteria are often called definition of done, defining the team’s
objective metrics for a releasable items. Entry criteria that a user story must fulfull to start the
development and/or testing activities are called Definition of Ready,
Estimation techniques
Involves predicting the amount of test-related work needed to meet the objectives of a test project.
It is important to make it clear to the stakeholders that the estimate is based on a number of
assumptions and is always subject to estimation error. Estimation for small tasks is usually more
accurate than for the large ones. Therefore when estimating a large task it can be decomposed into a
set of smaller tasks which then in turn can be estimated.
Risk-based prioritization: Where the order of test execution is base on the results of risk
analysis. Test cases covering the most important risks are executed first.
Coverage-based prioritization: Where the order of test execution is based on coverage. Test
achieving the highest coverage are executed first.
Requirements-based prioritization: Where the order of test execution is based on the
priorities of the requirements traced back to the corresponding test cases. Requirement
priorities are defined by stakeholders. Test cases related to the most important requirements
are executed first.
The order of the test execution must also take into account the availability of resources. For
example, the required test tools, test environments or people that may only be available for specific
time window.
Risk Management
Risk management allows the organization to increase the likelihood of achieving objectives,
improve the quality of their products and increase stakeholders confidence and trust.
The main risk management activities are:
User dissatisfaction
Loss of revenue, trust, reputation
Damage to third parties
High maintenance costs, overload of the helpdesk
Criminal penalties
In extreme cases, physical damage, injuries or even death.
Product Risk Analysis
The goal is to provide an awareness of product risk in order to focus the testing in a way that
minimizes the residual level of product risk. Ideally, product risk analysis begins early in the SDLC.
It consists of risk identification, and risk assessment. Risk identification is about generating a
comprehensive list of risks. Stakeholders can identify risks by using various techniques and tools,
like brainstorming, workshops, interviews, or cause-effect diagrams. Risk assessment involves
categorization of identified risks, determining their risk likelihood, risk impact and level,
prioritizing, and proposing ways to handle them. Categorization helps in assigning mitigation
actions, because usually risks falling into the same category can be mitigated using a similar
approach.
Product risk analysis may influence the thoroughness and scope of testing. Its results are used to:
Test period
Test progress, including any notable deviations
Impediments for testing
Test metrics
New and changed risks within testing period
Testing planned for the next period.
A test completion report is prepared during test completion, This report uses test progress reports
and other data, Typical test completion reports include:
Test summary
Testing and product quality evaluation based on the original test plan.
Deviations from the test plan
Testing impediments and workarounds
Test metrics based on test progress resports
Unmitigated risks, defects not fixed.
Lessons learned that are relevant to the testing.
Defect Management
Typical defect reports have the following objectives:
Provide those responsible for handling and resolving reported defects with sufficient
information to resolve the issue
Provide a means of tracking the quality of the work product.
Provide ideas for improvement of the development and test process.
A defect report logged during dynamic testing typically includes:
Unique identifier
Title with a shot summary of the anomaly being reported
Date when the anomaly was observed, issuing organization, and author, including their role.
Identification of the test object and test environment.
Context of the defect (test case being run, test activity being performed, SDLC phase, and
other relevant information such as the test technique, checklist or test data being used).
Description of the failure to enable reproduction and resolution including the steps that
detected the anomaly, and any relevant test logs, database dumps, screeshots, or recordings.
Expected results and actual results
Severity of the defect on the interests of the stakeholders or requirements
Priority to fix
Status of the defect: Open, deferred, duplicate, waiting to be fixed, awaiting confirmation
testing, re-opened, closed, rejected.
References to the test case.
Test Tools
Management tools: Increase the test process efficiency by facilitating management of the
SDLC, requirements, tests, defects, configuration.
Static Testing Tools: Support the tester in performing Reviews and static analysis
Test design and implementation tools: Facilitate the generation of test cases, test data and
test procedures.
Test execution and coverage tools: Facilitate automated test execution and coverage
measurement.
Non-functional testing tools: Allow the tester to perform non-functional testing that is
difficult or impossible to perform manually.
DevOps tools: Support the DevOps delivery pipeline, workflow tracking, automated build
process, CI/CD.
Collaboration tools: Facilitate communication.
Tools supporting scalability and deployment standardization: Virtual Machines,
Containerization tools.
Any other tool that assists in testing (a spreadsheet is a test tool in the context on testing).
Test Automation
Benefits of Test Automation
Time saved by reducing repetitive manual work: Execute regression tests, re-enter the same
test data, compare expected results to actual results, and check against coding standards.
Prevention of simple human errors through greater consistency and repeatability: Test are
consistently derived from requirements, test data is created in a systematic manner, and
tests are executed by a tool in the same order with the same frequency.
More objective assessment and providing measures that are too complicated for humans to
derive.
Easier access to information about testing to support test management and test reporting:
statistics, graphs, and aggregated data about test progress, defect rates, and test execution
duration.
Reduced risk test execution times to provide earlier defect detection, faster feedback and
faster time to market.
More time for testers to design new, deeper and more effective tests.
Potential risks of using automation:
Unrealistic expectations about the benefits of a tool: including functionality and ease of use.
Inaccurate estimations of time, costs, effort required to introduce a tool, maintain test
scripts and change existing manual scripts process.
Using a test tool when manual testing is more appropriate.
Relying on a tool too much: Ignoring the need of human critical thinking.
The dependency on the tool vendor which may go out of business, retire the too, sell the
tool to a different vendor or provide poor support.
Using an open-source software which might be abandoned, meaning that no further updates
are available, or its internal components may require quite frequent updates as a further
development.
The automation tool is not compatible with the development platform.
Choosing an unsuitable tool that did not comply with the regulatory requirements and/or
safety standards.