Project Monitoring and Evaluation and Its Importance
Project Monitoring and Evaluation and Its Importance
IMPORTANCE
MONITORING
This is the regular systematic collection and analysis of information to track the
progress of program implementation against pre-set targets and objectives. It
aims to answer the question “did we deliver?”
EVALUATION
This is the objective assessment of an ongoing or recently completed project,
program or policy, its design, implementation and results. It answers the
question “What has happened as a result?”
Evaluation Analyzes why intended results were or were not achieved, Assesses
specific casual contributions of activities to results, Examines implementation
process, Explores unintended results, Provides lessons, highlights significant
accomplishments or program potential and offers recommendations for
improvement
Monitoring and evaluation (M&E) helps those involved with any type of
projects to assess if progress desired is being achieved.
Six of the primary mirrors of the James Webb Space Telescope being prepared for acceptance testing
In engineering and its various subdisciplines, acceptance testing is a test conducted to determine if
the requirements of a specification or contract are met. It may involve chemical tests, physical tests,
or performance tests.
In systems engineering it may involve black-box testing performed on a system (for example: a piece
of software, lots of manufactured mechanical parts, or batches of chemical products) prior to its
delivery.[1]
In software testing the ISTQB defines acceptance as: formal testing with respect to user needs,
requirements, and business processes conducted to determine whether a system satisfies the
acceptance criteria and to enable the user, customers or other authorized entity to determine
whether or not to accept the system.[2] Acceptance testing is also known as user acceptance testing
(UAT), end-user testing, operational acceptance testing (OAT) or field (acceptance) testing.
A smoke test may be used as an acceptance test prior to introducing a build of software to the main
testing process.[not verified in body]
Contents
[hide]
1Overview
2Process
3User acceptance testing
4Operational acceptance testing
5Acceptance testing in extreme programming
6Types of acceptance testing
7List of acceptance-testing frameworks
8See also
9References
10Further reading
11External links
Overview[edit]
Testing is a set of activities conducted to facilitate discovery and/or evaluation of properties of one or
more items under test.[3] Each individual test, known as a test case, exercises a set of predefined test
activities, developed to drive the execution of the test item to meet test objectives; including correct
implementation, error identification, quality verification and other valued detail.[3] The test
environment is usually designed to be identical, or as close as possible, to the anticipated production
environment. It includes all facilities, hardware, software, firmware, procedures and/or
documentation intended for or used to perform the testing of software.[3]
UAT and OAT test cases are ideally derived in collaboration with business customers, business
analysts, testers, and developers. It's essential that these tests include both business logic tests as
well as operational environment conditions. The business customers (product owners) are the
primary stakeholders of these tests. As the test conditions successfully achieve their acceptance
criteria, the stakeholders are reassured the development is progressing in the right direction.[4]
User acceptance test (UAT) criteria (in agile software development) are usually created by
business customers and expressed in a business domain language. These are high-level tests
to verify the completeness of a user story or stories 'played' during any sprint/iteration.
Operational acceptance test (OAT) criteria (regardless if using agile, iterative or sequential
development) are defined in terms of functional and non-functional requirements; covering key
quality attributes of functional stability, portability and reliability.
Process[edit]
The acceptance test suite may need to be performed multiple times, as all of the test cases may not
be executed within a single test iteration.[5]
The acceptance test suite is run using predefined acceptance test procedures to direct the testers
which data to use, the step-by-step processes to follow and the expected result following execution.
The actual results are retained for comparison with the expected results.[5] If the actual results match
the expected results for each test case, the test case is said to pass. If the quantity of non-passing
test cases does not breach the project's predetermined threshold, the test suite is said to pass. If it
does, the system may either be rejected or accepted on conditions previously agreed between the
sponsor and the manufacturer.
The anticipated result of a successful test execution:
Evaluation is a process of judging how well the system’s original intended goals
have been achieved.
Evaluation happens after the solution has been developed, and its users have
used it long enough to become familiar with it and can use it effectively.
Evaluation usually is carried out the by the solution’s owners and users.
CRITERIA
Testing and evaluation can use the same criteria, which are the qualities or
attributes you want to test or evaluate.
Sample effectiveness criteria:
accuracy
robustness
security
enjoyability
completeness
readability
attractiveness
clarity
accessibility (for those with disabilities)
timeliness (being available when it’s needed)
communication of message
relevance
usability
ease of use
cost (purchase price
speed
the amount of labour required to make it work
METHODS
Evaluation and testing methods are the actions taken to measure the chosen
criteria.
Manual recalculation
Penetration testing of networks
Stress testing – deliberately working the solution hard to see if it fails
under pressure
Testing that validation rules work properly
User acceptance testing by a typical user
Using typical, unusual and invalid test data to test the solution’s behaviour
Integration testing – to see if components of the solution interact properly
performing all available user actions to check that the system responds
appropriately
Interviewing users to get their opinions about the solution’s ease of use
Questionnaires
Surveys
Analysis of records and data to find differences between the old system
and the new one
Observation of the system in actual use
With software - testing would mean to ensure that the program doesn't give
any errors by looking at the internals. Evaluation occurs post-production and
deals with customers to make sure that software you develop complies with
their requirements.
Example – a company wants to increase their sales. They decide to create a fun
website to attract young people who may buy their products.
They create the website and test all of its links, menus, readability, relevance,
usability, accessibility.
They test every page in the major browsers and every test is 100% successful.
They put the website online and wait for the money to pour in.
They want to evaluate whether the site is successful in achieving their aims of
popularity and profit.
What criteria should they evaluate? Popularity with young people. Increased
profit because of the site.
So they look at the site statistics: they find large numbers of young visitors.
They find zero profit.
They realise that none of these young visitors have credit cards or Paypal.
So testing showed that the site was fully functional and 100% accurate.
The evaluation showed it was a complete failure.
Definition: Test and evaluation is the set of practices and processes used to determine if the product
under examination meets the design, if the design correctly reflects the functional requirements, and if the
product performance satisfies the usability needs of personnel in the field.
Keywords: acceptance test, integration test, operational test, peer reviews, system test
MITRE SE Roles & Expectations: MITRE systems engineers (SEs) are expected to be familiar with
different kinds of tests and know which group conducts the tests, how to evaluate test documents, and
what the developer's procedures are for test control. MITRE SEs are also expected to analyze test data.
Background
Testing is the way a product, system, or capability under development is evaluated for correctness and
robustness and is proved to meet the stated requirements. Testing is done at each stage of development
and has characteristics unique to the level of test being performed. At a macro level, testing can be
divided into developer testing conducted before the system undergoes configuration management and
testing conducted after the system undergoes configuration management. Testing done before
configuration management includes peer reviews (sometimes called human testing) and unit tests. Testing
done after configuration management includes integration test, system test, acceptance test, and
operational test. Government testing agencies normally conduct an operational test. The developer
conducts other tests; in some cases, such as an acceptance test, government observers are present.
The government program management office (PMO) is tasked with assessing the developer's test and
evaluation plans and procedures. Often MITRE plays a central role in helping the PMO perform this
assessment. The requirements on which the developer's test plans and procedures are based must be well
crafted. A valid requirement is one that is measurable and testable. If it is not measurable and testable, it
is a poor requirement. Developer test plans and procedures should be based on the functional
requirements, not on the software design. Both the test community within the developer organization and
the development community should base their products on the functional requirements.
Assessing the developer's test plans and procedures should focus on the purpose of the test—that is, to
assess the correctness and robustness of the product, system, or service. The tests should first prove that
the product can do what it is intended to do and second that it can withstand anomalous conditions that
may arise. This second point requires particular care because there are huge differences in how robustness
is validated in a COTS-based system versus software developed for a real-time embedded system. The
environment in many COTS-based business systems can be tightly bound. A name or address field can be
limited in terms of acceptable characters and field length. In a real-time embedded system, you know
what the software expects to receive if all is going as it should, but you do not always know what possible
input data might actually arrive, which can vary in terms of data type, data rate, and so on. Denial-of-
service attacks often try to overwhelm a system with data, and the developer's skill in building into the
system a robustness that allows it to handle data it is not intended to process has a great deal to do with
the eventual reliability and availability of the delivered product. It is not unusual for the error protection
logic in complex government systems to be as large as, or larger than, the operational software.
Assessment of the test plans and procedures must take all of these issues into account. The assessor must
understand the nature and purpose of the system and the kind of software involved, and the assessor must
have the experience to examine the test plans and procedures to ensure they do an appropriate job of
verifying that the software functions as intended. The assessor must also verify that, when faced with
anomalous data conditions, the software will respond and deal with the situation without crashing. The
test conditions in the test plans and procedures should present a wide variety of data conditions and record
the responses.
For software systems, especially real-time systems, it is impossible to test all possible paths through the
software, but it should be possible to test all independent paths to ensure that the tests exercise all
segments of the software. There are software tools to facilitate this, such as the McCabe suite that will
identify paths as well as the test conditions needed to put into a test case. However it is accomplished, this
level of rigor is necessary to ensure the requisite reliability has been built into the software.
Unlike the unit test, the integration test plans and procedures focus on the interfaces between program
elements. These tests must verify that the data being passed between program elements will allow the
elements to function as intended, while also ensuring that anomalous data conditions are dealt with at
their entry point and not passed to other programs within the system. The assessor must pay particular
attention to this when assessing the integration test plans and procedures. These tests must be driven by
the functional requirements because those drive what the software must do in order for the sponsor to
accept the system.
The two primary test practices conducted prior to configuration management are:
Peer Reviews: Peer reviews are performed to find as many errors as possible in the software
before the product enters the integration test. Peer reviews are one of the key performance
activities at Level 3 of the Software Engineering Institute's (SEI) Capability Maturity Model. The
SEI accepts two kinds of peer reviews:
o Software inspections (SEI preference)—have a well-defined process understood
throughout the industry. Done properly, they can remove as much as 87 percent of the
life-cycle errors in software. (They are sometimes called Fagan Inspections after their
developer, Mike Fagan.)
o Code walkthroughs—have no standard process. They can have widely differing levels of
rigor and effectiveness and at best will remove about 60 percent of the errors in software.
Unit Test: The developer conducts the unit test, typically on the individual modules under
development. The unit test often requires the use of drivers and stubs because other modules,
which are the source of input data or receive the output of the module being tested, are not ready
for test.
Testing conducted after the product is placed under developer configuration control includes all testing
beyond unit test. Once the system is under configuration management, a problem discovered during
testing is recorded as a trouble report. This testing phase becomes progressively more expensive because
it involves integrating more and more modules and functional units as they become available; the system
therefore becomes increasingly more complex. Each test requires a documented test plan and procedure,
and each problem encountered is recorded on a trouble report. Each proposed fix must be validated
against the test procedure during which it was discovered and must also verify that the code inserted to
correct the problem does not cause another problem elsewhere. With each change made to respond to a
problem, the associated documentation must be upgraded, the fix must be documented as part of the
configuration management process, and the fix must be included in the next system build so that testing is
not conducted with patches. The longer it takes to find a problem, the more rework is likely, and the more
impact the fix may have on other system modules; therefore, the expense can continue to increase. Thus
performing good peer reviews and unit tests is very important.
Integration Test: This is a developer test that is successively more complex. It begins by
integrating the component parts, which are either the modules that have completed the unit test or
COTS products, to form functional elements. The integration test progresses from integration of
modules to form entire functional elements, to integration between functional elements, to
software-hardware integration testing. Modeling and simulation are often used to provide an
operational-like testing environment. An integration test is driven by an integration test plan and a
set of integration test procedures. Typically an integration test will have embedded within it a
subset of tests identified as regression tests, which are conducted following a system build. Their
objective is to verify that the build process did not create a serious problem that would prevent
the system from being properly tested. Often regression tests can be automated.
Test Data Analysis: When conducting peer reviews, unit tests, integration testing, and system
tests, a significant amount of data is collected and metric analysis is conducted to show the
condition state of the system. Significant metric data is produced related to defect density, pass-
fail data on test procedures, error trend analysis, etc. MITRE SEs should be familiar with test
metrics and evaluate the test results to determine the likelihood that the system can meet the
requirements of performance delivered on time and within budget.
System Test: This is an operational-like test of the entire system being developed. Following a
successful system test, a determination is made as to whether the system is ready for acceptance
test. After the completed system test and before the acceptance test, a test readiness review (TRR)
may be conducted to assess the readiness of the system to enter the acceptance test.
Acceptance Test: Witnessed by the government, this is the last test before the government
formally accepts the system. Similar to the system test, the acceptance test is often a subset of the
procedures run during system test.
Operational Test: Performed by an operational unit of the government, this is the final test
before the system is declared ready for general distribution to the field.
Review the peer review and unit test results prior to the start of integration testing. Due to the
expense and time needed to correct problems discovered in the post-configuration management tests, SEs
should understand how thorough the prior tests were and whether there is a hint of any issues that need to
be addressed before the integration test starts. If peer reviews and unit tests are done properly, the error
density trend data during the integration test should show an error density of 0.2 to 1.2 defects per 1,000
source lines of code.
Consider modeling and simulation options. These could support or substitute for some aspects of
integration that are either of lower risk or extremely expensive or complex to perform with the actual
system.
Complete a thorough independent review of the test results to date prior to supporting the
TRR. This is especially true for performance or design areas deemed to be of the greatest risk during the
design phase. After the TRR is passed and the program enters acceptance testing, correcting problems is
extremely expensive and time consuming.
Involve the government Responsible Test Organization (RTO) early. Involve the RTO (during the
concept development phase is not too early) so they understand the programmatic and technical issues on
the program. Including the RTO as part of the team with the acquisition and engineering organizations
will lessen conflicts between the acquisition organization and RTO due to lack of communication and
misunderstanding of objectives.
Diagram 2.1.1
.
Disputes are inevitable in construction projects. Skills in dispute resolution should be part of the resources of any
professional in a managerial position. There are various dispute resolution mechanisms and procedures such as litigation,
negotiation, meditation, arbitration and adjudication. With consideration to the main features and characteristics it can be
determined which dispute resolution strategy should be adopted after deliberation of the nature of the dispute.
2.3.1 Negotiation
Negotiation has been defined as “the process we use to satisfy our needs when someone else controls what we want". It is
the most common form of dispute resolution. Brown and Marriot (1993). It forms an important part of everyday life not
only for lawyers but also for all people. It is based on informal discussions, speedy and economical way or resolving
disputes that are not complex in nature. Due to its simplicity, negotiation is often used by the parties as a 1st option before
considering alternative means of dispute resolution or more formal strategies.
Negotiation is also a form of ADR. For negotiation to be successful it is important for both parties to be willing to reach an
amicable agreement, which both parties feel is acceptable and once the process is complete the parties should feel that the
outcome has been beneficial. Mackie (1991).
Bingham (online) illustrates that negotiation is voluntary in the sense that negotiations can be broken off at any time by
simply refusing to continue, and no one can coerce a reluctant party to negotiate. This makes negotiation non-
determinative, as there is no guarantee that it will end with a mutual agreement that resolves the matter at hand.
There are many great benefits of negotiation, from the parties point of view, are firstly that the parties can do it on their
own, thus minimising costs, and secondly they retain control of how the dispute is resolved.
There may be different reasons for entering into negotiation. One may be due to the need to resolve a problematic area
which is accepted by the other party with a degree of understanding, which can then be resolved; this allows the parties to
retain mutual trust. There is also the opportunity for negotiations, which may be based on financial gain, the change of
specification or a change of the contract. Brown and Marriot (1993, pg 89) believe that “the problem solving approach to
negotiation is very helpful to ADR practice".
Before the negotiation process commences each party should know where their strengths and weaknesses lie. Once this
has been identified the others parties position should be evaluated and both arguments compared. This will determine how
the negotiation will proceed, if both parties have the same desire to settle and what the final outcome will be.
Both parties will leave the negotiation with a particular sense of the outcome of the event whether it is positive or
negative. The timing of the negotiation in relation to the project can also be significant in the future relations between the
parties.
It is important that both parties in the negotiation process have all relevant information available to refer to at negotiation
meetings. During negotiation both parties should be clear on their stance with relation to the dispute, stating it clearly and
concisely. This should include explaining the situation and how it arose as well as its affect on the party. The other party
should then provide a response.
Once this is complete the area of contention should be discussed by both parties and any areas of agreement identified. In
the event that neither party is able to reach an acceptable agreement, it is important that the reason for the failure to
negotiate is assessed, in order to pave the way to an amicable solution which satisfies both parties’ requirements.
For a negotiation to be successful the representatives of both parties should have a detailed understanding of the project
and allow for sufficient preparation of the case. Also positions of the party should be made clear and reasonable, with the
aim of ensuring mutual respect. Trust is also necessary for the process to succeed, with a willingness to negotiate, each
party should be able to speak with experience and know the intricate details of the dispute. Brown and Marriot (1993).
Mediation
Mediation is a structured form of ADR, with one respective from each parties side meeting with a mediator to resolve the
dispute. Mackie (1991, pg 87) defines mediation as the “intervention by a third party to assist disputing parties to reach a
settlement".
It is a voluntary, non-binding process in which a neutral party, known as a mediator, helps to guide the parties towards a
mutually beneficial resolution. The mediator plays a facultative role (see diagram below) in the resolution process by
assisting the parties to decide for themselves whether to settle and on what terms. A mediator does not impose a decision
or attempt to judge the case. The mediator’s skills help the parties to move away from trying to demonstrate that they are
‘right’ instead the emphasis is placed on the resolution of the problem. The mediator helps the parties to achieve a
settlement by focusing on what is to be done rather that pointing the blame. Bingham (online).
Mediation gives parties involved much more control over the way their dispute or difference of opinion is dealt with and
over the outcome. If negotiation as previously mentioned has failed, mediation provides alternative to pursuing litigation
etc. The scope for solutions is usually greater than the solutions available in courts and tribunals, or even prolonged
negotiation.
Adjudication
Similarly to ADR, Adjudication is a process of decision making which involves a neutral third party. The neutral has the
authority to determine a binding resolution through some form of judgement or award. Adjudication is commonly carried
out in the form of the court system. Brown and Marriot (1993). Although binding, but not necessarily final outcome to
disputes on construction contracts. It is mainly for disputes that will otherwise hold up cash flow, by obtaining a quick
decision by an impartial third party while the contract is in progress.
Court based Adjudication is found to be significantly more formal than arbitration and the ADR process which may be
used in the settlement of a dispute. Adjudication is not a voluntary process; each party presents their case which results in
a final win lose outcome. Once the Adjudication process commences both parties are obligated by law to participate in the
proceedings. If the case does go to trial each side must present evidence to support their argument and cause for dispute.
Once the evidence of each party has been presented, a judge or jury will then make their decision based on the cases
presented from both parties. Once a decision has been made the losing party has the right of appeal which may be filed and
taken to a higher court. Mackie (1991).
Construction generally offers its clients great flexibility. Changes are inevitable therefore there will be changes in price
and programme. This makes for disputes and if it is not efficiently managed it reduces itself to conflict. All other aspects
in the project will suffer. The adjudicator will quickly identify dodgy claims and the project will not be damaged with poor
relations.
Adjudication offers many advantages including the following:-
Cost Efficient - The parties split the cost of a highly credentialed Adjudicator who efficiently decides the dispute after
zealously researching the law and facts with the assistance of an expert legal staff. Intelligent Justice informs the parties of
the total cost at the very beginning of the process and payment is made through capped monthly installments.
Time reduced - Adjudication will commence immediately with an analysis of the applicable legal authorities and
identification of the evidence that supports the positions of each party to the dispute. The Adjudicator will promptly
examine material witnesses under oath in the locations where they are found. There is no need to repeat the process years
later in an expensive trial.
Adjudicators Experience –The parties jointly approve the selection of their Adjudicator in advance, Intelligent Justice
eliminates the risk and uncertainty that occurs when judges and jurors with unknown qualifications, backgrounds, biases
and abilities are used to decide complex legal disputes in the public court system.
Enforceable - When the adjudication is completed the verdict may be quickly confirmed as an enforceable court judgment
or appealed as permitted by the parties' agreement and applicable law. Brown and Marriot (1993).
Arbitration
Arbitration is a procedure for the settlement of disputes under which the resultant disputants agree to be bound by the
decision of an arbitrator who decision is final and enforced by the law.
Arbitration is a voluntary, binding alternative to court litigation. It is a consensual process based on an agreement between
parties to submit their disputes to an independent person or tribunal for final verdict of the disagreement between the two
parties.
MacRoberts, (2008) states “An arbitration has been defined as the method of procedure by which parties who are in
dispute with each other agree to submit their dispute to the decision of one or more persons, traditionally in Scotland
described as ‘arbiters’, rather than resort to the courts of law."
In Scotland arbitration has played a pivotal role in settling disputes and as a result Scotland has its own highly regarded
and developed body of arbitration law. The Scottish courts have always recognised the right of parties to agree to exclude
the jurisdiction of the courts to inquire into the merits of their disputes and instead to refer any disputes to arbitration.
Traditionally, the means of conflict resolution have been very simple. Construction disputes were resolved by arbitration.
Arbitration could have been regarded as the primary provisional response when in need of a conflict or dispute resolution.
The main purpose of arbitration was to avoid the flaws that litigation along with the court based proceedings had.
Arbitration became popular widely but in recent times has fell into the same bracket as litigation. Brooker & Lavers
(1997). According to Flood and Caiger (1993) arbitration is likely to lead to delays and high costs. The imitation of legal
procedures According to Flood and Caiger (1993) is described as the “jurisdication of arbitration".
According to Brown and Marriot (1993), there are many differences between Arbitration and the “alternatives". Firstly
mediation and negotiation is are both consensual and rest on agreement, similar to arbitration, however arbitration is more
legalistic and enforced by the courts whereas agreements in the ADR process will not be.
Strategic Alliances
Alliances In Construction: An Introduction
In the construction industry various forms of alliances are now being offered such as a series of major long term
construction commitments and opportunities. There are various forms of strategic alliances such as partnering and joint
ventures which have allowed separate companies to co-operate in developing new projects. The structure for such schemes
is simple its is wholly based on a risk/award structure i.e. each company sharing each other’s “pain and gain" during the
course of the project. Ngowi and Pienaar (2005). Ingirige and Sexton (2006) point out that alliances on the whole provide
opportunities for individuals and groups to gain mutual benefits such as reducing uncertainties, the product of such benefit
is now providing the concept of alliancing in the construction industry an increased momentum.
Harris and McCaffer (2001) concur with this idea stating that the construction and design team prepare targets and
objectives united to a structure which includes the risks and rewards based on the final outcome. The suggestion being that
further collaborations will be readily available if the final outcome meets the client’s approval.
According to Ngowi and Pienaar (2005) there are many reasons as to why firms form an alliance. Not only do they
provide pulling together the costs and risks they can also “trim product and service development times, penetrate new
markets, and gain access to new technologies and economics of scale." Ngowi and Pienaar (2005, p 268).
Alliances are loosely based on the important factor of trust, something which in practice should alleviate common
conflicts. Coleman (1990, p 78) illustrates trust as “committing to an exchange before you know how the other person will
reciprocate". The foundations of trust being confidence and dependence. Sabel (1993) as cited in Ngowi and Pienaar
(2005, p 268) puts it more pithily believing it be defined as “the mutual confidence that no party to an exchange will
exploit another’s vulnerabilities".
Ultimately the main characteristics of alliances are the combination of cooperation and competition. This combination of
joint reliance between the partners with elements of conflict can add risk and caution to the cooperation. This can be
overcome by the primary factor of trust as illustrated above by Coleman (1990) and Sabel (1993). The trust between
partners helps overcome this threat.
Harris and McCaffer (2004) makes it clear that originally partnering is primarily regarded as the standard form of strategic
alliancing. This form of construction alliance usually provides a long-term commitment between partners to work together
on a series of projects or developments. Fryer(2006) makes it clear that partnering as a process is fully and completely
aimed at prevention prior to a dispute evolving and one of the benefits attributable to partnering is reduced exposure to
litigation through issues being resolved through strategies and communication.
Partnering
For a number of years now partnering has become one of the most practical concepts in dispute avoidance. Fryer (2004)
illustrates that the increased promotion of partnering between the design and construction teams in the industry has
displayed effective tactics for profitability and cost effectiveness.
Egan’s report “Rethinking Construction" in 1998 contained promising developments in Section 12 of the report which
demonstrated “tools" to tackle fragmentation between parties. This was the idea of partnering and framework agreements
which at the time were picking up increased popularity by firms as a substitute for traditional contract based procurement
and project management. The main aspect for this was the opportunity for parties to share in the rewards of improved
performance. The report’s definition of partnering is as follows:
“Partnering involves two or more organisations working together to improve performance through agreeing mutual
objectives, devising a way for resolving any disputes and commiting themselves to continous improvement, measuring
progress and sharing gains".
The aspect of a partnering approach provides a harmonious agreement from the outset. The structure of which purely
focuses on co-operation and teamwork to avoid adversarial confrontation. Fryer (2004) supports this view further by
providing the view that partnering is increasingly becoming acknowledged as a way of addressing the industry’s
fragmentation and confrontation. Harris and McCaffer (2001) believes the way partnering reduces disputes is based on the
fact that problems can be foreseen and mutually resolved i.e disputes and claims. Fryer (2009) concludes that there are
three core principles providing the foundation for the partnering relationship: good communications, commitments and
conflict avoidance and resolution.
Gardiner (2005, p 152) states that “fostering of a mutual interdependence and trust rather than a blame culture" is
fundamentally the concept of the strategy. It is also a methodology for avoiding and resolving disputes, mitigation of
conflict and avoiding litigation.
Framework Agreements
A framework agreement is a general document which suits long-term collaborations between two parties or organisations.
Framework agreements are loosely based on an “umbrella" agreement in which the parties enter into a number of specific
contracts. Harris and McCaffer (2001).
Barr construction currently have numerous framework agreements in place in the retail industry with Tesco stores. The
agreement itself has according to Barr (Online) has resulted in over forty project with the retailer. The main reason
companies such as Barr will enter into framework agreements is to exchange innovations, reduce tendering from project to
project and ultimately improve costs. Fryer (2004).
The framework agreement established and was recommended in Sir John Egan’s report “Re-thinking Construction" and
“Accelerating Change". The JCT contract also contains a JCT 2005 Framework Agreement whereby it is understood that
through the encouragement of collaboration, the main objectives include team working and consideration of others and
also the avoidance of disputes. JCT (2005) acknowledges that “frameworking arrangements are really only likely to pay
dividends on larger, lengthier projects which give the project participants the opportunity and incentive to invest in people,
processes and products and develop as a team".
Klimt (2005) explains that in 2005 the Joint Contracts Tribunal published two versions of the Framework Agreement,
binding and non-binding. These framework agreements are based around the concept of replacing blame with cooperation.
The main purpose being providing “best practice". He explains that traditional contracts in the past have been poor and
impractical in dealing with problems and therefore they have festered unresolved and become more complex. The
Framework Agreement has been published so that identification of problems is early on and resolved. Any disputes or
conflict is recorded on a form. If the problem cannot be resolved it is passed up to the next hierarchical level. If the dispute
is still not resolved then formal dispute resolution procedures will be implemented as outlined in the Project Specific
Order.
Conclusion
This literature review has provided an extensive analysis into the main methods of dispute resolution and avoidance which
are currently available in the construction industry. This conflict resolution procedure is useful for both researchers and
practitioners to better deal with the dispute-prone nature of the construction industry.It would not be a mistake to conclude
that disputes are inevitable in any project and whether it should be resolved at the time through ADR or avoided
completely through Strategic alliances is determined by the scope of the project itself. In recent times negotiation,
mediation, arbitration and adjudication have provided professionals with a procedure which could prevent individuals
from the more formal hearings of litigation.
The literature review also examined the use of strategic alliances in today’s industry. In a construction alliance, there are
cooperation aspects between the parties involved. The study intended to determine the applicability of such alliances in
preventing disputes and conflict through the foundations of long-term commitments and collaborations. Partnering has
long been a proven method of avoiding conflict through the basis of mutual trust and the idea of sharing each other’s “pain
and gain" through the lifespan of their partnership. The author has also studied the use of framework agreements in the
construction industry. It is evident that framework agreements providing yet-again long term collaborations between
parties. The use of framework agreements in retail is immensely popular in today’s industry due to the concept being
loosely based on providing “best practice". Such purposes have allowed framework agreements to replace blame and the
natural “pointing the finger" notion with cooperation and understanding.
Adjudication is a simple and inexpensive way of resolving disputes between two parties, which appoints a neutral third party to
investigate the issues and make a decision. We have led the industry move towards adjudication, rather than more costly forms of
litigation, working with other institutions and bodies to set criteria to ensure high standards of competency in adjudicators.
We will identify and appoint the appropriate adjudicator for the particular case, for which we charge a fee of £300.
Arbitration is a more complex procedure, preferred by some for dispute resolution. The Arbitration Act 1996 aimed to address
some of the weakness in this method, giving the arbitrator new powers to speed up the process and save costs.
For a fee of £300, we will appoint an arbitrator who will have the qualifications and experience to arbitrate the matter at hand.
Share on Facebook
Tweet on Twitter
Negotiation: This is the least costly, and most informal, method of dispute resolution,
allowing the outcome to be agreed upon by the contracting parties themselves. A
lawyer is not required, but it does necessitate both parties entering into discussions
willingly.
How the aggrieved party will provide notice to the other party about the nature
of the dispute.
How the parties will commence dispute resolution, such as a formal meeting to
discuss the disagreement and attempt to negotiate a resolution.
How the parties will continue dispute resolution, if required. The clause should
outline who will act as the mediator or arbitrator, any guidelines as to how
mediation or arbitration will be conducted, and how any costs will be covered.
How to determine when the alternative dispute resolution has come to an end,
and formal litigation is needed to begin.
What jurisdiction any such litigation should be commenced in.
The Ajuve arbitration clause is completely free until you have a dispute. Just
copy and paste the following into all your contracts:
This agreement shall be governed by and will be construed in accordance with English
Law. The Parties agree to refer all disputes and differences of any kind (including any
dispute regarding the existence of this agreement) to the Ajuve Dispute Resolution
Rules which shall be read and construed as if set out in full herein. A full copy of those
rules is available at http://ajuve.com/dispute-resolution-rules.
Your LinkedIn
Connections at Firm
Making the decision to take your New Hampshire-based business international is a big step. You’ve found
the right partner, and now you are ready to embark on your international venture. But the long term health
of your international business relationships may depend on how well you plan ahead for trouble. While
deal-making may be the initial priority, providing for dispute resolution when trouble occurs should be just
as important.
Don’t overlook the simple things. Make sure that your international business relationships are
documented with a carefully drawn contract drafted with the help of an experienced lawyer. Moreover, all
international commercial contracts should, as a matter of course, have a dispute resolution clause that
includes an effective agreement to arbitrate.
Why Arbitrate?
Arbitration is the most widely used method of alternative dispute resolution. It is similar in process to a
court, but is private and generally less formal. Like litigation in courts, it is expensive, time consuming,
and distracting. But for most international disputes, there is no satisfactory substitute. There are two
major reasons for this. First, agreeing to arbitrate assures that your dispute will be heard in a neutral
forum, instead of the other party’s home court, where you may face a biased judge and an unfamiliar
language. Second, with arbitration you have a far better chance at having a judgment or award enforced.
Enforcing the judgment of a court (even a U.S. court) in another country isn’t a sure thing. International
arbitration is different. Under a multinational treaty known as the “New York Convention,” 142 signatory
countries have agreed to recognize and enforce foreign arbitration awards. There is no comparable treaty
governing the judgments of U.S. or foreign courts. Therefore, it is almost always to your advantage to
agree to arbitration when you are contracting with an overseas party, provided that each party is located
in a country, or has assets in a country, which is a signatory to the New York Convention. If you are
dealing with a party who is not located in a country that has signed the New York Convention, you must
proceed with extreme caution and take additional measures to protect yourself. Consult an experienced
attorney in such a case.
When beginning a commercial relationship, dispute resolution is often the last thing on everyone’s minds.
A principal decision maker might treat the arbitration clause as “boilerplate” and never give it a second
thought. But you ignore it at your peril. You must craft the arbitration clause to fit your business objectives
and accommodate the unique circumstances of your particular deal. Poorly drafted arbitration clauses
pose many dangers, and clever defendants can exploit an imprecise arbitration clause to delay or
completely avoid arbitration.
With that in mind, there are several essentials of any well-drafted arbitration agreement. These essentials
include the scope of the dispute, exclusivity, number of arbitrators, language of the arbitration (English, if
at all possible) and applicable substantive law, among others. In addition, every arbitration clause should
contain a provision referring to and incorporating the rules of a specific arbitration organization to govern
the arbitration. There are perhaps a half dozen pre-eminent arbitration organizations in the world. They
include the International Chamber of Commerce (“ICC”), the London Court of International Arbitration
(“LCIA”), the International Centre for Dispute Resolution or “ICDR”, the Stockholm Chamber of
Commerce (“SCC”) or the Singapore International Arbitration Centre (“SIAC”), among others. All of these
organizations are reputable, but experienced lawyers often recommend using the ICC as the arbitration
administrator in international disputes. It is generally believed that the ICC, although expensive, is the
most widely recognized organization and has the best panel of arbitrators to choose from. In some
situations, you may want to use an organization like the LCIA or the ICDR, which charge a fixed fee for
the arbitration versus a percentage of the claim, like the ICC. It is possible to draft an “ad hoc” arbitration
clause, which calls for the arbitration to he held privately between the parties, without the rules or
administrative oversight of an arbitral organization, but this is generally not recommended.
One of the most important decisions to make is designating the place of arbitration. Generally speaking,
the parties try to pick a place that is neutral. You should take care to choose a country that has a well-
developed body of arbitration law and a modern arbitration statute. The most reliable (and popular)
places of arbitration are London, Paris, Geneva, Zurich, Stockholm, the Hague, New York, Los Angeles,
Toronto, Rome, Brussels, Singapore, and Bermuda.
There are other provisions you may consider other than the essentials listed above. In deciding whether
to include such provisions, you and your lawyer should give a great deal of thought to whether such
provisions are necessary and/or whether they may provide strategic advantage in the context of your
deal. Examples of such other provisions include designating special qualifications of the arbitrators,
nationality of the arbitrators, special discovery procedures, limitations on types of damages, deadlines
and time limits, special summary disposition procedures, shifting costs and expenses to the losing party,
and assessing costs to collect a judgment.
Conclusion
In every international commercial contract a carefully crafted arbitration clause is a necessary part of any
process for resolving international disputes. Putting an effective dispute resolution process in place at the
beginning of the contract will avoid problems later, if the unthinkable happens.
Michael Tule is of counsel to the Corporate Department at the law firm of McLane, Graf, Raulerson &
Middleton, Professional Association.
The content of this article is intended to provide a general guide to the subject matter. Specialist advice
should be sought about your specific circumstances.
What is Assurance?
Assurance is provided by organization management, it
means giving a positive declaration on a product which
obtains confidence for the outcome. It gives a security that
the product will work without any glitches as per the
expectations or requests.
All team members are responsible for Testing team is responsible for QC.
QA.
QA means Planning for doing a process. QC Means Action for executing the
planned process.
QA makes sure you are doing the right QC makes sure the results of what
things. you’ve done are what you expected.
Key Points
In QA, processes are planned to evade the defects.
QC agreements with discovery the defects and
modifying them while making the product.
QA detects weakness.
QC detects defects.
QA is process oriented
QC is product oriented.
QA is failure prevention system.
QC is failure detection system.
Effective quality systems can contribute enormously to the success of projects, but the counterpoint is
that, when poorly understood, the quality systems are likely to be weak and ineffective in ensuring that
the delivered system is delivered on time, built by the team within their allocated budget, and satisfies the
customer’s requirements.
This article considers the difference between Quality Assurance and Quality Control. The concepts are
investigated by looking at guidance from key industry players.
Introduction
How many times has it struck you that many practitioners involved in the ICT field lack an understanding
of the difference between Quality Assurance and Quality Control? Often you will hear someone talk about
‘QA’, when what they actually mean is ‘QC’.
This ambiguity consistently throws up problems and is a sure way of undermining a project. Projects are
negatively affected as it tends to lead to strained conversations and makes reaching consensus difficult.
Although QA and QC are closely related concepts, and are both aspects of quality management, they are
fundamentally different in their focus:
Achieving success in a project requires both QA and QC. If we only apply QA, then we have a set of
processes that can be applied to ensure great quality in our delivered solution, but the delivered solution
itself is never actually quality-checked.
Likewise, if we only focus on QC then we are simply conducting tests without any clear vision for making
our tests repeatable, for understanding and eliminating problems in testing, and for generally driving
improvement into the means we use to deliver our ICT solutions.
In either case, the delivered solution is unlikely to meet the customer expectation or satisfy the business
needs that gave rise to the project in the first place.
A good point of reference for understanding the difference is the ISO 9000 family of standards. These
standards relate to quality management systems and are designed to help organisations meet the needs
of customers and other stakeholders.
In terms of this standard, a quality management system is comprised of quality planning and quality
improvement activities, the establishment of a set of quality policies and objectives that will act as
guidelines within an organisation, and QA and QC.
In the ISO 9000 standard, clause 3.2.10 defines Quality Control as:
“A part of quality management focused on fulfilling quality requirements”
These definitions lay a good foundation, but they are too broad and vague to be useful. NASA, one of the
most rigorous software engineering firms in the world, provides the following definitions
(www.hq.nasa.gov/office/codeq/software/umbrella_defs.htm):
Simply put, Quality Assurance focuses on the process of quality, while Quality Control focuses on the
quality of output.
The purpose of QA is to prevent defects from entering into the solution in the first place. In other
words, QA is a pro-active management practice that is used to assure a stated level of quality for an IT
initiative.
Undertaking QA at the beginning of a project is a key tool to mitigate the risks that have been identified
during the specification phases. Communication plays a pivotal role in managing project risk, and is
crucial for realising effective QA. Part of any risk mitigation strategy is the clear communication of both the
risks, and their associated remedies to the team or teams involved in the project.
QC involves verification of output conformance to desired quality levels. This means that the ICT solution
is checked against customer requirements, with various checks being conducted at planned points in the
development lifecycle. Teams will use, amongst other techniques, structured walkthroughs, testing and
code inspections to ensure that the solution meets the agreed set of requirements.
Quality Control is used, in conjunction with the quality improvement activity, to isolate and provide
feedback on the causes of quality problems. By using this approach consistently, across projects, the
feedback mechanism works towards identifying root-cause problems, and then developing strategies to
eliminating these problems. Using this holistic approach ensures that teams achieve ever higher levels of
quality.
As a consequence of formulating and executing a quality management plan the company can expect:
Greater levels of customer satisfaction, which will very likely result in both repeat
business, as well as referral business
A motivated team that not only understand the policy objectives of the quality
management plan, but who also actively participate in executing the plan
Elimination of waste by eliminating rework arising from either the need to address
bugs, or to address gaps in the solution’s ability to meet customer requirements
Higher levels of confidence in planning, since the tasks arising from unplanned
rework will fall away
Financial rewards for the company, which are a consequence of new projects
from existing and referral clients, as well as through the reduction of monies
spent on rework tasks.
As the company’s quality management plan matures, the confidence of all stakeholders will grow. The
company will be seen to be more effective and efficient in delivering an agreed ICT solution to clients.
Dialog Information Technology provides expertise in Quality Control and Quality Assurance and services
range from strategic IT consulting through full lifecycle application development and managed application
services to long-term operational support.
QA is QualityAssurance
but
QC is QualityControl
The difference is that QA is process oriented and QC is product oriented.
Testing, therefore is product oriented and thus is in the QC domain. Testing for quality
isn't assuring quality, it's controlling it.
Quality Assurance makes sure you are doing the right things, the right
way.
Quality Control makes sure the results of what you've done are what you
expected.
-- HillelGlazer
Please assure me that our QA team knows the right things and the right way. I'm
not even convinced they know up from down. And how did they find out the right
things and the right way for software engineering, when even AlistairCockburn is
still trying to figure it out? -- JohnFarrell
The QA team's job is to see that standards, processes, and policies (or other
governing/guiding "writ") are in place and carried out; to recommend and
implement improvements to them, and to ensure that the people that need to know
about them know about them. QA "audits" or "reviews" are intended to determine
the efficacy of these "writs."
It's often easier to understand QA by an example: In one place (a large company
with large projects) I've seen QA's role to help project managers plan their
projects -- so that the projects follow certain organizational procedures; so that
they include the required artifacts, events, and milestones; and so that the projects
know what is expected of them and when in terms of reports, reviews, and
documentation. As the project progressed, QA would conduct "checkpoints"
along the way to see where the project may be developing risks, for example
either by progressing beyond where they've got authorization to go, or where they
may have need to escalate issues to management. These checkpoints would also
be an opportunity to ensure that people who need to be involved are involved at
the right times. This approach to QA isn't one that really suits XP, but there are
easily ways to adapt such an approach to XP.
Maybe mixups like this are happening because ambiguous terms have been given
specific meanings (see: DeGeneralization).
Why would "assurance" mean process-related and "control" mean product-
related?
Because "assurance" means that you know you did everything needed to make
something that works right, while "control" means that you have no idea whether
any or all of your batch o'junk is worth anything until you examine each item.
"Just in time" production requires that you are assured your supplier is supplying
you with exactly what you need as you need it so you don't have to stop to test
each incoming item for defects. See ToyotaProductionSystem or DemingMethod.
(I plan to eventually write a book about how this revolution in manufacturing
could be applied to software.) In the case of software, QA means that you know
that the code you are making fulfills the requirements, QC is discovering that
your process of writing code was not adequate to assure that you didn't create a
lot of bugs along with your features. -- ChrisBaugh
One way to avoid confusion is to use more descriptive names: "Product-QA" and
"Process-QA".
I have always found the term "Quality Assurance" a curious choice. Why would
we want to be assured about quality (made to feel better about it) rather than have
quality ensured (guaranteed)?
Assuring quality is about confidence. It's about the *processes* by which we go
about doing what we do. Part of that is knowing that we're doing the right things
at the right time, and part of it is that we are doing them the right way. This all
should be done or known before we start working, not afterwards.
These terms have been defined for us. While it may be 'more right' to redefine
them so they're no longer 'degeneralized,' we ought to at least try to ensure
everyone has their terms straight when using them.
We may also look at the other areas in which these terms are used. 'Control' is
often a term used with statistics, and processes that can be "controlled" with
applying the analysis of the statistics. 'Control' is also an active term and has a
certain circumstantial connotation. The need to test ("control") is a function of the
circumstance of the existence of bugs or lack of confidence. Whereas 'assurance'
is a passive term with a more universal connotation. The need to "assure" is a
function of building confidence regardless of the existence of bugs and isn't just
about bugs, but of the entire effort.
The difference between QA and QC is largely one of power and control. QC is
usually as service provided to development (i.e. under the control of
development) and is responsible for providing that service. QA expects
development to provide services to it (control development) and is not
responsible for any result.
The above statement is a symptom of one of the biggest problems in the entire
software industry. People don't know what "QA" is supposed to do (as defined in
this page). As a result many that do QA, do it wrongly, and many that work with
those doing QA incorrectly wind up hating QA because of the lousy experience
they had with it.
QA, when done effectively in a project, should not even cross paths with
developers except perhaps at the highest levels of management. Or to teach new
developers the organizational processes, or teach existing developers about
changes to existing organizational processes. And then later QA may interact
with development to figure out whether the processes work and what needs to
happen to make sure they work better. QA should not be causing developers to do
any more work than they need to be doing to develop the work. In fact, most of
QA's role should occur with project managers and leaders of other disciplines
such as SCM, and QC. (In XP, the idea of SCM or a CCB, or CCA, or a QC
"group" doesn't apply in the same way, but there's still a QA role that can be
incorporated.)
Another unfortunate characteristic of poorly executed QA is that often in
organizations that don't know the first thing about QA the people assigned to QA
are rejects from other skills which simply compounds the problem. Again, I'm
talking about QA as described here... not QC.
QA ... should not even cross paths with development.
Anyone care to explain this one? What is the purpose of QA if it does not deal
with development? I'm afraid "QA as described here" leaves a lot to the
imagination.
An edit was made to the paragraph above that may clarify the intent. QA's job is
not to inspect code or run tests or test programmers' know-how. If QA is
"process" centric, QA shouldn't fit in with the work developers are doing. QA
folks should be "process" folks not "product" folks. As such, QA's job should
happen before and in the aftermath of product development, and less to do during
development. The purpose of this page is to separate QA from the concept of
"testing." A discussion of what QA (QualityAssurance) really is seems to be
brewing. Which is something this page is not a good place for. Without proper
definition, QA rightly so "leaves a lot to the imagination."
So, define it already. It is not enough to say QaIsNotQc, you need to also say
what QA is. Also, is there really some deep distinction between the two, or is this
just some nit-picking intended to preserve a QA role when none is needed?
I have to agree with the prior comment that the terms are very confusing. Even
though they are commonly used, I think we'd be much better off if they were not.
Part of the problem is that these terms were taken from manufacturing paradigms.
QA dealt with manufacturing processes, while QC inspected the manufacturing
output to verify that it was acceptable. In the manufacturing environment you
analyze, design, and implement once and then crank out many essentially
identical products. This is very different from the software environment where the
analysis, design, and implementation happen on each product, and then you start
over for the next iteration.
In a software environment, "Quality Assurance" does not assure quality. Rather,
they assure that a process is being followed, which is very different. The process,
if it's a good process, will increase the probability that there is a quality result. It
will, presumably, make sure that requirements are communicated, traced, planned
for, scheduled, etc. The "Quality Assurance" role will likely also collect metrics
that it can then use for process improvement. However, this group could do its job
and still have the organization's product be low-quality. Therefore this function
would be better termed "Process Management."
Similarly, "Quality Control" does not control quality. Rather they measure
quality, by verifying that what was implemented matches the requirements.
Certainly there's a feedback loop, as any defects will be reported and presumably
fixed, thus increasing quality. But this group doesn't really control anything. It
would thus be better termed "Requirements Verification," although I find the term
"Product Test" more concise.
Calling these functions Process Management and Product Test removes any
ambiguity or confusion. And calling them by these names does not impact your
ability to get ISO 9000 certification or achieve a CMM level. Those require
certain things to be done or documented, and require certain functions to be
executed by the organization, but absolutely do not require specific organizational
names.
-- JimSeidman
Well Said!!! -- HillelGlazer
I'm currently performing a quality assurance role on a software project. This has
been a first for me and I haven't spent a great deal of effort find out what quality
assurance is "supposed" to mean, because to me it is clear. I'm supposed to assure
the quality of the product and customer satisfaction. That doesn't mean reassuring
people that quality exists, it means stepping in wherever I see an opportunity to
add or ensure quality. For this project it has meant a lot of things - clarifying
requirements, documenting new requirements, facilitating communication within
the development team, and yes, testing. But not just testing software, testing
requirements, testing understanding of requirements, testing the schedule, etc.
If it is difficult to define what quality assurance is or should do, perhaps part of that is
because it likely will vary from project to project. If you spend a great deal of effort
trying to apply some "this is what I'm supposed to be doing" process to a project I'd
guess a lot of your effort is going to go into dotting i's and crossing t's that could have
better been spent contributing something useful to the project. Which is not to say you
shouldn't learn QA techniques or processes, only that QA requires the ability to
analyze and respond logically to the needs of the environment and situation in which
it is being performed.
Comparison chart
Quality Assurance versus Quality Control comparison chart
Focus on QA aims to prevent defects with a focus on QC aims to identify (and correct)
the process used to make the product. It is a defects in the finished product.
proactive quality process. Quality control, therefore, is a
reactive process.
Statistical Statistical Tools & Techniques can be When statistical tools & techniques
Techniques applied in both QA & QC. When they are are applied to finished products
applied to processes (process inputs & (process outputs), they are called as
operational parameters), they are called Statistical Quality Control (SQC) &
Statistical Process Control (SPC); & it comes under QC.
becomes the part of QA.
Quality Assurance (QA) refers to the process used to create the deliverables, and can
be performed by a manager, client, or even a third-party reviewer. Examples of quality
assurance include process checklists, project audits and methodology and standards
development.
Quality Control (QC) refers to quality related activities associated with the creation of
project deliverables. Quality control is used to verify that deliverables are of acceptable
quality and that they are complete and correct. Examples of quality control activities
include inspection, deliverable peer reviews and the testing process.
Quality control is about adherence to requirements. Quality assurance is generic and
does not concern the specific requirements of the product being developed.
Quality assurance activities are determined before production work begins and these
activities are performed while the product is being developed. In contrast, Quality
control activities are performed after the product is developed.
Sport can add value for the development of individuals, of organisations and of whole
communities irrespective of the level of development. Despite this broadly shared conviction,
there is still a lack of substantiated evidence to support the purported potential of sport.
Effective, transparent and (if possible) comparable M&E must therefore take place to further
determine the inherent benefits, risks and limitations of sport and physical activity.
This is critically important because while the global evidence base on the proportion of women
having ever experienced various forms of abuse is strong, evidence on what kinds of strategies are
effective in preventing such violence and offering adequate support to victims and survivors is still
weak. This is especially relevant in resource poor areas, where difficult decisions need to be made
with respect to funding priorities.
At the programme level, the purpose of monitoring and evaluation is to track implementation and
outputs systematically, and measure the effectiveness of programmes. It helps determine exactly
when a programme is on track and when changes may be needed. Monitoring and evaluation forms
the basis for modification of interventions and assessing the quality of activities being conducted.
Monitoring and evaluation can be used to demonstrate that programme efforts have had a
measurable impact on expected outcomes and have been implemented effectively. It is essential in
helping managers, planners, implementers, policy makers and donors acquire the information and
understanding they need to make informed decisions about programme operations.
Monitoring and evaluation helps with identifying the most valuable and efficient use of resources. It is
critical for developing objective conclusions regarding the extent to which programmes can be
judged a “success”. Monitoring and evaluation together provide the necessary data to guide strategic
planning, to design and implement programmes and projects, and to allocate, and re-allocate
resources in better ways.
(Adapted from Gage and Dunn 2009, Frankel and Gage 2007)
For initiatives addressing violence against women, monitoring and evaluation is more
than a costing or cost-effectiveness exercise. It is a way of ensuring women and girls are
able to live their lives free from violence and abuse.
What can be learned in general from monitoring and evaluation of initiatives on violence
against women?
What interventions and strategies are effective at preventing and responding to violence against
women and girls?
What services are needed to help women and girls recover from violence?
What could be the role of different sectors in addressing and preventing violence?
What other factors (social, economic, political, cultural etc.) play a role in perpetuating vulnerability to
violence or hindering access to services?
What kinds of investments produce more promising results/ how much do they cost? (Adapted from
Watts 2008)
Are the proposed activities being carried out in the manner outlined? Why/ why not?
What services are provided, to whom, when, how often, for how long, in what context?
Are services accessible? Is the quality adequate? Is the target population being reached?
If yes, what actual difference the intervention is making; how it is making this difference and for whom.
The extent to which the intervention is responsible for the measured or observed changes.
What are some important questions that an evaluation can help answer?
Can it be scaled up? That is, can the intervention be adapted, replicated or built on to increase its
reach or scope (for a larger population or a different region)?
If yes, how can it be scaled up? What aspects can be scaled up?