AIGP Note
AIGP Note
PROFESSIONAL - AIGP
Study Notes by Caiky Avellar (2025)
English Version - Translated with DeepL Pro
DOMAIN I
Understand the foundations of AI governance
EXAMPLE: In the same fruit example as above, in this case the system naturally
identifies similar patterns and characteristics in groups of fruit. For example, it
manages to representatively group together images of fruit that are elongated and
yellow (bananas) and, in another group, fruit that are rounded and red (apples).
▪ Decision tree: a model that resembles a flowchart in the form of a tree. It's
like a flowchart of questions. At each "node", the tree asks a question (for
example: "Is the salary higher than X?"). Depending on the answer (yes or no),
the flow goes to a new node until it reaches a sheet with the final decision.
▪ Random Forest: a set of several decision trees (a "forest"). Each tree asks
its own questions and, in the end, they all vote to arrive at the most stable
and accurate answer.
▪ SVM (Support Vector Machine): a method that looks for a "line" (or plane)
that best separates two categories. Imagine your dots are drawn on a piece
of paper: the SVM tries to put a band (margin) of maximum distance between
the classes. If the data is not separable by a straight line, it "projects" these
points into another dimension to create a separation.
▪ Neural Networks: These are layers of "nodes" (like little neurons) that learn
from examples. Each layer makes non-linear transformations to the data,
making it possible to capture complex patterns that simple methods would
miss. Used, for example, in speech recognition - transforming audio into text
using deep neural networks (Deep Learning) or in image classification -
categorizing whether an image reflects a dog or a cat.
o Input Data: Set of data supplied to the model. It is the raw material.
Data can be structured (tables with defined fields) or unstructured
(free text, images and audio). The quality and representativeness of
this data is crucial to obtaining more positive and less biased results.
❖ Discriminative Models: These are models that learn to classify input data
into categories or decisions - in other words, they learn to discriminate
between different classes, focusing on the input characteristics to predict a
label.
Models map input attributes to output labels and are widely used in classification
tasks.
EXAMPLE: a discriminative vision model can identify whether an image contains a
cat or not by detecting patterns such as "four legs, pointed ears = cat".
Prompt: A horizontal, mid-range shot of a photorealistic orange tabby cat standing on its hind legs
in a Michelin-star restaurant kitchen. The cat wears a crisp white chef's jacket and tall toque
blanche, holding a stainless-steel sauté pan with one paw and a wooden spoon in the other,
actively stirring a colorful vegetable-and-herb dish. Surrounding it are gleaming stainless-steel pots
and pans, high-end cookware hanging from a rack, and a wooden prep table laden with fresh
EXAMPLE II: GPT 4o is a generative language model that, when given a prompt,
produces original text (such as answers to questions, essays) because it has
learned the distribution of human language from a vast corpus. Another example
would be DALL-E and ElevenLabs.
What's more, there are many advanced models based on Deep Learning, so
Generative AI can be highly complex, with millions or billions of parameters. It can
be difficult to understand exactly how or why the model generated a specific piece
of content or made a certain decision.
EXAMPLE: A classic example is the LLM models known on the market, such as
ChatGPT, Gemini, Claude, DALL-E, among others.
From a Governance point of view, generative models present unique challenges:
considering their probabilistic nature, they can produce "HALUCINATIONS",
i.e. content that seems plausible but is false or incorrect, and "DEEP FAKES",
highly realistic synthetic content (images, videos and audio) that can be used
maliciously.
https://www.bloomberg.com/news/newsletters/2023-04-06/pope-francis-
white-puffer-coat-ai-image-sparks-deep-fake-concerns
Similarly, discriminative models could also, for example, unduly disqualify
candidates from a CV screening process, considering, for example, poor
representation in the corpus and input data from their learning, favoring the risk
of undue discrimination.
❖ Neural Networks / Deep Learning: The system aims to behave like the
human brain, inspired by the neurons themselves connecting. This network
of "neurons" is organized into interconnected artificial layers, which have
specific functions.
Systems can also contain hidden layers, one or several. In general, the more layers,
the greater the ability to learn more complex and abstract representations of the
data at different levels.
Deep Learning refers to the number of hidden layers in a neural network. In other
words, "classical" or "non-deep" neural networks have one or a few hidden layers,
while a deep learning model has several hidden layers.
Deep learning models tend to be real 'black boxes' because their decisions emerge
from a highly complex internal logic that cannot be directly interpreted. This
"opacity" compromises transparency, makes audits difficult, makes it more
expensive to explain results and hinders the identification of biases.
EXAMPLES: One of the first "classic" Neural Network models was the " Perceptron
", demonstrated in 1957, which has only one hidden layer between the input and
output layers.
Today, given the huge computing power (especially advanced GPUs and TPUs), it is
now possible to train gigantic neural networks with BILLIONS OF PARAMETERS
on massive data sets. These large-scale models are called FOUNDATION
MODELS.
One of the main examples of Foundational Models are LARGE LANGUAGE
MODELS (LLMs), which operate like neural networks trained on huge volumes of
text to predict the next word. They capture syntactic and semantic structures of the
language, enabling tasks such as coherent text generation, translation,
summarization, chatbots, etc.
In addition, we also have VISION MODELS, which can recognize thousands of
categories of objects and bodies; SCIENTIFIC MODELS, which are specialized
networks trained with scientific data and purpose (for example, Google DeepMind's
AlphaFold) and AUDIO MODELS, which are networks trained by huge collections of
speech and music, capable of generating realistic synthetic voices, songs or
transcribing speech into text (for example, Suno and ElevenLabs).
One of the most relevant stages related to models is the adjustment/tuning stage,
also known as fine tuning, which is the process of taking a pre-trained foundational
model and training it further (improving/tuning it) on a specific data set of the target
task in order to specialize it for better results. Fine tuning customizes the model for
purposes such as summarizing documents, or imitating writing styles, without
having to rebuild the entire language capacity from scratch. IBM - Fine Tuning.
• Speed and scale: AI models can operate at high speed and scale operations
globally. This means that both positive and negative impacts can be spread
very quickly and widely.
This capability highlights the importance of systems being rigorously tested before
they are launched or marketed to the public, and of kill switch mechanisms being in
place to shut down the model when unexpected behavior or behavior that is harmful
to humans is detected.
Speed also makes it difficult for human supervision to detect anomalies or
incidents in real time.
REAL EXAMPLE: In 2010, there was a phenomenon called a flash crash that
generated a trillion-dollar crash in the US stock markets and lasted 36 minutes.
After investigations, it was identified that the trader Navinder Sarao launched
million-dollar orders and quickly canceled them using investment software, with
the aim of fooling the high-frequency investment algorithms, which were
responsible for a large part of the trading volume.
• Rogues Ais - Rogue AI: With the rapid advance of technology, there is a
catastrophic risk that Artificial Intelligence can no longer be controlled or
used in a beneficial way.
One example would be that future AIs could pursue goals humans do not endorse.
Furthermore, despite sounding entirely like a sci-fi narrative, AIs could also
generate a sense of existential self-preservation and the pursuit of human power
and disempowerment - a hypothetical example would be an Artificial Intelligence
creating multiple copies of itself, making it difficult to shut down.
B) GENERAL RISKS
EXAMPLE: The "Social Credit System" (SC) that has been built by the Chinese
government since 2014 has the main objective of evaluating the "trustworthiness"
of individuals, companies and public agencies, containing varied information such
as banking and debt history, court decisions and/or administrative
sanctions/environmental violations/tax fraud, as well as information on civic
behavior such as recycling, volunteering and traffic fines. This database
concentrates detailed information on hundreds of millions of people living in
Chinese cities. With the development and study of the use of Artificial Intelligence
models in the country, it is hoped that this technology can further promote social
scoring and, consequently, become an accelerator of state power, promoting a
system subject to internal contradictions.
Papers: Demystifying the Chinese Social Credit System: A Case Study on AI-
Powered Control Systems in China / AI as a Tool for Surveillance: China's Concave
Trilemma.
EXAMPLE: Considering that the vast majority of the data used to train the models is
in English. In this regard, Lelapa recently launched an AI model called VULAVULA,
which converts voice into text and detects names of people and places in written
text using four languages spoken in South Africa - Isizulu, Afrikaans, Sesotho and
English. In this context, ChatGPT was unable to do this accurately.
Internally, AI has embraced this goal in such a literal and monomaniacal way,
without the proper implicit moral safeguards that a human being would (or should)
have.
This hypothetical example illustrates how even a seemingly trivial goal can lead to
catastrophic consequences if AI is not properly aligned with human values and
governed.
Considering the technological advances and the discussions around the risks and
need for regulation of Artificial Intelligence, organizations and companies have
been adopting common principles for the ethical and responsible establishment
and development of Artificial Intelligence. In general, the most common principles
include:
• Fairness: Artificial Intelligence must be fair and equal to all individuals, and
must not establish or propagate discriminatory or biased decisions.
Likewise, AIs must respect the rule of law, human rights, fundamental rights
and democratic values.
EXAMPLE: A financial institution adopts an AI system that performs credit and
default analysis that disadvantages black people to the detriment of white people.
In this context, a person's race should not be considered a factor in credit analysis,
and loan approval rates should be equivalent between different comparable
demographic groups, preventing minorities or marginalized groups from being
unjustifiably disadvantaged.
EXAMPLE: OpenAI has specialized teams for red teaming and feedback on its
models, assessing their security before making them publicly accessible. Human
and automatic evaluations are carried out to mitigate risks and incidents that
• Privacy: Considering the fact that AI models need large volumes of data and
information, among which personal data is also included, it is vital that
developers and applicators, in the course of developing, using and improving
AI models, respect and protect personal data, including adopting
compliance measures and actions in line with local regulations and
international best practices in Privacy. This also includes minimizing the
collection of sensitive personal data and data from vulnerable groups
(children, adolescents and the elderly), adopting security and administrative
techniques and measures such as anonymization, encryption and PETs
(Privacy Enhancing Technologies), transparency regarding the use of
personal data used and its purposes, as well as respecting the rights of data
subjects.
EXAMPLE: When training an AI model on patient health data, use anonymization or
pseudonymization on the records (removal of direct identifiers such as names, social
security numbers, etc.) so that the model can learn medical patterns without directly
exposing patients' identities.
In the same way, AI projects in healthcare are guided by the quest to benefit patients
without adding unnecessary risks. An AI system for diagnostic medical imaging can
autonomously detect and flag suspicious lesions on X-rays, but requires confirmation
from a specialist doctor (radiologist) before issuing a final diagnosis to the patient. In this
way, the human remains in the decision-making loop to ensure safety and ethical
compliance.
One of the first steps is to clearly define who does what in the organization's AI
ecosystem. Three main roles often mentioned, which are also present in emerging
legislation, are: Developer, Deployer and AI User. These terms may have specific
definitions depending on the context, but generically:
• USER / FINAL USER: This is the end user or recipient of the decisions or
assistance provided by the AI system. This could be an internal collaborator
using the tool (for example, a credit analyst who receives an AI score and
makes a decision based on it) or an end consumer/customer (for example, a
driver using a smart GPS, a patient using a diagnostic app).
The user has a responsibility to understand the guidelines for use - to follow the
terms of acceptable use, not to use the system for improper purposes, and to
provide feedback or report problems. In the case of professional users within the
company, they are expected to understand at least at a high level how AI works and
its limitations (which is why training programs are important).
Users should also be encouraged to maintain critical thinking and not completely
delegate decisions without evaluation (avoiding the "authority effect" where they
blindly trust the machine).
It's important to note that the same organization or person can play more than
one role. For example, a startup that develops an AI and uses it in its own service is
both a developer and a Deployer; while a corporate client that buys this AI and uses
it internally may only be a Deployer and have internal users; an end consumer of a
service may only be a user. Understanding these roles helps to allocate governance
obligations in the right way:
• Users: must be trained and integrated into the governance loop, providing
feedback and operating according to policies.
And beware: an entity can be a developer and then also a Deployer of the same
system or another (life cycle). Governance must take these transitions into account
- e.g. during R&D, area X is a developer; when moving to production, area Y
(operation) becomes a Deployer - how to ensure the transfer of knowledge and
responsibilities? Documentation and internal training are key.
• Diversity & Inclusion in the AI team: Ensuring that the teams that develop
and govern AI are themselves diverse in terms of gender, race, background,
etc. broadens the perspectives considered. Homogeneity in the team can
lead to blindness to certain problems (e.g. lack of perception of algorithmic
racism). So involving collaborators from different backgrounds, and even
consulting external interest groups (affected communities) in some cases,
enriches governance.
Stakeholder mapping and stakeholder engagement are useful tools: map who is
affected or interested (customers, regulators, society, employees, shareholders)
and have engagement strategies (communications, public consultations,
publication of white papers and openness to comments, etc.).
Implementing AI governance isn't just about creating rules - it's also about shaping
the organizational culture so that everyone understands what AI is, its benefits
and risks, and acts according to ethical principles. Thus, an essential part of I.B is
to create and deliver training and awareness programs on AI for all stakeholders.
Elements of a good AI awareness program:
Explain the different types of bias (implicit, sampling, etc.) and how they can arise.
Train teams to think critically about AI outputs and detect possible problems.
The aim is to create a culture of responsibility: every employee who deals with AI
feels responsible for its correct use (not just the AI team).
Business teams, on the other hand, need to understand capabilities and limitations
in order to calibrate expectations and plan processes (e.g. the credit team learns
that the AI model provides a score but it's up to them to make the final decision and
documentation to comply with regulation). Executives can have executive sessions
on strategic opportunities and risk governance (higher level). Even the
communications/marketing department could train itself to communicate about AI
responsibly, avoiding exaggerated promises that could mislead consumers (don't
"embellish" AI as infallible magic).
Creating this ethical culture in AI is an ongoing process. But the fruits are valuable:
informed and aligned employees make fewer mistakes, perceive risks earlier and
are more engaged in AI efforts (they don't see it as "IT's black box" but something
that is also their responsibility). In addition, from the point of view of certifications
or regulations, being able to demonstrate that the company has responsible AI
training and culture mitigates penalties and builds trust with regulators and
customers.
In short, educated and aware people are the first line of defense in AI governance.
Not everyone needs to be an expert, but everyone involved should have situational
awareness: knowing that there is a model there, that it follows certain rules, and
that there are policies and support if they encounter a problem or have questions.
This way, the whole organization "rows together" to use AI effectively and safely.
There is no single AI governance model that fits all companies. Approaches need to
be proportionate and adapted to the characteristics of the organization: its size,
digital maturity, industry, product/service portfolio, strategic objectives and risk
appetite. The BoK indicates that an I.B competency is to "differentiate AI governance
approaches based on size, maturity, industry, products/services, objectives and
risk tolerance".
There, AI governance needs to be agile and light enough not to stifle innovation, but
still cover the critical points. For example, in a 10-person startup, a formal monthly
committee might not make sense, but there should be at least one risk-aware
person and some simple checklists.
They can integrate AI governance into existing frameworks (for example, extending
the DevOps pipeline to DevAIOps with ethics gates). Traditional companies at the
beginning of their AI journey, on the other hand, may not even have organized data
or data governance.
The governance approach should meet the organization where it is: if it is still in its
infancy in AI, start with basic governance and get more sophisticated as the
company learns (e.g. first AI project - implement a post-mortem at the end to extract
lessons and improve processes for the second).
In these sectors, compliance and risk are critical - AI governance can even be
audited by regulators. Therefore, these companies need detailed policies,
documentation ready for inspection and perhaps certifications. In less regulated
sectors (marketing, retail, entertainment), governance can be guided more by self-
regulation and reputation.
Still, even without a specific law, ethical principles and general risks apply. For
example, a social network (hardly regulated directly) faces enormous public
pressure about how its recommendation AIs affect society - so it implements robust
voluntary governance to avoid harm (or in anticipation of future regulation).
Also consider the potential failure impact of each application: AI that decides
something life or death (such as medical diagnosis) justifies much stricter controls
than AI for recommending movies. In terms of criticality classification, governance
must be proportional to the risk of the specific use of the AI (risk-based approach
principle).
For example, a customer support chatbot can have lighter governance (occasional
monitoring, content correction if problems are noticed), while a candidate
• Strategic Objectives: The reason why the company adopts AI will influence
governance. If the main objective is rapid innovation for market gain, there
may be pressure for speed, and governance needs to balance this with
caution (perhaps delineating "safe zones" for experimentation and "high-risk
zones" that require extra validation).
If the goal is operational efficiency and cost reduction, governance will focus on
ensuring that AI really does bring reliability and doesn't cause interruptions (a
mistake could negate the savings).
On the other hand, companies merely following fashion may lack conviction -
governance will have to sell the importance internally by showing how AI risks can
affect their own objectives (e.g. a major failure can ruin reputation and alienate
customers, undermining the goal of growth).
• Risk Appetite: Every organization has a risk profile. Some are conservative -
they avoid legal and reputational risks at all costs; others take on more risk
for the sake of innovation. Assessing risk appetite in an AI context helps
calibrate controls.
This may involve filling in an Assessment form answering: What business problem
will be solved? Is AI the best solution or are there simpler alternatives? What data
will be needed and where does it come from? Does it impact customers or
employees, and how? Are there potential ethical or compliance issues (e.g. does it
involve personal data? could it affect vulnerable groups?).
This evaluation makes it possible to filter out inappropriate ideas (for example, an
idea to use AI for decisions that by law must be human should be barred) and
prioritize projects with high value and controlled risk.
• Risk and Harms Matrix: As part of the assessment, produce (or update) a
case-specific risk analysis.
• Documentation from the start: Creation of the project dossier where all
this is recorded: case description, risk analysis, approvals, ethical
requirements. This initial documentation will form part of the project's
compliance file, which will be useful in the event of a later audit or review.
At the stage of obtaining data to train and test the AI, several governance policies
are relevant:
• Data Governance Policy for AI: Expand existing data governance policies to
cover AI needs. For example, if the company already had procedures for
using personal data, reinforce that any use of personal data in AI training
must undergo a Privacy Impact Assessment (PIA).
• Provenance and Data Lineage: Keeping track of where each set of data
came from, when it was extracted, who approved its use, and any
transformations carried out. This helps with accountability: if a problem
arises, you know which data source was responsible. Data lineage tools can
be integrated.
• Consent for Secondary Use: If data has been collected for a different
purpose, check whether the use in AI is compatible or whether additional
consent is required. E.g. customer images collected for security may not be
able to be used to train a marketing algorithm without consent.
In essence, this phase materializes the saying "garbage in, garbage out" -
governance ensures that the AI input meets standards, which reduces the chances
of problematic outputs. Many organizations create a Data Preparation Checklist
that must be signed off by the data and project manager before proceeding to
modelling.
In this phase, the data scientists/engineers build and train the AI model.
Governance policies to be applied:
Also check for overfitting and underfitting as expected - underfitted models may be
ignoring nuances (potentially leading to omitted variable bias), overfitted models
may be coding unwanted noise (potentially amplifying biases present).
And it's vital to involve the business and compliance parties at the end of
development for a cross-functional sign-off. For example, legal reviews whether the
conditions of use (terms and privacy) would cover the new model; compliance
checks whether, say, a credit model is in line with fair credit regulations (or whether
it will need approval from the regulator). This final documentation and sign-off
closes the development phase.
Part of the life cycle is termination. It can happen that an AI system is retired or
replaced. Governance should cover:
• Safe Transition: If you're switching to another model, run both in parallel for
a while to make sure the new one doesn't introduce problems, and phase out
the old one gradually. Notify stakeholders if the change is noticeable (e.g.
"we've updated our algorithm for better accuracy, we expect a better
experience").
It is specifically mentioned in the BoK: evaluate and update existing privacy and
security policies for AI. This is important because AI can introduce new scenarios
not covered by traditional IT policies:
Another part of C.I. is managing third-party risks in the context of AI. This includes:
In the end, an organization that follows these practices can show that:
That's the essence of "comprehensive AI governance": it's not just one document
or committee, but a series of interconnected procedures throughout the
lifecycle. Implementing all of this is challenging but, in proportion to the risks,
necessary - especially in contexts of critical use or high impact. Leading companies
already follow many of these practices, and emerging regulations are likely to make
them mandatory for certain cases (for example, the EU AI Act formalizes pre-use
compliance assessments for high-risk systems, registries, etc., very much in line
with what has been described).
Before the introduction of the main current Data Protection laws, such as the GDPR, the
CCPA, PIPEDA and the LGPD, there were two historical sets of principles that established
ethical and good practice guidelines to guide the responsible processing of personal data.
In the 1970s, with the advance of information technology and the increase in automated
processing of personal data, the need to establish standards and guidelines to protect
individuals against the misuse of their information became evident. In addition, the cross-
border flow of data between countries grew, which began to raise concerns.
In this context, FIPs emerged in the 1970s, and were initially referred to in a report called
"Records, Computers and the Rights of Citizens", published by the US Department of
Health, Education and Welfare (HEW). The general principles, translated into
recommendations and guidelines, generally include:
3. Authority. Agencies should only create, collect, use, process, store, maintain,
disseminate, or disclose PII if they have authority to do so, and should identify this
authority in the appropriate notice.
4. Minimization. Agencies should only create, collect, use, process, store, maintain,
disseminate, or disclose PII that is directly relevant and necessary to accomplish a
legally authorized purpose, and should only maintain PII for as long as is necessary
to accomplish the purpose.
5. Quality and Integrity. Agencies should create, collect, use, process, store,
maintain, disseminate, or disclose PII with such accuracy, relevance, timeliness,
and completeness as is reasonably necessary to ensure fairness to the individual.
7. Purpose Specification and Use Limitation. Agencies should provide notice of the
specific purpose for which PII is collected and should only use, process, store,
maintain, disseminate, or disclose PII for a purpose that is explained in the notice
and is compatible with the purpose for which the PII was collected, or that is
otherwise legally authorized.
Later, in 1980, the OECD refined these recommendations in its "Guidelines Governing the
Protection of Privacy and Transborder Flows of Personal Data", seeking to create a
consensus among member countries to deal with the emerging challenges of the
information age. The document was not a binding treaty, but rather recommendations to
member countries to adopt legislation and practices that comply with these
principles. The principles are practically the same as those set out in the American report
of 1970.
Later, in 2013, these principles were updated by the OECD itself to reflect the new
challenges in a more interconnected era.
• Privacy by Design: Developed in 1995 by Ann Cavoukian, who at the time was the
Information and Privacy Commissioner for Ontario, Canada.
1. Proactive, not reactive: The aim is to anticipate and prevent privacy problems
before they occur, rather than reacting to incidents after they have occurred;
2. Privacy by Default: By default, the settings of a system or service should offer the
highest level of privacy to the user, without the need for any individual action to
activate it.
4. Full Functionality - Positive Sum, not Zero-Sum: The goal here is to achieve not
only the organization's objectives, but also to combine them with privacy
requirements. It is not hoped to create false dichotomies of privacy vs. security or
privacy vs. functionality. The aim here is to achieve both.
5. End to end security: Robust security is essential for privacy. Data protection must
cover the entire life cycle of information, from collection to destruction/secure
disposal.
7. Respect for user privacy: Everything should be centered on the user - the design
should prioritize their interests, offering measures such as appropriate warnings
and user-friendly options for managing their privacy.
Ann Cavoukian also referenced the FIPs, but chose to establish guidelines more linked to
practice, propagating a methodology to incorporate these principles from conception, right
from the start of the process.
Thus, unlike the current broad framework in relation to Privacy and Data Protection, with
many countries having specific legislation on the matter, the regulation of Artificial
Intelligence is still growing and taking shape. To give you an idea, the EU AI ACT, which came
into force in August 2024, is considered the world's first comprehensive legislation
designed specifically to regulate the use of AI systems in EU member states.
With this in mind, as AI becomes ubiquitous and specific regulations are still being
discussed, various "parallel" laws have been and are still being used to establish the ethical
and responsible use of technology.
• Data Protection
The GDPR is considered to be the benchmark global legislation on the subject of Data
Protection and Privacy, having inspired several other pieces of legislation around the world,
including Brazil's own text, the LGPD - General Data Protection Law.
In force since May 2018, the GDPR imposes strict obligations on the use of personal data,
establishing principles (including in line with the FIPs) and rights for data subjects.
In the context of AI, as it is understood that many systems process personal data, it is also
applicable that companies using such systems are also responsible for applying and
complying with the rules established by the GDPR, if applicable to the hypotheses of legal
application of the standard.
In addition to the GDPR, we can also include other legislation in our comments, from
Canada's PIPEDA (Personal Information Protection and Electronic Documents Act),
Australia's Privacy Act, Chile's new data protection law and the LGPD - General Data
Protection Law, enacted in 2018.
Despite local peculiarities, it is clear that, in many contexts, the Data Protection and Privacy
regulatory model for various countries follows a similar model, establishing guiding
principles, rights for data subjects, obligations for organizations, the existence and
supervision of supervisory authorities/agencies, the need to report security incidents and
other normative characteristics that are considered "standardized" in an international
scenario under analysis of the matter.
As a result, both Deployers and Providers, when using AI systems, must also follow local
Data Protection and Privacy regulations, applying the necessary principles and analyzes.
• Intellectual Property
In this context, we can cite a recent case in the USA (Thale v. Vidal), which takes place in
2023. In 2019, Stephen Thaler submitted two patents to the USPTO listing only his AI system
"DABUS" as the inventor. The USPTO rejected the application, claiming that the Patent Act
states that the inventor must be a "natural person". Following the decision, Thaler
appealed. The District Court of Virginia and the Federal Circuit upheld the USPTO's
interpretation. The US Supreme Court refused to consider the issue. Thus, currently,
considering the case law, only patents created by HUMAN BEINGS / NATURAL PERSONS
are allowed to be registered in the US, although the agency has already issued specific
guidelines for AI "assisted" inventions in 2024.
In other countries, such as Europe, patent agencies have also rejected applications with AI
as the inventor, stressing that the invention requires human legal personality.
B) COPYRIGHT AND WORKS GENERATED BY AI: This is undoubtedly one of the main
issues of conflict related to Artificial Intelligence.
Copyright aims to protect original works of authorship fixed in a tangible medium (text,
images, music, etc.). Thus, a relevant question arises: can content created by AI (such
as a generated image) be protected by Copyright?
In several countries, the answer is no, because there is no human author. The US
Copyright Office has been denying registration to works purely generated by Artificial
Intelligence without human creative involvement.
In addition to this discussion, there are extensive debates regarding the use of copyrighted
materials in training Generative AI models. There are several relevant real-life cases taking
this discussion as their object:
A group of authors, such as Sarah Silverman, Christopher Golden and Richard Kadrey, have
filed a class action lawsuit against OpenAI, the developer of ChatGPT, claiming that OpenAI
has used their books, obtained from pirated repositories, to train GPT models, in violation
of copyright and the California Unfair Competition Law and the DMCA (Digital Millennium
Copyright Act). Currently, OpenAI is facing other similar lawsuits in the US and, as a result,
several are being evaluated together in the Southern District of New York, under MDL No.
3143, assigning Judge Sidney H. Stein to conduct the case and suspending the hearings in
California.
In February 2024, Judge Araceli Martinez-Olguín upheld the central charge of direct
infringement (and a limited version of the unfair competition claim), but rejected the theses
In December 2023, The New York Times filed a lawsuit in New York alleging that Microsoft
Corp and OpenAI had used millions of paywall articles to train ChatGPT and Bing AI,
violating copyright and causing a loss of subscriptions and advertising revenue.
In March 2025, Judge Sidney H. Stein rejected most of the defendants' motion to dismiss,
keeping alive the charges of direct and contributory infringement, rejecting the statute of
limitations argument. This is one of the cases centered in Stein's hands, as indicated in
SILVERMAN v. OPENAI.
In both cases, the main legal debate focuses on the extent to which Fair Use, argued by the
big-tech companies and considered an exception to the application of US Copyright Law,
would also apply to this case.
The big-tech companies compare the analysis mainly in a relevant case law involving
Google Books - in 2005, a company sued Google for digitizing millions of books for its search
engine without authorization from the rights holders. Google claimed that the copies served
merely to index and display brief "previews", and did not make complete works available to
the public. On November 14, 2013, there was a summary decision in favor of Google,
declaring that the Google Books project fell under Fair Use, classifying it as "highly
transformative", being beneficial for research and preservation, and harmless to the
publishing market. The Second Circuit, a year later, confirmed the understanding,
reinforcing the four factors that favored Google in the use of Fair Use. The Supreme Court
refused to consider the Authors Guild's appeal, ending the case and consolidating the
understanding that the digitization and display of previews by Google Books constitute
legitimate use under US law.
In addition, a relevant challenge here is precisely what is meant by trade secrets. One
challenge is that AI models, such as Deep Learning, are often associated with "black
boxes", the exact know-how learned being difficult to delineate, which makes protection
complex.
• Health
In the US, the AFFORDABLE CARE ACT (ACA), enacted in 2010, more specifically the
interpretation given to Section 1557, has become one of the main federal instruments to
combat algorithmic bias in US healthcare.
In 2024, HHS/OCR expanded the scope of the rule, making it clear that patient-care
decision-support tools - including machine learning algorithms purchased from third
parties - cannot discriminate against patients or policyholders on the basis of race, color,
national origin, sex, disability or age. The rule requires hospitals, health plans and
insurers to inventory, test for and mitigate any biases, under penalty of losing federal
funding or facing sanctions.
The rule began to apply to algorithms implemented as of July 5, 2024 and to legacy systems
as of May 1, 2025 in the US.
In this context, below are the main practical obligations for entities covered by the ACA:
b- Bias testing: Auditing performance stratified by race, gender, age and disability:
comparing error rates and impacts.
d- Integrated governance: Align these steps with frameworks such as NIST AI RMF
and the value requirements in Medicare / Medicaid programs.
One of the main concerns about the use of AI systems in everyday business processes is
the risk of discrimination. Thus, with the intensification of the use of technology in recent
years, specific legislation has emerged to mitigate and establish controls on discrimination
in selection processes, credit analysis, housing, immigration, among other impact
scenarios.
In New York, there is New York Local Law 144, specifically designed to regulate the use of
algorithmic tools in employment-related decisions - AEDT - Automated Employment
Decision Tool. This law, which came into force in 2023, requires companies that use AI in
hiring processes to carry out annual independent audits to assess and identify possible
discriminatory biases before use and to notify the candidate when a decision has been
automated.
The state of Illinois (USA) has enacted the AI Video Interview Act, which obliges employers
to inform candidates when AI is used to analyze recorded interviews and delete the videos
upon request.
To discuss this topic, the following question is valid: if an autonomous vehicle causes an
accident, who should be held accountable - the manufacturer, the developer of the
vehicle's autonomous driving model, the "driver" present in the vehicle or the vehicle
itself (in this case, the AI)?
Currently, civil liability laws apply to AI indirectly, via theories that have already been
consolidated: negligence, liability for defective products, professional error, etc.
Likewise, it is also necessary to assess the cause and the link - in any incidents involving AI,
technical investigations must be carried out to identify whether the error was the system's
or the person operating it.
In the EU, for example, in September 2022, the European Commission released the
proposed AI Liability Directive, which deals with claims for damage caused by AI systems
or the use of AI, adapting the rules of non-contractual civil liability to AI. In this way, the
standard would complement the EU AI ACT by introducing a new liability regime, with the
aim of establishing greater legal certainty, increasing consumer/user confidence in the use
of technology and helping with liability claims for damage caused.
In any case, the proposal was shelved and the European Commission has no plans to
resume debate on the bill, citing a lack of agreement, as the technology industry has been
pushing for simpler regulations that promote innovation. In this same context, it is also
important to note the harsh criticism made by US Vice-President J.D Vance at the AI Action
Summit promoted by the French government in Paris in February 2025, who called for
Specifically with regard to criminal liability, there is no specific information on whether the
AI system itself has been indicted or held directly responsible, but there is an investigation
into the human beings behind the system. Debates about giving AI autonomous legal
personality have arisen academically, but have not yet been adopted in legal frameworks,
and it is a more accepted trend to frame AI as a tool for which humans (developers,
operators and companies) are accountable.
• California laws
In addition to the indirect application of the CCPA (California Consumer Privacy Act) for
issues related to privacy and data protection, California also has regulations that aim to
establish obligations and guidelines related to the governance and implementation/use of
AI systems.
The central objective is to establish public transparency of the datasets used to train any
GenAI or AI services made available to California residents. The scope of the law is
generative AI systems launched and/or updated on or after 01/01/2022, and public or
private developers who develop free or paid AI systems for the Californian public.
As for penalties, there is no indication of a fixed amount, but it does authorize civil action
by the Attorney General, including the possibility of preventing the service/system
from being offered.
The requirement to disclose data sources allows for external audits of biases,
representativeness and potential discrimination built into the training of models.
Developers will need to assess and justify the appropriate diversity/quality of data to avoid
litigation.
The central aim of the regulation is to guarantee the provenance and labeling of AI-
generated content (text, image, audio and video).
The regulation applies to any person or entity that develops a GenAI model with more than
1 million monthly visitors and that is publicly accessible in California.
➢ Offer the option of manifest disclosure (visible mark) and oblige latent disclosure
(permanent metadata) that identifies that the content is generated by AI and the
name of the provider/time and unique identifier.
➢ Revoke license within 96 hours if the third party removes the disclosure capability.
As for sanctions, a civil fine of USD 5,000 per day is applicable for Covered Providers, in
addition to the possibility of injunctions and costs against third parties.
Although the focus of the law is content transparency, the law facilitates the detection of
discriminatory and disinformative deepfakes by creating technical mechanisms that
support accountability for biased use.
The central aim of the regulation is to ban the use of unidentified bots posing as humans
to influence elections or induce purchases. The law came into force in 2019.
The scope of the law is any "bot" operating online, performing mostly automated
interactions and applicable to websites/applications with more than 10 million unique
users in the US in the last 12 months.
There are no set amounts for non-compliance, but violations can be prosecuted by the
Attorney General as a deceptive practice.
• I AI ACT
The European Union has established itself at the forefront of regulation on the subject with
the introduction of the EU AI ACT, the world's first horizontal regime dedicated to AI.
B) OPERATORS: The EU AI ACT defines specific agents that relate distinctly to the AI
system:
D) PROHIBITED IA PRACTICES
E) HIGH-RISK AI SYSTEMS
➢ Go-live
G) Supervision
For the entire European Union, the European AI Office will be the authority responsible for
the coordination and uniform application of the EU AI ACT. Compared to the Privacy
Framework, it resembles the EDPB - European Data Protection Board.
For the context of the EU AI ACT, GPAI with high impact potential have 10^25 Flops.
At the national level, each member state will be able to establish THREE TYPES OF
AUTHORITIES as part of the implementation of the EU AI Act.
1- Market Surveillance Authority: must carry out activities and take measures
relating to market surveillance and product compliance. In general, e;a will have the
task of ensuring that only products that comply with European laws are made
available.
3- National Public Authority: member states must identify their National Public
Authorities that enforce the obligation to respect fundamental rights in relation to
high-risk systems, with powers to request or access any documentation created or
stored in accordance with the EU AI ACT, when such documentation is necessary
for the effective fulfillment of their mandate, within legal limits.
In the Netherlands, supervision and monitoring will be the responsibility of the Dutch
Digital Infrastructure Authority (RDI) and the Dutch Data Protection Authority (AP). The
supervision of high-risk systems will be the responsibility of the AP and the RDI. The AP will
also be responsible for overseeing transparency obligations. The RDI will be responsible for
coordination and support on technical aspects.
H) Sanctions
Like the GDPR, the EU AI ACT also provides for administrative sanctions in the event of
infringements.
In addition to fines, the authorities can order technical corrections or adaptations to the
model; temporary or permanent bans on use; demand product recalls and publish the
identity of the offender - the effect of naming and shaming (similar to what happens with
the LGPD with the sanction of publicization).
For startups and SMEs, the EU AI Act allows for a reduction in the amount of sanctions
(Recital 103), while maintaining the percentage for large groups.
The sanctions relating to prohibited practices and AI Literacy requirements have been
active since February 2025 in the EU, and the remaining sanctions, applicable to
obligations on GPAI, EU AI Office duties and all high-risk requirements will apply from
August 2025.
In addition to the EU AI Act, which established the global basis for AI governance, there are
other regulations being developed around the world that are worth mentioning:
1- Canada - Artificial Intelligence and Data Act (AIDA): Inserted in Bill C-27, AIDA
establishes a federal "obligations-based compliance" regime for "righ risk
systems", requiring risk assessments, documentation and internal audits. It also
makes it possible to impose fines of up to CAD 25m or 5% of the company's global
revenue.
The text was repealed on January 20, 2025 by the Donald Trump
administration, which in turn issued another Executive Order - Removing
Barriers to American Leadership in Artificial Intelligence. This was a clear
backroom deal between the new Trump administration and Big Tech.
iv. Advances in fairness and civil rights: Seeks to ensure the non-
discriminatory use of AI by combating algorithmic discrimination in
the legal and criminal system. Requires the issuance of guidance to
federal agencies on discrimination and privacy violations arising
from the use of AI. Promotes the responsible development of AI in
health, human services, transportation and education sectors.
Main points:
ii. Review and revocation of policies: The order directs the review of
all policies, guidelines, regulations and other actions taken
pursuant to Executive Order 14110. Any actions that are found to be
inconsistent with or present obstacles to the AI incentive policy
should be suspended, revised or rescinded.
c. California SB 1047: Senate Bill (SB 1047), known as the Safe and Secure
Innovation for Frontier Artificial Intelligence Models Act, was a bill passed
by the California legislature with the aim of regulating advanced AI models.
Obligations:
4- United Kingdom: The British government has opted to adopt a "light touch"
regulatory model, which requires 12 sector regulators to publish plans on how to
apply five guiding principles regarding the implementation and use of Artificial
Intelligence - security; transparency; accountability; contestability and
proportionality - in each of their areas of competence.
Other countries also discussing and evaluating the issue: the United Arab Emirates,
Australia, Nigeria, South Korea, Japan and India.
While laws define what must be done (obligations, prohibitions and rights), technical norms
and standards provide HOW TO - specifications, metrics, practical guidelines for
implementing AI in a safe, interoperable and ethical way.
Several international standardization bodies and international working groups have been
working on international AI standards, including ISO/IEC and IEE, as well as other sectoral
initiatives.
These guidelines complement legislation - for example, a law can require an AI system
to be "auditable and secure" and a technical standard can offer measurable
requirements for auditability and security.
• ISO/IEC
Through the IEEE Standards Association, the IEEE has also established initiatives for ethical
standards in autonomous and intelligent systems.
Within this scope, the IEEE 7000 series family stands out, focused on aspects of human
values in the engineering of intelligent systems.
In addition to the 7000 family, the IEEE produces technical standards in basic areas for AI,
such as neural networks, sensor interoperability, etc.
In 2022, the IEEE launched IEEE 2801-2022, a standard for evaluating the document
transparency of machine learning systems - more focused on Model Cards-type
documentation.
As with ISO/IEC, companies can voluntarily adopt these standards to guide their practices
and demonstrate adoption/respect for the obligations laid down in regulations.
In addition to ISO/IEC and IEEE, there are consortia and industry groups defining their own
guidelines for responsible AI, such as:
b) W3C (World Wide Web Consortium): discusses web standards that can influence
AI, such as standardized representations for explaining automated decisions or
ontologies for learning data.
• OECD AI Principles
The AI principles established by the OECD are considered a global framework. In 2019, the
member countries of the OECD (Organization for Economic Cooperation and
Development), including the US and many EU member states, along with other countries
(totaling 42 countries), adopted five guiding principles for trustworthy AI:
In addition to the 5 guiding principles, the OECD also sets out 5 recommendations for
policymakers, including:
UNESCO - the UN agency for education, science and culture - approved its global standard
Recommendation on the Ethics of Artificial Intelligence in 2021, agreed by 193 member
states. It is an extensive document outlining values, principles and recommended actions
for countries to implement to ensure that AI is developed in accordance with human rights
and universal values.
The document:
▪ It defines fundamental values that should permeate the entire AI, such as respect,
protection and promotion of human rights, diversity and inclusion, harmony with
the environment (sustainability) and guaranteeing peace;
▪ In addition to the principles, UNESCO also details concrete policy measures that
countries should pursue, such as: ethical impact assessments before
implementing AI systems on a large scale (especially in public
▪ It also addresses specific areas, such as ensuring that AI does not exacerbate
gender disparities, such as stereotypes in virtual assistants, and that digital cultural
heritage is respected.
In October 2022, the OSTP - Office of Science and Technology Policy - issued the
Blueprint for an AI Bill of Rights, a set of non-binding principles that aim to protect the
public from potential harm from AI. This blueprint defines 5 rights or expectations that
citizens should have in relation to AI systems:
1- SAFE AND EFFECTIVE SYSTEMS: The right to be protected against unsafe or faulty
AI systems. Authorized systems must be tested for risks, designed to consider
contexts of use and accompanied by continuous post-deployment monitoring to
ensure safety throughout their use and that they do not cause unexpected and
harmful results;
3- DATA PRIVACY: The right to protection and privacy over your personal data used in
AI systems. This means the option of consent where appropriate, limits on the
collection of data beyond what is necessary (data minimization) and incorporating
privacy-enhancing techniques (PETs, anonymization, synthetic data, etc.).
4- NOTICE AND EXPLANATION: The right to know when one is interacting with an AI
(and not a human being) and to receive understandable explanations of what main
criteria led to an adverse automated decision. This echoes the principle of
transparency: for example, if a loan applicant is turned down by an algorithm, they
should be informed that the decision was automated and given a basic explanation
- insufficient income (example).
The blueprint did not create legal rights in the US - it only serves as a federal guideline
and recommendation for good practice. In other words, these rights were not legally
guaranteed to the US population. Government agencies were encouraged to apply it in
their spheres. For companies and developers, the blueprint is a sign of what American
regulators consider to be the basis of responsible AI.
The Human Rights, Democracy and Rule of Law Assurance (HUDERAF) is a framework
proposed by the Council of Europe to ensure that AI systems comply with human rights,
democratic principles and the rule of law.
HUDERAF combines human rights due diligence processes with technical governance.
Broadly speaking, it suggests that organizations conduct AI impact assessments
focused on human rights, democracy and the rule of law - AI HUMAN RIGHTS IMPACT
ASSESSMENT. This includes:
➢ Identify the risks that an AI system can pose to rights (such as privacy, freedom of
expression, equality, etc.);
In 2022, the Council of Europe implemented a tool called HUDERIA (Human Rights,
Democracy and Rule of Law Impact Assessment) aligned with HUDERAF, offering a
structured guide to assessing the impacts of AI on these pillars.
NIST AI Risk Management is a framework designed to help organizations manage the risks
associated with AI and promote the development and use of reliable and responsible AI
systems. Launched in January 2023, the framework aims to provide a flexible, non-sector-
specific and use-case agnostic approach for organizations of all sizes.
Considering the transformative potential and the risks that can negatively impact
individuals, organizations and society, the AI RMF emphasizes the need for proactive
management to build a trustworthy AI culture.
1- VALID AND RELIABLE: Guarantees that the AI is fit for purpose and the specific
intended use/application has been fulfilled. In addition, the AI has the ability to
function as intended, without failure, for a given time interval, under given
conditions. In this sense, AI systems that prove to be inaccurate or do not generalize
well to data and scenarios beyond their training, potentiate negative risks.
2- SAFE: A safe AI system is one that, under defined conditions, does not lead to a
state in which human life, health, property or the environment could be threatened.
Safe operation is enhanced through responsible practices by design. Safety risks
that represent the potential for serious harm or death require the most urgent
prioritization and the most complex risk management processes.
6- PRIVACY-ENHANCED: AI systems must respect the norms and practices that help
safeguard human autonomy, identity and dignity. Privacy values such as
anonymity, confidentiality and control should guide choices by design. Privacy
Enhancing Technologies (PETs) for AI, as well as data minimization methods such
as de-identification and aggregation for certain model outputs, can support the
design of privacy-enhanced AI systems.
7- FAIR: Fairness in AI includes concerns about equality and equity, addressing issues
such as prejudicial bias and discrimination. Fairness standards can be complex
and difficult to define, as perceptions of what would be "fair" can vary between
cultures and can change depending on the application. A system in which
prejudicial biases are mitigated is not necessarily fair - it may still be, for example,
not very accessible to individuals with disabilities or exacerbate existing disparities.
There are three categories of bias that must be considered and managed:
In addition, the AI RMF also establishes 4 macro functions that organizations should adopt
when assessing risks applicable to AI systems:
➢ GOVERN: permeates all other functions and levels of the organization. An AI risk
management culture must be cultivated and implemented within the organization.
This involves establishing policies, processes, procedures, organizational
structures and responsibilities for anticipating, identifying and managing risks.
o Establish and document the risk management process and its results in a
transparent manner;
➢ MAP: Establish the necessary context for identifying and framing AI-related risks.
o Identify the negative and positive impacts of the potential use of AI systems
on individuals, communities, organizations and society;
o Document the knowledge limits of the AI system and how its output can be
used and supervised by humans;
o Select and implement approaches and metrics for measuring the AI risks
listed in the MAP stage, prioritizing the significant risks;
o Establish feedback with users and society who report problems and ask for
automated decisions to be reviewed, integrating this feedback into
evaluation and monitoring metrics.
o Determine whether the AI system achieves its stated aims and objectives
and whether or not its development or deployment should continue;
o Plan and prepare strategies to maximize the benefits of AI and minimize the
negative impacts;
o Monitor the risks and benefits of third-party AI and apply risk controls;
• PLANNING: lays the foundations for the project. Governance is manifested in the
clear definition of the business problem that AI aims to solve, alignment with the
organization's mission and objectives, identification of gaps that AI can fill,
preliminary assessment of data availability and suitability, definition of the project
scope and establishment of the initial governance structure, including
identification of a Champion or "sponsor" to ensure support and resources.
• DEVELOP: Here, the model is actually built and trained. Governance focuses on
ensuring that data is collected and used in accordance with established legal
standards and ethical principles, that testing is comprehensive (unit, integration,
validation, performance, security, bias and interpretability), that data lineage and
provenance are maintained, and that risks identified during training are monitored
and mitigated. Documentation of the training and testing processes is an essential
part of this.
• DEPLOYMENT: Involves releasing the model into production and its ongoing
monitoring and maintenance. Governance includes the final assessment of
readiness for release, establishing maintenance and upgrade schedules, carrying
out periodic audits, managing incidents and risks and analyzing root causes of
failures, and ensuring transparency through public disclosure and technical
documentation.
In all phases, it is also recommended to identify the correct stakeholders for each phase -
Legal, Compliance, Ethics, Privacy, Info Security, Data Science, Engineering, IT and so on.
In addition, gate reviews (checkpoints between phases) also help to ensure that
governance requirements are met.
Firstly, it is vital that organizations understand precisely the problem they are trying to solve
by developing the AI system. Without a thorough understanding of the problem and the
associated objectives, there is a risk of developing a solution that is technically impressive
but fails to deliver real value or, worse, introduces new, unanticipated risks.
From the point of view of governance, the aim at this stage is for this definition to be
structured, collaborative and aligned with the organization's overall strategy.
Methodologies such as Design Thinking can be useful for empathizing with end users and
deeply understanding their needs and pains. Similarly, user interviews, journey mapping
and root cause analysis (such as the 5 whys?) help the organization to go beyond superficial
symptoms and identify the fundamental problem. Here, the following questions can be
asked:
Likewise, it is essential that the project's objectives are directly linked to the organization's
broader strategic objectives. This ensures that the project contributes to the company's
overall mission and thus makes it easier to get support for its execution. With this, a clear
demonstration of the expected business value (ROI, cost reduction, revenue increase,
improved customer experience) becomes essential to justify the investments,
resources allocated and, consequently, the time that will be spent.
• Technical feasibility: Does the organization have the necessary data (in quantity
and quality), technical expertise (data scientists and AI engineers) and
infrastructure capable of supporting the project?
• ECONOMIC VIABILITY: Is the cost of developing and maintaining the AI system over
the long term justified by the expected business value? A detailed cost-benefit
analysis should be carried out.
• LEGAL AND ETHICAL VIABILITY: Does the intended use of AI comply with
applicable laws and regulations? Are there significant ethical risks that need to be
addressed and eradicated before implementation?
Another aspect that should be assessed is the definition of objectives and aligned KPIs. In
this way, organizations can, for example, adopt the S.M.A.R.T. framework:
Specific
Measurable
Achievable
Relevant
Time-bound
In this case, instead of having the objective of "improving customer service", a SMART
objective would be "reducing the average waiting time on the phone by 20% over the next
12 months by implementing an AI chatbot to answer frequently asked questions on our
website".
Finally, a question that must be asked internally is whether to DEVELOP vs. BUY.
Organizations have the option of either developing the AI models they deem necessary
internally or acquiring them ready-made from suppliers/partners. Governance requires a
careful assessment of the risks and benefits of each of these approaches. Developing the
AI system in-house, despite being more complex, offers more control and monitoring over
all phases, but ends up requiring resources and expertise. Acquisition by a partner/supplier
tends to be faster, but introduces security, privacy and compliance risks for the
supplier/partner which, depending on the internal governance of the contracting company,
may become more latent. For the purchase/acquisition of AI systems by
partners/suppliers, it is recommended to carry out rigorous due diligence, information
security practices and controls and to formalize appropriate contractual clauses.
Once the problem has been addressed and the objectives clearly defined, the next crucial
stage in the design is the selection of the appropriate type of AI solution and its underlying
technical architecture. The choice should not be based solely on technical sophistication,
but rather on the suitability of the problem, the available data, resource constraints and,
crucially, the governance considerations associated with each approach.
• GENERATIVE SYSTEMS: Create new content, such as text, images, audio and
code. Examples include the most popular LLMs, such as ChatGPT, Gemini,
Claude, DALL-E, Midjourney and ElevenLabs.
In addition to the choice of system and its respective governance-related risks, the specific
architecture must also be chosen, which could be CNN, RNN, Transformer, etc.
• Feed-Forward (FNN / MLP): Data travels through dense layers in a single direction,
without cycles.
In other words, the data enters through one door, passes through a row of
tables and leaves through another door. There are no cycles or other paths.
Generally used to predict credit risk, classify emails as spam or not or any data in a
table.
• Convolutional (CNN): Layers apply local filters that are repeated over the entire
input.
Imagine a team of researchers from antiquity using magnifying glasses - each one
analyzes small pieces of an image and highlights patterns (edges, colors, shapes).
Then each of them puts it all together and makes a final assessment.
Generally used for photo recognition (e.g. detecting cracks or potholes in roads),
automatic reading of x-rays or other images for medical purposes, quality control
by camera in a factory.
• Recurrent (RNN / LSTM / GRU): It not only processes current information, but also
takes into account what it has learned from previous information in the sequence.
Information can "circulate" within the network, allowing it to remember past events
when processing new ones. In other words, it has a "memory" to facilitate current
data processing.
• Graph Neural Network (GNN): model entities as nodes and their relationships as
edges, letting information propagate across the graph. It doesn't just look at single
items, but at how they are connected in a kind of network or "graph". It analyzes
each "node" and "edge" - the information from one "node" can influence its
neighbors, causing the GNN to learn from this relationship structure.
These are technical points that require evaluation by data scientists and AI engineers, but
AI governance can also influence the choice of architecture, taking into account the
requirements of explainability (simpler models, such as decision trees, are preferable in
some regulatory contexts), robustness and computational efficiency.
The decision on the chosen architecture must therefore be taken into account:
A central pillar in the governance of AI in the design and development phase is the proactive
conduct of impact assessments. The aim of these assessments is to understand and
mitigate the potential negative effects of the AI system on individuals, groups, organizations
and society as a whole, covering ethical, legal and social dimensions.
• Data Protection Impact Assessment (DPIA): Required by the GDPR and other
regulations of the same nature, the DPIA focuses its analysis on the risks applicable
to the processing of personal data in high-risk activities.
• Human Rights, Democracy and the Rule of Law Impact Assessment (HUDERIA):
Proposed by the Council of Europe and aligned with HUDERAF - Human Rights,
Democracy and The Rule of Law Assurance Framework - this assessment seeks to
evaluate the potential impacts of AI systems in these fundamental areas. There are
also tools such as the RIN - Risk Index Number and the Mitigation Hierarchy.
• Ethical Impact Assessment (EIA): Evaluates the system's alignment with ethical
principles (justice, autonomy, non-maleficence, beneficence, explainability, etc.),
identifying potential ethical dilemmas.
In general, conducting an effective impact assessment generally follows a process that can
be structured as follows:
➢ Risk analysis: Assess the probability and severity of the risks identified (using
methodologies such as HUDERIA's risk matrix or RIN);
In addition, with regard to the structure of these risk assessments, organizations can use
ready-made and publicly accessible templates provided by either supervisory
authorities/agencies (such as the iconic DPIA template provided by the ICO) and/or
use/adapt technical risk management standards (such as ISO/IEC and IEEE), as well as
established frameworks such as the NIST AI RISK MANAGEMENT FRAMEWORK. Many
technology companies also provide such documentation publicly, such as Microsoft and
Google.
In any case, it is important to note that, even with the ease of use of the accessible
template, it is recommended that organizations adapt their impact assessments to the
existing regulatory requirements in each region, the organization's own risk strategy, sector
of operation and other contexts that may vary.
After identifying the potential negative impacts during the impact assessment, the next step
in AI system design governance is systematic risk assessment. This involves analyzing the
probability of occurrence and the severity of occurrence, allowing the most critical
➢ Probability and severity matrix (e.g. 3x3 or 5x5): visual and widely used tool. It
classifies risks in two dimensions: PROBABILITY (low, medium, high and very high)
and SEVERITY (insignificant, marginal, moderate, critical or catastrophic).
The combination of these two factors positions the risk in the matrix, which generally uses
colors to identify risk levels.
➢ HUDERIA's Risk Index Number (RIN): this analysis, included in HUDERIA, offers a
granular approach focused specifically on the risks applicable to human rights. It
consists of:
Finally, after identifying the RIN and the applicable risks, the organization must conduct a
Mitigation Plan, which uses a hierarchy to choose the measures in proportion to the
severity/RIN. To do this, 4 GRADUAL levels are defined (the first being the ideal and the last
the least recommended):
(i) AVOID: Modify the design or development to completely eliminate the risk.
Here, you want to prevent the impact before it exists;
(iii) RESTORE: After one year, to return people to the same or equivalent situation
as before (EXAMPLE: public retraction, correction of registration, restoration of
a denied right).
➢ FAIRNESS: Ensuring that the AI System does not produce results that are
systematically biased or discriminatory against individuals or groups, especially
those belonging to protected categories. This involves the careful analysis of
training data, the selection of appropriate fairness metrics (EX: demographic parity,
(i) Implement clear policies and guidelines: Internal documents that define the
organization's commitment to ethical AI and provide specific guidance for
development teams.
(ii) Training and awareness-raising: Educating teams about ethical principles and
how to apply them in their daily work.
In addition, it is important for the AI Governance team to bear in mind that applying ethical
principles can involve trade-offs. For example, increasing explainability may, in some
cases, slightly reduce the accuracy of the model. Maximizing fairness for multiple groups
simultaneously may be mathematically impossible in certain situations. Governance
involves navigating these trade-offs in a transparent and justified way, documenting
decision-making and assessing the context, risks, legal/regulatory analysis and the
company's strategy on the subject.
1) Identification and analysis: the first step is to identify who the relevant
stakeholders are.
INTERNAL stakeholders can be executive leadership (senior managers and the board),
product teams, Legal, Compliance, Privacy, Information Security, HR, Marketing, Sales,
Operations, Customer Support, Communications and other areas of the organization.
After the first identification, you should analyze the relationship of each of these
stakeholders in the project (whether they support, oppose or act neutrally) and the risks
and opportunities they represent.
a. Stakeholders with HIGH power and HIGH interest: they should be closely
involved, consulted regularly and included in key decisions;
With the identification and engagement of stakeholders, in conducting the AI project and
governance for the system under development, it is natural for conflicts to arise, which
can occur due to divergent priorities (e.g. speed of development vs. ethical rigor and privacy
risks). Governance must establish mechanisms to identify, analyze and manage these
conflicts constructively, seeking solutions that balance the different interests or escalating
the decision to a higher level when necessary.
Thus, adopting an evaluation framework can help ensure that all relevant aspects are
considered. As a minimum, this framework should cover:
(i) IMPACT: What is the potential positive impact of the AI solution on the identified
business problem? Will it solve the problem completely or only partially? What
is the expected quantifiable value (cost reduction, revenue increase, efficiency
improvement, increase in customer satisfaction); Is the impact strategic for the
organization?
(ii) EFFORT: What resources are needed to implement and maintain the solution?
This includes financial costs (development, infrastructure, capacity), time
(projected schedule), technical expertise and data.
(iii) FITNESS: How does the proposed AI solution fit into the organizational context
and with the company's strategic objectives? How is the AI aligned with the
organization's culture and values? Is it compatible with the organization's
existing logical infrastructure and other systems/processes? Does the
organization have the necessary maturity and readiness to adopt and manage
this technology? What are the associated risks (technical, ethical, legal and
operational) and does the organization have the capacity to mitigate them?
In addition, with the AI hype and the various ideas and solutions that aim to solve various
internal problems, it is common for organizations' AI Governance teams to receive several
queries for case ideas. As a result, a prioritization process must be carried out to
identify the projects that are most essential and feasible. This can be done using the
following methods:
• SCORECARDS: Assign scores to each use case based on predefined criteria (e.g.
IMPACT, EFFORT, FITNESS and RISK) and rank the use cases based on the total
score;
• MoSCow (Must Have, Should have, Could have, Won't have): Classifying use
cases based on their strategic importance and urgency, always varying from
organization to organization.
No less important, this process of analyzing and evaluating the intended use case must
also include the participation of relevant stakeholders, who will bring insights,
Another relevant action can be to carry out case studies and benchmarking with other
organizations in the same sector or with similar challenges, to understand how they
have implemented and evaluated similar (or identical) use cases to provide valuable
insights.
The absence of adequate documentation can expose the organization to significant risks
and undermine confidence in the system, even before its final implementation.
But what should be documented? Documentation at this stage should cover all key
decisions and processes, including:
(i) DEFINITION OF THE PROBLEM AND OBJECTIVES: The business case, the
SMART objectives, the defined KPIs;
(iii) LEGAL AND ETHICAL REQUIREMENTS: The applicable laws and regulations
identified, the ethical principles adopted and how they have been incorporated
into the design, the security and robustness measures of the model, the legal
assumptions for data processing, etc;
(v) DATA: Sources of data, legal rights regarding its use, collection, preparation
and cleaning processes, quality analysis and possible biases, description of
training sets, validation and testing;
(vi) DESIGN AND ARCHITECTURE: Justification for choosing the type of solution
and architecture, detailed description and design of the architecture,
algorithms used, model parameters, keys and weights;
(viii) PILOT TESTS: Plan, results and decisions based on the pilot carried out;
(ix) KEY DESIGN DECISIONS: Justifications for important choices made during the
process.
Still in the context of documentation, some standardized structures and tools may be
advisable:
➢ Model Cards: concise documents that summarize the essential information about
an AI model, including its purpose, performance, limitations, data used and ethical
considerations;
o EXAMPLES:
https://modelcards.withgoogle.com/
https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templ
ates/modelcard_template.md
➢ System Cards: Used by organizations such as OpenAI, they provide a broad view of
the AI system as a whole, including not only the model, but also the infrastructure,
interfaces, risks and usage policies, addressing safety and limitations in a more
systemic context.
o EXAMPLES:
Finally, it should be noted that this documentation should be stored in an accessible and
organized place, also following the recommendations:
▪ Review and approval: Implement formal review and approval processes for key
documents;
o Log keeping: Ensure that the system automatically generates logs to track
its operation and enable it to be monitored and audited;
The first step in data governance for AI is to ensure that the organization has the legal and
unambiguous right to collect, process and use the data for the intended training and testing
purposes. Failure to establish a solid legal foundation can result in violations of privacy,
intellectual property/copyright, contractual disputes and severe regulatory sanctions. This
whole process must be well documented.
Under the Privacy and Data Protection aspects of the use of personal data, depending on
the specific jurisdiction, there may be different legal bases. Under the GDPR, for example,
the most common legal bases include:
(i) Consent
(ii) Contract execution
(iii) Legal obligation
(iv) Public authorities
(v) Legitimate Interest - often invoked to use data for AI training.
Also, with regard to the use of personal data, it is essential to demonstrate transparency to
users regarding the use of their information. To this end, it is necessary to include specific
passages in the Privacy Policies and Terms of Use (TCUs) that indicate in clear language
that your personal data can be used, for example, to train AI models.
Still in the context of personal data processing, the use of sensitive personal data generally
requires stricter legal bases and stricter controls, considering the greater impact.
Now, under the Intellectual Property sphere, the use of data for AI training raises complex
and still uncertain issues, especially related to Copyright, considering the use of protected
content in generative models without permission, potentially posing significant litigation
risks, similar to what happened between The New York Times v. Open AI and The Wall Street
Journal v. Perplexity. In addition, there are also risks related to trade secrets, with the use
of proprietary data that may contain confidential and restricted information about
organizations.
The maxim "garbage in, garbage out" highlights the critical importance of data quality in the
development of AI systems. Models trained with inaccurate, incomplete, inconsistent,
outdated, irrelevant and poorly represented data have produced unreliable results,
potentially leading the system to make wrong, biased decisions and, consequently, harm
users.
Data governance for AI must therefore implement rigorous processes to guarantee and
maintain high data quality throughout the system's lifecycle.
o Relevance: The degree to which the data is appropriate and useful for the
specific purpose of the model. Using irrelevant data sets can introduce
noise and degrade performance.
Data preparation, often called Data Wrangling or Data Munging, is the process of
transforming raw and often disorganized data into a clean, structured format suitable for
analysis and training AI models. It is notoriously one of the most time-consuming and labor-
intensive stages, consuming many final hours of the project.
Effective governance of this stage is crucial to guaranteeing the quality, traceability and
compliance of the process.
• Why is preparation necessary? Raw data is rarely ready for use. It can come from
a variety of sources, in a variety of formats (structured, semi-structured and
unstructured), contain errors, missing values, inconsistencies and inaccuracies, or
simply not be in the format required by the algorithms.
o Structuring: Organizing the data into a tabular or other format suitable for
analysis (EX: converting unstructured logs into rows and columns);
o Versioning: Control the versions of the preparation scripts and the resulting
data sets;
The concept of the "5Vs" (Volume, Velocity, Variety, Veracity and Value) is often used to
describe the characteristics and challenges in Big Data. In the context of AI development,
understanding and managing each of these dimensions can also be key to effective data
governance.
• VOLUME: Refers to the enormous amount of data generated and collected. For AI,
large volumes of data are often needed to train complex models, especially deep
learning.
• Speed: Describes how quickly data is generated and needs to be processed. Many
AI applications require real-time or near-real-time processing.
• Variety: This refers to the different types of data that need to be managed, including
structured data (relational databases, spreadsheets), semi-structured data (JSON,
XML, Logs) and unstructured data (text, images, audio and video).
• Veracity: This concerns the quality, reliability and accuracy of the data. Data can
be uncertain, imprecise, ambiguous or contain biases.
• Value: Refers to the usefulness and business value that can be extracted from the
data. Not all the data collected is valuable or relevant to a given problem.
All five of these dimensions are interconnected and often involve trade-offs. For example,
to increase processing speed, veracity can be compromised if quality checks are
simplified. Dealing with high variety can increase the complexity and cost involved.
Effective governance requires a holistic understanding of these dimensions and
making conscious decisions about how to balance them according to the
requirements of the specific use case.
EXAMPLE: An AI system for monitoring urban traffic. VOLUME: data from sensors,
cameras, GPS from millions of vehicles. SPEED: Data arriving in real time. VARIETY:
Numerical data from sensors, images from cameras, route data from traffic apps, weather
data. VERACITY: Accuracy of sensors, quality of images in different conditions, reliability
of user reports. VALUE: Ability to predict congestion, optimize traffic lights and inform
drivers of routes and events/alerts.
o Missing values: missing data in certain fields of the data set. Missing values
break the traceability of decisions and can violate rules of fairness if the
absence is concentrated in minority groups.
o Outliers: Extreme values that deviate significantly from the rest of the data
set. An example would be a database with customer demographic
information and there is a finding of people aged 200.
o Typing and formatting errors: Data entered incorrectly in the set or that
does not follow the expected pattern.
In addition, where applicable, when the data set contains personal data, anonymization
and pseudonymization techniques are always recommended, for example:
Considering that data flows from multiple sources, undergoes various transformations and
is used to train and evaluate different versions of models, the ability to trace the origin
and history of this data becomes fundamental. Data lineage and provenance are key
concepts for this traceability, essential for governance, auditing, debugging and
reproducibility.
• PROVENANCE: Refers to the origin and historical record of the data. It answers
questions such as - Where did this data come from? Who created or modified it?
When? What processes were involved in its creation?
• Lineage: Describes the complete life cycle of the data, mapping its flow through
different systems, processes and transformations over time. It answers questions
such as - What sources fed this dataset? What transformations have been applied
to this dataset? What models or applications use this dataset?
(ii) Reproducibility: Makes it possible to recreate data sets and model results -
essential for scientific validation and debugging.
(iv) Debugging: Facilitates the identification of the root cause of errors or biases in
the model, tracing them back to the original problematic data or processes.
(vi) Trust and transparency: Increases confidence in data and models by providing
visibility into their history.
Testing models goes far beyond measuring accuracy or performance. Given the complexity
and potential impact of AI systems, a comprehensive, multi-faceted testing strategy is also
essential to ensure robustness, security, fairness and compliance with requirements and
purposes. Governance requires that testing is planned, executed and documented on a
regular basis.
• Factors that guide testing: the testing strategy should be informed by:
o Identified risks: focus tests on the areas of greatest risk identified in the
impact assessments;
o Review and approval: Formal process for reviewing the results and
approving the move to the next stage/phase of model
development/implementation.
o ROBUSTNESS TESTS - Tests that assess the model's resilience in the face
of non-ideal or unexpected situations.
▪ Testing with Outliers and Extreme Values: Check how the model
reacts to rare, extreme or anomalous inputs. Depending on the
inclusion of these types of "dirt", the model may simply "break" or
generate inaccurate or completely absurd answers.
o FAIRNESS TESTS - Essential to ensure that the model does not produce
discriminatory or unfair decisions, either directly or indirectly.
• Ensemble methods
Ensemble is the practice of combining two or more machine learning models to produce a
single prediction or decision. Instead of betting on just one "champion model", ensemble
treats each model (learner) as a "voter" or "expert" and uses combination rules (vote,
average, meta-model etc) to arrive at the final output.
o Voter - A learner whose output goes into a voting scheme - majority, average
and weighting. Focuses on the mechanics of the decision - each person
casts a vote.
In other words, by evaluating these "terms", we can see that all the "voters" and "experts"
are learners.
o Boosting: Each new model receives special focus on the mistakes that the
previous one made, like a chain review. It is relevant in situations where a
single model has become "myopic" and needs to reduce bias rates, and can
improve coverage in minority groups.
An inadequate division of data can lead to optimistic (or pessimistic) estimates of actual
performance and promote the occurrence of overfitting. Therefore, the way in which the
available data is divided up for different purposes during the life cycle is a relevant
methodological decision.
o Training Set: The bulk of the data (usually 60-80%), used to teach the model
to learn the patterns and relationships present in the data. The model
adjusts its internal parameters (weights, biases) based on this data;
o Validation Set: A smaller portion of the data (usually 10-20%), used during
the training process to:
o Test Set: A final portion of the data (usually 10-20%), kept completely
separate throughout the training and hyperparameter tuning process. It is
used once, at the end, to provide an unbiased estimate of the model's
performance on completely new and unseen data. The result on the test set
represents the best estimate of how the model will behave in the real world
when it is implemented.
• Governance considerations:
o Separate data sets by purpose and define restricted use: Ensure that no
information from the validation or test set is used, directly or indirectly,
during model training. This includes not adjusting hyperparameters based
on the test set.
o Reproducibility: Use fixed Random seeds during splitting to ensure that the
same sets can be recreated if necessary in the future.
A practical example would be the creation of an AI system for classifying images of cats and
dogs. Using the 70/15/15 stratified split ensures that the proportion of dog and cat images
is the same in the training, validation and test sets, preventing the model from being unfairly
evaluated if, by chance, the test set contains mainly images of one type.
In many AI development scenarios, obtaining real, sufficient and high-quality data can be a
significant challenge. Data can be scarce, expensive, difficult to collect, or contain
sensitive information that restricts its use. In such cases, synthetic data - data artificially
generated by algorithms or models to mimic the statistical properties of real data - emerges
as a valuable approach, although its use requires careful governance considerations.
The aim, when creating synthetic data, is for them to share statistical characteristics
(distributions and correlations) with the real data they are intended to simulate.
Synthetic data can be generated in different ways, whether (i) based on statistical rules and
models, from known distributions, business rules; (ii) based on machine learning models,
which can use generative adversarial networks, variational autoencorders; (iii) flow-based
models, through diffusion models and (iv) simulation/agent-based, with the creation of
virtual environments or agendas that integrate for data generation.
o Test and development: Create realistic data sets to test systems or train
models without relying on production data.
o Statistical fidelity: How well does the synthetic data replicate the
statistical properties (marginal distributions, correlations, etc.) of the
synthetic data replicate the statistical properties (marginal distributions,
correlations, etc.) of the real data? Statistical metrics and comparative
visualizations can be used.
o Privacy: How well does synthetic data protect the privacy of real records?
Tests such as membership inference attacks can assess the residual risk.
o Biases: Synthetic data can inherit or even amplify biases present in real
data. Fairness assessment is also necessary;
10. Monitoring and risk mitigation during model training and testing
The process of training and testing AI models also requires risk assessments and controls.
Various problems can arise, compromising the performance, fairness or safety of the
resulting model. Effective governance requires the implementation of continuous
monitoring mechanisms to detect these risks early and the application of appropriate
mitigation techniques.
• Common risks
o Overfitting: The model learns the training data excessively well, including
noise and sample-specific patterns, but fails to generalize to new and
unseen data. In other words, it is so adherent to the training that, when run
in practice with new data, it performs poorly on the validation and test set.
• Monitoring strategies:
o Learning curves: Plot the model's performance (e.g. error, accuracy) on the
training and validation sets over the training epochs. A large gap between
the curves indicates overfitting - both curves stagnating at low performance
indicate underfitting.
o Error analysis: Investigate the errors made by the model to identify patterns
or subgroups where it fails the most.
This gatekeeping process ensures that the model meets all the established technical,
legal, ethical and business requirements, and that the residual risks are acceptable and
manageable. Governance at this stage involves defining clear acceptance criteria and a
formal review and approval process.
• Evaluation frameworks
o Fairness and Ethics: Have the risks of bias been adequately mitigated?
Does the model operate in accordance with the ethical principles of the
organization? Are the results of the fairness audit acceptable?
Model Cards have emerged as a tool to promote transparency and accountability in the
development and deployment of AI models. Inspired by nutrition labels and fact sheets,
they provide a structured and accessible summary of the most important information
about a specific model, its data, its performance, its limitations and its ethical
considerations. The creation of a comprehensive Model Card is often associated as a key
requirement in the assessment of readiness to release a system for deployment.
o Usage considerations:
▪ Primary use cases (what are the specific tasks that the model has
been designed and validated to perform?);
o Risks and biases: What are the potential biases identified in the model's
behavior and what are the potential ethical, social and security risks
regarding its execution.
It is created during the development of the model and is updated as new information
becomes available. In addition, it is also recommended that it be versioned and updated to
reflect new changes in operation, risks and considerations to be made about the model.
The Model Card focuses on the individual model. The System Card focuses on the
complete AI system.
- System Card: This has a broader scope, addressing the AI system as a whole, which can
include multiple models, software components, hardware and even policies and
processes that govern its use. It describes how the system works at a higher level, its
capabilities, possible interactions and the mitigations put in place for potential risks and
damage.
A pilot aims to test the system in a real or semi-real environment, but on a limited basis, to
collect data on its performance, feedback from users, identify unforeseen problems,
validate assumptions and refine its deployment strategy before full launch.
• Methodology
o Planning: Clearly define the specific objectives of the pilot and its scope
(which functionalities, which users/segments, which environment), the
success metrics, its duration and the resources needed to carry it out.
o Analyze the data: Analyze the data and feedback collected, in relation to
the defined objectives and metrics. Identify successes, failures, lessons
learned and areas for improvement.
• Types of pilots
o A/B Testing: Two or more groups of users are exposed to different versions
(e.g. old system vs. new AI system) and their results are statistically
compared to determine which one performed better.
The implementation of an AI model is not the end of the line for governance - it is the
beginning of its operational life, which requires continuous maintenance and monitoring,
as well as updates and often periodic training to ensure that it continues to operate
effectively, safely and relevantly.
o Data drift and concept: The real world changes. The statistical distribution
of input data or the relationship between inputs and the desired output can
change over time, degrading the performance of a model trained on older
data.
o Periodic Re-Training: Schedule set to re-train the model with more recent
data. The frequency depends on the volatility of the environment and the
observed drift rate (it can be daily, biannually, monthly, quarterly, etc.).
o System criticality: the more critical/higher risk the system, the more
intensive the monitoring and updates must be;
o Cost vs. Benefit: Balance the cost of re-training/updating with the expected
benefit in terms of performance and risk mitigation.
o Data availability: The frequency with which relevant new data becomes
available.
➢ Process governance
New vulnerabilities can be discovered after the system has been deployed/released. In
addition, threats can evolve and the system's own behavior can change unexpectedly.
Therefore, continuous governance requires periodic security audits and tests to verify that
the system continues to operate securely, effectively and in compliance with internal
policies and requirements.
o Internal audit: Carried out by a specialized team from within the company.
The frequency of the audit can vary according to the risk level of the system, regulatory
requirements and the results of previous audits (it can be annual, biannual or more
frequent for critical systems).
➢ Governance practices
In addition to security audits and tests, two more proactive techniques focused on
simulating real adversaries are: Red Teaming and Threat Modeling. They help identify
vulnerabilities and risks that may not be apparent through more conventional approaches.
➢ Red Teaming for AI: Inspired by cybersecurity exercises, Red Teaming for AI
involves a dedicated team simulating attacks by adversaries with the aim of finding
weaknesses in the AI system, the data or the associated processes.
The aim is not only to identify vulnerabilities, but also to assess the effectiveness of
defenses, detection and response processes and the overall resilience of the system. The
Red Team process can be carried out in the following steps:
o Planning: Define the objectives (e.g. trick the model, extract data, cause
denial of service), the scope (which components of the system), the rules of
engagement (what is allowed to be done) and the composition of the team
(expertise in AI, security, ethics, privacy, etc.).
(I) Evasion attacks: Creating subtly modified inputs (images and text) that are
classified as incorrect by the model;
(II) Poisoning Attacks: Inserting malicious data into the training set (during initial
training or through feedback loops) to compromise the model;
The difference lies in when they are applied and the purpose of each approach.
Threat Modeling: This is a proactive approach that aims to identify and mitigate possible
threats during the early stages of system development. It involves analyzing the
architecture, data flows and possible attack vectors to anticipate vulnerabilities before
the system goes into final operation. It involves theoretical analysis and modeling.
Red Teaming: This is a practical, offensive approach that simulates real attacks to test
the effectiveness of an organization's defences. In other words, it is practical and reactive,
focusing on identifying flaws that can be exploited by real attackers in already operational
environments, applying offensive tests.
Despite all prevention and testing efforts, incidents and failures can occur when the AI
system is operating in the real world. These can be technical failures, performance
degradation, unexpected or biased behavior, security or privacy breaches and unintended
consequences.
o Eradication: Identifying and eliminating the root cause of the incident. This
may involve bug fixing, security re-training.
In addition, like other aspects of AI governance, documentation is also a relevant action for
incident management, response and containment, and there should be a detailed record
of all incidents that have occurred, including timestamps, description, impact, actions
taken, resolution and lessons learned.
No less important, the organization must define the protocols for communicating this
incident both to the responsible and applicable authorities/agencies, but also to the
impacted owners/users (if applicable), in addition to internal communication to the teams.
This can include senior management, Legal, Compliance, Public Relations and
Communications, among other teams.
To ensure that an AI system operates consistently, safely and in line with the policies and
objectives defined on a day-to-day basis, it is necessary to implement a set of operational
controls. These are systematic mechanisms and procedures focused on the ongoing
management and supervision of the system in production. Governance must ensure that
these controls are adequate, implemented correctly and monitored for their effectiveness.
➢ Types of controls
o Preventive: Designed to prevent errors, fraud or unwanted events from
happening (e.g. input validation, access control, segregation of duties,
training);
o User understanding: Helps users understand how the system works, how
to interact with it and what its limitations and associated risks are;
As for the components of this technical documentation, after the system has been
implemented, the ongoing documentation should include:
(i) Version history: Record of all versions of the model deployed, with dates and
description of changes;
(v) Audit and test reports: Results of internal and external audits, as well as
periodic security tests.
Organizations must also take into account the balance of transparency, assessing the
expected balance between full transparency and possible impacts. For example, by
adopting total transparency in documentation about the AI system, its infrastructure,
operation and parameters, organizations can expose Intellectual Property (trade secrets,
proprietary algorithms) or create attack vectors (reveal details and flaws to be exploited) by
malicious third parties/attackers. Governance must define the appropriate level of
transparency for each type of information and audience, justifying its decisions not to
disclose certain details on the basis of legitimate risks.
Planning the controlled and responsible decommissioning and end-of-life of the system is
the final stage of AI lifecycle governance.
The first and perhaps most crucial step in the pre-deployment evaluation of an AI system is
a thorough understanding of the specific context of the use case.
➢ Business objectives:
It must be assessed what specific business problem the AI wants to solve or what
opportunity it seeks to exploit.
The implementation of the system must be directly linked to measurable strategic goals.
EXAMPLE: A bank that implements an AI model for fraud detection aims to reduce financial
losses and increase the security of transactions.
Clear objectives are essential for defining success metrics and justifying the business case.
➢ Performance requirements
Evaluate which metrics - accuracy, latency, scalability, reliability - define success for the
model.
A medical diagnostic system, for example, requires a very high level of precision, while a
chatbot for customer service can understand that the requirements for speed of response
are more critical.
Defining these requirements is vital for model selection, training and monitoring.
EXAMPLE: An algorithmic trading system in the financial market needs extremely low
latency, while a demand forecasting system can tolerate longer processing times.
When we see a drop in the model's performance, we say that we are dealing with a MODEL
DECAY - a gradual performance drop, typically driven by model - or data-drift, overfitting,
or underfitting.
o High Variance: The model over-learns the details of the training set, even
the noises and exceptions. In other words, it does very well on training, but
very badly on new data. The common cause is OVERFITTING - the model is
memorizing rather than actually learning general patterns.
o High Bias: The model doesn't learn well enough - it makes very simple
assumptions. In other words, it gets a lot wrong in training and with new
o Data Drift: When new data changes over time - it is no longer similar to the
data used for training and, statistically, there have been changes in users'
consumption/use patterns. Even a good model starts to make mistakes
because the context / reality has changed.
The lack of suitable data or the presence of bias in existing data can severely compromise
the performance and fairness of the model.
Data governance is crucial here, including the assessment of lineage and provenance.
➢ Workforce Readiness
Should the organization and its employees be prepared to adopt and interact with the new
AI system?
This question involves assessing the need for training, redefining roles and processes, and
managing the organization's cultural change. Resilience in adoption or a lack of
understanding of how to use the AI system correctly can undermine the expected benefits.
Open-source models have their source code publicly accessible, allowing inspection,
modification and distribution.
The choice has an impact on transparency, costs, flexibility, information security and third-
party risk management.
For example, open source models often improve transparency and allow on-prem
deployment, though they require stronger internal security oversight. However, their use
requires greater internal expertise for implementation and security, as well as specific
licenses which, in certain situations, need to be evaluated.
Large models, such as many LLMs, require robust infrastructure (often in the cloud) and
can be difficult to interpret (Explainability).
Smaller models can be more efficient for deployment on devices with limited resources -
Edge Computing.
Multimodal models can process and combine text, images, audio, or video - for example,
ChatGPT can evaluate an image and describe it in text.
Multimodality expands the possible applications, but also increases the complexity of data
management and risk assessment, as different types of biases and failures can arise for
each of these modalities.
How an AI model is technically deployed and made available for use is also a critical
decision with governance implications.
The choice depends on the use case - for example, an AI system for analyzing medical
images can be deployed on-premise for privacy, control and latency reasons, while a
recommendation system on an e-commerce website usually resides in the cloud for
scalability.
In general, these are ready-to-use systems, made by a company and sold as a product.
A kind of off-the-shelf software.
Custom-built systems, on the other hand, are those that are tailor-made to the specific
needs of the company. In this case, the organization needs to further refine / adapt the
system to its use case and purpose, also taking into account the peculiarities of its
operations, infrastructure and capacity. They are generally more expensive and time-
consuming to implement, considering customization.
➢ Use "as is" vs. Fine Tuning vs. RAG vs. Other Techniques
In addition to RAG and Fine Tuning, there are other techniques/forms for adapting or
integrating models, such as Prompt Engineering, Ensembles - discussed earlier, which
consists of putting several models together - etc.
The aim is to proactively identify, analyze and document the potential risks and
benefits associated with using the AI model in the specific context defined in the use
case. This assessment is not a one-off event, but part of an iterative risk management
process.
(i) Describe the system and the context: Detail how the model works, the data
used, the purpose (use case) and the deployment environment;
(ii) Identify risks and benefits: Map potential risks (legal, ethical, operational,
reputational, security and social) and expected benefits for the organization,
individuals and society;
(iii) Assess necessity and proportionality: Justify why the use of the AI system is
necessary and whether the risks are proportional to the benefits.
(vi) Document: Record the entire process, findings and decisions for compliance
and accountability purposes and future reviews.
In addition, it is possible to evaluate the types of assessments that can be carried out on
AI systems, both by external teams (specialized consultancies) and by the AI Governance
team itself:
➢ Data Usage: Focuses on what the system does with the data, such as collecting,
using, sharing, storing and protecting it.
o Example: Does the system use personal data? Is third-party data shared?
What categories of data are used in the system?
➢ Security and Safety: Evaluates whether the system is secure against attacks,
failures, misuse and produces safe and reliable results.
o Examples Are there security controls in place to protect against data leaks?
Can it generate offensive or discriminatory content?
It is important to identify and understand the complex regulatory landscape that applies to
the specific AI model and its use case. Legal compliance is a pillar of responsible
governance.
Thus, organizations should map not only the specific regulations/standards on Artificial
Intelligence (e.g. EU AI ACT and Japan AI Law), but also identify laws related to it:
➢ Sectoral regulations: Sectors such as finance and health (HIPPA) have specific
regulations;
➢ Specific challenges
o Transparency vs. Trade Secrets: Balancing the need for transparency for
regulators and end users with the protection of their intellectual property
and the trade secrets invested in developing the model - Transparency vs.
Intellectual Property. Thus, deciding what to document and disclose
becomes a complex strategic decision.
o Training data governance: The organization must ensure that the data
used to train the model has been obtained legally and ethically, especially
if it involves external sources (e.g. webscraping) or personal user/owner
data. Lies about the provenance of this data are a growing risk and several
legal debates are underway about this issue.
The transition of the model from the development/test environment to the production
environment (deployment) must be actively managed, applying the established governance
guidelines.
This includes ensuring that only necessary data is used - Data Minimization, that data
security is maintained and that privacy requirements are respected in the operating
environment.
In other words, the organization must define a performance baseline at the time of
implementation to compare with future measurements.
➢ Performance monitoring
Implement tools and processes to continuously track the defined metrics. This can involve
automated logs, visualization dashboards and alerts for significant deviations (Deviations,
Irregular decisions).
➢ Drift monitoring
Specifically monitoring changes in the statistical distributions of input and output data to
detect Data Drifts or Model Drifts. Specialized tools can be used for this.
Retraining may be necessary to adapt the model to new patterns in the data or to correct
performance degradation. The frequency will therefore depend on the volatility of the
environment - e.g. market prediction models may need to be retrained more frequently
than image recognition models, which are usually more stable.
➢ Versioning
Keep clear records of the different versions of the model and the data used to train them.
This is crucial for traceability and for reverting to previous versions if necessary.
➢ Prevention of unwanted secondary uses: The model has been designed for a
specific use case. Controls (technical and political) must be implemented to
prevent it from being used for unintended or inappropriate purposes, which could
generate new risks.