Ai Specialist Exam Prep
Ai Specialist Exam Prep
Examen
60 pregunta
105 mins
73% para aprobar
Contenido
Problems that can be faced with Generative AI due to it being LLM generated:
Prompt → Set of instructions to a LLM (Large Language Model) to generate a specific output or generated
response. The prompt can come from any of the CRM apps. You can create it on Prompt Builder and invoke it
from Apex or a Flow. This step is mandatory. For LLM to generate personalized email content, it needs
context about your customer, their preferences and other relevant data. Example of masked data:
Grounding → Process of adding additional context to the prompt, because the LLM needs additional context.
You can ground your prompt using merge fields with CRM data (ex: record fields, flows, Apex, Data Cloud
DMOs and related lists). It’s dynamic because happens at run time and depends on the user’s access.
Secure data retrieval First step in the Trust Layer. Means that the prompt is grounded only with data that
the executing user has access to, because data retrieval process respects existing access control and
permissions.
The Einstein Trust Layer uses data masking to prevent sensitive data from being exposed to the LLM (Large
Language Model). Data masking involves replacing the sensitive data with placeholder text based on what it
represents. We identify sensitive data using two methods:
Pattern-based: We use patterns and context to identify sensitive data in the prompt text.
Field-Based: We use the metadata in the fields that are classified using Shield Platform Encryption or
data classification to identify sensitive fields. Field-Based masking supports only merge fields that are
referenced in record merge fields and related lists.
Once identified, the data is then masked with a placeholder text to prevent the data from being exposed to
external models.
Einstein Trust Layer temporarily stores the relationship between the original entities and their respective
placeholders. The relationship is used later to demask the data in the generated response.
Prompt Defense refers to system policies that help limit hallucinations and decrease the likelihood of
harmful outputs. To help decrease the likelihood of the LLM generating something unintended or harmful,
Prompt Builder and Prompt Template Connect API use system policies (a set of instructions to the LLM for
how to behave in a certain manner).
Models built or fine-tuned by Salesforce are hosted in the Salesforce trust boundary. External Models built
and maintained by third-party providers, such as OpenAI, are in a shared trust boundary. Models that you
build and maintain are hosted on your infrastructure.
When the prompt is ready, it’s sent through the LLM gateway, which generates a response and sends it back
to Salesforce.
Zero-data retention policy in place with external partner model providers, such as OpenAI or Azure Open
AI. The policy states that data sent to the LLM from Salesforce isn’t retained and is deleted after a response is
sent back to Salesforce.
Next, a set of instructions and policies are applied to the prompt to reduce inaccuracies and prevent
unintended consequences.
But before sending it back to you, Einstein scans the response for any toxicity (toxicity detection) and then
de-masks the data (the placeholders that were created for masking the data during the prompt journey are
now replaced with the actual data that was behind them.
). The toxicity detection process includes a toxicity confidence score, which reflects the probability of the
response including harmful or inappropriate content.
Finally, you can edit further the response or send it as it is. Give feedback to improve the tool. Data and
feedback are stored in Data Cloud for 30 days (included the toxicity score). Einstein generative AI audit and
feedback data is stored in Data Cloud.
Audit Trail also includes the original prompt, masked prompt, scores logged during toxicity detection, the
original output from the LLM, and the demasked output.
Use reports in Data Cloud to check how Einstein Trust Layer is impacting and feedback to improve prompt
design.
1. From Setup, in the Quick Find box, enter Einstein, and then select Einstein Trust Layer.
2. Select Go to Einstein Trust Layer
3. Turn on large language model data masking.
4. Review the list of data types included in the pattern-based masking section and make changes as
needed. Some data types are turned on for data masking by default.
5. Turn on data masking for Shield Platform Encryption, compliance categories, and data sensitivity levels.
Confirm that the sensitive fields that must be masked are tagged with the correct compliance categories
and data sensitivity levels in Object Manager.
You see the Shield Platform Encryption option in Einstein Trust Layer setup only if you enabled it in your org.
When the isToxicityDetected field is true, it indicates a high level of confidence that the content contains
toxic language. However, when the isToxicityDetected field is false, it doesn’t necessarily mean there isn’t
toxicity, but rather, that the model didn’t detect toxicity in the content. The model is trained to provide
scores from 0 through 1.
The score for the safety category ranges from 0 through 1 with 1 being the safest. We consider a
safety score that is between 0.5 and 1 as safe.
The scores for all other categories indicate the toxicity in each category, and range from 0 through 1,
1 being the most toxic. We consider toxicity scores of 0.5 and above as toxic for that category.
Generative AI audit data (also known as audit trail) includes data about the Einstein Trust Layer features such
as data masking and toxicity scores.
Audit trail along with feedback data is stored in Data Cloud.
DMO: Data Model Object.
Request Tag: captures any custom data points as key-value pairs.
Gateway Request: captures prompt input, request parameters, and model details and parameters.
Gateway Response: enables joins between the GenAIRequest DMO and GenAIGenerations DMO.
Content Category DMO: captures Einstein Trust Layer detector result values by detector type and sub-
category. This DMO includes the safety and toxicity scores of output from the LLM.
Content Quality DMO: captures if a request or response is safe or unsafe.
Generation: captures generated responses returned for generation requests. This DMO also includes
masked prompts if masking is enabled and may also include sensitive user data.
Feedback Detail DMO: details of user feedback.
Feedback: captures feedback for a specific generation or part of a generation.
App Generation: captures feature-specific changes made to the original generated text.