OCI Answers
OCI Answers
Which statement best describes the role of encoder and decoder models in natural language
processing?
A. Encoder models are used only for numerical calculations, whereas decoder models are
used to interpret the calculated numerical values back into text.
B. Encoder models convert a sequence of words into a vector representation, and decoder
models take this vector representation to generate a sequence of words.
C. Encoder models and decoder models both convert sequences of words into vector
representations without generating new text.
D. Encoder models take a sequence of words and predict the next word in the sequence,
whereas decoder models convert a sequence of words into a numerical representation.
Answer: 2
2. Which is the main characteristic of greedy decoding in the context of language model word
prediction?
Answer: 1
3. Which is NOT a category of pretrained foundational models available in the OCI Generative
AI service?
A. Translation models.
B. Generation models
C. Summarization models
D. Embedding models
Answer: 3
4. How are fine-tuned customer models stored to enable strong data privacy and security in the
OCI Generative AI service?
Answer: 2
5. You create a fine-tuning dedicated AI cluster to customize a foundational model with your
custom training data.
How many unit hours are required for fine-tuning if the cluster is active for 10 hours?
A. 20 unit hours
B. 30 unit hours
C. 40 unit hours
D. 25 unit hours
Answer: 1
6. How does the architecture of dedicated AI clusters contribute to minimizing GPU memory
overhead for T- Few fine-tuned model inference?
A. By loading the entire model into GPU memory for efficient processing
B. By sharing base model weights across multiple fine-tuned models on the same group of
GPUS
Answer: 2
Answer: 4
Answer: 3
9. Which technique involves prompting the Large Language Model (LLM) to emit intermediate
reasoning steps as part of its response?
A. Chain-of-Thought
B. In-context Learning
C. Step-Back Prompting
D. Least-to-most Prompting
Answer: 1
10. Analyze the user prompts provided to a language model. Which scenario exemplifies prompt
injection (jailbreaking)?
Answer: 1
11. What does "k-shot prompting" refer to when using Large Language Models for task-specific
applications?
A. Limiting the model to only k possible outcomes or answers for a given task
B. Providing the exact k words in the prompt to guide the model's response
C. Explicitly providing k examples of the intended task in the prompt to guide the model's
output
D. The process of training the model on k different tasks simultaneously to improve its
versatility
Answer: 3
12. Given the following prompts used with a Large Language Model, classify each as employing
the Chain-of- Thought, Least-to-most, or Step-Back prompting technique.
1. Calculate the total number of wheels needed for 3 cars. Cars have 4 wheels each. Then, use
the total number of wheels to
determine how many sets of wheels we can buy with $200 if one set (4 wheels) costs $50. 2.
Solve a complex math problem by first identifying the formula needed, and then solve a simpler
version of the problem before tackling the full question.
3. To understand the impact of greenhouse gases on climate change, let's start by defining what
greenhouse gases are. Next, we'll explore how they trap heat in the Earth's atmosphere.
A. 1: Least-to-most, 2: Chain-of-Thought, 3: Step-Back
Answer: 3
Answer: 4
Answer: 3
15. Which is NOT a built-in memory type in LangChain?
A. Conversation TokenBufferMemory
B. ConversationSummaryMemory
C. ConversationImageMemory
D. ConversationBufferMemory
Answer: 3
B. After user input but before chain execution, and again after core logic but before output
Answer: 2
17. Why is normalization of vectors important before indexing in a hybrid search system?
B. It standardizes vector lengths for meaningful comparison using metrics such as Cosine
Similarity.
Answer: 2
B. PEFT modifies all parameters and is typically used when no training data exists.
C. PEFT involves only a few or new parameters and uses labeled, task-specific data.
D. PEFT does not modify any parameters but uses soft prompting with unlabeled data.
Answer: 3
19. How does the utilization of T-Few transformer layers contribute to the efficiency of the
fine-tuning process?
Answer: 1
20. Which is a key characteristic of the annotation process used in T-Few fine-tuning?
C. T-Few fine-tuning involves updating the weights of all layers in the model.
Answer: 4
21. What issue might arise from using small data sets with the Vanilla fine-tuning method in the
OCI Generative AI service?
A. Overfitting
B. Underfitting
C. Data Leakage
D. Model Drift
Answer: 1
22. What does "Loss" measure in the evaluation of OCI Generative Al fine-tuned models?
A. The difference between the accuracy of the model at the beginning of training and the
accuracy of the deployed model
B. The percentage of incorrect predictions made by the model compared with the total
number of predictions in the evaluation
D. The level of incorrectness in the model's predictions, with lower values indicating better
performance
Answer: 4
23. When should you use the T-Few fine-tuning method for training a model?
Answer: 4
24. Which is a key advantage of using T-Few over Vanilla fine-tuning in the OCI Generative AI
service?
Considering the capabilities, which type of model would the company likely focus on integrating
into their AI assistant?
B. A Large Language Model based agent that focuses on generating textual responses
C. A Retrieval-Augmented Generation (RAG) model that uses text as input and output
Answer: 1
26. Which component of Retrieval-Augmented Generation (RAG) evaluates and prioritizes the
information retrieved by the retrieval system?
A. Retriever
B. Encoder-decoder
C. Ranker
D. Generator
Answer: 3
27. How does the Retrieval-Augmented Generation (RAG) Token technique differ from RAG
Sequence when generating a model's response?
A. RAG Token does not use document retrieval but generates responses based on
pre-existing knowledge only.
B. RAG Token retrieves relevant documents for each part of the response and constructs
the answer incrementally.
C. RAG Token retrieves documents only at the beginning of the response generation and
uses those for the entire content.
D. Unlike RAG Sequence, RAG Token generates the entire response at once without
considering individual parts.
Answer: 2
28. How do Dot Product and Cosine Distance differ in their application to comparing text
embeddings in natural language processing?
A. Dot Product is used for semantic analysis, whereas Cosine Distance is used for syntactic
comparisons.
B. Dot Product assesses the overall similarity in content, whereas Cosine Distance
measures topical relevance.
C. Dot Product measures the magnitude and direction of vectors, whereas Cosine Distance
focuses on the orientation regardless of magnitude.
D. Dot Product calculates the literal overlap of words, whereas Cosine Distance evaluates
the stylistic similarity
Answer: 3
29. Which is a cost-related benefit of using vector databases with Large Language Models
(LLMs)?
B. They offer real-time updated knowledge bases and are cheaper than fine-tuned LLMs.
D. They increase the cost due to the need for real-time updates.
Answer: 2
30. How does the integration of a vector database into Retrieval-Augmented Generation
(RAG)-based Large Language Models (LLMs) fundamentally alter their responses?
A. It enables them to bypass the need for pretraining on large text corpora.
D. It shifts the basis of their responses from pretrained internal knowledge to real-time data
retrieval.
Answer: 4
31. What is the purpose of the "stop sequence" parameter in the OCI Generative AI Generation
models?
A. It specifies a string that tells the model to stop generating more content.
B. It determines the maximum number of tokens the model can generate per response.
Answer: 1
32. What does a higher number assigned to a token signify in the "Show Likelihoods" feature of
the language model token generation?
B. The token is unrelated to the current token and will not be used.
C. The token will be the only one considered in the next generation step.
Answer: 4
33. What is the primary function of the "temperature" parameter in the OCI Generative AI
Generation models?
A. Determines the maximum number of tokens the model can generate per response
C. Specifies a string that tells the model to stop generating more content
D. Assigns a penalty to tokens that have already appeared in the preceding text
Answer: 2
34. What distinguishes the Cohere Embed v3 model from its predecessor in the OCI Generative
AI service?
Answer: 4
35. Which statement describes the difference between "Top k ^ prime prime and "Top p" in
selecting the next token in theOCI Generative AI Generation models?
A. "Top k" selects the next token based on its position in the list of probable tokens,
whereas "Top p" selects based on the cumulative probability of the top tokens.
B. "Top k" considers the sum of probabilities of the top tokens, whereas "Top p" selects
from the"Top k ^ prime prime tokens sorted by probability.
C. "Top k ^ prime prime and "Top p" are identical in their approach to token selection but
differ in theirapplication of penalties to tokens.
D. "Top k ^ prime prime and "Top P ^ n both select from the same set of tokens but use
different methods toprioritize them based on frequency.
Answer: 1
36. Which statement is true about the "Top p" parameter of the OCI Generative AI Generation
models?
A. "Top p" limits token selection based on the sum of their probabilities.
D. "Top p" selects tokens from the "Top k" tokens sorted by probability.
Answer: 1
37. Which Oracle Accelerated Data Science (ADS) class can be used to deploy a Large
Language Model (LLM) application to OCI Data Science model deployment?
A. RetrievalQA
B. GenerativeAl
C. ChainDeployment
D. TextLoader
Answer: 3
38. What does a dedicated RDMA cluster network do during model fine-tuning and inference?
A. It limits the number of fine-tuned models deployable on the same GPU cluster.
Answer: 2
39. Which role does a "model endpoint" serve in the inference workflow of the OCI Generative
AI service?
C. Updates the weights of the base model during the fine-tuning process
Answer: 1
40. In LangChain, which retriever search type is used to balance between relevancy and
diversity?
A. similarity_score_threshold
B. mmr
C. top k
D. similarity
Answer: 2