Weekend Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: netbudy65

1z0-1127-25 Oracle Cloud Infrastructure 2025 Generative AI Professional Questions and Answers

Questions 4

Which statement is true about the "Top p" parameter of the OCI Generative AI Generation models?

Options:

A.

"Top p" selects tokens from the "Top k" tokens sorted by probability.

B.

"Top p" assigns penalties to frequently occurring tokens.

C.

"Top p" limits token selection based on the sum of their probabilities.

D.

"Top p" determines the maximum number of tokens per response.

Buy Now
Questions 5

Which statement is true about Fine-tuning and Parameter-Efficient Fine-Tuning (PEFT)?

Options:

A.

Fine-tuning requires training the entire model on new data, often leading to substantial computational costs, whereas PEFT involves updating only a small subset of parameters, minimizing computational requirements and data needs.

B.

PEFT requires replacing the entire model architecture with a new one designed specifically for the new task, making it significantly more data-intensive than Fine-tuning.

C.

Both Fine-tuning and PEFT require the model to be trained from scratch on new data, making them equally data and computationally intensive.

D.

Fine-tuning and PEFT do not involve model modification; they differ only in the type of data used for training, with Fine-tuning requiring labeled data and PEFT using unlabeled data.

Buy Now
Questions 6

How does the structure of vector databases differ from traditional relational databases?

Options:

A.

It stores data in a linear or tabular format.

B.

It is not optimized for high-dimensional spaces.

C.

It uses simple row-based data storage.

D.

It is based on distances and similarities in a vector space.

Buy Now
Questions 7

What does a higher number assigned to a token signify in the "Show Likelihoods" feature of the language model token generation?

Options:

A.

The token is less likely to follow the current token.

B.

The token is more likely to follow the current token.

C.

The token is unrelated to the current token and will not be used.

D.

The token will be the only one considered in the next generation step.

Buy Now
Questions 8

What happens if a period (.) is used as a stop sequence in text generation?

Options:

A.

The model ignores periods and continues generating text until it reaches the token limit.

B.

The model generates additional sentences to complete the paragraph.

C.

The model stops generating text after it reaches the end of the current paragraph.

D.

The model stops generating text after it reaches the end of the first sentence, even if the token limit is much higher.

Buy Now
Questions 9

Why is normalization of vectors important before indexing in a hybrid search system?

Options:

A.

It ensures that all vectors represent keywords only.

B.

It significantly reduces the size of the database.

C.

It standardizes vector lengths for meaningful comparison using metrics such as Cosine Similarity.

D.

It converts all sparse vectors to dense vectors.

Buy Now
Questions 10

How are prompt templates typically designed for language models?

Options:

A.

As complex algorithms that require manual compilation

B.

As predefined recipes that guide the generation of language model prompts

C.

To be used without any modification or customization

D.

To work only with numerical data instead of textual content

Buy Now
Questions 11

What is LCEL in the context of LangChain Chains?

Options:

A.

A programming language used to write documentation for LangChain

B.

A legacy method for creating chains in LangChain

C.

A declarative way to compose chains together using LangChain Expression Language

D.

An older Python library for building Large Language Models

Buy Now
Questions 12

What is the purpose of Retrieval Augmented Generation (RAG) in text generation?

Options:

A.

To generate text based only on the model's internal knowledge without external data

B.

To generate text using extra information obtained from an external data source

C.

To store text in an external database without using it for generation

D.

To retrieve text from an external source and present it without any modifications

Buy Now
Questions 13

How does the temperature setting in a decoding algorithm influence the probability distribution over the vocabulary?

Options:

A.

Increasing the temperature removes the impact of the most likely word.

B.

Decreasing the temperature broadens the distribution, making less likely words more probable.

C.

Increasing the temperature flattens the distribution, allowing for more varied word choices.

D.

Temperature has no effect on probability distribution; it only changes the speed of decoding.

Buy Now
Questions 14

Which is a key characteristic of Large Language Models (LLMs) without Retrieval Augmented Generation (RAG)?

Options:

A.

They always use an external database for generating responses.

B.

They rely on internal knowledge learned during pretraining on a large text corpus.

C.

They cannot generate responses without fine-tuning.

D.

They use vector databases exclusively to produce answers.

Buy Now
Questions 15

Which is a characteristic of T-Few fine-tuning for Large Language Models (LLMs)?

Options:

A.

It updates all the weights of the model uniformly.

B.

It does not update any weights but restructures the model architecture.

C.

It selectively updates only a fraction of the model’s weights.

D.

It increases the training time as compared to Vanilla fine-tuning.

Buy Now
Questions 16

How does a presence penalty function in language model generation?

Options:

A.

It penalizes all tokens equally, regardless of how often they have appeared.

B.

It penalizes only tokens that have never appeared in the text before.

C.

It applies a penalty only if the token has appeared more than twice.

D.

It penalizes a token each time it appears after the first occurrence.

Buy Now
Questions 17

What is the function of the Generator in a text generation system?

Options:

A.

To collect user queries and convert them into database search terms

B.

To rank the information based on its relevance to the user's query

C.

To generate human-like text using the information retrieved and ranked, along with the user's original query

D.

To store the generated responses for future use

Buy Now
Questions 18

When does a chain typically interact with memory in a run within the LangChain framework?

Options:

A.

Only after the output has been generated.

B.

Before user input and after chain execution.

C.

After user input but before chain execution, and again after core logic but before output.

D.

Continuously throughout the entire chain execution process.

Buy Now
Questions 19

How does the utilization of T-Few transformer layers contribute to the efficiency of the fine-tuning process?

Options:

A.

By incorporating additional layers to the base model

B.

By allowing updates across all layers of the model

C.

By excluding transformer layers from the fine-tuning process entirely

D.

By restricting updates to only a specific group of transformer layers

Buy Now
Questions 20

What does in-context learning in Large Language Models involve?

Options:

A.

Pretraining the model on a specific domain

B.

Training the model using reinforcement learning

C.

Conditioning the model with task-specific instructions or demonstrations

D.

Adding more layers to the model

Buy Now
Questions 21

What do prompt templates use for templating in language model applications?

Options:

A.

Python's list comprehension syntax

B.

Python's str.format syntax

C.

Python's lambda functions

D.

Python's class and object structures

Buy Now
Questions 22

You create a fine-tuning dedicated AI cluster to customize a foundational model with your custom training data. How many unit hours are required for fine-tuning if the cluster is active for 10 days?

Options:

A.

480 unit hours

B.

240 unit hours

C.

744 unit hours

D.

20 unit hours

Buy Now
Questions 23

How can the concept of "Groundedness" differ from "Answer Relevance" in the context of Retrieval Augmented Generation (RAG)?

Options:

A.

Groundedness pertains to factual correctness, whereas Answer Relevance concerns query relevance.

B.

Groundedness refers to contextual alignment, whereas Answer Relevance deals with syntactic accuracy.

C.

Groundedness measures relevance to the user query, whereas Answer Relevance evaluates data integrity.

D.

Groundedness focuses on data integrity, whereas Answer Relevance emphasizes lexical diversity.

Buy Now
Questions 24

How does a presence penalty function in language model generation when using OCI Generative AI service?

Options:

A.

It penalizes all tokens equally, regardless of how often they have appeared.

B.

It only penalizes tokens that have never appeared in the text before.

C.

It applies a penalty only if the token has appeared more than twice.

D.

It penalizes a token each time it appears after the first occurrence.

Buy Now
Questions 25

An AI development company is working on an advanced AI assistant capable of handling queries in a seamless manner. Their goal is to create an assistant that can analyze images provided by users and generate descriptive text, as well as take text descriptions and produce accurate visual representations. Considering the capabilities, which type of model would the company likely focus on integrating into their AI assistant?

Options:

A.

A diffusion model that specializes in producing complex outputs.

B.

A Large Language Model-based agent that focuses on generating textual responses

C.

A language model that operates on a token-by-token output basis

D.

A Retrieval Augmented Generation (RAG) model that uses text as input and output

Buy Now
Questions 26

Which is a cost-related benefit of using vector databases with Large Language Models (LLMs)?

Options:

A.

They require frequent manual updates, which increase operational costs.

B.

They offer real-time updated knowledge bases and are cheaper than fine-tuned LLMs.

C.

They increase the cost due to the need for real-time updates.

D.

They are more expensive but provide higher quality data.

Buy Now
Exam Code: 1z0-1127-25
Exam Name: Oracle Cloud Infrastructure 2025 Generative AI Professional
Last Update: Oct 15, 2025
Questions: 88

PDF + Testing Engine

$140

Testing Engine

$105

PDF (Q&A)

$90