Spring Sale 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: best70

Databricks-Generative-AI-Engineer-Associate Databricks Certified Generative AI Engineer Associate Questions and Answers

Questions 4

A small and cost-conscious startup in the cancer research field wants to build a RAG application using Foundation Model APIs.

Which strategy would allow the startup to build a good-quality RAG application while being cost-conscious and able to cater to customer needs?

Options:

A.

Limit the number of relevant documents available for the RAG application to retrieve from

B.

Pick a smaller LLM that is domain-specific

C.

Limit the number of queries a customer can send per day

D.

Use the largest LLM possible because that gives the best performance for any general queries

Buy Now
Questions 5

Generative AI Engineer at an electronics company just deployed a RAG application for customers to ask questions about products that the company carries. However, they received feedback that the RAG response often returns information about an irrelevant product.

What can the engineer do to improve the relevance of the RAG’s response?

Options:

A.

Assess the quality of the retrieved context

B.

Implement caching for frequently asked questions

C.

Use a different LLM to improve the generated response

D.

Use a different semantic similarity search algorithm

Buy Now
Questions 6

Which indicator should be considered to evaluate the safety of the LLM outputs when qualitatively assessing LLM responses for a translation use case?

Options:

A.

The ability to generate responses in code

B.

The similarity to the previous language

C.

The latency of the response and the length of text generated

D.

The accuracy and relevance of the responses

Buy Now
Questions 7

A Generative AI Engineer just deployed an LLM application at a digital marketing company that assists with answering customer service inquiries.

Which metric should they monitor for their customer service LLM application in production?

Options:

A.

Number of customer inquiries processed per unit of time

B.

Energy usage per query

C.

Final perplexity scores for the training of the model

D.

HuggingFace Leaderboard values for the base LLM

Buy Now
Questions 8

A Generative AI Engineer is creating an agent-based LLM system for their favorite monster truck team. The system can answer text based questions about the monster truck team, lookup event dates via an API call, or query tables on the team’s latest standings.

How could the Generative AI Engineer best design these capabilities into their system?

Options:

A.

Ingest PDF documents about the monster truck team into a vector store and query it in a RAG architecture.

B.

Write a system prompt for the agent listing available tools and bundle it into an agent system that runs a number of calls to solve a query.

C.

Instruct the LLM to respond with “RAG”, “API”, or “TABLE” depending on the query, then use text parsing and conditional statements to resolve the query.

D.

Build a system prompt with all possible event dates and table information in the system prompt. Use a RAG architecture to lookup generic text questions and otherwise leverage the information in the system prompt.

Buy Now
Questions 9

What is an effective method to preprocess prompts using custom code before sending them to an LLM?

Options:

A.

Directly modify the LLM’s internal architecture to include preprocessing steps

B.

It is better not to introduce custom code to preprocess prompts as the LLM has not been trained with examples of the preprocessed prompts

C.

Rather than preprocessing prompts, it’s more effective to postprocess the LLM outputs to align the outputs to desired outcomes

D.

Write a MLflow PyFunc model that has a separate function to process the prompts

Buy Now
Questions 10

A Generative Al Engineer is developing a RAG application and would like to experiment with different embedding models to improve the application performance.

Which strategy for picking an embedding model should they choose?

Options:

A.

Pick an embedding model trained on related domain knowledge

B.

Pick the most recent and most performant open LLM released at the time

C.

pick the embedding model ranked highest on the Massive Text Embedding Benchmark (MTEB) leaderboard hosted by HuggingFace

D.

Pick an embedding model with multilingual support to support potential multilingual user questions

Buy Now
Questions 11

A Generative Al Engineer is tasked with developing an application that is based on an open source large language model (LLM). They need a foundation LLM with a large context window.

Which model fits this need?

Options:

A.

DistilBERT

B.

MPT-30B

C.

Llama2-70B

D.

DBRX

Buy Now
Questions 12

A Generative Al Engineer is building a production-ready LLM system which replies directly to customers. The solution makes use of the Foundation Model API via provisioned throughput. They are concerned that the LLM could potentially respond in a toxic or otherwise unsafe way. They also wish to perform this with the least amount of effort.

Which approach will do this?

Options:

A.

Host Llama Guard on Foundation Model API and use it to detect unsafe responses

B.

Add some LLM calls to their chain to detect unsafe content before returning text

C.

Add a regex expression on inputs and outputs to detect unsafe responses.

D.

Ask users to report unsafe responses

Buy Now
Questions 13

A Generative AI Engineer I using the code below to test setting up a vector store:

Databricks-Generative-AI-Engineer-Associate Question 13

Assuming they intend to use Databricks managed embeddings with the default embedding model, what should be the next logical function call?

Options:

A.

vsc.get_index()

B.

vsc.create_delta_sync_index()

C.

vsc.create_direct_access_index()

D.

vsc.similarity_search()

Buy Now
Questions 14

Which TWO chain components are required for building a basic LLM-enabled chat application that includes conversational capabilities, knowledge retrieval, and contextual memory?

Options:

A.

(Q)

B.

Vector Stores

C.

Conversation Buffer Memory

D.

External tools

E.

Chat loaders

F.

React Components

Buy Now
Questions 15

A Generative Al Engineer needs to design an LLM pipeline to conduct multi-stage reasoning that leverages external tools. To be effective at this, the LLM will need to plan and adapt actions while performing complex reasoning tasks.

Which approach will do this?

Options:

A.

Tram the LLM to generate a single, comprehensive response without interacting with any external tools, relying solely on its pre-trained knowledge.

B.

Implement a framework like ReAct which allows the LLM to generate reasoning traces and perform task-specific actions that leverage external tools if necessary.

C.

Encourage the LLM to make multiple API calls in sequence without planning or structuring the calls, allowing the LLM to decide when and how to use external tools spontaneously.

D.

Use a Chain-of-Thought (CoT) prompting technique to guide the LLM through a series of reasoning steps, then manually input the results from external tools for the final answer.

Buy Now
Questions 16

A Generative Al Engineer is building an LLM-based application that has an

important transcription (speech-to-text) task. Speed is essential for the success of the application

Which open Generative Al models should be used?

Options:

A.

L!ama-2-70b-chat-hf

B.

MPT-30B-lnstruct

C.

DBRX

D.

whisper-large-v3 (1.6B)

Buy Now
Questions 17

A team uses Mosaic AI Vector Search to retrieve documents for their Retrieval-Augmented Generation (RAG) pipeline. The search query returns five relevant documents, and the first three are added to the prompt as context. Performance evaluation with Agent Evaluation shows that some lower-ranked retrieved documents have higher context relevancy scores than higher-ranked documents. Which option should the team consider to optimize this workflow?

Options:

A.

Use a reranker to order the documents based on the relevance scores.

B.

Modify the prompt to instruct the LLM to order the documents based on the relevance scores.

C.

Use a different embedding model for computing document embeddings.

D.

Increase the number of documents added to the prompt to improve context relevance.

Buy Now
Questions 18

A Generative Al Engineer is working with a retail company that wants to enhance its customer experience by automatically handling common customer inquiries. They are working on an LLM-powered Al solution that should improve response times while maintaining a personalized interaction. They want to define the appropriate input and LLM task to do this.

Which input/output pair will do this?

Options:

A.

Input: Customer reviews; Output Group the reviews by users and aggregate per-user average rating, then respond

B.

Input: Customer service chat logs; Output Group the chat logs by users, followed by summarizing each user's interactions, then respond

C.

Input: Customer service chat logs; Output: Find the answers to similar questions and respond with a summary

D.

Input: Customer reviews: Output Classify review sentiment

Buy Now
Questions 19

A Generative AI Engineer is designing a chatbot for a gaming company that aims to engage users on its platform while its users play online video games.

Which metric would help them increase user engagement and retention for their platform?

Options:

A.

Randomness

B.

Diversity of responses

C.

Lack of relevance

D.

Repetition of responses

Buy Now
Questions 20

A Generative AI Engineer is designing a RAG application for answering user questions on technical regulations as they learn a new sport.

What are the steps needed to build this RAG application and deploy it?

Options:

A.

Ingest documents from a source –> Index the documents and saves to Vector Search –> User submits queries against an LLM –> LLM retrieves relevant documents –> Evaluate model –> LLM generates a response –> Deploy it using Model Serving

B.

Ingest documents from a source –> Index the documents and save to Vector Search –> User submits queries against an LLM –> LLM retrieves relevant documents –> LLM generates a response -> Evaluate model –> Deploy it using Model Serving

C.

Ingest documents from a source –> Index the documents and save to Vector Search –> Evaluate model –> Deploy it using Model Serving

D.

User submits queries against an LLM –> Ingest documents from a source –> Index the documents and save to Vector Search –> LLM retrieves relevant documents –> LLM generates a response –> Evaluate model –> Deploy it using Model Serving

Buy Now
Questions 21

A Generative AI Engineer has a provisioned throughput model serving endpoint as part of a RAG application and would like to monitor the serving endpoint’s incoming requests and outgoing responses. The current approach is to include a micro-service in between the endpoint and the user interface to write logs to a remote server.

Which Databricks feature should they use instead which will perform the same task?

Options:

A.

Vector Search

B.

Lakeview

C.

DBSQL

D.

Inference Tables

Buy Now
Exam Name: Databricks Certified Generative AI Engineer Associate
Last Update: Mar 5, 2026
Questions: 73

PDF + Testing Engine

$134.99

Testing Engine

$99.99

PDF (Q&A)

$84.99