Spring Sale 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: exams65

ExamsBrite Dumps

Databricks Certified Generative AI Engineer Associate Question and Answers

Databricks Certified Generative AI Engineer Associate

Last Update Feb 28, 2026
Total Questions : 73

We are offering FREE Databricks-Generative-AI-Engineer-Associate Databricks exam questions. All you do is to just go and sign up. Give your details, prepare Databricks-Generative-AI-Engineer-Associate free exam questions and then go for complete pool of Databricks Certified Generative AI Engineer Associate test questions that will help you more.

Databricks-Generative-AI-Engineer-Associate pdf

Databricks-Generative-AI-Engineer-Associate PDF

$36.75  $104.99
Databricks-Generative-AI-Engineer-Associate Engine

Databricks-Generative-AI-Engineer-Associate Testing Engine

$43.75  $124.99
Databricks-Generative-AI-Engineer-Associate PDF + Engine

Databricks-Generative-AI-Engineer-Associate PDF + Testing Engine

$57.75  $164.99
Questions 1

A Generative Al Engineer is developing a RAG system for their company to perform internal document Q&A for structured HR policies, but the answers returned are frequently incomplete and unstructured It seems that the retriever is not returning all relevant context The Generative Al Engineer has experimented with different embedding and response generating LLMs but that did not improve results.

Which TWO options could be used to improve the response quality?

Choose 2 answers

Options:

A.  

Add the section header as a prefix to chunks

B.  

Increase the document chunk size

C.  

Split the document by sentence

D.  

Use a larger embedding model

E.  

Fine tune the response generation model

Discussion 0
Questions 2

A Generative AI Engineer is building a Generative AI system that suggests the best matched employee team member to newly scoped projects. The team member is selected from a very large team. The match should be based upon project date availability and how well their employee profile matches the project scope. Both the employee profile and project scope are unstructured text.

How should the Generative Al Engineer architect their system?

Options:

A.  

Create a tool for finding available team members given project dates. Embed all project scopes into a vector store, perform a retrieval using team member profiles to find the best team member.

B.  

Create a tool for finding team member availability given project dates, and another tool that uses an LLM to extract keywords from project scopes. Iterate through available team members’ profiles and perform keyword matching to find the best available team member.

C.  

Create a tool to find available team members given project dates. Create a second tool that can calculate a similarity score for a combination of team member profile and the project scope. Iterate through the team members and rank by best score to select a team member.

D.  

Create a tool for finding available team members given project dates. Embed team profiles into a vector store and use the project scope and filtering to perform retrieval to find the available best matched team members.

Discussion 0
Questions 3

A team uses Mosaic AI Vector Search to retrieve documents for their Retrieval-Augmented Generation (RAG) pipeline. The search query returns five relevant documents, and the first three are added to the prompt as context. Performance evaluation with Agent Evaluation shows that some lower-ranked retrieved documents have higher context relevancy scores than higher-ranked documents. Which option should the team consider to optimize this workflow?

Options:

A.  

Use a reranker to order the documents based on the relevance scores.

B.  

Modify the prompt to instruct the LLM to order the documents based on the relevance scores.

C.  

Use a different embedding model for computing document embeddings.

D.  

Increase the number of documents added to the prompt to improve context relevance.

Discussion 0
Questions 4

A team wants to serve a code generation model as an assistant for their software developers. It should support multiple programming languages. Quality is the primary objective.

Which of the Databricks Foundation Model APIs, or models available in the Marketplace, would be the best fit?

Options:

A.  

Llama2-70b

B.  

BGE-large

C.  

MPT-7b

D.  

CodeLlama-34B

Discussion 0
Questions 5

A Generative Al Engineer has successfully ingested unstructured documents and chunked them by document sections. They would like to store the chunks in a Vector Search index. The current format of the dataframe has two columns: (i) original document file name (ii) an array of text chunks for each document.

What is the most performant way to store this dataframe?

Options:

A.  

Split the data into train and test set, create a unique identifier for each document, then save to a Delta table

B.  

Flatten the dataframe to one chunk per row, create a unique identifier for each row, and save to a Delta table

C.  

First create a unique identifier for each document, then save to a Delta table

D.  

Store each chunk as an independent JSON file in Unity Catalog Volume. For each JSON file, the key is the document section name and the value is the array of text chunks for that section

Discussion 0
Questions 6

A small and cost-conscious startup in the cancer research field wants to build a RAG application using Foundation Model APIs.

Which strategy would allow the startup to build a good-quality RAG application while being cost-conscious and able to cater to customer needs?

Options:

A.  

Limit the number of relevant documents available for the RAG application to retrieve from

B.  

Pick a smaller LLM that is domain-specific

C.  

Limit the number of queries a customer can send per day

D.  

Use the largest LLM possible because that gives the best performance for any general queries

Discussion 0
Questions 7

A Generative AI Engineer is developing an agent system using a popular agent-authoring library. The agent comprises multiple parallel and sequential chains. The engineer encounters challenges as the agent fails at one of the steps, making it difficult to debug the root cause. They need to find an appropriate approach to research this issue and discover the cause of failure. Which approach do they choose?

Options:

A.  

Enable MLflow tracing to gain visibility into each agent's behavior and execution step.

B.  

Run MLflow.evaluate to determine root cause of failed step.

C.  

Implement structured logging within the agent's code to capture detailed execution information.

D.  

Deconstruct the agent into independent steps to simplify debugging.

Discussion 0
Questions 8

An AI developer team wants to fine-tune an open-weight model to have exceptional performance on a code generation use case. They are trying to choose the best model to start with. They want to minimize model hosting costs and are using Hugging Face model cards and spaces to explore models. Which TWO model attributes and metrics should the team focus on to make their selection?

Options:

A.  

Big Code Models Leaderboard

B.  

Number of model parameters

C.  

MTEB Leaderboard

D.  

Chatbot Arena Leaderboard

E.  

Number of model downloads last month

Discussion 0
Questions 9

A Generative Al Engineer is using an LLM to classify species of edible mushrooms based on text descriptions of certain features. The model is returning accurate responses in testing and the Generative Al Engineer is confident they have the correct list of possible labels, but the output frequently contains additional reasoning in the answer when the Generative Al Engineer only wants to return the label with no additional text.

Which action should they take to elicit the desired behavior from this LLM?

Options:

A.  

Use few snot prompting to instruct the model on expected output format

B.  

Use zero shot prompting to instruct the model on expected output format

C.  

Use zero shot chain-of-thought prompting to prevent a verbose output format

D.  

Use a system prompt to instruct the model to be succinct in its answer

Discussion 0
Questions 10

A Generative AI Engineer has been reviewing issues with their company's LLM-based question-answering assistant and has determined that a technique called prompt chaining could help alleviate some performance concerns. However, to suggest this to their team, they have to clearly explain how it works and how it can benefit their question-answering assistant. Which explanation do they communicate to the team?

Options:

A.  

It allows you to break down complex tasks into multiple independent subtasks. This enables the assistant to generate more comprehensive and accurate responses.

B.  

It allows you to reduce the latency of your applications. By having multiple chains participating in the response as a chain, you increase the rate at which the response is generated.

C.  

It allows you to decrease the effort involved in crafting a prompt. Chains make it possible to reuse prompt text across multiple different use cases.

D.  

It reduces the average cost of a typical request. Chains make more efficient use of the tokens produced to generate higher quality responses with fewer tokens.

Discussion 0
Questions 11

A Generative Al Engineer is ready to deploy an LLM application written using Foundation Model APIs. They want to follow security best practices for production scenarios

Which authentication method should they choose?

Options:

A.  

Use an access token belonging to service principals

B.  

Use a frequently rotated access token belonging to either a workspace user or a service principal

C.  

Use OAuth machine-to-machine authentication

D.  

Use an access token belonging to any workspace user

Discussion 0
Questions 12

A Generative AI Engineer received the following business requirements for an external chatbot.

The chatbot needs to know what types of questions the user asks and routes to appropriate models to answer the questions. For example, the user might ask about upcoming event details. Another user might ask about purchasing tickets for a particular event.

What is an ideal workflow for such a chatbot?

Options:

A.  

The chatbot should only look at previous event information

B.  

There should be two different chatbots handling different types of user queries.

C.  

The chatbot should be implemented as a multi-step LLM workflow. First, identify the type of question asked, then route the question to the appropriate model. If it’s an upcoming event question, send the query to a text-to-SQL model. If it’s about ticket purchasing, the customer should be redirected to a payment platform.

D.  

The chatbot should only process payments

Discussion 0
Questions 13

A Generative AI Engineer is developing a chatbot designed to assist users with insurance-related queries. The chatbot is built on a large language model (LLM) and is conversational. However, to maintain the chatbot’s focus and to comply with company policy, it must not provide responses to questions about politics. Instead, when presented with political inquiries, the chatbot should respond with a standard message:

“Sorry, I cannot answer that. I am a chatbot that can only answer questions around insurance.”

Which framework type should be implemented to solve this?

Options:

A.  

Safety Guardrail

B.  

Security Guardrail

C.  

Contextual Guardrail

D.  

Compliance Guardrail

Discussion 0
Questions 14

A Generative Al Engineer would like an LLM to generate formatted JSON from emails. This will require parsing and extracting the following information: order ID, date, and sender email. Here’s a sample email:

They will need to write a prompt that will extract the relevant information in JSON format with the highest level of output accuracy.

Which prompt will do that?

Options:

A.  

You will receive customer emails and need to extract date, sender email, and order ID. You should return the date, sender email, and order ID information in JSON format.

B.  

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in JSON format.

Here’s an example: {“date”: “April 16, 2024”, “sender_email”: “sarah.lee925@gmail.com”, “order_id”: “RE987D”}

C.  

You will receive customer emails and need to extract date, sender email, and order ID. Return the extracted information in a human-readable format.

D.  

You will receive customer emails and need to extract date, sender email, and order I

D.  

Return the extracted information in JSON format.

Discussion 0
Questions 15

Generative AI Engineer at an electronics company just deployed a RAG application for customers to ask questions about products that the company carries. However, they received feedback that the RAG response often returns information about an irrelevant product.

What can the engineer do to improve the relevance of the RAG’s response?

Options:

A.  

Assess the quality of the retrieved context

B.  

Implement caching for frequently asked questions

C.  

Use a different LLM to improve the generated response

D.  

Use a different semantic similarity search algorithm

Discussion 0
Questions 16

A Generative AI Engineer is tasked with deploying an application that takes advantage of a custom MLflow Pyfunc model to return some interim results.

How should they configure the endpoint to pass the secrets and credentials?

Options:

A.  

Use spark.conf.set ()

B.  

Pass variables using the Databricks Feature Store API

C.  

Add credentials using environment variables

D.  

Pass the secrets in plain text

Discussion 0
Questions 17

A Generative Al Engineer interfaces with an LLM with prompt/response behavior that has been trained on customer calls inquiring about product availability. The LLM is designed to output “In Stock” if the product is available or only the term “Out of Stock” if not.

Which prompt will work to allow the engineer to respond to call classification labels correctly?

Options:

A.  

Respond with “In Stock” if the customer asks for a product.

B.  

You will be given a customer call transcript where the customer asks about product availability. The outputs are either “In Stock” or “Out of Stock”. Format the output in JSON, for example: {“call_id”: “123”, “label”: “In Stock”}.

C.  

Respond with “Out of Stock” if the customer asks for a product.

D.  

You will be given a customer call transcript where the customer inquires about product availability. Respond with “In Stock” if the product is available or “Out of Stock” if not.

Discussion 0
Questions 18

What is an effective method to preprocess prompts using custom code before sending them to an LLM?

Options:

A.  

Directly modify the LLM’s internal architecture to include preprocessing steps

B.  

It is better not to introduce custom code to preprocess prompts as the LLM has not been trained with examples of the preprocessed prompts

C.  

Rather than preprocessing prompts, it’s more effective to postprocess the LLM outputs to align the outputs to desired outcomes

D.  

Write a MLflow PyFunc model that has a separate function to process the prompts

Discussion 0
Questions 19

A Generative Al Engineer needs to design an LLM pipeline to conduct multi-stage reasoning that leverages external tools. To be effective at this, the LLM will need to plan and adapt actions while performing complex reasoning tasks.

Which approach will do this?

Options:

A.  

Tram the LLM to generate a single, comprehensive response without interacting with any external tools, relying solely on its pre-trained knowledge.

B.  

Implement a framework like ReAct which allows the LLM to generate reasoning traces and perform task-specific actions that leverage external tools if necessary.

C.  

Encourage the LLM to make multiple API calls in sequence without planning or structuring the calls, allowing the LLM to decide when and how to use external tools spontaneously.

D.  

Use a Chain-of-Thought (CoT) prompting technique to guide the LLM through a series of reasoning steps, then manually input the results from external tools for the final answer.

Discussion 0
Questions 20

A generative AI engineer is deploying an AI agent authored with MLflow’s ChatAgent interface for a retail company's customer support system on Databricks. The agent must handle thousands of inquiries daily, and the engineer needs to track its performance and quality in real-time to ensure it meets service-level agreements. Which metrics are automatically captured by default and made available for monitoring when the agent is deployed using the Mosaic AI Agent Framework?

Options:

A.  

Operational metrics like request volume, latency, and errors

B.  

Quality metrics like correctness and guideline adherence

C.  

Both operational and quality metrics

D.  

No metrics are automatically captured

Discussion 0
Questions 21

A Generative AI Engineer is testing a simple prompt template in LangChain using the code below, but is getting an error:

Python

from langchain.chains import LLMChain

from langchain_community.llms import OpenAI

from langchain_core.prompts import PromptTemplate

prompt_template = "Tell me a {adjective} joke"

prompt = PromptTemplate(input_variables=["adjective"], template=prompt_template)

# ... (Error-prone section)

Assuming the API key was properly defined, what change does the Generative AI Engineer need to make to fix their chain?

Options:

A.  

(Incorrect structure)

B.  

(Incorrect structure)

C.  

prompt_template = "Tell me a {adjective} joke"

prompt = PromptTemplate(input_variables=["adjective"], template=prompt_template)

llm = OpenAI()

llm_chain = LLMChain(prompt=prompt, llm=llm)

llm_chain.generate([{"adjective": "funny"}])

D.  

(Incorrect structure)

Discussion 0