Summer Special Discount 60% Offer - Ends in 0d 00h 00m 00s - Coupon code: brite60

ExamsBrite Dumps

Databricks Certified Generative AI Engineer Associate Question and Answers

Databricks Certified Generative AI Engineer Associate

Last Update Jul 15, 2025
Total Questions : 61

We are offering FREE Databricks-Generative-AI-Engineer-Associate Databricks exam questions. All you do is to just go and sign up. Give your details, prepare Databricks-Generative-AI-Engineer-Associate free exam questions and then go for complete pool of Databricks Certified Generative AI Engineer Associate test questions that will help you more.

Databricks-Generative-AI-Engineer-Associate pdf

Databricks-Generative-AI-Engineer-Associate PDF

$42  $104.99
Databricks-Generative-AI-Engineer-Associate Engine

Databricks-Generative-AI-Engineer-Associate Testing Engine

$50  $124.99
Databricks-Generative-AI-Engineer-Associate PDF + Engine

Databricks-Generative-AI-Engineer-Associate PDF + Testing Engine

$66  $164.99
Questions 1

What is the most suitable library for building a multi-step LLM-based workflow?

Options:

A.  

Pandas

B.  

TensorFlow

C.  

PySpark

D.  

LangChain

Discussion 0
Questions 2

A Generative Al Engineer is tasked with improving the RAG quality by addressing its inflammatory outputs.

Which action would be most effective in mitigating the problem of offensive text outputs?

Options:

A.  

Increase the frequency of upstream data updates

B.  

Inform the user of the expected RAG behavior

C.  

Restrict access to the data sources to a limited number of users

D.  

Curate upstream data properly that includes manual review before it is fed into the RAG system

Discussion 0
Questions 3

A Generative Al Engineer is building a RAG application that answers questions about internal documents for the company SnoPen AI.

The source documents may contain a significant amount of irrelevant content, such as advertisements, sports news, or entertainment news, or content about other companies.

Which approach is advisable when building a RAG application to achieve this goal of filtering irrelevant information?

Options:

A.  

Keep all articles because the RAG application needs to understand non-company content to avoid answering questions about them.

B.  

Include in the system prompt that any information it sees will be about SnoPenAI, even if no data filtering is performed.

C.  

Include in the system prompt that the application is not supposed to answer any questions unrelated to SnoPen Al.

D.  

Consolidate all SnoPen AI related documents into a single chunk in the vector database.

Discussion 0
Questions 4

A Generative AI Engineer has created a RAG application which can help employees retrieve answers from an internal knowledge base, such as Confluence pages or Google Drive. The prototype application is now working with some positive feedback from internal company testers. Now the Generative Al Engineer wants to formally evaluate the system’s performance and understand where to focus their efforts to further improve the system.

How should the Generative AI Engineer evaluate the system?

Options:

A.  

Use cosine similarity score to comprehensively evaluate the quality of the final generated answers.

B.  

Curate a dataset that can test the retrieval and generation components of the system separately. Use MLflow’s built in evaluation metrics to perform the evaluation on the retrieval and generation components.

C.  

Benchmark multiple LLMs with the same data and pick the best LLM for the job.

D.  

Use an LLM-as-a-judge to evaluate the quality of the final answers generated.

Discussion 0
Questions 5

A Generative Al Engineer is setting up a Databricks Vector Search that will lookup news articles by topic within 10 days of the date specified An example query might be "Tell me about monster truck news around January 5th 1992". They want to do this with the least amount of effort.

How can they set up their Vector Search index to support this use case?

Options:

A.  

Split articles by 10 day blocks and return the block closest to the query.

B.  

Include metadata columns for article date and topic to support metadata filtering.

C.  

pass the query directly to the vector search index and return the best articles.

D.  

Create separate indexes by topic and add a classifier model to appropriately pick the best index.

Discussion 0
Questions 6

A Generative AI Engineer is developing a patient-facing healthcare-focused chatbot. If the patient’s question is not a medical emergency, the chatbot should solicit more information from the patient to pass to the doctor’s office and suggest a few relevant pre-approved medical articles for reading. If the patient’s question is urgent, direct the patient to calling their local emergency services.

Given the following user input:

“I have been experiencing severe headaches and dizziness for the past two days.”

Which response is most appropriate for the chatbot to generate?

Options:

A.  

Here are a few relevant articles for your browsing. Let me know if you have questions after reading them.

B.  

Please call your local emergency services.

C.  

Headaches can be tough. Hope you feel better soon!

D.  

Please provide your age, recent activities, and any other symptoms you have noticed along with your headaches and dizziness.

Discussion 0
Questions 7

A Generative Al Engineer at an automotive company would like to build a question-answering chatbot for customers to inquire about their vehicles. They have a database containing various documents of different vehicle makes, their hardware parts, and common maintenance information.

Which of the following components will NOT be useful in building such a chatbot?

Options:

A.  

Response-generating LLM

B.  

Invite users to submit long, rather than concise, questions

C.  

Vector database

D.  

Embedding model

Discussion 0
Questions 8

A Generative AI Engineer received the following business requirements for an external chatbot.

The chatbot needs to know what types of questions the user asks and routes to appropriate models to answer the questions. For example, the user might ask about upcoming event details. Another user might ask about purchasing tickets for a particular event.

What is an ideal workflow for such a chatbot?

Options:

A.  

The chatbot should only look at previous event information

B.  

There should be two different chatbots handling different types of user queries.

C.  

The chatbot should be implemented as a multi-step LLM workflow. First, identify the type of question asked, then route the question to the appropriate model. If it’s an upcoming event question, send the query to a text-to-SQL model. If it’s about ticket purchasing, the customer should be redirected to a payment platform.

D.  

The chatbot should only process payments

Discussion 0
Questions 9

A small and cost-conscious startup in the cancer research field wants to build a RAG application using Foundation Model APIs.

Which strategy would allow the startup to build a good-quality RAG application while being cost-conscious and able to cater to customer needs?

Options:

A.  

Limit the number of relevant documents available for the RAG application to retrieve from

B.  

Pick a smaller LLM that is domain-specific

C.  

Limit the number of queries a customer can send per day

D.  

Use the largest LLM possible because that gives the best performance for any general queries

Discussion 0
Questions 10

A Generative Al Engineer is developing a RAG application and would like to experiment with different embedding models to improve the application performance.

Which strategy for picking an embedding model should they choose?

Options:

A.  

Pick an embedding model trained on related domain knowledge

B.  

Pick the most recent and most performant open LLM released at the time

C.  

pick the embedding model ranked highest on the Massive Text Embedding Benchmark (MTEB) leaderboard hosted by HuggingFace

D.  

Pick an embedding model with multilingual support to support potential multilingual user questions

Discussion 0
Questions 11

A Generative AI Engineer has been asked to build an LLM-based question-answering application. The application should take into account new documents that are frequently published. The engineer wants to build this application with the least cost and least development effort and have it operate at the lowest cost possible.

Which combination of chaining components and configuration meets these requirements?

Options:

A.  

For the application a prompt, a retriever, and an LLM are required. The retriever output is inserted into the prompt which is given to the LLM to generate answers.

B.  

The LLM needs to be frequently with the new documents in order to provide most up-to-date answers.

C.  

For the question-answering application, prompt engineering and an LLM are required to generate answers.

D.  

For the application a prompt, an agent and a fine-tuned LLM are required. The agent is used by the LLM to retrieve relevant content that is inserted into the prompt which is given to the LLM to generate answers.

Discussion 0
Questions 12

A Generative Al Engineer interfaces with an LLM with prompt/response behavior that has been trained on customer calls inquiring about product availability. The LLM is designed to output “In Stock” if the product is available or only the term “Out of Stock” if not.

Which prompt will work to allow the engineer to respond to call classification labels correctly?

Options:

A.  

Respond with “In Stock” if the customer asks for a product.

B.  

You will be given a customer call transcript where the customer asks about product availability. The outputs are either “In Stock” or “Out of Stock”. Format the output in JSON, for example: {“call_id”: “123”, “label”: “In Stock”}.

C.  

Respond with “Out of Stock” if the customer asks for a product.

D.  

You will be given a customer call transcript where the customer inquires about product availability. Respond with “In Stock” if the product is available or “Out of Stock” if not.

Discussion 0
Questions 13

A Generative Al Engineer is ready to deploy an LLM application written using Foundation Model APIs. They want to follow security best practices for production scenarios

Which authentication method should they choose?

Options:

A.  

Use an access token belonging to service principals

B.  

Use a frequently rotated access token belonging to either a workspace user or a service principal

C.  

Use OAuth machine-to-machine authentication

D.  

Use an access token belonging to any workspace user

Discussion 0
Questions 14

A Generative Al Engineer is building a production-ready LLM system which replies directly to customers. The solution makes use of the Foundation Model API via provisioned throughput. They are concerned that the LLM could potentially respond in a toxic or otherwise unsafe way. They also wish to perform this with the least amount of effort.

Which approach will do this?

Options:

A.  

Host Llama Guard on Foundation Model API and use it to detect unsafe responses

B.  

Add some LLM calls to their chain to detect unsafe content before returning text

C.  

Add a regex expression on inputs and outputs to detect unsafe responses.

D.  

Ask users to report unsafe responses

Discussion 0
Questions 15

A Generative AI Engineer is testing a simple prompt template in LangChain using the code below, but is getting an error.

Assuming the API key was properly defined, what change does the Generative AI Engineer need to make to fix their chain?

A)

B)

C)

D)

Options:

A.  

Option A

B.  

Option B

C.  

Option C

D.  

Option D

Discussion 0
Questions 16

A Generative Al Engineer is responsible for developing a chatbot to enable their company’s internal HelpDesk Call Center team to more quickly find related tickets and provide resolution. While creating the GenAI application work breakdown tasks for this project, they realize they need to start planningwhich data sources (either Unity Catalog volume or Delta table) they could choose for this application. They have collected several candidate data sources for consideration:

call_rep_history: a Delta table with primary keys representative_id, call_id. This table is maintained to calculate representatives’ call resolution from fields call_duration and call start_time.

transcript Volume: a Unity Catalog Volume of all recordings as a *.wav files, but also a text transcript as *.txt files.

call_cust_history: a Delta table with primary keys customer_id, cal1_id. This table is maintained to calculate how much internal customers use the HelpDesk to make sure that the charge back model is consistent with actual service use.

call_detail: a Delta table that includes a snapshot of all call details updated hourly. It includes root_cause and resolution fields, but those fields may be empty for calls that are still active.

maintenance_schedule – a Delta table that includes a listing of both HelpDesk application outages as well as planned upcoming maintenance downtimes.

They need sources that could add context to best identify ticket root cause and resolution.

Which TWO sources do that? (Choose two.)

Options:

A.  

call_cust_history

B.  

maintenance_schedule

C.  

call_rep_history

D.  

call_detail

E.  

transcript Volume

Discussion 0
Questions 17

A Generative AI Engineer is developing a chatbot designed to assist users with insurance-related queries. The chatbot is built on a large language model (LLM) and is conversational. However, to maintain the chatbot’s focus and to comply with company policy, it must not provide responses to questions about politics. Instead, when presented with political inquiries, the chatbot should respond with a standard message:

“Sorry, I cannot answer that. I am a chatbot that can only answer questions around insurance.”

Which framework type should be implemented to solve this?

Options:

A.  

Safety Guardrail

B.  

Security Guardrail

C.  

Contextual Guardrail

D.  

Compliance Guardrail

Discussion 0
Questions 18

A Generative AI Engineer is creating an LLM-powered application that will need access to up-to-date news articles and stock prices.

The design requires the use of stock prices which are stored in Delta tables and finding the latest relevant news articles by searching the internet.

How should the Generative AI Engineer architect their LLM system?

Options:

A.  

Use an LLM to summarize the latest news articles and lookup stock tickers from the summaries to find stock prices.

B.  

Query the Delta table for volatile stock prices and use an LLM to generate a search query to investigate potential causes of the stock volatility.

C.  

Download and store news articles and stock price information in a vector store. Use a RAG architecture to retrieve and generate at runtime.

D.  

Create an agent with tools for SQL querying of Delta tables and web searching, provide retrieved values to an LLM for generation of response.

Discussion 0