databricks certified data engineer associate practice test

Exam Title: certified data engineer associate

Last update: Dec 11 ,2025
Question 1

Which of the following describes the type of workloads that are always compatible with Auto Loader?

  • A. Streaming workloads
  • B. Machine learning workloads
  • C. Serverless workloads
  • D. Batch workloads
  • E. Dashboard workloads
Answer:

a

vote your answer:
A
B
C
D
E
A 4 B 1 C 0 D 0 E 0
Comments
Question 2

A new data engineering team has been assigned to work on a project. The team will need access to database customers in order to see what tables already exist. The team has its own group team.
Which of the following commands can be used to grant the necessary permission on the entire database to the new team?

  • A. GRANT VIEW ON CATALOG customers TO team;
  • B. GRANT CREATE ON DATABASE customers TO team;
  • C. GRANT USAGE ON CATALOG team TO customers;
  • D. GRANT CREATE ON DATABASE team TO customers;
  • E. GRANT USAGE ON DATABASE customers TO team;
Answer:

e -

vote your answer:
A
B
C
D
E
A 0 B 0 C 0 D 0 E 2
Comments
Question 3

Which of the following is stored in the Databricks customer's cloud account?

  • A. Databricks web application
  • B. Cluster management metadata
  • C. Repos
  • D. Data
  • E. Notebooks
Answer:

d

vote your answer:
A
B
C
D
E
A 0 B 0 C 0 D 2 E 0
Comments
Question 4

Which of the following can be used to simplify and unify siloed data architectures that are specialized for specific use cases?

  • A. None of these
  • B. Data lake
  • C. Data warehouse
  • D. All of these
  • E. Data lakehouse
Answer:

e

vote your answer:
A
B
C
D
E
A 0 B 0 C 0 D 0 E 2
Comments
Question 5

Which of the following data lakehouse features results in improved data quality over a traditional data lake?

  • A. A data lakehouse provides storage solutions for structured and unstructured data.
  • B. A data lakehouse supports ACID-compliant transactions.
  • C. A data lakehouse allows the use of SQL queries to examine data.
  • D. A data lakehouse stores data in open formats.
  • E. A data lakehouse enables machine learning and artificial Intelligence workloads.
Answer:

c

vote your answer:
A
B
C
D
E
A 0 B 2 C 1 D 0 E 0
Comments
Question 6

A data engineer is attempting to drop a Spark SQL table my_table and runs the following command:

DROP TABLE IF EXISTS my_table;

After running this command, the engineer notices that the data files and metadata files have been deleted from the file system.

Which of the following describes why all of these files were deleted?

  • A. The table was managed
  • B. The table's data was smaller than 10 GB
  • C. The table's data was larger than 10 GB
  • D. The table was external
  • E. The table did not have a location
Answer:

a

vote your answer:
A
B
C
D
E
A 1 B 0 C 0 D 0 E 0
Comments
Question 7

A data engineer has been given a new record of data:

id STRING = 'a1'
rank INTEGER = 6
rating FLOAT = 9.4

Which of the following SQL commands can be used to append the new record to an existing Delta table my_table?

  • A. INSERT INTO my_table VALUES ('a1', 6, 9.4)
  • B. my_table UNION VALUES ('a1', 6, 9.4)
  • C. INSERT VALUES ( 'a1' , 6, 9.4) INTO my_table
  • D. UPDATE my_table VALUES ('a1', 6, 9.4)
  • E. UPDATE VALUES ('a1', 6, 9.4) my_table
Answer:

a

vote your answer:
A
B
C
D
E
A 2 B 0 C 0 D 0 E 0
Comments
Question 8

Which of the following Structured Streaming queries is performing a hop from a Silver table to a Gold table?

  • E. None
Answer:

e

vote your answer:
E
E 1
Comments
Question 9

A data engineer has created a new database using the following command:

CREATE DATABASE IF NOT EXISTS customer360;

In which of the following locations will the customer360 database be located?

  • A. dbfs:/user/hive/database/customer360
  • B. dbfs:/user/hive/warehouse
  • C. dbfs:/user/hive/customer360
  • D. More information is needed to determine the correct response
  • E. dbfs:/user/hive/database
Answer:

d

vote your answer:
A
B
C
D
E
A 0 B 0 C 0 D 2 E 0
Comments
Question 10

A data engineer is designing a data pipeline. The source system generates files in a shared directory that is also used by other processes. As a result, the files should be kept as is and will accumulate in the directory. The data engineer needs to identify which files are new since the previous run in the pipeline, and set up the pipeline to only ingest those new files with each run.
Which of the following tools can the data engineer use to solve this problem?

  • A. Unity Catalog
  • B. Delta Lake
  • C. Databricks SQL
  • D. Data Explorer
  • E. Auto Loader
Answer:

e

vote your answer:
A
B
C
D
E
A 0 B 0 C 0 D 1 E 3
Comments
Page 1 out of 8
Viewing questions 1-10 out of 90
Go To
page 2