Certification Practice Test | PDF Questions | Actual Questions | Test Engine | Pass4Sure
DCDEP : Databricks Certified Data Engineer Professional Exam
Databrick DCDEP Questions & Answers
Full Version: 134 Q&A
DCDEP Dumps
DCDEP Braindumps
DCDEP Real Questions
DCDEP Practice Test
DCDEP Actual Questions
Databrick
DCDEP
Databricks Certified Data Engineer Professional
https://killexams.com/pass4sure/exam-detail/DCDEP
Question: 21
A DELTA LIVE TABLE pipelines can be scheduled to run in two different modes, what are these two different
modes?
A. Triggered, Incremental
B. Once, Continuous
C. Triggered, Continuous
D. Once, Incremental
E. Continuous, Incremental
Answer: C
Explanation:
The answer is Triggered, Continuous
https://docs.microsoft.com/en-us/azure/databricks/data-engineering/delta-live-tables/delta-live-tables-concepts#--
continuous-and-triggered-pipelines
βΆ Triggered pipelines update each table with whatever data is currently available and then stop the cluster running the
pipeline. Delta Live Tables automatically analyzes the dependencies between your tables and starts by computing
those that read from external sources. Tables within the pipeline are updated after their dependent data sources have
been updated.
βΆ Continuous pipelines update tables continuously as input data changes. Once an update is started, it continues to run
until manually stopped. Continuous pipelines require an always-running cluster but ensure that downstream consumers
have the most up-to-date data.
Question: 22
Which of the following developer operations in CI/CD flow can be implemented in Databricks Re-pos?
A. Merge when code is committed
B. Pull request and review process
C. Trigger Databricks Repos API to pull the latest version of code into production folder
D. Resolve merge conflicts
E. Delete a branch
Answer: C
Explanation:
See the below diagram to understand the role Databricks Repos and Git provider plays when building a CI/CD
workflow.
All the steps highlighted in yellow can be done Databricks Repo, all the steps highlighted in Gray are done in a git
provider like Github or Azure DevOps
Question: 23
Identify one of the below statements that can query a delta table in PySpark Dataframe API
A. Spark.read.mode("delta").table("table_name")
B. Spark.read.table.delta("table_name")
C. Spark.read.table("table_name")
D. Spark.read.format("delta").LoadTableAs("table_name")
E. Spark.read.format("delta").TableAs("table_name")
Answer: C
Question: 24
How VACCUM and OPTIMIZE commands can be used to manage the DELTA lake?
A. VACCUM command can be used to compact small parquet files, and the OP-TIMZE command can be used to
delete parquet files that are marked for dele-tion/unused.
B. VACCUM command can be used to delete empty/blank parquet files in a delta table.
OPTIMIZE command can be used to update stale statistics on a delta table.
C. VACCUM command can be used to compress the parquet files to reduce the size of the table, OPTIMIZE
command can be used to cache frequently delta tables for better performance.
D. VACCUM command can be used to delete empty/blank parquet files in a delta table, OPTIMIZE command can be
used to cache frequently delta tables for better performance.
E. OPTIMIZE command can be used to compact small parquet files, and the VAC-CUM command can be used to
delete parquet files that are marked for deletion/unused. (Correct)
Answer: E
Explanation:
VACCUM:
You can remove files no longer referenced by a Delta table and are older than the retention thresh-old by running the
vacuum command on the table. vacuum is not triggered automatically. The de-fault retention threshold for the files is 7
days. To change this behavior, see Configure data retention for time travel.
OPTIMIZE:
Using OPTIMIZE you can compact data files on Delta Lake, this can improve the speed of read queries on the table.
Too many small files can significantly degrade the performance of the query.
Question: 25
Which of the following statements are correct on how Delta Lake implements a lake house?
A. Delta lake uses a proprietary format to write data, optimized for cloud storage
B. Using Apache Hadoop on cloud object storage
C. Delta lake always stores meta data in memory vs storage
D. Delta lake uses open source, open format, optimized cloud storage and scalable meta data
E. Delta lake stores data and meta data in computes memory
Answer: D
Explanation:
Delta lake is
βΆ Open source
βΆ Builds up on standard data format
βΆ Optimized for cloud object storage
βΆ Built for scalable metadata handling Delta lake is not
βΆ Proprietary technology
βΆ Storage format
βΆ Storage medium
βΆ Database service or data warehouse
Question: 26
What are the different ways you can schedule a job in Databricks workspace?
A. Continuous, Incremental
B. On-Demand runs, File notification from Cloud object storage
C. Cron, On Demand runs
D. Cron, File notification from Cloud object storage
E. Once, Continuous
Answer: C
Explanation:
The answer is, Cron, On-Demand runs
Supports running job immediately or using can be scheduled using CRON syntax
Question: 27
Which of the following type of tasks cannot setup through a job?
A. Notebook
B. DELTA LIVE PIPELINE
C. Spark Submit
D. Python
E. Databricks SQL Dashboard refresh
Answer: E
Question: 28
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks
Lakehouse Platform?
A. Databricks Repos can facilitate the pull request, review, and approval process before merging branches
B. Databricks Repos can merge changes from a secondary Git branch into a main Git branch
C. Databricks Repos can be used to design, develop, and trigger Git automation pipelines
D. Databricks Repos can store the single-source-of-truth Git repository
E. Databricks Repos can commit or push code changes to trigger a CI/CD process
Answer: E
Explanation:
Answer is Databricks Repos can commit or push code changes to trigger a CI/CD process See below diagram to
understand the role Databricks Repos and Git provider plays when building a CI/CD workdlow.
All the steps highlighted in yellow can be done Databricks Repo, all the steps highlighted in Gray are done in a git
provider like Github or Azure Devops.
Diagram
Description automatically generated
User: Maryam***** I passed the DCDEP exam on the first try, thanks to killexams.com practice tests team. I used my work knowledge within the question and answer format to answer the exam papers with an exam simulator and got a good grasp of the exam paper. I would like to thank killexams.com for their excellent study material. |
User: Seryozha***** Despite finding the dcdep exam difficult, I managed to score 87% and pass it, thanks to Killexams.com. Initially, I was about to give up on taking this exam, but my friend suggested using Killexams.com questions and answers. Within just four weeks, I was thoroughly prepared for the exam. |
User: Jood***** I would recommend killexams.com question bank to anyone preparing for the DCDEP exam. It was very helpful in providing a concept of the type of questions that would come up and which areas to focus on. The practice exam provided was also excellent in getting a sense of what to expect on the actual exam day. The answer keys supplied were a great help in recollecting what I had learned, and the explanations provided were clear and easy to understand. |
User: Maksim***** There were many approaches for me to reach my goal of a high score within the DCDEP, but I was not having great success in that area. So, I took the excellent decision of using the online DCDEP practice help of the Killexams.com website, and I found that this decision was a sweet one to be remembered for a long time. The reason for my high score in the DCDEP exam was the Killexams.com practice exam, which was available online. |
User: Julianna***** I have renewed my membership with killexams.com for the Databrick DCDEP exam because their assistance is vital to my success. I am confident that their practice exams will help me obtain my accreditation and secure more than 95% marks. The team at killexams.com is doing an outstanding job, and I hope they continue to maintain their high standards. |
Features of iPass4sure DCDEP Exam
- Files: PDF / Test Engine
- Premium Access
- Online Test Engine
- Instant download Access
- Comprehensive Q&A
- Success Rate
- Real Questions
- Updated Regularly
- Portable Files
- Unlimited Download
- 100% Secured
- Confidentiality: 100%
- Success Guarantee: 100%
- Any Hidden Cost: $0.00
- Auto Recharge: No
- Updates Intimation: by Email
- Technical Support: Free
- PDF Compatibility: Windows, Android, iOS, Linux
- Test Engine Compatibility: Mac / Windows / Android / iOS / Linux
Premium PDF with 134 Q&A
Get Full VersionAll Databrick Exams
Databrick ExamsCertification and Entry Test Exams
Complete exam list