What steps need to be taken to set up a DELTA LIVE PIPELINE as a job using the workspace UI?
What steps need to be taken to set up a DELTA LIVE PIPELINE as a job using the workspace UI?A . DELTA LIVE TABLES do not support job cluster B. Select Workflows UI and Delta live tables tab, under task type select Delta live tables pipeline and select the notebook...
table("uncleanedSales")
table("uncleanedSales")View AnswerAnswer: B Explanation: The answer is
FROM carts GROUP BY cartId
FROM carts GROUP BY cartId Expected result: cartId items 1 [1,100,200,300,250]A . FLATTEN, COLLECT_UNION B. ARRAY_UNION, FLATTEN C. ARRAY_UNION, ARRAY_DISTINT D. ARRAY_UNION, COLLECT_SET E. ARRAY_DISTINCT, ARRAY_UNIONView AnswerAnswer: D Explanation: COLLECT SET is a kind of aggregate function that combines a column value from all rows into a unique list ARRAY_UNION...
The current ELT pipeline is receiving data from the operations team once a day so you had setup an AUTO LOADER process to run once a day using trigger (Once = True) and scheduled a job to run once a day, operations team recently rolled out a new feature that allows them to send data every 1 min, what changes do you need to make to AUTO LOADER to process the data every 1 min.
The current ELT pipeline is receiving data from the operations team once a day so you had setup an AUTO LOADER process to run once a day using trigger (Once = True) and scheduled a job to run once a day, operations team recently rolled out a new feature that...
as total_sales from sales
as total_sales from salesView AnswerAnswer: C Explanation: The answer is
ELSE (temp C 33 ) * 5/9 5.END
ELSE (temp C 33 ) * 5/9 5.ENDView AnswerAnswer: D Explanation: The answer is
You noticed that colleague is manually copying the notebook with _bkp to store the previous versions, which of the following feature would you recommend instead.
You noticed that colleague is manually copying the notebook with _bkp to store the previous versions, which of the following feature would you recommend instead.A . Databricks notebooks support change tracking and versioning B. Databricks notebooks should be copied to a local machine and setup source control locally to version...
You are asked to setup an AUTO LOADER to process the incoming data, this data arrives in JSON format and get dropped into cloud object storage and you are required to process the data as soon as it arrives in cloud storage, which of the following statements is correct
You are asked to setup an AUTO LOADER to process the incoming data, this data arrives in JSON format and get dropped into cloud object storage and you are required to process the data as soon as it arrives in cloud storage, which of the following statements is correctA ....
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?A . Databricks Repos can facilitate the pull request, review, and approval process before merging branches B. Databricks Repos can merge changes from a secondary Git branch into a main Git branch C....
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?
Which of the following describes how Databricks Repos can help facilitate CI/CD workflows on the Databricks Lakehouse Platform?A . Databricks Repos can facilitate the pull request, review, and approval process before merging branches B. Databricks Repos can merge changes from a secondary Git branch into a main Git branch C....