Certification Practice Test | PDF Questions | Actual Questions | Test Engine | Pass4Sure
DEA-C01 : SnowPro Advanced Data Engineer Exam

SnowFlake DEA-C01 Questions & Answers
Full Version: 133 Q&A
DEA-C01 Dumps DEA-C01 Braindumps
DEA-C01 Real Questions DEA-C01 Practice Test DEA-C01 Actual Questions
killexams.com SnowFlake DEA-C01
SnowPro Advanced Data Engineer
https://killexams.com/pass4sure/exam-detail/DEA-C01
Question: 62
Each micro partition contains between 50 mb and 500 MB of uncompressed data
TRUE
FALSE
Answer: A Explanation
What are Micro-partitions?
All data in Snowflake tables is automatically divided into micro-partitions, which are contiguous units of storage. Each micro-partition contains between 50 MB and 500 MB of uncompressed data (note that the actual size in Snowflake is smaller because data is always stored compressed). Groups of rows in tables are mapped into individual micro-partitions, organized in a columnar fashion. This size and structure allows for extremely granular pruning of very large tables, which can be comprised of millions, or even hundreds of millions, of micro-partitions.
Snowflake stores metadata about all rows stored in a micro-partition, including:
Question: 63
An existing clustering key is copied in which of the below scenarios
CREATE TABLE…CLONE
CREATE TABLE…LIKE
CREATE TABLE…AS SELECT
Answer: A Explanation
https://docs.snowflake.com/en/sql-reference/functions/system_estimate_search_optimization_costs.html#out put
BuildCosts
This object describes the predicted costs of building the search access path for the table. If search optimization has already been added to the table, this object contains no cost information. StorageCosts
This object describes the predicted amount of storage space (in TB) needed for the search access path for the table. If search optimization has already been added to the table, this object shows the current amount of space used by the search access path.
Benefit
This object does not contain any cost information at this time. MaintenanceCosts
This object describes the predicted costs of maintaining the search access path for the table. If this table has been created recently, this object does not contain any cost information.
Question: 64
CREATE OR REPLACE TABLE TIME_TRAVEL_SCHEMA.TIME_TRAVEL_TABLE (ID NUMBER) DATA_RETENTION_TIME_IN_DAYS =20;
Later you dropped the schema. In this scenario what data retention value will be honored for the table, if we need to retrieve the table data
10
20
30
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/data-time-travel.html#dropped-containers-and-object-retention-inh eritance Dropped Containers and Object Retention Inheritance
Currently, when a database is dropped, the data retention period for child schemas or tables, if explicitly set to be different from the retention of the database, is not honored. The child schemas or tables are retained for the same period of time as the database.
Similarly, when a schema is dropped, the data retention period for child tables, if explicitly set to be different from the retention of the schema, is not honored. The child tables are retained for the same period of time as the schema.
To honor the data retention period for these child objects (schemas or tables), drop them explicitly before you drop the database or schema.
Question: 65
Which system table will you use to get the total credit consumption over a specific time period?
WAREHOUSE_METERING_HISTORY
WAREHOUSE_CREDIT_USAGE_HISTORY
WAREHOUSE_USAGE_HISTORY
Answer: A Explanation
The WAREHOUSE_METERING_HISTORY table in the ACCOUNT_USAGE Schema can be used to get the desired information. Run the below query to try this out.
SELECT WAREHOUSE_NAME, SUM(CREDITS_USED_COMPUTE) AS CREDITS_USED_COMPUTE_SUM FROM ACCOUNT_USAGE.WAREHOUSE_METERING_HISTORY
GROUP BY 1 ORDER BY 2 DESC;
Question: 66
Snowpipe supports loading from both internal and external stage
FALSE
TRUE
Answer: B Explanation
Step 1: Create a Stage (If Needed)
Snowpipe supports loading from the following stage types:
Question: 67
Each micro partition contains between 50 mb and 500 MB of uncompressed data
TRUE
FALSE
Answer: A
Explanation
What are Micro-partitions?
All data in Snowflake tables is automatically divided into micro-partitions, which are contiguous units of storage. Each micro-partition contains between 50 MB and 500 MB of uncompressed data (note that the actual size in Snowflake is smaller because data is always stored compressed). Groups of rows in tables are mapped into individual micro-partitions, organized in a columnar fashion. This size and structure allows for extremely granular pruning of very large tables, which can be comprised of millions, or even hundreds of millions, of micro-partitions.
Snowflake stores metadata about all rows stored in a micro-partition, including:
Question: 68
Which of the below statements are true?
ACCOUNT USAGE includes dropped objects but INFORMATION SCHEMA does not
INFORMATION SCHEMA includes dropped objects but ACCOUNT USAGE does not
BOTH includes dropped object
BOTH does not include dropped object
Answer: A Explanation
https://docs.snowflake.com/en/sql-reference/account-usage.html#differences-between-account-usage-and-information-schema Dropped Object Records
Account usage views include records for all objects that have been dropped. An additional DELETED column displays the timestamp when the object was dropped.
In addition, because objects can be dropped and recreated with the same name, to differentiate between objects records that have the same name, the account usage views include ID columns, where appropriate, that display the internal IDs generated and assigned to each record by the system.
Question: 69
Which of the below statements are true?
ACCOUNT USAGE includes dropped objects but INFORMATION SCHEMA does not
INFORMATION SCHEMA includes dropped objects but ACCOUNT USAGE does not
BOTH includes dropped object
BOTH does not include dropped object
Answer: A Explanation
https://docs.snowflake.com/en/sql-reference/account-usage.html#differences-between-account-usage-and-information-schema Dropped Object Records
Account usage views include records for all objects that have been dropped. An additional DELETED column displays the timestamp when the object was dropped.
In addition, because objects can be dropped and recreated with the same name, to differentiate between objects records that have the same name, the account usage views include ID columns, where appropriate, that display the internal IDs generated and assigned to each record by the system.
Question: 70
CREATE OR REPLACE TABLE TIME_TRAVEL_SCHEMA.TIME_TRAVEL_TABLE (ID NUMBER) DATA_RETENTION_TIME_IN_DAYS =20;
Later you dropped the schema. In this scenario what data retention value will be honored for the table, if we need to retrieve the table data
10
20
30
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/data-time-travel.html#dropped-containers-and-object-retention-inh eritance Dropped Containers and Object Retention Inheritance
Currently, when a database is dropped, the data retention period for child schemas or tables, if explicitly set to be different from the retention of the database, is not honored. The child schemas or tables are retained for the same period of time as the database.
Similarly, when a schema is dropped, the data retention period for child tables, if explicitly set to be different from the retention of the schema, is not honored. The child tables are retained for the same period of time as the schema.
To honor the data retention period for these child objects (schemas or tables), drop them explicitly before you drop the database or schema.
Question: 71
Time travel cannot be disabled for an account, but it can be disabled for individual databases, schemas and tables by specifying DATA_RETENTION_TIME_IN_DAYS with a value of 0 for the object
TRUE
FALSE
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/data-time-travel.html#enabling-and-disabling-time-travel Enabling and Disabling Time Travel
No tasks are required to enable Time Travel. It is automatically enabled with the standard, 1-day retention period.
However, you may wish to upgrade to Snowflake Enterprise Edition to enable configuring longer data retention periods of up to 90 days for databases, schemas, and tables. Note that extended data retention requires additional storage which will be reflected in your monthly storage charges. For more information about storage charges, see Storage Costs for Time Travel and Fail-safe.
Time Travel cannot be disabled for an account; however, it can be disabled for individual databases, schemas, and tables by specifying DATA_RETENTION_TIME_IN_DAYS with a value of 0 for the object. Also, users with the ACCOUNTADMIN role can set DATA_RETENTION_TIME_IN_DAYS to 0 at the account level, which means that all databases (and subsequently all schemas and tables) created in the account have no retention period by default; however, this default can be overridden at any time for any database, schema, or table.
Question: 72
CREATE OR REPLACE TABLE TIME_TRAVEL_SCHEMA.TIME_TRAVEL_TABLE (ID NUMBER) DATA_RETENTION_TIME_IN_DAYS =20;
Later you dropped the schema. In this scenario what data retention value will be honored for the table, if we need to retrieve the table data
10
20
30
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/data-time-travel.html#dropped-containers-and-object-retention-inh eritance Dropped Containers and Object Retention Inheritance
Currently, when a database is dropped, the data retention period for child schemas or tables, if explicitly set to be different from the retention of the database, is not honored. The child schemas or tables are retained for the same period of time as the database.
Similarly, when a schema is dropped, the data retention period for child tables, if explicitly set to be different from the retention of the schema, is not honored. The child tables are retained for the same period of time as the schema.
To honor the data retention period for these child objects (schemas or tables), drop them explicitly before you drop the database or schema.
Question: 73
Snowpipe supports loading from both internal and external stage
FALSE
TRUE
Answer: B Explanation
Step 1: Create a Stage (If Needed)
Snowpipe supports loading from the following stage types:
Question: 74
Snowflake charges a per-byte fee when users transfer data from your snowflake account into cloud storages in another region on the same cloud platform or into cloud storage in another cloud platform
TRUE
FALSE
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/billing-data-transfer.html#understanding-snowflake-data-transfer-billing Cloud providers apply data egress charges in either of the following use cases:
Data is transferred from one region to another within the same cloud platform. Data is transferred out of the cloud platform.
To recover these expenses, Snowflake charges a per-byte fee when users transfer data from your Snowflake account (hosted on AWS, Google Cloud Platform, or Microsoft Azure) into cloud storage in another region on the same cloud platform, or into cloud storage in another cloud platform.
The amount charged per byte depends on the region where your Snowflake account is hosted. For data transfer pricing, see the pricing guide (on the Snowflake website):
Question: 75
;
call sp1();
SELECT * FROM EMPLOYEE ORDER BY ID;
1 MOHAN
2 RON
1 MOHAN
RON
RANJAN
1 MOHAN
3 RANJAN
1 MOHAN
Answer: A Explanation
https://docs.snowflake.com/en/sql-reference/transactions.html#scoped-transactions Scoped Transactions
A stored procedure that contains a transaction can be called from within another transaction. For example, a transaction inside a stored procedure can include a call to another stored procedure that contains a transaction.
Snowflake does not treat the inner transaction as nested; instead, the inner transaction is a separate transaction. Snowflake calls these “autonomous scoped transactions†(or simply “scoped transactionsâ€). The starting point and ending point of each scoped transaction determine which statements are included in the transaction. The start and end can be explicit or implicit. Each SQL statement is part of only one transaction. An enclosing ROLLBACK or COMMIT does not undo an enclosed COMMIT or ROLLBACK.
Question: 76
Which of the below statements are true for API integration object?
Only Snowflake users who have the ACCOUNTADMIN role or who have a role with the global CREATE INTEGRATION privilege can execute CREATE API INTEGRATION.
Only Snowflake roles with OWNERSHIP or USAGE privileges on the API integration can use the API integration directly (e.g. by creating an external function that specifies that API integration).
An API integration object is tied to a specific cloud platform account and role within that account, but not to a specific HTTPS proxy URL. You can create more than one instance of an HTTPS proxy service in a cloud provider account, and you can use the same API integration to authenticate to multiple proxy services in that account.
Your Snowflake account can have multiple API integration objects, for example, for different cloud platform accounts.
Multiple external functions can use the same API integration object, and thus the same HTTPS proxy service.
ALL OF THE ABOVE
Answer: F Explanation
API integration is an important topic to focus on for the certification. Please read this topic thoroughly. https://docs.snowflake.com/en/sql-reference/sql/create- api-integration.html#create-api-integration Usage Notes
Only Snowflake users who have the ACCOUNTADMIN role or who have a role with the global CREATE INTEGRATION privilege can execute CREATE API INTEGRATION.
Only Snowflake roles with OWNERSHIP or USAGE privileges on the API integration can use the API integration directly (e.g. by creating an external function that specifies that API integration).
An API integration object is tied to a specific cloud platform account and role within that account, but not to a specific HTTPS proxy URL. You can create more than one instance of an HTTPS proxy service in a cloud provider account, and you can use the same API integration to authenticate to multiple proxy services in that account.
Your Snowflake account can have multiple API integration objects, for example, for different cloud platform accounts. Multiple external functions can use the same API integration object, and thus the same HTTPS proxy service. Question: 77
Time travel cannot be disabled for an account, but it can be disabled for individual databases, schemas and tables by specifying
DATA_RETENTION_TIME_IN_DAYS with a value of 0 for the object
TRUE
FALSE
Answer: A Explanation
https://docs.snowflake.com/en/user-guide/data-time-travel.html#enabling-and-disabling-time-travel Enabling and Disabling Time Travel
No tasks are required to enable Time Travel. It is automatically enabled with the standard, 1-day retention period.
However, you may wish to upgrade to Snowflake Enterprise Edition to enable configuring longer data retention periods of up to 90 days for databases, schemas, and tables. Note that extended data retention requires additional storage which will be reflected in your monthly storage charges. For more information about storage charges, see Storage Costs for Time Travel and Fail-safe.
Time Travel cannot be disabled for an account; however, it can be disabled for individual databases, schemas, and tables by specifying DATA_RETENTION_TIME_IN_DAYS with a value of 0 for the object. Also, users with the ACCOUNTADMIN role can set DATA_RETENTION_TIME_IN_DAYS to 0 at the account level, which means that all databases (and subsequently all schemas and tables) created in the account have no retention period by default; however, this default can be overridden at any time for any database, schema, or table.
User: Agustina*****![]() ![]() ![]() ![]() ![]() I highly recommend this package to anyone planning to take the DEA-C01 exam. The certification exams are difficult, and it requires a lot of work to pass them. Killexams.com does most of the heavy lifting for you. The DEA-C01 exam questions I received from this website were very similar to those on the actual exam. Without these practice tests, I would have failed, and that is why many people do not pass the DEA-C01 exam on their first attempt. |
User: Agnes*****![]() ![]() ![]() ![]() ![]() Comprehensive testprep resources for the dea-c01 exam allowed me to complete all questions within the scheduled time. Their Questions and Answers were invaluable, ensuring secure coaching and a successful pass. I am grateful for their outstanding support. |
User: Polly*****![]() ![]() ![]() ![]() ![]() With Killexams.com, I was able to prepare for the DEA-C01 exam in just one day and achieved remarkable success. Their preparation pack included real exam questions, the latest updates, and much more, which helped me focus on precisely what I needed to know and avoid wasting time on unnecessary topics. |
User: Evie*****![]() ![]() ![]() ![]() ![]() I renewed my Killexams.com subscription for the DEA-C01 exam, and it was a wise decision. Their resources are invaluable, offering practice tests that helped me score above 95%. The team at Killexams.com provides exceptional support, and their materials are critical for anyone aiming for high marks in the DEA-C01 exam. |
User: Atharv*****![]() ![]() ![]() ![]() ![]() I want to inform everyone that I achieved a solid score on the DEA-C01 exam thanks to Killexams.com. Their material is a reliable exam practice test that I highly recommend to anyone working towards an IT certification. Everyone in my IT organization has either used or heard of Killexams.com, as they not only help you pass but also ensure that you become a successful expert in your field. |
Features of iPass4sure DEA-C01 Exam
- Files: PDF / Test Engine
- Premium Access
- Online Test Engine
- Instant download Access
- Comprehensive Q&A
- Success Rate
- Real Questions
- Updated Regularly
- Portable Files
- Unlimited Download
- 100% Secured
- Confidentiality: 100%
- Success Guarantee: 100%
- Any Hidden Cost: $0.00
- Auto Recharge: No
- Updates Intimation: by Email
- Technical Support: Free
- PDF Compatibility: Windows, Android, iOS, Linux
- Test Engine Compatibility: Mac / Windows / Android / iOS / Linux
Premium PDF with 133 Q&A
Get Full VersionAll SnowFlake Exams
SnowFlake ExamsCertification and Entry Test Exams
Complete exam list