Snowpro Advanced Data Engineer
Snowpro Advanced Data Engineer
Total: 65 Questions
Link: https://certyiq.com/papers/snowflake/snowpro-advanced-data-engineer
Question: 1 CertyIQ
A Data Engineer is investigating a query that is taking a long time to return. The Query Profile shows the following:
What step should the Engineer take to increase the query performance?
Answer: B
Explanation:
Question: 2 CertyIQ
How can the following relational data be transformed into semi-structured data using the LEAST amount of
operational overhead?
Answer: C
Explanation:
Reference:
https://docs.snowflake.com/en/sql-reference/functions/object_construct.
Question: 3 CertyIQ
A Data Engineer executes a complex query and wants to make use of Snowflake’s query results caching
capabilities to reuse the results.
Which conditions must be met? (Choose three.)
Answer: DEF
Explanation:
D.The table structure contributing to the query result cannot have changed.
E.The new query must have the same syntax as the previously executed query.
F.The micro-partitions cannot have changed due to changes to other data in the table.
Question: 4 CertyIQ
A Data Engineer needs to load JSON output from some software into Snowflake using Snowpipe.
Which recommendations apply to this scenario? (Choose three.)
Answer: BDE
Explanation:
B.Ensure that data files are 100-250 MB (or larger) in size, compressed.
D.Verify each value of each unique element stores a single native data type (string or number).
E.Extract semi-structured data elements containing null values into relational columns before loading.
Question: 5 CertyIQ
Given the table SALES which has a clustering key of column CLOSED_DATE, which table function will return the
average clustering depth for the SALES_REPRESENTATIVE column for the North American region?
Answer: B
Explanation:
Question: 6 CertyIQ
A large table with 200 columns contains two years of historical data. When queried, the table is filtered on a single
day. Below is the Query Profile:
Using a size 2XL virtual warehouse, this query took over an hour to complete.
What will improve the query performance the MOST?
Answer: D
Explanation:
Add a date column as a cluster key on the table.
Question: 7 CertyIQ
A Data Engineer is working on a Snowflake deployment in AWS eu-west-1 (Ireland). The Engineer is planning to
load data from staged files into target tables using the COPY INTO command.
Which sources are valid? (Choose three.)
Answer: CDE
Explanation:
Question: 8 CertyIQ
A Data Engineer wants to create a new development database (DEV) as a clone of the permanent production
database (PROD). There is a requirement to disable Fail-safe for all tables.
Which command will meet these requirements?
CLONE PROD -
FAIL_SAFE = FALSE;
B.CREATE DATABASE DEV -
CLONE PROD;
C.CREATE TRANSIENT DATABASE DEV -
CLONE PROD;
D.CREATE DATABASE DEV -
CLONE PROD -
DATA_RETENTION_TIME_IN DAYS = 0;
Answer: C
Explanation:
CLONE PROD.
Reference:
https://docs.snowflake.com/en/user-guide/tables-temp-transient
Question: 9 CertyIQ
Which query will show a list of the 20 most recent executions of a specified task, MYTASK, that have been
scheduled within the last hour that have ended or are still running?
A.
B.
C.
D.
Answer: B
Explanation:
BTo query only those tasks that have already completed or are currently running, include WHERE query_id IS
NOT NULL as a filter. The QUERY_ID column in the TASK_HISTORY output is populated only when a task has
started running.https://docs.snowflake.com/en/sql-reference/functions/task_historyA - will give all the
schedules even the ones that have not run yetC - A schedule could be skipped, cancelled so it won't give all
the runsD - It won't return the most recent tasks.
Question: 10 CertyIQ
Which methods can be used to create a DataFrame object in Snowpark? (Choose three.)
A.session.jdbc_connection()
B.session.read.json()
C.session.table()
D.DataFrame.write()
E.session.builder()
F.session.sql()
Answer: BCF
Explanation:
B.session.read.json().
C.session.table().
F.session.sql().
Reference:
https://docs.snowflake.com/en/developer-guide/snowpark/python/working-with-dataframes
Question: 11 CertyIQ
A new CUSTOMER table is created by a data pipeline in a Snowflake schema where MANAGED ACCESS is
enabled.
Which roles can grant access to the CUSTOMER table? (Choose three.)
Answer: AEF
Explanation:
Question: 12 CertyIQ
What is the purpose of the BUILD_STAGE_FILE_URL function in Snowflake?
Answer: C
Explanation:
Question: 13 CertyIQ
The JSON below is stored in a VARIANT column named V in a table named jCustRaw:
Which query will return one row per team member (stored in the teamMembers array) along with all of the
attributes of each team member?
A.
B.
C.
D.
Answer: B
Explanation:
Reference:
https://docs.snowflake.com/user-guide/semistructured-considerations#using-flatten-to-list-distinct-key-
names
Question: 14 CertyIQ
A company has an extensive script in Scala that transforms data by leveraging DataFrames. A Data Engineer
needs to move these transformations to Snowpark.
What characteristics of data transformations in Snowpark should be considered to meet this requirement? (Choose
two.)
Answer: AB
Explanation:
A.It is possible to join multiple tables using DataFrames.
Question: 15 CertyIQ
The following is returned from SYSTEM$CLUSTERING_INFORMATION() for a table named ORDERS with a DATE
column named O_ORDERDATE:
A.The table is clustered very well on O_ORDERDATE, as there are 493 micro-partitions that could not be
significantly improved by reclustering.
B.The table is not clustered well on O_ORDERDATE, as there are 493 micro-partitions where the range of
values in that column overlap with every other micro-partition in the table.
C.The data in O_ORDERDATE does not change very often, as there are 493 micro-partitions containing rows
where that column has not been modified since the row was created.
D.The data in O_ORDERDATE has a very low cardinality, as there are 493 micro-partitions where there is only a
single distinct value in that column for all rows in the micro-partition.
Answer: A
Explanation:
The table is clustered very well on O_ORDERDATE, as there are 493 micro-partitions that could not be
significantly improved by reclustering.
Question: 16 CertyIQ
A company is building a dashboard for thousands of Analysts. The dashboard presents the results of a few
summary queries on tables that are regularly updated. The query conditions vary by topic according to what data
each Analyst needs. Responsiveness of the dashboard queries is a top priority, and the data cache should be
preserved.
How should the Data Engineer configure the compute resources to support this dashboard?
A.Assign queries to a multi-cluster virtual warehouse with economy auto-scaling. Allow the system to
automatically start and stop clusters according to demand.
B.Assign all queries to a multi-cluster virtual warehouse set to maximized mode. Monitor to determine the
smallest suitable number of clusters.
C.Create a virtual warehouse for every 250 Analysts. Monitor to determine how many of these virtual
warehouses are being utilized at capacity.
D.Create a size XL virtual warehouse to support all the dashboard queries. Monitor query runtimes to determine
whether the virtual warehouse should be resized.
Answer: B
Explanation:
Assign all queries to a multi-cluster virtual warehouse set to maximized mode. Monitor to determine the
smallest suitable number of clusters.
Question: 17 CertyIQ
A Data Engineer has developed a dashboard that will issue the same SQL select clause to Snowflake every 12
hours.
How long will Snowflake use the persisted query results from the result cache, provided that the underlying data
has not changed?
A.12 hours
B.24 hours
C.14 days
D.31 days
Answer: D
Explanation:
Each time the persisted result for a query is reused, Snowflake resets the 24-hour retention period for the
result, up to a maximum of 31 days from the date and time that the query was first executed. After 31 days,
the result is purged and the next time the query is submitted, a new result is generated and persisted.
https://docs.snowflake.com/en/user-guide/querying-persisted-results
Question: 18 CertyIQ
A Data Engineer ran a stored procedure containing various transactions. During the execution, the session abruptly
disconnected, preventing one transaction from committing or rolling back. The transaction was left in a detached
state and created a lock on resources.
What step must the Engineer take to immediately run a new transaction?
Answer: A
Explanation:
Question: 19 CertyIQ
A database contains a table and a stored procedure defined as:
The log_table is initially empty and a Data Engineer issues the following command:
CALL insert_log(NULL::VARCHAR);
No other operations are affecting the log_table.
What will be the outcome of the procedure call?
A.The log_table contains zero records and the stored procedure returned 1 as a return value.
B.The log_table contains one record and the stored procedure returned 1 as a return value.
C.The log_table contains one record and the stored procedure returned NULL as a return value.
D.The log_table contains zero records and the stored procedure returned NULL as a return value.
Answer: D
Explanation:
The log_table contains zero records and the stored procedure returned NULL as a return value.
Question: 20 CertyIQ
When would a Data Engineer use TABLE with the FLATTEN function instead of the LATERAL FLATTEN
combination?
A.When TABLE with FLATTEN requires another source in the FROM clause to refer to.
B.When TABLE with FLATTEN requires no additional source in the FROM clause to refer to.
C.When the LATERAL FLATTEN combination requires no other source in the FROM clause to refer to.
D.When TABLE with FLATTEN is acting like a sub-query executed for each returned row.
Answer: B
Explanation:
Reference:
https://docs.snowflake.com/en/sql-reference/functions/flatten
Thank you
Thank you for being so interested in the premium exam material.
I'm glad to hear that you found it informative and helpful.
But Wait
I wanted to let you know that there is more content available in the full version.
The full paper contains additional sections and information that you may find helpful,
and I encourage you to download it to get a more comprehensive and detailed view of
all the subject matter.
Total: 65 Questions
Link: https://certyiq.com/papers/snowflake/snowpro-advanced-data-engineer