Spring Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: scxmas70

ARA-C01 Exam Dumps - SnowPro Advanced: Architect Certification Exam

Searching for workable clues to ace the Snowflake ARA-C01 Exam? You’re on the right place! ExamCert has realistic, trusted and authentic exam prep tools to help you achieve your desired credential. ExamCert’s ARA-C01 PDF Study Guide, Testing Engine and Exam Dumps follow a reliable exam preparation strategy, providing you the most relevant and updated study material that is crafted in an easy to learn format of questions and answers. ExamCert’s study tools aim at simplifying all complex and confusing concepts of the exam and introduce you to the real exam scenario and practice it with the help of its testing engine and real exam dumps

Go to page:
Question # 9

An Architect is designing partitioned external tables for a Snowflake data lake. The data lake size may grow over time, and partition definitions may need to change in the future.

How can these requirements be met?

A.

Use the PARTITION BY () clause when creating the external table.

B.

Use partition_type = USER_SPECIFIED when creating the external table.

C.

Set METADATA$EXTERNAL_TABLE_PARTITION = MANUAL.

D.

Alter the table using ADD_PARTITION_COLUMN before defining a new partition column.

Full Access
Question # 10

The following DDL command was used to create a task based on a stream:

Assuming MY_WH is set to auto_suspend – 60 and used exclusively for this task, which statement is true?

A.

The warehouse MY_WH will be made active every five minutes to check the stream.

B.

The warehouse MY_WH will only be active when there are results in the stream.

C.

The warehouse MY_WH will never suspend.

D.

The warehouse MY_WH will automatically resize to accommodate the size of the stream.

Full Access
Question # 11

What is a key consideration when setting up search optimization service for a table?

A.

Search optimization service works best with a column that has a minimum of 100 K distinct values.

B.

Search optimization service can significantly improve query performance on partitioned external tables.

C.

Search optimization service can help to optimize storage usage by compressing the data into a GZIP format.

D.

The table must be clustered with a key having multiple columns for effective search optimization.

Full Access
Question # 12

A new user user_01 is created within Snowflake. The following two commands are executed:

Command 1→ SHOW GRANTS TO USER user_01;

Command 2→ SHOW GRANTS ON USER user_01;

What inferences can be made about these commands?

A.

Command 1 defines which user owns user_01Command 2 defines all the grants which have been given to user_01

B.

Command 1 defines all the grants which are given to user_01Command 2 defines which user owns user_01

C.

Command 1 defines which role owns user_01Command 2 defines all the grants which have been given to user_01

D.

Command 1 defines all the grants which are given to user_01Command 2 defines which role owns user_01

Full Access
Question # 13

A user has the appropriate privilege to see unmasked data in a column.

If the user loads this column data into another column that does not have a masking policy, what will occur?

A.

Unmasked data will be loaded in the new column.

B.

Masked data will be loaded into the new column.

C.

Unmasked data will be loaded into the new column but only users with the appropriate privileges will be able to see the unmasked data.

D.

Unmasked data will be loaded into the new column and no users will be able to see the unmasked data.

Full Access
Question # 14

A media company needs a data pipeline that will ingest customer review data into a Snowflake table, and apply some transformations. The company also needs to use Amazon Comprehend to do sentiment analysis and make the de-identified final data set available publicly for advertising companies who use different cloud providers in different regions.

The data pipeline needs to run continuously ang efficiently as new records arrive in the object storage leveraging event notifications. Also, the operational complexity, maintenance of the infrastructure, including platform upgrades and security, and the development effort should be minimal.

Which design will meet these requirements?

A.

Ingest the data using COPY INTO and use streams and tasks to orchestrate transformations. Export the data into Amazon S3 to do model inference with Amazon Comprehend and ingest the data back into a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.

B.

Ingest the data using Snowpipe and use streams and tasks to orchestrate transformations. Create an external function to do model inference with Amazon Comprehend and write the final records to a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.

C.

Ingest the data into Snowflake using Amazon EMR and PySpark using the Snowflake Spark connector. Apply transformations using another Spark job. Develop a python program to do model inference by leveraging the Amazon Comprehend text analysis API. Then write the results to a Snowflake table and create a listing in the Snowflake Marketplace to make the data available to other companies.

D.

Ingest the data using Snowpipe and use streams and tasks to orchestrate transformations. Export the data into Amazon S3 to do model inference with Amazon Comprehend and ingest the data back into a Snowflake table. Then create a listing in the Snowflake Marketplace to make the data available to other companies.

Full Access
Question # 15

How is the change of local time due to daylight savings time handled in Snowflake tasks? (Choose two.)

A.

A task scheduled in a UTC-based schedule will have no issues with the time changes.

B.

Task schedules can be designed to follow specified or local time zones to accommodate the time changes.

C.

A task will move to a suspended state during the daylight savings time change.

D.

A frequent task execution schedule like minutes may not cause a problem, but will affect the task history.

E.

A task schedule will follow only the specified time and will fail to handle lost or duplicated hours.

Full Access
Question # 16

Which technique will efficiently ingest and consume semi-structured data for Snowflake data lake workloads?

A.

IDEF1X

B.

Schema-on-write

C.

Schema-on-read

D.

Information schema

Full Access
Go to page: