GES-C01 Online Practice Questions

Home / Snowflake / GES-C01

Latest GES-C01 Exam Practice Questions

The practice questions for GES-C01 exam was last updated on 2025-11-20 .

Viewing page 1 out of 24 pages.

Viewing questions 1 out of 124 questions.

Question#1

A data engineering team is setting up an automated pipeline in Snowflake to process call center transcripts. These transcripts, once loaded into a raw table, need to be enriched by extracting specific entities like the customer's name, the primary issue reported, and the proposed resolution. The extracted data must be stored in a structured JSON format in a processed table. The pipeline leverages a SQL task that processes new records from a stream.
Which of the following SQL snippets and approaches, utilizing Snowflake Cortex LLM functions, would most effectively extract this information and guarantee a structured JSON output for each transcript?


A. Option A
B. Option B
C. Option C
D. Option D
E. Option E

Explanation:
To guarantee a structured JSON output for entity extraction, (the updated version of 'COMPLETE()') with the response_format’ argument and a specified JSON schema is the most effective approach. This mechanism enforces that the LLM's output strictly conforms to the predefined structure, including data types and required fields, significantly reducing the need for post-processing and improving data quality within the pipeline.
Option A requires multiple calls and manual JSON assembly, which is less efficient.
Option B relies on the LLM's 'natural ability' to generate JSON, which might not be consistently structured without explicit ‘response_format’.
Option D uses, which is for generating summaries, not structured entity extraction.
Option E involves external LLM API calls and Python UDFs, which, while possible, is less direct than using native 'AI_COMPLETE structured outputs within a SQL pipeline in Snowflake Cortex for this specific goal.

Question#2

An ML Engineer has developed a custom PyTorch model for image processing that requires GPU acceleration and specific PyPl packages ('torch', 'torchvision'). They want to deploy it as a service on Snowpark Container Services (SPCS) using the Snowflake Model Registry.
Which of the following statements are true regarding the deployment of this model to SPCS and its requirements? (Select all that apply.)


A. Option A
B. Option B
C. Option C
D. Option D
E. Option E

Explanation:
StatementA is incorrect. While Snowflake recommends using only ‘conda_dependencies’ or only 'pip_requirements’ (not both) to avoid package conflicts, the scenario explicitly mentions PyPl packages. If using ‘pip_requirements’, all required packages should be listed there. The example incorrectly assumes ‘torchvision’ would necessarily be best sourced from Conda and dictates avoiding 'pip_requirements’ entirely, which is an oversimplification of the recommendation. Statement B is correct. To utilize GPU acceleration in SPCS, a compute pool configured with a GPU instance family (e.g., *GPU must be created and then referenced by name in the ‘service_compute_poor’ argument when creating the service. Statement C is correct. Snowflake's warehouse nodes have restricted directory access, and '/tmpP is recommended as a safe and writeable location for models that need to write files during execution. This principle extends to SPCS containers. Statement D is correct. The ‘create_service' method for deploying models to SPCS takes a gpu_requests argument, which specifies the number of GPUs to allocate to the service. Setting this (e.g., to "s) is crucial for ensuring the model runs on GPU hardware. Statement E is incorrect. The 'relax_version’ option, which modifies version constraints, defaults to 'True' in 'log_moder’ While often beneficial, it is not mandatory to explicitly set it to ‘True' for every deployment scenario.

Question#3

A data engineer is building a robust pipeline to process customer feedback. They need to extract specific sentiment categories (food_quality, food_taste, wait_time, food _cost) from text reviews and ensure the output is always a valid JSON object matching a predefined schema, even for complex reviews. They also want to control the determinism of the LLM responses.
Which of the following SQL statements or considerations are correct for achieving this using Snowflake Cortex AI functions?

A. The following SQL statement uses the response_format argument and temperature setting to achieve structured output and determinism:

B. The response_format argument with a JSON schema is primarily for OpenAl (GPT) models; for other models like Mistral, a strong prompt instruction such as 'Respond in strict JSON' is generally more effective.
C. To ensure the model explicitly attempts to extract all specified fields, the 'required' array in the JSON schema is critical; AI_COMPLETE will raise an error if any required field cannot be extracted.
D. Using AI_COMPLETE with response_format incurs additional compute cost for the overhead of verifying each token against the supplied JSON schema, in addition to standard token costs.
E. For the most consistent structured output, especially in complex reasoning tasks, setting the temperature option to 0 when calling AI_COMPLETE is recommended.

Explanation:
Option A is correct because it demonstrates the proper use of the 'AI_COMPLETE function with the ‘response_format’ argument to specify a JSON schema and sets ‘temperature’ to 0 for consistent output, as per the documentation.
Option C is correct as the "required'’ field in the JSON schema ensures that specific fields must be extracted, and 'COMPLETE (or 'AI_COMPLETE) will raise an error if these fields cannot be found.
Option E is correct because for the most consistent results, setting the ‘temperature* option to O is recommended when calling 'COMPLETE (or "AI_COMPLETE) with structured outputs, regardless of the task or model.
Option B is incorrect because all models supported by support structured output, and specifying the ‘response_format’ is the direct mechanism to enforce a schema, although for complex tasks, adding 'Respond in JSON' to the prompt can improve accuracy.
Option D is incorrect as 'AI_COMPLETE Structured Outputs incurs compute cost based on the number of tokens processed, but it does not incur additional compute cost for the overhead of verifying each token against the supplied JSON schema.

Question#4

A new Gen AI specialist is setting up Document AI. They have successfully created the necessary database, schema, and a custom role named 'doc_ai specialist_role'. This custom role has been granted the 'SNOWFLAKE.DOCUMENT INTELLIGENCE CREATOR database role. However, when the specialist attempts to create a new Document AI model build in Snowsight, they receive the error: Unable to create a build on the specified database and schema. Please check the documentation to learn more.
What is the most likely missing privilege for the that is preventing the model build creation?

A. 

B. 

C. 

D. 

E. The virtual warehouse assigned to the Document AI environment is not large enough to support model build creation.

Explanation:
The error message 'Unable to create a build on the specified database and schema' points directly to insufficient privileges for creating the model build itself. While the database role enables working on Document AI models and using SQL for extraction, specific object privileges are required to create the model ‘build’ in a given schema. The documentation explicitly states that 'CREATE SNOWFLAKE.ML.DOCUMENT_INTELLIGENCE on the schema is required to create model builds (instances of the 'DOCUMENT INTELLIGENCE' class), along with 'CREATE MODEL’ on the schema.
Option A (USAGE on database) is a general prerequisite but not the direct cause of this specific 'unable to create a build' error.
Option B CCREATE TABLE) is for creating tables, not model builds.
Option D (s EXECUTE TASK) is required for running tasks in a pipeline, not for model build creation.
Option E is incorrect as warehouse size primarily impacts cost and performance, not the ability to create a model build due to privileges.

Question#5

A data engineering team is implementing a solution using Snowflake Cortex's AI_COMPLETE function to process customer support tickets. They are concerned about sensitive information and ensuring the model's responses are safe, while adhering to Snowflake's data governance principles.
Which of the following statements correctly describe the functionality of Cortex Guard and Snowflake's data privacy commitments in this context?


A. Option A
B. Option B
C. Option C
D. Option D
E. Option E

Explanation:
Option B is correct because Cortex Guard evaluates the LLM's responses, not just the prompts, to filter unsafe content. Additionally, only the input tokens for Cortex Guard are counted for billing, and this cost is in addition to the function's cost.
Option E is correct because Snowflake's trust and safety principles explicitly state that your Usage and Customer Data (including inputs and outputs of Snowflake AI Features) are NOT available to other customers and are NOT used to train, re-train, or fine-tune Models made available to others.
Option A is incorrect as customer data is not anonymised and used to improve general model performance for others.
Option C is incorrect as the sources do not mention fine-tuning a customer-owned guardrail model.
Option D is incorrect as the sources do not mention a ‘CORTEX GUARD ADMIN’ role or logging of filtered content to a customer-accessible audit table.

Exam Code: GES-C01Q & A: 353 Q&AsUpdated:  2025-11-20

 Get All GES-C01 Q&As