Categories
Top 44 Snowflake Interview Questions and Answers For Freshers
Snowflake is attaining momentum as the best cloud data warehouse solution because of its innovative features like separation of computing and storage, data sharing, and data cleaning. It gives support for popular programming languages like Java, Golang, .Net, Python, etc. Tech giants like Adobe systems, AWS, Informatica, Logitech, Looker are using the Snowflake platform to build data-intensive applications. Therefore, there is always a demand for Snowflake professionals.
1. What is a Snowflake cloud data warehouse?
2. What are the cloud platforms currently supported by Snowflake?
2) Google Cloud Platform (GCP)
3) Microsoft Azure (Azure)
3. Describe Snowflake computing.
4. What are the features of Snowflake?
1) Database and Object Closing
2) Support for XML
3) External tables
4) Hive meta store integration
5) Supports geospatial data
6) Security and data protection
7) Data sharing
8) Search optimization service
9) Table streams on external tables and shared tables
10) Result Caching
5. Explain Snowflake architecture
Three main layers make the Snowflake architecture - database storage, query processing, and cloud services.
1) Data storage - In Snowflake, the stored data is reorganized into its internal optimized, columnar, and optimized format.
2) Query processing - Virtual warehouses process the queries in Snowflake.
3) Cloud services - This layer coordinates and handles all activities across the Snowflake. It provides the best results for Authentication, Metadata management, Infrastructure management, Access control, and Query parsing.
6. What kind of SQL does Snowflake use?
7. What type of database is Snowflake?
8. What ETL tools do you use with Snowflake?
1) Matillion
2) Blendo
3) Hevo Data
4) StreamSets
5) Etleap
9. Is Snowflake an ETL tool?
In data engineering, new tools and self-service pipelines are displacing traditional tasks such as manual ETL coding and data cleaning. With Snowflake's simple ETL and ELT options, data engineers can spend more time focusing on essential data strategy and pipeline improvement initiatives. Furthermore, using the Snowflake Cloud Platform as your data lake and data warehouse, extract, convert, and load may be efficiently avoided, as no pre-transformations or pre-schemas are needed.
10. What is the use of the Cloud Services layer in Snowflake?
11. Explain Columnar database
12. Is Snowflake OLTP or OLAP?
13. Explain Virtual warehouse
1) Execute the SQL SELECT statements that necessitate the use of computing resources (e.g. retrieving rows from tables and views).
14. How many editions of Snowflake are available?
1) Standard edition - Its introductory level offering provides unlimited access to Snowflake’s standard features.
2) Enterprise edition - Along with Standard edition features and services, offers additional features required for large-scale enterprises.
3) Business-critical edition - Also, called Enterprise for Sensitive Data (ESD). It offers high-level data protection for sensitive data to organization needs.
4) Virtual Private Snowflake (VPS) - Provides high-level security for organizations dealing with financial activities.
15. How is data stored in Snowflake?
16. How do we secure the data in the Snowflake?
17. Why is Snowflake highly successful?
1) It assists a wide variety of technology areas like data integration, business intelligence, advanced analytics, security, and governance.
2) It offers cloud infrastructure and supports advanced design architectures ideal for dynamic and quick usage developments.
3) Snowflake supports predetermined features like data cloning, data sharing, division of computing and storage, and directly scalable computing.
4) Snowflake eases data processing.
5) Snowflake provides extendable computing power.
6) Snowflake suits various applications like ODS with the staged data, data lakes with data warehouse, raw marts, and data marts with acceptable and modelled data.
18. What are the different ways to access the Snowflake Cloud data warehouse?
1) A web-based user interface from which all aspects of Snowflake management and usage can be accessed.
2) Command-line clients (such as SnowSQL) that can access all parts of Snowflake management and use.
3) Snowflake has ODBC and JDBC drivers, which allow other applications (like Tableau) to connect to it.
4) Native connectors (e.g., Python, Spark) for developing programmes that connect to Snowflake.
5) Third-party connectors can be used to link applications such as ETL tools (e.g., Informatica) and BI tools (e.g., ThoughtSpot) to Snowflake.
19. What is the use of the Compute layer in Snowflake?
20. What is the use of a database storage layer?
21. What are Micro Partitions?
22. Can AWS glue connect to Snowflake?
23. Tell me something about Snowflake AWS?
24. Describe Snowflake Schema
The benefits of using Snowflake schemas are it provides structured data and uses small disk space.
25. Explain Snowpipe in Snowflake
The data is loaded using the COPY command defined in a connected pipe. Snowpipe can use a pipe, which is a named, first-class Snowflake object containing a COPY statement. The COPY statement specifies the location of the data files (i.e., a stage) as well as the target table. All data types, including semi-structured data types like JSON and Avro, are supported.
There are several ways for detecting staged files:
1) Using cloud messaging to automate Snowpipe
2) REST endpoints in Snowpipe
The Snowpipe benefits are as follows:
1) Real-time insights
2) User-friendly
3)Cost-efficient
4) Resilience
26. Describe Snowflake Schema
The benefits of using Snowflake schemas are it provides structured data and uses small disk space.
An example of Snowflake Schema is shown below:
27. Explain Snowflake Time Travel
1) Restore the data-associated objects that may have lost unintentionally.
2) For examining the data utilization and changes done to the data in a specific time period.
3) Duplicating and backing up the data from the essential points in history.
28. What is the difference between Star Schema and Snowflake Schema?
Star Schema | Snowflake Schema |
The fact tables and dimension tables are both contained in the star schema. | The fact tables, dimension tables, and sub dimension tables are all contained in the snowflake schema. |
The star schema is a top-down model. | While it is a bottom-up model. |
The star schema takes up more space. | While it takes up less space. |
Queries are executed in less time. | Here query execution takes longer than with the star schema. |
Normalization is not employed in the star schema. | Both normalisation and denormalization are employed in this. |
It has a very simple design. | While its design is complex. |
Star schema has a low query complexity. | Snowflake schema has a higher query complexity than star schema. |
It contains fewer foreign keys. | It has a larger number of foreign keys. |
It has a high level of data redundancy. | While it has a minimal level of data redundancy. |
29. What is zero-copy Cloning in Snowflake?
Advantages:
1) There are no additional storage costs associated with data replication.
2) There is no waiting time for copying data from production to non-production contexts.
3) There is no need for administrative efforts since cloning is as simple as a click of a button.
4) No copy, only clone: Data exists only in one place.
5) Promote corrected/fixed data to production instantly.
30. What is Data Retention Period in Snowflake?
When data in a table is modified, such as deletion or discarding an object holding data, Snowflake saves the data's previous state. The data retention period determines the number of days that this historical data is kept and, as a result, Time Travel operations (SELECT, CREATE... CLONE, UNDROP) can be performed on it.
The standard retention period is one day (24 hours) and is enabled by default for all Snowflake accounts.
31. What is SnowSQL used for?
SnowSQL (snowsql executable) can be operated as an interactive shell or in batch mode via stdin or with the -f option.
32. What is the use of Snowflake Connectors?
The Snowflake connector can be used to execute the following tasks:
1) Read data from or publish data to tables in the Snowflake data warehouse.
2) Load data in bulk into a Snowflake data warehouse table.
3) You can insert or bulk load data into numerous tables at the same time by using the Numerous input connections functionality.
4) To lookup records from a table in the Snowflake data warehouse.
Following are the types of Snowflake Connectors:
1) Snowflake Connector for Kafka
2) Snowflake Connector for Spark
3) Snowflake Connector for Python
33. What are Snowflake views?
Non-materialized views (often referred to as "views") - The results of a non-materialized view are obtained by executing the query at the moment the view is referenced in a query. When compared to materialised views, performance is slower.
Materialized views - Although named as a type of view, a materialised view behaves more like a table in many aspects. The results of a materialised view are saved in a similar way to that of a table. This allows for faster access, but it necessitates storage space and active maintenance, both of which incur extra expenses.
34. Describe Snowflake Clustering
A clustering key is a subset of columns in a table (or expressions on a database) that are deliberately intended to co-locate the table's data in the same micro-partitions. This is beneficial for very large tables where the ordering was not perfect (at the time the data was inserted/loaded) or if extensive DML has weakened the table's natural clustering.
Some general indicators that can help determine whether a clustering key should be defined for a table are as follows:
1) Table queries are running slower than expected or have degraded noticeably over time.
2) The table's clustering depth is large.
35. Explain Data Shares
36. Does Snowflake use Indexes?
37. Where do we store data in Snowflake?
38. What is “Stage” in the Snowflake?
Internal Stages are further divided as below
1) Table Stage
2) User Stage
3) Internal Named Stage
39. Does Snowflake maintain stored procedures?
40. How do we execute the Snowflake procedure?
1) Run a SQL statement
2) Extract the query results
3) Extract the result set metadata
41. Explain Snowflake Compression
1) Storage expenses are lesser than original cloud storage because of compression.
2) No storage expenditure for on-disk caches.
3) Approximately zero storage expenses for data sharing or data cloning.
42. How to create a Snowflake task?
1) CREATE TASK in the schema.
2) USAGE in the warehouse on task definition.
3) Run SQL statement or stored procedure in the task definition.
43. Differentiate Fail-Safe and Time-Travel in Snowflake
According to the Snowflake edition, account or object particular time travel setup, users can retrieve and set the data reverting to the history.
Fail-Safe
Fail-Safe, the User does not have control over the recovery of data valuable merely after completing the period. In this context, only Snowflake assistance can help for 7 days. Therefore if you set time travel as six days, we retrieve the database objects after executing the transaction + 6 days duration.
44. How do we create temporary tables?
Create temporary table mytable (id number, creation_date date);