This space is dedicated to sharing insights, lessons, and thought leadership from real-world work in cloud, data, and AI. From architecting scalable pipelines on Google Cloud Platform to optimizing performance and data sharing in Snowflake, our posts aim to deepen technical competencies and spark meaningful conversations. Whether you’re a data enthusiast or a cloud architect, we hope you’ll find knowledge worth applying — and sharing.
Snowflake Certification Guides
Snowflake SnowPro Core Advanced Architecture
This exam evaluates architectural and design skills with understanding of the data ecosystem. Snowflake experience is essential to start preparing for this certtification.
Snowflake SnowPro Core Certification Guide
This guide is summary of key areas required to be studied for certification exam. The first part of blog shares details on exam , pattern, pre-requisites before you start preparing for certification.
Snowflake Blogs
Snowflake Query Insights
I am excited to learn about Snowflake Query_insights view and how this can be of a help to improve performance of queries. Snowflake’s focus on simplicity and ease of use always excites me most about the platform, features, and services.
Snowflake Summit 2025 Announcements
It was snowing in SFO in the month of June! I attended Snowflake summit this year and had an amazing experience listening to keynotes, sessions with lots of takeaways. It’s tough to cover everything in one blog! This is summary blog of the annoucements. Look out for upcoming blogs to get deeper insights of each of the managed services, services, and features announced this year.
Getting started with Snowflake AI Assistant
All about understanding the Snowflake AI Assistant — Snowflake Co-Pilot and using it to develop, optimize, get recommendations, generate sql queries and more. Lets begin with series of questions to understand more!
Snowflake Cortex LLM Function updates
This is the next blog in the series about understanding and using Snowflake Cortex functions. The first blog helps you to understand the remaining Snowflake Cortex functions and sample use cases to implement them.
Using Snowflake’s Data Metric Functions
This blog is the next blog in the series of “Data Quality” and “Data Governance” series. Refer to my earlier blog on understanding data metrics functions.
Understanding Snowflake’s Data Metric Functions
This blog is the next blog in the series of “Data Quality” and “Data Governance” series. Refer to my earlier blog on “Data Quality Essentials” that talks about necessity of Data Quality and metrics considered to validate data quality frameworks.
Using Snowflake Cortex-LLM Functions
This is the next blog in the series about understanding and using Snowflake Cortex functions. The first blog helps you to understand the remaining Snowflake Cortex functions and sample use cases to implement them.
Using Snowflake Cortex — COMPLETE function to generate responses to prompt
This blog in about understanding and using Snowflake Cortex functions. The first blog helps you to understand the Snowflake Cortex and its functions offered as part of the Cortex.
Understanding Snowflake Cortex functions
Snowflake Cortex is a managed service that offers AI and ML solutions to users. There are two types of functions offered by Snowflake:
- LLM Functions: These are SQL and Python-based functions that can be used to develop an understanding, query, translate, summarize, and generate free-form text.
2. ML-Based Functions: These are SQL functions based on the ML algorithms that are used to derive predictions.
Data Import and export using COPY
Data integrations are integral part of the data platform implementations. There are various integrators and tools available that can be used to integrate the applications or systems with Data platform. In a typical data platform architecture, you would have the source integrations that are required to be integrated to bring in the data to as source to the platform. You will also have bunch of consumer integrations and applications that runs on top of your “data” platform or consume “data” from your platform. In this blog, you will learn about the import and export of data required as part of source and consumer application integrations.
Data use cases with Snowflake
The data use cases blog series focuses on data analytics, consumption of the data, and data accessibility across regions, platforms, and users. This blog series starts with Snowflake offerings for data consumers.
Implement Cost Monitoring Dashboards using Snowsight
This blog focuses more on implementing the dashboards and monitoring techniques that will help to optimize the performance and cost as well as help in implementing the controls to monitor and report the utilization.
Implementing Snowflake’s Cost Optimization techniques
This blog focuses more on implementing the optimization techniques which will help to optimize the performance and cost as well as help in implementing the controls to monitor and report the utilization.
Understanding Snowflake Optimization services
Learn more on the optimization services available in Snowflake.
Undetstanding Warehouse Cost & Optimization
Snowflake’s compute or query processing layer is one of the layers of its architecture. Compute is referred to as Warehouse. Warehouses are available in various sizes and their consumption is calculated based on the usage in the form of credits. The earlier blog refers to the credit-to-cost computation. This blog refers to computing the warehouse cost and components that are considered as part of compute cost.
Understanding Snowflake Cost
This blog focuses more on understanding the various components of Snowflake that contribute to the total cost. You will learn more about the cost in terms of Q&A.
Snowflake Cost Optimization Series
Snowflake’s costing model is very transparent and easy to understand. You will be charged only for the services being used and only for the time they are up and running. You need to understand Snowflake’s architecture — three layered architecture and their features to understand the costing model well. You can refer to my earlier blogs to understand some of the foundational concepts, features required for the cost series.
Query Tagging in Snowflake
This blog focuses more on query tags , defining query tags, using query tags , use cases and benefits of using query tags.
Snowflake Row Access Policy
Data security is one of the critical pillar of Data governance architecture. Column level and Row level security is part of data security implementation. Earlier blogs in Data Governance series focus on column level security. This blog focuses on implementing Row level security in Snowflake.
Tag-based Masking Policy
This is the third blog in the Data Governance series focusing more on implementing tag-based policies in Snowflake.
Object Tagging in Snowflake
This blog helps you understand tagging in Snowflake. This also covers tagging details — how you can create them, use them, assign to the database objects, and track them for usage.
Capture Logs & Events using Event Tables in Snowflake
This blog is designed to share information on Snowflake Event tables. This blog covers details of Event tables — creation, usage, use cases to implement them, querying event tables, etc.
Building DataLake with Snowflake
This is the next blog in the series that helps you learn more about data lake, typical data lake challenges, designing a data lake and implementing it with Snowflake.
Dynamic Data Masking in Snowflake
Data governance implementation has various pillars. The data governance umbrella consist of many subcategories. Some of the pillars include — Access control, Data protection (classification), Data security (encryption ) , Data Quality, Data sharing etc. This blog will help you learn one of the most important security feature — Data Protection.
Cross cloud computing with Snowflake Snowgrid
You must have come across many scenarios, use cases or business requirements where you need to generate data , share or send over to other applications for their consumption.
In a typical implementations, you might have developed and deployed set of jobs, pipelines to generate aggregates or consumer data to be shared either by FTP or ETL jobs etc.
Snowflake Snowsight – New Features
Snowsight is Snowflake’s web user interface (UI). Snowflake also had Classic console — earlier Web UI which is available to old accounts as an option in Snowsight to access previous version of Web UI. With the recent release, Snowflake announced that Snowsight will be the default user interface for all accounts that are crated post May 2023.
Designing DataMesh with Snowflake
Data Mesh is an architectural approach and organizational model for data management that aims to address the challenges of scalability, autonomy, and agility in large-scale data systems. It proposes a paradigm shift from centralized data platforms to a decentralized and domain-oriented approach.
Architecting Data Warehouse Solutions with Snowflake
A data warehouse is a centralized repository of integrated data that is used for reporting, data analysis, and business intelligence purposes. It is designed to support the decision-making process by providing a single, comprehensive view of a business’s data from various sources.
Implementing ELT with Snowflake
You are designing a centralized data platform and keen on leveraging the platform capabilities to implement data integrations, develop data pipelines, and implement data analytics or AI/ML.
Setting up alerting with Snowflake Platform
Alerting or Alerts play a crucial role in any application, system implementation. Alerts are notifications, messages setup to notify existing issues as well as proactive notification upon reaching up given threshold values, limits of utilizations. We can refer these alerts as part of reactive and pro-active alerting mechanisms.
Change Data Capture using Snowflake Streams
We are going to learn more about implementing slowly changing dimensions or Change data capture in Snowflake using Snowflake — Streams.
Snowflake Data Protection – Part II
In this blog, we are going to learn more about data protection features. Snowflake has two different features — Time Travel and Failsafe.
Snowflake Data Protection – Part I
In this blog, we are going to learn more about data protection features. Snowflake has 2 different features — Time Travel and Failsafe.
Automated Account Management with Snowflake Snowsight
We are going to learn more about account management and how we can automate account management process using Snowsight dashboards.
Building Dashboards with Snowflake Snowsight
we are going to learn more about Snowsight — new Web UI. Snowsight is used to build dashboards and used natively with Snowflake.
Data Cloning in Snowflake
In this blog, we are going to learn about Data Cloning — one of distinguishing feature of Snowflake.
Snowflake – Data Sharing
Snowflake feature — Data Sharing is feature which allows to create/share objects required to be accessed or utilized by other teams. Data sharing enables users to share snowflake data objects like tables, external tables, secure views, secure UDFs, materialized views etc.
Snowflake Features
Snowflake is data on cloud with shared data architecture. There are many features of Snowflake while some of them are key features which makes Snowflake different than any other Datawarehouse offerings on any other cloud.
Snowflake Data On Cloud
Snowflake offers Data on Cloud offered as SaaS — Software as a Service . Snowflake is not build on any other Database or Hadoop/Bigdata platform. This is data offering built from scratch for cloud offering.
Google Cloud Platform Blogs
Bringing AI to Google BigQuery
GCP BigQuery is a serverless and cost-efficient data service of GCP. You can use GCP BQ to implement a data platform — data warehouse, or data lake on GCP. GCP BQ also offers ML implementation to integrate ML use cases in an SQLish way. BQ ML allows users to create and maintain models using BQ SQL. There are pre-trained models available to be used in BQ ML.
Data Fusion – ETL on Google Cloud
Cloud Data Fusion is a fully managed, cloud-native, enterprise data integration service for quickly building and managing data pipelines. Business users, developers, and data scientists can easily and reliably build scalable data integration solutions to cleanse, prepare, blend, transfer, and transform data without having to wrestle with infrastructure.
ELT with Google Cloud
ELT — Extract, Load and Transform — Building ELT pipelines are little different than building an ETL pipeline. There is difference in approach of implementation with these 2 types — ETL & ELT.
Serverless Implementations with GCP
Google Cloud’s serverless platform help us to build, develop, and deploy functions and applications, as source code or containers, while simplifying the developer experience by eliminating all infrastructure management.
ELT with Bigquery – Part II
In this blog, we are going to see some of the important aspects of designing/architecting ELT on GCP specifically when we have chosen Bigquery as DW/Data Lake to store data. Before explaining the specifics, lets look at the Bigquery and try to find out answer to many questions like access to data, retrieve data, data security, data loss prevention, run data analysis , data analytics, generate stats , building dashboards, exploring data through reporting and the most importantly how can we get ML models built/run on Bigquery?
Implementing Slowly changing dimensions with Google cloud services
We have seen different types of change data capture and ways to implement change data capture. Slowly Changing Dimensions(here after referred as SCD) can be implemented in many ways out of which SCD Type II is widely used enterprise solution. Let’s quickly recap on SCD Type II — This is Row versioning , Track changes as version records with current flag & active dates and other metadata. Type II implementations carry few of audit columns with each of its implementation to represent audit trail and active records.
Implementing Slowly Changing Dimensions using GCP Datastream
Datastream is a serverless and easy-to-use change data capture (CDC) and replication service. It allows you to synchronize data across heterogeneous databases and applications reliably, and with minimal latency and downtime.
Data Warehouse design recommendations using Google BigQuery
This blog shares some of design recommendations based on my previous experiences of GCP Data migrations , DW implementation with Google Bigquery.
Data Governance and Security with GCP BigQuery
Data security is securing data wherever its located, whether its in transit or stored or at rest, whether its managed/maintained by internal or external teams. This is critical to implement and ensure data security is in place whether data is in cloud or at on-premises systems.
Performance Optimization with Google BigQuery
This blog shares recommendations to optimize BigQiery performance.
BigQuery Pricing Model and Cost Optimization Recommendations
This blog shares BigQuery pricing models and optimization techniques for BQ pricing.
Migrating workloads to Google Cloud – Part I (Planning Migration)
Migrations are always challenging. I have worked on various migrations — Legacy to Legacy, Legacy to Bigdata and now from past few years, working on migrations to Cloud platform. Based on my past experiences, I am coming up with series of blogs on Migrations. Hope this blog series will help to plan, design, and implement the processes to migrate your applications/workloads to Google cloud.
Migrating workloads to Google Cloud – Part II (phase of migration)
This blog is second chapter of migration series, we are going to discuss more about migration path and phases of migration. In earlier blog , we have learnt that the assessment of existing system , defining goal, defining type of migration applicable to current project/application.
Migrating Workloads to Google Cloud – Part III (Reference Use cases)
This blog is third and last but one chapter of migration series, we are going to discuss some of the reference architectures and use cases of GCP Migration. In earlier blogs, we have learnt that the assessment of existing system , defining goal, defining type of migration applicable to current project/application. We also learnt the phases and designing phases of GCP Migrations.