Requirement - Snowflake Data Engineer

0 views
Skip to first unread message

vamsi s

unread,
Jan 29, 2026, 9:06:20 AM (18 hours ago) Jan 29
to vam...@sritechsolutions.com

Tittle : Snowflake Data Engineer

Location: Dallas TX

Onsite position 

Experience:12+


Key Responsibilities

  • Design and implement Snowflake schemas (star, snowflake, data vault) optimized with micro-partitioning, clustering keys, materialized views, and search optimization services.
  • Build real-time and batch ingestion pipelines into Snowflake using Snowpipe, Kafka Connect, Fivetran, Matillion, Informatica, or dbt.
  • Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture).
  • Use Zero-Copy Cloning for environment management, testing, and sandboxing.
  • Apply Time Travel and Fail-safe features for data recovery and auditing.
  • Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake.
  • Design integrations with cloud storage (S3, Azure ADLS, GCS) for staging and external tables.
  • Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace.
  • Enable semi-structured data handling (JSON, Avro, Parquet, ORC, XML) using VARIANT columns and lateral flattening.
  • Integrate Snowflake with BI tools (Power BI, Tableau) via live connections and semantic layers.
  • Implement RBAC (Role-Based Access Control), Row Access Policies, and Dynamic Data Masking for data security.
  • Optimize compute usage with multi-cluster warehouses, resource monitors, and query performance tuning.
  • Manage cost optimization strategies (warehouse auto-suspend, query profiling, storage/compute separation).
  • Integrate with data catalog & governance platforms (Collibra, Alation, Informatica CDGC) using Snowflake metadata and APIs.
  • Work with domain teams to deliver data products leveraging Snowflake’s data mesh-friendly features.
  • Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers.
  • Support CI/CD automation for Snowflake code deployment using GitHub Actions, Azure DevOps, or dbt Cloud.

Qualifications

Education: Bachelor’s or Master’s in Computer Science, Data Engineering, or related field.

Experience:

  • 10+ years of data engineering experience, with 5+ years in Snowflake Data Cloud.
  • Expertise in SQL optimization and Snowflake performance tuning.
  • Hands-on with Snowpipe, Streams & Tasks, Snowpark, Zero-Copy Cloning, and Secure Data Sharing.
  • Proficiency in Python, Scala, or Java for Snowpark development.
  • Experience integrating with cloud platforms like AWS.
  • Exposure to ETL/ELT tools (Informatica, Matillion, Fivetran).
  • Familiarity with CI/CD, Git, DevOps practices for data operations.
  • Preferred Certifications:
  • SnowPro Core Certification

Key Skills

  • Snowflake-native feature design and implementation (Snowpark, Streams, Time Travel, Secure Data Sharing)
  • Data ingestion (Snowpipe, CDC, Kafka, Fivetran)
  • Semi-structured data handling (VARIANT, JSON, Avro, Parquet)
  • Advanced SQL and performance tuning
  • Data governance (RBAC, masking, lineage, catalogs)
  • Cloud data platform integrations (AWS S3, Azure ADLS, GCP GCS)
  • BI and analytics tool integration
  • Cost optimization and warehouse orchestration.

 With Regards,  


Vamsi Sattaru US IT Technical Recruiter

Reply all
Reply to author
Forward
0 new messages