Hello,
Hope you are doing well.
Please have a look at the below job description and share some relevant profiles for this opportunity.
Role : Snowflake Developer
Location : NYC, NY (Onsite from Day 1 with Hybrid)
Experience : 11+ Years
Duration : Contract
Note: No H1b and Looking for Locals candidates of NY.
Primary Responsibilities:
· Architecture, design, implementation and operationalization of large-scale data and analytics solutions on Snowflake Cloud Data Warehouse.
· Hands-on experience with Snowflake utilities such as SnowSQL, SnowPipe, Python, Tasks, Streams, Time travel, Optimizer, Metadata Manager, data sharing, and stored procedures.
· Experience in Data warehousing - OLTP, OLAP, Dimensions, Facts, and Data modeling.
· Need to have working knowledge of MS Azure configuration items with respect to Snowflake.
· Developing EL pipelines in and out of data warehouse using combination of Databricks, Python and SnowSQL.
· Developing scripts UNIX, Python etc. to do Extract, Load and Transform data.
· Provide production support for Data Warehouse issues such data load problems, transformation translation problems
· Translate mapping specifications to data transformation design and development strategies and code, incorporating standards and best practices for optimal execution.
· Understanding data pipelines and modern ways of automating data pipeline using cloud based testing and clearly document implementations, so others can easily understand the requirements, implementation, and test conditions.
· Perform code reviews to ensure fit to requirements, optimal execution patterns and adherence to established standards.
· Establishing and monitoring Operational Level Agreements for the health and performance/cost of the warehouse environment (Loads, queries, data quality)
Preferred Skills:
· Bachelor’s degree in computer science, Engineering, Technical Science
· 3 years of technical architecture and build experience with large-scale data warehouse solutions.
· Code optimization expertise
· Experience building data ingestion pipelines using Python and Databricks in working with MS Azure.
· 3 years’ experience in Finance / Banking industry – some understanding of Securities and Banking products and their data footprints.
Basic Qualifications:
· Minimum 3 years of designing and implementing an operational production grade large-scale data solution on MA Azure Snowflake Data Warehouse.
· Including hands on experience with productionized data ingestion and processing pipelines using Python, Databricks, SnowSql
· Excellent understanding of Snowflake Internals and integration of Snowflake with other data processing and reporting technologies
· Excellent presentation and communication skills, both written and verbal ability to problem solve and design in an environment with unclear requirements.
· Ability to lead and drive performance of a team of developers – local and offshore.
Warm Regards!
Sandeep Aggarwal
Direct: +1 (848-668-9626)
Email: sand...@centraprise.com
Connect with me: www.linkedin.com/in/sandeep-aggarwal-434413245