Job Title | : Lead Data Engineer |
Location | : San Jose, CA (Onsite) |
Type | : Long term contract |
Note: Need candidates local to CA
Must have minimum 13 years of experience
Job Description:
Successfully lead the design, development, and execution of scalable data engineering solutions and migration strategies. Ensure seamless data movement from legacy systems to modern platforms with minimal downtime and data integrity. Deliver optimized data pipelines, enforce data governance standards, and enable analytics readiness. Drive technical excellence, mentor engineering teams, and collaborate with stakeholders to align data solutions with business goals.
Required Skills:
- 5+ years of experience in data engineering and data platform development.
- 5+ years of experience contributing enterprise data projects involving Azure-based ETL solutions.
- 5+ years of experience with Databricks & SQL
- 5+ years of in building ETL pipelines using Azure Data Factory
- Minimum of 8 years of experience in data engineering and architecture roles.
- Must have led 3 or more end-to-end enterprise data projects using Databricks, and Azure technologies.
Secondary Skills:
- Experience using Fivetran to automate data pipeline builds.
- Understanding of Databricks ML and analytics tools.
- Experience resolving networking/VPN issues related to data flow.
- Familiarity with data governance, security, and compliance frameworks.
Responsibilities:
- Lead the design and implementation of ETL solutions using SAP Data Services and Azure Data Factory.
- Leverage Fivetran for automated data ingestion from SAP S4 source systems into the Bronze layer.
- Analyze and migrate stored procedures from SAP HANA using SQL / PL/SQL to Databricks-based logic.
- Guide and mentor team members on data engineering best practices.
- Develop and maintain complex ETL pipelines using Python
- Identify and resolve performance bottlenecks and network-related issues.
- Ensure adherence to data governance and compliance standards across all data flows.
- Participate in performance tuning, issue resolution, and data validation tasks.
- Document data flows, pipeline logic, and lineage as part of project delivery.