Hi Team,
This side Utkarsh from 1Point System we have an opening for
Sr Data Engineer
(AWS, ETL, Python) with SAP HANA Database experience at Atlanta, GA,
Please have a look at complete JD and do let me know so that we can proceed further.
Job: SETL / AWS / Python
Developer
Sr Data Engineer
(AWS, ETL, Python) with SAP HANA Database experience
Long-Term Contract
Atlanta, GA | Onsite from day one
Visa: No H1B
Note from Client:
This This is a Sr. Data Engineer role but associated with Data Engineering
along with experience handling Linux systems and stored procedures in SAP
HANA database.
They won’t consider resumes without SAP HANA Database experience.
This person will assist in design, development, and implementation of
end-to-end complex ETL system using Informatica , Alteryx
and SAP HANA tools.
Data ingestion to Azure/ AWS and data lake and lake formation is
a big PLUS.
Qualifications:
- 5
years of relevant experience.
- Bachelor’s
Degree in computer science, Information Systems, or related field
- Good
experience in Python & SQL
- Good
Understanding of cloud computing and AWS architecture and best practices
- Development
experience on Lambda with Python or Java
- Must
have experience in using AWS services API, AWS CLI and SDK
- Working
experience on AWS services like EC2, S3, Route 53, Cloud watcher
- Python/ETL
Developer with Strong AWS experience
- Use
of AWS technologies for building, deploying, and operating applications
-Very Critical Requirement
- Understanding
of core AWS services and basic AWS architecture best practices (S3, EBS,
EC2, SQS/SNS, CloudFront, Route53, Lambda, CloudWatch, ECS Fargate, API
Gateway)
- Proficiency
in developing, deploying, and debugging cloud-based applications using AWS
ETL native tools
- Understanding
of the use of containers in the development process
- Ability
to understand ETL process
- Expertise
in AWS ETL tool Glue, EMR, Redshift, RDS, SNS/SQS
- Ability
to design technical workflow end to end using AWS services
- Experience
with working in Data Lake and data pipeline
- Python,
SQL, PySpark, Spark SQL, NoSQL such as Document DB/Mongo DB
- Ability
to use a CI/CD pipeline to deploy applications on AWS
- Be
able and willing to mentor team members
Responsibilities:
- Architecting
and supporting ETL and data load processes
- Designing
processes that can extract/receive data from various heterogeneous source
systems
- Perform
data cleansing and transforming the data according to business rules
- Developing
reusable frameworks for data extraction, loading and cleansing
- Designing
and building change data capture processes and updating the DataMart accordingly
- Parquet
/ Avro to file formats - Building capabilities to exchange the data with
external partners in various formats (Parquet, Avro, XML, CSV, Flat File)
--
Thanks and Regards
Utkarsh Dwivedi| 1Point System LLC
Direct: __________ • utk...@1pointsys.com
115 Stone Village Drive • Suite C • Fort Mill, SC • 29708
An E-Verified company | An Equal Opportunity Employer
DISCLAIMER: If you have received this email in error or prefer not to receive such emails in the future, please notify by replying with a ''REMOVE'' in the subject line and your email address shall be removed immediately from the mailer list.