Need Only Locals Boston MA / Nashville TN
We are currently prioritizing GC candidates who can work directly (no third-party employers).
Job Title: Data
Engineer
Location: Hybrid-
Foxboro- Andover-Boston (limited seating).. Nashville TN will work but prefers
mass locations 3 DAYS ONSITE lOCALS ONLY
Duration: 3-5
Months
Interview process: 1-round
Start Date: ASAP
Top skills:
Python
AWS Glue
Cloud Technologies
Data Movement
Roles & Responsibilities
- Optimize Amazon Redshift performance (distribution keys, sort keys,
query tuning) and support Athena optimization.
- Build and operate ETL/ELT pipelines with AWS Glue; orchestrate
workflows with Airflow.
- Manage semantic layers and metadata to enable trustworthy analytics
and AI.
- Apply best practices for partitioning, compression, and columnar
storage.
- Monitor and troubleshoot data workflows for high availability and
reliability; automate observability and reporting.
- Automate data processes using Python, SQL, and AWS‑native tools.
- Enforce data security and governance (Lake Formation, IAM),
including row/column‑level controls and least‑privilege access.
- Support monitoring, auditing, and compliance using CloudWatch,
CloudTrail, and related services.
- Continuously improve architecture by adopting relevant AWS best
practices and emerging patterns.
- Collaborate with Operations, Data Governance, and PMO to meet
standards and delivery goals.
Qualifications
- Bachelor’s in Computer Science/IT or equivalent experience in data
management, integration, or data warehousing.
- Hands‑on with AWS: S3, KMS, Lambda, Glue/Spark, SQS, EventBridge,
Step Functions.
- Strong AWS networking fundamentals: VPC, subnets, routing, NAT
gateways, security groups.
- Expertise with Redshift concurrency scaling and Athena tuning.
- IAM proficiency: roles, policies, assume‑role, cross‑account
access.
- Advanced SQL (joins, window functions, aggregations) with
distributed engines (Redshift/Athena).
- Experience writing and maintaining Terraform.
- Solid understanding of Spark architecture, execution plans, and
performance tuning.
- Familiarity with serverless and event‑driven pipelines.
- NoSQL experience (DynamoDB, MongoDB).
- Strong communication, ability to juggle priorities in a fast‑paced
environment, and a collaborative, data‑driven mindset.