Please send me the profiles at dee...@godigitive.com
Position : Data Lead Developer (AZURE)
Location: Omaha, NE
Duration: 6 months to start with (immediate position)
Responsibilities:
• Designing and implementing highly performant data ingestion pipelines from multiple sources using Apache Spark and/or Azure Databricks ensuring that the infrastructure is highly-available and secure.
• Develop scalable and re-usable frameworks for ingestion of data from centralized and federated hubs
• Architect and build security compliant user management framework for multi-tenant big data platform
• Performance-tune Apache Spark applications to optimize cluster configurations
• Set up and maintain infrastructure for RESTful microservices layer
• Integrate the end to end data pipeline to take data from source systems to target data repositories ensuring the quality and consistency of data is maintained at all times
• Collaborate across internal functional groups and external vendors to incorporate emerging/novel big data tools
• Help maintain and support the platform on a day-to-day basis including on-boarding new members
• Work closely with the product management and development teams to rapidly translate the understanding of customer data and requirements to product and solutions
• Enable ETL/ELT solutions and perform DB tuning, table partitioning, shell scripting, drive prototypes and POCs
Qualifications:
• Bachelor's Degree in Computer Science
• Minimum 8 years of experience as a technology leader with strong knowledge of Data Management principles and dimensional modelling concepts
• Advanced applied knowledge of Spark – a must.
• Proven knowledge of designing, developing and deploying federated data architecture – highly desirable
• Hands on experience designing and delivering solutions using with Microsoft Azure data services (Azure Data Lake, Azure Data Factory, Azure ML, Azure SQL, Azure Datawarehouse, Azure DataBricks, Azure Cosmos DB) -– highly desirable
• Experience with event based / streaming technologies to ingest and process data (Apache KAFKA, Nifi, Azure Event Hub) – highly desirable
• Experience with OS non-relational/NoSQL data repositories (incl. MongoDB, Cassandra, Neo4J) - desirable
• Familiarity Open Source big data products Hadoop (incl. Hive, Pig, Impala) – desirable
• Broad based architecture acumen: Database architecture, ETL, SOA, Cloud, etc - desirable
• Ability to work successfully in a distributed team environment
Deepak Gulia | Digitive LLC – cloud. made simple
Fax: 408-935-8696 | Email: dee...@godigitive.com
GTALK :- deepakguli...@gmail.com