AWS Data Engineer
Brooksource - charlotte, NC
Apply NowJob Description
Location: Charlotte, NC (Hybrid 2-3 days a week on site) 3-year contract with opportunity for extension or full-time hire W-2 Only, No Corp-to-Corp or 1099 Brooksource is searching for an AWS Data Engineer with experience in data warehousing using AWS Redshift to join our Fortune 500 Energy & Utilities client in Charlotte, NC. RESPONSIBILITIES: Provides technical direction, guides the team on key technical aspects and responsible for product tech delivery Lead the Design, Build, Test and Deployment of components Where applicable in collaboration with Lead Developers (Data Engineer, Software Engineer, Data Scientist, Technical Test Lead) Understand requirements use case to outline technical scope and lead delivery of technical solution Confirm required developers and skillsets specific to product Provides leadership, direction, peer review and accountability to developers on the product (key responsibility) Works closely with the Product Owner to align on delivery goals and timing Assists Product Owner with prioritizing and managing team backlog Collaborates with Data and Solution architects on key technical decisions The architecture and design to deliver the requirements and functionality Skilled in developing data pipelines, focusing on long-term reliability and maintaining high data quality Designs data warehousing solutions with the end-user in mind, ensuring ease of use without compromising on performance Manage and resolve issues in production data warehouse environments on AWS TECHNICAL REQUIREMENTS: 5+ years of AWS experience, specifically including AWS Redshift AWS services - S3, EMR, Glue Jobs, Lambda, Athena, CloudTrail, SNS, SQS, CloudWatch, Step Functions, QuickSight Experience with KafkaMessaging preferably Confluent Kafka Experience with EMR databases such as Glue Catalog, Lake Formation, Redshift, DynamoDB and Aurora Experience with Amazon Redshift for AWS data warehousing tools such as Amazon Redshift and Amazon Athena Proven track record in the design and implementation of data warehouse solutions using AWS Skilled in data modeling and executing ETL processes tailored for data warehousing Competence in developing and refining data pipelines within AWS Proficient in handling both real-time and batch data processing tasks Extensive understanding of database management fundamentals Expertise in creating alerts and automated solutions for handling production problems Tools and Languages - Python, Spark, PySpark and Pandas Infrastructure as Code technology - TerraformCloudFormation Experience with Secrets Management Platform like Vault and AWS Secrets manager Experience with Event Driven Architecture DevOps pipeline (CICD); Bitbucket; Concourse Experience with RDBMS platforms and Strong proficiency with SQL Experience with Rest APIs and API gateway Deep knowledge of IAM roles and Policies Experience using AWS monitoring services like CloudWatch, CloudTrail ad CloudWatch events Deep understanding of networking DNS, TCPIP and VPN Experience with AWS workflow orchestration tool like Airflow or Step FunctionsPREFFERED SKILLS Experience with native AWS technologies for data and analytics such as Kinesis, OpenSearch Databases - Document DB, Mongo DB Hadoop platform (Hive; HBase; Druid) Java, Scala, Node JS Workflow Automation Experience transitioning on premise big data platforms into cloud-based platforms such as AWS Strong Background in Kubernetes, Distributed Systems, Microservice architecture and containersADDITIONAL REQUIREMENTS Ability to perform hands on development and peer review for certain components tech stack on the product Standing up of development instances and migration path (with required security, accessroles) Develop components and related processes (e.g. data pipelines and associated ETL processes, workflows) Lead implementation of integrated data quality framework Ensures optimal framework design and load testing scope to optimize performance (specifically for Big Data) Supports data scientist with test and validation of models Performs impact analysis and identifies risk to design changes Ability to build new data pipelines, identify existing data gaps and provide automated solutions to deliver analytical capabilities and enriched data to applications Ability to implement data pipelines with the right attentiveness to durability and data quality Implements data warehousing products thinking of the end users experience (ease of use with right performance) Ensures Test Driven development 5+ years of Experience leading teams to deliver complex products Strong technical skills and communication skills Strong skills with business stakeholder interactions Strong solutioning and architecture skills 5+ years of Experience building real time data ingestion streams (event driven) Ensure data security and permissions solutions, including data encryption, user access controls and logging
Created: 2024-10-19