Data Engineer - Analytics/Modelling
Location: Birmingham/London, UK
Mode: Hybrid
Responsibilities
- Lead the design and implementation of AWS-based data products that leverage reusable datasets.
- Collaborate on creating scalable data solutions using a range of new and emerging technologies from the AWS platform.
- Demonstrate AWS data expertise when communicating with stakeholders and translating requirements into technical data solutions.
- Manage both real-time and batch data pipelines. Our technology stack includes a wide variety of technologies such as Kafka, AWS Kinesis, Redshift, and DBT.
- Design and model data workflows from ingestion to presentation, ensuring data security, privacy, and cost-effective solutions.
- Create a showcase environment to demonstrate data engineering best practices and cost-effective solutions on AWS.
- Build a framework suitable for stakeholders with low data fluency. This framework should enable easy access to data insights and facilitate informed decision-making.
Requirements
- Expertise in the full data lifecycle: project setup, data pipeline design, data modelling and serving, testing, deployment, monitoring, and maintenance.
- Strong data architecture background in cloud-based architectures (SaaS, PaaS, IaaS).
- Proven engineering skills with experience in Python, SQL, Spark, and DBT, or similar frameworks for large-scale data processing.
- Deep knowledge of AWS services relevant to data engineering, including AWS Glue, AWS EMR, Amazon S3, Redshift.
- Experience with Infrastructure-as-Code (IaC) using Terraform or AWS CloudFormation.
- Proven ability to design and optimize data models to address data quality and performance issues.
- Excellent communication and collaboration skills to work effectively with stakeholders across various teams.
- Ability to create user-friendly data interfaces and visualizations that cater to stakeholders with varying levels of data literacy.