- Company Name
- The Nationwide Group
- Job Title
- Data Engineer – AI & AWS Focus
- Job Description
-
**Job Title:** Data Engineer – AI & AWS Focus
**Role Summary:**
Design, build, and maintain scalable data pipelines and cloud data warehouses on AWS to support AI/ML and analytics workloads. Collaborate with product, analytics, and AI teams in an Agile environment to deliver reliable, secure, and cost‑effective data solutions.
**Expectations:**
- Deliver end‑to‑end ETL/ELT pipelines for structured and unstructured data.
- Optimize data storage, compute, and query performance on AWS services.
- Ensure data quality, governance, and compliance with security standards.
- Support AI/ML model training, feature engineering, and deployment workflows.
- Continuously improve architecture based on emerging technologies and best practices.
**Key Responsibilities:**
1. Design and maintain ETL/ELT pipelines using AWS Glue, Kinesis, Lambda, and S3.
2. Develop and manage cloud data warehouses (Amazon Redshift, Snowflake).
3. Support relational (PostgreSQL, MySQL) and NoSQL (DynamoDB, MongoDB) databases; optimize schemas and queries.
4. Implement automated data quality checks, monitoring, and governance frameworks.
5. Optimize pipelines for scalability and cost efficiency on AWS.
6. Apply security best practices and ensure regulatory compliance.
7. Partner with product managers, analysts, and AI teams to align data infrastructure with business goals.
8. Stay current on AI, data engineering, and cloud technologies; recommend improvements.
**Required Skills:**
- AWS services: Glue, Redshift, S3, Lambda (strong proficiency)
- SQL and advanced database design; experience with MySQL
- Programming: Python, Java, or Scala for data pipelines and AI integration
- ETL tools: Glue, Informatica, DataStage
- Version control (Git) and CI/CD for data workflows
- Knowledge of AI/ML lifecycle (model training, feature engineering, deployment, monitoring)
- Familiarity with data visualization/reporting tools (Tableau, Power BI, Jaspersoft) – a plus
- Nice‑to‑have: GoAnywhere, SageMaker, EMR
**Required Education & Certifications:**
- Bachelor’s degree in Computer Science, Data Engineering, or a related field
- 4–6 years of data engineering experience, with ≥2 years focused on AWS cloud services
- Demonstrated experience enabling AI/ML workflows (formal certifications not required but AWS Certified Data Analytics or Machine Learning credentials are advantageous).