- Company Name
- Saki Soft Limited
- Job Title
- AWS Data Engineer
- Job Description
-
Job Title: AWS Data Engineer
Role Summary:
Design, develop, and maintain scalable ETL/ELT pipelines in AWS, leveraging Python, PySpark, and SQL to transform data for analytics. Build and optimize data models in Amazon Redshift, integrate multi-source data (batch and streaming), and ensure high-quality, secure, and compliant data services.
Expectations:
- 5+ years of data engineering experience in cloud environments.
- Strong proficiency in Python and PySpark.
- Advanced SQL skills with performance tuning and window functions.
- Expertise with Amazon Redshift, RDS, Glue, Lambda, S3, EventBridge, and Kinesis.
- Proven record of building large‑scale, distributed data pipelines.
- Familiarity with Git, CI/CD principles, and workflow orchestration (Airflow preferred).
- Commitment to data security, governance, and best‑practice compliance in AWS.
Key Responsibilities:
- Engineer robust ETL/ELT pipelines using Python and PySpark.
- Design and refine star/snowflake schemas for Redshift analytics.
- Optimize Redshift performance through tuning and query optimization.
- Write complex SQL queries and develop data transformation logic.
- Integrate batch and streaming data sources into the data warehouse.
- Ensure pipeline reliability, scalability, and data quality.
- Collaborate with analytics, product, and engineering teams to capture data requirements.
- Implement monitoring, logging, and error‑handling for all pipelines.
- Enforce AWS data security, governance, and compliance standards.
Required Skills:
- Python (data engineering), PySpark/Apache Spark.
- Advanced SQL (CTEs, window functions, query optimization).
- Amazon Redshift, RDS, Glue, Lambda, S3, EventBridge, Kinesis.
- Version control using Git.
- Familiarity with CI/CD for data platforms.
- Knowledge of data warehousing concepts (fact/dimension tables, star/snowflake schemas).
- Experience with large‑scale data and distributed processing.
- Optional: Airflow, Kafka, data visualization/BI tools, CI/CD pipelines.
Required Education & Certifications:
- Bachelor’s degree in Computer Science, Information Systems, Engineering, or related field.
- AWS certifications (e.g., AWS Certified Solutions Architect or Data Engineer) preferred.