cover image
Saki Soft Limited

Saki Soft Limited

www.sakisoft.co.uk

1 Job

11 Employees

About the Company

Welcome to Saki Soft Limited. Saki Soft provides Business Intelligence,strategic planning of the resources which are not abundant to keep its client KPIs upward through it highly qualified and dedicated Team. We strive for Quality and Customer Satisfaction. We provide recruitment and HR services.

Listed Jobs

Company background Company brand
Company Name
Saki Soft Limited
Job Title
AWS Data Engineer
Job Description
Job Title: AWS Data Engineer Role Summary: Design, develop, and maintain scalable ETL/ELT pipelines in AWS, leveraging Python, PySpark, and SQL to transform data for analytics. Build and optimize data models in Amazon Redshift, integrate multi-source data (batch and streaming), and ensure high-quality, secure, and compliant data services. Expectations: - 5+ years of data engineering experience in cloud environments. - Strong proficiency in Python and PySpark. - Advanced SQL skills with performance tuning and window functions. - Expertise with Amazon Redshift, RDS, Glue, Lambda, S3, EventBridge, and Kinesis. - Proven record of building large‑scale, distributed data pipelines. - Familiarity with Git, CI/CD principles, and workflow orchestration (Airflow preferred). - Commitment to data security, governance, and best‑practice compliance in AWS. Key Responsibilities: - Engineer robust ETL/ELT pipelines using Python and PySpark. - Design and refine star/snowflake schemas for Redshift analytics. - Optimize Redshift performance through tuning and query optimization. - Write complex SQL queries and develop data transformation logic. - Integrate batch and streaming data sources into the data warehouse. - Ensure pipeline reliability, scalability, and data quality. - Collaborate with analytics, product, and engineering teams to capture data requirements. - Implement monitoring, logging, and error‑handling for all pipelines. - Enforce AWS data security, governance, and compliance standards. Required Skills: - Python (data engineering), PySpark/Apache Spark. - Advanced SQL (CTEs, window functions, query optimization). - Amazon Redshift, RDS, Glue, Lambda, S3, EventBridge, Kinesis. - Version control using Git. - Familiarity with CI/CD for data platforms. - Knowledge of data warehousing concepts (fact/dimension tables, star/snowflake schemas). - Experience with large‑scale data and distributed processing. - Optional: Airflow, Kafka, data visualization/BI tools, CI/CD pipelines. Required Education & Certifications: - Bachelor’s degree in Computer Science, Information Systems, Engineering, or related field. - AWS certifications (e.g., AWS Certified Solutions Architect or Data Engineer) preferred.
London, United kingdom
On site
29-12-2025