- Company Name
- pass Culture
- Job Title
- Data Engineer - CDI - Paris
- Job Description
-
**Job title**
Data Engineer
**Role Summary**
Design, develop, and deploy scalable ELT pipelines that feed a BigQuery data warehouse. Ensure data integrity, reliability, and performance through automated testing, monitoring, and infrastructure optimization. Support data analysts and scientists with services that streamline data access, reverse‑ETL, and machine‑learning workflows.
**Expectations**
* Minimum 3 years of experience as a Data Engineer or ML Engineer.
* Proficient in Python, Docker, Airflow, and dbt.
* Demonstrated expertise building modern data pipelines on cloud platforms, especially GCP (BigQuery, GCS, GKE, Vertex AI).
* Experience with infrastructure automation tools such as Terraform, GitHub Actions, and ArgoCD.
* Familiarity with Kubernetes and API development in production settings.
* Strong analytical ability, prioritisation skills, and rigorous problem‑diagnosis.
* Excellent communication in French; technical English proficiency.
* Passion for documentation, knowledge sharing, and collaborative teamwork.
**Key Responsibilities**
1. Contribute to the design, development, and deployment of robust, scalable ELT pipelines feeding BigQuery.
2. Implement automated tests, monitoring, and alerting (Dbt, elementary, Grafana) to guarantee data quality and reliability.
3. Build and maintain services that empower analysts and scientists, including reverse‑ETL, external API integration, and ML‑feature provision.
4. Optimize infrastructure to support growth, ensuring efficient processing times and cost‑effective resource utilisation.
5. Participate in data‑strategy rituals (stand‑ups, demos, retrospectives) and mentor data‑team members on engineering practices.
**Required Skills**
* Python, Docker, Airflow, dbt
* GCP services: BigQuery, Cloud Storage, GKE, Vertex AI, CloudSQL
* Infrastructure‑as‑code: Terraform, GitHub Actions, ArgoCD
* Monitoring & observability: Grafana, Cloud Monitoring, elementary
* API design & deployment (Rest, Cloud Endpoints or equivalent)
* Knowledge of ClickHouse, Metabase, Looker, ML frameworks (TensorFlow/PyTorch, MLFlow) is a plus
* Strong command of Linux, version control (Git), and CI/CD pipelines
**Required Education & Certifications**
* Bachelor’s or Master’s degree in Computer Science, Data Engineering, Software Engineering, or a related field.
* Certifications such as Google Cloud Certified – Professional Data Engineer are highly desirable.