Data Engineer
- Full-time
Company Description
Inetum is a global leader in IT services, dedicated to providing innovative solutions to our clients. We are committed to fostering a dynamic, inclusive workplace that values diversity, where creativity and collaboration thrive. We operate in 19 countries with more than 28,000 employees worldwide.
If you are looking for a dynamic, innovative, and technology-driven company, Inetum is the place for you! Come be Inetum!
Job Description
The Data Engineer will be responsible for designing, developing, and maintaining scalable and reliable data pipelines for a financial services project. The role focuses on backend data processing, data quality, and integration of multiple data sources in a cloud-based environment, working closely with international teams.
Key Responsibilities
- Design, develop, and maintain end-to-end ETL/ELT data pipelines to process large volumes of structured and semi-structured data.
- Implement backend data solutions using Python and SQL, applying Object-Oriented Programming (OOP) to ensure modularity, reusability, and maintainability.
- Orchestrate data workflows using Apache Airflow, including scheduling, monitoring, and failure handling.
- Process and transform large datasets using PySpark in distributed environments.
- Integrate data from multiple sources, including APIs, relational databases, and cloud storage systems.
- Manage and utilize AWS S3 for data storage and data lake architectures.
- Apply data quality checks, validation rules, and deduplication logic to ensure data consistency and accuracy.
- Develop, maintain, and support CI/CD pipelines using Bitbucket, ensuring controlled deployments, versioning, and code quality.
- Collaborate with cross-functional and international teams, contributing to technical discussions and documentation in English.
- Support downstream data consumers by ensuring datasets are well-structured, documented, and ready for analytics or reporting.
- Troubleshoot and resolve data pipeline issues, performance bottlenecks, and data inconsistencies.
Qualifications
- Programming Languages: Python, SQL
- Programming Paradigms: Object-Oriented Programming (OOP)
- Data Processing: PySpark
- Orchestration: Apache Airflow
- CI/CD: Bitbucket
- Cloud & Storage: AWS (S3)
- Data Sources: APIs, relational databases, parquet files
- Data Architecture: ETL/ELT pipelines, data lakes
Required Skills & Experience
- Strong experience in data engineering and backend data development.
- Solid knowledge of Python and SQL, with practical application of OOP principles.
- Experience building and maintaining production-grade ETL/ELT pipelines.
- Hands-on experience with Apache Airflow for workflow orchestration.
- Experience with CI/CD practices
- Experience working with distributed data processing frameworks such as Spark / PySpark.
- Familiarity with cloud-based data platforms, preferably AWS.
- Ability to work autonomously while collaborating with remote, international teams.
- Professional working proficiency in English.
Nice to Have
- Experience in financial services or regulated environments.
- Familiarity with data quality frameworks, monitoring, or observability tools.
- Exposure to Oracle Apex.
- Experience working in agile and/or DevOps-oriented teams.
Additional Information
The candidate is expected to work in a Hybrid model, 50/50 frame work.