Senior Data Engineer

  • Full-time

Company Description

Homzmart is an end to end community that connects home professionals with homeowners who can browse home designs, furniture, accessories and products and also can find reviews, referring professionals, take advice from professionals and read articles about home improvement.

Job Description

We are seeking experienced data engineers to help us with rapid expansion of our Data Cloud, constructing new data pipelines and buildout of the Snowflake cloud data warehouse structures for various stakeholders and analytics applications, using most modern techniques and technologies. Our modern solution features standardized data ingestion pipelines that support low latency of data updates from the sources and enable a variety of analytics applications.

What you’ll be doing:

  • Build a new data ingestion pipeline into our data cloud, while helping guide the decisions about the future of our data infrastructure.
  • Be the key contributor in the assigned domains for the data pipelines and new data lake and cloud data warehouse structures being created for analytics and reporting or AI/ML applications.
  • Contribute to the development of the data models, data dictionaries and catalog references in the application domains of the primary responsibility.
  • Maintain and monitor our data pipeline processes and data cloud zones to ensure timely, quality and accurate data.
  • Steward our data cloud, through the definition of standards and best practices around data usage, including metadata cataloging and organization.
  • Contribute to our team's thought leadership for data cloud and operational processes lifecycle and tooling.

Qualifications

  • BS in an Engineering or Science discipline, or equivalent experience.
  • 3-5+ years of software/data engineering experience using Java, Scala, and/or Python, with at least 5 years experience in a data focused role.
  • Experience in data integration (ETL/ELT) development using multiple languages (e.g., Java, Scala, Python, PySpark, SparkSQL).
  • Experience building and maintaining data pipelines supporting a variety of integration patterns (batch, replication/CDC, event streaming) and data lake/warehouse in production environments.
  • Experience with AWS-based data services technologies (e.g., Kinesis, Glue, RDS, Athena, Snowflake, etc.)
  • Experience of working in the larger initiatives building and rationalizing large scale data environments with a large variety of data pipelines, possibly with internal and external partner integrations, would be a plus.
  • Familiarity with frameworks for Audit, Balance and Control (ABC) for managing data pipelines operations and data quality, would be a plus.
  • Willingness to experiment and learn new approaches and technology applications
  • Knowledge and experience with various databases (e.g., SQL Server, MySQL, PostgreSQL, MongoDB, Teradata) and SQL.
  • Knowledge of software engineering and agile development best practices.
  • Excellent written and verbal communication skills.

Additional Information

location: Katameya - Cairo, Egypt