Data Engineer Unidad editorial

Descripción de la empresa

Unidad Editorial, S.A. is one of the leading Spanish multimedia publishing groups, active in all areas of publishing, from newspapers to magazines, from digital to books, as well as being among the main players in the advertising market and to organize events and sports formats of primary importance.

Descripción del empleo

 

Perfil necesario para la correcta prestación del servicio:

The Data Department of Unidad Editorial, with the aim of being a partner in the strategic and digital evolution of business areas with a data-driven logic, is strengthening its structure with selection of talents in data science, data engineering and data analysis.

In particular, within the Data Engineering team, we are looking for a Senior Data Engineer who wants to get involved in order to support our reality in this rapid digital development process.

The identified person will collaborate with the team ensuring all the activities needed to implement and build the data ecosystem so that the data can be used by data scientists, data visualization experts and data analysts to support the business. In particular, it will have to deal with:

  • Implementation of data pipeline for ingestion, processing, enrichment and aggregation of the same in Data Lake or Data warehouse environment.
  • Implementation of Data Quality checks on all data pipelines.
  • Cloud computing platform performance and cost monitoring.
  • Standardization and automation of data management processes.

Activity detail:

  • Ingestion of enterprise and third party systems data flows, in the cloud or on-prem.
  • Creation of enrichment and aggregation data pipeline for data model, report, etc.
  • Maintenance and evolution of data architecture.
  • Production of pipelines produced by Data Scientists.
  • Industrialization of ML models.
  • Application of Data Quality techniques for formal and quantitative data checks.
  • Efficient use of the cloud computing platform.
  • Monitoring the use of resources and related costs.
  • Maintenance and code evolution.
  • Creating shared code libraries.
  • Definition of standardized development processes.

 

It requires at least 3-5 years of experience in the role and in particular:

Requisitos

  • Experience in building data-driven solutions in cloud-native environments.
  • Experience with data quality and data governance tools and techniques.
  • Experience with code version control tools such as Git.
  • Knowledge and use of at least one of the cloud computing environments: GCP, AWS or Azure.
  • In-depth knowledge of Python, SQL and pyspark languages, welcome knowledge of Bash.
  • Knowledge of the main Big Data frameworks in particular Apache Spark, Hadoop, Hive, Airflow.
  • Knowledge of the principles of modeling and management of data batch and in real-time.
  • Long-term vision of how to advance the data infrastructure to the next level by exploring and evaluating new technologies as appropriate.
  • Good ability to communicate in English.
  • Appreciated knowledge of Computer Science.
  • Welcome knowledge of visualization tools infrastructure (QlikSense/Tableau/Looker)
  • Welcome knowledge of technologies, concepts and analysis of the world’s own Data Scientist.
Política de privacidad