Junior Data Engineer

  • 3401 North Thanksgiving Way, Lehi, UT 84043, USA
  • Full-time

Company Description

Ranked in the Inc. 500 Fastest Growing companies in the US, Jane is more than just a workplace; it’s a lifestyle. We’re a boutique marketplace featuring the latest in women's fashion trends, home decor, and children's clothing at prices that are too good to pass up. We work hard because we care about what we do and whether it’s done right. We are Jane.com.

Job Description

We are looking for someone to become a builder and trusted advisor to Jane by helping the company make better, evidence-based decisions that optimize our growth.  Jane’s Junior Data Engineer is expected to assist in preparing, cleaning, modeling, and loading data into our data warehouse and data take.  A data engineer at Jane is responsible for the development, construction, maintenance and testing of architectures, such as databases and large-scale processing systems, and is also expected to assist the analytics team in fulfilling data requests as needed.

Qualifications

  • Proficient and comfortable with SQL.
  • Knowledge of data warehouse dimensional modeling methods (star schema).
  • Proficient in at least one high-level programming language using data such as Python.
  • Strong desire to learn best-practice data engineering and data warehousing principles.
  • Able to understand business processes and how to measure and improve them in an organization.

WHAT YOU'LL BE HELPING TO DO

  • Ensure data platforms performance meets needs of reporting tools.

  • Performance tuning slow queries and reporting data structures.

  • Create and maintain data pipelines and monitoring data movement.

  • Profile data sources, create dimensional models, implement ETL, and load into data warehouse and data lake.

  • Work closely with teams to identify high-value data and integrate it into the data platform to empower each team to make data-driven decisions.

  • Help develop and maintain applications built using solutions such as Redshift, Apache Airflow, S3, Presto, Python, and Hive.

  • Ensure data quality and establish best practices across data infrastructure.

  • Ensure the data is well documented, high quality, and security is properly administered.

  • Deliver clean, scalable, maintainable code and monitoring.

  • Ensure SLAs are met for data pipelines.

  • Help manage our reporting platform by collaborating on data models, data structure, and visualizations.

  • Help the company make better use of data through training on dashboards, data insights, and analytical thinking generally.

  • Extract data from a variety of data sources for ad-hoc analysis and integration into data warehouse.

  • Troubleshoot data quality issues and work to resolve them with the data team and stakeholders.

  • Able to communicate findings in data through clear, consumable, audience-centric way regardless of whether the medium is written, oral, or visual.

BONUS

  • Experience with configuration and maintenance of distributed computing systems such as Hadoop.
  • Python
  • Amazon Web Services (AWS), especiallly Redshift, EMR, Kinesis, and Lambda
  • Microsoft SQL Server
  • Linux shell

Additional Information

All your information will be kept confidential according to EEO guidelines.

Videos To Watch