Principal Data Platform Engineer, Data Management Engineering

  • Full-time
  • Business Function: Data Management Engineering
  • Work Location: Penang SDSM Office--LOC_SNDK_Seberang Perai Office

Company Description

Sandisk understands how people and businesses consume data and we relentlessly innovate to deliver solutions that enable today’s needs and tomorrow’s next big ideas. With a rich history of groundbreaking innovations in Flash and advanced memory technologies, our solutions have become the beating heart of the digital world we’re living in and that we have the power to shape.

Sandisk meets people and businesses at the intersection of their aspirations and the moment, enabling them to keep moving and pushing possibility forward. We do this through the balance of our powerhouse manufacturing capabilities and our industry-leading portfolio of products that are recognized globally for innovation, performance and quality.

Sandisk has two facilities recognized by the World Economic Forum as part of the Global Lighthouse Network for advanced 4IR innovations. These facilities were also recognized as Sustainability Lighthouses for breakthroughs in efficient operations. With our global reach, we ensure the global supply chain has access to the Flash memory it needs to keep our world moving forward.

Job Description

Principal Data Platform Engineer

(Data Engineering | Platform | Streaming | Cloud)

ESSENTIAL DUTIES AND RESPONSIBILITIES:

We are seeking a Principal Data Platform Engineer to define, evolve, and scale our enterprise data platform. This role is data-first and architecture-driven, with hands-on impact across ETL/ELT, big data, streaming, and cloud data platforms. You will act as a technical authority, shaping long-term platform direction, setting engineering standards, and mentoring senior engineers across teams.

Deep expertise in core data engineering is required. Candidates are not expected to be experts in every listed technology.

What You’ll Do

  • Own and define data platform architecture, standards, and long-term technical roadmap.
  • Design and oversee scalable ETL/ELT pipelines using Python across multiple data domains.
  • Establish data ingestion and data access APIs using Python and FastAPI for platform consistency and reuse.
  • Lead design and optimization of batch and streaming pipelines using Spark and Apache Kafka.
  • Drive architecture decisions for relational databases (MySQL, Oracle), cloud data warehouses (AWS Redshift), and NoSQL systems (Elasticsearch).
  • Guide large-scale data processing using Hive, Trino, and Hadoop distributed computing.
  • Define standards for object and file storage integrations (AWS S3, Dell ECS, SFTP).
  • Enable data quality, lineage, governance, and reliability at platform scale.
  • Support analytics and BI enablement (Power BI, Spotfire) through well-modeled datasets.
  • Contribute to lightweight internal UIs using React for data observability, configuration, or platform tooling (custom, not product UI).
  • Mentor senior and junior engineers; influence architecture across teams.

Why Join Us:

  • Principal-level ownership with hands-on technical impact
  • Opportunity to define enterprise-scale data architecture
  • High trust, high autonomy engineering culture
  • Focus on long-term platform quality, not short-term delivery

Qualifications

Required

  • 10+ years of experience in data engineering, data platform, or large-scale distributed data systems.
  • Demonstrated ownership of data platform architecture in complex environments.
  • Deep expertise in Python, SQL, and ETL/ELT design.
  • Strong hands-on experience with distributed data systems, including:
    • Big data: Spark, Hive, Trino, Hadoop
    • Streaming: Apache Kafka
  • Experience designing platforms using:
    • MySQL, Oracle, AWS Redshift
    • Elasticsearch or other NoSQL systems
  • Experience building data-focused APIs using FastAPI.
  • Familiarity with AWS cloud, DockerKubernetes and CI/CD at production scale.
  • Proven ability to influence without authority, mentor senior engineers, and drive cross-team alignment.

Preferred

  • Lakehouse technologies (Delta Lake, Iceberg)
  • Workflow orchestration (Airflow, Dagster, Prefect)
  • MLOps platforms (MLflow, Kubeflow, SageMaker)
  • Real-time processing frameworks (Flink, Kafka Streams)
  • Knowledge of data governance and compliance (GDPR, CCPA)
  • Experience building internal tooling UIs using React

Additional Information

Why Join Us?

You’ll shape the future of data and AI strategy within an organization committed to innovation and sustainability. This is an opportunity to lead impactful, enterprise-scale data initiatives and influence how the company leverages data-driven intelligence to drive growth and efficiency.

Sandisk thrives on the power and potential of diversity. As a global company, we believe the most effective way to embrace the diversity of our customers and communities is to mirror it from within. We believe the fusion of various perspectives results in the best outcomes for our employees, our company, our customers, and the world around us. We are committed to an inclusive environment where every individual can thrive through a sense of belonging, respect and contribution.

Sandisk is committed to offering opportunities to applicants with disabilities and ensuring all candidates can successfully navigate our careers website and our hiring process. Please contact us at [email protected] to advise us of your accommodation request. In your email, please include a description of the specific accommodation you are requesting as well as the job title and requisition number of the position for which you are applying.

NOTICE TO CANDIDATES: Sandisk has received reports of scams where a payment is requested on Sandisk’s behalf as a condition for receiving an offer of employment. Please be aware that Sandisk and its subsidiaries will never request payment as a condition for applying for a position or receiving an offer of employment. Should you encounter any such requests, please report it immediately to Sandisk Ethics Helpline or email [email protected].

Privacy Policy