Solution Architect_Local candidates only(W2 role)
- Contract
Job Description
Role: Solution Architect
Location: Phoenix, AZ – (Remote) but Locals only for in-person interview
Type: Contract
Job Description
- Data is responsible for contributing to the design, modernization, and optimization of enterprise-scale data systems, as well as the maintenance and operations strategy for client.
- This role involves designing and implementing data systems that organize, store, and manage data within our cloud data platform.
- The architect will perform continuous maintenance, and operations work for client in the cloud environment.
- They will review and analyze client’s data infrastructure, plan future database solutions, and implement systems to support data management for client users.
- Additionally, this role is accountable for ensuring data integrity, making sure the client team adheres to data governance standards to maintain accuracy, consistency, and reliability across all systems.
- The architect will identify data discrepancies and quality issues, and work to resolve them. This position requires a strong blend of architectural leadership, technical depth, and the ability to collaborate with business stakeholders, data engineers, machine learning practitioners, and domain experts to deliver scalable, secure, and reliable AI-driven solutions.
- The ideal candidate will have a proven track record of delivering end-to-end ETL/ELT pipelines across Databricks, Azure, and AWS environments.
Key Responsibilities
- Design scalable data lake and data architectures using Databricks and cloud-native services.
- Develop metadata-driven, parameterized ingestion frameworks and multi-layer data architectures.
- Optimize data workloads and performance.
- Define data governance frameworks for client.
- Design and develop robust data pipelines.
- Architect AI systems, including RAG workflows and prompt engineering.
- Lead cloud migration initiatives from legacy systems to modern data platforms.
- Provide architectural guidance, best practices, and technical leadership across teams.
- Build documentation, reusable modules, and standardized patterns. Required Skills and Experience
- Strong expertise in cloud platforms, primarily Azure or AWS.
- Hands-on experience with Databricks.
- Deep proficiency in Python and SQL.
- Expertise in building ETL/ELT pipelines and ADF workflows.
- Experience architecting data lakes and implementing data governance frameworks. Hands-on experience with CI/CD, DevOps, and Git-based development.
- Ability to translate business requirements into technical architecture. Technical
Additional Information
All your information will be kept confidential according to EEO guidelines.