Job Details

Job Information

AI Data Engineer
AWM-182-AI Data Engineer
5/8/2026
5/13/2026
Negotiable
Permanent

Other Information

www.apple.com
Cupertino, CA, 95015, USA
Cupertino
California
United States
95015

Job Description

No Video Available
 

Weekly Hours: 40

Role Number: 200660980-0836

Summary

Imagine what you could do here. At Apple, new ideas have a way of becoming great products, services, and customer experiences very quickly. Bring passion and dedication to your job and there's no telling what you could accomplish.

Are you passionate about building the data pipelines that make AI systems fast, accurate, and reliable?
Do you thrive on engineering clean data flows that connect enterprise systems to intelligent applications?
Can you build infrastructure that's both production-grade and purpose-built for AI consumption?

The Applied Data Science team within Legal Operations is building production-grade AI for a global legal organization — and every AI system is only as good as the data flowing into it. The AI Data Engineer owns the pipelines, data feeds, and integration infrastructure that ensure AI applications have the right data, in the right form, at the right time.

Description

The AI Data Engineer builds and maintains the data infrastructure that powers AI applications across Legal Operations. You will design and implement data pipelines that ingest from legal systems, transform data into AI-ready formats, load vector databases and other AI stores, and expose data services through APIs. This role is embedded within the AI team and works in close partnership with AI and data colleagues to ensure AI systems have reliable, high-quality data at every stage.

• Design and implement data pipelines that ingest, transform, and deliver data from legal systems (matter management, eBilling, CLM, document management) to AI applications
• Build and maintain pipelines that load and refresh vector databases, document stores, and graph databases used by AI retrieval systems
• Engineer data transformations that prepare legal data for AI consumption — chunking, embedding generation, metadata enrichment, and schema normalization
• Build upstream and downstream integrations with MCP (Model Context Protocol), vector databases, and knowledge graphs to support context engineering and AI retrieval systems
• Develop and maintain APIs that expose structured and unstructured data to AI applications and analytics tools
• Implement data quality checks and validation at pipeline ingestion points to ensure AI systems receive reliable, complete data
• Build monitoring and alerting for pipeline health, data freshness, and load failures
• Understand AI data access patterns and optimize data delivery for AI performance
• Integrate with the semantic layer — consuming entity resolution outputs, taxonomy mappings, and enriched datasets to ground AI applications
• Implement ETL/ELT processes using dbt, Fivetran, or similar tools with a focus on reliability and maintainability
• Document pipeline designs, data contracts, and operational runbooks

Minimum Qualifications

  • Bachelor's degree in Computer Science, Data Science, Information Systems, or related field (or equivalent experience); Master's degree preferred

  • 4+ years of experience in data engineering related to AI application

  • Strong proficiency in SQL and Python for data engineering and transformation

  • Experience with cloud data platforms (Snowflake, Databricks, BigQuery, or similar)

  • Experience with ETL/ELT tools (dbt, Fivetran, Airflow, or similar)

  • Experience building and maintaining REST APIs

  • Understanding of data modeling and data transformation best practices

  • Experience with version control (Git) and CI/CD practices

  • Ability to work closely with AI/ML teams and understand their data requirements

Preferred Qualifications

  • Experience with vector databases (Pinecone, Weaviate, Chroma), embedding generation pipelines, document stores (MongoDB or similar) and their integration patterns

  • Understanding of RAG, MCP architectures, context engineering principles, and how data quality affects retrieval performance

  • Experience with semantic layer technologies (dbt Semantic Layer, Cube, AtScale), knowledge graphs (Neo4j), or ontology design

  • Experience with streaming or event-driven data architectures (Kafka or similar)

  • Familiarity with legal operations data (matter management, eBilling, CLM, document management)

Other Details

No Video Available
--

About Organization

 
About Organization