Big Data Engineer - Manufacturing Systems & Infrastructure

Bengaluru, Karnataka, India
Operations and Supply Chain

Summary

Posted:
Role Number:200440631
Imagine what you could do here. At Apple, new ideas have a way of becoming extraordinary products, services, and customer experiences very quickly. Bring passion and dedication to your job and there's no telling what you could accomplish. The people here at Apple don’t just create products — they create the kind of wonder that’s revolutionized entire industries. It’s the diversity of those people and their ideas that inspires the innovation that runs through everything we do, from amazing technology to industry-leading environmental efforts. Join Apple, and help us leave the world better than we found it. Apple's Manufacturing Systems & Infrastructure (MSI) team is responsible for gathering, consolidating and tracking all manufacturing data for Apple’s products, modules and manufacturing equipment worldwide. This data is used throughout the company and the product's lifecycle, from the very beginning, to validate that units being built are fully tested and of high quality before leaving the factory, all of the way through to warranty support for customers. Apple's MSI team is seeking a Big Data Engineer who possesses a real passion for developing extraordinary products with a deep appreciation for user experience. We develop web applications, batch and streaming analytics solutions using Spring, Kafka, Hadoop, Spark, Flink, Trino and other state of the art technologies in a large scale infrastructure. We are looking for a passionate and results-oriented Sr. Software Engineer/Sr. Big Data Engineer to join our team and work on some of the highly visible data projects in Operations organization!

Key Qualifications

  • Deep experience developing large scale distributed computing systems.
  • In-depth knowledge and experience in one or more of the following technologies: Big Data ecosystem, Kafka, Distributed SQL Database (Snowflake, SingleStore), HBase, Cassandra, Iceberg, Spark, Flink and Trino
  • In-depth knowledge and experience in one or more of the following technologies: Big Data ecosystem, Kafka, Distributed SQL Database (Snowflake, SingleStore), HBase, Cassandra, Iceberg, Spark, Flink and Trino
  • Extensive experience working with one or more SQL Database systems like PostgreSQL - define schema & DDL, maintain Stored Procedures, perform query optimizations, tracking DB metrics, and more
  • Strong software development, problem-solving and debugging skills with experience in one or more of the following languages: Java, Python, Scala
  • Strong software development, problem-solving and debugging skills with experience in one or more of the following languages: Java, Python, Scala
  • Experience building cloud native platforms using containerization technologies like Kubernetes, docker, helm and well versed in AWS/GCP or Azure
  • Ideally 5+ years of experience in crafting, implementing and supporting highly scalable data pipelines systems with high volume of data.
  • Adept at quickly grasping and distilling highly complex matters into clean, understandable solutions
  • Nice to have experience with data governance, security & Privacy tools like Apache Ranger, Apache Atlas
  • Nice to have experience with large data reporting and data visualization tools
  • Organized, detail oriented, and thorough in every undertaking
  • Ability to thrive in a multi-functional team on high profile, critical projects
  • Candidate should be able to initiate and explore alternate technology and approaches to solving problems.

Description

In this role we look forward to your designing and developing software architecture for large scale data pipelines and core data storage solutions to enable analytics and ML use cases and management for the Machine Metrics system at Apple Product Operations - You will be responsible in crafting, developing and maintaining our large-scale data pipelines, storage, and processing services - Build self-service analytics and ML tools to help engineering teams derive actionable metrics out of large volumes of raw data - Partner with different Ops teams, data science and engineering teams and develop platforms/apps/algorithms to answer complex questions on quality of our products. - You will work closely with the DevOps team and develop monitoring and alerting scripts on various web, data pipelines and jobs - You will have the opportunity to learn and work on the latest Big Data technologies, lead POCs to demonstrate new ideas and influence the future direction of our technology stack - Review hardware infrastructure planning and provide support when required for software upgrades.

Education & Experience

Bachelor’s / Masters degree in Computer Science Engineering

Additional Requirements