Data Platform Engineer

Santa Clara Valley (Cupertino), California, United States
Software and Services

Summary

Posted: Jan 15, 2019
Weekly Hours: 40
Role Number: 200019706
At Apple, we work every day to create products that enrich people’s lives. Our Ad Platforms group makes it possible for people around the world to easily access informative and imaginative content on their devices while helping publishers and developers promote and monetize their work. Our technology and services power Search Ads in the App Store and the advertising found in Apple News. Our platforms are highly performant, deployed at scale, and set new standards for enabling effective advertising while protecting user privacy. The Ad Platforms Data Insights Engineering team is seeking a data platform engineer to join in developing the next generation of analytical solutions built to empower Sales, Product, and Executive teams. In this role you will be a key member of the team driving the strategy, development, execution, and continuous improvement of core analytical data products for Ad Platforms. You will build the data infrastructure, tools, and processes that enable our teams of data scientists, analysts, and engineers to rapidly and reliably build reproducible and effective pipelines, reports, analyses, and models. A successful candidate will have experience architecting and building data platforms using varied data storage and analysis toolsets.

Key Qualifications

  • Background in computer science, mathematics, or similar quantitative field with a minimum of 3-5 years professional experience
  • Data Platform design / build utilizing Hadoop, Spark, or similar technologies
  • Programming experience (Java, Scala preferred)
  • Advanced skills using one or more scripting languages (e.g., Python, bash, etc.)
  • SQL, data modeling, and relational database experience and knowledge of data warehousing concepts, methodologies, and frameworks
  • Extract Transform Load (ETL) experience using Spark, Kafka, Hadoop, or similar technologies
  • Presto, Hive, SparkSQL, Cassandra, Solr, or other big data query and transformation experience
  • UNIX admin and general server administration experience required
  • Solid understanding of software engineering concepts like version control, automated testing, and CI/CD required
  • Experience with workflow scheduling / orchestration such as Airflow or Oozie
  • Experience applying data encryption and data security standards
  • Hands-on experience developing or supporting machine learning and data science workloads a plus
  • Experience with containerization technologies like Docker and Kubernetes in a public / private cloud environment plus
  • Experience with automation (either custom or using technologies such as Puppet / Chef) a plus
  • Data visualization or web development skills a plus
  • Ability to communicate technical concepts to a business-focused audience
  • Ability to work with teams across functions such as engineering, operations, and analytics
  • Most importantly, a sense of humor and an eagerness to learn

Description

Design and build a next generation data platform to enhance existing analytics, data science, and reporting capabilities. Solve tough problems across the technology spectrum including architecting, designing, creating, and extending data storage, processing, and analytic solutions. Partner with business and analytics teams to understand specific requirements of the data platform. Automate and optimize existing analytic workloads by recognizing patterns of data and technology usage Implement solutions to automate and monitor the testing and deployment of large scale data products and analytics. Build tools to enable data scientist, analysts, and engineers to develop performant, robust, and reproducible analyses, models, and data products. Must be able to work in a rapidly changing environment and perform effectively in a sprint-based agile development environment.

Education & Experience

BSCS

Additional Requirements