AI/ML – ML/HW/SW Co-Design Engineer, DMLI – 200555473 -Seattle, Washington, United States

Apple

The Data and Machine Learning Innovation (DMLI) team employs ML/HW/SW co-design to achieve best-in-class performance and energy efficiency for numerous use cases that deploy neural networks. We seek a generalist to help define and implement features that would be realized in next-gen neural accelerators. Our team is comprised of HW, SW, and ML engineers working together in the area of Efficient ML. Our charter is to push the frontiers of perf and power for AI inference.

As a co-design engineer, you will be accountable for bridging the gap between ML and devices. This including working with ML, computer architecture and SW teams to understand the trade-offs from various viewpoints when it comes to ML inference, keeping abreast of advancements in the ML efficiency techniques space particularly for Generative AI, and come up with novel solution that addresses the ever growing needs for the future ML workloads.

On a day-to-day basis, your responsibilities include but not be limited to: architecting energy-efficient hardware designs, implementing changes in an in-house simulator, implementing new features in an in-house compiler and working with ML engineers to optimize their networks for HW acceleration. Our ideal team member is courageous when it comes to trying new things, is adept at reasoning about systems performance, and is willing to iterate on ideas. We value team members with strong communication skills with experience working cross-functionally with HW, SW, and ML teams.

As a member of this team, you will use your background to:

– Work with ML experts to co-design hardware and software solutions that further improve efficiency of neural workloads at inference time

– Implement features in our simulation engine and compiler for next-gen accelerators

– Benchmark and diagnose performance bottlenecks of deep learning models

– Work with a variety of partners from all parts of the stack — from Apps to Compilation, HW Arch, and Silicon Validation

Bachelor’s, Masters or PhD degree in Electrical Engineering, Computer Science, or a related field5+ years of experience working on ML inference performance optimization either in SW or HW spaceProficient in Python, working knowledge of C++Proficiency in algorithms and coding skills to optimize a workload and solve hard problems in scheduling, allocation and tune solutions to the details of an architecture platform

 

Job Overview