Apple
Sr. Machine Learning Engineer, ASR Infrastructure and Tools
Apple, Cambridge, Massachusetts, us, 02140
Sr. Machine Learning Engineer, ASR Infrastructure and Tools
Cambridge, Massachusetts, United States Want to join the team pushing the boundaries of AI and building an intelligent assistant that helps millions of people get things done? Join the Siri team at Apple. To build the best speech recognition models, we need to use the latest technology in distributed training and the best available data. We combine those needs into one team and are focused on blurring the lines between traditional "data processing" and "model training". Efficiently training on petabytes of audio data pushes us to consider the entire training stack while developing new models to extract useful signals from unprecedented volumes of data. By joining our team, you'll have the opportunity to work with large and diverse datasets, iterate with research and production teams, and deliver voice based experiences to millions of users worldwide. Description
The Siri Speech team is looking for exceptional individuals to extend the core technology that let Siri understand, learn, and remember. You will be part of a cross-functional team consisting of software engineers as well as data and machine learning engineers/scientists and having a large impact on the Siri product. This is a rare opportunity to apply distributed data engineering techniques at the intersection of various areas such as speech recognition, natural language processing, and dialogue management. In this role you will: Work with open source tools like PySpark, Jax, Ray and others Optimize how to move multi-modal data from various sources into complex model training pipelines Use open source models to extract signals from large volumes of speech data to drive modeling improvements Minimum Qualifications
Experience processing large, complex, unstructured data Knowledge of distributed data processing frameworks (Beam, Spark, Dask, Ray) Strong software engineering skills Preferred Qualifications
M.S. or Ph.D. degree in Computer Science, or equivalent experience Machine Learning experience a plus Speech understanding or generation experience a plus Strong data engineering background in speech and/or language/text/dialogue processing field Strong software engineering abilities, ideally Python Strong interpersonal skills to work well with engineering teams Excellent problem solving and critical thinking Ability to work in a fast-paced environment with rapidly changing priorities Passionate about building extraordinary products and experiences for our users
Cambridge, Massachusetts, United States Want to join the team pushing the boundaries of AI and building an intelligent assistant that helps millions of people get things done? Join the Siri team at Apple. To build the best speech recognition models, we need to use the latest technology in distributed training and the best available data. We combine those needs into one team and are focused on blurring the lines between traditional "data processing" and "model training". Efficiently training on petabytes of audio data pushes us to consider the entire training stack while developing new models to extract useful signals from unprecedented volumes of data. By joining our team, you'll have the opportunity to work with large and diverse datasets, iterate with research and production teams, and deliver voice based experiences to millions of users worldwide. Description
The Siri Speech team is looking for exceptional individuals to extend the core technology that let Siri understand, learn, and remember. You will be part of a cross-functional team consisting of software engineers as well as data and machine learning engineers/scientists and having a large impact on the Siri product. This is a rare opportunity to apply distributed data engineering techniques at the intersection of various areas such as speech recognition, natural language processing, and dialogue management. In this role you will: Work with open source tools like PySpark, Jax, Ray and others Optimize how to move multi-modal data from various sources into complex model training pipelines Use open source models to extract signals from large volumes of speech data to drive modeling improvements Minimum Qualifications
Experience processing large, complex, unstructured data Knowledge of distributed data processing frameworks (Beam, Spark, Dask, Ray) Strong software engineering skills Preferred Qualifications
M.S. or Ph.D. degree in Computer Science, or equivalent experience Machine Learning experience a plus Speech understanding or generation experience a plus Strong data engineering background in speech and/or language/text/dialogue processing field Strong software engineering abilities, ideally Python Strong interpersonal skills to work well with engineering teams Excellent problem solving and critical thinking Ability to work in a fast-paced environment with rapidly changing priorities Passionate about building extraordinary products and experiences for our users