Key job responsibilities
You will develop, improve and release our cutting-edge inference engines.
You will leverage advanced hardware, innovative software architecture, and distributed computing techniques to enable breakthrough research and product development across the company.
You will innovate in state-of-the-art inference and establish Amazon as the market leader in enterprise AI solutions, offering unbeatable price-performance.
You will lead our efforts to build the best inference performance on custom AWS Trainium and Inferentia silicon and the Trn1, Inf1/2 servers. Strong software development (Python and C++) and Machine Learning knowledge (Text and Multimodal) are both critical to this role.
You will understand current and future directions of ML framework development, with a focus on enabling the fastest and most price-performant inference.
You will collaborate with AWS Neuron, AWS Bedrock, and other teams within and outside Amazon to achieve the best outcome for AGI customers.
About the team
We are a Science and Engineering team working on the cutting edge of inference. We are interested in tackling the hardest and most impactful problems in AI inference. We explore inference-aware architectures, and compiler, kernel and runtime improvements to serve AI models of increasing size and performance.
BASIC QUALIFICATIONS
- 5+ years of non-internship professional software development experience
- 5+ years of programming with at least one software programming language experience
- 5+ years of leading design or architecture (design patterns, reliability and scaling) of new and existing systems experience
- 5+ years of full software development life cycle, including coding standards, code reviews, source control management, build processes, testing, and operations experience
- Experience as a mentor, tech lead or leading an engineering team
- Fundamentals of machine learning, prevalent model architectures including Transformers, familiarity of training and inference lifecycles, work experience with inference frameworks and inference optimizations.
PREFERRED QUALIFICATIONS
- Bachelor's degree in computer science or equivalent
Amazon is committed to a diverse and inclusive workplace. Amazon is an equal opportunity employer and does not discriminate on the basis of race, national origin, gender, gender identity, sexual orientation, protected veteran status, disability, age, or other legally protected status. For individuals with disabilities who would like to request an accommodation, please visit https://www.amazon.jobs/en/disability/us.
Our compensation reflects the cost of labor across several US geographic markets. The base pay for this position ranges from $151,300/year in our lowest geographic market up to $261,500/year in our highest geographic market. Pay is based on a number of factors including market location and may vary depending on job-related knowledge, skills, and experience. Amazon is a total compensation company. Dependent on the position offered, equity, sign-on payments, and other forms of compensation may be provided as part of a total compensation package, in addition to a full range of medical, financial, and/or other benefits. For more information, please visit https://www.aboutamazon.com/workplace/employee-benefits. This position will remain posted until filled. Applicants should apply via our internal or external career site.