Search jobs > San Jose, CA > Software engineer ai

AI Compiler Software Development Engineer

Advanced Micro Devices, Inc
San Jose, California, United States
Full-time

WHAT YOU DO AT AMD CHANGES EVERYTHING We care deeply about transforming lives with AMD technology to enrich our industry, our communities, and the world.

Our mission is to build great products that accelerate next-generation computing experiences the building blocks for the data center, artificial intelligence, PCs, gaming and embedded.

Underpinning our mission is the AMD culture. We push the limits of innovation to solve the world’s most important challenges.

We strive for execution excellence while being direct, humble, collaborative, and inclusive of diverse perspectives. AMD together we advance THE ROLE : Come join our team working on the open-source SHARK, IREE, Turbine, and torch-mlir projects.

You will be responsible for building distributed training / inference solutions and service layers on top of existing core ML compiler and runtime.

There will be no short of intriguing technical challenges to tackle, and there are abundant chances to collaborate with industry experts working at different layers of the stack.

If this sounds interesting to you, please don’t hesitate to reach out to us! THE PERSON : An ideal candidate should be familiar with ML model parallelism techniques, multi-GPU inference, sharding, collectives, and integrating ML compiler / runtime / libraries into ML model services.

He / she should be willing to learn and work across boundaries, and comfortable with fast paced iterations to bring in the newest features in ML serving.

KEY RESPONSIBILITIES : Develop and maintain ML serving solutions on top of SHARK / IREE compiler and runtime Enable various ML model parallelism techniques on top of SHARK / IREE Analyze whole system performance, identify bottlenecks, propose ideas to improve, prototype and productionize solutions Follow industry directions and adopt emerging technologies in model serving PREFERRED EXPERIENCE : Familiarity with various model parallelism techniques Experience with existing ML model serving frameworks like vLLM, TensorRT, etc.

Experience with multi-GPU inference Experience with collectives and communication mechanisms Familiarity with SHARK, IREE, MLIR, PyTorch, etc.

Open-source development ethos ACADEMIC CREDENTIALS : Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent #LI-EM1 At AMD, your base pay is one part of your total rewards package.

Your base pay will depend on where your skills, qualifications, experience, and location fit into the hiring range for the position.

You may be eligible for incentives based upon your role such as either an annual bonus or sales incentive. Many AMD employees have the opportunity to own shares of AMD stock, as well as a discount when purchasing AMD stock if voluntarily participating in AMD’s Employee Stock Purchase Plan.

You’ll also be eligible for competitive benefits described in more detail here. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services.

AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and / or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law.

We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.

THE ROLE : Come join our team working on the open-source SHARK, IREE, Turbine, and torch-mlir projects. You will be responsible for building distributed training / inference solutions and service layers on top of existing core ML compiler and runtime.

There will be no short of intriguing technical challenges to tackle, and there are abundant chances to collaborate with industry experts working at different layers of the stack.

If this sounds interesting to you, please don’t hesitate to reach out to us! THE PERSON : An ideal candidate should be familiar with ML model parallelism techniques, multi-GPU inference, sharding, collectives, and integrating ML compiler / runtime / libraries into ML model services.

He / she should be willing to learn and work across boundaries, and comfortable with fast paced iterations to bring in the newest features in ML serving.

KEY RESPONSIBILITIES : Develop and maintain ML serving solutions on top of SHARK / IREE compiler and runtime Enable various ML model parallelism techniques on top of SHARK / IREE Analyze whole system performance, identify bottlenecks, propose ideas to improve, prototype and productionize solutions Follow industry directions and adopt emerging technologies in model serving PREFERRED EXPERIENCE : Familiarity with various model parallelism techniques Experience with existing ML model serving frameworks like vLLM, TensorRT, etc.

Experience with multi-GPU inference Experience with collectives and communication mechanisms Familiarity with SHARK, IREE, MLIR, PyTorch, etc.

Open-source development ethos ACADEMIC CREDENTIALS : Bachelor’s or Master's degree in Computer Science, Computer Engineering, Electrical Engineering, or equivalent #LI-EM1At AMD, your base pay is one part of your total rewards package.

Your base pay will depend on where your skills, qualifications, experience, and location fit into the hiring range for the position.

You may be eligible for incentives based upon your role such as either an annual bonus or sales incentive. Many AMD employees have the opportunity to own shares of AMD stock, as well as a discount when purchasing AMD stock if voluntarily participating in AMD’s Employee Stock Purchase Plan.

You’ll also be eligible for competitive benefits described in more detail here. AMD does not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services.

AMD and its subsidiaries are equal opportunity, inclusive employers and will consider all applicants without regard to age, ancestry, color, marital status, medical condition, mental or physical disability, national origin, race, religion, political and / or third-party affiliation, sex, pregnancy, sexual orientation, gender identity, military or veteran status, or any other characteristic protected by law.

We encourage applications from all qualified candidates and will accommodate applicants’ needs under the respective laws throughout all stages of the recruitment and selection process.

30+ days ago
Related jobs
Promoted
TikTok
San Jose, California

Stay up-to-date with the latest iOS mobile development, AIGC and Generative AI technologies, and incorporate them into the product development process, ensuring our creative tools stay at the forefront of the industry. Collaborate in the development and maintenance of core components and frameworks ...

Promoted
Penn Foster
Santa Clara, California

Background with modular software development and software containerization. The ideal candidate will have a thorough experience of developing innovative solutions in sensor processing and AI, delivering product-quality software, with a provable track record of open-source software contributions. Sup...

Promoted
The Rundown AI, Inc.
Palo Alto, California

We are looking for people with strong Backend Data Engineering capabilities to build highly efficient, resilient systems & pipelines for large-scale data processing. Work with researchers to identify and implement technical data requirements, and optimize distributed loading for model training. ...

Promoted
Otter.ai
Mountain View, California

We are searching for a Senior Front End Engineer who will be responsible for driving feature development and communicating with key stakeholders across the organization. Write clean, maintainable, and reusable code that adheres to code standards. We are growing fast, last year we recorded a total of...

Promoted
MatX
Mountain View, California

We are developing vertically integrated full-stack solutions from silicon to systems including hardware and software to train and run the largest ML workloads for AGI. Design and implement compilers tailored for our proprietary hardware architecture. Possess outstanding software engineering skills w...

Promoted
jobmonkeyjobs.com - JobBoard
Sunnyvale, California

Knowledge of professional software engineering & best practices for full software development life cycle, including coding standards, software architectures, code reviews, source control management, continuous deployments, testing, and operational excellence. As a Software Development Engineer o...

TikTok
San Jose, California

Minimum Qualification:- Bachelor’s degree in computer science or a related STEM field- Specialized experience in one or more of the following machine learning/deep learning domains: hardware accelerator architectures, machine learning compilers or ML systems, AI infrastructure, high-performance comp...

Amazon Development Center U.S., Inc.
Palo Alto, California

These are core systems development positions where you will own the design and development of significant software components critical to our industry leading database services architected for the cloud. Design, implement, test, deploy and maintain innovative software solutions to transform service ...

Apple
Sunnyvale, California

Would you like to work in a fast-paced environment where your technical abilities will be challenged on a day to day basis? If so, Apple’s AI & Data Platform (AiDP) team is seeking a Software Engineer to work on building and scaling best in class data and reporting apps presenting metrics & performa...

Palo Alto Networks
Santa Clara, California

The AI security cloud service engineering team is the core engineering team to build a solid product to assure the runtime security of our customers when they are using AI especially LLM services. Collaborate with product managers, cybersecurity researchers, AI application researchers and infrastruc...