If you are interested in having your job openings posted, details on becoming a sponsor may be found here.

DIAMOND SPONSOR JOB LISTINGS

Company Description: WHAT YOU DO AT AMD CHANGES EVERYTHING. At AMD, our mission is to build great products that accelerate next-generation computing experiences—from AI and data centers, to PCs, gaming and embedded systems. Grounded in a culture of innovation and collaboration, we believe real progress comes from bold ideas, human ingenuity and a shared passion to create something extraordinary. When you join AMD, you’ll discover the real differentiator is our culture. We push the limits of innovation to solve the world’s most important challenges—striving for execution excellence, while being direct, humble, collaborative, and inclusive of diverse perspectives. Join us as we shape the future of AI and beyond.  Together, we advance your career.

Company Contact: Murray Hennessy - Murray.Hennessy@amd.com

Job Title: Software Development Engineer - Compilers

Job Description: The candidate will work on the next generation AMD LLVM Fortran compiler implementing language features to accelerate real-world scientific and engineering applications using OpenMP on AMD GPUs in exascale supercomputers.

The Person:
The successful candidate will work on language implementation and optimization in the new opensource LLVM Flang compiler. Prior experience with MLIR is ideal. Prior experience with applications written in Fortran and familiarity with the OpenMP programming model is highly desired.

Key Responsibilities:
- The candidate will implement newly defined language features and target offload support for both recent and upcoming Fortran and OpenMP standards.
- During the first year, the candidate will be fully up to speed and contributing to the next generation LLVM Flang compiler project.
 
Preferred Experience:
- Strong background in compilers
- Strong C/C++ programming skills
- LLVM/Clang/Flang/MLIR experience
- Prior experience with Fortran desired; modern Fortran dialects are a plus.
 
Academic Credentials:
BS in Computer Science, Electrical Engineering, or Math, a MS or PhD is strongly preferred.

Job Title: Senior PyTorch Engineer

Job Description: AMD is looking for a specialized software engineer who is passionate about improving the performance of key applications and benchmarks. You will be a member of a core team of incredibly talented industry specialists and will work with the very latest hardware and software technology. 

The Person:
Our team is seeking expert-level skilled and motivated senior developer with a strong technical foundation in C++/Python AI development in Linux environment. The ideal candidate combines deep technical expertise with strong analytical abilities, excelling both as a collaborative team member and as an independent contributor capable of defining goals, scoping work and leading development efforts end to end.

Key Responsibilities:
- Model Development and Optimization: Lead the development and optimization of Deep Learning models on AMD GPUs, ensuring high performance and scalability across training and inference workloads.
- Framework Enablement: Lead efforts to enhance PyTorch and its ecosystems on AMD GPUs, contributing performance-critical improvements to upstream open-source repositories.
- Open-Source Intergration: Engage directly with PyTorch and other open-source maintainers to align on requirements, propose enhancements and ensure AMD contributions are successfully adopted and maintained upstream.
- Community Feature Enablement: Collaborate with PyTorch community and ecosystem partners to design, enable and upstream new features, ensuring AMD GPUs remain first class for emerging AI workloads.
- Cross-Functional Collaboration: Partner closely with internal GPU Libraries, compilers, and runtime teams to identify performance bottlenecks and deliver optimized solutions for large-scale AI workloads.
- Scalable Systems Optimization: Developed innovative solutions for both scale-up (multi-GPU) and scale-out (multi-node) distributed training environments to optimize performance on AMD GPUs.
- Engineering Excellence: Apply rigorous software engineering best practices, including code review, testing, CI/CD integration and debugging, to deliver reliable and maintainable solutions at scale.

Preferred Experience:
- Software Engineering: Expert skills in Python and C++, with experience in debugging, performance tuning, and test design to - ensure high-quality, maintainable software solutions.
- AI Framework & Deep Learning: Strong understanding of AI frameworks such as PyTorch, Triton and vLLM, with applied - knowledge across domains such as Natural Language Processing, Vision, Audio and Recommendation Systems.
- GPU Computing: Strong experience with GPU Programming models (CUDA, HIP).
- Development Processes & Tools: Expert with modern software practices and tooling, including debuggers, profilers, and version control systems (e.g., GitHub).
- Collaboration & Problem Solving: Effective and efficient communicator with strong analytical and problem-solving skills, capable of working across diverse technical teams.

Academic Credentials:
- Master’s or PhD level degree in Computer Science, Computer Engineering, Electrical Engineering, or a related field.
- Senior Professional Experience in technical software development, with a focus on GPU optimization, performance engineering, and framework development.

Company Description: Apple’s investment in developer tools, security technologies, and performance optimizations powered by LLVM helps deliver products that impact billions of users worldwide. This work enhances applications used by everyday people and gives rise to new technologies. We actively participate in the LLVM open source project and are dedicated to supporting and mentoring our employees. We emphasize diversity, collaboration, and creativity!

Company Contact: Jonas Devliegherejdevlieghere@apple.com

Job Title: Debugger Engineer - London, UK

Job Description: The LLDB team is hiring an engineer to work on the debugger, which is used internally across Apple and externally by millions of developers. You'll drive features and improvements in LLDB, collaborating across teams working on the OS and toolchain. Responsibilities range from supporting new hardware and platforms to developing debugging features, with most work open-source (llvm.org & swift.org).

Job Title: Swift Compiler Backend Engineer - Cupertino, US

Job Description: You will be responsible for the quality and correctness of Swift’s generated code through ongoing development and management of the Swift backend. In this role, you will work closely with the Swift optimizer team to ensure that new optimizer concepts are correctly handled, with the Swift runtime team in their ongoing search for better ways to handle Swift’s runtime operations, with the LLVM team to steadily improve Swift’s usage of LLVM, and with platform maintainers both within and outside of Apple to ensure that Swift code runs well on a growing collection of target platforms.

Job Title: Compiler Optimization Engineer – London, UK

Job Description: The CPU Compiler Team is seeking senior engineers passionate about working on advancing compiler performance and optimization technology. We are responsible for optimizations and code generation for CPUs on all Apple platforms. Our team works across many different organizations on state-of-the art, industry impacting technology that enhances the user experience for all Apple customers with better run-time performance, battery life, compile-times, code size or enhanced security. A lot of the team's work happens directly in the LLVM open source project and gets presented at the LLVM Developer Meeting and the Apple’s WWDC conferences.

Job Title: Compiler Optimization Engineer – Israel

Job Description: The CPU Compiler Performance Team welcomes applications from engineers with different levels of experience, passionate about advancing compiler-based optimization technologies. Our team is responsible for optimizing the code generation for Apple CPUs. We work across many different organizations, such as Hardware Architecture and OS teams, to generate high-performance code, tuned to Apple’s unique microarchitecture, to provide the best run-time performance and battery life to our customers. Most of the team’s work happens directly in the LLVM open-source project and gets presented at the LLVM Developer Meeting. Our team embodies diversity, collaboration, and creativity.

Job Title: Compiler Security Engineer - C/C++ - Cupertino, US

Job Description: The Security Tools team at Apple is looking for software engineers to develop secure language features for C/C++ and enhance security features in the Clang compiler. Clang is a core part of Apple's developer tools, used internally to build Apple's software stack and externally by millions of developers, and is open source, powering the broader ecosystem. Your work will help eliminate entire classes of vulnerabilities at the language level, protecting billions of devices.

Job Title: Security Developer Tools Engineer (Static Analysis) - Seattle, US

Job Description: Apple’s static program analysis tools help millions of developers inside and outside of Apple find, fix, and prevent security and stability bugs. We are looking for a senior engineer with a strong background in compilers or static analysis tools to help develop tools to prevent classes of errors, support the adoption of safe programming models, and find bugs. This role will involve developing program analyses to prevent classes of errors and automate the adoption of safe programming models, collaborating with and mentoring engineers, and partnering with operating system and security teams to improve the security and reliability of software in the Apple ecosystem.  We have multiple talks on our work at EuroLLVM. If you are interested, please come talk to us.

Company Description: Our mission is to organize the world’s information and make it universally accessible and useful.

Company Contact: Shankar Subramanian - shankarks@google.com

Job Title: Software Engineer, TPU Compiler, PhD, Early Careers (London, UK)

Job Description: TPU team develops the Accelerated Linear Algebra (XLA) TPU parallelizing compiler used to partition, optimize, and run large-scale machine learning models across multiple TPU accelerators for internal (e.g. Google DeepMind) and external customers. It is a vital part of the Google Gemini software infrastructure.

Job Title: Software Engineer III, Tensor Processing Units, AI/ML (London, UK)

Job Description: The Core ML team contributes to frameworks and compilers that supports the service and related ML models/frameworks. It provides ML infrastructure to the customers (both external and internal to Google) with large-scale, cloud-based access to Google’s first-party ML supercomputers to run training and inference workloads.

In this role, you will work on AI framework development to successfully enable PyTorch models.

Job Title: Staff Software Engineer, ML Performance and Accelerators (London, UK)

Job Description: In this role, you will deliver credible machine learning (ML) and artificial intelligence (AI) performance and cost projections, you will influence the machine learning hardware and software (HW/SW) roadmap through co-design and key insights, and foster full stack innovation.
- Co-design hardware and system features, including compute, networking, and memory, for evolving workloads while identifying future hardware and software opportunities to improve ML performance.
- Build, maintain, and validate high-fidelity Hardware and Software (HW/SW) simulators in C++ and Python to enable rapid evaluation of solutions, alongside developing methods to measure, visualize, and analyze ML performance.

Job Title: Software Engineer, Compilers, Runtimes and Toolchains, Early Career (Mexico City, MX)

Job Description: As a Software Engineer on the Compilers, Runtimes and Toolchains (CRT) team, you will work very closely with other senior SWEs on the team in developing features in the space of LLVM optimization and code generation that support (but not limited to) build scalability, toolchain velocity, peak performance and safety. The SWE will develop changes (to tools and processes) that help us deliver state-of-the-art C++ toolchain efficiently and always correct and efficient + make the necessary changes upstream in LLVM toolchain that aligns with Google's needs and interests.

Job Title: Staff Software Engineer, Machine Learning Compiler, Google Research (Multiple Cities, USA)

Job Description: (USA locations: Austin, TX; Kirkland, WA; Mountain View, CA; Sunnyvale, CA)

The Machine Learning Compiler (MPACT) team in Google Research is developing open-source, retargetable compiler infrastructure in the Low Level Virtual Machines (LLVM) and MLIR frameworks to streamline iterative processor and system co-design for the Google Tensor Processing Unit (TPU) family of processors, and to enable C/C++/Cuda High Performance Computing (HPC) codes to run efficiently on TPUs.

Help build LLVM/(MLIR)-based C/C++/Cuda compiler and simulation tools for Google TPU family of ML accelerators, enabling conventional HPC development for TPUs.

Job Title: ML Compiler Software Engineer, TPU (Sunnyvale, CA, USA)

Job Description: Accelerated Linear Algebra (XLA) powers all ML workloads at Google. It’s also a choice of most external foundation model producers who value performance and reliability at large-scale. It is the most advanced ML compiler in the industry.

You will specialize in scaling capabilities of the compiler essential for supporting increasing model sizes. Your contributions as part of the team will be critical for achieving best performance and reliability for the most important and extremely large ML programs at Google and top external AI companies.

You will work with the world experts in ML hardware, compiler and performance optimization.

Our team operates across the layers of the compiler. You will have an opportunity to contribute across the stack from the higher level op rewrites to the low level emitters exercising specialized hardware features.

Job Title: Senior Silicon Engineer, GPU (Bengaluru, India)

Job Description: Be part of a team that pushes boundaries, developing custom silicon solutions that power the future of Google's direct-to-consumer products. You'll contribute to the innovation behind products loved by millions worldwide. Your expertise will shape the next generation of hardware experiences, delivering unparalleled performance, efficiency, and integration.

As a Senior Graphics Processing Unit (GPU) Silicon Engineer in the Google Silicon team, you will be at the forefront of defining how our custom-designed hardware interacts with the entire software stack. You will drive the architecture and implementation of kernel-mode and user-mode drivers that power everything from high-end mobile graphics. You aren't just writing code, you are bridging the gap between hardware intent and software execution.

Job Title: Software Engineer, Edge TPU Compilers, Silicon (Mountain View CA, USA; Kirkland WA, USA)

Job Description: As a part of this team, you will contribute to building Google Tensor TPU , Google’s custom System-on-Chip (SoC) ML accelerator that powers the latest Pixel phones. As a part of this team, you will help make transformative user experiences possible by leveraging cutting-edge Machine Learning (ML) running on Tensor TPU. You will be responsible for analysis, optimization, and compilation of ML models targeting the EdgeTPU. You will enable Gemini Nano, our efficient AI model for on-device tasks to run on Pixel phones and leverage the latest ML innovations and research by delivering computing hardware and software. 

PLATINUM SPONSOR JOB LISTINGS

Company Description:  Arm’s processors are shipped in billions of products, across a huge range of markets, each with unique code generation challenges. LLVM is a foundational code generator for all our processors: CPUs, GPUs and ML accelerators.

Company Contact: Kristof Beyls - Kristof.Beyls@arm.com

Job Title:  Many LLVM-related jobs at Arm

Job Description:  Are you a developer looking to work with technology that reaches 100% of the connected global population?

In the past year, about 100 Arm engineers contributed to LLVM, in areas such as performance optimization, security hardening, support for new instructions and many more. We contribute to most LLVM projects, including the mid- and back-ends, MLIR, Clang, Flang, LLDB, compiler-rt, LLD, Bolt, libc, …

Your skills and knowledge of compiler fundamentals, and your passion to learn from and contribute to the LLVM community will help us develop innovative technologies that improve the performance and security of the entire field of computing.

Arm always has a range of LLVM-related job vacancies open.

Company Description: Founded in 1987, Huawei is a leading global provider of information and communications technology (ICT) infrastructure and smart devices. We have 207,000 employees and operate in over 170 countries and regions, serving more than three billion people around the world.

Our vision and mission is to bring digital to every person, home and organization for a fully connected, intelligent world. To this end, we will drive ubiquitous connectivity and promote equal access to networks; bring cloud and artificial intelligence to all four corners of the earth to provide superior computing power where you need it, when you need it; build digital platforms to help all industries and organizations become more agile, efficient, and dynamic; redefine user experience with AI, making it more personalized for people in all aspects of their life, whether they’re at home, in the office, or on the go.

This spirit of innovation has led Huawei to work in close partnership with leading academic institutions in the UK to develop and refine the latest technologies. With a shared commitment to innovation and progress, both parties have worked together to achieve common goals and establish a strong partnership. The partnership between UK and Huawei help to develop the technologies of the future that will transform the way we all communicate, work and live.

Company Contact: Leo Huang - edinburghjobs@huawei.com

Job Title: AI Compiler Optimization Engineer (Cambridge, UK)

Job Description: We are seeking a skilled AI Compiler Optimization Engineer to optimize AI model inference performance through advanced compiler technologies. You will focus on performance tuning for CPU or hybrid CPU/XPU heterogeneous architectures, profiling AI frameworks to discover new optimization opportunities, and delivering cutting-edge insights from industry research.

Key Responsibilities:
Compiler-Based Performance Optimization:
• Implement compiler techniques (e.g., MLIR level optimizations, LLVM backend optimizations) to enhance inference performance on CPU and CPU/XPU hybrid systems.
• Optimize JIT level compute graphs with operator fusion, memory allocation and etc. for latency/throughput improvements.
AI Model Profiling & Framework Optimization
• Profile end-to-end inference workflows on frameworks like TensorFlow, PyTorch, ONNX, and llama.cpp to identify hotspots and bottlenecks.
• Propose and implement optimization strategies (e.g., kernel tuning, graph-level optimizations).
Research & Insight Development
• Track and analyze the latest advancements in AI & compiler research (academic papers, open-source projects).
• Produce actionable insight reports summarizing trends, benchmarks, and potential optimizations.

Requirements:
• Proficiency in C/C++ and compiler infrastructure (LLVM, MLIR, or similar)
• Deep understanding of AI model architectures and inference workflows
• Experience with performance profiling tools (e.g. perf, TensorBoard, VTune)
• Familiarity with CPU/XPU hardware architectures and optimization techniques

Preferred Qualifications:
• BSc/MSc/PhD in CS
• Contributions to open-source compiler projects (LLVM/MLIR communities)
• Experience with heterogeneous computing (CPU/GPU/XPU)

Job Title: Compiler Engineer - Internship (Cambridge, UK)

Job Description: You will join a team of compiler experts located in Cambridge and global locations to develop LLVM compiler and libraries. You will learn to resolve real-world server, HPC and ML compiler problems by analyzing performance issues on xPUs, applying state of art optimizations, debugging the compilation procedure and present the result. You will be encouraged to solve problem in innovative ways. You will work in a diversified team with different culture background but share common goals. Your contribution can be up-streamed to open-source communities.

Key Responsibilities:
Develop LLVM/MLIR based compilers:
• Optimize compiler for server, HPC and ML workloads
• Using compiler to automatically generate optimized operators
• Use AI technology to help compiler optimize
• Design & implement optimization for xPU architectures
• Analyze performance of compiler generated code
• Co-design with xPU architecture teams

Requirements:
• Passionate to develop compiler
• Advanced C/C++ programming skills
• Basic knowledge of CPU architecture

Desired:
• BSc/MSc/PhD in CS
• Experience of contributing to open-source projects
• Good at self-learning, courageous to explore new things, strong in practical skills
• Good communication skills and teamwork

Company Description: AI is powerful - but expensive, fragmented, and locked into a few hardware ecosystems. We believe everyone should have the freedom to build and run AI anywhere, without limits. Our mission: make AI’s compute layer unified, efficient, and accessible to all.

After decades working at the world’s largest tech companies, we saw the same barriers everywhere—high costs, complex tools, and closed platforms. These limit AI’s reach to a privileged few, stifling innovation and slowing real-world impact. Now, with offices in San Francisco, Los Altos, Boston, and Edinburgh, we're bringing powerful inference to everyone.

We’re building modular and composable infrastructure that simplifies AI development and deployment, and we're open sourcing the language and engine that makes it all possible. The Modular Platform unifies AI under a single framework, offering text, audio, and image inference - all with the state-of-the-art performance that you can deploy with shared endpoints, dedicated endpoints, in your cloud or ours, and with custom models. Modular's simple-to-use infrastructure and inference is designed to solve the problems that we knew developers were facing in the explosion of the AI industry.

Company Contact: Laura Nguyen - careers@modular.com

Job Title: Software Engineer, Hardware Enablement

Job Description: At Modular, we’re on a mission to revolutionize AI infrastructure by systematically rebuilding the AI software stack from the ground up. Our team, made up of industry leaders and experts, is building cutting-edge, modular infrastructure that simplifies AI development and deployment. By rethinking the complexities of AI systems, we’re empowering everyone to unlock AI’s full potential and tackle some of the world’s most pressing challenges.

If you’re passionate about shaping the future of AI and creating tools that make a real difference in people’s lives, we want you on our team. You can read about our culture and careers to understand how we work and what we value.

About the role:

ML developers today face significant friction in taking trained models into deployment. They work in a highly fragmented space, with incomplete and patchwork solutions that require significant performance tuning and non-generalizable, model-specific enhancements. At Modular, we are building the next generation AI platform that will radically improve the way developers build and deploy AI models.

As part of our mission to build AI's unified compute layer, we are expanding the Modular software stack to a variety of new and exciting hardware platforms.

We are looking for a motivated engineer to join the Hardware Enablement team at Modular. In this role you will work across the Modular software stack — from Mojo kernels and the graph compiler to MAX model serving — to bring up and optimize support for new accelerator platforms. You'll collaborate closely with internal teams and external hardware partners, and you'll develop deep expertise in novel architectures while contributing to our portability story.
 
LOCATION: Candidates based in and have authorization to work in Edinburgh, UK, Norway and the US are welcome to apply. We are unable to sponsor work visas at this time. To support growth and collaboration, those in earlier career stages work in a hybrid capacity at one of our office locations (minimum 2 days per week on-site). Onboarding for new hires is conducted in-person.

What you will do:

Implement and validate support for new hardware architectures across the Modular stack, working under the guidance of senior engineers on the team
Write and optimize Mojo kernels targeting novel accelerator architectures, with a focus on correctness first and performance iteration
Contribute to cross-team efforts improving portability infrastructure, tooling, and debugging workflows for new target hardware
Collaborate with hardware vendor engineers to understand target platforms, build integration tests, and triage platform-specific issues
Develop working knowledge of new hardware platforms — including ISA documentation, memory hierarchies, and vendor toolchains — and share findings with the team through demos and write-ups
Participate in company events such as on-sites and hackathons, contributing to a collaborative and open engineering culture

What you bring to the table:

 2+ years of experience in high-performance computing, compiler engineering, or related domains in industry or research
Proficiency in C++ and experience working in complex, multi-component software systems
Hands-on experience with at least one heterogeneous programming model (CUDA, SYCL, OpenCL, or similar), either as a user or contributor
Some exposure to non-GPU accelerator architectures (DSPs, NPUs, or other hardware accelerators) is a strong plus
Curiosity and willingness to learn new hardware platforms quickly, comfortable reading architecture manuals and vendor documentation
A collaborative, team-oriented attitude and alignment with our culture

Helpful, but not required:

Familiarity with how AI operators are implemented at a low level (e.g., experience writing or modifying GPU kernels, custom operators, or working with frameworks like PyTorch at the C++ layer)
Experience with GPU DSLs/DSELs such as Triton, CUTLASS, or CuTe
Familiarity with MLIR or LLVM compiler infrastructure
Experience working directly with hardware vendor teams or on platform bring-up efforts
Exposure to model serving or inference optimization workflows

What Modular brings to the table:
Amazing Team. We are a progressive and agile team with some of the industry’s best engineering and product leaders.
World-class Benefits. In order to attract the best, we need to offer the best. Premier insurance plans, up to 5% 401k matching, flexible paid time off, and more are available to you! Please note that specific benefit packages may vary based on your location.
Competitive Compensation. We offer very strong compensation packages, including stock options. We want people to be focused on their best work and believe in tailoring compensation plans to meet the needs of our workforce. 
Team Building Events. We organize regular team onsites and local meetups in Los Altos, CA as well as different cities. Traveling 2-4 times a year is expected for all roles. 

Working at Modular will enable you to grow quickly as you work alongside incredibly motivated and talented people who have high standards, possess a growth mindset, and a purpose to truly change the world.

Job Title: Lead Graph Compiler Engineer

Job Description: At Modular, we’re on a mission to revolutionize AI infrastructure by systematically rebuilding the AI software stack from the ground up. Our team, made up of industry leaders and experts, is building cutting-edge, modular infrastructure that simplifies AI development and deployment. By rethinking the complexities of AI systems, we’re empowering everyone to unlock AI’s full potential and tackle some of the world’s most pressing challenges.

If you’re passionate about shaping the future of AI and creating tools that make a real difference in people’s lives, we want you on our team. You can read about our culture and careers to understand how we work and what we value.

About the role:

The Modular graph compiler is a central piece of the MAX Framework which executes ML models at state-of-the-art performance across multiple hardware platforms. It uses a novel design based on integration with the Mojo programming language, provides best-in-class features (e.g. full parametric dynamic shapes), and uses advanced MLIR implementation techniques. Modular’s compiler stack is vertically integrated, from the API down to the kernel implementation - opening up opportunities for optimization not available in competing technologies.

This role involves extending and developing Modular’s novel graph compiler to support more features and hardwares. We’re looking for an ambitious engineer who is willing to grow their skills and be a technical lead to an incredible team. You will set strategic technical direction, contribute to the design and implementation of graph compiler features, and communicate with both leadership and users.

LOCATION: Candidates based in the US or Canada  are welcome to apply. You can work in our office in Los Altos, CA or remotely from home. Onboarding for new hires is conducted in-person in our Los Altos, CA office.

 
What you will do:

Set the technical direction for a team building the best heterogenous multi-device compiler and runtime system available, using MLIR and LLVM technologies.
Collaborate with the Mojo compiler team to influence and harness Mojo's powerful compile-time meta-programming capabilities.
Develop and disseminate deep expertise in new hardware platforms, partnering with Mojo kernel developers to achieve state-of-the-art performance on the latest hardware.
Tackle complex technical challenges to empower ML engineers and accelerate model development.
Strategically identify technical opportunities to grow the business.
Mentor and develop junior engineers, growing the capability of the team over time.
 

What you bring to the table:

5+ years of compiler engineering experience
A track record of challenging the status quo and delivering significant, measurable improvements.
Experience working with compilers for machine learning frameworks, such as PyTorch.
Knowledge of core compiler algorithms and data structures.
Knowledge of and experience working with MLIR and LLVM.
Knowledge of C++, as well as, knowledge of basic GitHub workflows like pull requests.
Creativity and curiosity for solving complex problems, a team-oriented attitude that enables you to work well with others, and alignment with our culture.
 

Helpful but not required:

Experience with ML graph optimizations, parallel / distributed programming, heterogeneous ML computation, and/or code generation.
Advanced degree in Computer Science or a related area is a plus.
 

What Modular brings to the table:

Amazing Team. We are a progressive and agile team with some of the industry’s best engineering and product leaders.
World-class Benefits. In order to attract the best, we need to offer the best. Premier insurance plans, up to 5% 401k matching, flexible paid time off, and more are available to you! Please note that specific benefit packages may vary based on your location.
Competitive Compensation. We offer very strong compensation packages, including stock options. We want people to be focused on their best work and believe in tailoring compensation plans to meet the needs of our workforce. 
Team Building Events. We organize regular team onsites and local meetups in Los Altos, CA as well as different cities. Traveling 2-4 times a year is expected for all roles. 
Working at Modular will enable you to grow quickly as you work alongside incredibly motivated and talented people who have high standards, possess a growth mindset, and a purpose to truly change the world.

Job Title: Mojo Compiler Engineer

Job Description: At Modular, we’re on a mission to revolutionize AI infrastructure by systematically rebuilding the AI software stack from the ground up. Our team, made up of industry leaders and experts, is building cutting-edge, modular infrastructure that simplifies AI development and deployment. By rethinking the complexities of AI systems, we’re empowering everyone to unlock AI’s full potential and tackle some of the world’s most pressing challenges.

If you’re passionate about shaping the future of AI and creating tools that make a real difference in people’s lives, we want you on our team. You can read about our culture and careers to understand how we work and what we value.

About the role:

Mojo is our secret sauce, spell-book, and Mjolnir — a new programming language that blends the soul of Python with the raw, elemental power of C. It’s designed to unlock every ounce of performance from modern AI hardware, while still being human-friendly and beautifully expressive.
We’re searching for a compiler engineer who sees programming languages as both art and architecture. You’ll help design and evolve the Mojo compiler — a next-generation toolchain built on MLIR — and push the boundaries of what’s possible in heterogeneous compute and AI systems.
You won’t be maintaining legacy. You’ll be inventing the future.

LOCATION: Candidates based in the US or Canada are welcome to apply. To support growth and collaboration, those in earlier career stages work in a hybrid capacity at our Los Altos, CA office (minimum 2 days per week on-site) with relocation assistance provided for out-of-state candidates. Senior members have both in office or remote flexibility. All new hires complete onboarding in-person.

What you will do:

Craft and implement new Mojo language features on top our MLIR-based compiler architecture.
Dream up and realize optimization passes that make GPUs, CPUs, and accelerators sing in harmony.
Shape the developer experience for Mojo — from syntax to performance semantics.
Collaborate with the Modular core teams, researchers, and an emerging open-source community.
Help define what "next-generation language design" means in the age of AI.


What you bring to the table:

3+ years working on compilers, runtimes, or language implementations.
Hands-on mastery of C++ (Python fluency a big plus).
Familiarity with compiler development frameworks such MLIR and LLVM.
Understanding of type systems, compiler optimization theory, and language semantics. (those who read the Pierce’s Types and Programming Languages, and can present a proof, will skip an interview)
Contributions to Clang, GCC, Swift, Rust, Dart, or another real-world programming language implementation.

Helpful, but not required:

Experience with HPC or accelerator ecosystems (CUDA, SYCL, OpenCL, etc.).
Experience working in type-system-heavy programming languages (Rust, Scala, Haskell).

What Modular brings to the table:

Amazing Team. We are a progressive and agile team with some of the industry’s best engineering and product leaders.
World-class Benefits. In order to attract the best, we need to offer the best. Premier insurance plans, up to 5% 401k matching, flexible paid time off, and more are available to you! Please note that specific benefit packages may vary based on your location.
Competitive Compensation. We offer very strong compensation packages, including stock options. We want people to be focused on their best work and believe in tailoring compensation plans to meet the needs of our workforce. 
Team Building Events. We organize regular team onsites and local meetups in Los Altos, CA as well as different cities. Traveling 2-4 times a year is expected for all roles. 
Working at Modular will enable you to grow quickly as you work alongside incredibly motivated and talented people who have high standards, possess a growth mindset, and a purpose to truly change the world.

Job Title: Head of Compiler and Runtime Engineering

Job Description: At Modular, we’re on a mission to revolutionize AI infrastructure by systematically rebuilding the AI software stack from the ground up. Our team, made up of industry leaders and experts, is building cutting-edge, modular infrastructure that simplifies AI development and deployment. By rethinking the complexities of AI systems, we’re empowering everyone to unlock AI’s full potential and tackle some of the world’s most pressing challenges.

If you’re passionate about shaping the future of AI and creating tools that make a real difference in people’s lives, we want you on our team. You can read about our culture and careers to understand how we work and what we value.

About the role:

ML developers today face significant friction in taking trained models into deployment. They work in a highly fragmented space, with incomplete and patchwork solutions that require significant performance tuning and non-generalizable/ model-specific enhancements. At Modular, we are building the next generation AI platform that will radically improve the way developers build and deploy AI models.

A core part of this offering is providing a platform that allows customers to achieve state-of-the-art performance across model families and hardware types. As the leader of the high-calibre MAX platform team, you will manage a team of engineers specializing in state of the art compiler technology, such as graph compilers and runtimes, and overlook the development of Mojo and MAX as the next generation technology platform for AI developers, and help revolutionize AI development and research.

This is a fantastic opportunity for a leader to help drive the core technology at Modular!

 
LOCATION: Candidates based in the US or Canada are welcome to apply. You can work in our office in Los Altos, CA or remotely from home. Onboarding for new hires is conducted in-person in our Los Altos, CA office.

What you will do:

 Leadership: Direct the MAX Platform team, focusing on the development of the Mojo programming language, AI compilers, runtimes, and associated tools.
Team Development: Cultivate and develop a high-performance team culture, fostering individual growth, collaboration, and innovation.
Cross-Functional Collaboration: Work closely with Modular’s engineering leaders, product and developer advocacy teams to align and drive initiatives.
Ecosystem Engagement: Drive the development of a technology platform that both customers and developers love.
Foster a vibrant developer community around MAX and Mojo, facilitating outreach and support for AI algorithm development on our platform, and
Assist customers in deploying AI inference solutions on the MAX platform.
Technical Oversight: Make strategic architectural decisions that influence the direction of large-scale projects and initiatives.


What you bring to the table:
 
 10+ years of experience in people management
15+ years of experience in the field of compilers, runtimes, performance, or related areas.
Proven experience in developing production-quality high-performance software.
Proven experience in managing large teams and as a manager of managers.
A robust understanding of the principles of AI compilers, languages and tools of various types of hardware.
Experience navigating a fast-paced startup environment.

What Modular brings to the table:
Amazing Team. We are a progressive and agile team with some of the industry’s best engineering and product leaders.
World-class Benefits. In order to attract the best, we need to offer the best. Premier insurance plans, up to 5% 401k matching, flexible paid time off, and more are available to you! Please note that specific benefit packages may vary based on your location.
Competitive Compensation. We offer very strong compensation packages, including stock options. We want people to be focused on their best work and believe in tailoring compensation plans to meet the needs of our workforce. 
Team Building Events. We organize regular team onsites and local meetups in Los Altos, CA as well as different cities. Traveling 2-4 times a year is expected for all roles. 
Working at Modular will enable you to grow quickly as you work alongside incredibly motivated and talented people who have high standards, possess a growth mindset, and a purpose to truly change the world.

Company Description:  NVIDIA pioneered accelerated computing. Today, our AI infrastructure powers global intelligence, transforming every industry. Discover how we’re transforming the world’s largest industries through our work and get your questions answered about starting a career with us. Whether you're exploring opportunities or learning more about our innovative work, our team is here to help.

Company Contact: Sarah Heraldo - sheraldo@nvidia.com

Job Title:  Senior Compiler Engineer - Backend GPU

Job Description: We are searching for a Senior Backend Compiler Engineer with experience in LLVM code generation for an exciting and fun role in our GPU Software organization. Our Compiler team is responsible for constructing and emitting the highest performance GPU machine instructions for Graphics (OpenGL, Vulkan, DX) and Compute (CUDA, PTX, OpenCL, Fortran, C++). This team is comprised of worldwide leading compiler engineering experts who provide leading edge performance and capabilities for NVIDIA's current and future complex parallel SIMT architectures.

What you will be doing:
• Guide the design and implementation of a new LLVM based backend JIT compiler targeting NVIDIA GPUs
• Design and develop new compiler passes and optimizations required to meet the high performance and low compilation time requirements of the compiler
• Work with global compiler, hardware and application teams to oversee improvements and problem resolutions
• Apply and adapt the latest compiler technologies to produce premier, robust, supportable compiler and tools
• Get a chance to influence and work on new GPU architectures

What we need to see:
• B.S. or higher degree in Computer Science/Engineering or equivalent experience
• 4+ years of compiler development experience including 3+ years of experience in LLVM with a focus in code generation
• Expertise in LLVM code generation including instruction scheduling, software pipelining, register allocation, GlobalISel, TableGen, LLVM IR, and Machine IR (MIR)
• Proficient in compiler design and in performance analysis tools
• Excellent hands-on C++ programming skills applied to industry standard C++ compilers and development frameworks
• Strong background in software engineering principles with a focus on crafting robust and maintainable solutions to challenging problems
• Good communication and documentation skills and self-motivated

Job Title:  Senior Compiler Engineer - Compute Front-End

Job Description: We are looking for a highly motivated Compiler Software Engineer to join this dynamic and innovative CUDA Compiler team. You will craft compiler capabilities that enable development and deployment of high performance parallel applications on NVIDIA's next generation GPU's. Our team delivers language extensions and performance improvements to better realize the potential of NVIDIA hardware for a growing range of computational workloads, ranging from deep learning, scientific computation, and self-driving cars to video games and virtual reality.

In this role, as a leader on our team, you have the opportunity of working on the optimizing CUDA C++ compiler, a core component of our platform for accelerating general purpose computation on the GPU. The CUDA programming language defines a unified programming model across a range of system configurations and hardware capabilities. The compiler is responsible for translating parallel programs written in the CUDA/C++ programming language into microcode that is optimized for fast execution on NVIDIA's family of industry-leading GPU architectures. You will be solving challenging problems alongside some of the top valued diverse minds in GPU computing and systems software, doing what you enjoy. If you’re excited by these challenges, we’d love to hear from you!

What you'll be doing:
• Design and implement enhancements to the CUDA C++ compiler driver and front-end, ensuring tight integration with industry-leading third-party compiler software.
• Contribute to the CUDA C++ programming language specification and its evolution factoring in support for new GPU hardware capabilities.
• Drive language conformance, front-end robustness, and compile-time improvements that benefit a wide range of users, from deep learning researchers to safety-critical software developers.
• Provide engineering support for the compiler and deliver targeted changes for production customer releases.
• Establish strong customer/partner relationships and trust through excellent execution and high-quality software development.

What we need to see:
• BS/MS/PhD degree in Computer Science, Computer Engineering, or related field (or equivalent experience)
• 4+ years of experience in developing production compilers or related software
• Knowledge of compiler front-end technology
• Advanced C++ development skills, including experience working with large, modularized codebases.
• Strong interpersonal, verbal and written communications skills
• Capability to achieve objectives under tight deadlines
• Experience executing tasks while managing competing priorities
• Experience writing and debugging multithreaded programs
• Deep understanding of technology and passion for what you do with strong teamwork, specifically a proven ability to effectively guide and influence within a dynamic matrix environment

Job Title:  Senior Deep Learning Compiler Engineer - XLA

Job Description: In this role, develop compiler optimization algorithms for deep learning workloads. You will optimize inference and training performance for the JAX framework and the OpenXLA compiler on NVIDIA GPUs at scale. You’ll collaborate with our partners in deep learning framework teams and our hardware architecture teams to accelerate the next generation of deep learning software. The scope of these efforts include:

• Crafting and implementing compiler optimization techniques for deep learning network graphs.
• Designing novel graph partitioning and tensor sharding techniques for distributed training and inference.
• Performance tuning and analysis.
• Code-generation for NVIDIA GPU backends using open-source compilers such as MLIR, LLVM and OpenAI Triton.
• Designing user facing features in JAX and related libraries and other general software engineering work.
• Working closely with GPU hardware engineering teams to design AI compiler software features for next-generation GPUs.

What we need to see:
• Bachelors, Masters or Ph.D. in Computer Science, Computer Engineering, related field (or equivalent experience).
• 4+ years of relevant work or research experience in performance analysis and compiler optimizations.
• Ability to work independently, define project goals and scope, and lead your own development effort adopting clean software engineering and testing practices.
• Excellent C/C++ programming and software design skills, including debugging, performance analysis, and test design.
• Strong foundation in architecture of CPU, GPUs or other high performance hardware accelerators. Knowledge of high-performance computing and distributed programming.
• CUDA or OpenCL programming experience is desired but not required.
• Experience with the following technologies is a huge plus: XLA, TVM, MLIR, LLVM, OpenAI Triton, deep learning models and algorithms, and deep learning framework design.
• Strong interpersonal skills are required along with the ability to work in a dynamic product-oriented team. A history of mentoring junior engineers and interns is a bonus.

Job Title:  Senior DL Compiler Engineer- CUDA Tile

Job Description: We are hiring software engineers for the CUDA Tile team. NVIDIA GPUs are at the center of the deep learning revolution and continue to enable breakthroughs in generative AI, large language models, recommendation systems, speech recognition, image classification and other areas. Come join us to work with a top-notch team and have broad impact across the entire deep learning community.

What you’ll be doing:
• In this role, you will be working on CUDA Tile, a new tile-based programming model for our GPUs. CUDA Tile shipped with CUDA 13.1 and is a major addition to CUDA (https://developer.nvidia.com/cuda/tile).
• You will design and implement compiler transformations, develop MLIR-based dialects and lowering passes, and optimize the performance of tile-based kernels to ensure they execute efficiently across multiple generations of NVIDIA GPU architectures.
• The scope of these efforts includes defining public APIs, crafting and implementing compiler and optimization techniques, performance optimization, and other general software engineering work.

What we need to see:
• Bachelors, Masters or Ph.D. in Computer Science, Computer Engineering or a related field (or equivalent experience)
• 3+ years of relevant work or research experience in compiler optimization, performance analysis and IR design.
• Ability to work independently, define project goals and scope, and lead your own development effort.
• Excellent C/C++ programming and software design skills, including debugging, performance analysis, and test design.
• Strong interpersonal skills are required along with the ability to work in a dynamic product-oriented team.

GOLD SPONSOR JOB LISTINGS

Company Description: We are a global leader in technology and business solutions that transform organizations and the world around us.

Our team’s mission is to develop toolchain for FUJITSU processors, such as the Arm-based FUJITSU-MONAKA CPU, to drive performance and quality improvement in Datacenter, AI and HPC applications.

Company Contact: Naoyuki Shida - shidax@fujitsu.com

Job Title: Compiler Engineering Manager

Job Description: You will drive development of the LLVM as an engineering manager of the toolchain for FUJITSU processors. You will also communicate with relevant departments, OSS community and partners to lead planning and performing development. 

Job Requirements:
- Compiler development experience
- Management experience
- Business level of Japanese language skill

Job Title: Compiler Engineering

Job Description: You will lead the development of the LLVM as a lead engineer to maximize the performance of FUJITSU processors. You will also communicate with relevant departments, OSS community and partners to contribute to team's results and value.  

Job Requirements:
- Compiler development experience
- Business level of Japanese language skill

SUPPORTER SPONSOR JOB LISTINGS

Company Description: Join Graphcore - Explore Opportunities in AI & Advanced Compute

Location: Bristol, UK | London, UK | Cambridge, UK | Gdańsk, Poland | Austin, TX, US | Bengaluru, India 

At Graphcore, we’re building the compute that will power the next generation of AI.  

From designing cutting-edge silicon to developing world-class software and systems, our teams are pushing the boundaries of what’s possible in machine intelligence. As part of the SoftBank Group, and following significant investment in our future, we’re working towards a bold vision: enabling AI at global scale. 

We’re growing - fast - and we’re looking for curious, ambitious people who want to work on problems that genuinely matter. 

Company Contact: Moe Ahmed - moea@graphcore.ai

Job Title: Register your interest

Job Description: Rather than applying for a single role, this is your chance to explore the breadth of opportunities we have across Graphcore. Whether your background is in software, hardware, AI research, systems engineering, or something adjacent, we’d love to hear from you. 

Where you could make an impact: 

Software Engineering (C++, Python, compilers, systems, ML frameworks, LLVM/MLIR) 

Machine Learning and AI Research 

Silicon and Hardware Engineering 

Systems and Infrastructure 

Performance and Optimisation 

Developer Experience and Tools 

…and more. 

Why Graphcore? 

You’ll work on technology that enables the next generation of machine intelligence - with the potential to transform industries from healthcare to science and beyond. 

You’ll collaborate with people across silicon, systems and software, combining different perspectives to solve problems that don’t have obvious answers. You’ll be trusted to take ownership - contributing ideas, shaping decisions, and having real impact as we continue to grow. 

Interested? 

Register your interest in seconds - just share your name, email, and phone number, and we’ll be in touch to start the conversation. 

Company Description: The Intel NPU organization is dedicated to the development of hardware accelerators for the future of AI, enabling unprecedented machine intelligence on billions of edge, desktop and mobile computers. Doing what’s never been done before takes vision, innovation, and the world’s best talent. You’ll be immersed in a diverse, encouraging environment where everyone is encouraged to do their best work. Join the adventure of harnessing the potential of state-of-art AI workloads networks and the most advanced AI hardware accelerator in the world. 

Company Contact: John Brady - john.brady@intel.com

Job Title: NPU MLIR Compiler engineer

Job Description: Intel is hiring software engineers for its deep learning compiler team. Our customers around the world are using NPUs to power a revolution in deep learning, enabling breakthroughs in problems from image classification, speech recognition to natural language processing. As deep learning models are becoming more versatile, using compiler technologies to achieve high performance becomes essential.

We are looking for individuals who are interested in building the world's leading deep learning compiler for current and future Intel NPUs. This is a product development position with the end goal being high-quality, high-performance, secure product software. You will start optimizing pre-silicon and have access to hardware shortly after it is powered on.

You will be responsible for analyzing deep learning networks and developing compiler optimization algorithms. Collaborate with members of the deep learning software framework teams and the hardware architecture teams to accelerate the next generation of deep learning software. The scope of these efforts includes performance tuning and analysis, developing and implementing compiler and optimization techniques for neural networks, and other general software engineering work.

You will conduct software development and optimizations in the following areas:
- Develop MLIR based compiler technology for deep learning workloads on Intel NPUs.
- Develop AI execution middleware based on Intel OpenVINO toolkit.
- Develop large-scale production software with validation and continuous integration in mind.
- Collaborate with frameworks teams to develop compiler optimizations for the deep learning domain.
- Collaborate and coordinate internally and externally with cross geographical teams such as execution runtime software, NPU hardware, infrastructure, and front-end teams on the same project.

Qualifications: You must possess the below minimum qualifications to be initially considered for this position. Preferred qualifications are in addition to the minimum requirements and are considered a plus factor in identifying top candidates.

Minimal Qualifications:

  • BS/MS in Computer Science or a similar field
  • At least 4 years of experience in C++ programming, software design skills, including debugging, performance analysis, and test design.
  • Excellent data structure, algorithms and modern C++ programming skills
  • Strong interpersonal skills are required along with the ability to work in a multifaceted product-oriented team.
  • Experience in one the following: compiler technologies, deep-learning frameworks or algorithms, high-performance computing, computer vision, numerical modelling
  • Positive can-do attitude, desire to deliver results and winning products.
  • Excellent written and oral communication skills.
  • You should have a passion for optimization and performance, close to hardware, as well as for good software engineering practice and usability.

Preferred Qualifications:

  • Experience in developing modern compilers, especially LLVM and MLIR
  • Experience with formulating optimization problems and using ILP solvers (e.g. ORTOOLs, CPLEX, SCIP)
  • Experience in AI hardware accelerators, GPU, heterogeneous architectures software development
  • A background in performance analysis and optimization, particularly in machine-learning
  • A solid understanding of modern machine-learning primitives and LLMs
  • A background in Python, modern AI frameworks and ecosystem (e.g., Torch, HuggingFace, llama.cpp, etc.)

If you are interested in having your job openings posted, details on becoming a sponsor may be found here.