d-Matrix hiring ML Compiler Architect, Senior Principal • Remote (Work from Home) | Himalayas
d-MatrixD-

ML Compiler Architect, Senior Principal

d-Matrix is at the forefront of AI inference technology, providing innovative solutions that enhance the efficiency and scalability of AI applications.

d-Matrix

Employee count: 51-200

Canada only

At d-Matrix, we are focused on unleashing the potential of generative AI to power the transformation of technology. We are at the forefront of software and hardware innovation, pushing the boundaries of what is possible. Our culture is one of respect and collaboration.

We value humility and believe in direct communication. Our team is inclusive, and our differing perspectives allow for better solutions. We are seeking individuals passionate about tackling challenges and are driven by execution. Ready to come find your playground? Together, we can help shape the endless possibilities of AI.

Location:

Hybrid, working onsite at our Toronto, Ontario, Canada headquarters 3-5 days per week.

Role: Software Compiler Architect – MLIR/LLVM for Cloud Inference

What You Will Do:

As a hands-on Front-End Software Compiler Architect focused on cloud-based AI inference, you will drive the design and implementation of a scalable MLIR-based compiler framework optimized for deploying large-scale NLP and transformer models in cloud environments. You will architect the end-to-end software pipeline that translates high-level AI models into efficient, low-latency executables on a distributed, multi-chiplet hardware platform featuring heterogeneous compute elements such as in-memory tensor processors, vector engines, and hierarchical memory.

Your compiler designs will enable dynamic partitioning, scheduling, and deployment of inference workloads across a cloud-scale infrastructure, supporting both statically compiled and runtime-optimized execution paths. You will focus on compiler strategies that minimize inference latency, maximize throughput, and efficiently utilize compute and memory resources in data center environments, in addition to your work on developing the compiler.

You will collaborate cross-functionally with systems architects, ML framework teams, runtime developers, performance engineers, and cloud orchestration groups to ensure seamless integration and optimized inference delivery at scale.

Key Responsibilities:

  • Architect the MLIR-based compiler for cloud inference workloads, focusing on efficient mapping of large-scale AI models (e.g., LLMs, Transformers, Torch-MLIR) onto distributed compute and memory hierarchies.
  • Lead the development of compiler passes for model partitioning, operator fusion, tensor layout optimization, memory tiling, and latency-aware scheduling.
  • Design support for hybrid offline/online compilation and deployment flows with runtime-aware mapping, allowing for adaptive resource utilization and load balancing in cloud scenarios.
  • Define compiler abstractions that interoperate efficiently with runtime systems, orchestration layers, and cloud deployment frameworks.
  • Drive scalability, reproducibility, and performance through well-designed IR transformations and distributed execution strategies.
  • Mentor and guide a team of compiler engineers to deliver high-performance inference-optimized software stacks.

What You Will Bring:

  • BS 15+ Yrs / MS 12+ Yrs / PhD 10+ Yrs Computer Science or Electrical Engineering, with 12+ years of experience in Front End Compiler and systems software development, with a focus on ML inference.
  • Deep experience in designing or leading compiler efforts using MLIR, LLVM, Torch-MLIR, or similar frameworks.
  • Strong understanding of model optimization for inference: quantization, fusion, tensor layout transformation, memory hierarchy utilization, and scheduling.
  • Expertise in deploying ML models to heterogeneous compute environments, with specific attention to latency, throughput, and resource scaling in cloud systems.
  • Proven track record working with AI frameworks (e.g., PyTorch, TensorFlow), ONNX, and hardware backends.
  • Experience with cloud infrastructure, including resource provisioning, distributed execution, and profiling tools.

Preferred Qualifications:

  • Experience targeting inference accelerators (AI ASICs, FPGAs, GPUs) in cloud-scale deployments.
  • Knowledge of cloud deployment orchestration (e.g., Kubernetes, containerized AI workloads).
  • Strong leadership skills with experience mentoring teams and collaborating with large-scale software and hardware organizations.
  • Excellent written and verbal communication; capable of presenting complex compiler architectures and trade-offs to both technical and executive stakeholders.

This role is a cornerstone of our cloud AI software strategy. You'll shape the way inference workloads are deployed, optimized, and scaled across data center infrastructure.

Equal Opportunity Employment Policy

d-Matrix is proud to be an equal opportunity workplace and affirmative action employer. We’re committed to fostering an inclusive environment where everyone feels welcomed and empowered to do their best work. We hire the best talent for our teams, regardless of race, religion, color, age, disability, sex, gender identity, sexual orientation, ancestry, genetic information, marital status, national origin, political affiliation, or veteran status. Our focus is on hiring teammates with humble expertise, kindness, dedication and a willingness to embrace challenges and learn together every day.

d-Matrix does not accept resumes or candidate submissions from external agencies. We appreciate the interest and effort of recruitment firms, but we kindly request that individual interested in opportunities with d-Matrix apply directly through our official channels. This approach allows us to streamline our hiring processes and maintain a consistent and fair evaluation of al applicants. Thank you for your understanding and cooperation.

About the job

Apply before

Posted on

Job type

Full Time

Experience level

Senior
Executive

Location requirements

Hiring timezones

Canada +/- 0 hours

About d-Matrix

Learn more about d-Matrix and their company culture.

View company profile

d-Matrix is a revolutionary company founded in 2019, focused on delivering cutting-edge AI computing solutions specifically tailored for generative AI applications at scale. The company has created its flagship product, Corsair™, which is recognized as the world’s most efficient AI inference platform for data centers. By taking a unique approach that combines high performance with sustainability, d-Matrix is dedicated to transforming the way AI is utilized in commercial settings.

The leadership of d-Matrix includes experienced entrepreneurs who have a history of success in the semiconductor industry, collectively boasting over 20 years of expertise and the accomplishment of shipping over 100 million chips while generating substantial revenues exceeding $1 billion. With significant backing from investors, including Microsoft, d-Matrix is positioned as a key player in the AI semiconductor landscape, aiming to make AI inference not only viable but also accessible to a wider range of applications and businesses.

Claim this profiled-Matrix logoD-

d-Matrix

View company profile

Similar remote jobs

Here are other jobs you might want to apply for.

View all remote jobs

22 remote jobs at d-Matrix

Explore the variety of open remote roles at d-Matrix, offering flexible work options across multiple disciplines and skill levels.

View all jobs at d-Matrix
d-Matrix logoD-
United States only

Principal Software Engineer, Distributed Systems

d-Matrix

Employee count: 51-200

Salary: 196k-327k USD

d-Matrix logoD-
United States only

AI / ML System Software Engineer, Senior Staff

d-Matrix

Employee count: 51-200

Salary: 180k-280k USD

d-Matrix logoD-
United States only

Machine Learning Engineer, Staff - Model Factory

d-Matrix

Employee count: 51-200

Salary: 155k-250k USD

d-Matrix logoD-
United States only

Staff Engineer, Developer and Qualification Tools

d-Matrix

Employee count: 51-200

Salary: 132k-234k USD

d-Matrix logoD-
United States only

Software Engineer, Senior Staff- AI/ML Kernels

d-Matrix

Employee count: 51-200

Salary: 180k-300k USD

Remote companies like d-Matrix

Find your next opportunity by exploring profiles of companies that are similar to d-Matrix. Compare culture, benefits, and job openings on Himalayas.

View all companies

Find your dream job

Sign up now and join over 100,000 remote workers who receive personalized job alerts, curated job matches, and more for free!

Sign up
Himalayas profile for an example user named Frankie Sullivan