Software Engineer, Modular

Cloud Inference

$157-299k

Salary for candidates based in Canada is $158,000.00 - $270,000.00 CAD. Plus annual target bonus and equity

Kubernetes
C++
PyTorch
Junior, Mid and Senior level
Remote in Canada, US
San Francisco Bay Area
Modular

AI infrastructure for developers

Open for applications

Modular

AI infrastructure for developers

201-500 employees

B2BArtificial IntelligenceMachine Learning

Open for applications

$157-299k

Salary for candidates based in Canada is $158,000.00 - $270,000.00 CAD. Plus annual target bonus and equity

Kubernetes
C++
PyTorch
Junior, Mid and Senior level
Remote in Canada, US
San Francisco Bay Area

201-500 employees

B2BArtificial IntelligenceMachine Learning

Company mission

To have real, positive impact in the world by reinventing the way AI technology is developed and deployed into production with a next-generation developer platform.

Role

Who you are

  • If you’re passionate about shaping the future of AI and creating tools that make a real difference in people’s lives, we want you on our team
  • We are looking for candidates based on their breadth and depth of experience in backend engineering, AI inference, and distributed systems development
  • If this sounds exciting, we invite you to join our world-leading AI infrastructure team and help drive our industry forward!
  • Experience working on high scale ML inference infrastructure (traditional AI or genAI)
  • Experience with kubernetes
  • Familiarity with HuggingFace API and workflows for using community models
  • Ability to create durable, reusable software tools and libraries that are leveraged across teams and functions
  • Experience in machine learning technologies and use cases
  • Creativity and curiosity for solving complex problems, a team-oriented attitude that enables you to work well with others, and alignment with our culture
  • Strongly identifies with our core company cultural values

Desirable

  • Experience building ML models in PyTorch
  • Familiarity with modern C++
  • Some experience with and interest in Mojo, our AI programming language!

What the job involves

  • In the Cloud Inference team, we are focused on building end to end distributed LLM inference deployments that are fully vertically integrated with the MAX stack
  • Work with Product and partner engineering teams to design and ship new inference server features
  • Collaborate with our kernels and genAI teams to achieve SOTA performance at the serving layer
  • Help design and develop helm charts and cloud services for scaled LLM inference (intelligent routing, distributed kvcache management, disaggregated inference, etc.)

Share this job

View 2 more jobs at Modular

Company

Company benefits

  • Leading medical, dental and vision packages
  • Equity packages
  • Generous maternity & paternity leave
  • 401K Plan
  • Work wherever you want
  • Unlimited Vacation & PTO
  • Corporate perks & epic team fun
  • Great set up

Funding (2 rounds)

Aug 2023

$100m

LATE VC

Jun 2022

$30m

EARLY VC

Total funding: $130m

Our take

Modular has been developing a programming language optimized for developing AI software, to make it easier for developers to complete their AI projects faster and more effectively.

Most of the time, developers will write AI models in Python because of its relatively simple and concise syntax. However, with this simplicity comes performance issues that can slow down programs and code, a problem that Modular hopes to address.

The company's language, named Mojo, combines the usability of Python with the performance of C; which, the company says, brings unparalleled programmability of AI hardware and extensibility of AI models.

With significant funding at its back, the company plans to enhance Mojo moving forward, as well as invest in its other product - a software tool called AI Engine that is designed to make companies’ neural networks faster.

Kirsty headshot

Kirsty

Company Specialist at Welcome to the Jungle