Tenstorrent

Inference Server – Product Software

Belgrade, Serbia Full Time

Tenstorrent is leading the industry on cutting-edge AI technology, revolutionizing performance expectations, ease of use, and cost efficiency. With AI redefining the computing paradigm, solutions must evolve to unify innovations in software models, compilers, platforms, networking, and semiconductors. Our diverse team of technologists have developed a high performance RISC-V CPU from scratch, and share a passion for AI and a deep desire to build the best AI platform possible. We value collaboration, curiosity, and a commitment to solving hard problems. We are growing our team and looking for contributors of all seniorities.

Join our Inference Server Technologies team, where we build the software layer that powers state-of-the-art AI inference on Tenstorrent hardware. This team develops APIs, deploys workloads, and benchmarks end-to-end model performance so developers can efficiently scale inference on our stack.

This role is hybrid based in Belgrade, Serbia.

You will work on a project under the guidance of experienced engineers and a dedicated mentor.

 
Who You Are
  • Final-year BSc or MSc student in Computer Science, Software Engineering, Electrical Engineering, or a related technical field
  • Strong programming fundamentals in Python, with familiarity in C++ considered a plus
  • Interested in backend systems, API design, and how ML models are deployed in production environments
  • Curious about performance optimization techniques such as batching, caching, and model parallelism
  • Motivated to learn and contribute in a collaborative engineering environment


What We Need
  • Contribute to backend features and APIs that support AI inference workloads
  • Assist in deploying, testing, and benchmarking models running on Tenstorrent hardware
  • Analyze inference performance and help identify optimization opportunities
  • Write clean, maintainable code with guidance from senior engineers
  • Collaborate with the team to improve reliability, usability, and performance of the inference server stack


What You Will Learn
  • How end-to-end ML inference is optimized on custom AI hardware
  • How scalable backend systems are designed to serve real-world AI applications
  • How APIs and infrastructure shape the developer experience for AI workloads
  • Practical performance analysis techniques in production-like environments
  • How modern AI software stacks integrate models, runtimes, and hardware
 
Tenstorrent offers a highly competitive compensation package and benefits, and we are an equal opportunity employer.

This offer of employment is contingent upon the applicant being eligible to access U.S. export-controlled technology. Due to U.S. export laws, including those codified in the U.S. Export Administration Regulations (EAR), the Company is required to ensure compliance with these laws when transferring technology to nationals of certain countries (such as EAR Country Groups D:1, E1, and E2). These requirements apply to persons located in the U.S. and all countries outside the U.S. As the position offered will have direct and/or indirect access to information, systems, or technologies subject to these laws, the offer may be contingent upon your citizenship/permanent residency status or ability to obtain prior license approval from the U.S. Commerce Department or applicable federal agency. If employment is not possible due to U.S. export laws, any offer of employment will be rescinded.