Roles

LLM Inference Optimization Engineer MN-Core LLM Serving Engine Engineer

Compensation Summary

Visa and relocation support provided.

Locations

Tokyo, Japan, Remote in Japan

Description

Preferred Networks is an AI company based in Tokyo working across the stack. They design in-house chips (MN-Core series) and train LLMs (PLaMo series). They are hiring two roles: (1) MN-Core LLM Serving Engine Engineer to build software infrastructure to serve LLMs using the upcoming inference accelerator MN-Core L1000. (2) LLM Inference Optimization Engineer to improve the inference engine powering their API service and maintain PLaMo implementations in open source projects such as vLLM. Both roles require relocation to Japan, and the company provides visa and relocation support.

Similar Jobs

Create your own personalized Job Alert