Posted 2 weeks ago
Roles
LLM Inference Optimization Engineer MN-Core LLM Serving Engine EngineerCompensation Summary
Visa and relocation support provided.
Locations
Tokyo, Japan, Remote in Japan
Description
Preferred Networks is an AI company based in Tokyo working across the stack. They design in-house chips (MN-Core series) and train LLMs (PLaMo series). They are hiring two roles: (1) MN-Core LLM Serving Engine Engineer to build software infrastructure to serve LLMs using the upcoming inference accelerator MN-Core L1000. (2) LLM Inference Optimization Engineer to improve the inference engine powering their API service and maintain PLaMo implementations in open source projects such as vLLM. Both roles require relocation to Japan, and the company provides visa and relocation support.
Similar Jobs
Create your own personalized Job Alert