
Preferred Networks is hiring a
LLM Inference Optimization Engineer
About Preferred Networks
Preferred Networks is an AI company based in Tokyo working across the stack, from AI chips and computing infrastructure to LLMs and products. You may already know us indirectly if you\'ve used software we\'ve built, such as Optuna or CuPy (or Chainer, back in the day). We are designing in-house chips (MN-Core series) and training LLMs (PLaMo series). Our team is actively hiring for two roles related to these endeavors.
Job Description
Improve the inference engine powering our API service and maintain PLaMo implementations in open source projects such as vLLM.Loading...
Share this job
Preferred Networks has 1 other job listed
Hiring engineers?
Reach thousands of tech candidates from the Hacker News community.
Post a Job β $99





