NEXA AI invented functional tokens and Octopus models for AI agents. We provide more accurate AI agent solutions, 4x faster and 10x cheaper than OpenAI GPT-4o API, with a latency of ~0.3s. Our product Octoverse enables developers to build AI Companions that understand and complete tasks for your users in apps. Learn more at nexa4ai.com.
Responsibilities:
Build, train and deploy LLMs.
Create and build LLM infra for training and inference.
Optimize LLM inference performance on edge devices.
Full stack development, if needed.
You may be a good fit if you:
Are enthusiastic about the AI and machine learning industry, demonstrating self-motivation and a proactive approach.
Possess strong technical skills and can effectively communicate with technical personnel.
Have experience in ML model optimization and performance tuning.
Are proficient in Python, and have experience with C or C++.
Have a solid understanding of cloud platforms like Google Cloud and AWS.
Can collaborate effectively with cross-functional teams to deliver high-quality ML solutions.