Machine Learning Engineer
Our company is an AI company developing a personalized operating system through a natural language interface. Our key technologies speed up multimedia language model applications by an order of magnitude and enable real-time, application-programming-interface-free agency on any platform. We are focused on developing a dedicated hardware device for our operating system, which will be one of the first of its kind.
We are a small, fast-moving team based in Los Angeles. We are well-funded and backed by Khosla Ventures, led by Vinod Khosla, co-founder of Sun Microsystems.
At our company, we believe that the best way to realize the value of cutting-edge research is by creating delightful, end-user-oriented experiences with taste. We pride ourselves on being uniquely positioned to capture the opportunities during this technology renaissance with our product focus. Our team has created mature software, hardware, and designs that have impacted millions of people, making us one of the few capable of all components of the vertically integrated chain.
Our Core Values
We value individuals with:
- a deep curiosity about technology
- out-of-the-box problem-solving skills
- an understanding of building and operating complex systems in production
- an obsession with customers and crafting experiences
- a sense of responsibility and teamwork
About The Role
The machine learning team at our prototypes and builds core machine learning systems that make our operating system experience unique, ranging from large language models to multi-modal foundation models. Your work will be both pragmatic and pioneering: the models you develop will support the day-to-day functioning of our operating system in areas such as computer vision, conversational intelligence, speech recognition, and synthesis. Furthermore, your work will expand our understanding of how AI interacts with the world in a way that is similar to human interaction.
Your day-to-day responsibilities will look like:
- Maintain, iterate, and deploy machine learning algorithms in production on our cloud systems.
- Research and propose novel methods for some of our most challenging problems that require a machine-learning-based solution. Oversee experimental progress through large-scale training and benchmarking pipelines, and formalize them into alpha and beta features of our operating system.
- Evaluate third-party machine learning services and cloud computing vendors for their performance. Provide a perspective on which to build in-house, which to benefit from external partnerships.
- Independently execute tasks, solve problems, and report to technical leadership, with the goal of leading a broader team of full-time employees and interns in the future.
We’d love to see you comfortable with:
Ideal candidates should be familiar with at least two parts of the full lifecycle of model development:
- Acquiring large-scale data, either through scraping or proprietary acquisition, of traditional multi-modal data (text, image, code) and data of a more esoteric nature.
- Curating subsets of large datasets to meet product needs while remaining IP compliant, and serving such data to the training compute infrastructure.
- Understanding and characterizing the behavior and limitations of existing models.
- Developing, implementing, and testing plans to improve such models with cutting-edge, original research.
- Utilizing distributed training frameworks on large compute clusters with multi-node GPU and high-speed interconnect.
- Tuning inter-node bandwidth over communication protocols like RDMA over Converged Ethernet (RoCE) with software integrations including NVIDIA Collective Communications Library (NCCL), Gloo, or Message Passing Interface (MPI).
- Developing training and fine-tuning pipelines for large models at scale.
- Optimizing and deploying models on a variety of homogeneous and heterogeneous hardware on the cloud or at the edge.
This role is ideal for candidates who have two or more years of industry experience.
Nice to haves:
Here is a brief overview of the models we are using, training, and deploying. Any experience working with them is valued and appreciated:
- Large language models for conversation, structured data processing, tool usage, and reasoning
- Streaming speech recognition models
- Streaming speech synthesis models (concatenative and neural)
- Empowering language models with long-term memory and real-time information: Retrieval Augmented Generation, integration with vector database
- Multi-modal foundation model for interface understanding (UI & UX), from rasterized image to DOM
- Named entity recognition, utterance and sentence segmentation, and zero-shot text classification and embedding models
- Open-world instance segmentation and object-tracking models
- Landmark detection of man-made and natural structures (tourist attractions, vendors)
- Object re-identification models
This role is based in our Los Angeles headquarters. However, we may consider remote candidates in exceptional circumstances.
Compensation And Benefits
The annual salary range for this role is $150,000 — $200,000. Total compensation also includes generous equity, ranging from 0.2% to 0.75%. The final package may vary depending on job-related knowledge, skills, candidate location, and experience. Benefits include medical, dental, and vision insurance for you and your family.
At our company, we are committed to creating an inclusive and diverse workplace. We welcome and encourage applicants from all backgrounds, and do not discriminate on the basis of race, religion, national origin, gender, sexual orientation, age, veteran status, disability, or any other legally protected status.