Would you be interested in exploring a perm full-time role for our start-up in Palo Alto, CA who are specialized in building GenAI infrastructure concentrating on the Voice/Audio/Speech, Vision, Multi-modal platforms.
If you are an expertise in any of this space: design, develop, train, fine-tune, implement state-of-art optimizing techniques and deploy these LLMs, Speech, vision, multi-modal (multilingual) foundational models using cutting edge technologies and frameworks, then you are at the right place!!
Qualifications:
- Ph.D. or MS with 2+ years of research / applied research experience in LLMs, NLP, CV, Reinforcement Learning, Voice, and Generative models
- Demonstrated expertise in high-performance computing with proficiency in Python, C/C++, CUDA, and kernel-level programming for AI applications
- Extensive experience in the optimization of training and inference for large-scale AI models, including practical knowledge of quantization, distillation, and LLMOps
- Prior experience with large-scale distributed training and fine-tuning of foundation models such as GPT-3, LLaMA2, AlphaFold, and DALL-E
- Experience with language modeling evaluation, prompt tuning and engineering, instruction tuning, and/or RLHF
- Research contributions in NLP, generative modeling, LLMs demonstrated through publications and products
- Strong programming skills and proficiency in Python, TensorFlow/PyTorch, and other ML frameworks and tools
- Experience in Information Extraction, Question Answering, Conversational Agents (Chatbots), Data Visualization and/or text-to-image models
Please reach out to Jia for more information.