The Role
As an AI Scientist Intern, you will assist in fine-tuning large language models, conduct research in AI, and collaborate with teams on model implementation.
Summary Generated by Built In
About Mistral
At Mistral AI, we believe in the power of AI to simplify tasks, save time, and enhance learning and creativity. Our technology is designed to integrate seamlessly into daily working life.
We democratize AI through high-performance, optimized, open-source and cutting-edge models, products and solutions. Our comprehensive AI platform is designed to meet enterprise needs, whether on-premises or in cloud environments. Our offerings include le Chat, the AI assistant for life and work.
We are a dynamic, collaborative team passionate about AI and its potential to transform society.
Our diverse workforce thrives in competitive environments and is committed to driving innovation. Our teams are distributed between France, USA, UK, Germany and Singapore. We are creative, low-ego and team-spirited.
Join us to be part of a pioneering company shaping the future of AI. Together, we can make a meaningful impact. See more about our culture on https://mistral.ai/careers.
Mistral AI are hiring experts in the role of pre-training and fine-tuning large language models.
Role Summary
-You will be working with the fine tuning team on making state-of-the-art generative models.
-You will run autonomous work streams under the supervision of experienced scientists.
-The role is based in our Bay area offices
-Internship duration : 3 to 6 months. We will only consider candidates looking for end of studies internships (Phd)
What you will do
-Explore state-of-the-art LLM algorithms for fine tuning LLMs, with the supervision of top level scientists.
-Assist in the design and implementation of machine learning models and algorithms.
-Conduct research on the latest advancements in natural language processing and LLMs.
-Contribute to the development and optimization of our LLM systems.
-Collaborate with cross-functional teams to integrate LLM technologies into various applications.
-Perform data analysis and visualization to support research and development efforts.
-Document research findings and contribute to technical reports and publications.
-Participate in team meetings and brainstorming sessions to share ideas and insights
About you
-Currently doing a Phd from tier 1 engineering schools / Universities.
-High scientific understanding of the field of generative AI.
-Broad knowledge of the field of AI, and specific knowledge or interest in fine-tuning and using language models for applications.
-Strong programming skills in Python, with experience in libraries such as TensorFlow, PyTorch, or similar.
-Familiarity with natural language processing techniques and machine learning algorithms.
-Design complex software and make them usable in production.
-Navigate the full MLOps technical stack, with a focus on architecture development and model evaluation and usage.
-Previous experience with LLMs or related technologies.
-Knowledge of deep learning frameworks and techniques..Experience with version control systems (e.g., Git) and linux shell environment.
Now, it would be ideal if you :
-Have experience in fine tuning LLMs.
-Have used complex HPC infrastructure with full autonomy.
Top Skills
Machine Learning
Natural Language Processing
Python
PyTorch
TensorFlow
Am I A Good Fit?
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.
Success! Refresh the page to see how your skills align with this role.
The Company
What We Do
Fast, open-source and secure language models. Facilitated specialisation of models on business use-cases, leveraging private data and usage feedback.
Built from a world-class team in Europe, targeting global market. Join the team ! https://jobs.lever.co/mistral/