INTERNSHIP DETAILS

Research Intern, Model Shaping (Summer 2026)

CompanyTogether AI
LocationSan Francisco
Work ModeOn Site
PostedJanuary 6, 2026
Internship Information
Core Responsibilities
Research and implement novel techniques in focus areas such as advanced post-training methods and efficient training systems. Document findings and integrate research results into Together AI products.
Internship Type
full time
Salary Range
$58 - $63
Company Size
304
Visa Sponsorship
No
Language
English
Working Hours
40 hours
Apply Now →

You'll be redirected to
the company's application page

About The Company
Together AI is a research-driven AI cloud infrastructure provider. Our purpose-built GPU cloud platform empowers AI engineers and researchers to train, fine-tune, and run frontier class AI models. Our customers include leading SaaS companies such as Salesforce, Zoom, and Zomato, as well as pioneering AI startups like ElevenLabs, Hedra, and Cartesia. We advocate for open source AI and believe that transparent AI systems will drive innovation and create the best outcomes for society.
About the Role
<h3><strong>About Together AI</strong></h3> <p>Together AI is a research-driven artificial intelligence company. We believe open and transparent AI systems will drive innovation and create the best outcomes for society, and together we are on a mission to significantly lower the cost of modern AI systems by co-designing software, algorithms, and models. We have contributed to leading open-source research, models, and datasets to advance the frontier of AI, and our team has been behind technological advancements such as FlashAttention, Mamba, FlexGen, SWARM Parallelism, Mixture of Agents, and RedPajama.</p> <h3><strong>Role Overview</strong></h3> <p>As a Research Intern in the Model Shaping team, you will work on one or more of the following areas:</p> <ul> <li>Advanced post-training methods across supervised learning, preference optimization, and reinforcement learning</li> <li>New techniques and systems for efficient training of neural networks (e.g., distributed training, algorithmic improvements, optimization methods)</li> <li>Robust and reliable evaluation of foundation model capabilities</li> </ul> <p>The Model Shaping team at Together AI works on products and research for tailoring open foundation models to downstream applications. We build services that allow machine learning developers to choose the best models for their tasks and further improve these models using domain-specific data. In addition to that, we develop new methods for more efficient model training and evaluation, drawing inspiration from a broad spectrum of ideas across machine learning, natural language processing, and ML systems.</p> <p>Past research led by Model Shaping interns resulted in the following papers:</p> <ul> <li><a href="https://arxiv.org/abs/2511.21667">Escaping the Verifier: Learning to Reason via Demonstrations</a></li> <li>​​<a href="https://arxiv.org/abs/2505.17967">FFT-based Dynamic Subspace Selection for Low-Rank Adaptive Optimization of Large Language Models</a></li> </ul> <h3><strong>Responsibilities</strong></h3> <ul> <li>Research and implement novel techniques in one or more of our focus areas</li> <li>Design and conduct rigorous experiments to validate hypotheses</li> <li>Document findings in scientific publications and blog posts</li> <li>Integrate the research results into Together products</li> <li>Communicate the plans, progress, and results of projects to the broader team</li> </ul> <h3><strong>Requirements</strong></h3> <ul> <li>Currently pursuing a Bachelor's, Master's, or Ph.D. degree in Computer Science, Electrical Engineering, or a related field</li> <li>Strong knowledge of Machine Learning and Deep Learning fundamentals</li> <li>Experience with deep learning frameworks (PyTorch, JAX, etc.)</li> <li>Strong programming skills in Python</li> <li>Familiarity with Transformer architectures and recent developments in foundation models</li> </ul> <h3><strong>Preferred Qualifications</strong></h3> <ul> <li>Prior research experience with foundation models or efficient machine learning</li> <li>Publications at leading ML and NLP conferences (such as NeurIPS, ICML, ICLR, ACL, or EMNLP)</li> <li>Understanding of model optimization techniques and hardware acceleration approaches</li> <li>Contributions to open-source machine learning projects</li> </ul> <h3><strong>Internship Details</strong></h3> <ul> <li>Duration: ~12 weeks (Summer 2026)</li> <li>Location: San Francisco, Amsterdam</li> </ul> <h3><strong>Internship Program Details</strong></h3> <p>Our summer internship program spans over 12 weeks where you’ll have the opportunity to work with industry-leading engineers building a cloud from the ground up and possibly contribute to influential open source projects. Our internship dates are May 18th to August 7th or June 15th to September 4th.&nbsp;</p> <h3><strong>Compensation</strong></h3> <p>We offer competitive compensation, housing stipends, and other competitive benefits. The estimated US hourly rate for this role is $58-63/hr. Our hourly rates are determined by location, level and role. Individual compensation will be determined by experience, skills, and job-related knowledge.</p> <h3><strong>Equal Opportunity</strong></h3> <p>Together AI is an Equal Opportunity Employer and is proud to offer equal employment opportunity to everyone regardless of race, color, ancestry, religion, sex, national origin, sexual orientation, age, citizenship, marital status, disability, gender identity, veteran status, and more.</p> <p>Please see our privacy policy at <a href="https://www.together.ai/privacy">https://www.together.ai/privacy</a></p>
Key Skills
Machine LearningDeep LearningPythonPyTorchJAXTransformer ArchitecturesModel OptimizationReinforcement LearningPreference OptimizationNeural NetworksScientific PublicationsAlgorithmic ImprovementsDistributed TrainingEvaluation TechniquesOpen SourceData
Categories
TechnologyScience & ResearchEngineeringData & AnalyticsSoftware
Benefits
Housing StipendsCompetitive Benefits