Research Lead Training Insights jobs in San Francisco – Browse 1,392 openings on RoboApply Jobs
Research Lead Training Insights jobs in San Francisco
Open roles matching “Research Lead Training Insights” with location signals for San Francisco. 1,392 active listings on RoboApply Jobs.
1,392 jobs found
Research Lead, Training Insights
AnthropicRemote-Friendly (Travel Required) | San Francisco, CA; San Francisco, CA | New York City, NY
Hybrid Full-time
Clicking Apply Now takes you to AutoApply where you can tailor your resume and apply.
Experience Level
Manager
About the job
Join Anthropic as a Research Lead for Training Insights, where you'll spearhead innovative research initiatives that shape the future of AI training methodologies. As part of our dynamic team, you will collaborate with cross-functional experts to extract meaningful insights from training data, driving improvements in AI models. Your expertise will be vital in enhancing our understanding of AI performance and guiding strategic decisions.
Full-time|Hybrid|Remote-Friendly (Travel Required) | San Francisco, CA; San Francisco, CA | New York City, NY
Join Anthropic as a Research Lead for Training Insights, where you'll spearhead innovative research initiatives that shape the future of AI training methodologies. As part of our dynamic team, you will collaborate with cross-functional experts to extract meaningful insights from training data, driving improvements in AI models. Your expertise will be vital i…
Full-time|$176K/yr - $220K/yr|On-site|United States
Founded in 2007, Airbnb has transformed travel, connecting over 5 million hosts with more than 2 billion guests worldwide, providing unique stays and experiences that foster authentic community connections.The Community You Will JoinBecome a vital member of our innovative Research Team at Airbnb, dedicated to generating actionable insights that drive strategic decision-making. As part of the Innovation Insights division, you will be at the forefront of understanding cultural trends, competitive dynamics, key audience segments, and new growth opportunities, all while utilizing advanced AI tools and methodologies.The Impact You Will MakeAs the AI-Driven Insights Lead, you will spearhead AI-powered analyses to reveal actionable insights about global traveler behavior and trends, as well as the needs of current and potential Airbnb guests and hosts. Your collaboration with various teams—Research, Strategy, Product, Engineering, Data Science, Analytics, and Marketing—will enhance our competitive edge through pioneering AI-driven research that informs Airbnb’s innovation strategy.A Day in Your LifeCollaborate with a world-class research team to generate insights that shape business strategies.Engage with internal and cross-functional partners to pinpoint opportunities for enhancing research efficiency and business outcomes.Lead the development of innovative AI-powered methodologies, utilizing both in-house and third-party tools to create tailored insights-driven analytical pipelines.Analyze extensive structured and unstructured datasets from internal sources (e.g., guest searches and reviews), primary research (e.g., global interviews), and external channels (e.g., social media interactions).Employ Natural Language Processing (NLP) alongside advanced data analysis and statistical modeling to identify relationships, test hypotheses, and extract novel insights.Enhance the team’s use of AI as a complement to human insights and expertise.Oversee the program for identifying valuable third-party AI tools that expand our research capabilities.
Join Our Innovative TeamAt OpenAI, our Training team is at the forefront of developing advanced language models that drive our research and products, getting us closer to achieving Artificial General Intelligence (AGI). This mission demands a blend of cutting-edge research to enhance our architecture, datasets, and optimization methods, alongside strategic long-term initiatives that boost the efficiency and capabilities of future models. We ensure that our models, including recent breakthroughs like GPT-4-Turbo and GPT-4o, adhere to the highest standards of excellence.Your RoleAs an integral member of our architecture team, you will spearhead architectural advancements for OpenAI’s leading models, enhancing their intelligence and efficiency while introducing novel capabilities. Your expertise in large language model (LLM) architectures and model inference will be crucial as you adopt a hands-on, empirical approach to problem-solving. Whether brainstorming creative breakthroughs, refining foundational systems, designing evaluations, or diagnosing performance issues, your diverse skill set will be invaluable.This position is located in San Francisco, where we embrace a hybrid work environment of three days in the office each week, and we provide relocation support for new hires.Your Key Responsibilities:Innovate, prototype, and upscale new architectures to elevate model intelligence.Conduct and evaluate experiments both independently and collaboratively.Analyze, debug, and enhance both model performance and computational efficiency.Contribute to the development of training and inference infrastructure.Who You Are:You possess experience with significant contributions to major LLM training projects.You excel at independently evaluating and enhancing deep learning architectures.You are driven to responsibly implement LLMs in real-world applications.You are knowledgeable about state-of-the-art transformer modifications aimed at improving efficiency.About OpenAIOpenAI is a pioneering AI research and deployment organization committed to ensuring that artificial general intelligence benefits humanity. We focus on developing safe and effective AI technologies that empower individuals and organizations across the globe.
Join Outset as a Sales Director specializing in Research & Insights. You'll lead our sales strategies and initiatives, driving growth through informed decision-making and data-driven insights. This pivotal role requires a visionary leader who can harness market research to elevate our sales efforts, develop and implement innovative strategies, and foster strong relationships with clients.
About Our TeamThe Alignment Training team at OpenAI focuses on understanding how advanced models develop lasting behavioral patterns throughout the training process. We investigate which behaviors can be influenced during the pre-training, mid-training, and post-training phases; create the necessary data, objectives, and evaluations to guide these behaviors; and assess whether the resulting actions represent a general capability or a byproduct of the training environment.Our research encompasses synthetic data development, various training stages, model behavior analysis, and performance evaluation. We explore how models grasp user intentions, adhere to instructions, reason effectively, demonstrate honesty, and maintain reliability in novel situations. Our ultimate aim is to foster desirable behaviors early in training, reinforce them throughout, and ensure their consistency in real-world applications.About This PositionWe are seeking a seasoned researcher with profound expertise in large-scale model training, synthetic data creation, or evaluation processes, who is passionate about exploring how training decisions influence aligned behaviors in state-of-the-art models.In this role, you will define the research agenda for alignment training: outlining the behaviors we aspire for models to acquire, designing data and training strategies to cultivate them, and developing evaluation mechanisms to verify the breadth, strength, and durability of those behaviors. The ideal candidate will excel at translating vague behavioral inquiries into structured experimental plans: devising hypotheses, creating interventions, establishing pipelines, conducting experiments, and analyzing results for authenticity.This position is particularly suited for individuals eager to engage closely with the core model training framework, where decisions regarding data, objectives, and evaluations critically influence the alignment of deployed systems.Key Responsibilities:Innovate synthetic data methods that instill higher-level behavioral tendencies in models, such as comprehending user intent, consistently following instructions, clear reasoning, honesty, and alignment with defined goals and constraints.Analyze the impact of pre-training, mid-training, and post-training on subsequent model behavior, identifying the most effective interventions for each phase.Develop evaluation loops that link model behavior back to training data and objectives, enabling quicker iterations and clearer feedback.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We strive to build a future where everyone has access to the knowledge and tools essential for making AI work effectively for their unique objectives.Our team comprises scientists, engineers, and innovators who have contributed to some of the most widely adopted AI products, including ChatGPT and Character.ai, as well as notable open-weight models like Mistral and popular open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe Post-Training Researcher position is pivotal to our roadmap. It serves as a crucial connection between raw model intelligence and a system that is genuinely beneficial, safe, and collaborative for human users.This role uniquely combines fundamental research with practical engineering, as we do not differentiate between these functions internally. Candidates will be expected to produce high-performance code and analyze technical reports. This position is ideal for individuals who relish both deep theoretical inquiry and hands-on experimentation, aiming to influence the foundational aspects of AI learning.Note: This position is classified as an 'evergreen role', meaning we continuously accept applications in this research domain. Given the high volume of applications, an immediate match for your skills and experience may not always be available. However, we encourage you to apply; we regularly review submissions and reach out as new opportunities arise. You are welcome to apply again after gaining more experience, but we ask that you refrain from applying more than once every six months. Additionally, specific postings for singular roles may be available for distinct projects or team needs, in which case you are welcome to apply directly in conjunction with this evergreen role.What You’ll DoDevelop and Optimize Recipes: Refine post-training recipes, encompassing various datasets, training stages, and hyperparameters, while assessing their impact on multiple performance metrics.Iterate on Evaluations: Engage in a continuous process of defining evaluation metrics, optimizing them, and recognizing their limitations. You will be accountable for enhancing performance metrics and ensuring they are meaningful.Debug and Analyze: During the fine-tuning of training configurations, you may encounter results that appear inconsistent. You will be responsible for troubleshooting and cultivating a deeper understanding to apply to subsequent challenges.Scale and Investigate: Assess and expand the capabilities of our models while exploring potential improvements.
abundant seeks a Research Lead based in San Francisco. This position steers research activities that help shape the company’s direction. The Research Lead partners with colleagues to analyze data, draw meaningful insights, and support projects where research has a clear business impact. Key responsibilities Plan, manage, and execute research initiatives from start to finish Work with team members to analyze data and spot important trends Turn research results into practical recommendations for the business Support projects that guide company strategy Collaboration and impact This role involves close teamwork and communication across departments. Research findings directly inform business decisions and contribute to the company’s ongoing growth.
Role overview OpenAI is looking for a Researcher focused on Agentic Post-Training, based in San Francisco. This role centers on analyzing and improving how AI systems behave after their initial training. The goal is to broaden the capabilities of AI and refine how models respond in complex situations. What you will do Study and assess agentic behaviors in trained AI models Create new approaches to strengthen these behaviors after training Collaborate with a talented team on projects that shape the future of artificial intelligence research Collaboration and impact This position involves hands-on research with other specialists at OpenAI. The work directly supports the advancement of AI capabilities and helps define new benchmarks for agentic performance in artificial intelligence.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We envision a future where everyone can harness the knowledge and tools necessary for AI to serve their unique needs and aspirations. Our team comprises scientists, engineers, and builders who have developed some of the most widely utilized AI products, such as ChatGPT and Character.ai, as well as open-weight models like Mistral and popular open-source projects including PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe role of a Post-Training Researcher is pivotal to our strategic vision. This position serves as the essential link between raw model intelligence and a practical, safe, and collaborative system for human users.Our research in post-training data sits at the intersection of human insights and machine learning. By integrating human and synthetic data techniques alongside innovative methodologies, we capture the subtleties of human behavior to inform and guide our models. We investigate and model the mechanisms that derive value for individuals, enabling us to articulate, predict, and enhance human preferences, behaviors, and satisfaction. Our objective is to translate research concepts into actionable data through meticulously planned data labeling and collection initiatives, while also understanding the science behind high-quality data that effectively trains our models. Additionally, we develop and assess quantitative metrics to evaluate the success and impact of our data and training strategies.Beyond execution, we explore new paradigms for human-AI interaction and scalable oversight, experimenting with optimal ways for humans to supervise, guide, and collaborate with models. This interdisciplinary role merges research, data operations, and technical implementation, pushing the boundaries of aligned, human-centered AI systems.This position combines foundational research and practical engineering, as we do not differentiate between these roles internally. You will be expected to write high-performance code and comprehend technical reports. This role is perfect for individuals who thrive on deep theoretical exploration and hands-on experimentation, eager to shape the foundational aspects of AI learning.Note: This is an evergreen role that we maintain continuously to express interest in this research area. We receive a high volume of applications, and while there may not always be an immediate fit for your skills and experience, we encourage you to apply. We regularly review applications and reach out to candidates as new opportunities arise. You are welcome to reapply after gaining more experience, but please limit applications to once every six months. You may also notice postings for specific roles for targeted positions.
OpenAI is hiring a Software Engineer for Post-Training Research in San Francisco. This position centers on improving the performance and capabilities of advanced machine learning models after their initial training phase. Role overview Work closely with a skilled team to explore new ways of strengthening AI systems. The focus is on researching and developing methods that push the boundaries of what these models can achieve once training is complete. Collaboration Expect to contribute to ongoing research efforts and share insights with colleagues who are passionate about advancing AI. Teamwork and knowledge exchange are key parts of this role. Location This position is based in San Francisco.
Full-time|$250K/yr - $450K/yr|On-site|San Francisco
About AfterQuery AfterQuery builds training data and evaluation frameworks used by leading AI labs around the world. The team partners with advanced research groups to create high-quality datasets and run detailed evaluations that go beyond standard benchmarks. As a small, post-Series A company based in San Francisco, every team member plays a key role in shaping how future AI models learn and improve. Role Overview The Post-Training Research Scientist focuses on proving the impact of AfterQuery's datasets. This work involves designing and running training experiments to isolate how specific data influences model performance. Projects span Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) post-training, with an emphasis on measuring effects on capability, generalization, and alignment. Working closely with partner labs, the scientist turns data into clear, verifiable results: showing exactly how a dataset leads to measurable improvements under defined conditions. The work is experimental and directly shapes the value of AfterQuery's products. What You Will Do Run controlled SFT and RL experiments to measure how datasets affect model outcomes. Quantify gains in areas like reasoning, tool use, long-horizon tasks, and specialized workflows. Share findings with partner labs to support sales and demonstrate value. Work with internal subject matter experts to improve data quality based on experimental results. What We Look For Strong background in LLM training and evaluation methods. Curiosity about how data structure, selection, and quality shape model behavior. Skill in designing experiments, executing quickly, and drawing practical insights from complex results. Comfort working across fields such as finance, software engineering, and policy. Focus on real-world implementation, not just theory. Research experience at the undergraduate or master's level is preferred; a PhD is not required. Compensation $250,000 - $450,000 total compensation plus equity
Full-time|$71K/yr - $125K/yr|Hybrid|Atlanta, Georgia, United States; Austin, Texas, United States; Boston, Massachusetts, United States; Charlotte, North Carolina, United States; Chicago, Illinois, United States; Los Angeles, California, United States; New York, New York, United States; San Francisco, California, United States; St. Louis, Missouri, United States; Washington, District of Columbia, United States
Overview FleishmanHillard, a leading global integrated public relations agency, is on the lookout for an innovative Research Director to join our True Global Intelligence Team. This hybrid role can be based in one of our multiple U.S. locations, allowing for a blend of in-office and remote work. The True Global Intelligence Practice at FleishmanHillard employs a fusion of primary and secondary research methodologies, combined with communications measurement and data analytics. We are at the forefront of integrating artificial intelligence tools and techniques across our diverse service offerings. We seek a Research Director to spearhead the expansion of our AI capabilities in an ever-evolving landscape where traditional research methods intersect with groundbreaking technologies. This role requires a deep understanding of established research practices and a proven track record in incorporating AI into research projects, programs, and client engagements. The ideal candidate will lead strategic research initiatives, positioning FleishmanHillard at the forefront of methodical research and innovative AI applications. In this position, you will help drive analytics and manage client relations across multiple accounts. Your focus will be on optimizing research programs through the integration and development of AI methodologies that maintain our research's rigor while addressing our clients' needs. This position is ideal for individuals who are already experimenting with AI across various research and insights applications, enjoy customizing prompts, and thrive in a dynamic, technology-driven environment. FleishmanHillard, headquartered in St. Louis, is among the largest integrated communications firms worldwide. Our client services leverage expertise across more than 25 disciplines, including B2C and B2B marketing, corporate reputation management, CSR, creative services, digital and social media, and technology. We believe that a diverse team enriches our perspectives and enhances our ability to serve clients effectively. We are dedicated to fostering impact and inclusion within our organization and the communities we serve. Our commitment to building a diverse workforce is unwavering, and we encourage candidates who share our passion for enhancing our impact and inclusion objectives to apply.
OverviewPluralis Research is at the forefront of Protocol Learning, innovating a decentralized approach to train and deploy AI models that democratizes access beyond just well-funded corporations. By aggregating computational resources from diverse participants, we incentivize collaboration while safeguarding against centralized control of model weights, paving the way for a truly open and cooperative environment for advanced AI.We are seeking a talented Machine Learning Training Platform Engineer to design, develop, and scale the core infrastructure that powers our decentralized ML training platform. In this role, you will have ownership over essential systems including infrastructure orchestration, distributed computing, and service integration, facilitating ongoing experimentation and large-scale model training.ResponsibilitiesMulti-Cloud Infrastructure: Create resource management systems that provision and orchestrate computing resources across AWS, GCP, and Azure using infrastructure-as-code tools like Pulumi or Terraform. Manage dynamic scaling, state synchronization, and concurrent operations across hundreds of diverse nodes.Distributed Training Systems: Design fault-tolerant infrastructure for distributed machine learning, including GPU clusters, NVIDIA runtime, S3 checkpointing, large dataset management and streaming, health monitoring, and resilient retry strategies.Real-World Networking: Develop systems that simulate and manage real-world network conditions—such as bandwidth shaping, latency injection, and packet loss—while accommodating dynamic node churn and ensuring efficient data flow across workers with varying connectivity, as our training occurs on consumer nodes and non-co-located infrastructure.
Join aiedu as a Senior Lead in Research & Evaluation, where you will drive impactful research initiatives that shape educational practices and policies. In this role, you will lead a team of researchers in designing and executing comprehensive evaluations that inform our strategic direction. Your expertise will be critical in analyzing data, generating insights, and communicating findings to stakeholders.
Full-time|On-site|San Francisco Bay Area (San Mateo) or Boston (Somerville)
About the RoleIn the realm of machine learning, pretraining lays the foundation for a general model, while post-training refines that model, enhancing its utility, controllability, safety, and performance in real-world applications. As a Post-Training Research Scientist, you will transform large pretrained robot models into production-ready systems through methodologies such as fine-tuning, reinforcement learning, steering, human feedback, task specialization, evaluation, and on-robot validation at scale. This position offers a unique opportunity for individuals from diverse backgrounds to evolve into full-stack ML roboticists, adept at swiftly identifying challenges across machine learning and control domains. This is where innovative research converges with practical implementation.Your Responsibilities Include:Crafting fine-tuning and adaptation strategies tailored for specific robotic tasks and embodiments.Developing methodologies to enhance reliability, robustness, and controllability of robotic systems.Establishing evaluation frameworks to assess real-world robot performance beyond just offline metrics.Collaborating with ML infrastructure teams to optimize inference-time performance, including latency, stability, and memory usage.Utilizing advanced techniques such as imitation learning, reinforcement learning, distillation, synthetic data, and curriculum learning.Bridging the gap between model outputs and tangible outcomes in the physical world.You Might Excel in This Role If You:Possess experience in fine-tuning large models for downstream applications, including RLHF, imitation learning, reinforcement learning, distillation, and domain adaptation.Have a background in embodied AI, robotics, or real-world machine learning systems.Demonstrate a strong commitment to evaluation, benchmarking, and failure analysis.Are comfortable troubleshooting and debugging across the entire ML stack, from analyzing loss curves to understanding robot behavior.Enjoy rapid iteration and thrive on real-world feedback loops.Aspire to connect foundational models with practical deployment scenarios.About GeneralistAt Generalist, we are dedicated to realizing the vision of general-purpose robots. We envision a future where industries and homes benefit from collaborative interactions between humans and machines, enabling us to achieve more than ever before. Our focus is on building embodied foundation models, starting with dexterity, and advancing the frontiers of data, models, and hardware to empower robots to intelligently engage with their environments.
Full-time|$252K/yr - $315K/yr|On-site|San Francisco, CA; Seattle, WA; New York, NY
At Scale AI, we collaborate with leading AI laboratories to supply high-quality data and foster advancements in Generative AI research. We seek innovative Research Scientists and Research Engineers with a strong focus on post-training techniques for Large Language Models (LLMs), including Supervised Fine-Tuning (SFT), Reinforcement Learning from Human Feedback (RLHF), and reward modeling. This position emphasizes optimizing data curation and evaluation processes to boost LLM performance across text and multimodal formats. In this pivotal role, you will pioneer new methods to enhance the alignment and generalization of extensive generative models. You will work closely with fellow researchers and engineers to establish best practices in data-driven AI development. Additionally, you will collaborate with top foundation model labs, providing critical technical and strategic insights for the evolution of next-generation generative AI models.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, we are dedicated to empowering humanity through the advancement of collaborative general intelligence. Our vision is to create a future where everyone can harness the power of AI to meet their individual needs and aspirations.Our team is composed of passionate scientists, engineers, and innovators who have developed some of the most influential AI technologies, such as ChatGPT and Character.ai, as well as cutting-edge open-weight models like Mistral and acclaimed open-source projects including PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe role of Pre-Training Researcher is pivotal to our strategic roadmap, focused on enhancing our understanding of how large models learn from data. You will investigate novel pre-training methodologies, architectures, and learning objectives aimed at making model training more efficient, robust, and aligned with human values.This position combines fundamental research with practical engineering, as we seamlessly integrate both disciplines within our team. You will be expected to produce high-performance code and engage with technical literature. This is an ideal opportunity for individuals who thrive on theoretical exploration as well as hands-on experimentation, and who aspire to influence the foundational methods by which AI learns.This is an evergreen role, meaning we keep this position open to welcome expressions of interest in this research field. We receive numerous applications, and while there may not always be an immediate fit, we encourage you to apply. We consistently review applications and will reach out as new opportunities arise. If you gain additional experience, you are welcome to reapply, but please limit your applications to once every six months. We may also post specific openings for project or team needs, where direct applications are welcome in addition to this evergreen role.What You’ll DoResearch and innovate new methodologies for pre-training.Engage in areas such as scaling, architecture, algorithms, or optimization of large-scale training runs based on your research interests and expertise.Design data curricula and sampling strategies that enhance learning dynamics and model generalization.Collaborate with infrastructure and data teams to conduct large-scale experiments in an efficient and reproducible manner.Publish and present research that propels the entire community forward, sharing code, datasets, and insights to accelerate progress across both industry and academia.
Role Overview Lyft is hiring a Lead Analyst for Market Insights in San Francisco, CA. This role centers on using analytics to identify trends and deliver insights that shape Lyft’s strategic direction. The Lead Analyst works closely with teams across the company to deepen understanding of market shifts and customer patterns, supporting Lyft’s position in the rideshare sector.
Full-time|$252K/yr - $315K/yr|On-site|San Francisco, CA; New York, NY
Join Scale's innovative Large Language Model (LLM) post-training platform team, where you will contribute to the development of our internal distributed framework designed specifically for LLM training. This sophisticated platform empowers Machine Learning Engineers (MLEs), researchers, data scientists, and operators to perform rapid and automated training and evaluation of LLMs. Additionally, it underpins the training framework for our data quality evaluation pipeline.Scale is at the forefront of the Artificial Intelligence sector, acting as a vital provider of training and evaluation data, as well as comprehensive solutions for the entire machine learning lifecycle. In this role, you will collaborate closely with Scale’s ML teams and researchers to construct the foundational platform that supports all our ML research and development initiatives. Your work will involve building and optimizing this platform to facilitate the training, inference, and data curation of next-generation LLMs.If you are passionate about driving the future of AI through groundbreaking innovations, we invite you to connect with us!
Join Baseten as a Post-Training Research Engineer and contribute to groundbreaking advancements in machine learning and AI. In this role, you will leverage your engineering skills to analyze and enhance models post-training, ensuring optimal performance and efficiency.
Full-time|Hybrid|Remote-Friendly (Travel Required) | San Francisco, CA; San Francisco, CA | New York City, NY
Join Anthropic as a Research Lead for Training Insights, where you'll spearhead innovative research initiatives that shape the future of AI training methodologies. As part of our dynamic team, you will collaborate with cross-functional experts to extract meaningful insights from training data, driving improvements in AI models. Your expertise will be vital i…
Full-time|$176K/yr - $220K/yr|On-site|United States
Founded in 2007, Airbnb has transformed travel, connecting over 5 million hosts with more than 2 billion guests worldwide, providing unique stays and experiences that foster authentic community connections.The Community You Will JoinBecome a vital member of our innovative Research Team at Airbnb, dedicated to generating actionable insights that drive strategic decision-making. As part of the Innovation Insights division, you will be at the forefront of understanding cultural trends, competitive dynamics, key audience segments, and new growth opportunities, all while utilizing advanced AI tools and methodologies.The Impact You Will MakeAs the AI-Driven Insights Lead, you will spearhead AI-powered analyses to reveal actionable insights about global traveler behavior and trends, as well as the needs of current and potential Airbnb guests and hosts. Your collaboration with various teams—Research, Strategy, Product, Engineering, Data Science, Analytics, and Marketing—will enhance our competitive edge through pioneering AI-driven research that informs Airbnb’s innovation strategy.A Day in Your LifeCollaborate with a world-class research team to generate insights that shape business strategies.Engage with internal and cross-functional partners to pinpoint opportunities for enhancing research efficiency and business outcomes.Lead the development of innovative AI-powered methodologies, utilizing both in-house and third-party tools to create tailored insights-driven analytical pipelines.Analyze extensive structured and unstructured datasets from internal sources (e.g., guest searches and reviews), primary research (e.g., global interviews), and external channels (e.g., social media interactions).Employ Natural Language Processing (NLP) alongside advanced data analysis and statistical modeling to identify relationships, test hypotheses, and extract novel insights.Enhance the team’s use of AI as a complement to human insights and expertise.Oversee the program for identifying valuable third-party AI tools that expand our research capabilities.
Join Our Innovative TeamAt OpenAI, our Training team is at the forefront of developing advanced language models that drive our research and products, getting us closer to achieving Artificial General Intelligence (AGI). This mission demands a blend of cutting-edge research to enhance our architecture, datasets, and optimization methods, alongside strategic long-term initiatives that boost the efficiency and capabilities of future models. We ensure that our models, including recent breakthroughs like GPT-4-Turbo and GPT-4o, adhere to the highest standards of excellence.Your RoleAs an integral member of our architecture team, you will spearhead architectural advancements for OpenAI’s leading models, enhancing their intelligence and efficiency while introducing novel capabilities. Your expertise in large language model (LLM) architectures and model inference will be crucial as you adopt a hands-on, empirical approach to problem-solving. Whether brainstorming creative breakthroughs, refining foundational systems, designing evaluations, or diagnosing performance issues, your diverse skill set will be invaluable.This position is located in San Francisco, where we embrace a hybrid work environment of three days in the office each week, and we provide relocation support for new hires.Your Key Responsibilities:Innovate, prototype, and upscale new architectures to elevate model intelligence.Conduct and evaluate experiments both independently and collaboratively.Analyze, debug, and enhance both model performance and computational efficiency.Contribute to the development of training and inference infrastructure.Who You Are:You possess experience with significant contributions to major LLM training projects.You excel at independently evaluating and enhancing deep learning architectures.You are driven to responsibly implement LLMs in real-world applications.You are knowledgeable about state-of-the-art transformer modifications aimed at improving efficiency.About OpenAIOpenAI is a pioneering AI research and deployment organization committed to ensuring that artificial general intelligence benefits humanity. We focus on developing safe and effective AI technologies that empower individuals and organizations across the globe.
Join Outset as a Sales Director specializing in Research & Insights. You'll lead our sales strategies and initiatives, driving growth through informed decision-making and data-driven insights. This pivotal role requires a visionary leader who can harness market research to elevate our sales efforts, develop and implement innovative strategies, and foster strong relationships with clients.
About Our TeamThe Alignment Training team at OpenAI focuses on understanding how advanced models develop lasting behavioral patterns throughout the training process. We investigate which behaviors can be influenced during the pre-training, mid-training, and post-training phases; create the necessary data, objectives, and evaluations to guide these behaviors; and assess whether the resulting actions represent a general capability or a byproduct of the training environment.Our research encompasses synthetic data development, various training stages, model behavior analysis, and performance evaluation. We explore how models grasp user intentions, adhere to instructions, reason effectively, demonstrate honesty, and maintain reliability in novel situations. Our ultimate aim is to foster desirable behaviors early in training, reinforce them throughout, and ensure their consistency in real-world applications.About This PositionWe are seeking a seasoned researcher with profound expertise in large-scale model training, synthetic data creation, or evaluation processes, who is passionate about exploring how training decisions influence aligned behaviors in state-of-the-art models.In this role, you will define the research agenda for alignment training: outlining the behaviors we aspire for models to acquire, designing data and training strategies to cultivate them, and developing evaluation mechanisms to verify the breadth, strength, and durability of those behaviors. The ideal candidate will excel at translating vague behavioral inquiries into structured experimental plans: devising hypotheses, creating interventions, establishing pipelines, conducting experiments, and analyzing results for authenticity.This position is particularly suited for individuals eager to engage closely with the core model training framework, where decisions regarding data, objectives, and evaluations critically influence the alignment of deployed systems.Key Responsibilities:Innovate synthetic data methods that instill higher-level behavioral tendencies in models, such as comprehending user intent, consistently following instructions, clear reasoning, honesty, and alignment with defined goals and constraints.Analyze the impact of pre-training, mid-training, and post-training on subsequent model behavior, identifying the most effective interventions for each phase.Develop evaluation loops that link model behavior back to training data and objectives, enabling quicker iterations and clearer feedback.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We strive to build a future where everyone has access to the knowledge and tools essential for making AI work effectively for their unique objectives.Our team comprises scientists, engineers, and innovators who have contributed to some of the most widely adopted AI products, including ChatGPT and Character.ai, as well as notable open-weight models like Mistral and popular open-source projects such as PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe Post-Training Researcher position is pivotal to our roadmap. It serves as a crucial connection between raw model intelligence and a system that is genuinely beneficial, safe, and collaborative for human users.This role uniquely combines fundamental research with practical engineering, as we do not differentiate between these functions internally. Candidates will be expected to produce high-performance code and analyze technical reports. This position is ideal for individuals who relish both deep theoretical inquiry and hands-on experimentation, aiming to influence the foundational aspects of AI learning.Note: This position is classified as an 'evergreen role', meaning we continuously accept applications in this research domain. Given the high volume of applications, an immediate match for your skills and experience may not always be available. However, we encourage you to apply; we regularly review submissions and reach out as new opportunities arise. You are welcome to apply again after gaining more experience, but we ask that you refrain from applying more than once every six months. Additionally, specific postings for singular roles may be available for distinct projects or team needs, in which case you are welcome to apply directly in conjunction with this evergreen role.What You’ll DoDevelop and Optimize Recipes: Refine post-training recipes, encompassing various datasets, training stages, and hyperparameters, while assessing their impact on multiple performance metrics.Iterate on Evaluations: Engage in a continuous process of defining evaluation metrics, optimizing them, and recognizing their limitations. You will be accountable for enhancing performance metrics and ensuring they are meaningful.Debug and Analyze: During the fine-tuning of training configurations, you may encounter results that appear inconsistent. You will be responsible for troubleshooting and cultivating a deeper understanding to apply to subsequent challenges.Scale and Investigate: Assess and expand the capabilities of our models while exploring potential improvements.
abundant seeks a Research Lead based in San Francisco. This position steers research activities that help shape the company’s direction. The Research Lead partners with colleagues to analyze data, draw meaningful insights, and support projects where research has a clear business impact. Key responsibilities Plan, manage, and execute research initiatives from start to finish Work with team members to analyze data and spot important trends Turn research results into practical recommendations for the business Support projects that guide company strategy Collaboration and impact This role involves close teamwork and communication across departments. Research findings directly inform business decisions and contribute to the company’s ongoing growth.
Role overview OpenAI is looking for a Researcher focused on Agentic Post-Training, based in San Francisco. This role centers on analyzing and improving how AI systems behave after their initial training. The goal is to broaden the capabilities of AI and refine how models respond in complex situations. What you will do Study and assess agentic behaviors in trained AI models Create new approaches to strengthen these behaviors after training Collaborate with a talented team on projects that shape the future of artificial intelligence research Collaboration and impact This position involves hands-on research with other specialists at OpenAI. The work directly supports the advancement of AI capabilities and helps define new benchmarks for agentic performance in artificial intelligence.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, our mission is to empower humanity by advancing collaborative general intelligence. We envision a future where everyone can harness the knowledge and tools necessary for AI to serve their unique needs and aspirations. Our team comprises scientists, engineers, and builders who have developed some of the most widely utilized AI products, such as ChatGPT and Character.ai, as well as open-weight models like Mistral and popular open-source projects including PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe role of a Post-Training Researcher is pivotal to our strategic vision. This position serves as the essential link between raw model intelligence and a practical, safe, and collaborative system for human users.Our research in post-training data sits at the intersection of human insights and machine learning. By integrating human and synthetic data techniques alongside innovative methodologies, we capture the subtleties of human behavior to inform and guide our models. We investigate and model the mechanisms that derive value for individuals, enabling us to articulate, predict, and enhance human preferences, behaviors, and satisfaction. Our objective is to translate research concepts into actionable data through meticulously planned data labeling and collection initiatives, while also understanding the science behind high-quality data that effectively trains our models. Additionally, we develop and assess quantitative metrics to evaluate the success and impact of our data and training strategies.Beyond execution, we explore new paradigms for human-AI interaction and scalable oversight, experimenting with optimal ways for humans to supervise, guide, and collaborate with models. This interdisciplinary role merges research, data operations, and technical implementation, pushing the boundaries of aligned, human-centered AI systems.This position combines foundational research and practical engineering, as we do not differentiate between these roles internally. You will be expected to write high-performance code and comprehend technical reports. This role is perfect for individuals who thrive on deep theoretical exploration and hands-on experimentation, eager to shape the foundational aspects of AI learning.Note: This is an evergreen role that we maintain continuously to express interest in this research area. We receive a high volume of applications, and while there may not always be an immediate fit for your skills and experience, we encourage you to apply. We regularly review applications and reach out to candidates as new opportunities arise. You are welcome to reapply after gaining more experience, but please limit applications to once every six months. You may also notice postings for specific roles for targeted positions.
OpenAI is hiring a Software Engineer for Post-Training Research in San Francisco. This position centers on improving the performance and capabilities of advanced machine learning models after their initial training phase. Role overview Work closely with a skilled team to explore new ways of strengthening AI systems. The focus is on researching and developing methods that push the boundaries of what these models can achieve once training is complete. Collaboration Expect to contribute to ongoing research efforts and share insights with colleagues who are passionate about advancing AI. Teamwork and knowledge exchange are key parts of this role. Location This position is based in San Francisco.
Full-time|$250K/yr - $450K/yr|On-site|San Francisco
About AfterQuery AfterQuery builds training data and evaluation frameworks used by leading AI labs around the world. The team partners with advanced research groups to create high-quality datasets and run detailed evaluations that go beyond standard benchmarks. As a small, post-Series A company based in San Francisco, every team member plays a key role in shaping how future AI models learn and improve. Role Overview The Post-Training Research Scientist focuses on proving the impact of AfterQuery's datasets. This work involves designing and running training experiments to isolate how specific data influences model performance. Projects span Supervised Fine-Tuning (SFT) and Reinforcement Learning (RL) post-training, with an emphasis on measuring effects on capability, generalization, and alignment. Working closely with partner labs, the scientist turns data into clear, verifiable results: showing exactly how a dataset leads to measurable improvements under defined conditions. The work is experimental and directly shapes the value of AfterQuery's products. What You Will Do Run controlled SFT and RL experiments to measure how datasets affect model outcomes. Quantify gains in areas like reasoning, tool use, long-horizon tasks, and specialized workflows. Share findings with partner labs to support sales and demonstrate value. Work with internal subject matter experts to improve data quality based on experimental results. What We Look For Strong background in LLM training and evaluation methods. Curiosity about how data structure, selection, and quality shape model behavior. Skill in designing experiments, executing quickly, and drawing practical insights from complex results. Comfort working across fields such as finance, software engineering, and policy. Focus on real-world implementation, not just theory. Research experience at the undergraduate or master's level is preferred; a PhD is not required. Compensation $250,000 - $450,000 total compensation plus equity
Full-time|$71K/yr - $125K/yr|Hybrid|Atlanta, Georgia, United States; Austin, Texas, United States; Boston, Massachusetts, United States; Charlotte, North Carolina, United States; Chicago, Illinois, United States; Los Angeles, California, United States; New York, New York, United States; San Francisco, California, United States; St. Louis, Missouri, United States; Washington, District of Columbia, United States
Overview FleishmanHillard, a leading global integrated public relations agency, is on the lookout for an innovative Research Director to join our True Global Intelligence Team. This hybrid role can be based in one of our multiple U.S. locations, allowing for a blend of in-office and remote work. The True Global Intelligence Practice at FleishmanHillard employs a fusion of primary and secondary research methodologies, combined with communications measurement and data analytics. We are at the forefront of integrating artificial intelligence tools and techniques across our diverse service offerings. We seek a Research Director to spearhead the expansion of our AI capabilities in an ever-evolving landscape where traditional research methods intersect with groundbreaking technologies. This role requires a deep understanding of established research practices and a proven track record in incorporating AI into research projects, programs, and client engagements. The ideal candidate will lead strategic research initiatives, positioning FleishmanHillard at the forefront of methodical research and innovative AI applications. In this position, you will help drive analytics and manage client relations across multiple accounts. Your focus will be on optimizing research programs through the integration and development of AI methodologies that maintain our research's rigor while addressing our clients' needs. This position is ideal for individuals who are already experimenting with AI across various research and insights applications, enjoy customizing prompts, and thrive in a dynamic, technology-driven environment. FleishmanHillard, headquartered in St. Louis, is among the largest integrated communications firms worldwide. Our client services leverage expertise across more than 25 disciplines, including B2C and B2B marketing, corporate reputation management, CSR, creative services, digital and social media, and technology. We believe that a diverse team enriches our perspectives and enhances our ability to serve clients effectively. We are dedicated to fostering impact and inclusion within our organization and the communities we serve. Our commitment to building a diverse workforce is unwavering, and we encourage candidates who share our passion for enhancing our impact and inclusion objectives to apply.
OverviewPluralis Research is at the forefront of Protocol Learning, innovating a decentralized approach to train and deploy AI models that democratizes access beyond just well-funded corporations. By aggregating computational resources from diverse participants, we incentivize collaboration while safeguarding against centralized control of model weights, paving the way for a truly open and cooperative environment for advanced AI.We are seeking a talented Machine Learning Training Platform Engineer to design, develop, and scale the core infrastructure that powers our decentralized ML training platform. In this role, you will have ownership over essential systems including infrastructure orchestration, distributed computing, and service integration, facilitating ongoing experimentation and large-scale model training.ResponsibilitiesMulti-Cloud Infrastructure: Create resource management systems that provision and orchestrate computing resources across AWS, GCP, and Azure using infrastructure-as-code tools like Pulumi or Terraform. Manage dynamic scaling, state synchronization, and concurrent operations across hundreds of diverse nodes.Distributed Training Systems: Design fault-tolerant infrastructure for distributed machine learning, including GPU clusters, NVIDIA runtime, S3 checkpointing, large dataset management and streaming, health monitoring, and resilient retry strategies.Real-World Networking: Develop systems that simulate and manage real-world network conditions—such as bandwidth shaping, latency injection, and packet loss—while accommodating dynamic node churn and ensuring efficient data flow across workers with varying connectivity, as our training occurs on consumer nodes and non-co-located infrastructure.
Join aiedu as a Senior Lead in Research & Evaluation, where you will drive impactful research initiatives that shape educational practices and policies. In this role, you will lead a team of researchers in designing and executing comprehensive evaluations that inform our strategic direction. Your expertise will be critical in analyzing data, generating insights, and communicating findings to stakeholders.
Full-time|On-site|San Francisco Bay Area (San Mateo) or Boston (Somerville)
About the RoleIn the realm of machine learning, pretraining lays the foundation for a general model, while post-training refines that model, enhancing its utility, controllability, safety, and performance in real-world applications. As a Post-Training Research Scientist, you will transform large pretrained robot models into production-ready systems through methodologies such as fine-tuning, reinforcement learning, steering, human feedback, task specialization, evaluation, and on-robot validation at scale. This position offers a unique opportunity for individuals from diverse backgrounds to evolve into full-stack ML roboticists, adept at swiftly identifying challenges across machine learning and control domains. This is where innovative research converges with practical implementation.Your Responsibilities Include:Crafting fine-tuning and adaptation strategies tailored for specific robotic tasks and embodiments.Developing methodologies to enhance reliability, robustness, and controllability of robotic systems.Establishing evaluation frameworks to assess real-world robot performance beyond just offline metrics.Collaborating with ML infrastructure teams to optimize inference-time performance, including latency, stability, and memory usage.Utilizing advanced techniques such as imitation learning, reinforcement learning, distillation, synthetic data, and curriculum learning.Bridging the gap between model outputs and tangible outcomes in the physical world.You Might Excel in This Role If You:Possess experience in fine-tuning large models for downstream applications, including RLHF, imitation learning, reinforcement learning, distillation, and domain adaptation.Have a background in embodied AI, robotics, or real-world machine learning systems.Demonstrate a strong commitment to evaluation, benchmarking, and failure analysis.Are comfortable troubleshooting and debugging across the entire ML stack, from analyzing loss curves to understanding robot behavior.Enjoy rapid iteration and thrive on real-world feedback loops.Aspire to connect foundational models with practical deployment scenarios.About GeneralistAt Generalist, we are dedicated to realizing the vision of general-purpose robots. We envision a future where industries and homes benefit from collaborative interactions between humans and machines, enabling us to achieve more than ever before. Our focus is on building embodied foundation models, starting with dexterity, and advancing the frontiers of data, models, and hardware to empower robots to intelligently engage with their environments.
Full-time|$252K/yr - $315K/yr|On-site|San Francisco, CA; Seattle, WA; New York, NY
At Scale AI, we collaborate with leading AI laboratories to supply high-quality data and foster advancements in Generative AI research. We seek innovative Research Scientists and Research Engineers with a strong focus on post-training techniques for Large Language Models (LLMs), including Supervised Fine-Tuning (SFT), Reinforcement Learning from Human Feedback (RLHF), and reward modeling. This position emphasizes optimizing data curation and evaluation processes to boost LLM performance across text and multimodal formats. In this pivotal role, you will pioneer new methods to enhance the alignment and generalization of extensive generative models. You will work closely with fellow researchers and engineers to establish best practices in data-driven AI development. Additionally, you will collaborate with top foundation model labs, providing critical technical and strategic insights for the evolution of next-generation generative AI models.
Full-time|$350K/yr - $475K/yr|On-site|San Francisco
At Thinking Machines Lab, we are dedicated to empowering humanity through the advancement of collaborative general intelligence. Our vision is to create a future where everyone can harness the power of AI to meet their individual needs and aspirations.Our team is composed of passionate scientists, engineers, and innovators who have developed some of the most influential AI technologies, such as ChatGPT and Character.ai, as well as cutting-edge open-weight models like Mistral and acclaimed open-source projects including PyTorch, OpenAI Gym, Fairseq, and Segment Anything.About the RoleThe role of Pre-Training Researcher is pivotal to our strategic roadmap, focused on enhancing our understanding of how large models learn from data. You will investigate novel pre-training methodologies, architectures, and learning objectives aimed at making model training more efficient, robust, and aligned with human values.This position combines fundamental research with practical engineering, as we seamlessly integrate both disciplines within our team. You will be expected to produce high-performance code and engage with technical literature. This is an ideal opportunity for individuals who thrive on theoretical exploration as well as hands-on experimentation, and who aspire to influence the foundational methods by which AI learns.This is an evergreen role, meaning we keep this position open to welcome expressions of interest in this research field. We receive numerous applications, and while there may not always be an immediate fit, we encourage you to apply. We consistently review applications and will reach out as new opportunities arise. If you gain additional experience, you are welcome to reapply, but please limit your applications to once every six months. We may also post specific openings for project or team needs, where direct applications are welcome in addition to this evergreen role.What You’ll DoResearch and innovate new methodologies for pre-training.Engage in areas such as scaling, architecture, algorithms, or optimization of large-scale training runs based on your research interests and expertise.Design data curricula and sampling strategies that enhance learning dynamics and model generalization.Collaborate with infrastructure and data teams to conduct large-scale experiments in an efficient and reproducible manner.Publish and present research that propels the entire community forward, sharing code, datasets, and insights to accelerate progress across both industry and academia.
Role Overview Lyft is hiring a Lead Analyst for Market Insights in San Francisco, CA. This role centers on using analytics to identify trends and deliver insights that shape Lyft’s strategic direction. The Lead Analyst works closely with teams across the company to deepen understanding of market shifts and customer patterns, supporting Lyft’s position in the rideshare sector.
Full-time|$252K/yr - $315K/yr|On-site|San Francisco, CA; New York, NY
Join Scale's innovative Large Language Model (LLM) post-training platform team, where you will contribute to the development of our internal distributed framework designed specifically for LLM training. This sophisticated platform empowers Machine Learning Engineers (MLEs), researchers, data scientists, and operators to perform rapid and automated training and evaluation of LLMs. Additionally, it underpins the training framework for our data quality evaluation pipeline.Scale is at the forefront of the Artificial Intelligence sector, acting as a vital provider of training and evaluation data, as well as comprehensive solutions for the entire machine learning lifecycle. In this role, you will collaborate closely with Scale’s ML teams and researchers to construct the foundational platform that supports all our ML research and development initiatives. Your work will involve building and optimizing this platform to facilitate the training, inference, and data curation of next-generation LLMs.If you are passionate about driving the future of AI through groundbreaking innovations, we invite you to connect with us!
Join Baseten as a Post-Training Research Engineer and contribute to groundbreaking advancements in machine learning and AI. In this role, you will leverage your engineering skills to analyze and enhance models post-training, ensuring optimal performance and efficiency.