Boston Dynamics

Atlas VLA Research Intern

Waltham Office (POST) Full time

Are you passionate about the intersection of foundation models and physical AI? As an intern on the Atlas VLA Research Team, you will join a world-class group of engineers and scientists dedicated to giving Atlas the intelligence to perceive, reason, and act.

Our team focuses on scaling Vision-Language-Action (VLA) models, leveraging diverse data sources, and building accurate spatial perception that keep Atlas grounded and capable of performing challenging dexterous tasks. We are looking for PhD-level interns to lead research projects that push the boundaries of what humanoid robots can do. You won't work on pure theory; you'll be deploying your ideas directly onto one of the most sophisticated pieces of hardware on the planet.

How you will make an impact:

  • Lead a high-stakes research project focused on either VLA scaling/training or spatial perception (SLAM/Calibration).

  • Prototype and deploy your algorithms directly on Atlas, moving from simulation to hardware.

  • Architect data pipelines that ingest alternative data sources to improve robot robustness.

  • Write production-grade code (Python/C++) that integrates with our existing systems.

  • Collaborate across teams To integrate learned policies with low-level robot control.

We are looking for:

  • Actively pursuing a PhD (preferred) or a research-heavy Master’s in Computer Science, Robotics, Machine Learning, or a related field.

  • Expertise in ONE or more of the following tracks:

    • Training large-scale multimodal models (VLMs/LLMs), imitation learning, or generative world models for robotics.

    • Classical and learned SLAM, visual odometry, or extrinsic/intrinsic camera calibration at scale.

Nice to have:

  • Experience troubleshooting and deploying algorithms on physical robot platforms, especially mobile and humanoid form factors.

  • Experience working with and contributing to large-scale datasets (e.g., Open X-Embodiment) or specialized data collection approaches like Universal Manipulation Interface.

  • Experience with large-scale cluster training (SLURM, distributed GPU training) and maintaining high-quality codebases.

  • Strong grasp of Lie groups, optimization, or transformer architectures.

The hourly pay range for this position is between $30-$45/ hour. Base pay will depend on multiple individualized factors including, but not limited to internal equity, job related knowledge, skills, education and experience. This range represents a good faith estimate of compensation at the time of posting.