cover image
Mercedes-Benz Research & Development North America, Inc.

Mercedes-Benz Research & Development North America, Inc.

www.mbrdna.com

1 Job

825 Employees

About the Company

Mercedes-Benz Research & Development North America (MBRDNA) is continuously striving to innovate premium mobility. It is a place for exceptional people with outstanding ideas and the absolute willingness to bring them to life. We are not about cars only - we are also about the latest and greatest software, cutting-edge technology, and groundbreaking innovation: Embedded in a worldwide network MBRDNA is headquartered in Silicon Valley, California, with key areas of Autonomous Driving, Advanced Interaction Design, Digital User Experience, Machine Learning, Consumer Research, and Open Innovation. In Redford, Michigan, the focus is on Powertrain and eDrive technology as well as in Long Beach, where the teams test durability of the latest driver assistant and telematic systems. The Digital Hub in Seattle focusses on developing a cloud architecture and building out the cloud platform for the next generation of connected car services. The Testing and Regulatory Affairs Division in Ann Arbor and the Advanced Vehicle Design in Carlsbad complete the competence center.

Listed Jobs

Company background Company brand
Company Name
Mercedes-Benz Research & Development North America, Inc.
Job Title
Intern, Machine Learning
Job Description
**Job title:** Intern, Machine Learning **Role Summary:** A research‑focused internship that applies state‑of‑the‑art vision‑language models (VLMs) to autonomous driving problems. The role involves developing, refining, and evaluating multi‑modal models for tasks such as scene captioning, visual question answering, and aligning textual navigation guidance with visual perception. **Expectations:** - Strong foundation in machine learning, computer vision, and natural language processing. - Experience training deep learning models on multi‑modal inputs (images, video, text). - Proficiency in Python and deep learning frameworks (PyTorch, TensorFlow). - Knowledge of large‑scale transformer architectures and VLMs such as CLIP, BLIP, or ViT‑based adaptations. - Ability to conduct rigorous evaluation, benchmark against datasets, and identify performance gaps. - Excellent written communication for internal research reports and potential conference submissions. **Key Responsibilities:** - Investigate and implement advanced VLM techniques for autonomous driving scenarios. - Develop models for captioning complex driving scenes, visual question answering, and multimodal route planning. - Collaborate with engineering teams to integrate VLM solutions into existing perception and planning pipelines. - Benchmark model performance on internal and public datasets; propose and test improvements. - Document research findings, prepare internal reports, and contribute to scholarly publications when results warrant. **Required Skills:** - Proficient in Python programming. - Hands‑on experience with PyTorch or TensorFlow for building and training deep neural networks. - Familiarity with large‑scale vision‑language models (e.g., CLIP, BLIP, ViT‑based VLMs). - Understanding of transformer architectures and multi‑modal pre‑training strategies. - Ability to evaluate models using metrics appropriate to captioning, VQA, and intent prediction tasks. - Strong analytical and problem‑solving abilities. **Required Education & Certifications:** - Bachelor’s or Master’s degree in Computer Science, Electrical Engineering, Robotics, or a related field with emphasis on machine learning, computer vision, and/or natural language processing. - PhD candidate or recent graduate preferred, with a publication record in reputable AI/ML/ CV/ NLP venues. ---
San jose, United states
On site
Fresher
18-11-2025