SMART
Intranet Share Print

SMART office will be closed today as the PSI reading is >300. For enquiries, please email the respective person. SMART office will be open tomorrow should the PSI reading be below 300.

  • Home
  • Research
    • LEES
      • LEES Careers
    • AMR
      • AMR Careers
    • CAMP
      • CAMP Careers
    • DiSTAP
      • DiSTAP Careers
    • FM
      • FM Careers
    • BioSyM
      • BioSyM Careers
    • CENSAM
      • CENSAM Careers
    • ID
      • ID Careers
    • M3S
      • M3S Careers
  • Fellowships
    • Undergraduate (SMURF)
      • Apply for SMURF Fellowship
    • Graduate (SMART Graduate)
      • Apply for SMART Graduate Fellowship
    • Postdoctoral (SMART Scholars)
      • Apply for SMART Scholars Fellowship
  • News & Events
    • News
  • Careers
    • Career Opportunities
      • Job Application
    • Student Jobs
banner
Career Opportunities
  • Career Opportunities
  • Student Jobs
  • Home
  • Career Opportunities
  • Job Application
  • Research Intern (LLM Agents, World Models & Embodied AI)

Processing

Please wait, form submission in progress

Research Intern (LLM Agents, World Models & Embodied AI)

IRG_M3S_T4I_2025_019

<< Back to Job Listing
Posted on 17 December 2025
Group: M3S

Project Overview

We are seeking 2-3 highly motivated research interns interested in advancing the state of the art in Embodied AI, Large Language Model (LLM) Agents, and Autonomous Systems at SMART.

With the rapid evolution of Generative AI, research interns will assist researchers in investigating how to develop next-generation LLM-based Agents and World Models that enable safe and trustworthy autonomy. Our goal is to bridge the gap between high-level reasoning and low-level control in complex urban environments. We are specifically interested in, but not limited to, the following research topics:

  • LLM Agents for Autonomy: Utilizing LLMs for planning, reasoning, and decision-making in autonomous driving and robotics.
  • World Models & Generative Simulation: Building data-driven world models for video generation, future prediction, and closed-loop simulation.
  • Embodied AI: Integrating vision, language, and action to create intelligent agents capable of interacting with the physical world.

Responsibilities

  • Develop and benchmark LLM-based decision agents, including chain-of-thought reasoning, structured memory, and planning modules.
  • Implement and train world models for spatiotemporal prediction, simulation, and autonomous driving or mobility applications.
  • Build reproducible research pipelines and assist with experiments, ablation studies, and model diagnostics.
  • Collaborate with research staff and graduate students to prepare publications for top conferences and journals in AI, robotics, and transportation.

Requirements

  • Currently pursuing a Bachelor’s or Master’s degree in Engineering, Computer Science, Data Science, Robotics, AI/ML, or related fields.
  • Strong interest in LLM agents, autonomous systems, world models, or risk-aware decision making.
  • Proficiency in Python, deep learning frameworks (PyTorch, Transformers, JAX preferred), and modern AI toolchains.
  • Experience in one or more areas such as LLM/MLLM reasoning, world models, multimodal learning, etc. is a plus.
  • We welcome students of all experience levels, as long as you are motivated, proactive, and willing to learn.

The internship must be approved by the University.

Apply

Please download and complete our SMART Job Application Form and upload in the field below. Thank you.

Personal Details

First Name*
Last Name*
Phone No*
Email Address*
How did you get to know of this job?
Notification on personal data protection

File Attachments

Please upload one of these formats PDF & Word Doc only. File size limited to 2MB

Upload Completed SMART Job Application Form *
Upload Your CV *
Upload Cover Letter
Any Other Required Documents

Back to Top
Last Updated 01/03/2024 Privacy Policy | Terms of Use | Open Access Articles | Sitemap | 2016 All Rights Reserved. Singapore-MIT Alliance for Research and Technology Last Updated 01/03/2024