
Multimodal Interaction & World Model
The Seed Multimodal Interaction and World Model team is dedicated to developing models that boast human-level multimodal understanding and interaction capabilities. The team also aspires to advance the exploration and development of multimodal assistant products.
Latest advancements
Selected papers
May 20, 2025
Emerging Properties in Unified Multimodal Pretraining
Computer Vision
May 13, 2025
Seed1.5-VL Technical Report
LLM
Jan 21, 2025
UI-TARS: Pioneering Automated GUI Interaction with Native Agents
Computer Vision
Featured roles
Research Scientist/Engineer - Multimodal Interaction & World Model
Singapore
Experienced Hiring
Apply Now
Research Scientist- Foundation Model, Vision and Language
San Jose / Seattle
Experienced Hiring
Apply Now
Research Scientist, Multimodal Interaction & World Model - 2025 Start
Singapore
Campus Recruitment
Apply Now
Research Scientist Graduate- (Foundation Model, Vision and Language) - 2025 Start (PhD)
San Jose / Seattle
Campus Recruitment
Apply Now
Research Scientist Intern - Multimodal Interaction & World Model - 2025 Start
Singapore
Internship
Apply Now
Student Researcher (Seed - Foundation Model - Vision and Language) - 2025 Start (PhD)
San Jose / Seattle
Internship
Apply Now