Skip to content
View nssmd's full-sized avatar

Block or report nssmd

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don’t include any personal information such as legal names or email addresses. Markdown is supported. This note will only be visible to you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
nssmd/README.md

Hi, I'm Zimo Wen

I am an undergraduate researcher in Computer Science at Shanghai Jiao Tong University, in the Zhiyuan Honors Program and MVIG Lab.

I work on:

  • embodied intelligence
  • multimodal models
  • VLA systems
  • multimodal evaluation and research tooling

Current focus

  • Building embodied systems with wearable gripper fingertips, force-aware sensing, teleoperation, and VLA-style training.
  • Working on multimodal generation, unified understanding, and world-model-adjacent systems.
  • Building benchmarks and evaluation pipelines for multimodal models.

Selected work

Selected papers

Links

Pinned Loading

  1. EvolvingLMMs-Lab/lmms-engine EvolvingLMMs-Lab/lmms-engine Public

    A simple, unified multimodal models training engine. Lean, flexible, and built for hacking at scale.

    Python 774 35

  2. Physical-Intelligence/openpi Physical-Intelligence/openpi Public

    Python 11.8k 1.9k

  3. EvolvingLMMs-Lab/lmms-eval EvolvingLMMs-Lab/lmms-eval Public

    One-for-All Multimodal Evaluation Toolkit Across Text, Image, Video, and Audio Tasks

    Python 4.1k 583

  4. AlenjandroWang/ASVR AlenjandroWang/ASVR Public

    Autoregressive Semantic Visual Reconstruction Helps VLMs Understand Better

    Python 190 18

  5. waltstephen/ArgusBot waltstephen/ArgusBot Public

    ArgusBot: A 24/7 supervisor Agent for Codex CLI and Claude Code CLI that keeps agents running, reviewing, and planning until the job is actually done.

    Python 301 28