About Me

Here is a virtual greeting from Ziqiao. As many of my friends found it hard to pronounce my name (马子乔, pronounced as /ma˨˩˦ tsɨ˧˥ tɕʰiɑʊ˧˥/ in Mandarin), it's absolutely fine to just call me Martin alternatively.

CV (2025) / Google Scholar / Research / CSE595 NLP / GrowAI.org / Chat?

For fun...

Bio / Profile Photo

Ziqiao (Martin) Ma is a 4th year Ph.D. candidate at the University of Michigan in Computer Science and Engineering advised by Professor Joyce Chai. His work has been supported in part by the Weinberg Cognitive Science Fellowship. He is also a part-time researcher at Adobe Research. Previously, he worked with Amazon Science. He received an Outstanding Paper Award at ACL 2023, and an Amazon Alexa Prize Award. He taught Natural Language Processing and won an Outstanding Graduate Student Instructor Award. He co-organized Bi-Align @ ICLR/CHI 2025, SpLU-RoboNLP @ ACL 2024, and co-instructed tutorial on Learning Language through Grounding @ NAACL 2025.

My Research (TL;DR)

I am interested in Grounded Language Processing and Computational Psycholinguistics. The three constant themes of my research are language, interaction, and embodiment, from a scalable and cognitive angle.

Learning with natural supervision: language grounding and alignment
Scientific inquiry: computational linguistics and psycholinguistics
Applications: multimodal interactive agents

I include a more dynamic and spontaneous document of my thoughts in my Research Blueprint.

Selected Awards/Recognitions

Selected Fellowships/Scholarships

Updates

News

Archived news...
  • [May. 2023] I started my intern with Amazon Alexa AI (Amazon AGI)!!
  • [Sep. 2022] I will serve as the Poster/Demo Session Chair for Michigan AI Symposium 2022.
  • [Aug. 2022] I will be the Graduate Student Instructor for EECS 595 (NLP) in Fall 2022 at Michigan.
  • [Mar. 2021] I will join the family of the Michigan AI as a Ph.D. student this fall. Go Blue!
  • [Dec. 2020] I will be the Instructional Aide for EECS 492 (Intro. AI) in Winter 2021 at Michigan.

Paper Alerts

Archived news...
  • [Nov. 2024] Our survey on vision-language navigation is accepted to TMLR with a survey certificate :)
  • [Sep. 2024] One paper to appear in NeurIPS 2024, see you in Vancouver :)
  • [Jun. 2024] One paper to appear in IROS 2024, see you in Abu Dhabi :)
  • [Feb. 2024] Two papers to appear in CVPR 2024, see you in Seattle :)
  • [Oct. 2023] One paper to appear in EMNLP 2023, see you in Singapore :)
  • [Sep. 2023] One paper to appear in NeurIPS 2023, see you in New Orleans :)
  • [May. 2023] Two papers to appear in ACL 2023, and I will serve as an on-site volunteer in Toronto :)
  • [Apr. 2023] One paper to appear in IJCAI 2023, see you in Macau :)
  • [Oct. 2022] Two papers to appear in EMNLP 2022, and I will serve as an on-site volunteer in Abu Dhabi :)

Seminar Talks

Previous talks...
  • [20241203] Seeing What You See: Perceptual Perspective-Taking Towards a Situated Machine Theory of Mind @ Cognitive Science Seminar Series, UMich.
  • [20240712] Babysit A Language Model From Scratch: Interactive Language Learning by Trials and Demonstrations @ Deep Learning: Classics and Trends (DLCT).
  • [20240705] Language Grounding to the Visual World and Human Interactions: How Far Are We from Embodied Dialogue Agents @ Data Science Group, KAIST.
  • [20240627] Babysit A Language Model From Scratch: Interactive Language Learning by Trials and Demonstrations @ CoCoDev Seminar.
  • [20240529] Language Grounding to the Visual World and Human Interactions: How Far Are We from Embodied Dialogue Agents @ University of Maryland.

Experiences

Education

Industry

Current Teaching

Guest Lectures

Academic Services

The 1st Workshop / Special Interest Group on Bidirectional Human-AI Alignment (Bi-Align @ ICLR 2025 / CHI 2025)

Co-organizer

[Homepage/CFP] [OpenReview]

The 4th International Combined Workshop on Spatial Language Understanding and Grounded Communication for Robotics (SpLU-RoboNLP @ ACL 2024)

Co-organizer

[Homepage/CFP] [OpenReview] [Proceedings]

The 5th Michigan AI Symposium: AI & Accessibility (2022)

Poster/Demo Chair

[Homepage/CFP]

Publications [.bib]

Show by... ( Recent Selection / Cognitive AI Selection / Recognition / Year / Topics )

Research Topics: Multimodal Learning and Generation / (Inter)active Learning and Alignment / Embodiment and Situated Intelligence / Teaching & Community Services

* indicates equal contributions; § indicates correspondence and mentoring.

VEGGIE: Instructional Editing and Reasoning Video Concepts with Grounded Generation
Shoubin Yu*, Difan Liu*, Ziqiao Ma*, Yicong Hong, Yang Zhou, Hao Tan, Joyce Chai, Mohit Bansal

Preprint, 2025

Paper / Homepage / Dataset

TL;DR...
  • We introduce VEGGIE, a diffusion-loss only video generative model that handles various tasks for both video concept grounding and editing from user instructions.
  • Pixel-level grounded training helps various video concept editing task in multi-task learning.
  • VEGGIE shows emergent zero-shot multimodal instructional and in-context video editing.
Do Vision-Language Models Represent Space and How? Evaluating Spatial Frame of Reference under Ambiguities
Zheyuan Zhang*, Fengyuan Hu*, Jayjun Lee*, Freda Shi, Parisa Kordjamshidi, Joyce Chai, Ziqiao Ma§

ICLR 2025 (Oral) / The 1st Pluralistic Alignment Workshop @ NeurIPS 2024

Paper / Homepage / GitHub / Dataset / Poster

TL;DR...
  • We introduce COMFORT, a protocol to evaluate spatial reasoning in VLMs across multilingual and ambiguous frames of reference (FoR);
  • VLMs exhibit poor robustness and consistency, lack the flexibility to accommodate multiple FoRs, and fail to adhere to language-specific or culture-specific conventions in cross-lingual tests.
Babysit A Language Model From Scratch: Interactive Language Learning by Trials and Demonstrations
Ziqiao Ma*, Zekun Wang*, Joyce Chai

NAACL 2025 / The 1st Workshop on LLMs and Cognition (LLMCog) @ ICML 2024 (Oral)

Paper / GitHub / Poster

TL;DR...
  • We introduce a trial-and-demonstration (TnD) learning framework that incorporates three components: student trials, teacher demonstrations, and a reward conditioned on language competence at various developmental stages;
  • TnD accelerates word representation learning for student models of equal and smaller numbers of parameters, and both trials and demonstrations matter.
  • We further show that the teacher's choices of words influence students' word-specific learning efficiency, and a practice-makes-perfect effect is evident by a strong correlation between the frequency of words in trials and their respective learning curves.
GroundHog: Grounding Large Language Models to Holistic Segmentation
Yichi Zhang, Ziqiao Ma, Xiaofeng Gao, Suhaila Shakiah, Qiaozi Gao, Joyce Chai

CVPR 2024

Paper / Homepage / Model (Coming) / Dataset / Poster

TL;DR...
  • We introduce GroundHog, a multimodal large language model grounded in holistic segmentation, using a masked feature extractor and unified grounding masks for fine-grained visual understanding.
  • Trained on the curated M3G2 dataset, GroundHog outperforms in language grounding tasks, reduces object hallucination, and offers improved diagnosis for complex visual inputs.
Inversion-Free Image Editing with Language-Guided Diffusion Models
Sihan Xu*, Yidong Huang*, Jiayi Pan, Ziqiao Ma§, Joyce Chai

CVPR 2024

Paper / Homepage / GitHub / Live Demo / Poster

TL;DR...
  • We derive Denoising Diffusion Consistent Model (DDCM), showing that when the initial sample is known, a special variance schedule reduces the denoising step to the same form as the multi-step consistency sampling;
  • DDCM implies a inversion-free strategy without explicit inversion in sampling for image editing;
  • We further unify the attention control mechanisms in an inference time algorithm for text-guided editing, taking less than 3 seconds per edit.
Towards A Holistic Landscape of Situated Theory of Mind in Large Language Models
Ziqiao Ma, Jacob Sansom, Run Peng, Joyce Chai

EMNLP 2023 (Findings)

Paper / GitHub / Dataset / Poster

TL;DR...
  • We taxonomize machine ToM into 7 mental state categories and delineate existing benchmarks to identify under-explored aspects of ToM;
  • Pilot studies for a holistic and situated evaluation of ToM to break ToM into individual components and treat LLMs as an agent who is physically situated in environments and socially situated in interactions with humans.
World-to-Words: Grounded Open Vocabulary Acquisition through Fast Mapping in Vision-Language Models
Ziqiao Ma*, Jiayi Pan*, Joyce Chai

ACL 2023 (🏆 Outstanding Paper Award)

Paper / GitHub / Model / Dataset / Poster

TL;DR...
  • We introduce OctoBERT, a visually grounded language model designed to acquire grounding ability during pre-training and enable fast mapping of new words through few-shot learning without explicit grounding supervision;
  • Visual grounding accelerates grounded word representation learning;
  • Imageability aligns positively with human intuition and prediction metrics, while concreteness shows opposite correlations -> need for language learning agents to acquire word meanings through physical interactions!

If you like my figures here, I highly recommend you also visit SiX's homepage.

Misc

Fun Facts

Game Design (More)

I seriously considered a career in game design, and although I ultimately chose a different path, it provided excellent preparation for my work in embodied AI research, which often involves intensive programming with simulators.

Here are some of the projects we worked on:

Contracts
Zekai Fan, Shiyu Qu, Juan Rivera Plata, Yihao Huang, Ziqiao Martin Ma

[trailer][itch.io][indidb][tigsource]

  • A turn-based tactic video game.

Mentoring

I understand that access to research oppotunities can be hard, particularly for beginners and the underrepresented. If there is a match in research interests, I am happy to collaborate with undergrads and masters when I have the bandwidth. Please find more details here.
I've been fortunate to have (co-)mentered and collaborated with these amazingly talented young researchers:

Random Tours

Chat?

If you would like to have a random (virtual) coffee chat with me, please visit my calendly page. I am happy to talk if you want to share your stress or just want to chat about life in general (when I have time), but be sure to check out the On-Campus Mental Health Resources @ Michigan.

Get In Touch

You are welcome to drop me a message :)

  • Phone

    xxx-xxx-xxxx
  • marstin0607
  • ziqiao_ma
  • Address

    Bob and Betty Beyster Building 4909,
    2260 Hayward Street,
    Ann Arbor, MI 48109.