Our mission is to enable trust between humans and AI agents through language technology, making human-machine teaming possible in real-world applications. Our research centers on natural language processing, with broader contributions to machine learning, multimodal models, and artificial intelligence.

AI for Formal Mathematics

AI for Science In progress

We aim to develop an agentic AI system that automates key aspects of mathematical reasoning, including theorem decomposition, lemma identification, novel proof strategy generation, seamless translation between natural language and formal proof systems, and automatic conjecture discovery.

Multimodal Interactive Learning

Multimodal In progress

We develop multimodal systems that learn new concepts through language, visual grounding, interaction, and feedback, with emphasis on recognizing unfamiliar objects, understanding richer descriptions, robust video-language understanding, and more reliable human-machine collaboration.

Related Projects

  • NOVA: A Neuro-Symbolic Vision-Language Framework for Multimodal Human-Machine Interactions
  • MIRACLE: Multimodal InteRActive Conceptual Learning

Sponsors

DARPA ONR

Neuro-Symbolic Reasoning

Reasoning In progress

We integrate neural prediction with probabilistic inference and symbolic constraints so models can produce structured outputs that are accurate and verifiable. We apply these approaches to problems ranging from improving LLM safety to enhancing model reasoning capabilities.

Related Projects

  • PYLON: An Integrated Semantic Framework for Probabilistic Neuro-Symbolic Learning and Reasoning
  • CRII: Learning Structured Prediction Model with Auxiliary Supervision

Sponsors

DARPA NSF

LLM Safety and Guardrails

TrustNLP In progress

We build detection, evaluation, guardrail, and mitigation methods for unsafe behavior, over-refusal, sleeper agents, jailbreaks, and customized multimodal safety policies.

Related Projects

  • SLES: Verifying and Enforcing Safety Constraints in AI-based Sequential Generation
  • Customized robust and controllable text processing
  • Safety reasoning and red-teaming for LLMs and multimodal systems

Sponsors

NSF Amazon Google

Mulitmodal AI Agents, Tool Use, and Long-Term Memory

Agent In progress

We develop open-source language agents, tool-use workflows, long-term memory benchmarks, and data-analysis agents for complex mulitmodal interactive tasks.

Related Projects

  • Enhancing the Reasoning Capabilities of Multimodal Large Language Models
  • Learning to Reason Better Than Your Teacher for Adaptive Multimodal Agents

Sponsors

Apple Amazon Google

Trustwrothy Medical AI Agent

Information In progress

We develop trustworthy AI solutions for healthcare applications, from matching patients in clinical trails, to clinical report analysis, radiology summarization, and patient-centered medical decision support.

Related Projects

  • Medical Vision-Language Foundation Models for Clinical Report Analysis
  • Co-designing ethical multimodal AI systems for mapping T1D progression

Sponsors

NIH Optum Labs

Information Extraction for Pandemic Prevention

Information Expired

We develop NLP approaches to detect early signs of emerging infectious diseases, predict their spread, and detect and monitor risk factors through multilingual social media posts.

Related Projects

  • PIPP Phase 1: An end-to-end pandemic early warning system by harnessing open-source intelligence
  • Online news trend-watching via linguistic analysis

Sponsors

NSF Taboola

Governing Bias and Human-Centered AI

TrustNLP Expired

We study how bias appears in representations, generation, recommendations, and social text, and design human-centered interventions for more equitable AI systems.

Related Projects

  • AI-DCL: Governing Bias in AI System with Humans in the Decision Loop
  • Discerning Group Biases in Online Communities via Linguistic Analysis
  • Sloan Research Fellowship on fairness, robustness, and inclusion

Sponsors

NSF DARPA Sloan Foundation Okawa Foundation

Machine Common Sense

Multimodal Expired

This project studies commonsense knowledge from video, images, text, and knowledge bases, with benchmarks and models for multimodal social and scientific reasoning.

Related Projects

  • Discovering Common Sense from Video, Images, Text and Knowledge Bases

Sponsors

DARPA