Paper-Conference

TherapyGym: Evaluating and Aligning Clinical Fidelity and Safety in Therapy Chatbots featured image

TherapyGym: Evaluating and Aligning Clinical Fidelity and Safety in Therapy Chatbots

ICML 2026 submission on evaluating clinical fidelity and safety in therapy chatbots.

fangrui-huang

Developing and Utilizing a Large-Scale Cantonese Dataset for Multi-Tasking in Large Language Models

EMNLP 2025 Findings paper on developing a large-scale Cantonese dataset for LLM multi-tasking.

jiyue-jiang

UAlign: Leveraging Uncertainty Estimations for Factuality Alignment on Large Language Models

A main conference paper at ACL 2025 presenting UAlign for factuality alignment using uncertainty estimations.

boyang-xue

MlingConf: A Comprehensive Study of Multilingual Confidence Estimation on Large Language Models

ACL 2025 Findings paper on multilingual confidence estimation in LLMs.

boyang-xue
TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning featured image

TreeSynth: Synthesizing Diverse Data from Scratch via Tree-Guided Subspace Partitioning

A spotlight paper at NeurIPS 2025 presenting TreeSynth, a method for synthesizing diverse data through tree-guided subspace partitioning.

avatar
Sheng Wang
QSpec: Speculative decoding with complementary quantization schemes featured image

QSpec: Speculative decoding with complementary quantization schemes

EMNLP 2025 main conference paper on efficient inference through speculative decoding.

juntao-zhao
ProReason: Multi-modal proactive reasoning with decoupled eyesight and wisdom featured image

ProReason: Multi-modal proactive reasoning with decoupled eyesight and wisdom

EMNLP 2025 main conference paper on multi-modal proactive reasoning (Co-Corresponding Author).

jingqi-zhou
MoS: Unleashing parameter efficiency of low-rank adaptation with mixture of shards featured image

MoS: Unleashing parameter efficiency of low-rank adaptation with mixture of shards

A paper at ICLR 2025 presenting MoS, a method for more parameter-efficient LoRA through mixture of shards.

avatar
Sheng Wang
How Well Do LLMs Handle Cantonese? Benchmarking Cantonese Capabilities of Large Language Models featured image

How Well Do LLMs Handle Cantonese? Benchmarking Cantonese Capabilities of Large Language Models

NAACL 2025 Findings paper benchmarking Cantonese capabilities of LLMs.

jiyue-jiang
Forewarned is Forearmed: Harnessing LLMs for Data Synthesis via Failure-Induced Exploration featured image

Forewarned is Forearmed: Harnessing LLMs for Data Synthesis via Failure-Induced Exploration

ICLR 2025 paper on data synthesis through failure-induced exploration.

qintong-li