Skip to main content

Spring Deadline: Sunday, March 1 @ 11:59pm PT. Click here to apply.

Back to Research
Accepted to MathAI @ NeurIPS 2024

Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting

Tim Knappe, Ryan Li, Ayush Chauhan, Kaylee Chhua

Abstract

We propose Semantic Self-Consistency (SSC), a novel framework for evaluating the reasoning capabilities of large language models. SSC measures whether a model produces semantically equivalent answers when presented with logically equivalent formulations of the same question. Unlike traditional consistency metrics that focus on exact string matching, SSC captures deeper semantic alignment through learned embeddings. Our experiments reveal significant inconsistencies in state-of-the-art models, with performance dropping by 15-30% on semantically rephrased questions. We release a benchmark of 10,000 question pairs for evaluating SSC.

Citation

Tim Knappe, Ryan Li, Ayush Chauhan, Kaylee Chhua. "Semantic Self-Consistency: Enhancing Language Model Reasoning via Semantic Weighting". Accepted to MathAI @ NeurIPS 2024.

Details

Conference
Accepted to MathAI @ NeurIPS 2024
Authors
4 authors

Publish Your Research

Join Algoverse and work with world-class mentors to publish at top AI conferences.

Start Your Application