LLM Analysis

6papers
2.7viability

State of the Field

Current research on large language models (LLMs) is increasingly focused on understanding and improving their reasoning capabilities and knowledge representation. Recent work has highlighted the significance of reasoning traces, revealing that the accuracy of LLM responses improves as more of these traces are provided, suggesting a need for better trace management in deployment contexts. Additionally, the exploration of long-tail knowledge has uncovered persistent gaps in LLM performance, particularly for infrequent or domain-specific information, prompting calls for more robust evaluation practices and interventions to enhance fairness and accountability. Studies on the differential encoding of syntactic and semantic information are shedding light on how these elements are represented within models, while investigations into motivated reasoning reveal discrepancies between LLM behavior and human reasoning patterns. Collectively, these insights are guiding the development of more reliable and interpretable LLMs, with potential applications in areas such as automated survey analysis and decision support systems.

Last updated Feb 26, 2026

Papers

1–6 of 6
Research Paper·Jan 30, 2026

Probing the Trajectories of Reasoning Traces in Large Language Models

Large language models (LLMs) increasingly solve difficult problems by producing "reasoning traces" before emitting a final response. However, it remains unclear how accuracy and decision commitment ev...

5.0 viability
Research Paper·Feb 18, 2026

Long-Tail Knowledge in Large Language Models: Taxonomy, Mechanisms, Interventions and Implications

Large language models (LLMs) are trained on web-scale corpora that exhibit steep power-law distributions, in which the distribution of knowledge is highly long-tailed, with most appearing infrequently...

3.0 viability
Research Paper·Jan 8, 2026

Differential syntactic and semantic encoding in LLMs

We study how syntactic and semantic information is encoded in inner layer representations of Large Language Models (LLMs), focusing on the very large DeepSeek-V3. We find that, by averaging hidden-rep...

2.0 viability
Research Paper·Jan 22, 2026

Replicating Human Motivated Reasoning Studies with LLMs

Motivated reasoning -- the idea that individuals processing information may be motivated to reach a certain conclusion, whether it be accurate or predetermined -- has been well-explored as a human phe...

2.0 viability
Research Paper·Feb 16, 2026

The Potential of CoT for Reasoning: A Closer Look at Trace Dynamics

Chain-of-thought (CoT) prompting is a de-facto standard technique to elicit reasoning-like responses from large language models (LLMs), allowing them to spell out individual steps before giving a fina...

2.0 viability
Research Paper·Mar 4, 2026

Causality Elicitation from Large Language Models

Large language models (LLMs) are trained on enormous amounts of data and encode knowledge in their parameters. We propose a pipeline to elicit causal relationships from LLMs. Specifically, (i) we samp...

2.0 viability