Nudging Hidden States: Training-Free Model Steering for Chain-of-Thought Reasoning in Large Audio-Language Models
Abstract
Training-free inference-time model steering enhances large audio-language model reasoning through diverse information sources and cross-modal transfer from text to speech, achieving improved accuracy and data efficiency.
Chain-of-thought (CoT) prompting has been extended to large audio-language models (LALMs) to elicit reasoning, yet enhancing its effectiveness without training remains challenging. We study inference-time model steering as a training-free approach to improve LALM reasoning. We introduce three strategies using diverse information sources and evaluate them across four LALMs and four benchmarks. Results show general accuracy gains up to 4.4% over CoT prompting. Notably, we identify a cross-modal transfer where steering vectors derived from few text samples effectively guide speech-based reasoning, demonstrating high data efficiency. We also examine hyperparameter sensitivity to understand the robustness of these approaches. Our findings position model steering as a practical direction for strengthening LALM reasoning.
Community
Steering LALMs for better CoT Reasoning
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Listen First, Then Answer: Timestamp-Grounded Speech Reasoning (2026)
- ALARM: Audio-Language Alignment for Reasoning Models (2026)
- RankSteer: Activation Steering for Pointwise LLM Ranking (2026)
- ConPress: Learning Efficient Reasoning from Multi-Question Contextual Pressure (2026)
- Rethinking Token-Level Policy Optimization for Multimodal Chain-of-Thought (2026)
- ALICE: A Multifaceted Evaluation Framework of Large Audio-Language Models'In-Context Learning Ability (2026)
- Learning When to Sample: Confidence-Aware Self-Consistency for Efficient LLM Chain-of-Thought Reasoning (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2603.14636 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper