Enhancing Zero-Shot Reasoning in Language Models Via Hybrid Instruction Marginalization
DOI:
https://doi.org/10.70764/gdpu-bit.2025.1(2)-01Keywords:
Zero-Shot Generalization, Marginalization in NLP, Instruction Tuning, Reasoning in LLMs, Dynamic Prompt SelectionAbstract
Objective: The study aims to enhance the reasoning abilities of Large Language Models (LLMs), which often remain shallow, inconsistent, and error-prone in complex multi-step tasks. It introduces the Hybrid Instruction Tuning Framework (HITF) to improve zero-shot reasoning through a task-aware hybrid selector that integrates both human-annotated and automatically generated examples.
Research Design & Methods: HITF strengthens reasoning performance using three main techniques: synthesizing transitional results, context-aware prompt merging, and recurrent optimization, all executed without model recalibration. The framework is empirically evaluated using rigorous cognitive benchmarks, including SuperGLUE, MMLU, GSM8K, and FermiQA. Component isolation tests examine the independent contribution of the example selector, output synthesizer, and instruction combiner. Statistical variability assessments further validate result reliability.
Findings: Results show that HITF consistently outperforms state-of-the-art methods across multiple metrics, demonstrating higher measurement accuracy, stronger argumentative quality, and deeper analytical processing. All core modules exhibit significant and measurable contributions, supported by stable statistical outcomes.
Implications & Recommendations: Findings suggest that combining context-driven instruction selection with statistical consolidation techniques can substantially improve deductive reasoning in LLMs, particularly in data-scarce and example-free settings. Future research should explore HITF’s integration with larger models and its application in real-world reasoning-intensive domains.
Contribution & Value Added: This study offers an innovative framework that enhances zero-shot reasoning without retraining. By merging hybrid instruction selection and iterative optimization strategies, HITF narrows the reasoning gap between LLMs and humans and provides a scalable, reliable approach for advancing high-level reasoning in modern language models
Downloads
Published
Issue
Section
License
Copyright (c) 2025 Breakthroughs Information Technology

This work is licensed under a Creative Commons Attribution 4.0 International License.






