Improving Multi-hop Logical Reasoning in Small LMs with LoRA Training

Authors

  • Onur Bilgin Department of Computer Science and Engineering, University of South Florida https://orcid.org/0009-0002-1690-4779
  • Abdullah As Sami PhD student, University of South Florida
  • Suraj Kumar Research Volunteer, University of South Florida
  • John Licato Associate Professor, University of South Florida

DOI:

https://doi.org/10.32473/flairs.38.1.138643

Keywords:

logical reasoning, multi-hop reasoning, Navset, LoRA, small language models

Abstract

Language models show increasing performance in reasoning tasks. However, logical reasoning in complex tasks remains a challenge. This challenge is more apparent when resources are limited, such as using smaller language models or small datasets for knowledge extraction. How can language models be used in this case to generalize and solve complex logical reasoning tasks? In this work, we show that LoRA training of language models with small datasets can improve logical reasoning and transferability for fact extraction. In our tests, we extracted facts with CoT-prompting to use them as input to the rule set. We explored our experiments with the StepGame, Navset, Comparison, and TriviaQA datasets and evaluated our results with precision, recall, and accuracy metrics. We compared the results against untrained language models. Our results show that LoRA training improves logical reasoning even for out-of-distribution samples.

Downloads

Published

14-05-2025

How to Cite

Bilgin, O., Sami, A. A., Kumar, S., & Licato, J. (2025). Improving Multi-hop Logical Reasoning in Small LMs with LoRA Training. The International FLAIRS Conference Proceedings, 38(1). https://doi.org/10.32473/flairs.38.1.138643