A Comparative Performance Analysis of Locally Deployed Large Language Models through a Retrieval-Augmented Generation Educational Assistant Application for Textual Data Extraction
Background: Rapid advancements in large language models (LLMs) have significantly enhanced Retrieval-Augmented Generation (RAG) techniques, leading to more accurate and context-aware information retrieval systems.
Methods: This article presents the creation of a RAG-based chatbot tailored for university course catalogs, aimed at answering queries related to course details and other essential academic information, and investigates its performance by testing it on several locally deployed large language models. By leveraging multiple LLM architectures, we evaluate performance of the models under test in terms of context length, embedding size, computational efficiency, and relevance of responses.
Results: The experimental analysis obtained by this research, which builds on recent comparative studies, reveals that while larger models achieve higher relevance scores, they incur greater response times than smaller, more efficient models.
Conclusions: The findings underscore the importance of balancing accuracy and efficiency for real-time educational applications. Overall, this work contributes to the field by offering insights into optimal RAG configurations and practical guidelines for deploying AI-powered educational assistants.
Files and links (2)
pdf
A Comparative Performance Analysis of Locally Deployed Large Language Models...1.47 MBDownloadView
Published (Version of record)Article pdfCC BY V4.0, Open Access
url
A Comparative Performance Analysis of Locally Deployed Large Language Models...View
Published (Version of record)link to articleCC BY V4.0, Open
Related links
Details
Title
A Comparative Performance Analysis of Locally Deployed Large Language Models through a Retrieval-Augmented Generation Educational Assistant Application for Textual Data Extraction