Vol.13, No.4, November 2024.                                                                                                                                                                           ISSN: 2217-8309

                                                                                                                                                                                                                       eISSN: 2217-8333

 

TEM Journal

 

TECHNOLOGY, EDUCATION, MANAGEMENT, INFORMATICS

Association for Information Communication Technology Education and Science


Development of Open Large Language Models for Artificial Intelligence Digital Textbooks

 

Youngho Lee

 

© 2024 Youngho Lee, published by UIKTEN. This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. (CC BY-NC-ND 4.0)

 

Citation Information: TEM Journal. Volume 13, Issue 4, Pages 2773-2783, ISSN 2217-8309, DOI: 10.18421/TEM134-14, November 2024.

 

Received: 19 June 2024.

Revised: 19 September 2024.
Accepted: 04 November 2024.
Published: 27 November 2024.

 

Abstract:

 

Artificial Intelligence (AI) is being utilized in various fields, and research on generative AI, particularly within natural language processing (NLP) technology, is actively being conducted. Currently, research related to generative AI in the education sector utilizes closed large language models (LLMs) like GPT. However, these models have limitations, as they are difficult to fine-tune and incur high costs. This study aims to explore the potential educational applications of Open LLMs by fine-tuning and comparing the performance of Llama2 and Polyglot, which are built on a Korean-based model, with Llama3, which is not based on a Korean model. The experimental results, using a question-and-answer dataset from elementary school social studies and science subjects, showed that the Llama2 13B model exhibited the highest performance, followed by the Polyglot 12.8B model. The Llama3 8B model achieved approximately 93.08% of the performance of the Llama2 13B model and about 98.63% of the performance of the Polyglot 12.8B model. This indicates that even relatively small, non-Koreanbased models can demonstrate high performance. These results suggest that future development of LLMs base models may omit the process of converting them into language-specific base models. Additionally, fine-tuning Open LLMs for educational applications shows potential for providing personalized education.

 

Keywords – Open LLMs, fine-tuning, performance of LLMs, AI digital textbooks, Llama3, Llama2, Polyglot.

 

-----------------------------------------------------------------------------------------------------------

Full text PDF >  

-----------------------------------------------------------------------------------------------------------

 


Copyright © 2024 UIKTEN
Copyright licence: All articles are licenced via Creative Commons CC BY-NC-ND 4.0 licence