Identificador persistente para citar o vincular este elemento: https://accedacris.ulpgc.es/jspui/handle/10553/154908
Título: Fine-Tuning a Local LLM for Thermoelectric Generators with QLoRA: From Generalist to Specialist
Autores/as: Monzón Verona, José Miguel 
Garcia-Alonso Montoya, Santiago 
Santana Martin, Francisco Jorge 
Clasificación UNESCO: Investigación
Palabras clave: Modules
Llm
Qlora
Janv1-4B
Fine-Tuning, et al.
Fecha de publicación: 2025
Publicación seriada: Applied Sciences 
Resumen: This work establishes a large language model (LLM) specialized in the domain of thermoelectric generators (TEGs), for deployment on local hardware. Starting with the generalist JanV1-4B model and Qwen3-4B-Thinking-2507 models, an efficient fine-tuning (FT) methodology using quantized low-rank adaptation (QLoRA) was employed, modifying only 3.18% of the total parameters of thee base models. The key to the process is the use of a custom-designed dataset, which merges deep theoretical knowledge with rigorous instruction tuning to refine behavior and mitigate catastrophic forgetting. The dataset employed for FT contains 202 curated questions and answers (QAs), strategically balanced between domain-specific knowledge (48.5%) and instruction-tuning for response behavior (51.5%). Performance of the models was evaluated using two complementary benchmarks: a 16-question multilevel cognitive benchmark (94% accuracy) and a specialized 42-question TEG benchmark (81% accuracy), scoring responses as excellent, correct with difficulties, or incorrect, based on technical accuracy and reasoning quality. The model's utility is demonstrated through experimental TEG design guidance, providing expert-level reasoning on thermal management strategies. This study validates the specialization of LLMs using QLoRA as an effective and accessible strategy for developing highly competent engineering support tools, eliminating dependence on large-scale computing infrastructures, achieving specialization on a consumer-grade NVIDIA RTX 2070 SUPER GPU (8 GB VRAM) in 263 s.
URI: https://accedacris.ulpgc.es/jspui/handle/10553/154908
ISSN: 2076-3417
DOI: 10.3390/app152413242
Fuente: Applied Sciences-Basel,v. 15 (24), (Diciembre 2025)
Colección:Artículos
Adobe PDF (2,36 MB)
Vista completa

Google ScholarTM

Verifica

Altmetric


Comparte



Exporta metadatos



Los elementos en ULPGC accedaCRIS están protegidos por derechos de autor con todos los derechos reservados, a menos que se indique lo contrario.