Domain-specific knowledge and context in large language models: challenges, concerns, and solutions

Kiran Mayee Adavala, Om Adavala

Abstract


Large language models (LLMs) are ubiquitous today with major usage in the fields of industry, research, and academia. LLMs involve unsupervised learning with large natural language data, obtained mostly from the internet. There are several challenges that arise because of these data sources. One such challenge is with respect to domain-specific knowledge and context. This paper deals with the major challenges faced by LLMs due to data sources, such as, lack of domain expertise, understanding specialized terminology, contextual understanding, data bias, and the limitations of transfer learning. This paper also discusses some solutions for the mitigation of these challenges such as pre-training LLMs on domain-specific corpora, expert annotations, improving transformer models with enhanced attention mechanisms, memory-augmented models, context-aware loss functions, balanced datasets, and the use of knowledge distillation techniques.


Keywords


Bias; Contextual understanding; Domain-specific knowledge; Expert annotations; Large language models; Memory augmented models; Transfer learning

Full Text:

PDF


DOI: http://doi.org/10.11591/ijai.v14.i4.pp2568-2578

Refbacks

  • There are currently no refbacks.


Copyright (c) 2025 Institute of Advanced Engineering and Science

Creative Commons License
This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

IAES International Journal of Artificial Intelligence (IJ-AI)
ISSN/e-ISSN 2089-4872/2252-8938 
This journal is published by the Institute of Advanced Engineering and Science (IAES).

View IJAI Stats