About cookies on this site Our websites require some cookies to function properly (required). In addition, other cookies may be used with your consent to analyze site usage, improve the user experience and for advertising. For more information, please review your options. By visiting our website, you agree to our processing of information as described in IBM’sprivacy statement. To provide a smooth navigation, your cookie preferences will be shared across the IBM web domains listed here.
Publication
IJCAI 2024
Workshop paper
InLegalLLaMA: Indian Legal Knowledge Enhanced Large Language Models
Abstract
Large Language Models (LLM) are being increasingly used in many domains including legal and justice. General purpose models trained on web data are not performant enough on legal text analytics (LTA) tasks while fine tuning task specific models is expensive because of the annotation and compute costs. Pre-training domain or application specific models is increasingly popular. However pre-training LLMs in small domain corpora like Indian legal documents is stymied by repetitive and less relevant information in court judgements and records. We introduce InLegalLlama models and show that pre-training LLMs on knowledge graph triples significantly reduces the training effort while retaining comparable performance on LTA tasks.