Skip navigation
Please use this identifier to cite or link to this item: https://repositorio.ufpe.br/handle/123456789/58332

Share on

Full metadata record
DC FieldValueLanguage
dc.contributor.advisorRen, Tsang Ing-
dc.contributor.authorSILVA, Weybson Alves da-
dc.date.accessioned2024-10-30T11:38:23Z-
dc.date.available2024-10-30T11:38:23Z-
dc.date.issued2024-10-09-
dc.date.submitted2024-10-29-
dc.identifier.citationSILVA, Weybson Alves da. Evaluation of Large Language Models in Contract Information Extraction. 2024. Trabalho de Conclusão de Curso (Ciência da Computação) – Universidade Federal de Pernambuco, Recife, 2024.pt_BR
dc.identifier.urihttps://repositorio.ufpe.br/handle/123456789/58332-
dc.description.abstractDespite the rapid advancement of Large Language Models (LLMs), there is limited research focused on their effectiveness in extracting specific information from contracts. This study evaluates the effectiveness of state-of-the-art models—GPT-3.5-Turbo, Gemini-1.5-Pro, Claude-3.5-Sonnet, and Llama-3-70B-Instruct—in extracting key clauses from contracts using the Contract Understanding Atticus Dataset (CUAD). We explore the impact of prompting strategies and input context configurations across two scenarios: one covering all 41 clause categories and another focusing on a subset of three. Our findings reveal that LLMs can extract contract information efficiently, outperforming traditional human review in terms of time and cost. Performance, however, varies significantly depending on context size and task specificity, with reduced context approaches and focused extractions often improving recall at the expense of precision. Notably, Claude-3.5-Sonnet, with zero-shot with output example and reduced context, achieved a recall of 0.77 and precision of 0.66, surpassing prior benchmarks on full-category extraction. However, performance is inconsistent across clause types. Models like Llama-3-70B-Instruct, while less robust, demonstrated strong performance on simpler tasks, highlighting their potential in targeted use cases. Additionally, retrieval-augmented generation shows potential for improving extraction and efficiency in long documents, though its performance is constrained by retriever accuracy. Our experiments suggest that with further refinement, LLMs could be vital in automating complex legal tasks, particularly in efficiently handling dense legal texts such as contracts.pt_BR
dc.format.extent28p.pt_BR
dc.language.isoengpt_BR
dc.rightsopenAccesspt_BR
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/br/*
dc.subjectInformation Extractionpt_BR
dc.subjectContract Reviewpt_BR
dc.subjectLarge Language Modelspt_BR
dc.subjectNatural Language Processingpt_BR
dc.titleEvaluation of Large Language Models in Contract Information Extractionpt_BR
dc.typebachelorThesispt_BR
dc.contributor.authorLatteshttps://lattes.cnpq.br/5639027966274673pt_BR
dc.degree.levelGraduacaopt_BR
dc.contributor.advisorLatteshttp://lattes.cnpq.br/3084134533707587pt_BR
dc.subject.cnpqÁreas::Ciências Exatas e da Terra::Ciência da Computaçãopt_BR
dc.degree.departament::(CIN-DCC) - Departamento de Ciência da Computaçãopt_BR
dc.degree.graduation::CIn-Curso de Ciência da Computaçãopt_BR
dc.degree.grantorUniversidade Federal de Pernambucopt_BR
dc.degree.localRecifept_BR
Appears in Collections:(TCC) - Ciência da Computação

Files in This Item:
File Description SizeFormat 
TCC Weybson Alves da Silva.pdf3.15 MBAdobe PDFThumbnail
View/Open


This item is protected by original copyright



This item is licensed under a Creative Commons License Creative Commons