GENERATION2 [논문] Compressing Context to Enhance Inference Efficiency of Large Language Models Compressing Context to Enhance Inference Efficiency of Large Language ModelsYucheng Li, Bo Dong, Frank Guerin, Chenghua Lin. Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing. 2023.aclanthology.orgAbstractLarge language models (LLMs) achieved remarkable performance across various tasks. However, they face challenges in managing long documents and extended con.. 2025. 1. 9. [논문] Learning to Filter Context for Retrieval-Augmented Generation Learning to Filter Context for Retrieval-Augmented GenerationOn-the-fly retrieval of relevant knowledge has proven an essential element of reliable systems for tasks such as open-domain question answering and fact verification. However, because retrieval systems are not perfect, generation models are required to genarxiv.orgAbstractOn-the-fly retrieval of relevant knowledge has proven an essenti.. 2025. 1. 9. 이전 1 다음 반응형