FltLM: An Intergrated Long-Context Large Language Model for Effective Context Filtering and Understanding

Type: Preprint

Publication Date: 2024-10-09

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2410.06886

Abstract

The development of Long-Context Large Language Models (LLMs) has markedly advanced natural language processing by facilitating the process of textual data across long documents and multiple corpora. However, Long-Context LLMs still face two critical challenges: The lost in the middle phenomenon, where crucial middle-context information is likely to be missed, and the distraction issue that the models lose focus due to overly extended contexts. To address these challenges, we propose the Context Filtering Language Model (FltLM), a novel integrated Long-Context LLM which enhances the ability of the model on multi-document question-answering (QA) tasks. Specifically, FltLM innovatively incorporates a context filter with a soft mask mechanism, identifying and dynamically excluding irrelevant content to concentrate on pertinent information for better comprehension and reasoning. Our approach not only mitigates these two challenges, but also enables the model to operate conveniently in a single forward pass. Experimental results demonstrate that FltLM significantly outperforms supervised fine-tuning and retrieval-based methods in complex QA scenarios, suggesting a promising solution for more accurate and reliable long-context natural language understanding applications.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ PDF Chat FltLM: An Intergrated Long-Context Large Language Model for Effective Context Filtering and Understanding 2024 Jingyang Deng
Zhengyang Shen
Boyang Wang
Lixin Su
Suqi Cheng
Ying Nie
Junfeng Wang
Dawei Yin
Jinwen Ma
+ LooGLE: Can Long-Context Language Models Understand Long Contexts? 2023 Jiaqi Li
Mengmeng Wang
Zilong Zheng
Muhan Zhang
+ PDF Chat Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA 2024 Minzheng Wang
Longze Chen
Cheng Fu
Shengyi Liao
Xinghua Zhang
Bingli Wu
Haiyang Yu
Nan Xu
Lei Zhang
Run Luo
+ PDF Chat ALR$^2$: A Retrieve-then-Reason Framework for Long-context Question Answering 2024 Huayang Li
Pat Verga
Priyanka Sen
Bowen Yang
Vijay Viswanathan
Patrick Lewis
Taro Watanabe
Yixuan Su
+ PDF Chat LongRAG: A Dual-Perspective Retrieval-Augmented Generation Paradigm for Long-Context Question Answering 2024 Qi Zhao
Ruobing Wang
Yukuo Cen
Daren Zha
Shicheng Tan
Yuxiao Dong
Junwang Tang
+ Marathon: A Race Through the Realm of Long Context with Large Language Models 2023 Nevin L. Zhang
Yunshui Li
Ziqiang Liu
Jiaxi yang
Junhao Liu
Min Yang
+ PDF Chat LongSkywork: A Training Recipe for Efficiently Extending Context Length in Large Language Models 2024 Liang Zhao
Tianwen Wei
Liang Zeng
Cheng Cheng
Yang Liu
Peng Cheng
Lijie Wang
Chenxia Li
Xuejie Wu
Bo Zhu
+ PDF Chat LLoCO: Learning Long Contexts Offline 2024 Sijun Tan
Xiuyu Li
Shishir G. Patil
Ziyang Wu
Tianjun Zhang
Kurt Keutzer
Joseph E. Gonzalez
Raluca Ada Popa
+ PDF Chat QUITO: Accelerating Long-Context Reasoning through Query-Guided Context Compression 2024 Wenshan Wang
Yihang Wang
Yixing Fan
Huaming Liao
Jiafeng Guo
+ LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding 2023 Yushi Bai
Xin Lv
Jiajie Zhang
Hongchang Lyu
Jiankai Tang
Zhidian Huang
Zhengxiao Du
Xiao Liu
Aohan Zeng
Lei Hou
+ BAMBOO: A Comprehensive Benchmark for Evaluating Long Text Modeling Capacities of Large Language Models 2023 Zican Dong
Tianyi Tang
Junyi Li
Wayne Xin Zhao
Ji-Rong Wen
+ Unlocking Context Constraints of LLMs: Enhancing Context Efficiency of LLMs with Self-Information-Based Content Filtering 2023 Yucheng Li
+ PDF Chat Reducing Distraction in Long-Context Language Models by Focused Learning 2024 Zijun Wu
Bingyuan Liu
Ran Yan
Lei Chen
Thomas Delteil
+ Lost in the Middle: How Language Models Use Long Contexts 2024 Nelson F. Liu
Kevin Lin
John Hewitt
Ashwin Paranjape
Michele Bevilacqua
Fabio Petroni
Percy Liang
+ PDF Chat Retrieval or Global Context Understanding? On Many-Shot In-Context Learning for Long-Context Evaluation 2024 Kaijian Zou
Muhammad Khalifa
Lu Wang
+ Lost in the Middle: How Language Models Use Long Contexts 2023 Nelson F. Liu
Kevin Lin
John K. Hewitt
Ashwin Paranjape
Michele Bevilacqua
Fabio Petroni
Percy Liang
+ PDF Chat An Effective Framework to Help Large Language Models Handle Numeric-involved Long-context Tasks 2024 Yijiong Yu
+ KALM: Knowledge-Aware Integration of Local, Document, and Global Contexts for Long Document Understanding 2022 Shangbin Feng
Zhaoxuan Tan
Wenā€Qian Zhang
Zhenyu Lei
Yulia Tsvetkov
+ PDF Chat ETHIC: Evaluating Large Language Models on Long-Context Tasks with High Information Coverage 2024 Taewhoo Lee
Chanwoong Yoon
Kyochul Jang
Donghyeon Lee
Moohyun Song
Hyunjae Kim
Jaewoo Kang
+ PDF Chat Make Your LLM Fully Utilize the Context 2024 Shengnan An
Zexiong Ma
Zeqi Lin
Nanning Zheng
Jianā€“Guang Lou

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors