Layered Query Retrieval: An Adaptive Framework for Retrieval-Augmented Generation in Complex Question Answering for Large Language Models
Abstract
:Featured Application
Abstract
1. Introduction
- (1)
- This study proposes a Layered Query Retrieval framework that adaptively handles both simple and complex queries, balancing both efficiency and accuracy in responses.
- (2)
- We designed a relevance classifier to filter out irrelevant documents from the retrieved ones, preventing the large model’s responses from being influenced by too much irrelevant information.
- (3)
- We conducted experiments on multiple public datasets, and the results show that our model improves accuracy and F1 score by over 10% compared to state-of-the-art methods.
2. Related Work
3. Method
3.1. Query Complexity Identification
3.2. Adaptive Retrieval
3.2.1. Single-Step Strategy
- : A term in the query .
- : The term frequency of in document .
- : The length of the document (in terms of number of words).
- avgdl: The average document length in the corpus.
- : A tunable parameter that controls the term frequency saturation.
- : A tunable parameter that controls the influence of document length normalization.
- : The inverse document frequency of term .
- : The total number of documents in the corpus.
- : The number of documents containing the term .
3.2.2. Multi-Step Strategy
3.3. LLM Inference
4. Experiment
4.1. Dataset
4.2. Evaluation Metrics
4.3. Experiment Details
4.4. Baseline
5. Experimental Results and Analyses
5.1. Main Results
5.2. Ablation Study
- The effect of the complexity classifier
- The effect of the first-stage retriever
- The effect of different LLMs
- The effect of the relevance classifier
5.3. Case Study
6. Discussion
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A
Dataset | Query and Answer | Single-Step | Our Method |
---|---|---|---|
SQuAD | Query: Who was the CEO of GE when it announced its Ecoimagination program? Answer: Jeff Immelt | Doc 1: GE Digital across field service processes. This acquisition closed in January 2017. In November 2016, GE Digital also launched its collaborative app development program. In January, GE completed its acquisition of ServiceMax, a cloud-based field service management company. In December, GE announced its intent to establish an independent IIoT company, taking over large parts of GE Digital. At the same time, GE announced its intended sale of a majority stake of ServiceMax. In February 2016, GE announced the GE Digital Alliance Program, a program dedicated to building the digital industrial ecosystem. This new program encompassed existing agreements with Accenture, AT&T, Cisco, SoftBank, Doc 2: GE Healthcare in the Technology Research for Independent Living Centre (TRIL). The Irish centre seeks to enhance independence for elderly people through technological innovation. In July 2015, GE Healthcare partnered with the 2015 CrossFit Games to provide athletes with mobile imaging equipment. In January 2016, it was announced GE Healthcare’s global headquarters will move to Chicago effective early 2016. In June 2017, GE announced Kieran Murphy as the new CEO of GE Healthcare, with former CEO John Flannery’s appointment as CEO of GE. In April 2018, GE announced the sale of several healthcare information technology assets for $1.05 billion to Veritas Capital. Doc 3: GE Healthcare of filtration products and technologies at 270p per share in cash for each Whatman share, valuing Whatman at approximately £363 million (approximately $713 million.) In July 2008, Joseph Hogan announced his intent to leave his post as CEO of GE Healthcare to take the role of CEO at ABB. On 17 July 2008, GE Healthcare announced John Dineen had been chosen to replace outgoing CEO Joseph Hogan. Mr. Dineen had been head of GE’s Transportation division since 2005. On 24 March 2010, GE Healthcare announced acquisition of MedPlexus. In late April 2010, GE Healthcare announced it was investing €3 million | Doc 1: Wiregrass Electric Cooperative first manager of the cooperative. He served until 1958. Subsequent leadership was as follows: • B.L Woodham, general manager, 1953–1968 • Bruce Dyess, general manager, 1969–1985 • Jerry Mosley, general manager, 1985–2005 • Michael McWaters, CEO, 2005–2013 • Les Moreland, CEO, 2014–present At Wiregrass Electric Cooperative, the consumer/member is the basis of the organization. Each person receiving electric service must be a member of the cooperative. Annual membership meetings are held where members receive reports on the operation of the system and elect trustees to handle the affairs of the Doc 2: James McNerney Information Services. McNerney competed with Bob Nardelli and Jeff Immelt to succeed the retiring Jack Welch as chairman and CEO of General Electric. When Immelt won the three-way race, McNerney and Nardelli left GE (as was Welch’s plan); McNerney was hired by 3M in 2001. From 2001 to 2005, McNerney held the position as chairman of the board and CEO of 3M, a $20 billion global industrial company with leading positions in electronics, telecommunications, industrial, consumer and office products, health care, safety and other businesses. On 30 June 2005 The Boeing Company hired McNerney as the Chairman, President and CEO. Doc 3: United States withdrawal from the Paris Agreement with the United Nations to submit a plan for limiting American climate-change emissions in accord with the Paris Agreement guidelines. Goldman Sachs CEO Lloyd Blankfein described Trump’s decision as “a setback for the environment and for the U.S.’s leadership position in the world”. General Electric CEO Jeff Immelt stated that “climate change is real”. Multiple tech company executives—including Google CEO Sundar Pichai, Microsoft President and Chief Legal Officer Brad Smith, Apple CEO Tim Cook, Facebook CEO Mark Zuckerberg, and General Electric CEO Jeff Immelt—condemned the decision. Microsoft’s Satya Nadella said Microsoft believes that “climate change is an urgent issue that’, Doc 4: Robert Nardelli often known as “Little Jack”, after his mentor Jack Welch, whom Nardelli had ambitions to succeed as CEO of GE. When Jack Welch retired as chairman and CEO of GE, a lengthy and well-publicized succession planning saga ensued. Nardelli competed with James McNerney and Jeffrey R. Immelt to succeed Welch. With Immelt winning the three-way race, Nardelli and McNerney left GE (as was Welch’s plan). About 10 min after Welch let him go, Nardelli received a job offer from Kenneth Langone, who at the time was on the boards of both GE and The Home Depot. Nardelli became CEO of Doc 5: Jeff Immelt when Jack Welch asked Gary Wendt, the head of GE Financial Services, Inc, then a $41 billion empire and GE’s most profitable division, to resign, on the theory that he would overshadow Welch’s other candidates. Immelt surpassed Robert Nardelli, who was so close to Welch that his nickname was “Little Jack”, and James McNerney; Nardelli would leave GE to head Home Depot and then Chrysler, and McNerney left GE to lead 3M (before going on to lead Boeing). Immelt was selected by the GE Board in November 2000, to succeed Welch as CEO of GE; he served as GE’s president |
Natural Questions | Query: who wrote just you and me by chicago Answer: James Pankow | Doc 1: “Just You ‘n’ Me ‘n’ Me” was written after a fight between Pankow and his future wife Karen: “Just You ‘n’ Me” was the very last song played by Chicago radio station WLS before switching to a talk radio format in 1989. Just You ‘n’ Me “Just You ‘n’ Me” is a song written by James Pankow for the group Chicago and recorded for their fifth studio album “Chicago VI” (1973). The lead vocals are sung by bassist Peter Cetera. The second single released from that album, it was more successful than the first single, “Feelin’ Stronger Every Day”, reaching #4 on the U.S. Doc 2: ‘Just You ‘n’ Me Just You ‘n’ Me “Just You ‘n’ Me” is a song written by James Pankow for the group Chicago and recorded for their fifth studio album “Chicago VI” (1973). The lead vocals are sung by bassist Peter Cetera. The second single released from that album, it was more successful than the first single, “Feelin’ Stronger Every Day”, reaching #4 on the U.S. “Billboard” Hot 100 and #1 on the “Cash Box” Top 100. Walter Parazaider plays a soprano saxophone solo during the instrumental section while guitarist Terry Kath uses a wah-wah pedal and phase shifter on his guitar. “Just You” Doc 3: Chicago VI would remain their recording base for the next four years. Robert Lamm authored half of the album’s tracks, including his response to some of Chicago’s negative reviewers in “Critics’ Choice”. James Pankow wrote the album’s two hits, “Just You ‘n’ Me” (#4) and “Feelin’ Stronger Every Day” (#10). The latter was co-composed with Peter Cetera, who also wrote “In Terms of Two”, and sang lead vocal on all three songs. Released in June 1973, “Chicago VI” was another commercial success, spending five weeks at #1 in the US. The band would not chart in the UK at all until 1976’s | Doc 1: “Just You ‘n’ Me ‘n’ Me” was written after a fight between Pankow and his future wife Karen: “Just You ‘n’ Me” was the very last song played by Chicago radio station WLS before switching to a talk radio format in 1989. Just You ‘n’ Me “Just You ‘n’ Me” is a song written by James Pankow for the group Chicago and recorded for their fifth studio album “Chicago VI” (1973). The lead vocals are sung by bassist Peter Cetera. The second single released from that album, it was more successful than the first single, “Feelin’ Stronger Every Day”, reaching #4 on the U.S. Doc 2: ‘Just You ‘n’ Me Just You ‘n’ Me “Just You ‘n’ Me” is a song written by James Pankow for the group Chicago and recorded for their fifth studio album “Chicago VI” (1973). The lead vocals are sung by bassist Peter Cetera. The second single released from that album, it was more successful than the first single, “Feelin’ Stronger Every Day”, reaching #4 on the U.S. “Billboard” Hot 100 and #1 on the “Cash Box” Top 100. Walter Parazaider plays a soprano saxophone solo during the instrumental section while guitarist Terry Kath uses a wah-wah pedal and phase shifter on his guitar. “Just You” Doc 3: Chicago VI would remain their recording base for the next four years. Robert Lamm authored half of the album’s tracks, including his response to some of Chicago’s negative reviewers in “Critics’ Choice”. James Pankow wrote the album’s two hits, “Just You ‘n’ Me” (#4) and “Feelin’ Stronger Every Day” (#10). The latter was co-composed with Peter Cetera, who also wrote “In Terms of Two”, and sang lead vocal on all three songs. Released in June 1973, “Chicago VI” was another commercial success, spending five weeks at #1 in the US. The band would not chart in the UK at all until 1976’s |
TriviaQA | Query: Which studio owns the rights to Bugs Bunny? Answer: Warner bros | Doc 1: Bugs Bunny’s Easter Special\nin an Easter bunny suit (finally getting it right), which neither surprises nor disappoints Bugs or Granny, having known all along it was Daffy. This special was released on DVD in 2011, in time for the Easter holiday. Bugs Bunny’s Easter Special Bugs Bunny’s Easter Special (also known as “The Bugs Bunny Easter Special” and “Bugs Bunny’s Easter Funnies”) is a “Looney Tunes” television special featuring a number of Warner Bros. cartoons. It originally debuted on the CBS network on 7 April 1977. The Easter bunny is ill, Granny needs to find a replacement for him and suggests Bugs Bunny. Doc 2: Bugs Bunny & Lola Bunny: Operation Carrot Patch\nBugs Bunny & Lola Bunny: Operation Carrot Patch Bugs Bunny & Lola Bunny: Operation Carrot Patch, known in North America as Looney Tunes: Carrot Crazy, is a 1998 Game Boy Color video game starring Bugs Bunny and Lola Bunny. Yosemite Sam, Daffy Duck, The Tasmanian Devil, Marvin the Martian and Elmer Fudd have stolen Bugs and Lola’s carrots and hid them in different sets in the nearby studio. Bugs and Lola head through five different “sets” in order to reclaim them. In the game, Bugs and Lola (the player can switch between them at any time) can traverse two levels Doc 3: Bugs Bunny He would not appear in new material on-screen again until “Bugs and Daffy’s Carnival of the Animals” aired in 1976. From the late 1970s through the early 1990s, Bugs was featured in various animated specials for network television, such as “Bugs Bunny’s Thanksgiving Diet”, “Bugs Bunny’s Easter Special”, “Bugs Bunny’s Looney Christmas Tales”, and “Bugs Bunny’s Bustin’ Out All Over”. Bugs also starred in several theatrical compilation features during this time, including the United Artists distributed documentary (1975) and Warner Bros.’ own releases: “The Bugs Bunny/Road Runner Movie” (1979), “The Looney Looney Looney Bugs Bunny Movie” (1981), (1982) | Doc 1: Bugs Bunny’s Easter Special in an Easter bunny suit (finally getting it right), which neither surprises nor disappoints Bugs or Granny, having known all along it was Daffy. This special was released on DVD in 2011, in time for the Easter holiday. Bugs Bunny’s Easter Special Bugs Bunny’s Easter Special (also known as “The Bugs Bunny Easter Special” and “Bugs Bunny’s Easter Funnies”) is a “Looney Tunes” television special featuring a number of Warner Bros. cartoons. It originally debuted on the CBS network on 7 April 1977. The Easter bunny is ill, Granny needs to find a replacement for him and suggests Bugs Bunny. Doc 2: Bugs Bunny & Lola Bunny: Operation Carrot Patch Bugs Bunny & Lola Bunny: Operation Carrot Patch Bugs Bunny & Lola Bunny: Operation Carrot Patch, known in North America as Looney Tunes: Carrot Crazy, is a 1998 Game Boy Color video game starring Bugs Bunny and Lola Bunny. Yosemite Sam, Daffy Duck, The Tasmanian Devil, Marvin the Martian and Elmer Fudd have stolen Bugs and Lola’s carrots and hid them in different sets in the nearby studio. Bugs and Lola head through five different “sets” in order to reclaim them. In the game, Bugs and Lola (the player can switch between them at any time) can traverse two levels Doc 3: Bugs Bunny He would not appear in new material on-screen again until “Bugs and Daffy’s Carnival of the Animals” aired in 1976. From the late 1970s through the early 1990s, Bugs was featured in various animated specials for network television, such as “Bugs Bunny’s Thanksgiving Diet”, “Bugs Bunny’s Easter Special”, “Bugs Bunny’s Looney Christmas Tales”, and “Bugs Bunny’s Bustin’ Out All Over”. Bugs also starred in several theatrical compilation features during this time, including the United Artists distributed documentary (1975) and Warner Bros.’ own releases: “The Bugs Bunny/Road Runner Movie” (1979), “The Looney Looney Looney Bugs Bunny Movie” (1981), (1982) |
Dataset | Query and Answer | Single-Step | Our Method |
---|---|---|---|
MusiQue | Query: What field does the author of Parallel Worlds work in? Answer: physics | Doc 1: Declaration of war by the United States For the United States, Article One, Section Eight of the Constitution says Congress shall have power to... declare War. ‘‘However, that passage provides no specific format for what form legislation must have in order to be considered a “declaration of war” nor does the Constitution itself use this term. In the courts, the United States Court of Appeals for the First Circuit, in Doe v. Bush, said: The text of the October Resolution itself spells out justifications for a war and frames itself as an ‘authorization’ of such a war. in effect saying an authorization suffices for declaration and what some may view as a formal Congressional “Declaration of War” was not required by the Constitution. Doc 2: Timoshenko Medal The Timoshenko Medal, widely regarded as the highest international award in the field of applied mechanics, was established in 1957 in honor of Stephen Timoshenko, world-renowned authority in the field. The Medal commemorates his contributions as author and teacher. | Doc 1: Parallel Worlds (Dave Douglas album) Parallel Worlds is the debut album by trumpeter Dave Douglas released on the Italian Soul Note label in 1993. It features six of Douglas’ compositions and compositions by Anton Webern, Kurt Weill, Duke Ellington and Igor Stravinsky performed by Douglas, Mark Feldman, Erik Friedlander, Mark Dresser and Michael Sarin, Doc 2: Parallel Worlds (book) Parallel Worlds: A Journey Through Creation, Higher Dimensions, and the Future of the Cosmos is a popular science book by Michio Kaku first published in 2004. Doc 3: José Antunes Sobrinho Brasília has also been the focus of modern-day literature. Published in 2008, The World In Grey: Dom Bosco’s Prophecy, by author Ryan J. Lucero, tells an apocalypticle story based on the famous prophecy from the late 19th century by the Italian saint Don Bosco. According to Don Bosco’s prophecy: “Between parallels 15 and 20, around a lake which shall be formed; A great civilization will thrive, and that will be the Promised Land”. Brasília lies between the parallels 15° S and 20° S, where an artificial lake (Paranoá Lake) was formed. Don Bosco is Brasília’s patron saint. Doc 4: Don Bosco (author) Don Bosco (born in June 1971) is a writer and publisher of fiction books from Singapore. In 2011, he founded Super Cool Books, a publishing company with interest in fantasy and mystery stories for children and young adults. As an active advocate of self-publishing and digital publishing, he has published ebooks, paperbacks and an iPad app. |
HotpotQA | Query: Gerd Neggo trained under the founder of which type of dance analysis? Answer: Laban Movement Analysis | Doc 1: Person Analysis Person Analysis is a phase of training needs analysis directed at identifying which individuals within an organization should receive training. Doc 2: Gerd Neggo Gerd Neggo (9 November 1891–1 September 1974) was an Estonian dancer, dance teacher and choreographer. She studied the musical response methods of É. Jaques-Dalcroze, trained under Rudolf von Laban in Hamburg, Germany, and in 1924 established her own dance studio at Tallinn, Estonia, and promoted modern dance and mime based on classical ballet. During the Soviet occupation of Estonia, she and her husband Paul Olak migrated to Sweden. Her contributions to the cultural heritage of Estonia, as the founder of modern dance and mime in her country, is recognised via a scholarship, awarded annually since 2011. Doc 3: Joyce K. Paul Joyce K. Paul is an Indian classical dancer and exercise physiologist. Classically trained in Bharat Natyam, she has also trained briefly in Mohiniattam, the classical dance of Kerala. Paul had her initial dance training under the Bharatanatyam guru, Leela Samson. She is the founder and the Artistic Director of Arpan set up in 2003 in Redmond, WA | Doc 1: Person Analysis Person Analysis is a phase of training needs analysis directed at identifying which individuals within an organization should receive training. Doc 2: Gerd Neggo Gerd Neggo (9 November 1891–1 September 1974) was an Estonian dancer, dance teacher and choreographer. She studied the musical response methods of É. Jaques-Dalcroze, trained under Rudolf von Laban in Hamburg, Germany, and in 1924 established her own dance studio at Tallinn, Estonia, and promoted modern dance and mime based on classical ballet. During the Soviet occupation of Estonia, she and her husband Paul Olak migrated to Sweden. Her contributions to the cultural heritage of Estonia, as the founder of modern dance and mime in her country, is recognised via a scholarship, awarded annually since 2011. Doc 3: Joyce K. Paul Joyce K. Paul is an Indian classical dancer and exercise physiologist. Classically trained in Bharat Natyam, she has also trained briefly in Mohiniattam, the classical dance of Kerala. Paul had her initial dance training under the Bharatanatyam guru, Leela Samson. She is the founder and the Artistic Director of Arpan set up in 2003 in Redmond, WA |
2Wiki | Query: Which country the founder of Team Red Bull (Nascar Team) is from? Answer: Austria | Doc 1: Team Red Bull (NASCAR team) Red Bull Racing Team, also known as Team Red Bull, was a NASCAR team owned by Red Bull founders Dietrich Mateschitz and Chaleo Yoovidhya. The team was based in Mooresville, North Carolina in the United States and was managed by Jay Frye. The team suspended operations on 8 December 2011 and their cars were sold to BK Racing. Doc 2: Helmut Marko Helmut Marko LL.D., (born 27 April 1943) is an Austrian former professional racing driver and current advisor to the Red Bull GmbH Formula One Teams and head of Red Bull’s driver development program. Doc 3: Red Bull GmbH Red Bull GmbH is an Austrian company which produces Red Bull energy drink. The company is also known for its sponsorship of a range of sporting events and teams. In 2018, a total of 6.8 billion cans of Red Bull were sold worldwide in over 171 countries. The headquarters of Red Bull GmbH are located in Fuschl am See, Austria. | Doc 1: Team Red Bull (NASCAR team) Red Bull Racing Team, also known as Team Red Bull, was a NASCAR team owned by Red Bull founders Dietrich Mateschitz and Chaleo Yoovidhya. The team was based in Mooresville, North Carolina in the United States and was managed by Jay Frye. The team suspended operations on 8 December 2011 and their cars were sold to BK Racing. Doc 2: Red Bull GmbH Red Bull GmbH is an Austrian company which produces Red Bull energy drink. The company is also known for its sponsorship of a range of sporting events and teams. In 2018, a total of 6.8 billion cans of Red Bull were sold worldwide in over 171 countries. The headquarters of Red Bull GmbH are located in Fuschl am See, Austria. Doc 3: Helmut Marko Helmut Marko LL.D., (born 27 April 1943) is an Austrian former professional racing driver and current advisor to the Red Bull GmbH Formula One Teams and head of Red Bull’s driver development program. |
References
- Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; Anadkat, S. Gpt-4 technical report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
- Dubey, A.; Jauhri, A.; Pandey, A.; Kadian, A.; Al-Dahle, A.; Letman, A.; Mathur, A.; Schelten, A.; Yang, A.; Fan, A. The llama 3 herd of models. arXiv 2024, arXiv:2407.21783. [Google Scholar]
- Team, G.; Anil, R.; Borgeaud, S.; Wu, Y.; Alayrac, J.-B.; Yu, J.; Soricut, R.; Schalkwyk, J.; Dai, A.M.; Hauth, A. Gemini: A family of highly capable multimodal models. arXiv 2023, arXiv:2312.11805. [Google Scholar]
- Kandpal, N.; Deng, H.; Roberts, A.; Wallace, E.; Raffel, C. Large language models struggle to learn long-tail knowledge. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 15696–15707. [Google Scholar]
- Gao, Y.; Xiong, Y.; Gao, X.; Jia, K.; Pan, J.; Bi, Y.; Dai, Y.; Sun, J.; Wang, H. Retrieval-augmented generation for large language models: A survey. arXiv 2023, arXiv:2312.10997. [Google Scholar]
- Zhang, Y.; Li, Y.; Cui, L.; Cai, D.; Liu, L.; Fu, T.; Huang, X.; Zhao, E.; Zhang, Y.; Chen, Y. Siren’s song in the AI ocean: A survey on hallucination in large language models. arXiv 2023, arXiv:2309.01219. [Google Scholar]
- Lewis, P.; Perez, E.; Piktus, A.; Petroni, F.; Karpukhin, V.; Goyal, N.; Küttler, H.; Lewis, M.; Yih, W.-t.; Rocktäschel, T. Retrieval-augmented generation for knowledge-intensive nlp tasks. Adv. Neural Inf. Process. Syst. 2020, 33, 9459–9474. [Google Scholar]
- Li, H.; Su, Y.; Cai, D.; Wang, Y.; Liu, L. A survey on retrieval-augmented text generation. arXiv 2022, arXiv:2202.01110. [Google Scholar]
- Izacard, G.; Lewis, P.; Lomeli, M.; Hosseini, L.; Petroni, F.; Schick, T.; Dwivedi-Yu, J.; Joulin, A.; Riedel, S.; Grave, E. Atlas: Few-shot learning with retrieval augmented language models. J. Mach. Learn. Res. 2023, 24, 1–43. [Google Scholar]
- Tang, Y.; Yang, Y. Multihop-rag: Benchmarking retrieval-augmented generation for multi-hop queries. arXiv 2024, arXiv:2401.15391. [Google Scholar]
- Lazaridou, A.; Gribovskaya, E.; Stokowiec, W.; Grigorev, N. Internet-augmented language models through few-shot prompting for open-domain question answering. arXiv 2022, arXiv:2203.05115. [Google Scholar]
- Ram, O.; Levine, Y.; Dalmedigos, I.; Muhlgay, D.; Shashua, A.; Leyton-Brown, K.; Shoham, Y. In-context retrieval-augmented language models. Trans. Assoc. Comput. Linguist. 2023, 11, 1316–1331. [Google Scholar] [CrossRef]
- Press, O.; Zhang, M.; Min, S.; Schmidt, L.; Smith, N.A.; Lewis, M. Measuring and narrowing the compositionality gap in language models. arXiv 2022, arXiv:2210.03350. [Google Scholar]
- Trivedi, H.; Balasubramanian, N.; Khot, T.; Sabharwal, A. Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions. arXiv 2022, arXiv:2212.10509. [Google Scholar]
- Vu, T.; Iyyer, M.; Wang, X.; Constant, N.; Wei, J.; Wei, J.; Tar, C.; Sung, Y.-H.; Zhou, D.; Le, Q. Freshllms: Refreshing large language models with search engine augmentation. arXiv 2023, arXiv:2310.03214. [Google Scholar]
- Zhao, P.; Zhang, H.; Yu, Q.; Wang, Z.; Geng, Y.; Fu, F.; Yang, L.; Zhang, W.; Cui, B. Retrieval-augmented generation for ai-generated content: A survey. arXiv 2024, arXiv:2402.19473. [Google Scholar]
- Liu, Y.; Peng, X.; Zhang, X.; Liu, W.; Yin, J.; Cao, J.; Du, T. RA-ISF: Learning to Answer and Understand from Retrieval Augmentation via Iterative Self-Feedback. arXiv 2024, arXiv:2403.06840. [Google Scholar]
- Yoran, O.; Wolfson, T.; Ram, O.; Berant, J. Making retrieval-augmented language models robust to irrelevant context. arXiv 2023, arXiv:2310.01558. [Google Scholar]
- Zhang, T.; Patil, S.G.; Jain, N.; Shen, S.; Zaharia, M.; Stoica, I.; Gonzalez, J.E. Raft: Adapting language model to domain specific rag. arXiv 2024, arXiv:2403.10131. [Google Scholar]
- Shao, Z.; Gong, Y.; Shen, Y.; Huang, M.; Duan, N.; Chen, W. Enhancing retrieval-augmented large language models with iterative retrieval-generation synergy. arXiv 2023, arXiv:2305.15294. [Google Scholar]
- Huang, J.; Ping, W.; Xu, P.; Shoeybi, M.; Chang, K.C.-C.; Catanzaro, B. Raven: In-context learning with retrieval augmented encoder-decoder language models. arXiv 2023, arXiv:2308.07922. [Google Scholar]
- Yu, W.; Zhang, H.; Pan, X.; Ma, K.; Wang, H.; Yu, D. Chain-of-note: Enhancing robustness in retrieval-augmented language models. arXiv 2023, arXiv:2311.09210. [Google Scholar]
- Jeong, S.; Baek, J.; Cho, S.; Hwang, S.J.; Park, J.C. Adaptive-rag: Learning to adapt retrieval-augmented large language models through question complexity. arXiv 2024, arXiv:2403.14403. [Google Scholar]
- Rajpurkar, P. Squad: 100,000+ questions for machine comprehension of text. arXiv 2016, arXiv:1606.05250. [Google Scholar]
- Kwiatkowski, T.; Palomaki, J.; Redfield, O.; Collins, M.; Parikh, A.; Alberti, C.; Epstein, D.; Polosukhin, I.; Devlin, J.; Lee, K. Natural questions: A benchmark for question answering research. Trans. Assoc. Comput. Linguist. 2019, 7, 453–466. [Google Scholar] [CrossRef]
- Joshi, M.; Choi, E.; Weld, D.; Zettlemoyer, L. TriviaQA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1601–1611. [Google Scholar]
- Trivedi, H.; Balasubramanian, N.; Khot, T.; Sabharwal, A. ♫ MuSiQue: Multihop Questions via Single-hop Question Composition. Trans. Assoc. Comput. Linguist. 2022, 10, 539–554. [Google Scholar] [CrossRef]
- Yang, Z.; Qi, P.; Zhang, S.; Bengio, Y.; Cohen, W.; Salakhutdinov, R.; Manning, C.D. HotpotQA: A Dataset for Diverse, Explainable Multi-hop Question Answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2369–2380. [Google Scholar]
- Ho, X.; Nguyen, A.-K.D.; Sugawara, S.; Aizawa, A. Constructing a multi-hop QA dataset for comprehensive evaluation of reasoning steps. arXiv 2020, arXiv:2011.01060. [Google Scholar]
- Beltagy, I.; Peters, M.E.; Cohan, A. Longformer: The long-document transformer. arXiv 2020, arXiv:2004.05150. [Google Scholar]
- Mallen, A.; Asai, A.; Zhong, V.; Das, R.; Khashabi, D.; Hajishirzi, H. When Not to Trust Language Models: Investigating Effectiveness of Parametric and Non-Parametric Memories. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics, Toronto, ON, Canada, 9–14 July 2023; pp. 9802–9822. [Google Scholar]
- Asai, A.; Wu, Z.; Wang, Y.; Sil, A.; Hajishirzi, H. Self-rag: Learning to retrieve, generate, and critique through self-reflection. arXiv 2023, arXiv:2310.11511. [Google Scholar]
Category | RAG Name | Limitations |
---|---|---|
Regular RAG | RAG |
|
Knowledge-filtered RAG | RAFT, IRCoT |
|
RAG adaptive to query complexity | Adaptive-RAG |
|
Type | Method | MuSiQue | HotpotQA | 2Wiki | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
F1 | Acc | Step | Time | F1 | Acc | Step | Time | F1 | Acc | Step | Time | ||
Simple | Non-Retrieval | 10.70 | 3.20 | 0.00 | 0.11 | 22.71 | 17.20 | 0.00 | 0.11 | 32.04 | 27.80 | 0.00 | 0.10 |
Single-Step Retrieval | 22.80 | 15.20 | 1.00 | 1.00 | 46.15 | 36.40 | 1.00 | 1.00 | 47.90 | 42.80 | 1.00 | 1.00 | |
Complex | Multi-Step Retrieval | 31.90 | 25.80 | 3.60 | 7.58 | 56.54 | 47.00 | 5.53 | 9.38 | 58.85 | 55.40 | 4.17 | 7.37 |
Adaptive | Adaptive Retrieval | 15.80 | 8.00 | 0.50 | 0.55 | 32.22 | 25.00 | 0.50 | 0.55 | 39.44 | 34.20 | 0.50 | 0.55 |
Self-RAG | 8.10 | 12.00 | 0.73 | 0.51 | 17.53 | 29.60 | 0.73 | 0.45 | 19.59 | 38.80 | 0.93 | 0.49 | |
Adaptive-RAG | 31.80 | 26.00 | 3.22 | 6.61 | 53.82 | 44.40 | 3.55 | 5.99 | 49.75 | 46.40 | 2.63 | 4.68 | |
Our Method | 41.97 | 26.55 | 3.88 | 4.45 | 69.96 | 53.8 | 3.42 | 3.92 | 54.65 | 37.60 | 3.88 | 4.32 |
Type | Method | SQuAD | Natural Questions | TriviaQA | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
F1 | Acc | Step | Time | F1 | Acc | Step | Time | F1 | Acc | Step | Time | ||
Simple | Non-Retrieval | 10.50 | 5.00 | 0.00 | 0.11 | 19.00 | 15.60 | 0.00 | 0.13 | 31.80 | 27.00 | 0.00 | 0.13 |
Single-Step Retrieval | 39.30 | 34.00 | 1.00 | 1.00 | 47.30 | 44.60 | 1.00 | 1.00 | 62.40 | 60.20 | 1.00 | 1.00 | |
Complex | Multi-Step Retrieval | 35.60 | 29.60 | 4.52 | 9.03 | 47.80 | 44.20 | 5.04 | 10.18 | 62.40 | 60.20 | 5.28 | 9.22 |
Adaptive | Adaptive Retrieval | 23.10 | 17.60 | 0.50 | 0.55 | 36.00 | 33.00 | 0.50 | 0.56 | 46.90 | 42.60 | 0.50 | 0.56 |
Self-RAG | 11.20 | 18.40 | 0.63 | 0.50 | 39.00 | 33.60 | 0.63 | 0.17 | 29.30 | 57.00 | 0.68 | 0.45 | |
Adaptive-RAG | 38.30 | 33.00 | 1.37 | 2.02 | 47.30 | 44.60 | 1.00 | 1.00 | 60.70 | 58.20 | 1.23 | 1.54 | |
Our Method | 46.50 | 43.40 | 3.73 | 4.30 | 60.16 | 58.8 | 3.24 | 3.41 | 70.20 | 68.80 | 3.84 | 4.50 |
Top-k | F1 | Acc | Step | Time |
---|---|---|---|---|
3 | 51.64 | 34.80 | 3.43 | 4.30 |
4 | 47.06 | 31.05 | 4.78 | 5.78 |
5 | 46.46 | 30.26 | 6.78 | 8.56 |
Model | MuSiQue | HotpotQA | 2Wiki | |||
---|---|---|---|---|---|---|
F1 | Acc | F1 | Acc | F1 | Acc | |
Llama-8B | 41.97 | 26.55 | 69.96 | 53.80 | 54.65 | 37.60 |
GPT4o | 59.94 | 42.80 | 79.37 | 68.60 | 59.55 | 42.40 |
Model | MuSiQue | HotpotQA | 2Wiki | |||
---|---|---|---|---|---|---|
F1 | Acc | F1 | Acc | F1 | Acc | |
Ours | 41.97 | 26.55 | 69.96 | 53.80 | 54.65 | 37.60 |
w/o rcls | 37.58 | 23.13 | 67.63 | 52.60 | 54.22 | 37.20 |
Dataset | Query and Answer | Llama3-8B | GPT-4o | Complexity Classification |
---|---|---|---|---|
SQuAD | Query: Who was the CEO of GE when it announced its Ecoimagination program? Answer: Jeff Immelt | Jeff Immelt | Jeff Immelt | Complex Q1: When did GE announce its Ecoimagination program? Q2: Who was the CEO of GE at that time? |
Natural Questions | Query: who wrote just you and me by chicago Answer: James Pankow | James Pankow | James Pankow | Simple |
TriviaQA | Query: Which studio owns the rights to Bugs Bunny? Answer: Warner Bros | Warner Bros | Warner Bros | Simple |
Dataset | Query and Answer | Llama3-8B | GPT-4o | Complexity Classification |
---|---|---|---|---|
MusiQue | Query: What field does the author of Parallel Worlds work in? Answer: physics | Physics | Science | Complex Q1: Who is the author of Parallel Worlds? Q2: What field does the author of Parallel Worlds work in? |
HotpotQA | Query: Gerd Neggo trained under the founder of which type of dance analysis? Answer: Laban Movement Analysis | Laban | Laban Movement Analysis | Simple |
2Wiki | Query: Which country the founder of Team Red Bull (Nascar Team) is from? Answer: Austria | Austrian | Austria | Complex Q1: Who founded the Team Red Bull NASCAR team? Q2: What country is the founder of the Team Red Bull NASCAR team from? |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Huang, J.; Wang, M.; Cui, Y.; Liu, J.; Chen, L.; Wang, T.; Li, H.; Wu, J. Layered Query Retrieval: An Adaptive Framework for Retrieval-Augmented Generation in Complex Question Answering for Large Language Models. Appl. Sci. 2024, 14, 11014. https://doi.org/10.3390/app142311014
Huang J, Wang M, Cui Y, Liu J, Chen L, Wang T, Li H, Wu J. Layered Query Retrieval: An Adaptive Framework for Retrieval-Augmented Generation in Complex Question Answering for Large Language Models. Applied Sciences. 2024; 14(23):11014. https://doi.org/10.3390/app142311014
Chicago/Turabian StyleHuang, Jie, Mo Wang, Yunpeng Cui, Juan Liu, Li Chen, Ting Wang, Huan Li, and Jinming Wu. 2024. "Layered Query Retrieval: An Adaptive Framework for Retrieval-Augmented Generation in Complex Question Answering for Large Language Models" Applied Sciences 14, no. 23: 11014. https://doi.org/10.3390/app142311014
APA StyleHuang, J., Wang, M., Cui, Y., Liu, J., Chen, L., Wang, T., Li, H., & Wu, J. (2024). Layered Query Retrieval: An Adaptive Framework for Retrieval-Augmented Generation in Complex Question Answering for Large Language Models. Applied Sciences, 14(23), 11014. https://doi.org/10.3390/app142311014