A quick guide to Amazon’s 30+ papers at NAACL 2024


In recent years, the fields of natural-language processing and computational linguistics, which were revolutionized a decade ago by deep learning, were revolutionized again by large language models (LLMs). Unsurprisingly, work involving LLMs, either as a subject of inquiry themselves or as tools for other natural-language-processing applications, predominates at this year’s meeting of the North American Chapter of the Association for Computational Linguistics (NAACL). This paper guide sorts Amazon’s NAACL papers into those that deal explicitly with LLMs and those that don’t — although in many cases, the ones that don’t present general techniques or datasets that could be used with either LLMs or more-traditional models.

LLM-related work

Agents

FLAP: Flow-adhering planning with constrained decoding in LLMs
Shamik Roy, Sailik Sengupta, Daniele Bonadiman, Saab Mansour, Arshit Gupta

Attribute value extraction

EIVEN: Efficient implicit attribute value extraction using multimodal LLM
Henry Peng Zou, Gavin Yu, Ziwei Fan, Dan Bu, Han Liu, Peng Dai, Dongmei Jia, Cornelia Caragea

Continual learning

Q-Tuning: Queue-based prompt tuning for lifelong few-shot language learning
Yanhui Guo, Shaoyuan Xu, Jinmiao Fu, Jia (Kevin) Liu, Chaosheng Dong, Bryan Wang

Dialogue

Leveraging LLMs for dialogue quality measurement
Jinghan Jia, Abi Komma, Timothy Leffel, Xujun Peng, Ajay Nagesh, Tamer Soliman, Aram Galstyan, Anoop Kumar

Hallucination mitigation

Less is more for improving automatic evaluation of factual consistency
Tong Wang, Ninad Kulkarni, Yanjun (Jane) Qi

TofuEval: Evaluating hallucinations of LLMs on topic-focused dialogue summarization
Liyan Tang, Igor Shalyminov, Amy Wong, Jon Burnsky, Jake Vincent, Yu’an Yang, Siffi Singh, Song Feng, Hwanjun Song, Hang Su, Justin Sun, Yi Zhang, Saab Mansour, Kathleen McKeown

Towards improved multi-source attribution for long-form answer generation
Nilay Patel, Shivashankar Subramanian, Siddhant Garg, Pratyay Banerjee, Amita Misra

Machine translation

A preference-driven paradigm for enhanced translation with large language models
Dawei Zhu, Sony Trenous, Xiaoyu Shen, Dietrich Klakow, Bill Byrne, Eva Hasler

Natural-language processing

Toward informal language processing: Knowledge of slang in large language models
Zhewei Sun, Qian Hu, Rahul Gupta, Richard Zemel, Yang Xu

Question answering

Bring your own KG: Self-supervised program synthesis for zero-shot KGQA
Dhruv Agarwal, Rajarshi (Raj) Das, Sopan Khosla, Rashmi Gangadharaiah

The universal question-answering model presented in “Bring your own KG: Self-supervised program synthesis for zero-shot KGQA” uses a three-stage method to efficiently adapt to a new knowledge graph — without any training data.

Reasoning

CoMM: Collaborative multi-agent, multi-reasoning-path prompting for complex problem solving
Pei Chen, Boran Han, Shuai Zhang

Recommender systems

RecMind: Large language model powered agent for recommendation
Yancheng Wang, Ziyan Jiang, Zheng Chen, Fan Yang, Yingxue Zhou, Eunah Cho, Xing Fan, Xiaojiang Huang, Yanbin Lu, Yingzhen Yang

Reinforcement learning from human feedback

RS-DPO: A hybrid rejection sampling and direct preference optimization method for alignment of large language models
Saeed Khaki, JinJin Li, Lan Ma, Liu Yang, Prathap Ramachandra

Responsible AI

ITERALIGN: Iterative constitutional alignment of large language models
Xiusi Chen, Hongzhi Wen, Sreyashi Nag, Chen Luo, Qingyu Yin, Ruirui Li, Zheng Li, Wei Wang

MICo: Preventative detoxification of large language models through inhibition control
Roy Siegelmann, Ninareh Mehrabi, Palash Goyal, Prasoon Goyal, Lisa Bauer, Jwala Dhamala, Aram Galstyan, Rahul Gupta, Reza Ghanadan

The steerability of large language models toward data-driven personas
Junyi Li, Charith Peris, Ninareh Mehrabi, Palash Goyal, Kai-Wei Chang, Aram Galstyan, Richard Zemel, Rahul Gupta

Retrieval-augmented generation

Enhancing contextual understanding in large language models through contrastive decoding
Zheng Zhao, Emilio Monti, Jens Lehmann, Haytham Assem

Text generation

Low-cost generation and evaluation of dictionary example sentences
Bill Cai, Clarence Ng, Daniel Tan, Shelvia Hotama

Multi-review fusion-in-context
Aviv Slobodkin, Ori Shapira, Ran Levy, Ido Dagan

Vision-language models

MAGID: An automated pipeline for generating synthetic multi-modal datasets
Hossein Aboutalebi, Justin Sun, Hwanjun Song, Yusheng Xie, Arshit Gupta, Hang Su, Igor Shalyminov, Nikolaos Pappas, Siffi Singh, Saab Mansour

Prompting vision-language models for aspect-controlled generation of referring expressions
Danfeng Guo, Sanchit Agarwal, Arpit Gupta, Jiun-Yu Kao, Emre Barut, Tagyoung Chung, Jing Huang, Mohit Bansal

General and classical techniques

Conversational agents

Leveraging interesting facts to enhance user engagement with conversational interfaces
Nikhita Vedula, Giuseppe Castellucci, Eugene Agichtein, Oleg Rokhlenko, Shervin Malmasi

Information extraction

Leveraging customer feedback for multi-modal insight extraction
Sandeep Sricharan Mukku, Abinesh Kanagarajan, Pushpendu Ghosh, Chetan Aggarwal

REXEL: An end-to-end model for document-level relation extraction and entity linking
Nacime Bouziani, Shubhi Tyagi, Joseph Fisher, Jens Lehmann, Andrea Pierleoni

Machine learning

DEED: Dynamic early exit on decoder for accelerating encoder-decoder transformer models
Peng Tang, Pengkai Zhu, Tian Li, Srikar Appalaraju, Vijay Mahadevan, R. Manmatha

Machine translation

How lexical is bilingual lexicon induction?
Harsh Kohli, Helian Feng, Nicholas Dronen, Calvin McCarter, Sina Moeini, Ali Kebarighotbi

M3T: A new benchmark dataset for multi-modal document-level machine translation
Benjamin Hsu, Xiaoyu Liu, Huayang Li, Yoshinari Fujinuma, Maria Nădejde, Xing Niu, Yair Kittenplon, Ron Litman, Raghavendra Pappagari

Responsible AI

Mitigating bias for question answering models by tracking bias influence
Mingyu Derek Ma, Jiun-Yu Kao, Arpit Gupta, Yu-Hsiang Lin, Wenbo Zhao, Tagyoung Chung, Wei Wang, Kai-Wei Chang, Nanyun Peng

Semantic retrieval

Extremely efficient online query encoding for dense retrieval
Nachshon Cohen, Yaron Fairstein, Guy Kushilevitz

Text summarization

CCSUM: A large-scale and high-quality dataset for abstractive news summarization
Xiang Jiang, Markus Dreyer

Semi-supervised dialogue abstractive summarization via high-quality pseudolabel selection
Jianfeng He, Hang Su, Jason Cai, Igor Shalyminov, Hwanjun Song, Saab Mansour

Visual question answering

Multiple-question multiple-answer text-VQA
Peng Tang, Srikar Appalaraju, R. Manmatha, Yusheng Xie, Vijay Mahadevan





Source link

We will be happy to hear your thoughts

Leave a reply

Rockstary Reviews
Logo
Shopping cart