-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathreference.bib
16 lines (16 loc) · 1.56 KB
/
reference.bib
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
@inproceedings{willemsen-skantze-2024-referring-expression,
title = "Referring Expression Generation in Visually Grounded Dialogue with Discourse-aware Comprehension Guiding",
author = "Willemsen, Bram and
Skantze, Gabriel",
editor = "Mahamood, Saad and
Minh, Nguyen Le and
Ippolito, Daphne",
booktitle = "Proceedings of the 17th International Natural Language Generation Conference",
month = sep,
year = "2024",
address = "Tokyo, Japan",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2024.inlg-main.38",
pages = "453--469",
abstract = "We propose an approach to referring expression generation (REG) in visually grounded dialogue that is meant to produce referring expressions (REs) that are both discriminative and discourse-appropriate. Our method constitutes a two-stage process. First, we model REG as a text- and image-conditioned next-token prediction task. REs are autoregressively generated based on their preceding linguistic context and a visual representation of the referent. Second, we propose the use of discourse-aware comprehension guiding as part of a generate-and-rerank strategy through which candidate REs generated with our REG model are reranked based on their discourse-dependent discriminatory power. Results from our human evaluation indicate that our proposed two-stage approach is effective in producing discriminative REs, with higher performance in terms of text-image retrieval accuracy for reranked REs compared to those generated using greedy decoding.",
}