File size: 1,204 Bytes
15a4366
 
 
293eb5f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
license: apache-2.0
---

# FutureTOD: Teaching Future Knowledge to Pre-trained Language Model for Task-Oriented Dialogue

We present our dialogue-pertaining model, FutureTOD, which distills future knowledge into the representation of the previous dialogue context using a self-training framework. Extensive experiments on diverse downstream dialogue tasks demonstrate the effectiveness of our model, especially its generalization, robustness, and ability to learn discriminative dialogue representations.

[This paper](https://arxiv.org/abs/2306.10315) has been accepted at the ACL 2023 Main Conference.

## Usage

We release our futuretod-base-v1.0 model here. You can use this model for downstream TOD tasks follow instructions in [FutureTOD](https://github.com/Zeng-WH/FutureTOD).

## Quotation

If you find our work helpful, please consider quoting the following papers.

```
@article{zeng2023futuretod,
  title={FutureTOD: Teaching Future Knowledge to Pre-trained Language Model for Task-Oriented Dialogue},
  author={Zeng, Weihao and He, Keqing and Wang, Yejie and Zeng, Chen and Wang, Jingang and Xian, Yunsen and Xu, Weiran},
  journal={arXiv preprint arXiv:2306.10315},
  year={2023}
}
```