File size: 2,353 Bytes
74cd113
 
 
 
 
 
 
 
3068fed
8234588
 
 
 
 
 
3068fed
 
8234588
fbb22f1
 
 
 
 
8234588
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2100f12
8234588
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: apache-2.0
datasets:
- Nuo97/Dolphin-DPO
language:
- zh
metrics:
- bleu
pipeline_tag: question-answering
---

# COMEDY: COmpressive Memory-Enhanced Dialogue sYstems framework.

Github: https://github.com/nuochenpku/COMEDY

Paper: https://arxiv.org/abs/2402.11975.pdf


<br>
<div align="center">
  <img src="comedy.png" width="40%" title="Introduction Figure">
</div>

### Task: Long-Term Conversation Dialogue Generation

Different from previous retrieval-based methods, COMEDY doesn't rely on any **retrieval module or database**. 

Instead, COMEDY adopts a groundbreaking ''**One-for-All**'' approach, utilizing a single, unified model to manage the entire process from memory generation, compression to final response generation for long-term memory dialogue generation.


 - COMEDY firstly involves distilling session-specific memory from past dialogues, encompassing fine-grained session summaries, including event recaps, and detailed user and bot portraits; 
 
 - In a break from traditional systems, COMEDY eschews the use of a memory database for storing these insights. Instead, it reprocesses and condenses memories from all past interactions, forming a *Compressive Memory*: The first part is the **concise events** that have occurred throughout all the conversations, creating a historical narrative that the system can draw upon. The second and third parts consist of a **detailed user profile** and the **dynamic relationship changes** between the user and chatbot across sessions, both derived from past conversational events.

- Finally, COMEDY skillfully integrates this compressive memory into ongoing conversations, enabling contextually memory-enhanced interactions.


### Training Dataset 

**Dolphin**, the biggest Chinese long-term conversation dataset, from actual online user-chatbot interactions. 

This dataset contains three tasks: 

**Session-Level Memory Summarization**;

**Memory Compression**;

**Memory-Grounded Response Generation**, 

comprising an extensive collection of 100k samples.

Dolphin is available at [**Dolphin**](https://huggingface.co/datasets/Nuo97/Dolphin-DPO)

### Training Strategy

Our training strategies include two stages: Mixed-task training and DPO Alignment.

<br>
<div align="center">
  <img src="training_strategy.png" width="100%" title="Introduction Figure">
</div>