Egocentric Memory Enhanced Mixed-Session Conversation Agent (EMMA)

Introduction

EMMA is a new conversation model designed for mixed-session conversations, incorporating Egocentric Memory trained on the MiSC dataset. This model focuses on handling dynamic interactions across sessions, where a main speaker engages with different partners.

🚨 This repository is for the adapter of EMMA's dialogue module, which is based on FLAN-T5-Large.

Model Description

Citation Information

If you use EMMA in your research, please cite the following paper:

@inproceedings{jang-etal-2024-mixed,
    title = "Mixed-Session Conversation with Egocentric Memory",
    author = "Jang, Jihyoung  and
      Kim, Taeyoung  and
      Kim, Hyounghun",
    editor = "Al-Onaizan, Yaser  and
      Bansal, Mohit  and
      Chen, Yun-Nung",
    booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
    month = nov,
    year = "2024",
    address = "Miami, Florida, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.findings-emnlp.689/",
    doi = "10.18653/v1/2024.findings-emnlp.689",
    pages = "11786--11815"
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for jihyoung/EMMA

Finetuned
(202)
this model

Dataset used to train jihyoung/EMMA

Paper for jihyoung/EMMA