| from zipfile import ZipFile, ZIP_DEFLATED |
| from shutil import rmtree |
| import json |
| import os |
| from tqdm import tqdm |
| from collections import Counter |
| from pprint import pprint |
| from nltk.tokenize import sent_tokenize, word_tokenize |
| from nltk.tokenize.treebank import TreebankWordDetokenizer |
| import re |
|
|
| topic_map = { |
| 1: "Ordinary Life", |
| 2: "School Life", |
| 3: "Culture & Education", |
| 4: "Attitude & Emotion", |
| 5: "Relationship", |
| 6: "Tourism", |
| 7: "Health", |
| 8: "Work", |
| 9: "Politics", |
| 10: "Finance" |
| } |
|
|
| act_map = { |
| 1: "inform", |
| 2: "question", |
| 3: "directive", |
| 4: "commissive" |
| } |
|
|
| emotion_map = { |
| 0: "no emotion", |
| 1: "anger", |
| 2: "disgust", |
| 3: "fear", |
| 4: "happiness", |
| 5: "sadness", |
| 6: "surprise" |
| } |
|
|
| def preprocess(): |
| original_data_dir = 'ijcnlp_dailydialog' |
| new_data_dir = 'data' |
|
|
| if not os.path.exists(original_data_dir): |
| original_data_zip = 'ijcnlp_dailydialog.zip' |
| if not os.path.exists(original_data_zip): |
| raise FileNotFoundError(f'cannot find original data {original_data_zip} in dailydialog/, should manually download ijcnlp_dailydialog.zip from http://yanran.li/files/ijcnlp_dailydialog.zip') |
| else: |
| archive = ZipFile(original_data_zip) |
| archive.extractall() |
|
|
| os.makedirs(new_data_dir, exist_ok=True) |
|
|
| dataset = 'dailydialog' |
| splits = ['train', 'validation', 'test'] |
| dialogues_by_split = {split:[] for split in splits} |
| dial2topics = {} |
| with open(os.path.join(original_data_dir, 'dialogues_text.txt')) as dialog_file, \ |
| open(os.path.join(original_data_dir, 'dialogues_topic.txt')) as topic_file: |
| for dialog, topic in zip(dialog_file, topic_file): |
| topic = int(topic.strip()) |
| dialog = dialog.replace(' __eou__ ', ' ') |
| if dialog in dial2topics: |
| dial2topics[dialog].append(topic) |
| else: |
| dial2topics[dialog] = [topic] |
|
|
| global topic_map, act_map, emotion_map |
|
|
| ontology = {'domains': {x:{'description': '', 'slots': {}} for x in topic_map.values()}, |
| 'intents': {x:{'description': ''} for x in act_map.values()}, |
| 'state': {}, |
| 'dialogue_acts': { |
| "categorical": [], |
| "non-categorical": [], |
| "binary": {} |
| }} |
|
|
| detokenizer = TreebankWordDetokenizer() |
|
|
| for data_split in splits: |
| archive = ZipFile(os.path.join(original_data_dir, f'{data_split}.zip')) |
| with archive.open(f'{data_split}/dialogues_{data_split}.txt') as dialog_file, \ |
| archive.open(f'{data_split}/dialogues_act_{data_split}.txt') as act_file, \ |
| archive.open(f'{data_split}/dialogues_emotion_{data_split}.txt') as emotion_file: |
| for dialog_line, act_line, emotion_line in tqdm(zip(dialog_file, act_file, emotion_file)): |
| if not dialog_line.strip(): |
| break |
| utts = dialog_line.decode().split("__eou__")[:-1] |
| acts = act_line.decode().split(" ")[:-1] |
| emotions = emotion_line.decode().split(" ")[:-1] |
| assert (len(utts) == len(acts) == len(emotions)), "Different turns btw dialogue & emotion & action" |
|
|
| topics = dial2topics[dialog_line.decode().replace(' __eou__ ', ' ')] |
| topic = Counter(topics).most_common(1)[0][0] |
| domain = topic_map[topic] |
| |
| dialogue_id = f'{dataset}-{data_split}-{len(dialogues_by_split[data_split])}' |
| dialogue = { |
| 'dataset': dataset, |
| 'data_split': data_split, |
| 'dialogue_id': dialogue_id, |
| 'original_id': f'{data_split}-{len(dialogues_by_split[data_split])}', |
| 'domains': [domain], |
| 'turns': [] |
| } |
|
|
| for utt, act, emotion in zip(utts, acts, emotions): |
| speaker = 'user' if len(dialogue['turns']) % 2 == 0 else 'system' |
| intent = act_map[int(act)] |
| emotion = emotion_map[int(emotion)] |
| |
| utt = ' '.join([detokenizer.detokenize(word_tokenize(s)) for s in sent_tokenize(utt)]) |
| |
| utt = utt.replace(' ’ ', "'") |
| |
| utt = re.sub('\.(?!com)(\w)', lambda x: '. '+x.group(1), utt) |
|
|
| dialogue['turns'].append({ |
| 'speaker': speaker, |
| 'utterance': utt.strip(), |
| 'utt_idx': len(dialogue['turns']), |
| 'dialogue_acts': { |
| 'binary': [{ |
| 'intent': intent, |
| 'domain': '', |
| 'slot': '' |
| }], |
| 'categorical': [], |
| 'non-categorical': [], |
| }, |
| 'emotion': emotion, |
| }) |
|
|
| ontology["dialogue_acts"]['binary'].setdefault((intent, '', ''), {}) |
| ontology["dialogue_acts"]['binary'][(intent, '', '')][speaker] = True |
|
|
| dialogues_by_split[data_split].append(dialogue) |
|
|
| ontology["dialogue_acts"]['binary'] = sorted([str({'user': speakers.get('user', False), 'system': speakers.get('system', False), 'intent':da[0],'domain':da[1], 'slot':da[2]}) for da, speakers in ontology["dialogue_acts"]['binary'].items()]) |
| dialogues = dialogues_by_split['train']+dialogues_by_split['validation']+dialogues_by_split['test'] |
| json.dump(dialogues[:10], open(f'dummy_data.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| json.dump(ontology, open(f'{new_data_dir}/ontology.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| json.dump(dialogues, open(f'{new_data_dir}/dialogues.json', 'w', encoding='utf-8'), indent=2, ensure_ascii=False) |
| with ZipFile('data.zip', 'w', ZIP_DEFLATED) as zf: |
| for filename in os.listdir(new_data_dir): |
| zf.write(f'{new_data_dir}/{filename}') |
| rmtree(original_data_dir) |
| rmtree(new_data_dir) |
| return dialogues, ontology |
|
|
|
|
| if __name__ == '__main__': |
| preprocess() |
|
|