hanqi-qi / Position-Bias-Mitigation-in-Emotion-Cause-Analysis

ACL2021:Position Bias Mitigation: A Knowledge-Aware Graph Model for EmotionCause Extraction
MIT License
21 stars 7 forks source link

关于代码文件缺失问题 #1

Open A-N-cmd opened 3 years ago

A-N-cmd commented 3 years ago

你好,最近想复现一下你的论文,按照你提供的代码跑,发现缺少一些文件,如wordCh.txt,wordEn.txt,paths.cfg,words.txt等,(顺便问一下path_data = open('').readlines()这个读取的是什么内容)如与提供,不胜感激!

hanqi-qi commented 3 years ago

Hi! The wordCh.txt and wordEn.txt are the separate words of the training corpus in Chinese and English. We first get use tokenization to get the Chinese words (wordCh.txt), then use translation API to get the wordEn.txt.

Paths.cfg is a file from another paper KagNet, links are provided in readme.md. It is used for subgraph extraction. words.txt is the word embedding file. path_data is the extracted path files in English, generated based on the above files.

A-N-cmd commented 3 years ago

非常感谢您的回复!请问一下:您说的中文语料库(training corpus in Chinese and English)具体是?还有words.txt具体是由哪个文件分词得到的呀?还有raw_paths.log文件是?最后,冒昧的咨询一下,您方便将源代码整体打包给我一下吗?拜托您了~

------------------ 原始邮件 ------------------ 发件人: @.>; 发送时间: 2021年8月12日(星期四) 下午4:52 收件人: @.>; 抄送: @.>; @.>; 主题: Re: [hanqi-qi/Position-Bias-Mitigation-in-Emotion-Cause-Analysis] 关于代码文件缺失问题 (#1)

Hi! The wordCh.txt and wordEn.txt are the separate words of the training corpus in Chinese and English. We first get use tokenization to get the Chinese words (wordCh.txt), then use translation API to get the wordEn.txt.

Paths.cfg is a file from another paper KagNet, links are provided in readme.md. It is used for subgraph extraction. words.txt is the word embedding file. path_data is the extracted path files in English, generated based on the above files.

— You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android.

hanqi-qi commented 3 years ago

非常感谢您的回复!请问一下:您说的中文语料库(training corpus in Chinese and English)具体是?还有words.txt具体是由哪个文件分词得到的呀?还有raw_paths.log文件是?最后,冒昧的咨询一下,您方便将源代码整体打包给我一下吗?拜托您了~ ------------------ 原始邮件 ------------------ 发件人: @.>; 发送时间: 2021年8月12日(星期四) 下午4:52 收件人: @.>; 抄送: @.>; @.>; 主题: Re: [hanqi-qi/Position-Bias-Mitigation-in-Emotion-Cause-Analysis] 关于代码文件缺失问题 (#1) Hi! The wordCh.txt and wordEn.txt are the separate words of the training corpus in Chinese and English. We first get use tokenization to get the Chinese words (wordCh.txt), then use translation API to get the wordEn.txt. Paths.cfg is a file from another paper KagNet, links are provided in readme.md. It is used for subgraph extraction. words.txt is the word embedding file. path_data is the extracted path files in English, generated based on the above files. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android.

Hi! Thanks for your interest in this work. I will upload these files (e.g. word.txt used as the model input) to this repository soon : )I used the package google translate to translate the original Chinese document to English. raw_paths.log is raw extracted paths (before filtering) from ConceptNet.

hanqi-qi commented 3 years ago

非常感谢您的回复!请问一下:您说的中文语料库(training corpus in Chinese and English)具体是?还有words.txt具体是由哪个文件分词得到的呀?还有raw_paths.log文件是?最后,冒昧的咨询一下,您方便将源代码整体打包给我一下吗?拜托您了~ ------------------ 原始邮件 ------------------ 发件人: @.>; 发送时间: 2021年8月12日(星期四) 下午4:52 收件人: @.>; 抄送: @.>; @.>; 主题: Re: [hanqi-qi/Position-Bias-Mitigation-in-Emotion-Cause-Analysis] 关于代码文件缺失问题 (#1) Hi! The wordCh.txt and wordEn.txt are the separate words of the training corpus in Chinese and English. We first get use tokenization to get the Chinese words (wordCh.txt), then use translation API to get the wordEn.txt. Paths.cfg is a file from another paper KagNet, links are provided in readme.md. It is used for subgraph extraction. words.txt is the word embedding file. path_data is the extracted path files in English, generated based on the above files. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android.

Hi, I have uploaded the word.txt and word_translation.txt. The generated tokens are generated by the tool "Jieba". Hope they can help.

xy-xiaotudou commented 3 years ago

Hello, can you upload wordEN.txt? I found that the version I got with Google Translate is not match with the case in the paper.

hanqi-qi commented 3 years ago

Hello, can you upload wordEN.txt? I found that the version I got with Google Translate is not match with the case in the paper.

I have already uploaded the translation results word_translation.txt and the origianal Chinese word file word.txt. Hope it can help.

Groot-tech commented 2 years ago

您好,最近读了您的文章《Position Bias Mitigation: A Knowledge-Aware Graph Model for Emotion Cause Extraction》,想学习一下这个工作,请问可以提供一份打包的源代码吗?上传的代码有一些缺失。谢谢!

xyyandxyy commented 2 years ago

Hello, can you upload wordEN.txt? I found that the version I got with Google Translate is not match with the case in the paper.

I have already uploaded the translation results word_translation.txt and the origianal Chinese word file word.txt. Hope it can help.

您好,最近读了您的文章《Position Bias Mitigation: A Knowledge-Aware Graph Model for Emotion Cause Extraction》,同样想学习一下这个工作,请问可以提供一份打包的源代码吗