Paper Group NANR 100
Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures. Analysis of English Spelling Errors in a Word-Typing Game. Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES). Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar …
Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures
Title | Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures |
Authors | Yasutomo Kimura, Keiichi Takamaru, Takuma Tanaka, Akio Kobayashi, Hiroki Sakaji, Yuzu Uchida, Hokuto Ototake, Shigeru Masuyama |
Abstract | This paper describes a Japanese political corpus created for interdisciplinary political research. The corpus contains the local assembly minutes of 47 prefectures from April 2011 to March 2015. This four-year period coincides with the term of office for assembly members in most autonomies. We analyze statistical data, such as the number of speakers, characters, and words, to clarify the characteristics of local assembly minutes. In addition, we identify problems associated with the different web services used by the autonomies to make the minutes available to the public. |
Tasks | |
Published | 2016-12-01 |
URL | https://www.aclweb.org/anthology/W16-5410/ |
https://www.aclweb.org/anthology/W16-5410 | |
PWC | https://paperswithcode.com/paper/creating-japanese-political-corpus-from-local |
Repo | |
Framework | |
Analysis of English Spelling Errors in a Word-Typing Game
Title | Analysis of English Spelling Errors in a Word-Typing Game |
Authors | Ryuichi Tachibana, Mamoru Komachi |
Abstract | The emergence of the web has necessitated the need to detect and correct noisy consumer-generated texts. Most of the previous studies on English spelling-error extraction collected English spelling errors from web services such as Twitter by using the edit distance or from input logs utilizing crowdsourcing. However, in the former approach, it is not clear which word corresponds to the spelling error, and the latter approach requires an annotation cost for the crowdsourcing. One notable exception is Rodrigues and Rytting (2012), who proposed to extract English spelling errors by using a word-typing game. Their approach saves the cost of crowdsourcing, and guarantees an exact alignment between the word and the spelling error. However, they did not assert whether the extracted spelling error corpora reflect the usual writing process such as writing a document. Therefore, we propose a new correctable word-typing game that is more similar to the actual writing process. Experimental results showed that we can regard typing-game logs as a source of spelling errors. |
Tasks | |
Published | 2016-05-01 |
URL | https://www.aclweb.org/anthology/L16-1060/ |
https://www.aclweb.org/anthology/L16-1060 | |
PWC | https://paperswithcode.com/paper/analysis-of-english-spelling-errors-in-a-word |
Repo | |
Framework | |
Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES)
Title | Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES) |
Authors | |
Abstract | |
Tasks | |
Published | 2016-12-01 |
URL | https://www.aclweb.org/anthology/W16-4300/ |
https://www.aclweb.org/anthology/W16-4300 | |
PWC | https://paperswithcode.com/paper/proceedings-of-the-workshop-on-computational |
Repo | |
Framework | |
Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction
Title | Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction |
Authors | Hiroshi Noji, Yusuke Miyao, Mark Johnson |
Abstract | |
Tasks | |
Published | 2016-11-01 |
URL | https://www.aclweb.org/anthology/D16-1004/ |
https://www.aclweb.org/anthology/D16-1004 | |
PWC | https://paperswithcode.com/paper/using-left-corner-parsing-to-encode-universal |
Repo | |
Framework | |
Large-Scale Semantic Indexing and Question Answering in Biomedicine
Title | Large-Scale Semantic Indexing and Question Answering in Biomedicine |
Authors | Eirini Papagiannopoulou, Yiannis Papanikolaou, Dimitris Dimitriadis, Sakis Lagopoulos, Grigorios Tsoumakas, Manos Laliotis, Nikos Markantonatos, Ioannis Vlahavas |
Abstract | |
Tasks | Question Answering |
Published | 2016-08-01 |
URL | https://www.aclweb.org/anthology/W16-3107/ |
https://www.aclweb.org/anthology/W16-3107 | |
PWC | https://paperswithcode.com/paper/large-scale-semantic-indexing-and-question |
Repo | |
Framework | |
KSAnswer: Question-answering System of Kangwon National University and Sogang University in the 2016 BioASQ Challenge
Title | KSAnswer: Question-answering System of Kangwon National University and Sogang University in the 2016 BioASQ Challenge |
Authors | Hyeon-gu Lee, Minkyoung Kim, Harksoo Kim, Juae Kim, Sunjae Kwon, Jungyun Seo, Yi-reun Kim, Jung-Kyu Choi |
Abstract | |
Tasks | Language Modelling, Morphological Analysis, Question Answering |
Published | 2016-08-01 |
URL | https://www.aclweb.org/anthology/W16-3106/ |
https://www.aclweb.org/anthology/W16-3106 | |
PWC | https://paperswithcode.com/paper/ksanswer-question-answering-system-of-kangwon |
Repo | |
Framework | |
Proceedings of the 19th Annual Conference of the European Association for Machine Translation
Title | Proceedings of the 19th Annual Conference of the European Association for Machine Translation |
Authors | |
Abstract | |
Tasks | Machine Translation |
Published | 2016-01-01 |
URL | https://www.aclweb.org/anthology/W16-3400/ |
https://www.aclweb.org/anthology/W16-3400 | |
PWC | https://paperswithcode.com/paper/proceedings-of-the-19th-annual-conference-of-1 |
Repo | |
Framework | |
Multiple birth least squares support vector machine for multi-class classification
Title | Multiple birth least squares support vector machine for multi-class classification |
Authors | Su-Gen Chen1, 2•Xiao-Jun Wu1 |
Abstract | Least squares twin support vector machine (LSTSVM) was initially designed for binary classification. However, practical problems often require the discrimina-tion more than two categories. To tackle multi-class classi-fication problem, a novel algorithm, called multiple birth least squares support vector machine (MBLSSVM), is proposed. Our MBLSSVM solves K quadratic programming problems (QPPs) to obtain K hyperplanes, each problem is similar to binary LSTSVM. Comparison against the Multi-LSTSVM, Multi-TWSVM, MBSVM and our MBLSSVM on both UCI datasets and ORL, YALE face datasets illustrates the effectiveness of the proposed method. |
Tasks | |
Published | 2016-06-10 |
URL | https://www.springer.com/cn |
https://www.springer.com/cn | |
PWC | https://paperswithcode.com/paper/multiple-birth-least-squares-support-vector |
Repo | |
Framework | |
Align Me: A framework to generate Parallel Corpus Using OCRs and Bilingual Dictionaries
Title | Align Me: A framework to generate Parallel Corpus Using OCRs and Bilingual Dictionaries |
Authors | Priyam Bakliwal, Devadath V V, C V Jawahar |
Abstract | Multilingual language processing tasks like statistical machine translation and cross language information retrieval rely mainly on availability of accurate parallel corpora. Manual construction of such corpus can be extremely expensive and time consuming. In this paper we present a simple yet efficient method to generate huge amount of reasonably accurate parallel corpus with minimal user efforts. We utilize the availability of large number of English books and their corresponding translations in other languages to build parallel corpus. Optical Character Recognizing systems are used to digitize such books. We propose a robust dictionary based parallel corpus generation system for alignment of multilingual text at different levels of granularity (sentence, paragraphs, etc). We show the performance of our proposed method on a manually aligned dataset of 300 Hindi-English sentences and 100 English-Malayalam sentences. |
Tasks | Active Learning, Information Retrieval, Machine Translation, Optical Character Recognition |
Published | 2016-12-01 |
URL | https://www.aclweb.org/anthology/W16-3719/ |
https://www.aclweb.org/anthology/W16-3719 | |
PWC | https://paperswithcode.com/paper/align-me-a-framework-to-generate-parallel |
Repo | |
Framework | |
Question-Answering with Logic Specific to Video Games
Title | Question-Answering with Logic Specific to Video Games |
Authors | Corentin Dumont, Ran Tian, Kentaro Inui |
Abstract | We present a corpus and a knowledge database aiming at developing Question-Answering in a new context, the open world of a video game. We chose a popular game called {`}Minecraft{'}, and created a QA corpus with a knowledge database related to this game and the ontology of a meaning representation that will be used to structure this database. We are interested in the logic rules specific to the game, which may not exist in the real world. The ultimate goal of this research is to build a QA system that can answer natural language questions from players by using inference on these game-specific logic rules. The QA corpus is partially composed of online quiz questions and partially composed of manually written variations of the most relevant ones. The knowledge database is extracted from several wiki-like websites about Minecraft. It is composed of unstructured data, such as text, that will be structured using the meaning representation we defined, and already structured data such as infoboxes. A preliminary examination of the data shows that players are asking creative questions about the game, and that the QA corpus can be used for clustering verbs and linking them to predefined actions in the game. | |
Tasks | Question Answering |
Published | 2016-05-01 |
URL | https://www.aclweb.org/anthology/L16-1734/ |
https://www.aclweb.org/anthology/L16-1734 | |
PWC | https://paperswithcode.com/paper/question-answering-with-logic-specific-to |
Repo | |
Framework | |
A method for Automatic Text Summarization using Consensus of Multiple Similarity Measures and Ranking Techniques
Title | A method for Automatic Text Summarization using Consensus of Multiple Similarity Measures and Ranking Techniques |
Authors | Mukesh Kumar Jadon, Ayush Pareek |
Abstract | |
Tasks | Text Summarization |
Published | 2016-12-01 |
URL | https://www.aclweb.org/anthology/W16-6318/ |
https://www.aclweb.org/anthology/W16-6318 | |
PWC | https://paperswithcode.com/paper/a-method-for-automatic-text-summarization |
Repo | |
Framework | |
Gaussian Visual-Linguistic Embedding for Zero-Shot Recognition
Title | Gaussian Visual-Linguistic Embedding for Zero-Shot Recognition |
Authors | Tanmoy Mukherjee, Timothy Hospedales |
Abstract | |
Tasks | Topic Models, Zero-Shot Learning |
Published | 2016-11-01 |
URL | https://www.aclweb.org/anthology/D16-1089/ |
https://www.aclweb.org/anthology/D16-1089 | |
PWC | https://paperswithcode.com/paper/gaussian-visual-linguistic-embedding-for-zero |
Repo | |
Framework | |
Sparse Bilingual Word Representations for Cross-lingual Lexical Entailment
Title | Sparse Bilingual Word Representations for Cross-lingual Lexical Entailment |
Authors | Yogarshi Vyas, Marine Carpuat |
Abstract | |
Tasks | Machine Translation |
Published | 2016-06-01 |
URL | https://www.aclweb.org/anthology/N16-1142/ |
https://www.aclweb.org/anthology/N16-1142 | |
PWC | https://paperswithcode.com/paper/sparse-bilingual-word-representations-for |
Repo | |
Framework | |
Equation Parsing : Mapping Sentences to Grounded Equations
Title | Equation Parsing : Mapping Sentences to Grounded Equations |
Authors | Subhro Roy, Shyam Upadhyay, Dan Roth |
Abstract | |
Tasks | Semantic Parsing |
Published | 2016-11-01 |
URL | https://www.aclweb.org/anthology/D16-1117/ |
https://www.aclweb.org/anthology/D16-1117 | |
PWC | https://paperswithcode.com/paper/equation-parsing-mapping-sentences-to-1 |
Repo | |
Framework | |
Accessing and Elaborating Walenty - a Valence Dictionary of Polish - via Internet Browser
Title | Accessing and Elaborating Walenty - a Valence Dictionary of Polish - via Internet Browser |
Authors | Bart{\l}omiej Nito{'n}, Tomasz Bartosiak, El{.z}bieta Hajnicz |
Abstract | This article presents Walenty - a new valence dictionary of Polish predicates, concentrating on its creation process and access via Internet browser. The dictionary contains two layers, syntactic and semantic. The syntactic layer describes syntactic and morphosyntactic constraints predicates put on their dependants. The semantic layer shows how predicates and their arguments are involved in a situation described in an utterance. These two layers are connected, representing how semantic arguments can be realised on the surface. Walenty also contains a powerful phraseological (idiomatic) component. Walenty has been created and can be accessed remotely with a dedicated tool called Slowal. In this article, we focus on most important functionalities of this system. First, we will depict how to access the dictionary and how built-in filtering system (covering both syntactic and semantic phenomena) works. Later, we will describe the process of creating dictionary by Slowal tool that both supports and controls the work of lexicographers. |
Tasks | |
Published | 2016-05-01 |
URL | https://www.aclweb.org/anthology/L16-1215/ |
https://www.aclweb.org/anthology/L16-1215 | |
PWC | https://paperswithcode.com/paper/accessing-and-elaborating-walenty-a-valence |
Repo | |
Framework | |