May 5, 2019

1544 words 8 mins read

Paper Group NANR 100

Paper Group NANR 100

Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures. Analysis of English Spelling Errors in a Word-Typing Game. Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES). Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar …

Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures

Title Creating Japanese Political Corpus from Local Assembly Minutes of 47 prefectures
Authors Yasutomo Kimura, Keiichi Takamaru, Takuma Tanaka, Akio Kobayashi, Hiroki Sakaji, Yuzu Uchida, Hokuto Ototake, Shigeru Masuyama
Abstract This paper describes a Japanese political corpus created for interdisciplinary political research. The corpus contains the local assembly minutes of 47 prefectures from April 2011 to March 2015. This four-year period coincides with the term of office for assembly members in most autonomies. We analyze statistical data, such as the number of speakers, characters, and words, to clarify the characteristics of local assembly minutes. In addition, we identify problems associated with the different web services used by the autonomies to make the minutes available to the public.
Tasks
Published 2016-12-01
URL https://www.aclweb.org/anthology/W16-5410/
PDF https://www.aclweb.org/anthology/W16-5410
PWC https://paperswithcode.com/paper/creating-japanese-political-corpus-from-local
Repo
Framework

Analysis of English Spelling Errors in a Word-Typing Game

Title Analysis of English Spelling Errors in a Word-Typing Game
Authors Ryuichi Tachibana, Mamoru Komachi
Abstract The emergence of the web has necessitated the need to detect and correct noisy consumer-generated texts. Most of the previous studies on English spelling-error extraction collected English spelling errors from web services such as Twitter by using the edit distance or from input logs utilizing crowdsourcing. However, in the former approach, it is not clear which word corresponds to the spelling error, and the latter approach requires an annotation cost for the crowdsourcing. One notable exception is Rodrigues and Rytting (2012), who proposed to extract English spelling errors by using a word-typing game. Their approach saves the cost of crowdsourcing, and guarantees an exact alignment between the word and the spelling error. However, they did not assert whether the extracted spelling error corpora reflect the usual writing process such as writing a document. Therefore, we propose a new correctable word-typing game that is more similar to the actual writing process. Experimental results showed that we can regard typing-game logs as a source of spelling errors.
Tasks
Published 2016-05-01
URL https://www.aclweb.org/anthology/L16-1060/
PDF https://www.aclweb.org/anthology/L16-1060
PWC https://paperswithcode.com/paper/analysis-of-english-spelling-errors-in-a-word
Repo
Framework

Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES)

Title Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES)
Authors
Abstract
Tasks
Published 2016-12-01
URL https://www.aclweb.org/anthology/W16-4300/
PDF https://www.aclweb.org/anthology/W16-4300
PWC https://paperswithcode.com/paper/proceedings-of-the-workshop-on-computational
Repo
Framework

Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction

Title Using Left-corner Parsing to Encode Universal Structural Constraints in Grammar Induction
Authors Hiroshi Noji, Yusuke Miyao, Mark Johnson
Abstract
Tasks
Published 2016-11-01
URL https://www.aclweb.org/anthology/D16-1004/
PDF https://www.aclweb.org/anthology/D16-1004
PWC https://paperswithcode.com/paper/using-left-corner-parsing-to-encode-universal
Repo
Framework

Large-Scale Semantic Indexing and Question Answering in Biomedicine

Title Large-Scale Semantic Indexing and Question Answering in Biomedicine
Authors Eirini Papagiannopoulou, Yiannis Papanikolaou, Dimitris Dimitriadis, Sakis Lagopoulos, Grigorios Tsoumakas, Manos Laliotis, Nikos Markantonatos, Ioannis Vlahavas
Abstract
Tasks Question Answering
Published 2016-08-01
URL https://www.aclweb.org/anthology/W16-3107/
PDF https://www.aclweb.org/anthology/W16-3107
PWC https://paperswithcode.com/paper/large-scale-semantic-indexing-and-question
Repo
Framework

KSAnswer: Question-answering System of Kangwon National University and Sogang University in the 2016 BioASQ Challenge

Title KSAnswer: Question-answering System of Kangwon National University and Sogang University in the 2016 BioASQ Challenge
Authors Hyeon-gu Lee, Minkyoung Kim, Harksoo Kim, Juae Kim, Sunjae Kwon, Jungyun Seo, Yi-reun Kim, Jung-Kyu Choi
Abstract
Tasks Language Modelling, Morphological Analysis, Question Answering
Published 2016-08-01
URL https://www.aclweb.org/anthology/W16-3106/
PDF https://www.aclweb.org/anthology/W16-3106
PWC https://paperswithcode.com/paper/ksanswer-question-answering-system-of-kangwon
Repo
Framework

Proceedings of the 19th Annual Conference of the European Association for Machine Translation

Title Proceedings of the 19th Annual Conference of the European Association for Machine Translation
Authors
Abstract
Tasks Machine Translation
Published 2016-01-01
URL https://www.aclweb.org/anthology/W16-3400/
PDF https://www.aclweb.org/anthology/W16-3400
PWC https://paperswithcode.com/paper/proceedings-of-the-19th-annual-conference-of-1
Repo
Framework

Multiple birth least squares support vector machine for multi-class classification

Title Multiple birth least squares support vector machine for multi-class classification
Authors Su-Gen Chen1, 2•Xiao-Jun Wu1
Abstract Least squares twin support vector machine (LSTSVM) was initially designed for binary classification. However, practical problems often require the discrimina-tion more than two categories. To tackle multi-class classi-fication problem, a novel algorithm, called multiple birth least squares support vector machine (MBLSSVM), is proposed. Our MBLSSVM solves K quadratic programming problems (QPPs) to obtain K hyperplanes, each problem is similar to binary LSTSVM. Comparison against the Multi-LSTSVM, Multi-TWSVM, MBSVM and our MBLSSVM on both UCI datasets and ORL, YALE face datasets illustrates the effectiveness of the proposed method.
Tasks
Published 2016-06-10
URL https://www.springer.com/cn
PDF https://www.springer.com/cn
PWC https://paperswithcode.com/paper/multiple-birth-least-squares-support-vector
Repo
Framework

Align Me: A framework to generate Parallel Corpus Using OCRs and Bilingual Dictionaries

Title Align Me: A framework to generate Parallel Corpus Using OCRs and Bilingual Dictionaries
Authors Priyam Bakliwal, Devadath V V, C V Jawahar
Abstract Multilingual language processing tasks like statistical machine translation and cross language information retrieval rely mainly on availability of accurate parallel corpora. Manual construction of such corpus can be extremely expensive and time consuming. In this paper we present a simple yet efficient method to generate huge amount of reasonably accurate parallel corpus with minimal user efforts. We utilize the availability of large number of English books and their corresponding translations in other languages to build parallel corpus. Optical Character Recognizing systems are used to digitize such books. We propose a robust dictionary based parallel corpus generation system for alignment of multilingual text at different levels of granularity (sentence, paragraphs, etc). We show the performance of our proposed method on a manually aligned dataset of 300 Hindi-English sentences and 100 English-Malayalam sentences.
Tasks Active Learning, Information Retrieval, Machine Translation, Optical Character Recognition
Published 2016-12-01
URL https://www.aclweb.org/anthology/W16-3719/
PDF https://www.aclweb.org/anthology/W16-3719
PWC https://paperswithcode.com/paper/align-me-a-framework-to-generate-parallel
Repo
Framework

Question-Answering with Logic Specific to Video Games

Title Question-Answering with Logic Specific to Video Games
Authors Corentin Dumont, Ran Tian, Kentaro Inui
Abstract We present a corpus and a knowledge database aiming at developing Question-Answering in a new context, the open world of a video game. We chose a popular game called {`}Minecraft{'}, and created a QA corpus with a knowledge database related to this game and the ontology of a meaning representation that will be used to structure this database. We are interested in the logic rules specific to the game, which may not exist in the real world. The ultimate goal of this research is to build a QA system that can answer natural language questions from players by using inference on these game-specific logic rules. The QA corpus is partially composed of online quiz questions and partially composed of manually written variations of the most relevant ones. The knowledge database is extracted from several wiki-like websites about Minecraft. It is composed of unstructured data, such as text, that will be structured using the meaning representation we defined, and already structured data such as infoboxes. A preliminary examination of the data shows that players are asking creative questions about the game, and that the QA corpus can be used for clustering verbs and linking them to predefined actions in the game. |
Tasks Question Answering
Published 2016-05-01
URL https://www.aclweb.org/anthology/L16-1734/
PDF https://www.aclweb.org/anthology/L16-1734
PWC https://paperswithcode.com/paper/question-answering-with-logic-specific-to
Repo
Framework

A method for Automatic Text Summarization using Consensus of Multiple Similarity Measures and Ranking Techniques

Title A method for Automatic Text Summarization using Consensus of Multiple Similarity Measures and Ranking Techniques
Authors Mukesh Kumar Jadon, Ayush Pareek
Abstract
Tasks Text Summarization
Published 2016-12-01
URL https://www.aclweb.org/anthology/W16-6318/
PDF https://www.aclweb.org/anthology/W16-6318
PWC https://paperswithcode.com/paper/a-method-for-automatic-text-summarization
Repo
Framework

Gaussian Visual-Linguistic Embedding for Zero-Shot Recognition

Title Gaussian Visual-Linguistic Embedding for Zero-Shot Recognition
Authors Tanmoy Mukherjee, Timothy Hospedales
Abstract
Tasks Topic Models, Zero-Shot Learning
Published 2016-11-01
URL https://www.aclweb.org/anthology/D16-1089/
PDF https://www.aclweb.org/anthology/D16-1089
PWC https://paperswithcode.com/paper/gaussian-visual-linguistic-embedding-for-zero
Repo
Framework

Sparse Bilingual Word Representations for Cross-lingual Lexical Entailment

Title Sparse Bilingual Word Representations for Cross-lingual Lexical Entailment
Authors Yogarshi Vyas, Marine Carpuat
Abstract
Tasks Machine Translation
Published 2016-06-01
URL https://www.aclweb.org/anthology/N16-1142/
PDF https://www.aclweb.org/anthology/N16-1142
PWC https://paperswithcode.com/paper/sparse-bilingual-word-representations-for
Repo
Framework

Equation Parsing : Mapping Sentences to Grounded Equations

Title Equation Parsing : Mapping Sentences to Grounded Equations
Authors Subhro Roy, Shyam Upadhyay, Dan Roth
Abstract
Tasks Semantic Parsing
Published 2016-11-01
URL https://www.aclweb.org/anthology/D16-1117/
PDF https://www.aclweb.org/anthology/D16-1117
PWC https://paperswithcode.com/paper/equation-parsing-mapping-sentences-to-1
Repo
Framework

Accessing and Elaborating Walenty - a Valence Dictionary of Polish - via Internet Browser

Title Accessing and Elaborating Walenty - a Valence Dictionary of Polish - via Internet Browser
Authors Bart{\l}omiej Nito{'n}, Tomasz Bartosiak, El{.z}bieta Hajnicz
Abstract This article presents Walenty - a new valence dictionary of Polish predicates, concentrating on its creation process and access via Internet browser. The dictionary contains two layers, syntactic and semantic. The syntactic layer describes syntactic and morphosyntactic constraints predicates put on their dependants. The semantic layer shows how predicates and their arguments are involved in a situation described in an utterance. These two layers are connected, representing how semantic arguments can be realised on the surface. Walenty also contains a powerful phraseological (idiomatic) component. Walenty has been created and can be accessed remotely with a dedicated tool called Slowal. In this article, we focus on most important functionalities of this system. First, we will depict how to access the dictionary and how built-in filtering system (covering both syntactic and semantic phenomena) works. Later, we will describe the process of creating dictionary by Slowal tool that both supports and controls the work of lexicographers.
Tasks
Published 2016-05-01
URL https://www.aclweb.org/anthology/L16-1215/
PDF https://www.aclweb.org/anthology/L16-1215
PWC https://paperswithcode.com/paper/accessing-and-elaborating-walenty-a-valence
Repo
Framework
comments powered by Disqus