October 15, 2019

1849 words 9 mins read

Paper Group NANR 56

Paper Group NANR 56

Using Object Information for Spotting Text. Semantically Aware Urban 3D Reconstruction with Plane-Based Regularization. Joint Learning for Emotion Classification and Emotion Cause Detection. Proceedings of the AMTA 2018 Workshop on Technologies for MT of Low Resource Languages (LoResMT 2018). Translation API Cases and Classes (TAPICC). Embedding Re …

Using Object Information for Spotting Text

Title Using Object Information for Spotting Text
Authors Shitala Prasad, Adams Wai Kin Kong
Abstract Text spotting, also called text detection, is a challenging computer vision task because of cluttered backgrounds, diverse imaging environments, various text sizes and similarity between some objects and characters, e.g., tyre and ’o’. However, text spotting is a vital step in numerous AI and computer vision systems, such as autonomous robots and systems for visually impaired. Due to its potential applications and commercial values, researchers have proposed various deep architectures and methods for text spotting. These methods and architectures concentrate only on text in images, but neglect other information related to text. There exists a strong relationship between certain objects and the presence of text, such as signboards or the absence of text, such as trees. In this paper, a text spotting algorithm based on text and object dependency is proposed. The proposed algorithm consists of two sub-convolutional neural networks and three training stages. For this study, a new NTU-UTOI dataset containing over 22k non-synthetic images with 277k bounding boxes for text and 42 text-related object classes is established. According to our best knowledge, it is the second largest non-synthetic text image database. Experimental results on three benchmark datasets with clutter backgrounds, COCO-Text, MSRA-TD500 and SVT show that the proposed algorithm provides comparable performance to state-of-the-art text spotting methods. Experiments are also performed on our newly established dataset to investigate the effectiveness of object information for text spotting. The experimental results indicate that the object information contributes significantly on the performance gain.
Tasks Text Spotting
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Shitala_Prasad_Using_Object_Information_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Shitala_Prasad_Using_Object_Information_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/using-object-information-for-spotting-text
Repo
Framework

Semantically Aware Urban 3D Reconstruction with Plane-Based Regularization

Title Semantically Aware Urban 3D Reconstruction with Plane-Based Regularization
Authors Thomas Holzmann, Michael Maurer, Friedrich Fraundorfer, Horst Bischof
Abstract We propose a method for urban 3D reconstruction, which incorporates semantic information and plane priors within the reconstruction process in order to generate visually appealing 3D models. We introduce a plane detection algorithm using 3D lines, which detects a more complete and less spurious plane set compared to point-based methods in urban environments. Further, the proposed normalized visibility-based energy formulation eases the combination of several energy terms within a tetrahedra occupancy labeling algorithm and, hence, is well suited for combining it with class specific smoothness terms. As a result, we produce visually appealing and detailed building models (i.e., straight edges and planar surfaces) and a smooth reconstruction of the surroundings.
Tasks 3D Reconstruction
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Thomas_Holzmann_Semantically_Aware_Urban_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Thomas_Holzmann_Semantically_Aware_Urban_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/semantically-aware-urban-3d-reconstruction
Repo
Framework

Joint Learning for Emotion Classification and Emotion Cause Detection

Title Joint Learning for Emotion Classification and Emotion Cause Detection
Authors Ying Chen, Wenjun Hou, Xiyao Cheng, Shoushan Li
Abstract We present a neural network-based joint approach for emotion classification and emotion cause detection, which attempts to capture mutual benefits across the two sub-tasks of emotion analysis. Considering that emotion classification and emotion cause detection need different kinds of features (affective and event-based separately), we propose a joint encoder which uses a unified framework to extract features for both sub-tasks and a joint model trainer which simultaneously learns two models for the two sub-tasks separately. Our experiments on Chinese microblogs show that the joint approach is very promising.
Tasks Emotion Classification, Emotion Recognition
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1066/
PDF https://www.aclweb.org/anthology/D18-1066
PWC https://paperswithcode.com/paper/joint-learning-for-emotion-classification-and
Repo
Framework

Proceedings of the AMTA 2018 Workshop on Technologies for MT of Low Resource Languages (LoResMT 2018)

Title Proceedings of the AMTA 2018 Workshop on Technologies for MT of Low Resource Languages (LoResMT 2018)
Authors
Abstract
Tasks Machine Translation
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2200/
PDF https://www.aclweb.org/anthology/W18-2200
PWC https://paperswithcode.com/paper/proceedings-of-the-amta-2018-workshop-on-1
Repo
Framework

Translation API Cases and Classes (TAPICC)

Title Translation API Cases and Classes (TAPICC)
Authors Alan Melby
Abstract
Tasks
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2006/
PDF https://www.aclweb.org/anthology/W18-2006
PWC https://paperswithcode.com/paper/translation-api-cases-and-classes-tapicc
Repo
Framework

Embedding Register-Aware MT into the CAT Workflow

Title Embedding Register-Aware MT into the CAT Workflow
Authors Corey Miller, Danielle Silverman, Vanesa Jurica, Elizabeth Richerson, Rodney Morris, Elisabeth Mallard
Abstract
Tasks Machine Translation
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-1920/
PDF https://www.aclweb.org/anthology/W18-1920
PWC https://paperswithcode.com/paper/embedding-register-aware-mt-into-the-cat
Repo
Framework

Translation Quality Standards

Title Translation Quality Standards
Authors Bill Rivers
Abstract
Tasks
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2004/
PDF https://www.aclweb.org/anthology/W18-2004
PWC https://paperswithcode.com/paper/translation-quality-standards
Repo
Framework

Proceedings of the AMTA 2018 Workshop on Translation Quality Estimation and Automatic Post-Editing

Title Proceedings of the AMTA 2018 Workshop on Translation Quality Estimation and Automatic Post-Editing
Authors
Abstract
Tasks Automatic Post-Editing, Machine Translation
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2100/
PDF https://www.aclweb.org/anthology/W18-2100
PWC https://paperswithcode.com/paper/proceedings-of-the-amta-2018-workshop-on
Repo
Framework

Automatic Post-Editing and Machine Translation Quality Estimation at eBay

Title Automatic Post-Editing and Machine Translation Quality Estimation at eBay
Authors Nicola Ueffing
Abstract
Tasks Automatic Post-Editing, Machine Translation
Published 2018-03-01
URL https://www.aclweb.org/anthology/W18-2101/
PDF https://www.aclweb.org/anthology/W18-2101
PWC https://paperswithcode.com/paper/automatic-post-editing-and-machine
Repo
Framework

Joint Camera Spectral Sensitivity Selection and Hyperspectral Image Recovery

Title Joint Camera Spectral Sensitivity Selection and Hyperspectral Image Recovery
Authors Ying Fu, Tao Zhang, Yinqiang Zheng, Debing Zhang, Hua Huang
Abstract Hyperspectral image (HSI) recovery from a single RGB image has attracted much attention, whose performance has recently been shown to be sensitive to the camera spectral sensitivity (CSS). In this paper, we present an efficient convolutional neural network (CNN) based method, which can jointly select the optimal CSS from a candidate dataset and learn a mapping to recover HSI from a single RGB image captured with this algorithmically selected camera. Given a specific CSS, we first present a HSI recovery network, which accounts for the underlying characteristics of the HSI, including spectral nonlinear mapping and spatial similarity. Later, we append a CSS selection layer onto the recovery network, and the optimal CSS can thus be automatically determined from the network weights under the nonnegative sparse constraint. Experimental results show that our HSI recovery network outperforms state-of-the-art methods in terms of both quantitative metrics and perceptive quality, and the selection layer always returns a CSS consistent to the best one determined by exhaustive search.
Tasks
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Ying_Fu_Joint_Camera_Spectral_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Ying_Fu_Joint_Camera_Spectral_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/joint-camera-spectral-sensitivity-selection
Repo
Framework

Saliency Detection in 360° Videos

Title Saliency Detection in 360° Videos
Authors Ziheng Zhang, Yanyu Xu, Jingyi Yu, Shenghua Gao
Abstract This paper presents a novel spherical convolutional neural network based scheme for saliency detection for 360° videos. Specifically, in our spherical convolution neural network definition, kernel is defined on a spherical crown, and the convolution involves the rotation of the kernel along the sphere. Considering that the 360° videos are usually stored with equirectangular panorama, we propose to implement the spherical convolution on panorama by stretching and rotating the kernel based on the location of patch to be convolved. Compared with existing spherical convolution, our definition has the parameter sharing property, which would greatly reduce the parameters to be learned. We further take the temporal coherence of the viewing process into consideration, and propose a sequential saliency detection by leveraging a spherical U-Net. To validate our approach, we construct a large-scale 360° videos saliency detection benchmark that consists of 104 360° videos viewed by 20+ human subjects. Comprehensive experiments validate the effectiveness of our spherical U-net for 360° video saliency detection.
Tasks Saliency Detection, Video Saliency Detection
Published 2018-09-01
URL http://openaccess.thecvf.com/content_ECCV_2018/html/Ziheng_Zhang_Saliency_Detection_in_ECCV_2018_paper.html
PDF http://openaccess.thecvf.com/content_ECCV_2018/papers/Ziheng_Zhang_Saliency_Detection_in_ECCV_2018_paper.pdf
PWC https://paperswithcode.com/paper/saliency-detection-in-360a-videos
Repo
Framework
Title BASHI: A Corpus of Wall Street Journal Articles Annotated with Bridging Links
Authors Ina R{"o}siger
Abstract
Tasks Coreference Resolution, Natural Language Inference
Published 2018-05-01
URL https://www.aclweb.org/anthology/L18-1058/
PDF https://www.aclweb.org/anthology/L18-1058
PWC https://paperswithcode.com/paper/bashi-a-corpus-of-wall-street-journal
Repo
Framework

Increasing In-Class Similarity by Retrofitting Embeddings with Demographic Information

Title Increasing In-Class Similarity by Retrofitting Embeddings with Demographic Information
Authors Dirk Hovy, Tommaso Fornaciari
Abstract Most text-classification approaches represent the input based on textual features, either feature-based or continuous. However, this ignores strong non-linguistic similarities like homophily: people within a demographic group use language more similar to each other than to non-group members. We use homophily cues to retrofit text-based author representations with non-linguistic information, and introduce a trade-off parameter. This approach increases in-class similarity between authors, and improves classification performance by making classes more linearly separable. We evaluate the effect of our method on two author-attribute prediction tasks with various training-set sizes and parameter settings. We find that our method can significantly improve classification performance, especially when the number of labels is large and limited labeled data is available. It is potentially applicable as preprocessing step to any text-classification task.
Tasks Semantic Textual Similarity, Text Classification
Published 2018-10-01
URL https://www.aclweb.org/anthology/D18-1070/
PDF https://www.aclweb.org/anthology/D18-1070
PWC https://paperswithcode.com/paper/increasing-in-class-similarity-by
Repo
Framework

A Reduction for Efficient LDA Topic Reconstruction

Title A Reduction for Efficient LDA Topic Reconstruction
Authors Matteo Almanza, Flavio Chierichetti, Alessandro Panconesi, Andrea Vattani
Abstract We present a novel approach for LDA (Latent Dirichlet Allocation) topic reconstruction. The main technical idea is to show that the distribution over the documents generated by LDA can be transformed into a distribution for a much simpler generative model in which documents are generated from {\em the same set of topics} but have a much simpler structure: documents are single topic and topics are chosen uniformly at random. Furthermore, this reduction is approximation preserving, in the sense that approximate distributions– the only ones we can hope to compute in practice– are mapped into approximate distribution in the simplified world. This opens up the possibility of efficiently reconstructing LDA topics in a roundabout way. Compute an approximate document distribution from the given corpus, transform it into an approximate distribution for the single-topic world, and run a reconstruction algorithm in the uniform, single topic world– a much simpler task than direct LDA reconstruction. Indeed, we show the viability of the approach by giving very simple algorithms for a generalization of two notable cases that have been studied in the literature, $p$-separability and Gibbs sampling for matrix-like topics.
Tasks
Published 2018-12-01
URL http://papers.nips.cc/paper/8012-a-reduction-for-efficient-lda-topic-reconstruction
PDF http://papers.nips.cc/paper/8012-a-reduction-for-efficient-lda-topic-reconstruction.pdf
PWC https://paperswithcode.com/paper/a-reduction-for-efficient-lda-topic
Repo
Framework

Improving Beam Search by Removing Monotonic Constraint for Neural Machine Translation

Title Improving Beam Search by Removing Monotonic Constraint for Neural Machine Translation
Authors Raphael Shu, Hideki Nakayama
Abstract To achieve high translation performance, neural machine translation models usually rely on the beam search algorithm for decoding sentences. The beam search finds good candidate translations by considering multiple hypotheses of translations simultaneously. However, as the algorithm produces hypotheses in a monotonic left-to-right order, a hypothesis can not be revisited once it is discarded. We found such monotonicity forces the algorithm to sacrifice some good decoding paths. To mitigate this problem, we relax the monotonic constraint of the beam search by maintaining all found hypotheses in a single priority queue and using a universal score function for hypothesis selection. The proposed algorithm allows discarded hypotheses to be recovered in a later step. Despite its simplicity, we show that the proposed decoding algorithm enhances the quality of selected hypotheses and improve the translations even for high-performance models in English-Japanese translation task.
Tasks Language Modelling, Machine Translation
Published 2018-07-01
URL https://www.aclweb.org/anthology/P18-2054/
PDF https://www.aclweb.org/anthology/P18-2054
PWC https://paperswithcode.com/paper/improving-beam-search-by-removing-monotonic
Repo
Framework
comments powered by Disqus