Paper Group AWR 249
OpenVSLAM: A Versatile Visual SLAM Framework. Time-warping invariants of multidimensional time series. A new evaluation framework for topic modeling algorithms based on synthetic corpora. Supervised online diarization with sample mean loss for multi-domain data. Learning Perspective Undistortion of Portraits. Scalable Kernel Learning via the Discri …
OpenVSLAM: A Versatile Visual SLAM Framework
Title | OpenVSLAM: A Versatile Visual SLAM Framework |
Authors | Shinya Sumikura, Mikiya Shibuya, Ken Sakurada |
Abstract | In this paper, we introduce OpenVSLAM, a visual SLAM framework with high usability and extensibility. Visual SLAM systems are essential for AR devices, autonomous control of robots and drones, etc. However, conventional open-source visual SLAM frameworks are not appropriately designed as libraries called from third-party programs. To overcome this situation, we have developed a novel visual SLAM framework. This software is designed to be easily used and extended. It incorporates several useful features and functions for research and development. OpenVSLAM is released at https://github.com/xdspacelab/openvslam under the 2-clause BSD license. |
Tasks | |
Published | 2019-10-02 |
URL | https://arxiv.org/abs/1910.01122v2 |
https://arxiv.org/pdf/1910.01122v2.pdf | |
PWC | https://paperswithcode.com/paper/openvslam-a-versatile-visual-slam-framework |
Repo | https://github.com/xdspacelab/openvslam |
Framework | none |
Time-warping invariants of multidimensional time series
Title | Time-warping invariants of multidimensional time series |
Authors | Joscha Diehl, Kurusch Ebrahimi-Fard, Nikolas Tapia |
Abstract | In data science, one is often confronted with a time series representing measurements of some quantity of interest. Usually, as a first step, features of the time series need to be extracted. These are numerical quantities that aim to succinctly describe the data and to dampen the influence of noise. In some applications, these features are also required to satisfy some invariance properties. In this paper, we concentrate on time-warping invariants. We show that these correspond to a certain family of iterated sums of the increments of the time series, known as quasisymmetric functions in the mathematics literature. We present these invariant features in an algebraic framework, and we develop some of their basic properties. |
Tasks | Time Series |
Published | 2019-06-13 |
URL | https://arxiv.org/abs/1906.05823v2 |
https://arxiv.org/pdf/1906.05823v2.pdf | |
PWC | https://paperswithcode.com/paper/time-warping-invariants-of-multidimensional |
Repo | https://github.com/diehlj/iterated-sums-signature-py |
Framework | none |
A new evaluation framework for topic modeling algorithms based on synthetic corpora
Title | A new evaluation framework for topic modeling algorithms based on synthetic corpora |
Authors | Hanyu Shi, Martin Gerlach, Isabel Diersen, Doug Downey, Luis A. N. Amaral |
Abstract | Topic models are in widespread use in natural language processing and beyond. Here, we propose a new framework for the evaluation of probabilistic topic modeling algorithms based on synthetic corpora containing an unambiguously defined ground truth topic structure. The major innovation of our approach is the ability to quantify the agreement between the planted and inferred topic structures by comparing the assigned topic labels at the level of the tokens. In experiments, our approach yields novel insights about the relative strengths of topic models as corpus characteristics vary, and the first evidence of an “undetectable phase” for topic models when the planted structure is weak. We also establish the practical relevance of the insights gained for synthetic corpora by predicting the performance of topic modeling algorithms in classification tasks in real-world corpora. |
Tasks | Topic Models |
Published | 2019-01-28 |
URL | http://arxiv.org/abs/1901.09848v1 |
http://arxiv.org/pdf/1901.09848v1.pdf | |
PWC | https://paperswithcode.com/paper/a-new-evaluation-framework-for-topic-modeling |
Repo | https://github.com/amarallab/synthetic_benchmark_topic_model |
Framework | none |
Supervised online diarization with sample mean loss for multi-domain data
Title | Supervised online diarization with sample mean loss for multi-domain data |
Authors | Enrico Fini, Alessio Brutti |
Abstract | Recently, a fully supervised speaker diarization approach was proposed (UIS-RNN) which models speakers using multiple instances of a parameter-sharing recurrent neural network. In this paper we propose qualitative modifications to the model that significantly improve the learning efficiency and the overall diarization performance. In particular, we introduce a novel loss function, we called Sample Mean Loss and we present a better modelling of the speaker turn behaviour, by devising an analytical expression to compute the probability of a new speaker joining the conversation. In addition, we demonstrate that our model can be trained on fixed-length speech segments, removing the need for speaker change information in inference. Using x-vectors as input features, we evaluate our proposed approach on the multi-domain dataset employed in the DIHARD II challenge: our online method improves with respect to the original UIS-RNN and achieves similar performance to an offline agglomerative clustering baseline using PLDA scoring. |
Tasks | Speaker Diarization |
Published | 2019-11-04 |
URL | https://arxiv.org/abs/1911.01266v3 |
https://arxiv.org/pdf/1911.01266v3.pdf | |
PWC | https://paperswithcode.com/paper/supervised-online-diarization-with-sample |
Repo | https://github.com/DonkeyShot21/uis-rnn-sml |
Framework | pytorch |
Learning Perspective Undistortion of Portraits
Title | Learning Perspective Undistortion of Portraits |
Authors | Yajie Zhao, Zeng Huang, Tianye Li, Weikai Chen, Chloe LeGendre, Xinglei Ren, Jun Xing, Ari Shapiro, Hao Li |
Abstract | Near-range portrait photographs often contain perspective distortion artifacts that bias human perception and challenge both facial recognition and reconstruction techniques. We present the first deep learning based approach to remove such artifacts from unconstrained portraits. In contrast to the previous state-of-the-art approach, our method handles even portraits with extreme perspective distortion, as we avoid the inaccurate and error-prone step of first fitting a 3D face model. Instead, we predict a distortion correction flow map that encodes a per-pixel displacement that removes distortion artifacts when applied to the input image. Our method also automatically infers missing facial features, i.e. occluded ears caused by strong perspective distortion, with coherent details. We demonstrate that our approach significantly outperforms the previous state-of-the-art both qualitatively and quantitatively, particularly for portraits with extreme perspective distortion or facial expressions. We further show that our technique benefits a number of fundamental tasks, significantly improving the accuracy of both face recognition and 3D reconstruction and enables a novel camera calibration technique from a single portrait. Moreover, we also build the first perspective portrait database with a large diversity in identities, expression and poses, which will benefit the related research in this area. |
Tasks | 3D Reconstruction, Calibration, Face Recognition |
Published | 2019-05-18 |
URL | https://arxiv.org/abs/1905.07515v1 |
https://arxiv.org/pdf/1905.07515v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-perspective-undistortion-of |
Repo | https://github.com/bearjoy730/Learning-Perspective-Undistortion-of-Portraits |
Framework | none |
Scalable Kernel Learning via the Discriminant Information
Title | Scalable Kernel Learning via the Discriminant Information |
Authors | Mert Al, Zejiang Hou, Sun-Yuan Kung |
Abstract | Kernel approximation methods create explicit, low-dimensional kernel feature maps to deal with the high computational and memory complexity of standard techniques. This work studies a supervised kernel learning methodology to optimize such mappings. We utilize the Discriminant Information criterion, a measure of class separability with a strong connection to Discriminant Analysis. By generalizing this measure to cover a wider range of kernel maps and learning settings, we develop scalable methods to learn kernel features with high discriminant power. Experimental results on several datasets showcase that our techniques can improve optimization and generalization performances over state of the art kernel learning methods. |
Tasks | |
Published | 2019-09-23 |
URL | https://arxiv.org/abs/1909.10432v2 |
https://arxiv.org/pdf/1909.10432v2.pdf | |
PWC | https://paperswithcode.com/paper/190910432 |
Repo | https://github.com/almert/di-kernel-optimization |
Framework | tf |
Normalizing Flows for Probabilistic Modeling and Inference
Title | Normalizing Flows for Probabilistic Modeling and Inference |
Authors | George Papamakarios, Eric Nalisnick, Danilo Jimenez Rezende, Shakir Mohamed, Balaji Lakshminarayanan |
Abstract | Normalizing flows provide a general mechanism for defining expressive probability distributions, only requiring the specification of a (usually simple) base distribution and a series of bijective transformations. There has been much recent work on normalizing flows, ranging from improving their expressive power to expanding their application. We believe the field has now matured and is in need of a unified perspective. In this review, we attempt to provide such a perspective by describing flows through the lens of probabilistic modeling and inference. We place special emphasis on the fundamental principles of flow design, and discuss foundational topics such as expressive power and computational trade-offs. We also broaden the conceptual framing of flows by relating them to more general probability transformations. Lastly, we summarize the use of flows for tasks such as generative modeling, approximate inference, and supervised learning. |
Tasks | |
Published | 2019-12-05 |
URL | https://arxiv.org/abs/1912.02762v1 |
https://arxiv.org/pdf/1912.02762v1.pdf | |
PWC | https://paperswithcode.com/paper/normalizing-flows-for-probabilistic-modeling |
Repo | https://github.com/mackelab/nflows |
Framework | pytorch |
CFEA: Collaborative Feature Ensembling Adaptation for Domain Adaptation in Unsupervised Optic Disc and Cup Segmentation
Title | CFEA: Collaborative Feature Ensembling Adaptation for Domain Adaptation in Unsupervised Optic Disc and Cup Segmentation |
Authors | Peng Liu, Bin Kong, Zhongyu Li, Shaoting Zhang, Ruogu Fang |
Abstract | Recently, deep neural networks have demonstrated comparable and even better performance with board-certified ophthalmologists in well-annotated datasets. However, the diversity of retinal imaging devices poses a significant challenge: domain shift, which leads to performance degradation when applying the deep learning models to new testing domains. In this paper, we propose a novel unsupervised domain adaptation framework, called Collaborative Feature Ensembling Adaptation (CFEA), to effectively overcome this challenge. Our proposed CFEA is an interactive paradigm which presents an exquisite of collaborative adaptation through both adversarial learning and ensembling weights. In particular, we simultaneously achieve domain-invariance and maintain an exponential moving average of the historical predictions, which achieves a better prediction for the unlabeled data, via ensembling weights during training. Without annotating any sample from the target domain, multiple adversarial losses in encoder and decoder layers guide the extraction of domain-invariant features to confuse the domain classifier and meanwhile benefit the ensembling of smoothing weights. Comprehensive experimental results demonstrate that our CFEA model can overcome performance degradation and outperform the state-of-the-art methods in segmenting retinal optic disc and cup from fundus images. \textit{Code is available at \url{https://github.com/cswin/AWC}}. |
Tasks | Domain Adaptation, Unsupervised Domain Adaptation |
Published | 2019-10-16 |
URL | https://arxiv.org/abs/1910.07638v1 |
https://arxiv.org/pdf/1910.07638v1.pdf | |
PWC | https://paperswithcode.com/paper/cfea-collaborative-feature-ensembling |
Repo | https://github.com/cswin/AWC |
Framework | pytorch |
Efficient Decoupled Neural Architecture Search by Structure and Operation Sampling
Title | Efficient Decoupled Neural Architecture Search by Structure and Operation Sampling |
Authors | Heung-Chang Lee, Do-Guk Kim, Bohyung Han |
Abstract | We propose a novel neural architecture search algorithm via reinforcement learning by decoupling structure and operation search processes. Our approach samples candidate models from the multinomial distribution on the policy vectors defined on the two search spaces independently. The proposed technique improves the efficiency of architecture search process significantly compared to the conventional methods based on reinforcement learning with the RNN controllers while achieving competitive accuracy and model size in target tasks. Our policy vectors are easily interpretable throughout the training procedure, which allows to analyze the search progress and the discovered architectures; the black-box characteristics of the RNN controllers hamper understanding training progress in terms of policy parameter updates. Our experiments demonstrate outstanding performance compared to the state-of-the-art methods with a fraction of search cost. |
Tasks | Neural Architecture Search |
Published | 2019-10-23 |
URL | https://arxiv.org/abs/1910.10397v1 |
https://arxiv.org/pdf/1910.10397v1.pdf | |
PWC | https://paperswithcode.com/paper/efficient-decoupled-neural-architecture |
Repo | https://github.com/logue311/EDNAS |
Framework | pytorch |
Semi-Supervised Semantic Segmentation with High- and Low-level Consistency
Title | Semi-Supervised Semantic Segmentation with High- and Low-level Consistency |
Authors | Sudhanshu Mittal, Maxim Tatarchenko, Thomas Brox |
Abstract | The ability to understand visual information from limited labeled data is an important aspect of machine learning. While image-level classification has been extensively studied in a semi-supervised setting, dense pixel-level classification with limited data has only drawn attention recently. In this work, we propose an approach for semi-supervised semantic segmentation that learns from limited pixel-wise annotated samples while exploiting additional annotation-free images. It uses two network branches that link semi-supervised classification with semi-supervised segmentation including self-training. The dual-branch approach reduces both the low-level and the high-level artifacts typical when training with few labels. The approach attains significant improvement over existing methods, especially when trained with very few labeled samples. On several standard benchmarks - PASCAL VOC 2012, PASCAL-Context, and Cityscapes - the approach achieves new state-of-the-art in semi-supervised learning. |
Tasks | Semantic Segmentation, Semi-Supervised Semantic Segmentation |
Published | 2019-08-15 |
URL | https://arxiv.org/abs/1908.05724v1 |
https://arxiv.org/pdf/1908.05724v1.pdf | |
PWC | https://paperswithcode.com/paper/semi-supervised-semantic-segmentation-with |
Repo | https://github.com/sud0301/semisup-semseg |
Framework | pytorch |
Unsupervised Domain Adaptation through Self-Supervision
Title | Unsupervised Domain Adaptation through Self-Supervision |
Authors | Yu Sun, Eric Tzeng, Trevor Darrell, Alexei A. Efros |
Abstract | This paper addresses unsupervised domain adaptation, the setting where labeled training data is available on a source domain, but the goal is to have good performance on a target domain with only unlabeled data. Like much of previous work, we seek to align the learned representations of the source and target domains while preserving discriminability. The way we accomplish alignment is by learning to perform auxiliary self-supervised task(s) on both domains simultaneously. Each self-supervised task brings the two domains closer together along the direction relevant to that task. Training this jointly with the main task classifier on the source domain is shown to successfully generalize to the unlabeled target domain. The presented objective is straightforward to implement and easy to optimize. We achieve state-of-the-art results on four out of seven standard benchmarks, and competitive results on segmentation adaptation. We also demonstrate that our method composes well with another popular pixel-level adaptation method. |
Tasks | Domain Adaptation, Unsupervised Domain Adaptation |
Published | 2019-09-26 |
URL | https://arxiv.org/abs/1909.11825v2 |
https://arxiv.org/pdf/1909.11825v2.pdf | |
PWC | https://paperswithcode.com/paper/unsupervised-domain-adaptation-through-self-1 |
Repo | https://github.com/yueatsprograms/uda_release |
Framework | pytorch |
Computer-aided diagnosis in histopathological images of the endometrium using a convolutional neural network and attention mechanisms
Title | Computer-aided diagnosis in histopathological images of the endometrium using a convolutional neural network and attention mechanisms |
Authors | Hao Sun, Xianxu Zeng, Tao Xu, Gang Peng, Yutao Ma |
Abstract | Uterine cancer, also known as endometrial cancer, can seriously affect the female reproductive organs, and histopathological image analysis is the gold standard for diagnosing endometrial cancer. However, due to the limited capability of modeling the complicated relationships between histopathological images and their interpretations, these computer-aided diagnosis (CADx) approaches based on traditional machine learning algorithms often failed to achieve satisfying results. In this study, we developed a CADx approach using a convolutional neural network (CNN) and attention mechanisms, called HIENet. Because HIENet used the attention mechanisms and feature map visualization techniques, it can provide pathologists better interpretability of diagnoses by highlighting the histopathological correlations of local (pixel-level) image features to morphological characteristics of endometrial tissue. In the ten-fold cross-validation process, the CADx approach, HIENet, achieved a 76.91 $\pm$ 1.17% (mean $\pm$ s. d.) classification accuracy for four classes of endometrial tissue, namely normal endometrium, endometrial polyp, endometrial hyperplasia, and endometrial adenocarcinoma. Also, HIENet achieved an area-under-the-curve (AUC) of 0.9579 $\pm$ 0.0103 with an 81.04 $\pm$ 3.87% sensitivity and 94.78 $\pm$ 0.87% specificity in a binary classification task that detected endometrioid adenocarcinoma (Malignant). Besides, in the external validation process, HIENet achieved an 84.50% accuracy in the four-class classification task, and it achieved an AUC of 0.9829 with a 77.97% (95% CI, 65.27%-87.71%) sensitivity and 100% (95% CI, 97.42%-100.00%) specificity. In summary, the proposed CADx approach, HIENet, outperformed three human experts and four end-to-end CNN-based classifiers on this small-scale dataset composed of 3,500 hematoxylin and eosin (H&E) images regarding overall classification performance. |
Tasks | |
Published | 2019-04-24 |
URL | http://arxiv.org/abs/1904.10626v1 |
http://arxiv.org/pdf/1904.10626v1.pdf | |
PWC | https://paperswithcode.com/paper/computer-aided-diagnosis-in-histopathological |
Repo | https://github.com/ssea-lab/DL4ETI |
Framework | tf |
Evaluating Rewards for Question Generation Models
Title | Evaluating Rewards for Question Generation Models |
Authors | Tom Hosking, Sebastian Riedel |
Abstract | Recent approaches to question generation have used modifications to a Seq2Seq architecture inspired by advances in machine translation. Models are trained using teacher forcing to optimise only the one-step-ahead prediction. However, at test time, the model is asked to generate a whole sequence, causing errors to propagate through the generation process (exposure bias). A number of authors have proposed countering this bias by optimising for a reward that is less tightly coupled to the training data, using reinforcement learning. We optimise directly for quality metrics, including a novel approach using a discriminator learned directly from the training data. We confirm that policy gradient methods can be used to decouple training from the ground truth, leading to increases in the metrics used as rewards. We perform a human evaluation, and show that although these metrics have previously been assumed to be good proxies for question quality, they are poorly aligned with human judgement and the model simply learns to exploit the weaknesses of the reward source. |
Tasks | Machine Translation, Policy Gradient Methods, Question Generation |
Published | 2019-02-28 |
URL | https://arxiv.org/abs/1902.11049v2 |
https://arxiv.org/pdf/1902.11049v2.pdf | |
PWC | https://paperswithcode.com/paper/evaluating-rewards-for-question-generation |
Repo | https://github.com/bloomsburyai/question-generation |
Framework | tf |
Compositionality decomposed: how do neural networks generalise?
Title | Compositionality decomposed: how do neural networks generalise? |
Authors | Dieuwke Hupkes, Verna Dankers, Mathijs Mul, Elia Bruni |
Abstract | Despite a multitude of empirical studies, little consensus exists on whether neural networks are able to generalise compositionally, a controversy that, in part, stems from a lack of agreement about what it means for a neural model to be compositional. As a response to this controversy, we present a set of tests that provide a bridge between, on the one hand, the vast amount of linguistic and philosophical theory about compositionality of language and, on the other, the successful neural models of language. We collect different interpretations of compositionality and translate them into five theoretically grounded tests for models that are formulated on a task-independent level. In particular, we provide tests to investigate (i) if models systematically recombine known parts and rules (ii) if models can extend their predictions beyond the length they have seen in the training data (iii) if models’ composition operations are local or global (iv) if models’ predictions are robust to synonym substitutions and (v) if models favour rules or exceptions during training. To demonstrate the usefulness of this evaluation paradigm, we instantiate these five tests on a highly compositional data set which we dub PCFG SET and apply the resulting tests to three popular sequence-to-sequence models: a recurrent, a convolution-based and a transformer model. We provide an in-depth analysis of the results, which uncover the strengths and weaknesses of these three architectures and point to potential areas of improvement. |
Tasks | |
Published | 2019-08-22 |
URL | https://arxiv.org/abs/1908.08351v2 |
https://arxiv.org/pdf/1908.08351v2.pdf | |
PWC | https://paperswithcode.com/paper/the-compositionality-of-neural-networks |
Repo | https://github.com/i-machine-think/am-i-compositional |
Framework | pytorch |
Correlating Twitter Language with Community-Level Health Outcomes
Title | Correlating Twitter Language with Community-Level Health Outcomes |
Authors | Arno Schneuwly, Ralf Grubenmann, Séverine Rion Logean, Mark Cieliebak, Martin Jaggi |
Abstract | We study how language on social media is linked to diseases such as atherosclerotic heart disease (AHD), diabetes and various types of cancer. Our proposed model leverages state-of-the-art sentence embeddings, followed by a regression model and clustering, without the need of additional labelled data. It allows to predict community-level medical outcomes from language, and thereby potentially translate these to the individual level. The method is applicable to a wide range of target variables and allows us to discover known and potentially novel correlations of medical outcomes with life-style aspects and other socioeconomic risk factors. |
Tasks | Sentence Embeddings |
Published | 2019-06-13 |
URL | https://arxiv.org/abs/1906.06465v2 |
https://arxiv.org/pdf/1906.06465v2.pdf | |
PWC | https://paperswithcode.com/paper/correlating-twitter-language-with-community |
Repo | https://github.com/epfml/correlating-tweets |
Framework | none |