Paper Group AWR 364
Investigating Sports Commentator Bias within a Large Corpus of American Football Broadcasts. Emulating Human Developmental Stages with Bayesian Neural Networks. 3D U-Net Based Brain Tumor Segmentation and Survival Days Prediction. Gromov-Wasserstein Averaging in a Riemannian Framework. Sparse Feature Selection in Kernel Discriminant Analysis via Op …
Investigating Sports Commentator Bias within a Large Corpus of American Football Broadcasts
Title | Investigating Sports Commentator Bias within a Large Corpus of American Football Broadcasts |
Authors | Jack Merullo, Luke Yeh, Abram Handler, Alvin Grissom II, Brendan O’Connor, Mohit Iyyer |
Abstract | Sports broadcasters inject drama into play-by-play commentary by building team and player narratives through subjective analyses and anecdotes. Prior studies based on small datasets and manual coding show that such theatrics evince commentator bias in sports broadcasts. To examine this phenomenon, we assemble FOOTBALL, which contains 1,455 broadcast transcripts from American football games across six decades that are automatically annotated with 250K player mentions and linked with racial metadata. We identify major confounding factors for researchers examining racial bias in FOOTBALL, and perform a computational analysis that supports conclusions from prior social science studies. |
Tasks | |
Published | 2019-09-07 |
URL | https://arxiv.org/abs/1909.03343v4 |
https://arxiv.org/pdf/1909.03343v4.pdf | |
PWC | https://paperswithcode.com/paper/investigating-sports-commentator-bias-within |
Repo | https://github.com/jmerullo/football |
Framework | none |
Emulating Human Developmental Stages with Bayesian Neural Networks
Title | Emulating Human Developmental Stages with Bayesian Neural Networks |
Authors | Marcel Binz, Dominik Endres |
Abstract | We compare the acquisition of knowledge in humans and machines. Research from the field of developmental psychology indicates, that human-employed hypothesis are initially guided by simple rules, before evolving into more complex theories. This observation is shared across many tasks and domains. We investigate whether stages of development in artificial learning systems are based on the same characteristics. We operationalize developmental stages as the size of the data-set, on which the artificial system is trained. For our analysis we look at the developmental progress of Bayesian Neural Networks on three different data-sets, including occlusion, support and quantity comparison tasks. We compare the results with prior research from developmental psychology and find agreement between the family of optimized models and pattern of development observed in infants and children on all three tasks, indicating common principles for the acquisition of knowledge. |
Tasks | |
Published | 2019-02-20 |
URL | http://arxiv.org/abs/1902.07579v1 |
http://arxiv.org/pdf/1902.07579v1.pdf | |
PWC | https://paperswithcode.com/paper/emulating-human-developmental-stages-with |
Repo | https://github.com/marcelbinz/Developmental-Stages-of-BNNs |
Framework | pytorch |
3D U-Net Based Brain Tumor Segmentation and Survival Days Prediction
Title | 3D U-Net Based Brain Tumor Segmentation and Survival Days Prediction |
Authors | Feifan Wang, Runzhou Jiang, Liqin Zheng, Chun Meng, Bharat Biswal |
Abstract | Past few years have witnessed the prevalence of deep learning in many application scenarios, among which is medical image processing. Diagnosis and treatment of brain tumors requires an accurate and reliable segmentation of brain tumors as a prerequisite. However, such work conventionally requires brain surgeons significant amount of time. Computer vision techniques could provide surgeons a relief from the tedious marking procedure. In this paper, a 3D U-net based deep learning model has been trained with the help of brain-wise normalization and patching strategies for the brain tumor segmentation task in the BraTS 2019 competition. Dice coefficients for enhancing tumor, tumor core, and the whole tumor are 0.737, 0.807 and 0.894 respectively on the validation dataset. These three values on the test dataset are 0.778, 0.798 and 0.852. Furthermore, numerical features including ratio of tumor size to brain size and the area of tumor surface as well as age of subjects are extracted from predicted tumor labels and have been used for the overall survival days prediction task. The accuracy could be 0.448 on the validation dataset, and 0.551 on the final test dataset. |
Tasks | Brain Tumor Segmentation |
Published | 2019-09-15 |
URL | https://arxiv.org/abs/1909.12901v2 |
https://arxiv.org/pdf/1909.12901v2.pdf | |
PWC | https://paperswithcode.com/paper/brain-wise-tumor-segmentation-and-patient |
Repo | https://github.com/woodywff/brats_2019 |
Framework | tf |
Gromov-Wasserstein Averaging in a Riemannian Framework
Title | Gromov-Wasserstein Averaging in a Riemannian Framework |
Authors | Samir Chowdhury, Tom Needham |
Abstract | We introduce a theoretical framework for performing statistical tasks—including, but not limited to, averaging and principal component analysis—on the space of (possibly asymmetric) matrices with arbitrary entries and sizes. This is carried out under the lens of the Gromov-Wasserstein (GW) distance, and our methods translate the Riemannian framework of GW distances developed by Sturm into practical, implementable tools for network data analysis. Our methods are illustrated on datasets of asymmetric stochastic blockmodel networks and planar shapes viewed as metric spaces. On the theoretical front, we supplement the work of Sturm by producing additional results on the tangent structure of this “space of spaces”, as well as on the gradient flow of the Fr'{e}chet functional on this space. |
Tasks | |
Published | 2019-10-10 |
URL | https://arxiv.org/abs/1910.04308v1 |
https://arxiv.org/pdf/1910.04308v1.pdf | |
PWC | https://paperswithcode.com/paper/gromov-wasserstein-averaging-in-a-riemannian |
Repo | https://github.com/trneedham/gromov-wasserstein-statistics |
Framework | none |
Sparse Feature Selection in Kernel Discriminant Analysis via Optimal Scoring
Title | Sparse Feature Selection in Kernel Discriminant Analysis via Optimal Scoring |
Authors | Alexander F. Lapanowski, Irina Gaynanova |
Abstract | We consider the two-group classification problem and propose a kernel classifier based on the optimal scoring framework. Unlike previous approaches, we provide theoretical guarantees on the expected risk consistency of the method. We also allow for feature selection by imposing structured sparsity using weighted kernels. We propose fully-automated methods for selection of all tuning parameters, and in particular adapt kernel shrinkage ideas for ridge parameter selection. Numerical studies demonstrate the superior classification performance of the proposed approach compared to existing nonparametric classifiers. |
Tasks | Feature Selection |
Published | 2019-02-12 |
URL | http://arxiv.org/abs/1902.04248v1 |
http://arxiv.org/pdf/1902.04248v1.pdf | |
PWC | https://paperswithcode.com/paper/sparse-feature-selection-in-kernel |
Repo | https://github.com/aflapan/sparseKOS |
Framework | none |
HINT: Hierarchical Invertible Neural Transport for Density Estimation and Bayesian Inference
Title | HINT: Hierarchical Invertible Neural Transport for Density Estimation and Bayesian Inference |
Authors | Jakob Kruse, Gianluca Detommaso, Robert Scheichl, Ullrich Köthe |
Abstract | A large proportion of recent invertible neural architectures is based on a coupling block design. It operates by dividing incoming variables into two sub-spaces, one of which parameterizes an easily invertible (usually affine) transformation that is applied to the other. While the Jacobian of such a transformation is triangular, it is very sparse and thus may lack expressiveness. This work presents a simple remedy by noting that (affine) coupling can be repeated recursively within the resulting sub-spaces, leading to an efficiently invertible block with dense triangular Jacobian. By formulating our recursive coupling scheme via a hierarchical architecture, HINT allows sampling from a joint distribution p(y,x) and the corresponding posterior p(xy) using a single invertible network. We demonstrate the power of our method for density estimation and Bayesian inference on a novel data set of 2D shapes in Fourier parameterization, which enables consistent visualization of samples for different dimensionalities. |
Tasks | Bayesian Inference, Density Estimation |
Published | 2019-05-25 |
URL | https://arxiv.org/abs/1905.10687v3 |
https://arxiv.org/pdf/1905.10687v3.pdf | |
PWC | https://paperswithcode.com/paper/hint-hierarchical-invertible-neural-transport |
Repo | https://github.com/VLL-HD/HINT |
Framework | pytorch |
Point-Based Multi-View Stereo Network
Title | Point-Based Multi-View Stereo Network |
Authors | Rui Chen, Songfang Han, Jing Xu, Hao Su |
Abstract | We introduce Point-MVSNet, a novel point-based deep framework for multi-view stereo (MVS). Distinct from existing cost volume approaches, our method directly processes the target scene as point clouds. More specifically, our method predicts the depth in a coarse-to-fine manner. We first generate a coarse depth map, convert it into a point cloud and refine the point cloud iteratively by estimating the residual between the depth of the current iteration and that of the ground truth. Our network leverages 3D geometry priors and 2D texture information jointly and effectively by fusing them into a feature-augmented point cloud, and processes the point cloud to estimate the 3D flow for each point. This point-based architecture allows higher accuracy, more computational efficiency and more flexibility than cost-volume-based counterparts. Experimental results show that our approach achieves a significant improvement in reconstruction quality compared with state-of-the-art methods on the DTU and the Tanks and Temples dataset. Our source code and trained models are available at https://github.com/callmeray/PointMVSNet . |
Tasks | |
Published | 2019-08-12 |
URL | https://arxiv.org/abs/1908.04422v1 |
https://arxiv.org/pdf/1908.04422v1.pdf | |
PWC | https://paperswithcode.com/paper/point-based-multi-view-stereo-network |
Repo | https://github.com/callmeray/PointMVSNet |
Framework | pytorch |
Discrete Flows: Invertible Generative Models of Discrete Data
Title | Discrete Flows: Invertible Generative Models of Discrete Data |
Authors | Dustin Tran, Keyon Vafa, Kumar Krishna Agrawal, Laurent Dinh, Ben Poole |
Abstract | While normalizing flows have led to significant advances in modeling high-dimensional continuous distributions, their applicability to discrete distributions remains unknown. In this paper, we show that flows can in fact be extended to discrete events—and under a simple change-of-variables formula not requiring log-determinant-Jacobian computations. Discrete flows have numerous applications. We consider two flow architectures: discrete autoregressive flows that enable bidirectionality, allowing, for example, tokens in text to depend on both left-to-right and right-to-left contexts in an exact language model; and discrete bipartite flows that enable efficient non-autoregressive generation as in RealNVP. Empirically, we find that discrete autoregressive flows outperform autoregressive baselines on synthetic discrete distributions, an addition task, and Potts models; and bipartite flows can obtain competitive performance with autoregressive baselines on character-level language modeling for Penn Tree Bank and text8. |
Tasks | Language Modelling |
Published | 2019-05-24 |
URL | https://arxiv.org/abs/1905.10347v1 |
https://arxiv.org/pdf/1905.10347v1.pdf | |
PWC | https://paperswithcode.com/paper/discrete-flows-invertible-generative-models |
Repo | https://github.com/TrentBrick/PyTorchDiscreteFlows |
Framework | pytorch |
Combining Sentiment Lexica with a Multi-View Variational Autoencoder
Title | Combining Sentiment Lexica with a Multi-View Variational Autoencoder |
Authors | Alexander Hoyle, Lawrence Wolf-Sonkin, Hanna Wallach, Ryan Cotterell, Isabelle Augenstein |
Abstract | When assigning quantitative labels to a dataset, different methodologies may rely on different scales. In particular, when assigning polarities to words in a sentiment lexicon, annotators may use binary, categorical, or continuous labels. Naturally, it is of interest to unify these labels from disparate scales to both achieve maximal coverage over words and to create a single, more robust sentiment lexicon while retaining scale coherence. We introduce a generative model of sentiment lexica to combine disparate scales into a common latent representation. We realize this model with a novel multi-view variational autoencoder (VAE), called SentiVAE. We evaluate our approach via a downstream text classification task involving nine English-Language sentiment analysis datasets; our representation outperforms six individual sentiment lexica, as well as a straightforward combination thereof. |
Tasks | Sentiment Analysis, Text Classification |
Published | 2019-04-05 |
URL | http://arxiv.org/abs/1904.02839v1 |
http://arxiv.org/pdf/1904.02839v1.pdf | |
PWC | https://paperswithcode.com/paper/combining-sentiment-lexica-with-a-multi-view |
Repo | https://github.com/ahoho/SentiVAE |
Framework | pytorch |
HoMM: Higher-order Moment Matching for Unsupervised Domain Adaptation
Title | HoMM: Higher-order Moment Matching for Unsupervised Domain Adaptation |
Authors | Chao Chen, Zhihang Fu, Zhihong Chen, Sheng Jin, Zhaowei Cheng, Xinyu Jin, Xian-Sheng Hua |
Abstract | Minimizing the discrepancy of feature distributions between different domains is one of the most promising directions in unsupervised domain adaptation. From the perspective of distribution matching, most existing discrepancy-based methods are designed to match the second-order or lower statistics, which however, have limited expression of statistical characteristic for non-Gaussian distributions. In this work, we explore the benefits of using higher-order statistics (mainly refer to third-order and fourth-order statistics) for domain matching. We propose a Higher-order Moment Matching (HoMM) method, and further extend the HoMM into reproducing kernel Hilbert spaces (RKHS). In particular, our proposed HoMM can perform arbitrary-order moment tensor matching, we show that the first-order HoMM is equivalent to Maximum Mean Discrepancy (MMD) and the second-order HoMM is equivalent to Correlation Alignment (CORAL). Moreover, the third-order and the fourth-order moment tensor matching are expected to perform comprehensive domain alignment as higher-order statistics can approximate more complex, non-Gaussian distributions. Besides, we also exploit the pseudo-labeled target samples to learn discriminative representations in the target domain, which further improves the transfer performance. Extensive experiments are conducted, showing that our proposed HoMM consistently outperforms the existing moment matching methods by a large margin. Codes are available at \url{https://github.com/chenchao666/HoMM-Master} |
Tasks | Domain Adaptation, Unsupervised Domain Adaptation |
Published | 2019-12-27 |
URL | https://arxiv.org/abs/1912.11976v1 |
https://arxiv.org/pdf/1912.11976v1.pdf | |
PWC | https://paperswithcode.com/paper/homm-higher-order-moment-matching-for |
Repo | https://github.com/chenchao666/HoMM-Master |
Framework | tf |
Coursera Corpus Mining and Multistage Fine-Tuning for Improving Lectures Translation
Title | Coursera Corpus Mining and Multistage Fine-Tuning for Improving Lectures Translation |
Authors | Haiyue Song, Raj Dabre, Atsushi Fujita, Sadao Kurohashi |
Abstract | Lectures translation is a case of spoken language translation and there is a lack of publicly available parallel corpora for this purpose. To address this, we examine a language independent framework for parallel corpus mining which is a quick and effective way to mine a parallel corpus from publicly available lectures at Coursera. Our approach determines sentence alignments, relying on machine translation and cosine similarity over continuous-space sentence representations. We also show how to use the resulting corpora in a multistage fine-tuning based domain adaptation for high-quality lectures translation. For Japanese–English lectures translation, we extracted parallel data of approximately 40,000 lines and created development and test sets through manual filtering for benchmarking translation performance. We demonstrate that the mined corpus greatly enhances the quality of translation when used in conjunction with out-of-domain parallel corpora via multistage training. This paper also suggests some guidelines to gather and clean corpora, mine parallel sentences, address noise in the mined data, and create high-quality evaluation splits. For the sake of reproducibility, we will release our code for parallel data creation. |
Tasks | Domain Adaptation, Machine Translation, Parallel Corpus Mining |
Published | 2019-12-26 |
URL | https://arxiv.org/abs/1912.11739v2 |
https://arxiv.org/pdf/1912.11739v2.pdf | |
PWC | https://paperswithcode.com/paper/coursera-corpus-mining-and-multistage-fine |
Repo | https://github.com/shyyhs/CourseraParallelCorpusMining |
Framework | none |
Adversarial Domain Adaptation with Domain Mixup
Title | Adversarial Domain Adaptation with Domain Mixup |
Authors | Minghao Xu, Jian Zhang, Bingbing Ni, Teng Li, Chengjie Wang, Qi Tian, Wenjun Zhang |
Abstract | Recent works on domain adaptation reveal the effectiveness of adversarial learning on filling the discrepancy between source and target domains. However, two common limitations exist in current adversarial-learning-based methods. First, samples from two domains alone are not sufficient to ensure domain-invariance at most part of latent space. Second, the domain discriminator involved in these methods can only judge real or fake with the guidance of hard label, while it is more reasonable to use soft scores to evaluate the generated images or features, i.e., to fully utilize the inter-domain information. In this paper, we present adversarial domain adaptation with domain mixup (DM-ADA), which guarantees domain-invariance in a more continuous latent space and guides the domain discriminator in judging samples’ difference relative to source and target domains. Domain mixup is jointly conducted on pixel and feature level to improve the robustness of models. Extensive experiments prove that the proposed approach can achieve superior performance on tasks with various degrees of domain shift and data complexity. |
Tasks | Domain Adaptation |
Published | 2019-12-04 |
URL | https://arxiv.org/abs/1912.01805v1 |
https://arxiv.org/pdf/1912.01805v1.pdf | |
PWC | https://paperswithcode.com/paper/adversarial-domain-adaptation-with-domain |
Repo | https://github.com/ChrisAllenMing/Mixup_for_UDA |
Framework | pytorch |
Cross-Domain Complementary Learning with Synthetic Data for Multi-Person Part Segmentation
Title | Cross-Domain Complementary Learning with Synthetic Data for Multi-Person Part Segmentation |
Authors | Kevin Lin, Lijuan Wang, Kun Luo, Yinpeng Chen, Zicheng Liu, Ming-Ting Sun |
Abstract | The success of supervised deep learning depends on the training labels. However, data labeling at pixel-level is very expensive, and people have been exploring synthetic data as an alternative. Even though it is easy to generate labels for synthetic data, the quality gap makes it challenging to transfer knowledge from synthetic data to real data. In this paper, we propose a novel technique, called cross-domain complementary learning that takes advantage of the rich variations of real data and the easily obtainable labels of synthetic data to learn multi-person part segmentation on real images without any human-annotated segmentation labels. To make sure the synthetic data and real data are aligned in a common latent space, we use an auxiliary task of human pose estimation to bridge the two domains. Without any real part segmentation training data, our method performs comparably to several supervised state-of-the-art approaches which require real part segmentation training data on Pascal-Person-Parts and COCO-DensePose datasets. We further demonstrate the generalizability of our method on predicting novel keypoints in the wild where no real data labels are available for the novel keypoints. |
Tasks | Domain Adaptation, Human Part Segmentation, Multi-Human Parsing, Pose Estimation |
Published | 2019-07-11 |
URL | https://arxiv.org/abs/1907.05193v1 |
https://arxiv.org/pdf/1907.05193v1.pdf | |
PWC | https://paperswithcode.com/paper/cross-domain-complementary-learning-with |
Repo | https://github.com/kevinlin311tw/CDCL-human-part-segmentation |
Framework | tf |
Random Projection in Neural Episodic Control
Title | Random Projection in Neural Episodic Control |
Authors | Daichi Nishio, Satoshi Yamane |
Abstract | End-to-end deep reinforcement learning has enabled agents to learn with little preprocessing by humans. However, it is still difficult to learn stably and efficiently because the learning method usually uses a nonlinear function approximation. Neural Episodic Control (NEC), which has been proposed in order to improve sample efficiency, is able to learn stably by estimating action values using a non-parametric method. In this paper, we propose an architecture that incorporates random projection into NEC to train with more stability. In addition, we verify the effectiveness of our architecture by Atari’s five games. The main idea is to reduce the number of parameters that have to learn by replacing neural networks with random projection in order to reduce dimensions while keeping the learning end-to-end. |
Tasks | |
Published | 2019-04-03 |
URL | http://arxiv.org/abs/1904.01790v2 |
http://arxiv.org/pdf/1904.01790v2.pdf | |
PWC | https://paperswithcode.com/paper/random-projection-in-neural-episodic-control |
Repo | https://github.com/dnishio/NEC-RP |
Framework | none |
Detecting Urban Changes with Recurrent Neural Networks from Multitemporal Sentinel-2 Data
Title | Detecting Urban Changes with Recurrent Neural Networks from Multitemporal Sentinel-2 Data |
Authors | Maria Papadomanolaki, Sagar Verma, Maria Vakalopoulou, Siddharth Gupta, Konstantinos Karantzalos |
Abstract | \begin{abstract} The advent of multitemporal high resolution data, like the Copernicus Sentinel-2, has enhanced significantly the potential of monitoring the earth’s surface and environmental dynamics. In this paper, we present a novel deep learning framework for urban change detection which combines state-of-the-art fully convolutional networks (similar to U-Net) for feature representation and powerful recurrent networks (such as LSTMs) for temporal modeling. We report our results on the recently publicly available bi-temporal Onera Satellite Change Detection (OSCD) Sentinel-2 dataset, enhancing the temporal information with additional images of the same region on different dates. Moreover, we evaluate the performance of the recurrent networks as well as the use of the additional dates on the unseen test-set using an ensemble cross-validation strategy. All the developed models during the validation phase have scored an overall accuracy of more than 95%, while the use of LSTMs and further temporal information, boost the F1 rate of the change class by an additional 1.5%. |
Tasks | |
Published | 2019-10-17 |
URL | https://arxiv.org/abs/1910.07778v1 |
https://arxiv.org/pdf/1910.07778v1.pdf | |
PWC | https://paperswithcode.com/paper/detecting-urban-changes-with-recurrent-neural |
Repo | https://github.com/mpapadomanolaki/UNetLSTM |
Framework | pytorch |