Paper Group ANR 1391
Inverse Projection Representation and Category Contribution Rate for Robust Tumor Recognition. Scalable Model Compression by Entropy Penalized Reparameterization. The Use of Mutual Coherence to Prove $\ell^1/\ell^0$-Equivalence in Classification Problems. Structure-adaptive manifold estimation. DGSAN: Discrete Generative Self-Adversarial Network. L …
Inverse Projection Representation and Category Contribution Rate for Robust Tumor Recognition
Title | Inverse Projection Representation and Category Contribution Rate for Robust Tumor Recognition |
Authors | Xiao-Hui Yang, Li Tian, Yun-Mei Chen, Li-Jun Yang, Shuang Xu, Wen-Ming Wu |
Abstract | Sparse representation based classification (SRC) methods have achieved remarkable results. SRC, however, still suffer from requiring enough training samples, insufficient use of test samples and instability of representation. In this paper, a stable inverse projection representation based classification (IPRC) is presented to tackle these problems by effectively using test samples. An IPR is firstly proposed and its feasibility and stability are analyzed. A classification criterion named category contribution rate is constructed to match the IPR and complete classification. Moreover, a statistical measure is introduced to quantify the stability of representation-based classification methods. Based on the IPRC technique, a robust tumor recognition framework is presented by interpreting microarray gene expression data, where a two-stage hybrid gene selection method is introduced to select informative genes. Finally, the functional analysis of candidate’s pathogenicity-related genes is given. Extensive experiments on six public tumor microarray gene expression datasets demonstrate the proposed technique is competitive with state-of-the-art methods. |
Tasks | Sparse Representation-based Classification |
Published | 2019-02-09 |
URL | https://arxiv.org/abs/1902.03510v2 |
https://arxiv.org/pdf/1902.03510v2.pdf | |
PWC | https://paperswithcode.com/paper/inverse-projection-representation-and |
Repo | |
Framework | |
Scalable Model Compression by Entropy Penalized Reparameterization
Title | Scalable Model Compression by Entropy Penalized Reparameterization |
Authors | Deniz Oktay, Johannes Ballé, Saurabh Singh, Abhinav Shrivastava |
Abstract | We describe a simple and general neural network weight compression approach, in which the network parameters (weights and biases) are represented in a “latent” space, amounting to a reparameterization. This space is equipped with a learned probability model, which is used to impose an entropy penalty on the parameter representation during training, and to compress the representation using a simple arithmetic coder after training. Classification accuracy and model compressibility is maximized jointly, with the bitrate–accuracy trade-off specified by a hyperparameter. We evaluate the method on the MNIST, CIFAR-10 and ImageNet classification benchmarks using six distinct model architectures. Our results show that state-of-the-art model compression can be achieved in a scalable and general way without requiring complex procedures such as multi-stage training. |
Tasks | Model Compression |
Published | 2019-06-15 |
URL | https://arxiv.org/abs/1906.06624v3 |
https://arxiv.org/pdf/1906.06624v3.pdf | |
PWC | https://paperswithcode.com/paper/model-compression-by-entropy-penalized |
Repo | |
Framework | |
The Use of Mutual Coherence to Prove $\ell^1/\ell^0$-Equivalence in Classification Problems
Title | The Use of Mutual Coherence to Prove $\ell^1/\ell^0$-Equivalence in Classification Problems |
Authors | Chelsea Weaver, Naoki Saito |
Abstract | We consider the decomposition of a signal over an overcomplete set of vectors. Minimization of the $\ell^1$-norm of the coefficient vector can often retrieve the sparsest solution (so-called “$\ell^1/\ell^0$-equivalence”), a generally NP-hard task, and this fact has powered the field of compressed sensing. Wright et al.‘s sparse representation-based classification (SRC) applies this relationship to machine learning, wherein the signal to be decomposed represents the test sample and columns of the dictionary are training samples. We investigate the relationships between $\ell^1$-minimization, sparsity, and classification accuracy in SRC. After proving that the tractable, deterministic approach to verifying $\ell^1/\ell^0$-equivalence fundamentally conflicts with the high coherence between same-class training samples, we demonstrate that $\ell^1$-minimization can still recover the sparsest solution when the classes are well-separated. Further, using a nonlinear transform so that sparse recovery conditions may be satisfied, we demonstrate that approximate (not strict) equivalence is key to the success of SRC. |
Tasks | Sparse Representation-based Classification |
Published | 2019-01-09 |
URL | http://arxiv.org/abs/1901.02783v1 |
http://arxiv.org/pdf/1901.02783v1.pdf | |
PWC | https://paperswithcode.com/paper/the-use-of-mutual-coherence-to-prove-ell1ell0 |
Repo | |
Framework | |
Structure-adaptive manifold estimation
Title | Structure-adaptive manifold estimation |
Authors | Nikita Puchkin, Vladimir Spokoiny |
Abstract | We consider a problem of manifold estimation from noisy observations. Many manifold learning procedures locally approximate a manifold by a weighted average over a small neighborhood. However, in the presence of large noise, the assigned weights become so corrupted that the averaged estimate shows very poor performance. We suggest a novel computationally efficient structure-adaptive procedure, which simultaneously reconstructs a smooth manifold and estimates projections of the point cloud onto this manifold. The proposed approach iteratively refines the weights on each step, using the structural information obtained at previous steps. After several iterations, we obtain nearly “oracle” weights, so that the final estimates are nearly efficient even in the presence of relatively large noise. In our theoretical study we establish tight lower and upper bounds proving asymptotic optimality of the method for manifold estimation under the Hausdorff loss. Our finite sample study confirms a very reasonable performance of the procedure in comparison with the other methods of manifold estimation. |
Tasks | |
Published | 2019-06-12 |
URL | https://arxiv.org/abs/1906.05014v3 |
https://arxiv.org/pdf/1906.05014v3.pdf | |
PWC | https://paperswithcode.com/paper/structure-adaptive-manifold-estimation |
Repo | |
Framework | |
DGSAN: Discrete Generative Self-Adversarial Network
Title | DGSAN: Discrete Generative Self-Adversarial Network |
Authors | Ehsan Montahaei, Danial Alihosseini, Mahdieh Soleymani Baghshah |
Abstract | Although GAN-based methods have received many achievements in the last few years, they have not been such successful in generating discrete data. The most important challenge of these methods is the difficulty of passing the gradient from the discriminator to the generator when the generator outputs are discrete. Despite several attempts done to alleviate this problem, none of the existing GAN-based methods has improved the performance of text generation (using measures that evaluate both the quality and the diversity of generated samples) compared to a generative RNN that is simply trained by the maximum likelihood approach. In this paper, we propose a new framework for generating discrete data by an adversarial approach in which we do not need to pass the gradient to the generator. In the proposed method, the update of either the generator or the discriminator can be accomplished straightforwardly. Moreover, we leverage the discreteness of data to explicitly model the data distribution and ensure the normalization of the generated distribution and consequently the convergence properties of the proposed method. Experimental results generally show the superiority of the proposed DGSAN method compared to the other GAN-based approaches for generating discrete sequential data. |
Tasks | Text Generation |
Published | 2019-08-24 |
URL | https://arxiv.org/abs/1908.09127v1 |
https://arxiv.org/pdf/1908.09127v1.pdf | |
PWC | https://paperswithcode.com/paper/dgsan-discrete-generative-self-adversarial |
Repo | |
Framework | |
Learning $\textit{Ex Nihilo}$
Title | Learning $\textit{Ex Nihilo}$ |
Authors | Selmer Bringsjord, Naveen Sundar Govindarajulu |
Abstract | This paper introduces, philosophically and to a degree formally, the novel concept of learning $\textit{ex nihilo}$, intended (obviously) to be analogous to the concept of creation $\textit{ex nihilo}$. Learning $\textit{ex nihilo}$ is an agent’s learning “from nothing,” by the suitable employment of schemata for deductive and inductive reasoning. This reasoning must be in machine-verifiable accord with a formal proof/argument theory in a $\textit{cognitive calculus}$ (i.e., roughly, an intensional higher-order multi-operator quantified logic), and this reasoning is applied to percepts received by the agent, in the context of both some prior knowledge, and some prior and current interests. Learning $\textit{ex nihilo}$ is a challenge to contemporary forms of ML, indeed a severe one, but the challenge is offered in the spirt of seeking to stimulate attempts, on the part of non-logicist ML researchers and engineers, to collaborate with those in possession of learning-$\textit{ex nihilo}$ frameworks, and eventually attempts to integrate directly with such frameworks at the implementation level. Such integration will require, among other things, the symbiotic interoperation of state-of-the-art automated reasoners and high-expressivity planners, with statistical/connectionist ML technology. |
Tasks | |
Published | 2019-03-04 |
URL | http://arxiv.org/abs/1903.03515v2 |
http://arxiv.org/pdf/1903.03515v2.pdf | |
PWC | https://paperswithcode.com/paper/learning-textitex-nihilo |
Repo | |
Framework | |
Automated Inline Analysis of Myocardial Perfusion MRI with Deep Learning
Title | Automated Inline Analysis of Myocardial Perfusion MRI with Deep Learning |
Authors | Hui Xue, Rhodri Davies, Louis AE Brown, Kristopher D Knott, Tushar Kotecha, Marianna Fontana, Sven Plein, James C Moon, Peter Kellman |
Abstract | Recent development of quantitative myocardial blood flow (MBF) mapping allows direct evaluation of absolute myocardial perfusion, by computing pixel-wise flow maps. Clinical studies suggest quantitative evaluation would be more desirable for objectivity and efficiency. Objective assessment can be further facilitated by segmenting the myocardium and automatically generating reports following the AHA model. This will free user interaction for analysis and lead to a ‘one-click’ solution to improve workflow. This paper proposes a deep neural network based computational workflow for inline myocardial perfusion analysis. Adenosine stress and rest perfusion scans were acquired from three hospitals. Training set included N=1,825 perfusion series from 1,034 patients. Independent test set included 200 scans from 105 patients. Data were consecutively acquired at each site. A convolution neural net (CNN) model was trained to provide segmentation for LV cavity, myocardium and right ventricular by processing incoming 2D+T perfusion Gd series. Model outputs were compared to manual ground-truth for accuracy of segmentation and flow measures derived on global and per-sector basis. The trained models were integrated onto MR scanners for effective inference. Segmentation accuracy and myocardial flow measures were compared between CNN models and manual ground-truth. The mean Dice ratio of CNN derived myocardium was 0.93 +/- 0.04. Both global flow and per-sector values showed no significant difference, compared to manual results. The AHA 16 segment model was automatically generated and reported on the MR scanner. As a result, the fully automated analysis of perfusion flow mapping was achieved. This solution was integrated on the MR scanner, enabling ‘one-click’ analysis and reporting of myocardial blood flow. |
Tasks | |
Published | 2019-11-02 |
URL | https://arxiv.org/abs/1911.00625v1 |
https://arxiv.org/pdf/1911.00625v1.pdf | |
PWC | https://paperswithcode.com/paper/automated-inline-analysis-of-myocardial |
Repo | |
Framework | |
Recognizing Variables from their Data via Deep Embeddings of Distributions
Title | Recognizing Variables from their Data via Deep Embeddings of Distributions |
Authors | Jonas Mueller, Alex Smola |
Abstract | A key obstacle in automated analytics and meta-learning is the inability to recognize when different datasets contain measurements of the same variable. Because provided attribute labels are often uninformative in practice, this task may be more robustly addressed by leveraging the data values themselves rather than just relying on their arbitrarily selected variable names. Here, we present a computationally efficient method to identify high-confidence variable matches between a given set of data values and a large repository of previously encountered datasets. Our approach enjoys numerous advantages over distributional similarity based techniques because we leverage learned vector embeddings of datasets which adaptively account for natural forms of data variation encountered in practice. Based on the neural architecture of deep sets, our embeddings can be computed for both numeric and string data. In dataset search and schema matching tasks, our methods outperform standard statistical techniques and we find that the learned embeddings generalize well to new data sources. |
Tasks | Meta-Learning |
Published | 2019-09-11 |
URL | https://arxiv.org/abs/1909.04844v1 |
https://arxiv.org/pdf/1909.04844v1.pdf | |
PWC | https://paperswithcode.com/paper/recognizing-variables-from-their-data-via |
Repo | |
Framework | |
PABO: Pseudo Agent-Based Multi-Objective Bayesian Hyperparameter Optimization for Efficient Neural Accelerator Design
Title | PABO: Pseudo Agent-Based Multi-Objective Bayesian Hyperparameter Optimization for Efficient Neural Accelerator Design |
Authors | Maryam Parsa, Aayush Ankit, Amirkoushyar Ziabari, Kaushik Roy |
Abstract | The ever increasing computational cost of Deep Neural Networks (DNN) and the demand for energy efficient hardware for DNN acceleration has made accuracy and hardware cost co-optimization for DNNs tremendously important, especially for edge devices. Owing to the large parameter space and cost of evaluating each parameter in the search space, manually tuning of DNN hyperparameters is impractical. Automatic joint DNN and hardware hyperparameter optimization is indispensable for such problems. Bayesian optimization-based approaches have shown promising results for hyperparameter optimization of DNNs. However, most of these techniques have been developed without considering the underlying hardware, thereby leading to inefficient designs. Further, the few works that perform joint optimization are not generalizable and mainly focus on CMOS-based architectures. In this work, we present a novel pseudo agent-based multi-objective hyperparameter optimization (PABO) for maximizing the DNN performance while obtaining low hardware cost. Compared to the existing methods, our work poses a theoretically different approach for joint optimization of accuracy and hardware cost and focuses on memristive crossbar-based accelerators. PABO uses a supervisor agent to establish connections between the posterior Gaussian distribution models of network accuracy and hardware cost requirements. The agent reduces the mathematical complexity of the co-optimization problem by removing unnecessary computations and updates of acquisition functions, thereby achieving significant speed-ups for the optimization procedure. PABO outputs a Pareto frontier that underscores the trade-offs between designing high-accuracy and hardware efficiency. Our results demonstrate a superior performance compared to the state-of-the-art methods both in terms of accuracy and computational speed (~100x speed up). |
Tasks | Hyperparameter Optimization |
Published | 2019-06-11 |
URL | https://arxiv.org/abs/1906.08167v1 |
https://arxiv.org/pdf/1906.08167v1.pdf | |
PWC | https://paperswithcode.com/paper/pabo-pseudo-agent-based-multi-objective |
Repo | |
Framework | |
Sequential Gaussian Processes for Online Learning of Nonstationary Functions
Title | Sequential Gaussian Processes for Online Learning of Nonstationary Functions |
Authors | Michael Minyi Zhang, Bianca Dumitrascu, Sinead A. Williamson, Barbara E. Engelhardt |
Abstract | Many machine learning problems can be framed in the context of estimating functions, and often these are time-dependent functions that are estimated in real-time as observations arrive. Gaussian processes (GPs) are an attractive choice for modeling real-valued nonlinear functions due to their flexibility and uncertainty quantification. However, the typical GP regression model suffers from several drawbacks: i) Conventional GP inference scales $O(N^{3})$ with respect to the number of observations; ii) updating a GP model sequentially is not trivial; and iii) covariance kernels often enforce stationarity constraints on the function, while GPs with non-stationary covariance kernels are often intractable to use in practice. To overcome these issues, we propose an online sequential Monte Carlo algorithm to fit mixtures of GPs that capture non-stationary behavior while allowing for fast, distributed inference. By formulating hyperparameter optimization as a multi-armed bandit problem, we accelerate mixing for real time inference. Our approach empirically improves performance over state-of-the-art methods for online GP estimation in the context of prediction for simulated non-stationary data and hospital time series data. |
Tasks | Gaussian Processes, Hyperparameter Optimization, Time Series |
Published | 2019-05-24 |
URL | https://arxiv.org/abs/1905.10003v2 |
https://arxiv.org/pdf/1905.10003v2.pdf | |
PWC | https://paperswithcode.com/paper/sequential-gaussian-processes-for-online |
Repo | |
Framework | |
Software Engineering for Fairness: A Case Study with Hyperparameter Optimization
Title | Software Engineering for Fairness: A Case Study with Hyperparameter Optimization |
Authors | Joymallya Chakraborty, Tianpei Xia, Fahmid M. Fahid, Tim Menzies |
Abstract | We assert that it is the ethical duty of software engineers to strive to reduce software discrimination. This paper discusses how that might be done. This is an important topic since machine learning software is increasingly being used to make decisions that affect people’s lives. Potentially, the application of that software will result in fairer decisions because (unlike humans) machine learning software is not biased. However, recent results show that the software within many data mining packages exhibits “group discrimination”; i.e. their decisions are inappropriately affected by “protected attributes”(e.g., race, gender, age, etc.). There has been much prior work on validating the fairness of machine-learning models (by recognizing when such software discrimination exists). But after detection, comes mitigation. What steps can ethical software engineers take to reduce discrimination in the software they produce? This paper shows that making \textit{fairness} as a goal during hyperparameter optimization can (a) preserve the predictive power of a model learned from a data miner while also (b) generates fairer results. To the best of our knowledge, this is the first application of hyperparameter optimization as a tool for software engineers to generate fairer software. |
Tasks | Hyperparameter Optimization |
Published | 2019-05-14 |
URL | https://arxiv.org/abs/1905.05786v2 |
https://arxiv.org/pdf/1905.05786v2.pdf | |
PWC | https://paperswithcode.com/paper/software-engineering-for-fairness-a-case |
Repo | |
Framework | |
Towards More Realistic Human-Robot Conversation: A Seq2Seq-based Body Gesture Interaction System
Title | Towards More Realistic Human-Robot Conversation: A Seq2Seq-based Body Gesture Interaction System |
Authors | Minjie Hua, Fuyuan Shi, Yibing Nan, Kai Wang, Hao Chen, Shiguo Lian |
Abstract | This paper presents a novel system that enables intelligent robots to exhibit realistic body gestures while communicating with humans. The proposed system consists of a listening model and a speaking model used in corresponding conversational phases. Both models are adapted from the sequence-to-sequence (seq2seq) architecture to synthesize body gestures represented by the movements of twelve upper-body keypoints. All the extracted 2D keypoints are firstly 3D-transformed, then rotated and normalized to discard irrelevant information. Substantial videos of human conversations from Youtube are collected and preprocessed to train the listening and speaking models separately, after which the two models are evaluated using metrics of mean squared error (MSE) and cosine similarity on the test dataset. The tuned system is implemented to drive a virtual avatar as well as Pepper, a physical humanoid robot, to demonstrate the improvement on conversational interaction abilities of our method in practice. |
Tasks | |
Published | 2019-05-05 |
URL | https://arxiv.org/abs/1905.01641v3 |
https://arxiv.org/pdf/1905.01641v3.pdf | |
PWC | https://paperswithcode.com/paper/towards-more-realistic-human-robot |
Repo | |
Framework | |
Hybrid Zero Dynamics Inspired Feedback Control Policy Design for 3D Bipedal Locomotion using Reinforcement Learning
Title | Hybrid Zero Dynamics Inspired Feedback Control Policy Design for 3D Bipedal Locomotion using Reinforcement Learning |
Authors | Guillermo A. Castillo, Bowen Weng, Wei Zhang, Ayonga Hereid |
Abstract | This paper presents a novel model-free reinforcement learning (RL) framework to design feedback control policies for 3D bipedal walking. Existing RL algorithms are often trained in an end-to-end manner or rely on prior knowledge of some reference joint trajectories. Different from these studies, we propose a novel policy structure that appropriately incorporates physical insights gained from the hybrid nature of the walking dynamics and the well-established hybrid zero dynamics approach for 3D bipedal walking. As a result, the overall RL framework has several key advantages, including lightweight network structure, short training time, and less dependence on prior knowledge. We demonstrate the effectiveness of the proposed method on Cassie, a challenging 3D bipedal robot. The proposed solution produces stable limit walking cycles that can track various walking speed in different directions. Surprisingly, without specifically trained with disturbances to achieve robustness, it also performs robustly against various adversarial forces applied to the torso towards both the forward and the backward directions. |
Tasks | |
Published | 2019-10-03 |
URL | https://arxiv.org/abs/1910.01748v1 |
https://arxiv.org/pdf/1910.01748v1.pdf | |
PWC | https://paperswithcode.com/paper/hybrid-zero-dynamics-inspired-feedback |
Repo | |
Framework | |
Advances in Online Audio-Visual Meeting Transcription
Title | Advances in Online Audio-Visual Meeting Transcription |
Authors | Takuya Yoshioka, Igor Abramovski, Cem Aksoylar, Zhuo Chen, Moshe David, Dimitrios Dimitriadis, Yifan Gong, Ilya Gurvich, Xuedong Huang, Yan Huang, Aviv Hurvitz, Li Jiang, Sharon Koubi, Eyal Krupka, Ido Leichter, Changliang Liu, Partha Parthasarathy, Alon Vinnikov, Lingfeng Wu, Xiong Xiao, Wayne Xiong, Huaming Wang, Zhenghao Wang, Jun Zhang, Yong Zhao, Tianyan Zhou |
Abstract | This paper describes a system that generates speaker-annotated transcripts of meetings by using a microphone array and a 360-degree camera. The hallmark of the system is its ability to handle overlapped speech, which has been an unsolved problem in realistic settings for over a decade. We show that this problem can be addressed by using a continuous speech separation approach. In addition, we describe an online audio-visual speaker diarization method that leverages face tracking and identification, sound source localization, speaker identification, and, if available, prior speaker information for robustness to various real world challenges. All components are integrated in a meeting transcription framework called SRD, which stands for “separate, recognize, and diarize”. Experimental results using recordings of natural meetings involving up to 11 attendees are reported. The continuous speech separation improves a word error rate (WER) by 16.1% compared with a highly tuned beamformer. When a complete list of meeting attendees is available, the discrepancy between WER and speaker-attributed WER is only 1.0%, indicating accurate word-to-speaker association. This increases marginally to 1.6% when 50% of the attendees are unknown to the system. |
Tasks | Speaker Diarization, Speaker Identification, Speech Separation |
Published | 2019-12-10 |
URL | https://arxiv.org/abs/1912.04979v1 |
https://arxiv.org/pdf/1912.04979v1.pdf | |
PWC | https://paperswithcode.com/paper/advances-in-online-audio-visual-meeting |
Repo | |
Framework | |
The Speed Submission to DIHARD II: Contributions & Lessons Learned
Title | The Speed Submission to DIHARD II: Contributions & Lessons Learned |
Authors | Md Sahidullah, Jose Patino, Samuele Cornell, Ruiqing Yin, Sunit Sivasankaran, Hervé Bredin, Pavel Korshunov, Alessio Brutti, Romain Serizel, Emmanuel Vincent, Nicholas Evans, Sébastien Marcel, Stefano Squartini, Claude Barras |
Abstract | This paper describes the speaker diarization systems developed for the Second DIHARD Speech Diarization Challenge (DIHARD II) by the Speed team. Besides describing the system, which considerably outperformed the challenge baselines, we also focus on the lessons learned from numerous approaches that we tried for single and multi-channel systems. We present several components of our diarization system, including categorization of domains, speech enhancement, speech activity detection, speaker embeddings, clustering methods, resegmentation, and system fusion. We analyze and discuss the effect of each such component on the overall diarization performance within the realistic settings of the challenge. |
Tasks | Action Detection, Activity Detection, Speaker Diarization, Speech Enhancement |
Published | 2019-11-06 |
URL | https://arxiv.org/abs/1911.02388v1 |
https://arxiv.org/pdf/1911.02388v1.pdf | |
PWC | https://paperswithcode.com/paper/the-speed-submission-to-dihard-ii |
Repo | |
Framework | |