Paper Group ANR 458
Using Videos to Evaluate Image Model Robustness. Sub-pixel matching method for low-resolution thermal stereo images. Neural Architecture based on Fuzzy Perceptual Representation For Online Multilingual Handwriting Recognition. A Collaborative Framework for High-Definition Mapping. Separating Argument Structure from Logical Structure in AMR. Taking …
Using Videos to Evaluate Image Model Robustness
Title | Using Videos to Evaluate Image Model Robustness |
Authors | Keren Gu, Brandon Yang, Jiquan Ngiam, Quoc Le, Jonathon Shlens |
Abstract | Human visual systems are robust to a wide range of image transformations that are challenging for artificial networks. We present the first study of image model robustness to the minute transformations found across video frames, which we term “natural robustness”. Compared to previous studies on adversarial examples and synthetic distortions, natural robustness captures a more diverse set of common image transformations that occur in the natural environment. Our study across a dozen model architectures shows that more accurate models are more robust to natural transformations, and that robustness to synthetic color distortions is a good proxy for natural robustness. In examining brittleness in videos, we find that majority of the brittleness found in videos lies outside the typical definition of adversarial examples (99.9%). Finally, we investigate training techniques to reduce brittleness and find that no single technique systematically improves natural robustness across twelve tested architectures. |
Tasks | |
Published | 2019-04-22 |
URL | https://arxiv.org/abs/1904.10076v3 |
https://arxiv.org/pdf/1904.10076v3.pdf | |
PWC | https://paperswithcode.com/paper/using-videos-to-evaluate-image-model |
Repo | |
Framework | |
Sub-pixel matching method for low-resolution thermal stereo images
Title | Sub-pixel matching method for low-resolution thermal stereo images |
Authors | Yannick Wend Kuni Zoetgnande, Geoffroy Cormier, Alain-Jérôme Fougères, Jean-Louis Dillenseger |
Abstract | In the context of a localization and tracking application, we developed a stereo vision system based on cheap low-resolution 80x60 pixels thermal cameras. We proposed a threefold sub-pixel stereo matching framework (called ST for Subpixel Thermal): 1) robust features extraction method based on phase congruency, 2) rough matching of these features in pixel precision, and 3) refined matching in sub-pixel accuracy based on local phase coherence. We performed experiments on our very low-resolution thermal images (acquired using a stereo system we manufactured) as for high-resolution images from a benchmark dataset. Even if phase congruency computation time is high, it was able to extract two times more features than state-of-the-art methods such as ORB or SURF. We proposed a modified version of the phase correlation applied in the phase congruency feature space for sub-pixel matching. Using simulated stereo, we investigated how the phase congruency threshold and the sub-image size of sub-pixel matching can influence the accuracy. We then proved that given our stereo setup and the resolution of our images, being wrong of 1 pixel leads to a 500 mm error in the Z position of the point. Finally, we showed that our method could extract four times more matches than a baseline method ORB + OpenCV KNN matching on low-resolution images. Moreover, our matches were more robust. More precisely, when projecting points of a standing person, ST got a standard deviation of 300 mm when ORB + OpenCV KNN gave more than 1000 mm. |
Tasks | Stereo Matching |
Published | 2019-11-30 |
URL | https://arxiv.org/abs/1912.00138v1 |
https://arxiv.org/pdf/1912.00138v1.pdf | |
PWC | https://paperswithcode.com/paper/sub-pixel-matching-method-for-low-resolution |
Repo | |
Framework | |
Neural Architecture based on Fuzzy Perceptual Representation For Online Multilingual Handwriting Recognition
Title | Neural Architecture based on Fuzzy Perceptual Representation For Online Multilingual Handwriting Recognition |
Authors | Hanen Akouaydi, Sourour Njah, Wael Ouarda, Anis Samet, Thameur Dhieb, Mourad Zaied, Adel M. Alimi |
Abstract | Due to the omnipresence of mobile devices, online handwritten scripts have become the most important feeding input to smartphones and tablet devices. To increase online handwriting recognition performance, deeper neural networks have extensively been used. In this context, our paper handles the problem of online handwritten script recognition based on extraction features system and deep approach system for sequences classification. Many solutions have appeared in order to facilitate the recognition of handwriting. Accordingly, we used an existent method and combined with new classifiers in order to get a flexible system. Good results are achieved compared to online characters and words recognition system on Latin and Arabic scripts. The performance of our two proposed systems is assessed by using five databases. Indeed, the recognition rate exceeds 98%. |
Tasks | |
Published | 2019-08-01 |
URL | https://arxiv.org/abs/1908.00634v1 |
https://arxiv.org/pdf/1908.00634v1.pdf | |
PWC | https://paperswithcode.com/paper/neural-architecture-based-on-fuzzy-perceptual |
Repo | |
Framework | |
A Collaborative Framework for High-Definition Mapping
Title | A Collaborative Framework for High-Definition Mapping |
Authors | Alexis Stoven-Dubois, Kuntima Kiala Miguel, Aziz Dziri, Bertrand Leroy, Roland Chapuis |
Abstract | For connected vehicles to have a substantial effect on road safety, it is required that accurate positions and trajectories can be shared. To this end, all vehicles must be accurately geo-localized in a common frame. This can be achieved by merging GNSS (Global Navigation Satellite System) information and visual observations matched with a map of geo-positioned landmarks. Building such a map remains a challenge, and current solutions are facing strong cost-related limitations. We present a collaborative framework for high-definition mapping, in which vehicles equipped with standard sensors, such as a GNSS receiver and a mono-visual camera, update a map of geo-localized landmarks. Our system is composed of two processing blocks: the first one is embedded in each vehicle, and aims at geo-localizing the vehicle and the detected feature marks. The second is operated on cloud servers, and uses observations from all the vehicles to compute updates for the map of geo-positioned landmarks. As the map’s landmarks are detected and positioned by more and more vehicles, the accuracy of the map increases, eventually converging in probability towards a null error. The landmarks’ geo-positions are estimated in a stable and scalable way, enabling to provide dynamic map updates in an automatic manner. |
Tasks | |
Published | 2019-10-14 |
URL | https://arxiv.org/abs/1910.06014v1 |
https://arxiv.org/pdf/1910.06014v1.pdf | |
PWC | https://paperswithcode.com/paper/a-collaborative-framework-for-high-definition |
Repo | |
Framework | |
Separating Argument Structure from Logical Structure in AMR
Title | Separating Argument Structure from Logical Structure in AMR |
Authors | Johan Bos |
Abstract | The AMR (Abstract Meaning Representation) formalism for representing meaning of natural language sentences was not designed to deal with scope and quantifiers. By extending AMR with indices for contexts and formulating constraints on these contexts, a formalism is derived that makes correct prediction for inferences involving negation and bound variables. The attractive core predicate-argument structure of AMR is preserved. The resulting framework is similar to that of Discourse Representation Theory. |
Tasks | |
Published | 2019-08-04 |
URL | https://arxiv.org/abs/1908.01355v1 |
https://arxiv.org/pdf/1908.01355v1.pdf | |
PWC | https://paperswithcode.com/paper/separating-argument-structure-from-logical |
Repo | |
Framework | |
Taking a Stance on Fake News: Towards Automatic Disinformation Assessment via Deep Bidirectional Transformer Language Models for Stance Detection
Title | Taking a Stance on Fake News: Towards Automatic Disinformation Assessment via Deep Bidirectional Transformer Language Models for Stance Detection |
Authors | Chris Dulhanty, Jason L. Deglint, Ibrahim Ben Daya, Alexander Wong |
Abstract | The exponential rise of social media and digital news in the past decade has had the unfortunate consequence of escalating what the United Nations has called a global topic of concern: the growing prevalence of disinformation. Given the complexity and time-consuming nature of combating disinformation through human assessment, one is motivated to explore harnessing AI solutions to automatically assess news articles for the presence of disinformation. A valuable first step towards automatic identification of disinformation is stance detection, where given a claim and a news article, the aim is to predict if the article agrees, disagrees, takes no position, or is unrelated to the claim. Existing approaches in literature have largely relied on hand-engineered features or shallow learned representations (e.g., word embeddings) to encode the claim-article pairs, which can limit the level of representational expressiveness needed to tackle the high complexity of disinformation identification. In this work, we explore the notion of harnessing large-scale deep bidirectional transformer language models for encoding claim-article pairs in an effort to construct state-of-the-art stance detection geared for identifying disinformation. Taking advantage of bidirectional cross-attention between claim-article pairs via pair encoding with self-attention, we construct a large-scale language model for stance detection by performing transfer learning on a RoBERTa deep bidirectional transformer language model, and were able to achieve state-of-the-art performance (weighted accuracy of 90.01%) on the Fake News Challenge Stage 1 (FNC-I) benchmark. These promising results serve as motivation for harnessing such large-scale language models as powerful building blocks for creating effective AI solutions to combat disinformation. |
Tasks | Language Modelling, Stance Detection, Transfer Learning, Word Embeddings |
Published | 2019-11-27 |
URL | https://arxiv.org/abs/1911.11951v1 |
https://arxiv.org/pdf/1911.11951v1.pdf | |
PWC | https://paperswithcode.com/paper/taking-a-stance-on-fake-news-towards |
Repo | |
Framework | |
Fast Algorithms for Computational Optimal Transport and Wasserstein Barycenter
Title | Fast Algorithms for Computational Optimal Transport and Wasserstein Barycenter |
Authors | Wenshuo Guo, Nhat Ho, Michael I. Jordan |
Abstract | We provide theoretical complexity analysis for new algorithms to compute the optimal transport (OT) distance between two discrete probability distributions, and demonstrate their favorable practical performance over state-of-art primal-dual algorithms and their capability in solving other problems in large-scale, such as the Wasserstein barycenter problem for multiple probability distributions. First, we introduce the \emph{accelerated primal-dual randomized coordinate descent} (APDRCD) algorithm for computing the OT distance. We provide its complexity upper bound $\bigOtil(\frac{n^{5/2}}{\varepsilon})$ where $n$ stands for the number of atoms of these probability measures and $\varepsilon > 0$ is the desired accuracy. This complexity bound matches the best known complexities of primal-dual algorithms for the OT problems, including the adaptive primal-dual accelerated gradient descent (APDAGD) and the adaptive primal-dual accelerated mirror descent (APDAMD) algorithms. Then, we demonstrate the better performance of the APDRCD algorithm over the APDAGD and APDAMD algorithms through extensive experimental studies, and further improve its practical performance by proposing a greedy version of it, which we refer to as \emph{accelerated primal-dual greedy coordinate descent} (APDGCD). Finally, we generalize the APDRCD and APDGCD algorithms to distributed algorithms for computing the Wasserstein barycenter for multiple probability distributions. |
Tasks | |
Published | 2019-05-23 |
URL | https://arxiv.org/abs/1905.09952v3 |
https://arxiv.org/pdf/1905.09952v3.pdf | |
PWC | https://paperswithcode.com/paper/accelerated-primal-dual-coordinate-descent |
Repo | |
Framework | |
Deep Ritz revisited
Title | Deep Ritz revisited |
Authors | Johannes Müller, Marius Zeinhofer |
Abstract | Recently, progress has been made in the application of neural networks to the numerical analysis of partial differential equations (PDEs). In the latter the variational formulation of the Poisson problem is used in order to obtain an objective function - a regularised Dirichlet energy - that was used for the optimisation of some neural networks. In this notes we use the notion of $\Gamma$-convergence to show that ReLU networks of growing architecture that are trained with respect to suitably regularised Dirichlet energies converge to the true solution of the Poisson problem. We discuss how this approach generalises to arbitrary variational problems under certain universality assumptions of neural networks and see that this covers some nonlinear stationary PDEs like the $p$-Laplace. |
Tasks | |
Published | 2019-12-09 |
URL | https://arxiv.org/abs/1912.03937v2 |
https://arxiv.org/pdf/1912.03937v2.pdf | |
PWC | https://paperswithcode.com/paper/deep-ritz-revisited |
Repo | |
Framework | |
Data augmentation approaches for improving animal audio classification
Title | Data augmentation approaches for improving animal audio classification |
Authors | Loris Nanni, Gianluca Maguolo, Michelangelo Paci |
Abstract | In this paper we present ensembles of classifiers for automated animal audio classification, exploiting different data augmentation techniques for training Convolutional Neural Networks (CNNs). The specific animal audio classification problems are i) birds and ii) cat sounds, whose datasets are freely available. We train five different CNNs on the original datasets and on their versions augmented by four augmentation protocols, working on the raw audio signals or their representations as spectrograms. We compared our best approaches with the state of the art, showing that we obtain the best recognition rate on the same datasets, without ad hoc parameter optimization. Our study shows that different CNNs can be trained for the purpose of animal audio classification and that their fusion works better than the stand-alone classifiers. To the best of our knowledge this is the largest study on data augmentation for CNNs in animal audio classification audio datasets using the same set of classifiers and parameters. Our MATLAB code is available at https://github.com/LorisNanni. |
Tasks | Audio Classification, Data Augmentation |
Published | 2019-12-16 |
URL | https://arxiv.org/abs/1912.07756v2 |
https://arxiv.org/pdf/1912.07756v2.pdf | |
PWC | https://paperswithcode.com/paper/data-augmentation-approaches-for-improving |
Repo | |
Framework | |
Adversarially Robust Learning Could Leverage Computational Hardness
Title | Adversarially Robust Learning Could Leverage Computational Hardness |
Authors | Sanjam Garg, Somesh Jha, Saeed Mahloujifar, Mohammad Mahmoody |
Abstract | Over recent years, devising classification algorithms that are robust to adversarial perturbations has emerged as a challenging problem. In particular, deep neural nets (DNNs) seem to be susceptible to small imperceptible changes over test instances. However, the line of work in provable robustness, so far, has been focused on information-theoretic robustness, ruling out even the existence of any adversarial examples. In this work, we study whether there is a hope to benefit from algorithmic nature of an attacker that searches for adversarial examples, and ask whether there is any learning task for which it is possible to design classifiers that are only robust against polynomial-time adversaries. Indeed, numerous cryptographic tasks can only be secure against computationally bounded adversaries, and are indeed impossible for computationally unbounded attackers. Thus, it is natural to ask if the same strategy could help robust learning. We show that computational limitation of attackers can indeed be useful in robust learning by demonstrating the possibility of a classifier for some learning task for which computational and information theoretic adversaries of bounded perturbations have very different power. Namely, while computationally unbounded adversaries can attack successfully and find adversarial examples with small perturbation, polynomial time adversaries are unable to do so unless they can break standard cryptographic hardness assumptions. Our results, therefore, indicate that perhaps a similar approach to cryptography (relying on computational hardness) holds promise for achieving computationally robust machine learning. On the reverse directions, we also show that the existence of such learning task in which computational robustness beats information theoretic robustness requires computational hardness by implying (average-case) hardness of NP. |
Tasks | |
Published | 2019-05-28 |
URL | https://arxiv.org/abs/1905.11564v2 |
https://arxiv.org/pdf/1905.11564v2.pdf | |
PWC | https://paperswithcode.com/paper/adversarially-robust-learning-could-leverage |
Repo | |
Framework | |
Architecture and evolution of semantic networks in mathematics texts
Title | Architecture and evolution of semantic networks in mathematics texts |
Authors | Nicolas H. Christianson, Ann Sizemore Blevins, Danielle S. Bassett |
Abstract | Knowledge is a network of interconnected concepts. Yet, precisely how the topological structure of knowledge constrains its acquisition remains unknown, hampering the development of learning enhancement strategies. Here we study the topological structure of semantic networks reflecting mathematical concepts and their relations in college-level linear algebra texts. We hypothesize that these networks will exhibit structural order, reflecting the logical sequence of topics that ensures accessibility. We find that the networks exhibit strong core-periphery architecture, where a dense core of concepts presented early is complemented with a sparse periphery presented evenly throughout the exposition; the latter is composed of many small modules each reflecting more narrow domains. Using tools from applied topology, we find that the expositional evolution of the semantic networks produces and subsequently fills knowledge gaps, and that the density of these gaps tracks negatively with community ratings of each textbook. Broadly, our study lays the groundwork for future efforts developing optimal design principles for textbook exposition and teaching in a classroom setting. |
Tasks | |
Published | 2019-08-14 |
URL | https://arxiv.org/abs/1908.04911v2 |
https://arxiv.org/pdf/1908.04911v2.pdf | |
PWC | https://paperswithcode.com/paper/architecture-and-evolution-of-semantic |
Repo | |
Framework | |
Deep Ensemble Learning for News Stance Detection
Title | Deep Ensemble Learning for News Stance Detection |
Authors | Wenjun Liao, Chenghua Lin |
Abstract | Stance detection in fake news is an important component in news veracity assessment because this process helps fact-checking by understanding stance to a central claim from different information sources. The Fake News Challenge Stage 1 (FNC-1) held in 2017 was setup for this purpose, which involves estimating the stance of a news article body relative to a given headline. This thesis starts from the error analysis for the three top-performing systems in FNC-1. Based on the analysis, a simple but tough-to-beat Multilayer Perceptron system is chosen as the baseline. Afterwards, three approaches are explored to improve baseline.The first approach explores the possibility of improving the prediction accuracy by adding extra keywords features when training a model, where keywords are converted to an indicator vector and then concatenated to the baseline features. A list of keywords is manually selected based on the error analysis, which may best reflect some characteristics of fake news titles and bodies. To make this selection process automatically, three algorithms are created based on Mutual Information (MI) theory: keywords generator based on MI stance class, MI customised class, and Pointwise MI algorithm. The second approach is based on word embedding, where word2vec model is introduced and two document similarities calculation algorithms are implemented: wor2vec cosine similarity and WMD distance. The third approach is ensemble learning. Different models are configured together with two continuous outputs combining algorithms. The 10-fold cross validation reveals that the ensemble of three neural network models trained from simple bag-of-words features gives the best performance. It is therefore selected to compete in FNC-1. After hyperparameters fine tuning, the selected deep ensemble model beats the FNC-1 winner team by a remarkable 34.25 marks under FNC-1’s evaluation metric. |
Tasks | Stance Detection |
Published | 2019-09-13 |
URL | https://arxiv.org/abs/1909.12233v1 |
https://arxiv.org/pdf/1909.12233v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-ensemble-learning-for-news-stance |
Repo | |
Framework | |
Empirical Evaluations of Preprocessing Parameters’ Impact on Predictive Coding’s Effectiveness
Title | Empirical Evaluations of Preprocessing Parameters’ Impact on Predictive Coding’s Effectiveness |
Authors | Rishi Chhatwal, Nathaniel Huber-Fliflet, Robert Keeling, Jianping Zhang, Haozhen Zhao |
Abstract | Predictive coding, once used in only a small fraction of legal and business matters, is now widely deployed to quickly cull through increasingly vast amounts of data and reduce the need for costly and inefficient human document review. Previously, the sole front-end input used to create a predictive model was the exemplar documents (training data) chosen by subject-matter experts. Many predictive coding tools require users to rely on static preprocessing parameters and a single machine learning algorithm to develop the predictive model. Little research has been published discussing the impact preprocessing parameters and learning algorithms have on the effectiveness of the technology. A deeper dive into the generation of a predictive model shows that the settings and algorithm can have a strong effect on the accuracy and efficacy of a predictive coding tool. Understanding how these input parameters affect the output will empower legal teams with the information they need to implement predictive coding as efficiently and effectively as possible. This paper outlines different preprocessing parameters and algorithms as applied to multiple real-world data sets to understand the influence of various approaches. |
Tasks | |
Published | 2019-04-03 |
URL | http://arxiv.org/abs/1904.01718v1 |
http://arxiv.org/pdf/1904.01718v1.pdf | |
PWC | https://paperswithcode.com/paper/empirical-evaluations-of-preprocessing |
Repo | |
Framework | |
Analyzing Cyber-Physical Systems from the Perspective of Artificial Intelligence
Title | Analyzing Cyber-Physical Systems from the Perspective of Artificial Intelligence |
Authors | Eric M. S. P. Veith, Lars Fischer, Martin Tröschel, Astrid Nieße |
Abstract | Principles of modern cyber-physical system (CPS) analysis are based on analytical methods that depend on whether safety or liveness requirements are considered. Complexity is abstracted through different techniques, ranging from stochastic modelling to contracts. However, both distributed heuristics and Artificial Intelligence (AI)-based approaches as well as the user perspective or unpredictable effects, such as accidents or the weather, introduce enough uncertainty to warrant reinforcement-learning-based approaches. This paper compares traditional approaches in the domain of CPS modelling and analysis with the AI researcher perspective to exploring unknown complex systems. |
Tasks | |
Published | 2019-08-21 |
URL | https://arxiv.org/abs/1908.11779v1 |
https://arxiv.org/pdf/1908.11779v1.pdf | |
PWC | https://paperswithcode.com/paper/analyzing-cyber-physical-systems-from-the |
Repo | |
Framework | |
tsmp: An R Package for Time Series with Matrix Profile
Title | tsmp: An R Package for Time Series with Matrix Profile |
Authors | Francisco Bischoff, Pedro Pereira Rodrigues |
Abstract | This article describes tsmp, an R package that implements the matrix profile concept for time series. The tsmp package is a toolkit that allows all-pairs similarity joins, motif, discords and chains discovery, semantic segmentation, etc. Here we describe how the tsmp package may be used by showing some of the use-cases from the original articles and evaluate the algorithm speed in the R environment. This package can be downloaded at https://CRAN.R-project.org/package=tsmp. |
Tasks | Semantic Segmentation, Time Series |
Published | 2019-04-18 |
URL | http://arxiv.org/abs/1904.12626v1 |
http://arxiv.org/pdf/1904.12626v1.pdf | |
PWC | https://paperswithcode.com/paper/190412626 |
Repo | |
Framework | |