Paper Group ANR 1148
Shapelets for earthquake detection. A greedy constructive algorithm for the optimization of neural network architectures. Improving 3D Object Detection for Pedestrians with Virtual Multi-View Synthesis Orientation Estimation. ORC Layout: Adaptive GUI Layout with OR-Constraints. Rethinking Exposure Bias In Language Modeling. Mean Field Limit of the …
Shapelets for earthquake detection
Title | Shapelets for earthquake detection |
Authors | Monica Arul, Ahsan Kareem |
Abstract | This paper introduces EQShapelets (EarthQuake Shapelets) a time-series shape-based approach embedded in machine learning to autonomously detect earthquakes. It promises to overcome the challenges in the field of seismology related to automated detection and cataloging of earthquakes. EQShapelets are amplitude and phase-independent, i.e., their detection sensitivity is irrespective of the magnitude of the earthquake and the time of occurrence. They are also robust to noise and other spurious signals. The detection capability of EQShapelets is tested on one week of continuous seismic data provided by the Northern California Seismic Network (NCSN) obtained from a station in central California near the Calaveras Fault. EQShapelets combined with a Random Forest classifier, detected all of the cataloged earthquakes and 281 uncataloged events with lower false detection rate thus offering a better performance than autocorrelation and FAST algorithms. The primary advantage of EQShapelets over competing methods is the interpretability and insight it offers. Shape-based approaches are intuitive, visually meaningful and offers immediate insight into the problem domain that goes beyond their use in accurate detection. EQShapelets, if implemented at a large scale, can significantly reduce catalog completeness magnitudes and can serve as an effective tool for near real-time earthquake monitoring and cataloging. |
Tasks | Time Series |
Published | 2019-11-20 |
URL | https://arxiv.org/abs/1911.09086v1 |
https://arxiv.org/pdf/1911.09086v1.pdf | |
PWC | https://paperswithcode.com/paper/shapelets-for-earthquake-detection |
Repo | |
Framework | |
A greedy constructive algorithm for the optimization of neural network architectures
Title | A greedy constructive algorithm for the optimization of neural network architectures |
Authors | Massimiliano Lupo Pasini, Junqi Yin, Ying Wai Li, Markus Eisenbach |
Abstract | In this work we propose a new method to optimize the architecture of an artificial neural network. The algorithm proposed, called Greedy Search for Neural Network Architecture, aims to minimize the complexity of the architecture search and the complexity of the final model selected without compromising the predictive performance. The reduction of the computational cost makes this approach appealing for two reasons. Firstly, there is a need from domain scientists to easily interpret predictions returned by a deep learning model and this tends to be cumbersome when neural networks have complex structures. Secondly, the use of neural networks is challenging in situations with compute/memory limitations. Promising numerical results show that our method is competitive against other hyperparameter optimization algorithms for attainable performance and computational cost. We also generalize the definition of adjusted score from linear regression models to neural networks. Numerical experiments are presented to show that the adjusted score can boost the greedy search to favor smaller architectures over larger ones without compromising the predictive performance. |
Tasks | Hyperparameter Optimization |
Published | 2019-09-07 |
URL | https://arxiv.org/abs/1909.03306v1 |
https://arxiv.org/pdf/1909.03306v1.pdf | |
PWC | https://paperswithcode.com/paper/a-greedy-constructive-algorithm-for-the |
Repo | |
Framework | |
Improving 3D Object Detection for Pedestrians with Virtual Multi-View Synthesis Orientation Estimation
Title | Improving 3D Object Detection for Pedestrians with Virtual Multi-View Synthesis Orientation Estimation |
Authors | Jason Ku, Alex D. Pon, Sean Walsh, Steven L. Waslander |
Abstract | Accurately estimating the orientation of pedestrians is an important and challenging task for autonomous driving because this information is essential for tracking and predicting pedestrian behavior. This paper presents a flexible Virtual Multi-View Synthesis module that can be adopted into 3D object detection methods to improve orientation estimation. The module uses a multi-step process to acquire the fine-grained semantic information required for accurate orientation estimation. First, the scene’s point cloud is densified using a structure preserving depth completion algorithm and each point is colorized using its corresponding RGB pixel. Next, virtual cameras are placed around each object in the densified point cloud to generate novel viewpoints, which preserve the object’s appearance. We show that this module greatly improves the orientation estimation on the challenging pedestrian class on the KITTI benchmark. When used with the open-source 3D detector AVOD-FPN, we outperform all other published methods on the pedestrian Orientation, 3D, and Bird’s Eye View benchmarks. |
Tasks | 3D Object Detection, Autonomous Driving, Depth Completion, Object Detection |
Published | 2019-07-15 |
URL | https://arxiv.org/abs/1907.06777v1 |
https://arxiv.org/pdf/1907.06777v1.pdf | |
PWC | https://paperswithcode.com/paper/improving-3d-object-detection-for-pedestrians |
Repo | |
Framework | |
ORC Layout: Adaptive GUI Layout with OR-Constraints
Title | ORC Layout: Adaptive GUI Layout with OR-Constraints |
Authors | Yue Jiang, Ruofei Du, Christof Lutteroth, Wolfgang Stuerzlinger |
Abstract | We propose a novel approach for constraint-based graphical user interface (GUI) layout based on OR-constraints (ORC) in standard soft/hard linear constraint systems. ORC layout unifies grid layout and flow layout, supporting both their features as well as cases where grid and flow layouts individually fail. We describe ORC design patterns that enable designers to safely create flexible layouts that work across different screen sizes and orientations. We also present the ORC Editor, a GUI editor that enables designers to apply ORC in a safe and effective manner, mixing grid, flow and new ORC layout features as appropriate. We demonstrate that our prototype can adapt layouts to screens with different aspect ratios with only a single layout specification, easing the burden of GUI maintenance. Finally, we show that ORC specifications can be modified interactively and solved efficiently at runtime. |
Tasks | |
Published | 2019-12-17 |
URL | https://arxiv.org/abs/1912.07827v1 |
https://arxiv.org/pdf/1912.07827v1.pdf | |
PWC | https://paperswithcode.com/paper/orc-layout-adaptive-gui-layout-with-or |
Repo | |
Framework | |
Rethinking Exposure Bias In Language Modeling
Title | Rethinking Exposure Bias In Language Modeling |
Authors | Yifan Xu, Kening Zhang, Haoyu Dong, Yuezhou Sun, Wenlong Zhao, Zhuowen Tu |
Abstract | Exposure bias describes the phenomenon that a language model trained under the teacher forcing schema may perform poorly at the inference stage when its predictions are conditioned on its previous predictions unseen from the training corpus. Recently, several generative adversarial networks (GANs) and reinforcement learning (RL) methods have been introduced to alleviate this problem. Nonetheless, a common issue in RL and GANs training is the sparsity of reward signals. In this paper, we adopt two simple strategies, multi-range reinforcing, and multi-entropy sampling, to amplify and denoise the reward signal. Our model produces an improvement over competing models with regards to BLEU scores and road exam, a new metric we designed to measure the robustness against exposure bias in language models. |
Tasks | Language Modelling |
Published | 2019-10-13 |
URL | https://arxiv.org/abs/1910.11235v2 |
https://arxiv.org/pdf/1910.11235v2.pdf | |
PWC | https://paperswithcode.com/paper/rethinking-exposure-bias-in-language-modeling |
Repo | |
Framework | |
Mean Field Limit of the Learning Dynamics of Multilayer Neural Networks
Title | Mean Field Limit of the Learning Dynamics of Multilayer Neural Networks |
Authors | Phan-Minh Nguyen |
Abstract | Can multilayer neural networks – typically constructed as highly complex structures with many nonlinearly activated neurons across layers – behave in a non-trivial way that yet simplifies away a major part of their complexities? In this work, we uncover a phenomenon in which the behavior of these complex networks – under suitable scalings and stochastic gradient descent dynamics – becomes independent of the number of neurons as this number grows sufficiently large. We develop a formalism in which this many-neurons limiting behavior is captured by a set of equations, thereby exposing a previously unknown operating regime of these networks. While the current pursuit is mathematically non-rigorous, it is complemented with several experiments that validate the existence of this behavior. |
Tasks | |
Published | 2019-02-07 |
URL | http://arxiv.org/abs/1902.02880v1 |
http://arxiv.org/pdf/1902.02880v1.pdf | |
PWC | https://paperswithcode.com/paper/mean-field-limit-of-the-learning-dynamics-of |
Repo | |
Framework | |
DSNet: An Efficient CNN for Road Scene Segmentation
Title | DSNet: An Efficient CNN for Road Scene Segmentation |
Authors | Ping-Rong Chen, Hsueh-Ming Hang, Sheng-Wei Chan, Jing-Jhih Lin |
Abstract | Road scene understanding is a critical component in an autonomous driving system. Although the deep learning-based road scene segmentation can achieve very high accuracy, its complexity is also very high for developing real-time applications. It is challenging to design a neural net with high accuracy and low computational complexity. To address this issue, we investigate the advantages and disadvantages of several popular CNN architectures in terms of speed, storage and segmentation accuracy. We start from the Fully Convolutional Network (FCN) with VGG, and then we study ResNet and DenseNet. Through detailed experiments, we pick up the favorable components from the existing architectures and at the end, we construct a light-weight network architecture based on the DenseNet. Our proposed network, called DSNet, demonstrates a real-time testing (inferencing) ability (on the popular GPU platform) and it maintains an accuracy comparable with most previous systems. We test our system on several datasets including the challenging Cityscapes dataset (resolution of 1024x512) with an mIoU of about 69.1 % and runtime of 0.0147 second per image on a single GTX 1080Ti. We also design a more accurate model but at the price of a slower speed, which has an mIoU of about 72.6 % on the CamVid dataset. |
Tasks | Autonomous Driving, Scene Segmentation, Scene Understanding |
Published | 2019-04-10 |
URL | http://arxiv.org/abs/1904.05022v1 |
http://arxiv.org/pdf/1904.05022v1.pdf | |
PWC | https://paperswithcode.com/paper/dsnet-an-efficient-cnn-for-road-scene |
Repo | |
Framework | |
Potential Field: Interpretable and Unified Representation for Trajectory Prediction
Title | Potential Field: Interpretable and Unified Representation for Trajectory Prediction |
Authors | Shan Su, Cheng Peng, Jianbo Shi, Chiho Choi |
Abstract | Predicting an agent’s future trajectory is a challenging task given the complicated stimuli (environmental/inertial/social) of motion. Prior works learn individual stimulus from different modules and fuse the representations in an end-to-end manner, which makes it hard to understand what are actually captured and how they are fused. In this work, we borrow the notion of potential field from physics as an interpretable and unified representation to model all stimuli. This allows us to not only supervise the intermediate learning process, but also have a coherent method to fuse the information of different sources. From the generated potential fields, we further estimate future motion direction and speed, which are modeled as Gaussian distributions to account for the multi-modal nature of the problem. The final prediction results are generated by recurrently moving past location based on the estimated motion direction and speed. We show state-of-the-art results on the ETH, UCY, and Stanford Drone datasets. |
Tasks | Trajectory Prediction |
Published | 2019-11-18 |
URL | https://arxiv.org/abs/1911.07414v1 |
https://arxiv.org/pdf/1911.07414v1.pdf | |
PWC | https://paperswithcode.com/paper/potential-field-interpretable-and-unified |
Repo | |
Framework | |
CBCL: Brain-Inspired Model for RGB-D Indoor Scene Classification
Title | CBCL: Brain-Inspired Model for RGB-D Indoor Scene Classification |
Authors | Ali Ayub, Alan Wagner |
Abstract | This paper contributes a novel method for RGB-D indoor scene classification. Recent approaches to this problem focus on developing increasingly complex pipelines that learn correlated features across the RGB and depth modalities. In contrast, this paper presents a simple method that first extracts features for the RGB and depth modalities using Places365-CNN and fine-tuned Places365-CNN on depth data, respectively and then clusters these features to generate a set of centroids representing each scene category from the training data. For classification a scene image is converted to CNN features and the distance of these features to the n closest learned centroids is used to predict the image’s category. We evaluate our method on two standard RGB-D indoor scene classification benchmarks: SUNRGB-D and NYU Depth V2 and demonstrate that our proposed classification approach achieves superior performance over the state-of-the-art methods on both datasets. |
Tasks | Scene Classification |
Published | 2019-11-01 |
URL | https://arxiv.org/abs/1911.00155v2 |
https://arxiv.org/pdf/1911.00155v2.pdf | |
PWC | https://paperswithcode.com/paper/centroid-based-scene-classification-cbsc |
Repo | |
Framework | |
Towards Egocentric Person Re-identification and Social Pattern Analysis
Title | Towards Egocentric Person Re-identification and Social Pattern Analysis |
Authors | Estefania Talavera, Alexandre Cola, Nicolai Petkov, Petia Radeva |
Abstract | Wearable cameras capture a first-person view of the daily activities of the camera wearer, offering a visual diary of the user behaviour. Detection of the appearance of people the camera user interacts with for social interactions analysis is of high interest. Generally speaking, social events, lifestyle and health are highly correlated, but there is a lack of tools to monitor and analyse them. We consider that egocentric vision provides a tool to obtain information and understand users social interactions. We propose a model that enables us to evaluate and visualize social traits obtained by analysing social interactions appearance within egocentric photostreams. Given sets of egocentric images, we detect the appearance of faces within the days of the camera wearer, and rely on clustering algorithms to group their feature descriptors in order to re-identify persons. Recurrence of detected faces within photostreams allows us to shape an idea of the social pattern of behaviour of the user. We validated our model over several weeks recorded by different camera wearers. Our findings indicate that social profiles are potentially useful for social behaviour interpretation. |
Tasks | Person Re-Identification |
Published | 2019-05-10 |
URL | https://arxiv.org/abs/1905.04073v1 |
https://arxiv.org/pdf/1905.04073v1.pdf | |
PWC | https://paperswithcode.com/paper/towards-egocentric-person-re-identification |
Repo | |
Framework | |
Machine Learning based detection of multiple Wi-Fi BSSs for LTE-U CSAT
Title | Machine Learning based detection of multiple Wi-Fi BSSs for LTE-U CSAT |
Authors | Vanlin Sathya, Adam Dziedzic, Monisha Ghosh, Sanjay Krishnan |
Abstract | According to the LTE-U Forum specification, a LTE-U base-station (BS) reduces its duty cycle from 50% to 33% when it senses an increase in the number of co-channel Wi-Fi basic service sets (BSSs) from one to two. The detection of the number of Wi-Fi BSSs that are operating on the channel in real-time, without decoding the Wi-Fi packets, still remains a challenge. In this paper, we present a novel machine learning (ML) approach that solves the problem by using energy values observed during LTE-U OFF duration. Observing the energy values (at LTE-U BS OFF time) is a much simpler operation than decoding the entire Wi-Fi packets. In this work, we implement and validate the proposed ML based approach in real-time experiments, and demonstrate that there are two distinct patterns between one and two Wi-Fi APs. This approach delivers an accuracy close to 100% compared to auto-correlation (AC) and energy detection (ED) approaches. |
Tasks | |
Published | 2019-11-21 |
URL | https://arxiv.org/abs/1911.09292v1 |
https://arxiv.org/pdf/1911.09292v1.pdf | |
PWC | https://paperswithcode.com/paper/machine-learning-based-detection-of-multiple |
Repo | |
Framework | |
Improved Hard Example Mining by Discovering Attribute-based Hard Person Identity
Title | Improved Hard Example Mining by Discovering Attribute-based Hard Person Identity |
Authors | Xiao Wang, Ziliang Chen, Rui Yang, Bin Luo, Jin Tang |
Abstract | In this paper, we propose Hard Person Identity Mining (HPIM) that attempts to refine the hard example mining to improve the exploration efficacy in person re-identification. It is motivated by following observation: the more attributes some people share, the more difficult to separate their identities. Based on this observation, we develop HPIM via a transferred attribute describer, a deep multi-attribute classifier trained from the source noisy person attribute datasets. We encode each image into the attribute probabilistic description in the target person re-ID dataset. Afterwards in the attribute code space, we consider each person as a distribution to generate his view-specific attribute codes in different practical scenarios. Hence we estimate the person-specific statistical moments from zeroth to higher order, which are further used to calculate the central moment discrepancies between persons. Such discrepancy is a ground to choose hard identity to organize proper mini-batches, without concerning the person representation changing in metric learning. It presents as a complementary tool of hard example mining, which helps to explore the global instead of the local hard example constraint in the mini-batch built by randomly sampled identities. Extensive experiments on two person re-identification benchmarks validated the effectiveness of our proposed algorithm. |
Tasks | Metric Learning, Person Re-Identification |
Published | 2019-05-06 |
URL | https://arxiv.org/abs/1905.02102v3 |
https://arxiv.org/pdf/1905.02102v3.pdf | |
PWC | https://paperswithcode.com/paper/improved-hard-example-mining-by-discovering |
Repo | |
Framework | |
Conformal calibrators
Title | Conformal calibrators |
Authors | Vladimir Vovk, Ivan Petej, Paolo Toccaceli, Alex Gammerman |
Abstract | Most existing examples of full conformal predictive systems, split-conformal predictive systems, and cross-conformal predictive systems impose severe restrictions on the adaptation of predictive distributions to the test object at hand. In this paper we develop split-conformal and cross-conformal predictive systems that are fully adaptive. Our method consists in calibrating existing predictive systems; the input predictive system is not supposed to satisfy any properties of validity, whereas the output predictive system is guaranteed to be calibrated in probability. It is interesting that the method may also work without the IID assumption, standard in conformal prediction. |
Tasks | |
Published | 2019-02-18 |
URL | http://arxiv.org/abs/1902.06579v1 |
http://arxiv.org/pdf/1902.06579v1.pdf | |
PWC | https://paperswithcode.com/paper/conformal-calibrators |
Repo | |
Framework | |
A Fast and Precise Method for Large-Scale Land-Use Mapping Based on Deep Learning
Title | A Fast and Precise Method for Large-Scale Land-Use Mapping Based on Deep Learning |
Authors | Xuan Yang, Zhengchao Chen, Baipeng Li, Dailiang Peng, Pan Chen, Bing Zhang |
Abstract | The land-use map is an important data that can reflect the use and transformation of human land, and can provide valuable reference for land-use planning. For the traditional image classification method, producing a high spatial resolution (HSR), land-use map in large-scale is a big project that requires a lot of human labor, time, and financial expenditure. The rise of the deep learning technique provides a new solution to the problems above. This paper proposes a fast and precise method that can achieve large-scale land-use classification based on deep convolutional neural network (DCNN). In this paper, we optimize the data tiling method and the structure of DCNN for the multi-channel data and the splicing edge effect, which are unique to remote sensing deep learning, and improve the accuracy of land-use classification. We apply our improved methods in the Guangdong Province of China using GF-1 images, and achieve the land-use classification accuracy of 81.52%. It takes only 13 hours to complete the work, which will take several months for human labor. |
Tasks | Image Classification |
Published | 2019-08-09 |
URL | https://arxiv.org/abs/1908.03438v1 |
https://arxiv.org/pdf/1908.03438v1.pdf | |
PWC | https://paperswithcode.com/paper/a-fast-and-precise-method-for-large-scale |
Repo | |
Framework | |
Value of Temporal Dynamics Information in Driving Scene Segmentation
Title | Value of Temporal Dynamics Information in Driving Scene Segmentation |
Authors | Li Ding, Jack Terwilliger, Rini Sherony, Bryan Reimer, Lex Fridman |
Abstract | Semantic scene segmentation has primarily been addressed by forming representations of single images both with supervised and unsupervised methods. The problem of semantic segmentation in dynamic scenes has begun to recently receive attention with video object segmentation approaches. What is not known is how much extra information the temporal dynamics of the visual scene carries that is complimentary to the information available in the individual frames of the video. There is evidence that the human visual system can effectively perceive the scene from temporal dynamics information of the scene’s changing visual characteristics without relying on the visual characteristics of individual snapshots themselves. Our work takes steps to explore whether machine perception can exhibit similar properties by combining appearance-based representations and temporal dynamics representations in a joint-learning problem that reveals the contribution of each toward successful dynamic scene segmentation. Additionally, we provide the MIT Driving Scene Segmentation dataset, which is a large-scale full driving scene segmentation dataset, densely annotated for every pixel and every one of 5,000 video frames. This dataset is intended to help further the exploration of the value of temporal dynamics information for semantic segmentation in video. |
Tasks | Scene Segmentation, Semantic Segmentation, Video Object Segmentation, Video Semantic Segmentation |
Published | 2019-03-21 |
URL | http://arxiv.org/abs/1904.00758v1 |
http://arxiv.org/pdf/1904.00758v1.pdf | |
PWC | https://paperswithcode.com/paper/value-of-temporal-dynamics-information-in |
Repo | |
Framework | |