Paper Group ANR 794
Joint Diacritization, Lemmatization, Normalization, and Fine-Grained Morphological Tagging. Trace-back Along Capsules and Its Application on Semantic Segmentation. Multi-Product Dynamic Pricing in High-Dimensions with Heterogenous Price Sensitivity. Predicting human decisions with behavioral theories and machine learning. Deep Reinforcement Learnin …
Joint Diacritization, Lemmatization, Normalization, and Fine-Grained Morphological Tagging
Title | Joint Diacritization, Lemmatization, Normalization, and Fine-Grained Morphological Tagging |
Authors | Nasser Zalmout, Nizar Habash |
Abstract | Semitic languages can be highly ambiguous, having several interpretations of the same surface forms, and morphologically rich, having many morphemes that realize several morphological features. This is further exacerbated for dialectal content, which is more prone to noise and lacks a standard orthography. The morphological features can be lexicalized, like lemmas and diacritized forms, or non-lexicalized, like gender, number, and part-of-speech tags, among others. Joint modeling of the lexicalized and non-lexicalized features can identify more intricate morphological patterns, which provide better context modeling, and further disambiguate ambiguous lexical choices. However, the different modeling granularity can make joint modeling more difficult. Our approach models the different features jointly, whether lexicalized (on the character-level), where we also model surface form normalization, or non-lexicalized (on the word-level). We use Arabic as a test case, and achieve state-of-the-art results for Modern Standard Arabic, with 20% relative error reduction, and Egyptian Arabic (a dialectal variant of Arabic), with 11% reduction. |
Tasks | Lemmatization, Morphological Tagging |
Published | 2019-10-05 |
URL | https://arxiv.org/abs/1910.02267v1 |
https://arxiv.org/pdf/1910.02267v1.pdf | |
PWC | https://paperswithcode.com/paper/joint-diacritization-lemmatization |
Repo | |
Framework | |
Trace-back Along Capsules and Its Application on Semantic Segmentation
Title | Trace-back Along Capsules and Its Application on Semantic Segmentation |
Authors | Tao Sun, Zhewei Wang, C. D. Smith, Jundong Liu |
Abstract | In this paper, we propose a capsule-based neural network model to solve the semantic segmentation problem. By taking advantage of the extractable part-whole dependencies available in capsule layers, we derive the probabilities of the class labels for individual capsules through a recursive, layer-by-layer procedure. We model this procedure as a traceback pipeline and take it as a central piece to build an end-to-end segmentation network. Under the proposed framework, image-level class labels and object boundaries are jointly sought in an explicit manner, which poses a significant advantage over the state-of-the-art fully convolutional network (FCN) solutions. Experiments conducted on modified MNIST and neuroimages demonstrate that our model considerably enhance the segmentation performance compared to the leading FCN variant. |
Tasks | Semantic Segmentation |
Published | 2019-01-09 |
URL | http://arxiv.org/abs/1901.02920v1 |
http://arxiv.org/pdf/1901.02920v1.pdf | |
PWC | https://paperswithcode.com/paper/trace-back-along-capsules-and-its-application |
Repo | |
Framework | |
Multi-Product Dynamic Pricing in High-Dimensions with Heterogenous Price Sensitivity
Title | Multi-Product Dynamic Pricing in High-Dimensions with Heterogenous Price Sensitivity |
Authors | Adel Javanmard, Hamid Nazerzadeh, Simeng Shao |
Abstract | We consider the problem of multi-product dynamic pricing, in a contextual setting, for a seller of differentiated products. In this environment, the customers arrive over time and products are described by high-dimensional feature vectors. Each customer chooses a product according to the widely used Multinomial Logit (MNL) choice model and her utility depends on the product features as well as the prices offered. Our model allows for heterogenous price sensitivities for products. The seller a-priori does not know the parameters of the choice model but can learn them through interactions with the customers. The seller’s goal is to design a pricing policy that maximizes her cumulative revenue. This model is motivated by online marketplaces such as Airbnb platform and online advertising. We measure the performance of a pricing policy in terms of regret, which is the expected revenue loss with respect to a clairvoyant policy that knows the parameters of the choice model in advance and always sets the revenue-maximizing prices. We propose a pricing policy, named M3P, that achieves a $T$-period regret of $O(\log(Td) ( \sqrt{T}+ d\log(T)))$ under heterogenous price sensitivity for products with features of dimension $d$. We also prove that no policy can achieve worst-case $T$-regret better than $\Omega(\sqrt{T})$. |
Tasks | |
Published | 2019-01-04 |
URL | https://arxiv.org/abs/1901.01030v2 |
https://arxiv.org/pdf/1901.01030v2.pdf | |
PWC | https://paperswithcode.com/paper/multi-product-dynamic-pricing-in-high |
Repo | |
Framework | |
Predicting human decisions with behavioral theories and machine learning
Title | Predicting human decisions with behavioral theories and machine learning |
Authors | Ori Plonsky, Reut Apel, Eyal Ert, Moshe Tennenholtz, David Bourgin, Joshua C. Peterson, Daniel Reichman, Thomas L. Griffiths, Stuart J. Russell, Evan C. Carter, James F. Cavanagh, Ido Erev |
Abstract | Behavioral decision theories aim to explain human behavior. Can they help predict it? An open tournament for prediction of human choices in fundamental economic decision tasks is presented. The results suggest that integration of certain behavioral theories as features in machine learning systems provides the best predictions. Surprisingly, the most useful theories for prediction build on basic properties of human and animal learning and are very different from mainstream decision theories that focus on deviations from rational choice. Moreover, we find that theoretical features should be based not only on qualitative behavioral insights (e.g. loss aversion), but also on quantitative behavioral foresights generated by functional descriptive models (e.g. Prospect Theory). Our analysis prescribes a recipe for derivation of explainable, useful predictions of human decisions. |
Tasks | |
Published | 2019-04-15 |
URL | http://arxiv.org/abs/1904.06866v1 |
http://arxiv.org/pdf/1904.06866v1.pdf | |
PWC | https://paperswithcode.com/paper/predicting-human-decisions-with-behavioral |
Repo | |
Framework | |
Deep Reinforcement Learning-Based Channel Allocation for Wireless LANs with Graph Convolutional Networks
Title | Deep Reinforcement Learning-Based Channel Allocation for Wireless LANs with Graph Convolutional Networks |
Authors | Kota Nakashima, Shotaro Kamiya, Kazuki Ohtsu, Koji Yamamoto, Takayuki Nishio, Masahiro Morikura |
Abstract | Last year, IEEE 802.11 Extremely High Throughput Study Group (EHT Study Group) was established to initiate discussions on new IEEE 802.11 features. Coordinated control methods of the access points (APs) in the wireless local area networks (WLANs) are discussed in EHT Study Group. The present study proposes a deep reinforcement learning-based channel allocation scheme using graph convolutional networks (GCNs). As a deep reinforcement learning method, we use a well-known method double deep Q-network. In densely deployed WLANs, the number of the available topologies of APs is extremely high, and thus we extract the features of the topological structures based on GCNs. We apply GCNs to a contention graph where APs within their carrier sensing ranges are connected to extract the features of carrier sensing relationships. Additionally, to improve the learning speed especially in an early stage of learning, we employ a game theory-based method to collect the training data independently of the neural network model. The simulation results indicate that the proposed method can appropriately control the channels when compared to extant methods. |
Tasks | |
Published | 2019-05-17 |
URL | https://arxiv.org/abs/1905.07144v1 |
https://arxiv.org/pdf/1905.07144v1.pdf | |
PWC | https://paperswithcode.com/paper/deep-reinforcement-learning-based-channel |
Repo | |
Framework | |
Attribute-Guided Coupled GAN for Cross-Resolution Face Recognition
Title | Attribute-Guided Coupled GAN for Cross-Resolution Face Recognition |
Authors | Veeru Talreja, Fariborz Taherkhani, Matthew C Valenti, Nasser M Nasrabadi |
Abstract | In this paper, we propose a novel attribute-guided cross-resolution (low-resolution to high-resolution) face recognition framework that leverages a coupled generative adversarial network (GAN) structure with adversarial training to find the hidden relationship between the low-resolution and high-resolution images in a latent common embedding subspace. The coupled GAN framework consists of two sub-networks, one dedicated to the low-resolution domain and the other dedicated to the high-resolution domain. Each sub-network aims to find a projection that maximizes the pair-wise correlation between the two feature domains in a common embedding subspace. In addition to projecting the images into a common subspace, the coupled network also predicts facial attributes to improve the cross-resolution face recognition. Specifically, our proposed coupled framework exploits facial attributes to further maximize the pair-wise correlation by implicitly matching facial attributes of the low and high-resolution images during the training, which leads to a more discriminative embedding subspace resulting in performance enhancement for cross-resolution face recognition. The efficacy of our approach compared with the state-of-the-art is demonstrated using the LFWA, Celeb-A, SCFace and UCCS datasets. |
Tasks | Face Recognition |
Published | 2019-08-05 |
URL | https://arxiv.org/abs/1908.01790v1 |
https://arxiv.org/pdf/1908.01790v1.pdf | |
PWC | https://paperswithcode.com/paper/attribute-guided-coupled-gan-for-cross |
Repo | |
Framework | |
CAMAL: Context-Aware Multi-scale Attention framework for Lightweight Visual Place Recognition
Title | CAMAL: Context-Aware Multi-scale Attention framework for Lightweight Visual Place Recognition |
Authors | Ahmad Khaliq, Shoaib Ehsan, Michael Milford, Klaus McDonald-Maier |
Abstract | In the last few years, Deep Convolutional Neural Networks (D-CNNs) have shown state-of-the-art performances for Visual Place Recognition (VPR). Their prestigious generalization power has played a vital role in identifying persistent image regions under changing conditions and viewpoints. However, against the computation intensive D-CNNs based VPR algorithms, lightweight VPR techniques are preferred for resource-constraints mobile robots. This paper presents a lightweight CNN-based VPR technique that captures multi-layer context-aware attentions robust under changing environment and viewpoints. Evaluation of challenging benchmark datasets reveals better performance at low memory and resources utilization over state-of-the-art contemporary VPR methodologies. |
Tasks | Visual Place Recognition |
Published | 2019-09-18 |
URL | https://arxiv.org/abs/1909.08153v1 |
https://arxiv.org/pdf/1909.08153v1.pdf | |
PWC | https://paperswithcode.com/paper/camal-context-aware-multi-scale-attention |
Repo | |
Framework | |
Attention-Aware Age-Agnostic Visual Place Recognition
Title | Attention-Aware Age-Agnostic Visual Place Recognition |
Authors | Ziqi Wang, Jiahui Li, Seyran Khademi, Jan van Gemert |
Abstract | A cross-domain visual place recognition (VPR) task is proposed in this work, i.e., matching images of the same architectures depicted in different domains. VPR is commonly treated as an image retrieval task, where a query image from an unknown location is matched with relevant instances from geo-tagged gallery database. Different from conventional VPR settings where the query images and gallery images come from the same domain, we propose a more common but challenging setup where the query images are collected under a new unseen condition. The two domains involved in this work are contemporary street view images of Amsterdam from the Mapillary dataset (source domain) and historical images of the same city from Beeldbank dataset (target domain). We tailored an age-invariant feature learning CNN that can focus on domain invariant objects and learn to match images based on a weakly supervised ranking loss. We propose an attention aggregation module that is robust to domain discrepancy between the train and the test data. Further, a multi-kernel maximum mean discrepancy (MK-MMD) domain adaptation loss is adopted to improve the cross-domain ranking performance. Both attention and adaptation modules are unsupervised while the ranking loss uses weak supervision. Visual inspection shows that the attention module focuses on built forms while the dramatically changing environment are less weighed. Our proposed CNN achieves state of the art results (99% accuracy) on the single-domain VPR task and 20% accuracy at its best on the cross-domain VPR task, revealing the difficulty of age-invariant VPR. |
Tasks | Domain Adaptation, Image Retrieval, Visual Place Recognition |
Published | 2019-09-11 |
URL | https://arxiv.org/abs/1909.05163v1 |
https://arxiv.org/pdf/1909.05163v1.pdf | |
PWC | https://paperswithcode.com/paper/attention-aware-age-agnostic-visual-place |
Repo | |
Framework | |
Automatic Fashion Knowledge Extraction from Social Media
Title | Automatic Fashion Knowledge Extraction from Social Media |
Authors | Yunshan Ma, Lizi Liao, Tat-Seng Chua |
Abstract | Fashion knowledge plays a pivotal role in helping people in their dressing. In this paper, we present a novel system to automatically harvest fashion knowledge from social media. It unifies three tasks of occasion, person and clothing discovery from multiple modalities of images, texts and metadata. A contextualized fashion concept learning model is applied to leverage the rich contextual information for improving the fashion concept learning performance. At the same time, to counter the label noise within training data, we employ a weak label modeling method to further boost the performance. We build a website to demonstrate the quality of fashion knowledge extracted by our system. |
Tasks | |
Published | 2019-08-12 |
URL | https://arxiv.org/abs/1908.04045v1 |
https://arxiv.org/pdf/1908.04045v1.pdf | |
PWC | https://paperswithcode.com/paper/automatic-fashion-knowledge-extraction-from |
Repo | |
Framework | |
Rapid Light Field Depth Estimation with Semi-Global Matching
Title | Rapid Light Field Depth Estimation with Semi-Global Matching |
Authors | Yuriy Anisimov, Oliver Wasenmüller, Didier Stricker |
Abstract | Running time of the light field depth estimation algorithms is typically high. This assessment is based on the computational complexity of existing methods and the large amounts of data involved. The aim of our work is to develop a simple and fast algorithm for accurate depth computation. In this context, we propose an approach, which involves Semi-Global Matching for the processing of light field images. It forms on comparison of pixels’ correspondences with different metrics in the substantially bounded light field space. We show that our method is suitable for the fast production of a proper result in a variety of light field configurations |
Tasks | Depth Estimation |
Published | 2019-07-31 |
URL | https://arxiv.org/abs/1907.13449v1 |
https://arxiv.org/pdf/1907.13449v1.pdf | |
PWC | https://paperswithcode.com/paper/rapid-light-field-depth-estimation-with-semi |
Repo | |
Framework | |
Learning to Recommend via Meta Parameter Partition
Title | Learning to Recommend via Meta Parameter Partition |
Authors | Liang Zhao, Yang Wang, Daxiang Dong, Hao Tian |
Abstract | In this paper we propose to solve an important problem in recommendation – user cold start, based on meta leaning method. Previous meta learning approaches finetune all parameters for each new user, which is both computing and storage expensive. In contrast, we divide model parameters into fixed and adaptive parts and develop a two-stage meta learning algorithm to learn them separately. The fixed part, capturing user invariant features, is shared by all users and is learned during offline meta learning stage. The adaptive part, capturing user specific features, is learned during online meta learning stage. By decoupling user invariant parameters from user dependent parameters, the proposed approach is more efficient and storage cheaper than previous methods. It also has potential to deal with catastrophic forgetting while continually adapting for streaming coming users. Experiments on production data demonstrates that the proposed method converges faster and to a better performance than baseline methods. Meta-training without online meta model finetuning increases the AUC from 72.24% to 74.72% (2.48% absolute improvement). Online meta training achieves a further gain of 2.46% absolute improvement comparing with offline meta training. |
Tasks | Meta-Learning |
Published | 2019-12-04 |
URL | https://arxiv.org/abs/1912.04108v1 |
https://arxiv.org/pdf/1912.04108v1.pdf | |
PWC | https://paperswithcode.com/paper/learning-to-recommend-via-meta-parameter |
Repo | |
Framework | |
Action Recognition in Untrimmed Videos with Composite Self-Attention Two-Stream Framework
Title | Action Recognition in Untrimmed Videos with Composite Self-Attention Two-Stream Framework |
Authors | Dong Cao, Lisha Xu, HaiBo Chen |
Abstract | With the rapid development of deep learning algorithms, action recognition in video has achieved many important research results. One issue in action recognition, Zero-Shot Action Recognition (ZSAR), has recently attracted considerable attention, which classify new categories without any positive examples. Another difficulty in action recognition is that untrimmed data may seriously affect model performance. We propose a composite two-stream framework with a pre-trained model. Our proposed framework includes a classifier branch and a composite feature branch. The graph network model is adopted in each of the two branches, which effectively improves the feature extraction and reasoning ability of the framework. In the composite feature branch, a 3-channel self-attention models are constructed to weight each frame in the video and give more attention to the key frames. Each self-attention models channel outputs a set of attention weights to focus on a particular aspect of the video, and a set of attention weights corresponds to a one-dimensional vector. The 3-channel self-attention models can evaluate key frames from multiple aspects, and the output sets of attention weight vectors form an attention matrix, which effectively enhances the attention of key frames with strong correlation of action. This model can implement action recognition under zero-shot conditions, and has good recognition performance for untrimmed video data. Experimental results on relevant data sets confirm the validity of our model. |
Tasks | Temporal Action Localization |
Published | 2019-08-04 |
URL | https://arxiv.org/abs/1908.04353v2 |
https://arxiv.org/pdf/1908.04353v2.pdf | |
PWC | https://paperswithcode.com/paper/action-recognition-in-untrimmed-videos-with |
Repo | |
Framework | |
Gaussian Process Learning via Fisher Scoring of Vecchia’s Approximation
Title | Gaussian Process Learning via Fisher Scoring of Vecchia’s Approximation |
Authors | Joseph Guinness |
Abstract | We derive a single pass algorithm for computing the gradient and Fisher information of Vecchia’s Gaussian process loglikelihood approximation, which provides a computationally efficient means for applying the Fisher scoring algorithm for maximizing the loglikelihood. The advantages of the optimization techniques are demonstrated in numerical examples and in an application to Argo ocean temperature data. The new methods are more accurate and much faster than an optimization method that uses only function evaluations, especially when the covariance function has many parameters. This allows practitioners to fit nonstationary models to large spatial and spatial-temporal datasets. |
Tasks | |
Published | 2019-05-20 |
URL | https://arxiv.org/abs/1905.08374v1 |
https://arxiv.org/pdf/1905.08374v1.pdf | |
PWC | https://paperswithcode.com/paper/gaussian-process-learning-via-fisher-scoring |
Repo | |
Framework | |
Adversarially Approximated Autoencoder for Image Generation and Manipulation
Title | Adversarially Approximated Autoencoder for Image Generation and Manipulation |
Authors | Wenju Xu, Shawn Keshmiri, Guanghui Wang |
Abstract | Regularized autoencoders learn the latent codes, a structure with the regularization under the distribution, which enables them the capability to infer the latent codes given observations and generate new samples given the codes. However, they are sometimes ambiguous as they tend to produce reconstructions that are not necessarily faithful reproduction of the inputs. The main reason is to enforce the learned latent code distribution to match a prior distribution while the true distribution remains unknown. To improve the reconstruction quality and learn the latent space a manifold structure, this work present a novel approach using the adversarially approximated autoencoder (AAAE) to investigate the latent codes with adversarial approximation. Instead of regularizing the latent codes by penalizing on the distance between the distributions of the model and the target, AAAE learns the autoencoder flexibly and approximates the latent space with a simpler generator. The ratio is estimated using generative adversarial network (GAN) to enforce the similarity of the distributions. Additionally, the image space is regularized with an additional adversarial regularizer. The proposed approach unifies two deep generative models for both latent space inference and diverse generation. The learning scheme is realized without regularization on the latent codes, which also encourages faithful reconstruction. Extensive validation experiments on four real-world datasets demonstrate the superior performance of AAAE. In comparison to the state-of-the-art approaches, AAAE generates samples with better quality and shares the properties of regularized autoencoder with a nice latent manifold structure. |
Tasks | Image Generation |
Published | 2019-02-14 |
URL | http://arxiv.org/abs/1902.05581v1 |
http://arxiv.org/pdf/1902.05581v1.pdf | |
PWC | https://paperswithcode.com/paper/adversarially-approximated-autoencoder-for |
Repo | |
Framework | |
Generative models as parsimonious descriptions of sensorimotor loops
Title | Generative models as parsimonious descriptions of sensorimotor loops |
Authors | Manuel Baltieri, Christopher L. Buckley |
Abstract | The Bayesian brain hypothesis, predictive processing and variational free energy minimisation are typically used to describe perceptual processes based on accurate generative models of the world. However, generative models need not be veridical representations of the environment. We suggest that they can (and should) be used to describe sensorimotor relationships relevant for behaviour rather than precise accounts of the world. |
Tasks | |
Published | 2019-04-29 |
URL | http://arxiv.org/abs/1904.12937v1 |
http://arxiv.org/pdf/1904.12937v1.pdf | |
PWC | https://paperswithcode.com/paper/generative-models-as-parsimonious |
Repo | |
Framework | |