• 936-A Review on Multi-Label Learning Algorithms.pdf

    Multi-label learning studies the problem where each example is represented by a single instance while associated with a set of labels simultaneously. During the past decade, significant amount of progresses have been made toward this emerging machine learning paradigm. This paper aims to provide a timely review on this area with emphasis on state-of-the-art multi-label learning algorithms. Firstly, fundamentals on multi-label learning including formal definition and evaluation metrics are given. Secondly and primarily, eight representative multi-label learning algorithms are scrutinized under common notations with relevant analyses and discussions. Thirdly, several related learning settings are briefly summarized. As a conclusion, online resources and open research problems on multi-label learning are outlined for reference purposes.

    0
    208
    2.21MB
    2020-03-29
    13
  • 771-A Survey on Concept Drift Adaptation.pdf

    Concept drift primarily refers to an online supervised learning scenario when the relation between the input data and the target variable changes over time. Assuming a general knowledge of supervised learning in this paper we characterize adaptive learning process, categorize existing strategies for handling concept drift, discuss the most representative, distinct and popular techniques and algorithms, discuss evaluation methodology of adaptive algorithms, and present a set of illustrative applications. This introduction to the concept drift adaptation presents the state of the art techniques and a collection of benchmarks for researchers, industry analysts and practitioners. The survey aims at covering the different facets of concept drift in an integrated way to reflect on the existing scattered state-of-the-art

    0
    150
    731KB
    2020-03-29
    24
  • 639-Multi-scale Orderless Pooling of Deep Convolutional Activation Features.pdf

    Deep convolutional neural networks (CNN) have shown their promise as a universal representation for recognition. However, global CNN activations lack geometric invariance, which limits their robustness for classification and matching of highly variable scenes. To improve the invariance of CNN activations without degrading their discriminative power, this paper presents a simple but effective scheme called multiscale orderless pooling (MOP-CNN). This scheme extracts CNN activations for local patches at multiple scale levels, performs orderless VLAD pooling of these activations at each level separately, and concatenates the result. The resulting MOP-CNN representation can be used as a generic feature for either supervised or unsupervised recognition tasks, from image classification to instance-level retrieval; it consistently outperforms global CNN activations without requiring any joint training of prediction layers for a particular target dataset. In absolute terms, it achieves stateof-the-art results on the challenging SUN397 and MIT Indoor Scenes classification datasets, and competitive results on ILSVRC2012/2013 classi- fication and INRIA Holidays retrieval datasets

    0
    110
    1.81MB
    2020-03-29
    9
  • Model-driven deep-learning.pdf

    With the arrival of the big data era, data requirements are gradually no longer an obstacle (at least for many areas), but the determination of network topology is still a bottleneck. This is mainly due to the lack of theoretical understandings of the relationship between the network topology and performance. In the current state, the selection of network topology is still an engineering practice instead of scientific research, leading to the fact that most of the existing deep-learning approaches lack theoretical foundations. The difficulties in network design and its interpretation, and a lack of understanding in its generalization ability are the common limitations of the deep-learning approach. These limitations may prevent its widespread use in the trends of ‘standardization, commercialization’ of machine learning and artificial intelligence technology

    0
    306
    226KB
    2020-03-29
    16
  • Learning representations on graphs.pdf

    Networks are everywhere. Popular examples include social networks, the hyperlinked World Wide Web, transportation networks, electricity power networks and biological gene networks. Networks are typically represented as a graph whose vertices represent entities and edges represent links or relationships between these entities. As the pervasiveness and scope of network data increase, there has been significant interest in developing statistical models to learn from networks for prediction or reasoning tasks

    0
    79
    109KB
    2020-03-29
    10
  • Generalization in Machine Learning via Analytical Learning Theory.pdf

    This paper introduces a novel measure-theoreticlearning theory to analyze generalization behaviors of practical interest. The proposed learningtheory has the following abilities: 1) to utilizethe qualities of each learned representation onthe path from raw inputs to outputs in representation learning, 2) to guarantee good generalization errors possibly with arbitrarily rich hypothesis spaces (e.g., arbitrarily large capacity andRademacher complexity) and non-stable/nonrobust learning algorithms, and 3) to clearly distinguish each individual problem instance fromeach other. Our generalization bounds are relative to a representation of the data, and hold true even if the representation is learned. We discuss several consequences of our results on deep learning, one-shot learning and curriculum learning. Unlike statistical learning theory, the proposed learning theory analyzes each probleminstance individually via measure theory, rather than a set of problem instances via statistics. Because of the differences in the assumptions and the objectives, the proposed learning theory is meant to be complementary to previous learning theory and is not designed to compete with it

    0
    149
    308KB
    2020-03-29
    1
  • Generalization in Deep Learning.pdf

    With a direct analysis of neural networks, this paper presents a mathematically tight generalization theory to partially address an open problem regarding the generalization of deep learning. Unlike previous bound-based theory, our main theory is quantitatively as tight as possible for every dataset individually, while producing qualitative insights competitively. Our results give insight into why and how deep learning can generalize well, despite its large capacity, complexity, possible algorithmic instability, nonrobustness, and sharp minima, answering to an open question in the literature. We also discuss limitations of our results and propose additional open problems

    0
    148
    323KB
    2020-03-29
    12
  • An overview of multi-task learning.pdf

    multiple related learning tasks by leveraging useful information among them. In this paper, we give an overview of MTL by first giving a definition of MTL. Then several different settings of MTL are introduced, including multi-task supervised learning, multi-task unsupervised learning, multi-task semi-supervised learning, multi-task active learning, multi-task reinforcement learning, multi-task online learning and multi-task multi-view learning. For each setting, representative MTL models are presented. In order to speed up the learning process, parallel and distributed MTL models are introduced. Many areas, including computer vision, bioinformatics, health informatics, speech, natural language processing, web applications and ubiquitous computing, use MTL to improve the performance of the applications involved and some representative works are reviewed. Finally, recent theoretical analyses for MTL are presented.

    0
    226
    280KB
    2020-03-29
    30
  • A brief introduction to weakly supervised learning.pdf

    Supervised learningtechniques construct predictive models by learning from a large number of training examples, where each training example has a label indicating its ground-truth output. Though current techniques have achieved great success, it is noteworthy that in many tasks it is difficult to get strong supervision information like fully ground-truth labels due to the high cost of the data-labeling process. Thus, it is desirable for machine-learning techniques to work with weak supervision. This article reviews some research progress of weakly supervised learning, focusing on three typical types of weak supervision: incomplete supervision, where only a subset of training data is given with labels; inexact supervision, where the training data are given with only coarse-grained labels; and inaccurate supervision, where the given labels are not always ground-truth

    0
    184
    866KB
    2020-03-29
    21
  • Streaming Graph Neural Networks.pdf

    Graphs are essential representations of many real-world data such as social networks. Recent years have witnessed the increasing efforts made to extend the neural network models to graph-structured data. These methods, which are usually known as the graph neural networks, have been applied to advance many graphs related tasks such as reasoning dynamics of the physical system, graph classification, and node classification. Most of the existing graph neural network models have been designed for static graphs, while many real-world graphs are inherently dynamic. For example, social networks are naturally evolving as new users joining and new relations being created. Current graph neural network models cannot utilize the dynamic information in dynamic graphs. However, the dynamic information has been proven to enhance the performance of many graph analytic tasks such as community detection and link prediction. Hence, it is necessary to design dedicated graph neural networks for dynamic graphs. In this paper, we propose DGNN, a new Dynamic Graph Neural Network model, which can model the dynamic information as the graph evolving. In particular, the proposed framework can keep updating node information by capturing the sequential information of edges (interactions), the time intervals between edges and information propagation coherently. Experimental results on various dynamic graphs demonstrate the effectiveness of the proposed framework.

    0
    390
    887KB
    2020-03-29
    31
  • 分享宗师

    成功上传21个资源即可获取
  • 分享精英

    成功上传11个资源即可获取
关注 私信
上传资源赚积分or赚钱