Ming Yang’s research while affiliated with Zhejiang University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (12)


A Survey of Multi-View Representation Learning
  • Article

September 2018

·

463 Reads

·

581 Citations

IEEE Transactions on Knowledge and Data Engineering

·

Ming Yang

·

Zhongfei Mark Zhang

Recently, multi-view representation learning has become a rapidly growing direction in machine learning and data mining areas. This paper introduces two categories for multi-view representation learning: multi-view representation alignment and multi-view representation fusion. Consequently, we first review the representative methods and theories of multi-view representation learning based on the perspective of alignment, such as correlation-based alignment. Representative examples are canonical correlation analysis (CCA) and its several extensions. Then from the perspective of representation fusion we investigate the advancement of multi-view representation learning that ranges from generative methods including multi-modal topic learning, multi-view sparse coding, and multi-view latent space Markov networks, to neural network-based methods including multi-modal autoencoders, multi-view convolutional neural networks and multi-modal recurrent neural networks. Further, we also investigate several important applications of multi-view representation learning. Overall, this survey aims to provide an insightful overview of theoretical foundation and state-of-the-art developments in the field of multi-view representation learning and to help researchers find the most appropriate tools for particular applications.


Learning with Feature Network and Label Network Simultaneously

February 2017

·

6 Reads

·

3 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

For many supervised learning problems, limited training samples and incomplete labels are two difficult challenges, which usually lead to degenerated performance on label prediction. To improve the generalization performance, in this paper, we propose Doubly Regularized Multi-Label learning (DRML) by exploiting feature network and label network regularization simultaneously. In more details, the proposed algorithm first constructs a feature network and a label network with marginalized linear denoising autoencoder in data feature set and label set, respectively, and then learns a robust predictor with the feature network and the label network regularization simultaneously. While DRML is a general method for multi-label learning, in the evaluations we focus on the specific application of multi-label text tagging. Extensive evaluations on three benchmark data sets demonstrate that DRML outstands with a superior performance in comparison with some existing multi-label learning methods.


Fig. 1. An illustrative example application of CCA in cross-modal retrieval (adapted from [25]). Left: Embedding of the text and image from their source spaces to a CCA space, Semantic Space and a Semantic space learned using CCA representation. Right: examples of cross-modal retrieval where both text and images are mapped to a common space. At the top is shown an example of retrieving text in response to an image query with a common semantic space. At the bottom is shown an example of retrieving images in response to a text query with a common subspace using CCA.
Fig. 2. The framework of deep CCA (adapted from [8]), in which the output layers of two deep networks are maximally correlated.  
Fig. 3. The graphical model representation of the Corr-LDA model (adapted from [11]).  
Fig. 4. The graphical model of deep multi-modal RBM (adapted from [17]), which models the joint distribution over image and text inputs.
Fig. 5. The bimodal deep autoencoder (adapted from [18]).

+4

Multi-View Representation Learning: A Survey from Shallow Methods to Deep Methods
  • Article
  • Full-text available

October 2016

·

3,515 Reads

·

38 Citations

Recently, multi-view representation learning has become a rapidly growing direction in machine learning and data mining areas. This paper first reviews the root methods and theories on multi-view representation learning, especially on canonical correlation analysis (CCA) and its several extensions. And then we investigate the advancement of multi-view representation learning that ranges from shallow methods including multi-modal topic learning, multi-view sparse coding, and multi-view latent space Markov networks, to deep methods including multi-modal restricted Boltzmann machines, multi-modal autoencoders, and multi-modal recurrent neural networks. Further, we also provide an important perspective from manifold alignment for multi-view representation learning. Overall, this survey aims to provide an insightful overview of theoretical basis and current developments in the field of multi-view representation learning and to help researchers find the most appropriate tools for particular applications.

Download

Learning with Marginalized Corrupted Features and Labels Together

February 2016

·

4 Reads

·

8 Citations

Proceedings of the AAAI Conference on Artificial Intelligence

Tagging has become increasingly important in many real-world applications noticeably including web applications, such as web blogs and resource sharing systems. Despite this importance, tagging methods often face difficult challenges such as limited training samples and incomplete labels, which usually lead to degenerated performance on tag prediction. To improve the generalization performance, in this paper, we propose Regularized Marginalized Cross-View learning (RMCV) by jointly modeling on attribute noise and label noise. In more details, the proposed model constructs infinite training examples with attribute noises from known exponential-family distributions and exploits label noise via marginalized denoising autoencoder. Therefore, the model benefits from its robustness and alleviates the problem of tag sparsity. While RMCV is a general method for learning tagging, in the evaluations we focus on the specific application of multi-label text tagging. Extensive evaluations on three benchmark data sets demonstrate that RMCV outstands with a superior performance in comparison with state-of-the-art methods.


Scientific articles recommendation with topic regression and relational matrix factorization

November 2014

·

34 Reads

·

2 Citations

Journal of Zhejiang University SCIENCE C

In this paper we study the problem of recommending scientific articles to users in an online community with a new perspective of considering topic regression modeling and articles relational structure analysis simultaneously. First, we present a novel topic regression model, the topic regression matrix factorization (tr-MF), to solve the problem. The main idea of tr-MF lies in extending the matrix factorization with a probabilistic topic modeling. In particular, tr-MF introduces a regression model to regularize user factors through probabilistic topic modeling under the basic hypothesis that users share similar preferences if they rate similar sets of items. Consequently, tr-MF provides interpretable latent factors for users and items, and makes accurate predictions for community users. To incorporate the relational structure into the framework of tr-MF, we introduce relational matrix factorization. Through combining tr-MF with the relational matrix factorization, we propose the topic regression collective matrix factorization (tr-CMF) model. In addition, we also present the collaborative topic regression model with relational matrix factorization (CTR-RMF) model, which combines the existing collaborative topic regression (CTR) model and relational matrix factorization (RMF). From this point of view, CTR-RMF can be considered as an appropriate baseline for tr-CMF. Further, we demonstrate the efficacy of the proposed models on a large subset of the data from CiteULike, a bibliography sharing service dataset. The proposed models outperform the state-of-the-art matrix factorization models with a significant margin. Specifically, the proposed models are effective in making predictions for users with only few ratings or even no ratings, and support tasks that are specific to a certain field, neither of which has been addressed in the existing literature.


Bayesian Multi-Task Relationship Learning with Link Structure

December 2013

·

34 Reads

·

2 Citations

In this paper, we study the multi-task learning problem with a new perspective of considering the link structure of data and task relationship modeling simultaneously. In particular, we first introduce the Matrix Generalized Inverse Gaussian (MGIG) distribution and define a Matrix Gaussian Matrix Generalized Inverse Gaussian (MG-MGIG) prior. Based on this prior, we propose a novel multi-task learning algorithm, the Bayesian Multi-task Relationship Learning (BMTRL) algorithm. To incorporate the link structure into the framework of BMTRL, we propose link constraints between samples. Through combining the BMTRL algorithm with the link constraints, we propose the Bayesian Multi-task Relationship Learning with Link Constraints (BMTRL-LC) algorithm. To make the computation tractable, we simultaneously use a convex optimization method and sampling techniques. In particular, we adopt two stochastic EM algorithms for BMTRL and BMTRL-LC, respectively. The experimental results on Cora dataset demonstrate the promise of the proposed algorithms.


Coordinate Ranking Regularized Non-negative Matrix Factorization

November 2013

·

38 Reads

Non-negative Matrix Factorization (NMF) has become increasingly popular in many applications that require data mining techniques such as information retrieval, computer vision, and pattern recognition. NMF aims at approximating the original data matrix in a high dimensional space with the product of two non-negative matrices in a lower dimensional space. In many applications with high dimensional data such as text, data often have a global geometric structure, which typically may not be directly derived from the local information. But the existing literature of NMF completely ignores this problem. This paper proposes a novel matrix factorization algorithm called Coordinate Ranking regularized NMF (CR-NMF) in order to address this problem. The idea of the proposed algorithm is to combine NMF and manifold ranking to encode both local and global geometric structures of the data. Experimental results on two real-world datasets demonstrate the superiority of this algorithm.


Scientific articles recommendation

October 2013

·

112 Reads

·

27 Citations

We study the problem of recommending scientific articles to users in an online community and present a novel matrix factorization model, the topic regression Matrix Factorization (tr-MF), to solve the problem. The main idea of tr-MF lies in extending the matrix factorization with a probabilistic topic modeling. Instead of regularizing item factors through the probabilistic topic modeling as in the framework of the CTR model, tr-MF introduces a regression model to regularize user factors through the probabilistic topic modeling under the basic hypothesis that users share the similar preferences if they rate similar sets of items. Consequently, tr-MF provides interpretable latent factors for users and items, and makes accurate predictions for community users. Specifically, it is effective in making predictions for users with only few ratings or even no ratings, and supports tasks that are specific to a certain field, neither of which is addressed in the existing literature. Further, we demonstrate the efficacy of tr-MF on a large subset of the data from CiteULike, a bibliography sharing service dataset. The proposed model outperforms the state-of-the-art matrix factorization models with a significant margin.


Learning with limited and noisy tagging

October 2013

·

12 Reads

·

7 Citations

With the rapid development of social networks, tagging has become an important means responsible for such rapid development. A robust tagging method must have the capability to meet the two challenging requirements: limited labeled training samples and noisy labeled training samples. In this paper, we investigate this challenging problem of learning with limited and noisy tagging and propose a discriminative model, called SpSVM-MC, that exploits both labeled and unlabeled data through a semi-parametric regularization and takes advantage of the multi-label constraints into the optimization. While SpSVM-MC is a general method for learning with limited and noisy tagging, in the evaluations we focus on the specific application of noisy image tagging with limited labeled training samples on a benchmark dataset. Theoretical analysis and extensive evaluations in comparison with state-of-the-art literature demonstrate that SpSVM-MC outstands with a superior performance.


Multi-view learning from imperfect tagging

October 2012

·

9 Reads

·

6 Citations

In many real-world applications, tagging is imperfect: incomplete, inconsistent, and error-prone. Solutions to this problem will generate societal and technical impacts. In this paper, we investigate this arguably new problem: learning from imperfect tagging. We propose a general and effective learning scheme called the Multi-view Imperfect Tagging Learning (MITL) to this problem. The main idea of MITL lies in extracting the information of the imperfectly tagged training dataset from multiple views to differentiate the data points in the role of classification. Further, a novel discriminative classification method is proposed under the framework of MITL, which explicitly makes use of the given multiple labels simultaneously as an additional feature to deliver a more effective classification performance than the existing literature where one label is considered at a time as the classification target while the rest of the given labels are completely ignored at the same time. The proposed methods can not only complete the incomplete tagging but also denoise the noisy tagging through an inductive learning. We apply the general solution to the problem with a more specific context - imperfect image annotation, and evaluate the proposed methods on a standard dataset from the related literature. Experiments show that they are superior to the peer methods on solving the problem of learning from imperfect tagging in cross-media.


Citations (8)


... Different from methods which improves the quality of data [39], noising scheme can be regarded as a kind of regularization [35] and has proven to be effective for training robust classifiers [8,23] or extracting better features [7,22]. Maaten et al. firstly propose the implicitly noising scheme for general learning problems [27]. ...

Reference:

Disturbance Grassmann Kernels for Subspace-Based Learning
Learning with Marginalized Corrupted Features and Labels Together
  • Citing Article
  • February 2016

Proceedings of the AAAI Conference on Artificial Intelligence

... In today's machine learning era, representation learning is essential for deriving meaningful features automatically from raw data, which is vital for tasks such as pattern recognition and classification [1][2][3]. As the complexity of tasks in ML grows, from image and speech recognition to natural language processing (NLP), the demand for more refined and adaptive representation learning techniques becomes critical. ...

A Survey of Multi-View Representation Learning
  • Citing Article
  • September 2018

IEEE Transactions on Knowledge and Data Engineering

... A composite image provides comprehensive information to humans and machines, supplying more detailed information than the single-model/view image [10]. In a similar manner to multimodal learning, multi-view learning which can be considered as an information fusion technique has been noticed in machine learning and data mining due to an increased rate of multi-view data provided by applications [22] [23]. Data from different views of a modal (e.g., image) have complementary information than single-view data. ...

Multi-View Representation Learning: A Survey from Shallow Methods to Deep Methods

... Dealing with noise in multilabel classification is a very important topic, as numerous applications use "soft labels": labels which are not assigned by a domain expert but are derived from automatic taggers [101] or from non-experts via crowdsourcing [92], which are known to introduce noise. However, the influence of noise in multilabel classification has so far been little studied. ...

Mining noisy tagging from multi-label space
  • Citing Article
  • October 2012

... Self-training [11] and co-training [12], [13] extract the most confidently classified examples from the unlabeled data, and add them into the labeled training set iteratively. Li et al. [14] explore the geometric structure of the marginal distribution of the whole data including the labeled and the unlabeled data through a specific family of parametric functions. Socher et al. [15] propose a recursive autoencoder trained on both labeled and unlabeled texts to predict sentiment distribution. ...

Learning with limited and noisy tagging
  • Citing Article
  • October 2013

... Many information processing tasks depend highly on the existence and the quality of topical annotations of content. For instance, academic search Xiong et al. (2017) and recommendation Li et al. (2013) engines, product review exploitation models Tsur and Rappoport (2009), or topic detection trend tracking Panem et al. (2014) are obvious examples. One indicator of the reliance on keyphrases in academic search and recommendation is that publishers often ask authors to label their publications with keyphrases manually. ...

Scientific articles recommendation
  • Citing Article
  • October 2013

... Matrix Generalized Inverse Gaussian (MGIG) distributions [3,10] are a flexible family of distributions over the space of symmetric positive definite matrices and has been recently applied as the prior for covariance matrix [21,32,33]. MGIG is a flexible prior since it contains Wishart, and Inverse Wishart distributions as special cases. ...

Bayesian Multi-Task Relationship Learning with Link Structure
  • Citing Conference Paper
  • December 2013

... Due to the capability of exploring the data distribution, probability graphical models (PGMs) have been popular for MLML problems since we can complement the missing labels in a generative manner. SSC-HDP [104] extends Corr-LDA [105] to a correspondence hierarchical Dirichlet process (Corr-HDP) that enables the dimension of latent factors can be chosen dynamically. Based on Corr-HDP, SSC-HDP iteratively updates the likelihood P (y j |x) for an instance x whose j-th label is missing, while the likelihood of remaining labels is fixed to 1. MPU [106] studies the largescale MLML tasks. ...

Mining partially annotated images
  • Citing Conference Paper
  • August 2011