International audienceNon-negative Matrix Factorization (NMF) and its variants have been successfully used for clustering text documents. However, NMF approaches like other models do not explicitly account for the contextual dependencies between words. To remedy this limitation, we draw inspiration from neural word embedding and posit that words that frequently co-occur within the same context (e.g., sentence or document) are likely related to each other in some semantic aspect. We then propose to jointly factorize the document-word and word-word co-occurrence matrices. The decomposition of the latter matrix encourages frequently co-occurring words to have similar latent representations and thereby reflecting the relationships among them. E...
Recent work incorporates pre-trained word embeddings such as BERT embeddings into Neural Topic Model...
Non-negative Matrix Factorization (NMF, [5]) and Probabilistic Latent Semantic Analysis (PLSA, [4]) ...
Clustering is a fundamental problem in unsupervised and semi-supervised machine learning. Besides cl...
Text data co-clustering is the process of partitioning the documents and words simultaneously. This ...
This paper proposes a deep hierarchical Non-negative Matrix Factorization (NMF) method with Skip-Gra...
Abstract. In this paper, we use non-negative matrix factorization (NMF) to refine the document clust...
In traditional text clustering, documents appear terms frequency without considering the semantic in...
Document clustering without any prior knowledge or background information is a challenging problem. ...
Abstract—Nonnegative Matrix Factorization (NMF) is one of the most promising techniques to reduce th...
This dissertation shows that nonnegative matrix factorization (NMF) can be extended to a general and...
International audienceCo-clustering of document-term matrices has proved to be more effective than o...
Most document clustering algorithms operate in a high dimensional bag-of-words space. The inherent p...
Non-negative Matrix Factorization (NMF) and Probabilistic Latent Semantic Indexing (PLSI) have been ...
Non-negative dyadic data, that is data representing observations which relate two finite sets of obj...
PACLIC 21 / Seoul National University, Seoul, Korea / November 1-3, 2007conference pape
Recent work incorporates pre-trained word embeddings such as BERT embeddings into Neural Topic Model...
Non-negative Matrix Factorization (NMF, [5]) and Probabilistic Latent Semantic Analysis (PLSA, [4]) ...
Clustering is a fundamental problem in unsupervised and semi-supervised machine learning. Besides cl...
Text data co-clustering is the process of partitioning the documents and words simultaneously. This ...
This paper proposes a deep hierarchical Non-negative Matrix Factorization (NMF) method with Skip-Gra...
Abstract. In this paper, we use non-negative matrix factorization (NMF) to refine the document clust...
In traditional text clustering, documents appear terms frequency without considering the semantic in...
Document clustering without any prior knowledge or background information is a challenging problem. ...
Abstract—Nonnegative Matrix Factorization (NMF) is one of the most promising techniques to reduce th...
This dissertation shows that nonnegative matrix factorization (NMF) can be extended to a general and...
International audienceCo-clustering of document-term matrices has proved to be more effective than o...
Most document clustering algorithms operate in a high dimensional bag-of-words space. The inherent p...
Non-negative Matrix Factorization (NMF) and Probabilistic Latent Semantic Indexing (PLSI) have been ...
Non-negative dyadic data, that is data representing observations which relate two finite sets of obj...
PACLIC 21 / Seoul National University, Seoul, Korea / November 1-3, 2007conference pape
Recent work incorporates pre-trained word embeddings such as BERT embeddings into Neural Topic Model...
Non-negative Matrix Factorization (NMF, [5]) and Probabilistic Latent Semantic Analysis (PLSA, [4]) ...
Clustering is a fundamental problem in unsupervised and semi-supervised machine learning. Besides cl...