%0 Unpublished work %T SHCoClust, a Scalable Similarity-based Hierarchical Co-clustering Method and its Application to Textual Collections %+ Entrepôts, Représentation et Ingénierie des Connaissances (ERIC) %A Wang, Xinyu %A Ah-Pine, Julien %A Darmont, Jérôme %Z French national project REQUEST %Z This paper is accepted as a long paper with an oral presentation by the IEEE international conference on fuzzy systems (FUZZ-IEEE 2017), which will take place in July 2017 in Naples, Italy. https://www.fuzzieee2017.org/ %8 2017-02 %D 2017 %K similarity-based agglomerative hierarchical clustering %K bipartite spectral co-clustering %K text clustering %Z Computer Science [cs]/Artificial Intelligence [cs.AI] %Z Computer Science [cs]/Computation and Language [cs.CL]Preprints, Working Papers, ... %X In comparison with flat clustering methods, such as K-means, hierarchical clustering and co-clustering methods are more advantageous, for the reason that hierarchical clustering is capable to reveal the internal connections of clusters, and co-clustering can yield clusters of data instances and features. Interested in organizing co-clusters in hierarchies and in discovering cluster hierarchies inside co-clusters, in this paper, we propose SHCoClust, a scalable similarity-based hierarchical co-clustering method. Except possessing the above-mentioned advantages in unison, SHCoClust is able to employ kernel functions, thanks to its utilization of inner product. Furthermore, having all similarities between 0 and 1, the input of SHCoClust can be sparsified by threshold values, so that less memory and less time are required for storage and for computation. This grants SHCoClust scalability, i.e, the ability to process relatively large datasets with reduced and limited computing resources. Our experiments demonstrate that SHCoClust significantly out-performs the conventional hierarchical clustering methods. In addition, with sparsifying the input similarity matrices obtained by linear kernel and by Gaussian kernel, SHCoClust is capable to guarantee the clustering quality, even when its input being largely sparsified. Consequently, up to 86% time gain and on average 75% memory gain are achieved. %G English %2 https://hal.science/hal-01504986v1/document %2 https://hal.science/hal-01504986v1/file/PID4744355%5Bcamera_read_submissiom_2%5D.pdf %L hal-01504986 %U https://hal.science/hal-01504986 %~ UNIV-LYON1 %~ UNIV-LYON2 %~ ERIC %~ LABEXIMU %~ LYON2 %~ UDL %~ UNIV-LYON