Skip to Main content Skip to Navigation
Journal articles

Coupled Ensembles of Neural Networks

Abstract : We investigate in this paper the architecture of deep convolutional networks. Building on existing state of the art models, we propose a reconfiguration of the model parameters into several parallel branches at the global network level, with each branch being a standalone CNN. We show that this arrangement is an efficient way to significantly reduce the number of parameters while at the same time improving the performance. The use of branches brings an additional form of regularization. In addition to splitting the parameters into parallel branches, we propose a tighter coupling of these branches by averaging their log-probabilities. The tighter coupling favours the learning of better representations, even at the level of the individual branches, as compared to when each branch is trained independently. We refer to this branched architecture as “coupled ensembles”. The approach is generic and can be applied to almost any neural network architecture. With coupled ensembles of DenseNet-BC and parameter budget of 25 M, we obtain error rates of 2.92%, 15.68% and 1.50% on CIFAR-10, CIFAR-100 and SVHN, respectively. For the same parameter budget, DenseNet-BC has an error rate of 3.46%, 17.18%, and 1.8%, respectively. With ensembles of coupled ensembles, of DenseNet-BC networks, with 50M total parameters, we obtain error rates of 2.72%, 15.13% and 1.42%, respectively on these tasks.
Complete list of metadatas

https://hal.archives-ouvertes.fr/hal-02088253
Contributor : Georges Quénot <>
Submitted on : Tuesday, April 2, 2019 - 4:53:58 PM
Last modification on : Saturday, August 29, 2020 - 3:29:44 AM

Links full text

Identifiers

Citation

Anuvabh Dutt, Georges Quénot, Denis Pellerin. Coupled Ensembles of Neural Networks. Neurocomputing, Elsevier, 2019, ⟨10.1016/j.neucom.2018.10.092⟩. ⟨hal-02088253⟩

Share

Metrics

Record views

234