Measuring (in)variances in Convolutional Networks

Convolutional neural networks (CNN) offer state-of-the-art performance in various computer vision tasks such as activity recognition, face detection, medical image analysis, among others. Many of those tasks need invariance to image transformations (i.e.. rotations, translations or scaling). This w...

Descripción completa

Detalles Bibliográficos
Autores principales: Quiroga, Facundo, Torrents-Barrena, Jordina, Lanzarini, Laura Cristina, Puig, Domenec
Formato: Objeto de conferencia
Lenguaje:Inglés
Publicado: 2019
Materias:
Acceso en línea:http://sedici.unlp.edu.ar/handle/10915/80387
Aporte de:
id I19-R120-10915-80387
record_format dspace
institution Universidad Nacional de La Plata
institution_str I-19
repository_str R-120
collection SEDICI (UNLP)
language Inglés
topic Ciencias Informáticas
transformation invariance
rotation invariance
Neural networks
variance measure
MNIST dataset
CIFAR10 dataset
Residual Network
VGG Network
AllConvolutional Network
spellingShingle Ciencias Informáticas
transformation invariance
rotation invariance
Neural networks
variance measure
MNIST dataset
CIFAR10 dataset
Residual Network
VGG Network
AllConvolutional Network
Quiroga, Facundo
Torrents-Barrena, Jordina
Lanzarini, Laura Cristina
Puig, Domenec
Measuring (in)variances in Convolutional Networks
topic_facet Ciencias Informáticas
transformation invariance
rotation invariance
Neural networks
variance measure
MNIST dataset
CIFAR10 dataset
Residual Network
VGG Network
AllConvolutional Network
description Convolutional neural networks (CNN) offer state-of-the-art performance in various computer vision tasks such as activity recognition, face detection, medical image analysis, among others. Many of those tasks need invariance to image transformations (i.e.. rotations, translations or scaling). This work proposes a versatile, straightforward and interpretable measure to quantify the (in)variance of CNN activations with respect to transformations of the input. Intermediate output values of feature maps and fully connected layers are also analyzed with respect to different input transformations. The technique is applicable to any type of neural network and/or transformation. Our technique is validated on rotation transformations and compared with the relative (in)variance of several networks. More specifically, ResNet, AllConvolutional and VGG architectures were trained on CIFAR10 and MNIST databases with and without rotational data augmentation. Experiments reveal that rotation (in)variance of CNN outputs is class conditional. A distribution analysis also shows that lower layers are the most invariant, which seems to go against previous guidelines that recommend placing invariances near the network output and equivariances near the input.
format Objeto de conferencia
Objeto de conferencia
author Quiroga, Facundo
Torrents-Barrena, Jordina
Lanzarini, Laura Cristina
Puig, Domenec
author_facet Quiroga, Facundo
Torrents-Barrena, Jordina
Lanzarini, Laura Cristina
Puig, Domenec
author_sort Quiroga, Facundo
title Measuring (in)variances in Convolutional Networks
title_short Measuring (in)variances in Convolutional Networks
title_full Measuring (in)variances in Convolutional Networks
title_fullStr Measuring (in)variances in Convolutional Networks
title_full_unstemmed Measuring (in)variances in Convolutional Networks
title_sort measuring (in)variances in convolutional networks
publishDate 2019
url http://sedici.unlp.edu.ar/handle/10915/80387
work_keys_str_mv AT quirogafacundo measuringinvariancesinconvolutionalnetworks
AT torrentsbarrenajordina measuringinvariancesinconvolutionalnetworks
AT lanzarinilauracristina measuringinvariancesinconvolutionalnetworks
AT puigdomenec measuringinvariancesinconvolutionalnetworks
bdutipo_str Repositorios
_version_ 1764820486914572294