Convolutional structures and marginal statistics--a study based on K-nearest neighbours

Authors

  • Jugurta Montalvão Federal University of Sergipe
  • Jânio Canuto Federal University of Sergipe
  • Elyson Carvalho Federal University of Sergipe

DOI:

https://doi.org/10.14209/jcis.2018.19

Keywords:

Data scarcity, explainable Artificial Intelligence, KNN Network.

Abstract

This paper addresses statistical tricks found in deep convolutive neural networks. First, the most relevant statistical tricks are studied under the perspective of data scarcity, then one of them, directly related to convolution-like structures, is regarded as a random variable marginalization. The same kind of marginalization is implemented in an ensemble of K-nearest neighbours cells, where each cell yields scores instead of class labels. Scores are then combined to improve classification accuracy, as compared to a conventional K-nearest neighbours classifier in experiments with two emblematic datasets---MNIST and CIFAR-10. This improvement is regarded as evidence of the variable marginalization effect over performance, whereas it is discussed the potential for further lessons learned from deep neural networks to be transferred to KNN based classifiers, whose advantage is to allow for explainable artificial intelligence.

Downloads

Download data is not yet available.

Downloads

Published

2018-07-03

How to Cite

Montalvão, J., Canuto, J., & Carvalho, E. (2018). Convolutional structures and marginal statistics--a study based on K-nearest neighbours. Journal of Communication and Information Systems, 33(1). https://doi.org/10.14209/jcis.2018.19

Issue

Section

Regular Papers
Received 2018-03-19
Accepted 2018-06-19
Published 2018-07-03

Most read articles by the same author(s)