Deep neural networks (DNNs) have been shown to perform well on exclusive,
multi-class classification tasks. However, when different classes have similar
visual features, it becomes challenging for human annotators to differentiate them.
This scenario necessitates the use of composite class labels. In this paper, we
propose a novel framework called Hyper-Evidential Neural Network (HENN)
that explicitly models predictive uncertainty due to composite class labels in
training data in the context of the belief theory called Subjective Logic (SL).
By placing a grouped Dirichlet distribution on the class probabilities, we treat
predictions of a neural network as parameters of hyper-subjective opinions and
learn the network that collects both single and composite evidence leading to
these hyper-opinions by a deterministic DNN from data. We introduce a new
uncertainty type called vagueness originally designed for hyper-opinions in SL to
quantify composite classification uncertainty for DNNs. Our results demonstrate
that HENN outperforms its state-of-the-art counterparts based on four image
datasets. The code and datasets are available at: https://github.com/
Hugo101/HyperEvidentialNN.
more »
« less
Conditional Classification: A Solution for Computational Energy Reduction
Deep convolutional neural networks have shown high efficiency in computer visions and other applications. However, with the increase in the depth of the networks, the computational complexity is growing exponentially. In this paper, we propose a novel solution to reduce the computational complexity of convolutional neural network models used for many class image classification. Our proposed technique breaks the classification task into two steps: 1) coarse-grain classification, in which the input samples are classified among a set of hyper-classes, 2) fine-grain classification, in which the final labels are predicted among those hyper-classes detected at the first step. We illustrate that our proposed classifier can reach the level of accuracy reported by the best in class classification models with less computational complexity (Flop Count) by only activating parts of the model that are needed for the image classification.
more »
« less
- NSF-PAR ID:
- 10298696
- Date Published:
- Journal Name:
- 22nd International Symposium on Quality Electronic Design (ISQED)
- Page Range / eLocation ID:
- 325 to 330
- Format(s):
- Medium: X
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Convolutional Neural Networks (CNNs) filter the input data using spatial convolution operators with compact stencils. Commonly, the convolution operators couple features from all channels, which leads to immense computational cost in the training of and prediction with CNNs. To improve the efficiency of CNNs, we introduce lean convolution operators that reduce the number of parameters and computational complexity, and can be used in a wide range of existing CNNs. Here, we exemplify their use in residual networks (ResNets), which have been very reliable for a few years now and analyzed intensively. In our experiments on three image classification problems, the proposed LeanResNet yields results that are comparable to other recently proposed reduced architectures using similar number of parameters.more » « less
-
null (Ed.)In this paper, we propose Code-Bridged Classifier (CBC), a framework for making a Convolutional Neural Network (CNNs) robust against adversarial attacks without increasing or even by decreasing the overall models' computational complexity. More specifically, we propose a stacked encoder-convolutional model, in which the input image is first encoded by the encoder module of a denoising auto-encoder, and then the resulting latent representation (without being decoded) is fed to a reduced complexity CNN for image classification. We illustrate that this network not only is more robust to adversarial examples but also has a significantly lower computational complexity when compared to the prior art defenses.more » « less
-
We demonstrate that image reconstruction can be achieved via a convolutional neural network for a “see-through” computational camera comprised of a transparent window and CMOS image sensor. Furthermore, we compared classification results using a classifier network for the raw sensor data against those with the reconstructed images. The results suggest that similar classification accuracy is likely possible in both cases with appropriate network optimizations. All networks were trained and tested for the MNIST (6 classes), EMNIST, and the Kanji49 datasets.
-
With Convolutional Neural Networks (CNN) becoming more of a commodity in the computer vision field, many have attempted to improve CNN in a bid to achieve better accuracy to a point that CNN accuracies have surpassed that of human's capabilities. However, with deeper networks, the number of computations and consequently the power needed per classification has grown considerably. In this paper, we propose Iterative CNN (ICNN) by reformulating the CNN from a single feed-forward network to a series of sequentially executed smaller networks. Each smaller network processes a sub-sample of input image, and features extracted from previous network, and enhances the classification accuracy. Upon reaching an acceptable classification confidence, ICNN immediately terminates. The proposed network architecture allows the CNN function to be dynamically approximated by creating the possibility of early termination and performing the classification with far fewer operations compared to a conventional CNN. Our results show that this iterative approach competes with the original larger networks in terms of accuracy while incurring far less computational complexity by detecting many images in early iterations.more » « less