Illustration of CIFAR-100 dataset examples. CIFAR-100 contains tiny images of 100 classes, with a resolution of 32x32. Shah et al. [28] managed to train CIFAR-100 using different ResNet models, including their variation, where ELU (Exponential Linear Unit) [29] was used as an activation function. Their test error on standard ResNet-101 achieved 27.23%. For this reason, we decided to use residual networks in our experiments.

Illustration of CIFAR-100 dataset examples. CIFAR-100 contains tiny images of 100 classes, with a resolution of 32x32. Shah et al. [28] managed to train CIFAR-100 using different ResNet models, including their variation, where ELU (Exponential Linear Unit) [29] was used as an activation function. Their test error on standard ResNet-101 achieved 27.23%. For this reason, we decided to use residual networks in our experiments.

Source publication
Preprint
Full-text available
This paper describes the transformation of a traditional in-silico classification network into an optical fully convolutional neural network with high-resolution feature maps and kernels. When using the free-space 4f system to accelerate the inference speed of neural networks, higher resolutions of feature maps and kernels can be used without the l...

Context in source publication

Context 1
... trained our network with the CIFAR-100 dataset (see Figure 2) and chose the ResNet-18 as the backbone network. The CIFAR-100 (Canadian Institute For Advanced Research) dataset consists of 60,000 images of 32x32 resolution. ...

Similar publications

Article
Full-text available
This paper describes the transformation of a traditional in silico classification network into an optical fully convolutional neural network with high-resolution feature maps and kernels. When using the free-space 4f system to accelerate the inference speed of neural networks, higher resolutions of feature maps and kernels can be used without the l...