Transcription of ImageNet Classification with Deep Convolutional Neural …
{{id}} {{{paragraph}}}
ImageNet Classification with Deep Convolutional Neural Networks Alex Krizhevsky Ilya Sutskever Geoffrey E. Hinton University of Toronto University of Toronto University of Toronto Abstract We trained a large, deep Convolutional Neural network to classify the million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 dif- ferent classes. On the test data, we achieved top-1 and top-5 error rates of and which is considerably better than the previous state-of-the-art. The Neural network, which has 60 million parameters and 650,000 neurons, consists of five Convolutional layers, some of which are followed by max-pooling layers, and three fully-connected layers with a final 1000-way softmax. To make train- ing faster, we used non-saturating neurons and a very efficient GPU implemen- tation of the convolution operation. To reduce overfitting in the fully-connected layers we employed a recently-developed regularization method called dropout.
five convolutional and three fully-connected. Below, we describe some of the novel or unusual features of our network’s architecture. Sections 3.1-3.4 are sorted according to our estimation of ... this work if we had used traditional saturating neuron models. ... this comparison is biased in favor of the one-GPU net, since it is bigger than ...
Domain:
Source:
Link to this page:
Please notify us if you found a problem with this document:
{{id}} {{{paragraph}}}