Comment on Can you think of any now?

<- View Parent
bluemellophone@lemmy.world ⁨1⁩ ⁨week⁩ ago

No, you are correct. Hinton began researching ReLUs in 2010 and his students Alex Krizhevsky and Ilya Sutskever used it to train a much deeper network (AlexNet) to win the 2012 ILSVRC. The reason AlexNet was so groundbreaking was because it brought all of the gradient optimization improvements (SGD with momentum as popularized by Schmidhuber, and dropout), better activation functions (ReLU), a deeper network (8 layers), supervised training on very large datasets (necessary to learn good general-purpose convolutional kernels), and GPU acceleration into a single approach.

NNs, and specifically CNNs, won out because they were able to create more expressive and superior image feature representations over the hand-crafted features of competing algorithms. The proof was in the vastly better performance, it was a major jump when the ILSVRC was becoming saturated. Nobody was making nearly +10% improvements on that challenge back the , it blew everybody out of the water and made NNs and deep learning impossible to ignore.

source
Sort:hotnewtop