With his students Sepp Hochreiter, Felix Gers, Fred Cummins, Alex Graves, and others, Schmidhuber published increasingly sophisticated versions of a type of recurrent neural network called the long short-term memory. First results were already reported in Hochreiter's diploma thesis which analyzed and overcame the famous vanishing gradient problem. The name LSTM was introduced in a tech report leading to the most cited LSTM publication. The standard LSTM architecture which is used in almost all current applications was introduced in 2000. Today's "vanilla LSTM" using backpropagation through time was published in 2005, and its connectionist temporal classification training algorithm in 2006. CTC enabled end-to-end speech recognition with LSTM. In 2015, LSTM trained by CTC was used in a new implementation of speech recognition in Google's software for smartphones. Google also used LSTM for the smart assistant Allo and for Google Translate. Apple used LSTM for the "Quicktype" function on the iPhone and for Siri. Amazon used LSTM for Amazon Alexa. In 2017, Facebook performed some 4.5 billion automatic translations every day using LSTM networks. Bloomberg Business Week wrote: "These powers make LSTM arguably the most commercial AI achievement, used for everything from predicting diseases to composing music." In 2011, Schmidhuber's team at IDSIA with his postdoc Dan Ciresan also achieved dramatic speedups of convolutional neural networks on fast parallel computers called GPUs. An earlier CNN on GPU by Chellapilla et al. was 4 times faster than an equivalent implementation on CPU. The deep CNN of Dan Ciresan et al. at IDSIA was already 60 times faster and achieved the first superhuman performance in a computer vision contest in August 2011. Between May 15, 2011 and September 10, 2012, their fast and deep CNNs won no fewer than four image competitions. They also significantly improved on the best performance in the literature for multiple image databases. The approach has become central to the field of computer vision. It is based on CNN designs introduced much earlier by Yann LeCun et al. who applied the backpropagation algorithm to a variant of Kunihiko Fukushima's original CNN architecture called neocognitron, later modified by J. Weng's method called max-pooling. In 2014, Schmidhuber formed a company, Nnaisense, to work on commercial applications of artificial intelligence in fields such as finance, heavy industry and self-driving cars. Sepp Hochreiter, Jaan Tallinn, and Marcus Hutter are advisers to the company. Sales were under US$11 million in 2016; however, Schmidhuber states that the current emphasis is on research and not revenue. Nnaisense raised its first round of capital funding in January 2017. Schmidhuber's overall goal is to create an all-purpose AI by training a single AI in sequence on a variety of narrow tasks; however, skeptics point out that companies such as Arago GmbH and IBM have applied AI to various different projects for years without showing any signs of artificial general intelligence.
Views
According to The Guardian, Schmidhuber complained in a "scathing 2015 article" that fellow deep learning researchers Geoffrey Hinton, Yann LeCun and Yoshua Bengio "heavily cite each other," but "fail to credit the pioneers of the field,” allegedly understating the contributions of Schmidhuber and other early machine learning pioneers including Alexey Grigorevich Ivakhnenko who published the first deep learning networks already in 1965. LeCun denies the charge, stating instead that Schmidhuber "keeps claiming credit he doesn't deserve".