Data Science by ODS.ai 🦜
51.6K subscribers
339 photos
30 videos
7 files
1.5K links
First Telegram Data Science channel. Covering all technical and popular staff about anything related to Data Science: AI, Big Data, Machine Learning, Statistics, general Math and the applications of former. To reach editors contact: @haarrp
Download Telegram
STMVis - Visual Analysis for Recurrent Neural Networks

LSTMVis a visual analysis tool for recurrent neural networks with a focus on understanding these hidden state dynamics. The tool allows a user to select a hypothesis input range to focus on local state changes, to match these states changes to similar patterns in a large data set, and to align these results with structural annotations from their domain. We provide data for the tool to analyze specific hidden state properties on dataset containing nesting, phrase structure, and chord progressions, and demonstrate how the tool can be used to isolate patterns for further statistical analysis.

http://lstm.seas.harvard.edu/

#harvard #video #dl #rnn
Another #CVPR2018 paper award: Global Pose Estimation with Attention-based RNNs.

Arxiv: https://arxiv.org/abs/1802.06857

#pose #rnn #dl
A cool paper from Facebook AI (not from FAIR!) about detecting and reading text in images, at scale. 

This is very useful for detecting inappropriate content on Facebook.

The system uses R-CNN/Detectron for detecting lines of text.

The OCR uses a ConvNet applied at the level of a whole line trained with CTC.

This concept of applying a ConvNet on a whole line of text, without prior segmentation, has roots in the early days of ConvNets, for example with this NIPS 1992 paper:
"Multi-Digit Recognition Using a Space Displacement Neural Network"
by Ofer Matan, Chris Burges, Yann LeCun and John Denker.

Link: https://papers.nips.cc/paper/557-multi-digit-recognition-using-a-space-displacement-neural-network
Youtuve video with short explanation: https://youtu.be/yl3P2tYewVg

#ocr #cv #dl #rnn #facebook #yannlecun #video
Reversible RNNs

Paper about how to reduce memory costs of GRU and LSTM networks by 10-15x without loss in performance. Also 5-10x for attention-based architectures. New paper with Matt MacKay, Paul Vicol, and Jimmy Ba, to appear at NIPS.

Link: https://arxiv.org/abs/1810.10999

#dl #RNN #NIPS2018
And the same for #ResNet, #RNN and feed-forward #nn without residual connections.

Gradient Descent Finds Global Minima of Deep Neural Networks
ArXiV: https://arxiv.org/pdf/1811.03804.pdf

On the Convergence Rate of Training Recurrent Neural Networks
ArXiV: https://arxiv.org/pdf/1810.12065.pdf

A Convergence Theory for Deep Learning via Over-Parameterization
ArXiV: https://arxiv.org/pdf/1811.03962.pdf

#dl
​​How Uber predicts prices

Engineering Uncertainty Estimation in Neural Networks for Time Series Prediction at Uber

Link: https://eng.uber.com/neural-networks-uncertainty-estimation/

#RNN #LSTM #Uber
Engineering Uncertainty Estimation in Neural Networks for Time Series Prediction at Uber

More complex article on #TS forecasting from #Uber team.

Link: https://eng.uber.com/neural-networks-uncertainty-estimation/

#RNN #LSTM #Uber
​​Unified rational protein engineering with sequence-only deep representation learning

UniRep predicts amino-acid sequences that form stable bonds. In industry, that’s vital for determining the production yields, reaction rates, and shelf life of protein-based products.

Link: https://www.biorxiv.org/content/10.1101/589333v1.full

#biolearning #rnn #Harvard #sequence #protein