Making floating point math highly efficient for AI hardware.
Fresh FB article on accelerating ML.
https://code.fb.com/ai-research/floating-point-math/
#facebook
Fresh FB article on accelerating ML.
https://code.fb.com/ai-research/floating-point-math/
Engineering at Meta
Making floating point math highly efficient for AI hardware
In recent years, compute-intensive artificial intelligence tasks have prompted creation of a wide variety of custom hardware to run these powerful new systems efficiently. Deep learning models, sucβ¦
Analyzing Experiment Outcomes: Beyond Average Treatment Effects
Cool piece from Uber's engineering department about why you can't just use the average customer experience to see if product changes are worth it. You have to consider the DISTRIBUTIONAL changes of the customer experience.
Link: https://eng.uber.com/analyzing-experiment-outcomes/
#statistics #uber #abtest
Cool piece from Uber's engineering department about why you can't just use the average customer experience to see if product changes are worth it. You have to consider the DISTRIBUTIONAL changes of the customer experience.
Link: https://eng.uber.com/analyzing-experiment-outcomes/
#statistics #uber #abtest
ββTL-GAN: transparent latent-space GAN
GANs to generate photo-realistic faces with some control over characteristics.
Demo: https://www.kaggle.com/summitkwan/tl-gan-demo
Medium Post: https://blog.insightdatascience.com/generating-custom-photo-realistic-faces-using-ai-d170b1b59255
Github: https://github.com/SummitKwan/transparent_latent_gan#1-instructions-on-the-online-demo
#GAN #cv
GANs to generate photo-realistic faces with some control over characteristics.
Demo: https://www.kaggle.com/summitkwan/tl-gan-demo
Medium Post: https://blog.insightdatascience.com/generating-custom-photo-realistic-faces-using-ai-d170b1b59255
Github: https://github.com/SummitKwan/transparent_latent_gan#1-instructions-on-the-online-demo
#GAN #cv
Amazonβs SageMaker Object2Vec, a highly customizable algorithm that can learn embeddings of various types high-dimensional objects.
Link: https://aws.amazon.com/ru/blogs/machine-learning/introduction-to-amazon-sagemaker-object2vec/
#Object2Vec #Amazon #Embeddings
Link: https://aws.amazon.com/ru/blogs/machine-learning/introduction-to-amazon-sagemaker-object2vec/
#Object2Vec #Amazon #Embeddings
Prototypical Clustering Networks for Dermatological Disease Diagnosis
Paper will be presented at the ML4D workshop at #NIPS2018
Link: https://arxiv.org/abs/1811.03066
#nn #bio #medical
Paper will be presented at the ML4D workshop at #NIPS2018
Link: https://arxiv.org/abs/1811.03066
#nn #bio #medical
Monitor Your PyTorch Models With Five Extra Lines of Code
Ever felt like manually managing your Visdom / TensorBoard server and logs is a pain across experiments, projects and teams?
Weights & Biases provides a simple cloud-based experiment logging and plotting system, with easy integration for PyTorch models.
Link: https://www.wandb.com/blog/monitor-your-pytorch-models-with-five-extra-lines-of-code
#pytorch
Ever felt like manually managing your Visdom / TensorBoard server and logs is a pain across experiments, projects and teams?
Weights & Biases provides a simple cloud-based experiment logging and plotting system, with easy integration for PyTorch models.
Link: https://www.wandb.com/blog/monitor-your-pytorch-models-with-five-extra-lines-of-code
#pytorch
wandb.ai
Monitor Your PyTorch Models With Five Extra Lines of Code on Weights & Biases
by Lukas Biewald β I love PyTorch and I love experiment tracking, here's how to do both!
New paper on Lipschitz neural net architectures. Uses sorting as an activation function, with matrix norm constrained weights. Universal Lipschitz function approx. Enforce adversarial robustness (margin) using hinge loss.
Link: https://arxiv.org/abs/1811.05381
#nn #lipschitz
Link: https://arxiv.org/abs/1811.05381
#nn #lipschitz
ββNeural network 3D visualization framework. Very nice in-depth visualizations.
Now you can actually see how the layers look.
Github: https://github.com/tensorspace-team/tensorspace
LiveDemo (!): https://tensorspace.org/html/playground/vgg16.html
#visualization #nn
Now you can actually see how the layers look.
Github: https://github.com/tensorspace-team/tensorspace
LiveDemo (!): https://tensorspace.org/html/playground/vgg16.html
#visualization #nn
ββReally interesting talk at MLconfSF by Franziska Bell on how #Uber uses NLP for customer experience. Most of what was described are recent advances in their COTA platform.
Link: https://eng.uber.com/cota/
Link: https://eng.uber.com/cota/
ββDeepMasterPrints: Generating MasterPrints for Dictionary Attacks via Latent Variable Evolution
Using GANs to generate MasterFingerPrints that unlock 22-78% phones sensors (dep. on security level of sensor). It doesn't get much more "adversarial" than that.
This work can be potentially used to create fingerprint which can be used to match 22-78% of fingerprints in the wild, creating Skeleton key, fitting any security system, including home alarm or phone lock.
ArXiV: https://arxiv.org/pdf/1705.07386.pdf
#GAN #security #fingerprint
Using GANs to generate MasterFingerPrints that unlock 22-78% phones sensors (dep. on security level of sensor). It doesn't get much more "adversarial" than that.
This work can be potentially used to create fingerprint which can be used to match 22-78% of fingerprints in the wild, creating Skeleton key, fitting any security system, including home alarm or phone lock.
ArXiV: https://arxiv.org/pdf/1705.07386.pdf
#GAN #security #fingerprint
Sptoify announced its new Data Science Challenge
Spotify Sequential Skip Prediction Challenge is a part of #WSDM Cup 2019. The dataset comprises 130M Spotify listening sessions, and the task is to predict if a track is skipped. The challenge is live today, and runs until Jan 4.
Link: https://www.crowdai.org/challenges/spotify-sequential-skip-prediction-challenge
#kaggle #CompetitiveDataScience #Spotify
Spotify Sequential Skip Prediction Challenge is a part of #WSDM Cup 2019. The dataset comprises 130M Spotify listening sessions, and the task is to predict if a track is skipped. The challenge is live today, and runs until Jan 4.
Link: https://www.crowdai.org/challenges/spotify-sequential-skip-prediction-challenge
#kaggle #CompetitiveDataScience #Spotify
ImageNet/ResNet-50 Training speed dramatically (6.6 min -> 224 sec) reduced
ResNet-50 on ImageNet now (allegedly) down to 224sec (3.7min) using 2176 V100s. Increasing batch size schedule, LARS, 5 epoch LR warmup, synch BN without mov avg. (mixed) fp16 training. "2D-Torus" all-reduce on NCCL2, with NVLink2 & 2 IB EDR interconnect.
1.28M images over 90 epochs with 68K batches, so the entire optimization is ~1700 updates to converge.
ArXiV: https://arxiv.org/abs/1811.05233
#ImageNet #ResNet
ResNet-50 on ImageNet now (allegedly) down to 224sec (3.7min) using 2176 V100s. Increasing batch size schedule, LARS, 5 epoch LR warmup, synch BN without mov avg. (mixed) fp16 training. "2D-Torus" all-reduce on NCCL2, with NVLink2 & 2 IB EDR interconnect.
1.28M images over 90 epochs with 68K batches, so the entire optimization is ~1700 updates to converge.
ArXiV: https://arxiv.org/abs/1811.05233
#ImageNet #ResNet
New open source suite of ODE solvers in PyTorch
Everything happens on the GPU and is differentiable. Now you can use ODEs in your deep learning models.
Github: https://github.com/rtqichen/torchdiffeq
Everything happens on the GPU and is differentiable. Now you can use ODEs in your deep learning models.
Github: https://github.com/rtqichen/torchdiffeq
GitHub
GitHub - rtqichen/torchdiffeq: Differentiable ODE solvers with full GPU support and O(1)-memory backpropagation.
Differentiable ODE solvers with full GPU support and O(1)-memory backpropagation. - rtqichen/torchdiffeq
Forwarded from βΟΓΖuΓ±Γ§ΟΓΟ ΓΟΟ¡çβ | πβπ¨βΊβΊβΊβΊ
Gradient Descent Provably Optimizes Over-parameterized Neural Networks
Paper shows that the loss of two-layer neural networks can be optimized to zero in polynomial time using gradient descent.
ArXiV: https://arxiv.org/pdf/1810.02054.pdf
#nn #dl
Paper shows that the loss of two-layer neural networks can be optimized to zero in polynomial time using gradient descent.
ArXiV: https://arxiv.org/pdf/1810.02054.pdf
#nn #dl
And the same for #ResNet, #RNN and feed-forward #nn without residual connections.
Gradient Descent Finds Global Minima of Deep Neural Networks
ArXiV: https://arxiv.org/pdf/1811.03804.pdf
On the Convergence Rate of Training Recurrent Neural Networks
ArXiV: https://arxiv.org/pdf/1810.12065.pdf
A Convergence Theory for Deep Learning via Over-Parameterization
ArXiV: https://arxiv.org/pdf/1811.03962.pdf
#dl
Gradient Descent Finds Global Minima of Deep Neural Networks
ArXiV: https://arxiv.org/pdf/1811.03804.pdf
On the Convergence Rate of Training Recurrent Neural Networks
ArXiV: https://arxiv.org/pdf/1810.12065.pdf
A Convergence Theory for Deep Learning via Over-Parameterization
ArXiV: https://arxiv.org/pdf/1811.03962.pdf
#dl
All the statistical distributions and how they relate to each other!
Source: http://www.math.wm.edu/~leemis/2008amstat.pdf
#distributions #visualization #cheatsheet #statistics
Source: http://www.math.wm.edu/~leemis/2008amstat.pdf
#distributions #visualization #cheatsheet #statistics
Dynamic relationship visualization project
Two kill two birds with one stone, we should also share this #visualization tool.
http://distributome.org/tools.html
#statistics #distributions
Two kill two birds with one stone, we should also share this #visualization tool.
http://distributome.org/tools.html
#statistics #distributions
www.distributome.org
Distributome Tools Carousel
Probability and Statistics Distributions Resource - Distributome