Music Transformer
arxiv.org/abs/1809.04281
#DeepLearning #Transformer #MachineLearning #SpeechProcessing #Music
arxiv.org/abs/1809.04281
#DeepLearning #Transformer #MachineLearning #SpeechProcessing #Music
arXiv.org
Music Transformer
Music relies heavily on repetition to build structure and meaning. Self-reference occurs on multiple timescales, from motifs to phrases to reusing of entire sections of music, such as in pieces...
MidiMe: Personalizing MusicVAE
Dinculescu et al.: https://magenta.tensorflow.org/midi-me
#ArtificialIntelligence #DeepLearning #Transformer
Dinculescu et al.: https://magenta.tensorflow.org/midi-me
#ArtificialIntelligence #DeepLearning #Transformer
Magenta
MidiMe: Personalizing MusicVAE
One of the areas of interest for the Magenta project is to empower individual expression. But how do you personalize a machine learning model and make it you...
MegatronLM: Training Billion+ Parameter Language Models Using GPU Model Parallelism
"... training an 8.3 billion parameter transformer language model with 8-way model parallelism and 64-way data parallelism on 512 GPUs, making it the largest transformer based language model ever trained at 24x the size of BERT and 5.6x the size of GPT-2."
Blog by NVIDIA Applied Deep Learning Research : https://nv-adlr.github.io/MegatronLM
Code: https://github.com/nvidia/megatron-lm
#ArtificialIntelligence #DeepLearning #NLP #PyTorch #Transformer
"... training an 8.3 billion parameter transformer language model with 8-way model parallelism and 64-way data parallelism on 512 GPUs, making it the largest transformer based language model ever trained at 24x the size of BERT and 5.6x the size of GPT-2."
Blog by NVIDIA Applied Deep Learning Research : https://nv-adlr.github.io/MegatronLM
Code: https://github.com/nvidia/megatron-lm
#ArtificialIntelligence #DeepLearning #NLP #PyTorch #Transformer
NVIDIA ADLR
MegatronLM: Training Billion+ Parameter Language Models Using GPU Model Parallelism
We train an 8.3 billion parameter transformer language model with 8-way model parallelism and 64-way data parallelism on 512 GPUs, making it the largest transformer based language model ever trained at 24x the size of BERT and 5.6x the size of GPT-2
Write With Transformer
See how a modern neural network auto-completes your text 🤗
With the brand new GPT-2 large!
Built by the phenomenal Hugging Face team : https://transformer.huggingface.co
H / T : Lysandre Debut
#GPT2 #NeuralNetwork #Transformer
See how a modern neural network auto-completes your text 🤗
With the brand new GPT-2 large!
Built by the phenomenal Hugging Face team : https://transformer.huggingface.co
H / T : Lysandre Debut
#GPT2 #NeuralNetwork #Transformer
banana-projects-transformer-autocomplete.hf.space
Write With Transformer
See how a modern neural network auto-completes your text
Write With Transformer
Hugging Face released a new version of their Write With Transformer app, using a language model trained directly on Arxiv to generate Deep Learning and NLP completions!
In addition, they add state-of-the-art NLP models such as GPT, GPT-2 and XLNet completions:
https://transformer.huggingface.co/
H / T : Lysandre Debut
#Transformer #Pytorch #NLP
@ArtificialIntelligenceArticles
Hugging Face released a new version of their Write With Transformer app, using a language model trained directly on Arxiv to generate Deep Learning and NLP completions!
In addition, they add state-of-the-art NLP models such as GPT, GPT-2 and XLNet completions:
https://transformer.huggingface.co/
H / T : Lysandre Debut
#Transformer #Pytorch #NLP
@ArtificialIntelligenceArticles
banana-projects-transformer-autocomplete.hf.space
Write With Transformer
See how a modern neural network auto-completes your text
On Extractive and Abstractive Neural Document Summarization with Transformer Language Models"
Sandeep Subramanian, Raymond Li, Jonathan Pilault, Christopher Pal : https://arxiv.org/abs/1909.03186
#transformer #naturallanguageprocessing #machinelearning
Sandeep Subramanian, Raymond Li, Jonathan Pilault, Christopher Pal : https://arxiv.org/abs/1909.03186
#transformer #naturallanguageprocessing #machinelearning
Forecaster: A Graph Transformer for Forecasting Spatial and Time-Dependent Data
Yang Li, José M. F. Moura : https://arxiv.org/abs/1909.04019v3
#MachineLearning #ArtificialIntelligence #Transformer
Yang Li, José M. F. Moura : https://arxiv.org/abs/1909.04019v3
#MachineLearning #ArtificialIntelligence #Transformer
Compressive Transformers for Long-Range Sequence Modelling
Anonymous : https://openreview.net/forum?id=SylKikSYDH
#ArtificialIntelligence #MachineLearning #Transformer
Anonymous : https://openreview.net/forum?id=SylKikSYDH
#ArtificialIntelligence #MachineLearning #Transformer
OpenReview
Compressive Transformers for Long-Range Sequence Modelling
Long-range transformer using a compressive memory, achieves sota in wikitext-103 and enwik8 LM benchmarks, release a new book-level LM benchmark PG-19.
The Illustrated GPT-2 (Visualizing Transformer Language Models)
Blog by Jay Alammar : https://jalammar.github.io/illustrated-gpt2/
#BERT #Transformer #ArtificialIntelligence
Blog by Jay Alammar : https://jalammar.github.io/illustrated-gpt2/
#BERT #Transformer #ArtificialIntelligence
jalammar.github.io
The Illustrated GPT-2 (Visualizing Transformer Language Models)
Discussions:
Hacker News (64 points, 3 comments), Reddit r/MachineLearning (219 points, 18 comments)
Translations: Simplified Chinese, French, Korean, Russian, Turkish
This year, we saw a dazzling application of machine learning. The OpenAI GPT…
Hacker News (64 points, 3 comments), Reddit r/MachineLearning (219 points, 18 comments)
Translations: Simplified Chinese, French, Korean, Russian, Turkish
This year, we saw a dazzling application of machine learning. The OpenAI GPT…
Q8BERT: Quantized 8Bit BERT
Zafrir et al.: https://arxiv.org/abs/1910.06188
#NaturalLanguageProcessing #NLP #Transformer
Zafrir et al.: https://arxiv.org/abs/1910.06188
#NaturalLanguageProcessing #NLP #Transformer
arXiv.org
Q8BERT: Quantized 8Bit BERT
Recently, pre-trained Transformer based language models such as BERT and GPT, have shown great improvement in many Natural Language Processing (NLP) tasks. However, these models contain a large...
CE7454 : Deep Learning for Data Science
Lecture 13: Attention Neural Networks
Xavier Bresson : https://dropbox.com/s/kbrsvhwe2lac1uo/lecture13_attention_neural_networks.pdf?dl=0
Demo :
https://github.com/xbresson/CE7454_2019/blob/master/codes/labs_lecture13/seq2seq_transformers_demo.ipynb
#DeepLearning #DataScience #Transformer
Lecture 13: Attention Neural Networks
Xavier Bresson : https://dropbox.com/s/kbrsvhwe2lac1uo/lecture13_attention_neural_networks.pdf?dl=0
Demo :
https://github.com/xbresson/CE7454_2019/blob/master/codes/labs_lecture13/seq2seq_transformers_demo.ipynb
#DeepLearning #DataScience #Transformer
Dropbox
lecture13_attention_neural_networks.pdf
Shared with Dropbox
DialoGPT: Large-Scale Generative Pre-training for Conversational Response Generation
Zhang et al.: https://arxiv.org/abs/1911.00536
#ArtificialIntelligence #MachineLearning #Transformer
Zhang et al.: https://arxiv.org/abs/1911.00536
#ArtificialIntelligence #MachineLearning #Transformer
arXiv.org
DialoGPT: Large-Scale Generative Pre-training for Conversational...
We present a large, tunable neural conversational response generation model, DialoGPT (dialogue generative pre-trained transformer). Trained on 147M conversation-like exchanges extracted from...
Reformer: The Efficient Transformer
Kitaev et al.: https://openreview.net/forum?id=rkgNKkHtvB
#Artificialintelligence #DeepLearning #Transformer
Kitaev et al.: https://openreview.net/forum?id=rkgNKkHtvB
#Artificialintelligence #DeepLearning #Transformer
OpenReview
Reformer: The Efficient Transformer
Efficient Transformer with locality-sensitive hashing and reversible layers
Multi-Graph Transformer for Free-Hand Sketch Recognition
Xu et al.: https://arxiv.org/abs/1912.11258
#ArtificialIntelligence #DeepLearning #Transformer
Xu et al.: https://arxiv.org/abs/1912.11258
#ArtificialIntelligence #DeepLearning #Transformer
Text classification with Transformer
Apoorv Nandan, Colab : https://colab.research.google.com/github/keras-team/keras-io/blob/master/examples/nlp/ipynb/text_classification_with_transformer.ipynb
#ArtificialIntelligence #DeepLearning #Transformer
Apoorv Nandan, Colab : https://colab.research.google.com/github/keras-team/keras-io/blob/master/examples/nlp/ipynb/text_classification_with_transformer.ipynb
#ArtificialIntelligence #DeepLearning #Transformer
Google
text_classification_with_transformer
Run, share, and edit Python notebooks
Transformer Reinforcement Learning
Leandro von Werra, GitHub : https://github.com/lvwerra/trl
#ProximalPolicyOptimization #Transformer #ReinforcementLearning
Leandro von Werra, GitHub : https://github.com/lvwerra/trl
#ProximalPolicyOptimization #Transformer #ReinforcementLearning
GitHub
GitHub - huggingface/trl: Train transformer language models with reinforcement learning.
Train transformer language models with reinforcement learning. - huggingface/trl
Efficient Transformers: A Survey
Tay et al.: https://arxiv.org/abs/2009.06732
#Transformer #DeepLearning #ArtificialIntelligence
Tay et al.: https://arxiv.org/abs/2009.06732
#Transformer #DeepLearning #ArtificialIntelligence