Write With Transformer
Hugging Face released a new version of their Write With Transformer app, using a language model trained directly on Arxiv to generate Deep Learning and NLP completions!
In addition, they add state-of-the-art NLP models such as GPT, GPT-2 and XLNet completions:
https://transformer.huggingface.co/
H / T : Lysandre Debut
#Transformer #Pytorch #NLP
@ArtificialIntelligenceArticles
Hugging Face released a new version of their Write With Transformer app, using a language model trained directly on Arxiv to generate Deep Learning and NLP completions!
In addition, they add state-of-the-art NLP models such as GPT, GPT-2 and XLNet completions:
https://transformer.huggingface.co/
H / T : Lysandre Debut
#Transformer #Pytorch #NLP
@ArtificialIntelligenceArticles
Are We Modeling the Task or the Annotator? An Investigation of Annotator Bias in Natural Language Understanding Datasets
Geva et al.: https://arxiv.org/abs/1908.07898
#ArtificialIntelligence #MachineLearning #NLP
Geva et al.: https://arxiv.org/abs/1908.07898
#ArtificialIntelligence #MachineLearning #NLP
What Kind of Language Is Hard to Language-Model?
Mielke et al.: https://arxiv.org/abs/1906.04726
#ArtificialIntelligence #MachineLearning #NLP
Mielke et al.: https://arxiv.org/abs/1906.04726
#ArtificialIntelligence #MachineLearning #NLP
Transformers: State-of-the-art Natural Language Processing for TensorFlow 2.0 and PyTorch
By 🤗 Hugging Face : https://huggingface.co/transformers
#Transformers #MachineLearning #NLP
By 🤗 Hugging Face : https://huggingface.co/transformers
#Transformers #MachineLearning #NLP
Extreme Language Model Compression with Optimal Subwords and Shared Projections
Zhao et al.: https://arxiv.org/abs/1909.11687
#neuralnetwork #bert #nlp
Zhao et al.: https://arxiv.org/abs/1909.11687
#neuralnetwork #bert #nlp
arXiv.org
Extremely Small BERT Models from Mixed-Vocabulary Training
Pretrained language models like BERT have achieved good results on NLP tasks, but are impractical on resource-limited devices due to memory footprint. A large fraction of this footprint comes from...
The State of Transfer Learning in NLP
By Sebastian Ruder : http://ruder.io/state-of-transfer-learning-in-nlp/
#TransferLearning #NaturalLanguageProcessing #NLP
By Sebastian Ruder : http://ruder.io/state-of-transfer-learning-in-nlp/
#TransferLearning #NaturalLanguageProcessing #NLP
ruder.io
The State of Transfer Learning in NLP
This post expands on the NAACL 2019 tutorial on Transfer Learning in NLP. It highlights key insights and takeaways and provides updates based on recent work.
Q8BERT: Quantized 8Bit BERT
Zafrir et al.: https://arxiv.org/abs/1910.06188
#NaturalLanguageProcessing #NLP #Transformer
Zafrir et al.: https://arxiv.org/abs/1910.06188
#NaturalLanguageProcessing #NLP #Transformer
arXiv.org
Q8BERT: Quantized 8Bit BERT
Recently, pre-trained Transformer based language models such as BERT and GPT, have shown great improvement in many Natural Language Processing (NLP) tasks. However, these models contain a large...
News classification using classic Machine Learning tools (TF-IDF) and modern NLP approach based on transfer learning (ULMFIT) deployed on GCP
By Imad El Hanafi
Live version: https://nlp.imadelhanafi.com/
Github: https://github.com/imadelh/NLP-news-classification
Blog: https://imadelhanafi.com/posts/text_classification_ulmfit/
#DeepLearning #MachineLearning #NLP
By Imad El Hanafi
Live version: https://nlp.imadelhanafi.com/
Github: https://github.com/imadelh/NLP-news-classification
Blog: https://imadelhanafi.com/posts/text_classification_ulmfit/
#DeepLearning #MachineLearning #NLP
Specializing Word Embeddings (for Parsing) by Information Bottleneck
Li et al.: https://www.aclweb.org/anthology/D19-1276.pdf
#ArtificialIntelligence #MachineLearning #NLP
Li et al.: https://www.aclweb.org/anthology/D19-1276.pdf
#ArtificialIntelligence #MachineLearning #NLP
The Illustrated GPT-2 (Visualizing Transformer Language Models)
Blog by Jay Alammar : https://jalammar.github.io/illustrated-gpt2/
#ArtificialIntelligence #NLP #UnsupervisedLearning
Blog by Jay Alammar : https://jalammar.github.io/illustrated-gpt2/
#ArtificialIntelligence #NLP #UnsupervisedLearning
Single Headed Attention RNN: Stop Thinking With Your Head
Stephen Merity : https://arxiv.org/abs/1911.11423
#ArtificialIntelligence #NeuralComputing #NLP
Stephen Merity : https://arxiv.org/abs/1911.11423
#ArtificialIntelligence #NeuralComputing #NLP
arXiv.org
Single Headed Attention RNN: Stop Thinking With Your Head
The leading approaches in language modeling are all obsessed with TV shows of my youth - namely Transformers and Sesame Street. Transformers this, Transformers that, and over here a bonfire worth...
A tutorial to implement state-of-the-art NLP models with Fastai for Sentiment Analysis
Maximilien Roberti : https://towardsdatascience.com/fastai-with-transformers-bert-roberta-xlnet-xlm-distilbert-4f41ee18ecb2
#FastAI #NLP #Transformers
Maximilien Roberti : https://towardsdatascience.com/fastai-with-transformers-bert-roberta-xlnet-xlm-distilbert-4f41ee18ecb2
#FastAI #NLP #Transformers
Single Headed Attention RNN: Stop Thinking With Your Head
Stephen Merity : https://arxiv.org/abs/1911.11423
#ArtificialIntelligence #NeuralComputing #NLP
Stephen Merity : https://arxiv.org/abs/1911.11423
#ArtificialIntelligence #NeuralComputing #NLP
arXiv.org
Single Headed Attention RNN: Stop Thinking With Your Head
The leading approaches in language modeling are all obsessed with TV shows of my youth - namely Transformers and Sesame Street. Transformers this, Transformers that, and over here a bonfire worth...
GitHub Typo Corpus: A Large-Scale Multilingual Dataset of Misspellings and Grammatical Errors
Masato Hagiwara, Masato Mita : https://arxiv.org/abs/1911.12893
Code & Dataset https://github.com/mhagiwara/github-typo-corpus
#ArtificialIntelligence #DeepLearning #NLP
Masato Hagiwara, Masato Mita : https://arxiv.org/abs/1911.12893
Code & Dataset https://github.com/mhagiwara/github-typo-corpus
#ArtificialIntelligence #DeepLearning #NLP
GitHub
GitHub - mhagiwara/github-typo-corpus: GitHub Typo Corpus: A Large-Scale Multilingual Dataset of Misspellings and Grammatical Errors
GitHub Typo Corpus: A Large-Scale Multilingual Dataset of Misspellings and Grammatical Errors - mhagiwara/github-typo-corpus
Major trends in #NLP : a review of 20 years of #ACL research
The 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019) is starting this week in Florence, Italy. We took the opportunity to review major research trends in the animated NLP space and formulate some implications from the business perspective. The article is backed by a statistical and — guess what — NLP-based analysis of ACL papers from the last 20 years
https://towardsdatascience.com/major-trends-in-nlp-a-review-of-20-years-of-acl-research-56f5520d473
The 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019) is starting this week in Florence, Italy. We took the opportunity to review major research trends in the animated NLP space and formulate some implications from the business perspective. The article is backed by a statistical and — guess what — NLP-based analysis of ACL papers from the last 20 years
https://towardsdatascience.com/major-trends-in-nlp-a-review-of-20-years-of-acl-research-56f5520d473
Medium
Major trends in NLP: a review of 20 years of ACL research
The 57th Annual Meeting of the Association for Computational Linguistics (ACL 2019) is starting this week in Florence, Italy. We took the…
Machine Learning on Graphs #NeurIPS2019
Michael Galkin : https://medium.com/@mgalkin/machine-learning-on-graphs-neurips-2019-875eecd41069
#GraphNeuralNetworks #NLP
Michael Galkin : https://medium.com/@mgalkin/machine-learning-on-graphs-neurips-2019-875eecd41069
#GraphNeuralNetworks #NLP
Medium
Machine Learning on Graphs @ NeurIPS 2019
If you still had any doubts — it’s time to admit. Machine Learning on Graphs becomes a first-class citizen at AI conferences while being…
SpeechBrain
A PyTorch-based Speech Toolkit
Video, by Mirco Ravanelli : https://youtube.com/watch?v=XETiKbN9ojE
: https://speechbrain.github.io
#speechbrain #NLP #DeepLearning
https://t.me/ArtificialIntelligenceArticles
A PyTorch-based Speech Toolkit
Video, by Mirco Ravanelli : https://youtube.com/watch?v=XETiKbN9ojE
: https://speechbrain.github.io
#speechbrain #NLP #DeepLearning
https://t.me/ArtificialIntelligenceArticles
YouTube
The SpeechBrain Project
SpeechBrain is an open-source and all-in-one speech toolkit relying on PyTorch.
The goal is to create a single, flexible, and user-friendly toolkit that can be used to easily develop state-of-the-art speech technologies
The goal is to create a single, flexible, and user-friendly toolkit that can be used to easily develop state-of-the-art speech technologies
exBERT- A Visual Analysis Tool to Explore Learned Representations in Transformers Models
Benjamin Hoover, Hendrik Strobelt, Sebastian Gehrmann : http://exbert.net
#NLP #BERT #LanguageModel
Benjamin Hoover, Hendrik Strobelt, Sebastian Gehrmann : http://exbert.net
#NLP #BERT #LanguageModel
Linguistic Knowledge and Transferability of Contextual Representations
Liu et al.: https://arxiv.org/abs/1903.08855
#ArtificialIntelligence #DeepLearning #NLP
Liu et al.: https://arxiv.org/abs/1903.08855
#ArtificialIntelligence #DeepLearning #NLP
arXiv.org
Linguistic Knowledge and Transferability of Contextual Representations
Contextual word representations derived from large-scale neural language models are successful across a diverse set of NLP tasks, suggesting that they encode useful and transferable features of...
JASS: Japanese-specific Sequence to Sequence Pre-training for Neural Machine Translation
Mao et al.: https://arxiv.org/abs/2005.03361
#ArtificialIntelligence #DeepLearning #NLP
Mao et al.: https://arxiv.org/abs/2005.03361
#ArtificialIntelligence #DeepLearning #NLP