@Machine_learn
More than 200 NLP datasets - this is gold (last update 21.01.202)
https://quantumstat.com/dataset/dataset.html
and also Google provided dataset search tool for publicly available datasets:
https://datasetsearch.research.google.com/
More than 200 NLP datasets - this is gold (last update 21.01.202)
https://quantumstat.com/dataset/dataset.html
and also Google provided dataset search tool for publicly available datasets:
https://datasetsearch.research.google.com/
سلام دوستان برای یه کار تحقیق نیاز به یسری دیتاست در زمینه تحلیل احساس فارسی داریم (به غیر از توییتر) ممنون میشم اگر کسی داره در پیوی برای بنده به اشتراک بزاره
@raminmousa
@raminmousa
Machine learning books and papers pinned «سلام دوستان برای یه کار تحقیق نیاز به یسری دیتاست در زمینه تحلیل احساس فارسی داریم (به غیر از توییتر) ممنون میشم اگر کسی داره در پیوی برای بنده به اشتراک بزاره @raminmousa»
@Machine_learn
MARKOV CHAIN MONTE CARLO (MCMC) SAMPLING
https://www.tweag.io/posts/2019-10-25-mcmc-intro1.html
Habr ru: https://habr.com/ru/company/piter/blog/491268/
MARKOV CHAIN MONTE CARLO (MCMC) SAMPLING
https://www.tweag.io/posts/2019-10-25-mcmc-intro1.html
Habr ru: https://habr.com/ru/company/piter/blog/491268/
@Machine_learn
#XGBoost
XGBoost: An Intuitive Explanation
Ashutosh Nayak :
https://towardsdatascience.com/xgboost-an-intuitive-explanation-88eb32a48eff
#XGBoost
XGBoost: An Intuitive Explanation
Ashutosh Nayak :
https://towardsdatascience.com/xgboost-an-intuitive-explanation-88eb32a48eff
Announcing TensorFlow Quantum: An Open Source Library for Quantum Machine Learning
@Machine_learn
https://ai.googleblog.com/2020/03/announcing-tensorflow-quantum-open.html
@Machine_learn
https://ai.googleblog.com/2020/03/announcing-tensorflow-quantum-open.html
Fresh picks from ArXiv
This week is accepted papers to CVPR and WebConf, submissions to ICML, 130-page survey on knowledge graphs and algorithms for rainbow vertex coloring 🌈
@Machine_learn
CVPR 20
* Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
* Bundle Adjustment on a Graph Processor
@Machine_learn
WebConf 20
* Just SLaQ When You Approximate: Accurate Spectral Distances for Web-Scale Graphs
* Heterogeneous Graph Transformer
* Learning to Hash with Graph Neural Networks for Recommender Systems
@Machine_learn
ICML 20
* Neural Enhanced Belief Propagation on Factor Graphs by group of Max Welling
@Machine_learn
Survey
* A Survey on The Expressive Power of Graph Neural Networks
@Machine_learn
by Ryoma Sato
* A Survey on Deep Hashing Methods
* Knowledge Graphs
* Knowledge Graphs and Knowledge Networks: The Story in Brief
@Machine_learn
Graph Theory
* Properties of Erdős-Rényi Graphs
* Algorithms for the rainbow vertex coloring problem on graph classes
* Direct Product Primality Testing of Graphs is GI-hard
This week is accepted papers to CVPR and WebConf, submissions to ICML, 130-page survey on knowledge graphs and algorithms for rainbow vertex coloring 🌈
@Machine_learn
CVPR 20
* Point-GNN: Graph Neural Network for 3D Object Detection in a Point Cloud
* Bundle Adjustment on a Graph Processor
@Machine_learn
WebConf 20
* Just SLaQ When You Approximate: Accurate Spectral Distances for Web-Scale Graphs
* Heterogeneous Graph Transformer
* Learning to Hash with Graph Neural Networks for Recommender Systems
@Machine_learn
ICML 20
* Neural Enhanced Belief Propagation on Factor Graphs by group of Max Welling
@Machine_learn
Survey
* A Survey on The Expressive Power of Graph Neural Networks
@Machine_learn
by Ryoma Sato
* A Survey on Deep Hashing Methods
* Knowledge Graphs
* Knowledge Graphs and Knowledge Networks: The Story in Brief
@Machine_learn
Graph Theory
* Properties of Erdős-Rényi Graphs
* Algorithms for the rainbow vertex coloring problem on graph classes
* Direct Product Primality Testing of Graphs is GI-hard
arXiv.org
Knowledge Graphs
In this paper we provide a comprehensive introduction to knowledge graphs, which have recently garnered significant attention from both industry and academia in scenarios that require exploiting...
1.Generative Adversarial Networks with python by Jason Brownlee
2.imbalanced classification with python by Jason Brownlee
I want these two books
@Raminmousa
2.imbalanced classification with python by Jason Brownlee
I want these two books
@Raminmousa
@machine_learn
A Survey on The Expressive Power of Graph Neural Networks
This is the best survey on the theory on GNNs I'm aware of. It produces so many illustrative examples on what GNN can and cannot distinguish.
It's funny, it's made by Ryoma Sato who I already saw from other works on GNNs and I thought it's one of these old Japanese professors with long beard and strict habits, but it turned out to be a 1st year MSc student 🇯🇵
A Survey on The Expressive Power of Graph Neural Networks
This is the best survey on the theory on GNNs I'm aware of. It produces so many illustrative examples on what GNN can and cannot distinguish.
It's funny, it's made by Ryoma Sato who I already saw from other works on GNNs and I thought it's one of these old Japanese professors with long beard and strict habits, but it turned out to be a 1st year MSc student 🇯🇵
❤1
"Deep learning for Computer Vision by Jason brownlee"
Please share it with me
@raminmousa
https://machinelearningmastery.com/deep-learning-for-computer-vision/
Please share it with me
@raminmousa
https://machinelearningmastery.com/deep-learning-for-computer-vision/
@Machine_learn
MaxUp: A Simple Way to Improve Generalization of Neural Network Training
A new approach to augmentation both images and text. The idea is to generate a set of augmented data with some random perturbations or transforms and minimize the maximum, or worst case loss over the augmented data. By doing so, the authors implicitly introduce a smoothness or robustness regularization against the random perturbations, and hence improve the generation performance. Testing MaxUp on a range of tasks, including image classification, language modeling, and adversarial certification, it is consistently outperforming the existing best baseline methods, without introducing substantial computational overhead.
.
.
.
paper: https://arxiv.org/abs/2002.09024
#augmentations #SOTA #ml
@Machine_learn
MaxUp: A Simple Way to Improve Generalization of Neural Network Training
A new approach to augmentation both images and text. The idea is to generate a set of augmented data with some random perturbations or transforms and minimize the maximum, or worst case loss over the augmented data. By doing so, the authors implicitly introduce a smoothness or robustness regularization against the random perturbations, and hence improve the generation performance. Testing MaxUp on a range of tasks, including image classification, language modeling, and adversarial certification, it is consistently outperforming the existing best baseline methods, without introducing substantial computational overhead.
.
.
.
paper: https://arxiv.org/abs/2002.09024
#augmentations #SOTA #ml
@Machine_learn
Paper:
https://arxiv.org/pdf/2003.05534.pdf
Github:
https://github.com/sniklaus/softmax-splatting
Short Summary:
https://www.marktechpost.com/2020/03/14/softmax-splatting-for-video-frame-interpolation/
Paper:
https://arxiv.org/pdf/2003.05534.pdf
Github:
https://github.com/sniklaus/softmax-splatting
Short Summary:
https://www.marktechpost.com/2020/03/14/softmax-splatting-for-video-frame-interpolation/
GitHub
GitHub - sniklaus/softmax-splatting: an implementation of softmax splatting for differentiable forward warping using PyTorch
an implementation of softmax splatting for differentiable forward warping using PyTorch - sniklaus/softmax-splatting
@Machine_learn
Grid Search Optimization Algorithm in Python
https://stackabuse.com/grid-search-optimization-algorithm-in-python/
Grid Search Optimization Algorithm in Python
https://stackabuse.com/grid-search-optimization-algorithm-in-python/
Stack Abuse
Grid Search Optimization Algorithm in Python
The article explains how to use the grid search optimization algorithm in Python for tuning hyper-parameters for deep learning algorithms.