New paper on much faster RL by learning graph representations of the world http://arxiv.org/abs/1907.00664 ! World graphs capture the structure of the world and can be used to focus exploration:
http://www-personal.umich.edu/~shangw/world_graph.html
✴️ @AI_Python_EN
http://www-personal.umich.edu/~shangw/world_graph.html
✴️ @AI_Python_EN
Training on only synthetic data, our new paper gets near the SOTA on the real-world 3DPW human pose dataset by focusing on motion. Simulation is underappreciated for representation learning; you just have to keep the nets from overfitting to the domain.
https://arxiv.org/abs/1907.02499
✴️ @AI_Python_EN
https://arxiv.org/abs/1907.02499
✴️ @AI_Python_EN
Current autopilot plane landings are not really autonomous: they rely on Instrument Landing Systems (ILS), only available at major airport runways. Computer Vision enables true autonomous landings: finding & landing on runway without ground instruments.
https://www.tum.de/nc/en/about-tum/news/press-releases/details/35556/
✴️ @AI_Python_EN
https://www.tum.de/nc/en/about-tum/news/press-releases/details/35556/
✴️ @AI_Python_EN
excited to announce the release of #Trafffic-Net, an image dataset for training real-time traffic analysis and accident detection #ArtificialIntelligence systems. https://github.com/OlafenwaMoses/Traffic-Net
✴️ @AI_Python_EN
✴️ @AI_Python_EN
#DeepLearning: What Is It Good For? - Prof. Ankit Patel Rice University [Video]
https://buff.ly/2U1irV9
#AI #MachineLearning
✴️ @AI_Python_EN
https://buff.ly/2U1irV9
#AI #MachineLearning
✴️ @AI_Python_EN
YouTube
Deep Learning: What is it good for? - Prof. Ankit Patel - Rice University
"In this talk, we will introduce deep learning and review some of the key advances in the field focusing on current attempts at a theoretical understanding. ...
Psychometricians make use of a very wide range of statistical models, for example:
Regression analysis
Path analysis
Exploratory factor analysis
Confirmatory factor analysis
Item response theory modeling
Structural equation modeling
Growth modeling
Discrete-time survival analysis
Continuous-time survival analysis
Time series analysis
Observed outcome variables may be continuous, censored, binary, ordered categorical (ordinal), unordered categorical (nominal), counts, or combinations of these variable types.
Some variations of these basic models are:
Regression mixture modeling
Path analysis mixture modeling
Latent class analysis
Latent class analysis with covariates and direct effects
Confirmatory latent class analysis
Latent class analysis with multiple categorical latent variables
Loglinear modeling
Multiple group analysis
Multilevel modeling
Finite mixture modeling
Complier Average Causal Effect (CACE) modeling
Latent transition analysis and hidden Markov modeling including mixtures and covariates
Latent class growth analysis
Discrete-time survival mixture analysis
Continuous-time survival mixture analysis
Over the years, I've used many of these models in my marketing-related work. They all have practical applications in many fields - it ain't just academic stuff.
✴️ @AI_Python_EN
Regression analysis
Path analysis
Exploratory factor analysis
Confirmatory factor analysis
Item response theory modeling
Structural equation modeling
Growth modeling
Discrete-time survival analysis
Continuous-time survival analysis
Time series analysis
Observed outcome variables may be continuous, censored, binary, ordered categorical (ordinal), unordered categorical (nominal), counts, or combinations of these variable types.
Some variations of these basic models are:
Regression mixture modeling
Path analysis mixture modeling
Latent class analysis
Latent class analysis with covariates and direct effects
Confirmatory latent class analysis
Latent class analysis with multiple categorical latent variables
Loglinear modeling
Multiple group analysis
Multilevel modeling
Finite mixture modeling
Complier Average Causal Effect (CACE) modeling
Latent transition analysis and hidden Markov modeling including mixtures and covariates
Latent class growth analysis
Discrete-time survival mixture analysis
Continuous-time survival mixture analysis
Over the years, I've used many of these models in my marketing-related work. They all have practical applications in many fields - it ain't just academic stuff.
✴️ @AI_Python_EN
A PyTorch implementation of "SimGNN: A Neural Network Approach to Fast Graph Similarity Computation" (WSDM 2019). This is a lightweight graph convolutional neural network for the fast calculation of approximate graph similarity at scale. Graph similarity search is among the most important graph-based applications, e.g. finding the chemical compounds that are most similar to a query compound.
https://lnkd.in/gA5tfuC
#datamining #machinelearning #deeplearning #datascience #bigdata
✴️ @AI_Python_EN
https://lnkd.in/gA5tfuC
#datamining #machinelearning #deeplearning #datascience #bigdata
✴️ @AI_Python_EN
Great Statistical software for Beginners.
Here is the Gretl Tutorial by Simone Gasperin
1)Simple Linear Regression
https://lnkd.in/ecfsV9c
2)Coding Dummy Variables
https://lnkd.in/ef7Yd7f
3)Forecasting New Observations
https://lnkd.in/eNKbxbU
4)Forecasting a Large Number of Observations
https://lnkd.in/eHmibGs
5)Logistic Regression
https://lnkd.in/eRfhQ87
6)Forecasting and Confusion Matrix
https://lnkd.in/eaqrFJr
7)Modeling and Forecasting Time Series Data
https://lnkd.in/e6fqKpF
8)Comparing Time Series Trend Models
https://lnkd.in/eKjEUAE
#datascience #machinelearning #statistics #dataanalytics #dataanalysis
✴️ @AI_Python_EN
Here is the Gretl Tutorial by Simone Gasperin
1)Simple Linear Regression
https://lnkd.in/ecfsV9c
2)Coding Dummy Variables
https://lnkd.in/ef7Yd7f
3)Forecasting New Observations
https://lnkd.in/eNKbxbU
4)Forecasting a Large Number of Observations
https://lnkd.in/eHmibGs
5)Logistic Regression
https://lnkd.in/eRfhQ87
6)Forecasting and Confusion Matrix
https://lnkd.in/eaqrFJr
7)Modeling and Forecasting Time Series Data
https://lnkd.in/e6fqKpF
8)Comparing Time Series Trend Models
https://lnkd.in/eKjEUAE
#datascience #machinelearning #statistics #dataanalytics #dataanalysis
✴️ @AI_Python_EN
What metrics to focus on, in confusionmatrix?
Ans-> It depends on the problem statement and data one is dealing with!
#examples
1) Spam Filter -> Consider +ve class as 'spam'. Optimize for Precision/Specificity, the reason for the same is...
False Negatives(spam emails in the primary) are more acceptable than False Positives(primary emails in Spam).
2) Fraud Transactions -> Consider +ve class as 'Fraud'. Optimize for Sensitivity, the reason for the same is...
False Positives(detecting fraud, but they are not) are more acceptable than False Negatives(detecting not as a fraud, but actually they are).
Will discuss more of Classification problem in our group
#datascience #machinelearning
✴️ @AI_Python_EN
Ans-> It depends on the problem statement and data one is dealing with!
#examples
1) Spam Filter -> Consider +ve class as 'spam'. Optimize for Precision/Specificity, the reason for the same is...
False Negatives(spam emails in the primary) are more acceptable than False Positives(primary emails in Spam).
2) Fraud Transactions -> Consider +ve class as 'Fraud'. Optimize for Sensitivity, the reason for the same is...
False Positives(detecting fraud, but they are not) are more acceptable than False Negatives(detecting not as a fraud, but actually they are).
Will discuss more of Classification problem in our group
#datascience #machinelearning
✴️ @AI_Python_EN
AI, Python, Cognitive Neuroscience
https://youtu.be/9M18rc9-VWU https://github.com/yuanming-hu/taichi_mpm ✴️ @AI_Python_EN
Implementation of Variational Auto-Encoder (VAE) and Deep Feature Consistent VAE for facial attribute manipulation using Keras and Tensorflow-dataset module.
https://github.com/iamsoroush/face_vae
✴️ @AI_Python_EN
https://github.com/iamsoroush/face_vae
✴️ @AI_Python_EN
Forwarded from AI, Python, Cognitive Neuroscience (Farzad)
A Full Hardware Guide to Deep Learning
🌎 http://bit.ly/2CBNi3W
#DeepLearning #MachineLearning #AI #DataScience
✴️ @AI_Python_EN
🌎 http://bit.ly/2CBNi3W
#DeepLearning #MachineLearning #AI #DataScience
✴️ @AI_Python_EN
Is this AI developing a sense of time?
https://www.zdnet.com/article/is-this-ai-developing-a-sense-of-time/
✴️ @AI_Python_EN
https://www.zdnet.com/article/is-this-ai-developing-a-sense-of-time/
✴️ @AI_Python_EN
https://arxiv.org/abs/1907.02544 show that GANs can be harnessed for unsupervised representation learning, with state-of-the-art results on ImageNet. Reconstructions, as shown below, tend to emphasise high-level semantics over pixel-level details.
These results showcase the potential of GANs and other generative models in unsupervised learning, as explained in our recent blog post:
https://deepmind.com/blog/unsupervised-learning/
✴️ @AI_Python_EN
These results showcase the potential of GANs and other generative models in unsupervised learning, as explained in our recent blog post:
https://deepmind.com/blog/unsupervised-learning/
✴️ @AI_Python_EN
BigBiGAN shows that "progress in image generation quality translates to substantially improved representation learning performance." Competitive w/self-supervised approaches on ImageNet. The cycle from generative models to other methods and back again continues.
Large Scale Adversarial Representation Learning. Jeff Donahue and Karen Simonyan http://arxiv.org/abs/1907.02544
✴️ @AI_Python_EN
Large Scale Adversarial Representation Learning. Jeff Donahue and Karen Simonyan http://arxiv.org/abs/1907.02544
✴️ @AI_Python_EN
arXiv.org
Large Scale Adversarial Representation Learning
Adversarially trained generative models (GANs) have recently achieved compelling image synthesis results. But despite early successes in using GANs for unsupervised representation learning, they...
Excited to share newest http://fast.ai
course: A Code-First Introduction to Natural Language Processing All code & videos are available for free online, please check it out!
✴️ @AI_Python_EN
course: A Code-First Introduction to Natural Language Processing All code & videos are available for free online, please check it out!
✴️ @AI_Python_EN
Ladies, if he:
- requires lots of supervision
- yet always wants more power
- can't explain decisions
- optimizes for the average outcome
- dismisses problems as edge cases
- forgets things catastrophically
He's not your man, he's a deep neural network. #AIFun
✴️ @AI_Python_EN
- requires lots of supervision
- yet always wants more power
- can't explain decisions
- optimizes for the average outcome
- dismisses problems as edge cases
- forgets things catastrophically
He's not your man, he's a deep neural network. #AIFun
✴️ @AI_Python_EN
Interesting work from Ross Wightman comparing something like EfficientNet / ResNet which uses only Imagenet data to the Facebook-IG ResNext that was trained on a lot of instagram public data. While their validation scores are close, the test scores seem to diverge more.
FacebookAI ResNeXt models pre-trained on Instagram hashtags stand out in their ability to generalized to the 'ImageNetV2' test set.
#PyTorch
https://colab.research.google.com/github/rwightman/pytorch-image-models/blob/master/notebooks/GeneralizationToImageNetV2.ipynb
✴️ @AI_Python_EN
FacebookAI ResNeXt models pre-trained on Instagram hashtags stand out in their ability to generalized to the 'ImageNetV2' test set.
#PyTorch
https://colab.research.google.com/github/rwightman/pytorch-image-models/blob/master/notebooks/GeneralizationToImageNetV2.ipynb
✴️ @AI_Python_EN
Towards Declarative Visual Reasoning… Or Not? https://buff.ly/2UkT3u4
#NeuralNetworks #DeepLearning #AI
#NeuralNetworks #DeepLearning #AI
SingularityNET
Towards Declarative Visual Reasoning. . . Or Not?
Compositional Visual Reasoning
Anyone working with data should be concerned about missing data...or they shouldn't be working with data.
One way to address missing data is multiple imputation. This is a complex topic, and many textbooks and articles have been published about it. Four I can recommend are:
- Applied Missing Data Analysis (Enders)
- Handbook of Missing Data Methodology (Molenberghs et al.)
- Flexible Imputation of Missing Data (van Buuren)
- Outlier Analysis (Aggarwal)
The Enders book is probably the most basic and readable of the ones I've listed.
The Stata 16 reference manual on multiple imputation, linked below, also provides a good overview of this subject, and includes many examples and a glossary of key terms.
✴️ @AI_Python_EN
One way to address missing data is multiple imputation. This is a complex topic, and many textbooks and articles have been published about it. Four I can recommend are:
- Applied Missing Data Analysis (Enders)
- Handbook of Missing Data Methodology (Molenberghs et al.)
- Flexible Imputation of Missing Data (van Buuren)
- Outlier Analysis (Aggarwal)
The Enders book is probably the most basic and readable of the ones I've listed.
The Stata 16 reference manual on multiple imputation, linked below, also provides a good overview of this subject, and includes many examples and a glossary of key terms.
✴️ @AI_Python_EN