How to deliver on Machine Learning projects
A guide to the ML Engineering Loop.
By Emmanuel Ameisen and Adam Coates:
https://blog.insightdatascience.com/how-to-deliver-on-machine-learning-projects-c8d82ce642b0
#ArtificialIntelligence #BigData #DataScience #DeepLearning #MachineLearning
❇️ @AI_Python_EN
A guide to the ML Engineering Loop.
By Emmanuel Ameisen and Adam Coates:
https://blog.insightdatascience.com/how-to-deliver-on-machine-learning-projects-c8d82ce642b0
#ArtificialIntelligence #BigData #DataScience #DeepLearning #MachineLearning
❇️ @AI_Python_EN
Can We Learn the Language of Proteins?
#DataScience #MachineLearning #ArtificialIntelligence
http://bit.ly/33jCof6
❇️ @AI_Python_EN
#DataScience #MachineLearning #ArtificialIntelligence
http://bit.ly/33jCof6
❇️ @AI_Python_EN
Research Guide: Advanced Loss Functions for Machine Learning Models
http://bit.ly/36HBefu
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_Python_EN
http://bit.ly/36HBefu
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_Python_EN
Visualizing an AI model’s blind spots
http://bit.ly/2CosFZn
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_PYTHON_EN
http://bit.ly/2CosFZn
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_PYTHON_EN
Machine ignoring = underfitting
Machine learning = optimal fitting
Machine memorization = overfitting
#datascience #machinelearning
❇️ @AI_Python_EN
Machine learning = optimal fitting
Machine memorization = overfitting
#datascience #machinelearning
❇️ @AI_Python_EN
Grid search vs randomized search?
💡 What are the pros and cons of grid search? Pros: • Grid search is great when you need to fine-tune hyperparameters over a small search space automatically. • For example, if you have 100 different datasets that you expect to be similar (e.g. solving the same problem repeatedly with different populations), you can use grid search to automatically fine-tune the hyperparameters for each model. Cons: • Grid search is computationally expensive and inefficient, often searching over parameter space that has very little chance of being useful, resulting it being extremely slow. It's especially slow if you need to search a large space since it's complexity increases exponentially as more hyperparameters are optimized.
💡 What are the pros and cons of randomized search? Pros: • Randomized search does a good job finding near-optimal hyperparameters over a very large search space relatively quickly and doesn't suffer from the same exponential scaling problem as grid search. Cons: • Randomized search does not fine-tune the results as much as grid search does since it typically does not test every possible combination of parameters.
#datascience
👉 Free training -> http://bit.ly/dsdj-webinar
❇️ @AI_Python_EN
💡 What are the pros and cons of grid search? Pros: • Grid search is great when you need to fine-tune hyperparameters over a small search space automatically. • For example, if you have 100 different datasets that you expect to be similar (e.g. solving the same problem repeatedly with different populations), you can use grid search to automatically fine-tune the hyperparameters for each model. Cons: • Grid search is computationally expensive and inefficient, often searching over parameter space that has very little chance of being useful, resulting it being extremely slow. It's especially slow if you need to search a large space since it's complexity increases exponentially as more hyperparameters are optimized.
💡 What are the pros and cons of randomized search? Pros: • Randomized search does a good job finding near-optimal hyperparameters over a very large search space relatively quickly and doesn't suffer from the same exponential scaling problem as grid search. Cons: • Randomized search does not fine-tune the results as much as grid search does since it typically does not test every possible combination of parameters.
#datascience
👉 Free training -> http://bit.ly/dsdj-webinar
❇️ @AI_Python_EN
Machine Learning w.r.t meditation routine.
Machine before meditation = underfitting
Machine after meditation = optimal fitting
Planning of meditation = overfitting
#datascience
❇️ @AI_Python_EN
Machine before meditation = underfitting
Machine after meditation = optimal fitting
Planning of meditation = overfitting
#datascience
❇️ @AI_Python_EN
4 Traits, qualities that a data scientist must seek ...
1) Technical bar: Data science teams work everyday in SQL, specifically in Postgres, and expect candidates to know Python/some fluency in some sort of statistical language. Also, someone who is really comfortable with querying really large datasets.
2) Communication: we’re in roles where a lot of our day-to-day is spent getting great insights or building models and communicating results of that to stakeholders, whether that’s product managers, marketing folks or finance. It’s super key that data science candidates have good communication skills.
3) Grit, tenacity and willingness to solve hard problems: Things that DS teams solve are generally hard problems. My hope is that anyone who joins the data science team is excited about hard problems and bumping against hard challenges.
4) Passion for the arts and passion for the mission: This is not the most important but great to have.
#datascience
❇️ @AI_Python_EN
1) Technical bar: Data science teams work everyday in SQL, specifically in Postgres, and expect candidates to know Python/some fluency in some sort of statistical language. Also, someone who is really comfortable with querying really large datasets.
2) Communication: we’re in roles where a lot of our day-to-day is spent getting great insights or building models and communicating results of that to stakeholders, whether that’s product managers, marketing folks or finance. It’s super key that data science candidates have good communication skills.
3) Grit, tenacity and willingness to solve hard problems: Things that DS teams solve are generally hard problems. My hope is that anyone who joins the data science team is excited about hard problems and bumping against hard challenges.
4) Passion for the arts and passion for the mission: This is not the most important but great to have.
#datascience
❇️ @AI_Python_EN
Microsoft: Actor critic method bests greedy exploration in #reinforcementlearning
http://bit.ly/2sfxt17
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_Python_EN
http://bit.ly/2sfxt17
#DataScience #MachineLearning #ArtificialIntelligence
❇️ @AI_Python_EN
In #datascience, you must understand context. There are times at work where looking at the data alone didn't help me from solving the problem.
It doesn't matter if your domain is in marketing, healthcare, product, etc... You need to understand the context first before diving into the data. Without background information about how the data was generated, it becomes really difficult to make accurate assumptions on what your data will show.
Taking the time to understand the context will not only benefit you in your analysis, but you may even help your colleagues tackle the problem better.
When you are informed about the data and problem, you increase your value because now you're in a position to communicate and identify other potential problems.
So do this:
On your next project, take the time to not just do EDA, but also document your understanding of the context behind the data.
This good practice will definitely help you in your career and is a valuable skill you can bring to any team.
Context first, data second.
❇️ @AI_Python_EN
It doesn't matter if your domain is in marketing, healthcare, product, etc... You need to understand the context first before diving into the data. Without background information about how the data was generated, it becomes really difficult to make accurate assumptions on what your data will show.
Taking the time to understand the context will not only benefit you in your analysis, but you may even help your colleagues tackle the problem better.
When you are informed about the data and problem, you increase your value because now you're in a position to communicate and identify other potential problems.
So do this:
On your next project, take the time to not just do EDA, but also document your understanding of the context behind the data.
This good practice will definitely help you in your career and is a valuable skill you can bring to any team.
Context first, data second.
❇️ @AI_Python_EN
A good introduction to #MachineLearning and its 4 approaches:
https://towardsdatascience.com/machine-learning-an-introduction-23b84d51e6d0?gi=10a5fcd4decd
#BigData #DataScience #AI #Algorithms #ReinforcementLearning
❇️ @AI_Python_EN
https://towardsdatascience.com/machine-learning-an-introduction-23b84d51e6d0?gi=10a5fcd4decd
#BigData #DataScience #AI #Algorithms #ReinforcementLearning
❇️ @AI_Python_EN
Named Entity Recognition Benchmark: spaCy, Flair, m-BERT and camemBERT on anonymizing French commercial legal cases
http://bit.ly/2rq1I5H
#DataScience #MachineLearning #ArtificialIntelligence #NLP
❇️ @AI_Python_EN
http://bit.ly/2rq1I5H
#DataScience #MachineLearning #ArtificialIntelligence #NLP
❇️ @AI_Python_EN
Medium
NER algo benchmark: spaCy, Flair, m-BERT and camemBERT on anonymizing French commercial legal cases
Does (model) size matters?
XGBoost: An Intuitive Explanation
Ashutosh Nayak :
https://towardsdatascience.com/xgboost-an-intuitive-explanation-88eb32a48eff
#MachineLearning #DataScience
Ashutosh Nayak :
https://towardsdatascience.com/xgboost-an-intuitive-explanation-88eb32a48eff
#MachineLearning #DataScience
Decision trees are extremely fast when it comes to classify unknown records. Watch this video to know how Decision Tree algorithm works, in an easy way - http://bit.ly/2Ggsb9l
#DataScience #MachineLearning #AI #ML #ReinforcementLearning #Analytics #CloudComputing #Python #DeepLearning #BigData #Hadoop
#DataScience #MachineLearning #AI #ML #ReinforcementLearning #Analytics #CloudComputing #Python #DeepLearning #BigData #Hadoop
Checkout these new free resources in #DataScience👇
1. Introduction to PyTorch for Deep Learning: https://lnkd.in/f7kqZS2
2. Pandas for Data Analysis in Python: https://lnkd.in/fvRQHww
3. Support Vector Machine (SVM) in Python and R: https://lnkd.in/faJcSHe
4. Fundamentals of Regression Analysis: https://lnkd.in/fnEDP78
5. Getting started with Decision Trees: https://bit.ly/2PuZRFB
6. Introduction to Neural Networks: https://lnkd.in/fYUnsYQ
1. Introduction to PyTorch for Deep Learning: https://lnkd.in/f7kqZS2
2. Pandas for Data Analysis in Python: https://lnkd.in/fvRQHww
3. Support Vector Machine (SVM) in Python and R: https://lnkd.in/faJcSHe
4. Fundamentals of Regression Analysis: https://lnkd.in/fnEDP78
5. Getting started with Decision Trees: https://bit.ly/2PuZRFB
6. Introduction to Neural Networks: https://lnkd.in/fYUnsYQ
This media is not supported in your browser
VIEW IN TELEGRAM
10 Useful ML Practices For Python Developers
Pratik Bhavsar:
https://medium.com/modern-nlp/10-great-ml-practices-for-python-developers-b089eefc18fc
#Python #MachineLearning #ArtificialIntelligence #DataScience #Programming
❇️ @AI_Python_EN
Pratik Bhavsar:
https://medium.com/modern-nlp/10-great-ml-practices-for-python-developers-b089eefc18fc
#Python #MachineLearning #ArtificialIntelligence #DataScience #Programming
❇️ @AI_Python_EN
Breast cancer classification with Keras and Deep Learning
To analyze the cellular structures in the breast histology images we were instead leveraging basic computer vision and image processing algorithms, but combining them in a novel way.
Researcher: Adrian Rosebrock
Paper & codes : http://ow.ly/yngq30qjLye
#artificialintelligence #ai #machinelearning #deeplearning #bigdata #datascience
❇️ @AI_Python_EN
To analyze the cellular structures in the breast histology images we were instead leveraging basic computer vision and image processing algorithms, but combining them in a novel way.
Researcher: Adrian Rosebrock
Paper & codes : http://ow.ly/yngq30qjLye
#artificialintelligence #ai #machinelearning #deeplearning #bigdata #datascience
❇️ @AI_Python_EN
ANNOUNCING PYCARET 1.0.0 - An amazingly simple, fast and efficient way to do machine learning in Python. NEW OPEN SOURCE ML LIBRARY If you are a DATA SCIENTIST or want to become one, then this is for YOU....
PyCaret is a NEW open source machine learning library to train and deploy ML models in low-code environment.
It allows you to go from preparing data to deploying a model within SECONDS.
PyCaret is designed to reduce time and efforts spent in coding ML experiments. It automates the following:
- Preprocessing (Data Preparation, Feature Engineering and Feature Selection)
- Model Selection (over 60 ready-to-use algorithms)
- Model Evaluation (50+ analysis plots)
- Model Deployment
- ML Integration and Monitoring (Power BI, Tableau, Alteryx, KNIME and more)
- ..... and much more!
Watch this 1 minute video to see how PyCaret can help you in your next machine learning project.
The easiest way to install pycaret is using pip. Just type "pip install pycaret" into your notebook.
To learn more about PyCaret, please visit the official website https://www.pycaret.org
#datascience #datascientist #machinelearning #ml #ai #artificialintelligence #analytics #pycaret
❇️ @AI_Python_EN
PyCaret is a NEW open source machine learning library to train and deploy ML models in low-code environment.
It allows you to go from preparing data to deploying a model within SECONDS.
PyCaret is designed to reduce time and efforts spent in coding ML experiments. It automates the following:
- Preprocessing (Data Preparation, Feature Engineering and Feature Selection)
- Model Selection (over 60 ready-to-use algorithms)
- Model Evaluation (50+ analysis plots)
- Model Deployment
- ML Integration and Monitoring (Power BI, Tableau, Alteryx, KNIME and more)
- ..... and much more!
Watch this 1 minute video to see how PyCaret can help you in your next machine learning project.
The easiest way to install pycaret is using pip. Just type "pip install pycaret" into your notebook.
To learn more about PyCaret, please visit the official website https://www.pycaret.org
#datascience #datascientist #machinelearning #ml #ai #artificialintelligence #analytics #pycaret
❇️ @AI_Python_EN