رابط کاربری مثل جوکه، اگه مجبور هستی توضیحش بدی یعنی آنقدرها هم خوب نیست!
@ai_python
@ai_python
Here are 10 common AI terms explained in an easily understandable way.
1. Classification
A Machine Learning task which seeks to classify data points into different groups (called targets or class labels) that are pre-determined by the training data. For example, if we have a medical dataset consisting of biological measurements...... (heart rate, body temperature, age, height, weight, etc.) and whether or not a person has a specific disease, we could train a classification model to predict whether or not a person has the disease given just the biological measurements.
2. Regression
A supervised learning task that tries to predict a numerical result given a data point. For example, giving the description of a house (location, number of rooms, energy label) and predicting the market price of the house.
3. Underfitting
A phenomenon in which a Machine Learning algorithm is not fitted well enough to the training data, resulting in low performance on both the training data and similar but distinct data.3. Underfitting cont. A common example of underfitting occurs when a neural network is not trained long enough or when there is not enough training data. The converse phenomenon is overfitting.
4. Overfitting
A phenomenon in which a Machine Learning algorithm is too fitted to the training data, making performance on the training data very high, but performance on similar but distinct data low due to poor generalizability.A common example of overfitting occurs when a neural network is trained for too long. The converse phenomenon is underfitting.
5. Cost function
This is what Machine Learning algorithms are trying to minimize to achieve the best performance. It is simply the error the algorithm makes over a given dataset. It is also sometimes referred to as “loss function.”
6. Loss function
A (generally continuous) value that is a computation-friendly proxy for the performance metric. It measures the error between values predicted by the model and the true values we want the model to predict.During training, this value is minimized. “Loss function” is sometimes used interchangeably with “cost function,” although the two are differentiated in some contexts.
7. Validation data
A subset of data that a model is not trained on but is used during training to verify that the model performs well on distinct data. Validation data is used for hyper parameter tuning in order to avoid over fitting.
8. Neural Network
A specific type of Machine Learning algorithm which can be represented graphically as a network, inspired by the way that biological brains work.The network represents many simple mathematical operations (addition, multiplication, etc.) that are combined to produce a complex operation that may perform a complicated task (e.g. identifying cars in an image).
9. Parameter
Generally refers to the numbers in a neural network or Machine Learning algorithm that are changed to alter how the model behaves (sometimes also called weights). If a neural network is analogous to a radio, providing the base structure of a system, then parameters are analogous to the knobs on the radio, which are tuned to achieve a specific behavior (like tuning in to a specific frequency). Parameters are not set by the creator of the model, rather, the values are determined by the training process automatically.
10. Hyperparameter
A value that takes part in defining the overall structure of a model or behavior of an algorithm. Hyperparameters are not altered by the model training process and are set ahead of time before training. Many potential values for hyperparameters are generally tested to find those that optimize the training process. E.g, in a neural network, the number of layers is a hyperparameter (not altered by training), whereas the values within the layers (“weights”) themselves are parameters (altered by training).If the model is a radio, then a hyperparameter would be the number of knobs on the radio, while the values of these knobs would be parameters.
#آموزش
✳️ @AI_Python
1. Classification
A Machine Learning task which seeks to classify data points into different groups (called targets or class labels) that are pre-determined by the training data. For example, if we have a medical dataset consisting of biological measurements...... (heart rate, body temperature, age, height, weight, etc.) and whether or not a person has a specific disease, we could train a classification model to predict whether or not a person has the disease given just the biological measurements.
2. Regression
A supervised learning task that tries to predict a numerical result given a data point. For example, giving the description of a house (location, number of rooms, energy label) and predicting the market price of the house.
3. Underfitting
A phenomenon in which a Machine Learning algorithm is not fitted well enough to the training data, resulting in low performance on both the training data and similar but distinct data.3. Underfitting cont. A common example of underfitting occurs when a neural network is not trained long enough or when there is not enough training data. The converse phenomenon is overfitting.
4. Overfitting
A phenomenon in which a Machine Learning algorithm is too fitted to the training data, making performance on the training data very high, but performance on similar but distinct data low due to poor generalizability.A common example of overfitting occurs when a neural network is trained for too long. The converse phenomenon is underfitting.
5. Cost function
This is what Machine Learning algorithms are trying to minimize to achieve the best performance. It is simply the error the algorithm makes over a given dataset. It is also sometimes referred to as “loss function.”
6. Loss function
A (generally continuous) value that is a computation-friendly proxy for the performance metric. It measures the error between values predicted by the model and the true values we want the model to predict.During training, this value is minimized. “Loss function” is sometimes used interchangeably with “cost function,” although the two are differentiated in some contexts.
7. Validation data
A subset of data that a model is not trained on but is used during training to verify that the model performs well on distinct data. Validation data is used for hyper parameter tuning in order to avoid over fitting.
8. Neural Network
A specific type of Machine Learning algorithm which can be represented graphically as a network, inspired by the way that biological brains work.The network represents many simple mathematical operations (addition, multiplication, etc.) that are combined to produce a complex operation that may perform a complicated task (e.g. identifying cars in an image).
9. Parameter
Generally refers to the numbers in a neural network or Machine Learning algorithm that are changed to alter how the model behaves (sometimes also called weights). If a neural network is analogous to a radio, providing the base structure of a system, then parameters are analogous to the knobs on the radio, which are tuned to achieve a specific behavior (like tuning in to a specific frequency). Parameters are not set by the creator of the model, rather, the values are determined by the training process automatically.
10. Hyperparameter
A value that takes part in defining the overall structure of a model or behavior of an algorithm. Hyperparameters are not altered by the model training process and are set ahead of time before training. Many potential values for hyperparameters are generally tested to find those that optimize the training process. E.g, in a neural network, the number of layers is a hyperparameter (not altered by training), whereas the values within the layers (“weights”) themselves are parameters (altered by training).If the model is a radio, then a hyperparameter would be the number of knobs on the radio, while the values of these knobs would be parameters.
#آموزش
✳️ @AI_Python
Paint-with-Words, Implemented with Stable diffusion
https://github.com/cloneofsimo/paint-with-words-sd
https://github.com/cloneofsimo/paint-with-words-sd
GitHub
GitHub - cloneofsimo/paint-with-words-sd: Implementation of Paint-with-words with Stable Diffusion : method from eDiff-I that let…
Implementation of Paint-with-words with Stable Diffusion : method from eDiff-I that let you generate image from text-labeled segmentation map. - cloneofsimo/paint-with-words-sd
اصطلاح data-drift یا dataset drift که در فارسی به جابجایی داده یا رانش داده ترجمه میشود زمانی اتفاق میافتد که مجموعه داده مورد استفاده در آموزش مدل تفاوت زیادی با دادههایی که در زمان استقرار یا محیط عملیاتی ( اصطلاحا deploy یا production) مشاهده خواهد شد دارد و در نتیجه مدل شما نتایج نامطلوب و عجیب ایجاد کرده و عملکرد ضعیفی دربرخواهد داشت.
در مقالهای جدید، تیمی از محققان روش خاصی را برای برخورد با این مشکل در زمینه دادههای تصویری ارائه کردند:
"Data Models for Dataset Drift Controls in Machine Learning With Images"
Paper: https://arxiv.org/abs/2211.02578
Code: https://github.com/aiaudit-org/raw2logit
Dataset: https://paperswithcode.com/dataset/raw-microscopy-and-raw-drone
#MachineLearning #DeepLearning #ArtificialIntelligence #ML #DL #AI
#یادگیری_ماشین #مقاله
✳️ @AI_Python
در مقالهای جدید، تیمی از محققان روش خاصی را برای برخورد با این مشکل در زمینه دادههای تصویری ارائه کردند:
"Data Models for Dataset Drift Controls in Machine Learning With Images"
Paper: https://arxiv.org/abs/2211.02578
Code: https://github.com/aiaudit-org/raw2logit
Dataset: https://paperswithcode.com/dataset/raw-microscopy-and-raw-drone
#MachineLearning #DeepLearning #ArtificialIntelligence #ML #DL #AI
#یادگیری_ماشین #مقاله
✳️ @AI_Python
👉 Foundations Of Machine Learning
Using this repo, you can learn the foundations of ML through intuitive explanations, clean code and visuals. Also, you can learn how to apply ML to build a production grade product to deliver value.
🔗 https://github.com/GokuMohandas/Made-With-ML
#یادگیری_ماشین #منابع
✳️ @AI_Python
Using this repo, you can learn the foundations of ML through intuitive explanations, clean code and visuals. Also, you can learn how to apply ML to build a production grade product to deliver value.
🔗 https://github.com/GokuMohandas/Made-With-ML
#یادگیری_ماشین #منابع
✳️ @AI_Python
دوره کلاسی جدید دانشگاه کارنگی ملون
Advanced NLP - Carnegie Mellon 2022
https://m.youtube.com/playlist?list=PL8PYTP1V4I8D0UkqW2fEhgLrnlDW9QK7z
#پردازش_زبان_طبیعی #منابع #فیلم #کلاس_آموزشی
#NLP
✳️ @AI_Python
Advanced NLP - Carnegie Mellon 2022
https://m.youtube.com/playlist?list=PL8PYTP1V4I8D0UkqW2fEhgLrnlDW9QK7z
#پردازش_زبان_طبیعی #منابع #فیلم #کلاس_آموزشی
#NLP
✳️ @AI_Python
https://www.linkedin.com/posts/sunghoon-ivan-lee-96617347_available-position-activity-6996495877104091136-Gtp8?utm_source=share&utm_medium=member_android
https://www.linkedin.com/posts/mmfouda_several-funded-ms-and-phd-positions-are-activity-6997296593292996609-cJvq?utm_source=share&utm_medium=member_android
#اپلای
https://www.linkedin.com/posts/mmfouda_several-funded-ms-and-phd-positions-are-activity-6997296593292996609-cJvq?utm_source=share&utm_medium=member_android
#اپلای
Linkedin
Sunghoon Ivan Lee on LinkedIn: Available Position
Please share! I am hiring a PhD student in deep learning for wearables, particularly related to semi-supervised, weak supervised, self-supervised, or active…
🎞 DOVER: the Disentangled Objective Video Quality Evaluator
🖥 Github: https://github.com/teowu/dover
🗒 Paper: https://arxiv.org/abs/2211.04894v1
➡️ Dataset: https://paperswithcode.com/dataset/youtube-ugc
#مقاله
✳️ @AI_Python
🖥 Github: https://github.com/teowu/dover
🗒 Paper: https://arxiv.org/abs/2211.04894v1
➡️ Dataset: https://paperswithcode.com/dataset/youtube-ugc
#مقاله
✳️ @AI_Python
Forwarded from Meysam
اسم من میثمه،
در این کانال فقط چیزهایی که به نظر خودم جالب هستند رو پست میکنم.
هوش مصنوعی یکی از موضوعاتی هست که در موردش مینویسم.
دوست داشتید دنبال کنید دوست نداشتید میوت نکنید لفت بدید.
مرسی.
@ai_person
در این کانال فقط چیزهایی که به نظر خودم جالب هستند رو پست میکنم.
هوش مصنوعی یکی از موضوعاتی هست که در موردش مینویسم.
دوست داشتید دنبال کنید دوست نداشتید میوت نکنید لفت بدید.
مرسی.
@ai_person
https://www.linkedin.com/jobs/view/3346079750
تویتر یک موقعیت شغلی ریموت داده که اصلا خوراک ایرانیاس که تویتهایی رو شناسایی و ازش حذف کنن و افرادی ک نباید باشند رو محدود کنن😂
پ.ن: ببینم چکار میکنین 😂😜
پ.ن: تویتر موقعیتهای شغلی Remote زیادی رو اعلام کرده که میتونین براساس چیزی ک پژوهش کردید و سابقه دارین اقدام کنید
تویتر یک موقعیت شغلی ریموت داده که اصلا خوراک ایرانیاس که تویتهایی رو شناسایی و ازش حذف کنن و افرادی ک نباید باشند رو محدود کنن😂
پ.ن: ببینم چکار میکنین 😂😜
پ.ن: تویتر موقعیتهای شغلی Remote زیادی رو اعلام کرده که میتونین براساس چیزی ک پژوهش کردید و سابقه دارین اقدام کنید
آموزش رایگان یادگیری ماشین از دکتر شریفی زارچی استاد دانشگاه شریف
https://mktb.me/b1zp
#آموزش_کلاسی #منابع #یادگیری_ماشین #فیلم
✳️ @AI_Python
https://mktb.me/b1zp
#آموزش_کلاسی #منابع #یادگیری_ماشین #فیلم
✳️ @AI_Python
مکتبخونه
آموزش رایگان مقدمه یادگیری ماشین | مکتبخونه
آموزش رایگان مقدمهای بر یادگیری ماشین با تدریس دکتر شریفی زارچی از دانشگاه شریف به زبان فارسی
This week was dominated by molecular ML and drug discovery events:
- Broad Institute published a YouTube playlist of talks from the recent Machine Learning and Drug Discovery Symposium and leading drug discovery researchers.
- ELLIS Machine Learning for Drug Discovery Workshop will take place online in Zoom and GatherTown on Nov 28th, registration is free!
- Valence Discovery launched a blog platform for Drug Discovery related posts, the inaugural post by Clemens Isert talks about Quantum ML for drug-like molecules.
And a new work from GemNet authors: How robust are modern graph neural network potentials in long and hot molecular dynamics simulations? The work supports a recent line of works on ML force fields particularly focusing on hot dynamics where low test MAE does not necessarily correspond to good simulations. The most important robustness factor seems to be more training data!
- Broad Institute published a YouTube playlist of talks from the recent Machine Learning and Drug Discovery Symposium and leading drug discovery researchers.
- ELLIS Machine Learning for Drug Discovery Workshop will take place online in Zoom and GatherTown on Nov 28th, registration is free!
- Valence Discovery launched a blog platform for Drug Discovery related posts, the inaugural post by Clemens Isert talks about Quantum ML for drug-like molecules.
And a new work from GemNet authors: How robust are modern graph neural network potentials in long and hot molecular dynamics simulations? The work supports a recent line of works on ML force fields particularly focusing on hot dynamics where low test MAE does not necessarily correspond to good simulations. The most important robustness factor seems to be more training data!
YouTube
Broad Institute Machine Learning in Drug Discovery (MLinDD) Symposium 2022
This playlist contains the talks and opening and closing remarks from the Broad MLinDD symposium held on Oct 24, 2022 For more information visit broad.io/mldd
Molecular dynamics is one of the booming Geometric DL areas where equivariant models show the best qualities. The two cool recent papers on that topic:
⚛️ Forces are not Enough: Benchmark and Critical Evaluation for Machine Learning Force Fields with Molecular Simulations by Fu et al. introduces a new benchmark for molecular dynamics - in addition to MD17, the authors add datasets on modeling liquids (Water), peptides (Alanine dipeptide), and solid-state materials (LiPS). More importantly, apart from Energy as the main metric, the authors consider a wide range of physical properties like Stability, Diffusivity, and Radial Distribution Functions. Most SOTA molecular dynamics models were probed including SchNet, ForceNet, DimeNet, GemNet (-T and -dT), NequIP.
Density Functional Theory (DFT) calculations are one of the main workhorses of molecular dynamics (and account for a great deal of computing time in big clusters). DFT is O(n^3) to the input size though, so can ML help here? Learned Force Fields Are Ready For Ground State Catalyst Discovery by Schaarschmidt et al. present the experimental study of models of learned potentials - turns out GNNs can do a very good job in O(n) time!
🪐 For astrophysics aficionados: Mangrove: Learning Galaxy Properties from Merger Trees by Jespersen et al. apply GraphSAGE to merger trees of dark matter to predict a variety of galactic properties like stellar mass, cold gas mass, star formation rate, and even black hole mass. The paper is heavy on the terminology of astrophysics but pretty easy in terms of GNN parameterization and training. Mangrove works 4-9 orders of magnitude faster than standard models (that is, 10 000 - 1 000 000 000 times faster). Experimental charts are pieces of art that you can hang on a wall.
🤖 Compositional Semantic Parsing with Large Language Models by Drozdov, Schärli et al. pretty much solve the compositional semantic parsing task (natural language query - structured query like SPARQL) using only
#مقاله
⚛️ Forces are not Enough: Benchmark and Critical Evaluation for Machine Learning Force Fields with Molecular Simulations by Fu et al. introduces a new benchmark for molecular dynamics - in addition to MD17, the authors add datasets on modeling liquids (Water), peptides (Alanine dipeptide), and solid-state materials (LiPS). More importantly, apart from Energy as the main metric, the authors consider a wide range of physical properties like Stability, Diffusivity, and Radial Distribution Functions. Most SOTA molecular dynamics models were probed including SchNet, ForceNet, DimeNet, GemNet (-T and -dT), NequIP.
Density Functional Theory (DFT) calculations are one of the main workhorses of molecular dynamics (and account for a great deal of computing time in big clusters). DFT is O(n^3) to the input size though, so can ML help here? Learned Force Fields Are Ready For Ground State Catalyst Discovery by Schaarschmidt et al. present the experimental study of models of learned potentials - turns out GNNs can do a very good job in O(n) time!
🪐 For astrophysics aficionados: Mangrove: Learning Galaxy Properties from Merger Trees by Jespersen et al. apply GraphSAGE to merger trees of dark matter to predict a variety of galactic properties like stellar mass, cold gas mass, star formation rate, and even black hole mass. The paper is heavy on the terminology of astrophysics but pretty easy in terms of GNN parameterization and training. Mangrove works 4-9 orders of magnitude faster than standard models (that is, 10 000 - 1 000 000 000 times faster). Experimental charts are pieces of art that you can hang on a wall.
🤖 Compositional Semantic Parsing with Large Language Models by Drozdov, Schärli et al. pretty much solve the compositional semantic parsing task (natural language query - structured query like SPARQL) using only
code-davinci-002
language model from OpenAI (which is InstructGPT fine-tuned on code). No need for hefty tailored semantic parsing models - turns out a smart extension of the Chain-of-thought prompting (aka "let's think step by step") devised as Least-to-Most prompting (where we first answer easy subproblems before generating a full query) yields whopping 95% accuracy even on hardest Compositional Freebase Questions (CFQ) dataset. CFQ was introduced at ICLR 2020, and just after two years LMs cracked this task - looks like it's time for the new, even more complex dataset.#مقاله
Forwarded from Meysam
نگارش نسخه دوم کتاب:
Mastering Transformers
رو شروع کردیم.
دوستانی که نسخه قبلی رو خوندن و اگر موضوعی یا مطلبی به نظرتون جذاب میاد و فکر میکنید که باید اضافه بشه بفرمایید خوشحال میشم نظراتتون رو بدونم.
@ai_person
Mastering Transformers
رو شروع کردیم.
دوستانی که نسخه قبلی رو خوندن و اگر موضوعی یا مطلبی به نظرتون جذاب میاد و فکر میکنید که باید اضافه بشه بفرمایید خوشحال میشم نظراتتون رو بدونم.
@ai_person
SSL4EO-S12: A Large-Scale Multi-Modal, Multi-Temporal Dataset for Self-Supervised Learning in Earth Observation
🖥 Github: https://github.com/zhu-xlab/ssl4eo-s12
📝 Paper: https://arxiv.org/abs/2211.07044v1
🖥 Dataset: https://mediatum.ub.tum.de/1660427
#مقاله
✳️ @AI_Python
🖥 Github: https://github.com/zhu-xlab/ssl4eo-s12
📝 Paper: https://arxiv.org/abs/2211.07044v1
🖥 Dataset: https://mediatum.ub.tum.de/1660427
#مقاله
✳️ @AI_Python
GitHub
GitHub - zhu-xlab/SSL4EO-S12: SSL4EO-S12: a large-scale dataset for self-supervised learning in Earth observation
SSL4EO-S12: a large-scale dataset for self-supervised learning in Earth observation - zhu-xlab/SSL4EO-S12
سلام به همه دوستان یک سایتی با عنوان 1500 در حال انتشار در کانالهای تلگرامی و اینستاگرامی هست به هیچ عنوان روی این سایت کلیک نکنید و تاکید میشود که اصلا روی این سایت کلیک نکنید که عواقب جبران ناپذیر دارد.
Apply for Ph.D. positions starting this year!
We have room for several outstanding machine learning Ph.D. students to join our elite group at the University of Cambridge.
We are looking for:
leaders who want to realise their full potential and shape the future of machine learning
explorers who think boldly and prefer pushing boundaries to sitting still
builders who see projects through to completion and want to make a real-world impact
If you’re interested, please click the URL below. Note that we have already received a lot of applications. If you’re considering applying, earlier is definitely better!
Join the van der Schaar Lab:
https://eur03.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.vanderschaar-lab.com%2Fjoin&data=05%7C01%7Ctl522%40universityofcambridgecloud.onmicrosoft.com%7C25eb48cdd35849dde6d208dac1c40d86%7C49a50445bdfa4b79ade3547b4f3986e9%7C1%7C0%7C638035345048429672%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=koklpmvj4TDjLmx8OXWKFUyWVlKRA%2B%2FzuRstdljgIbI%3D&reserved=0
#اپلای
We have room for several outstanding machine learning Ph.D. students to join our elite group at the University of Cambridge.
We are looking for:
leaders who want to realise their full potential and shape the future of machine learning
explorers who think boldly and prefer pushing boundaries to sitting still
builders who see projects through to completion and want to make a real-world impact
If you’re interested, please click the URL below. Note that we have already received a lot of applications. If you’re considering applying, earlier is definitely better!
Join the van der Schaar Lab:
https://eur03.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.vanderschaar-lab.com%2Fjoin&data=05%7C01%7Ctl522%40universityofcambridgecloud.onmicrosoft.com%7C25eb48cdd35849dde6d208dac1c40d86%7C49a50445bdfa4b79ade3547b4f3986e9%7C1%7C0%7C638035345048429672%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=koklpmvj4TDjLmx8OXWKFUyWVlKRA%2B%2FzuRstdljgIbI%3D&reserved=0
#اپلای
van der Schaar Lab
Join the lab as a Ph.D student // van der Schaar Lab
We have room for four outstanding machine learning Ph.D. students to join our elite group at the University of Cambridge.
This media is not supported in your browser
VIEW IN TELEGRAM
مدل زبانی خیلی بزرگ Galactica که میتواند:
Can summarize academic literature, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.
To accelerate science, we open source all models including the 120 billion model with no friction. You can access them here.
galactica.org/paper.pdf
Explore and get weights: galactica.org
#مقاله
✳️ @AI_Python
Can summarize academic literature, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.
To accelerate science, we open source all models including the 120 billion model with no friction. You can access them here.
galactica.org/paper.pdf
Explore and get weights: galactica.org
#مقاله
✳️ @AI_Python