DLeX: AI Python
22.9K subscribers
4.96K photos
1.22K videos
765 files
4.3K links
هوش‌مصنوعی و برنامه‌نویسی

ارتباط با نوید داریا در توییتر :
https://twitter.com/NaviDDariya

اراتباط با لی لی علوی در تلگرام :
@lilylawww
Download Telegram
رابط کاربری مثل جوکه، اگه مجبور هستی توضیحش بدی یعنی آنقدرها هم خوب نیست!
@ai_python
Here are 10 common AI terms explained in an easily understandable way.

1. Classification
A Machine Learning task which seeks to classify data points into different groups (called targets or class labels) that are pre-determined by the training data. For example, if we have a medical dataset consisting of biological measurements...... (heart rate, body temperature, age, height, weight, etc.) and whether or not a person has a specific disease, we could train a classification model to predict whether or not a person has the disease given just the biological measurements.

2. Regression
A supervised learning task that tries to predict a numerical result given a data point. For example, giving the description of a house (location, number of rooms, energy label) and predicting the market price of the house.

3. Underfitting
A phenomenon in which a Machine Learning algorithm is not fitted well enough to the training data, resulting in low performance on both the training data and similar but distinct data.3. Underfitting cont. A common example of underfitting occurs when a neural network is not trained long enough or when there is not enough training data. The converse phenomenon is overfitting.

4. Overfitting
A phenomenon in which a Machine Learning algorithm is too fitted to the training data, making performance on the training data very high, but performance on similar but distinct data low due to poor generalizability.A common example of overfitting occurs when a neural network is trained for too long. The converse phenomenon is underfitting.

5. Cost function
This is what Machine Learning algorithms are trying to minimize to achieve the best performance. It is simply the error the algorithm makes over a given dataset. It is also sometimes referred to as “loss function.”

6. Loss function
A (generally continuous) value that is a computation-friendly proxy for the performance metric. It measures the error between values predicted by the model and the true values we want the model to predict.During training, this value is minimized. “Loss function” is sometimes used interchangeably with “cost function,” although the two are differentiated in some contexts.

7. Validation data
A subset of data that a model is not trained on but is used during training to verify that the model performs well on distinct data. Validation data is used for hyper parameter tuning in order to avoid over fitting.

8. Neural Network
A specific type of Machine Learning algorithm which can be represented graphically as a network, inspired by the way that biological brains work.The network represents many simple mathematical operations (addition, multiplication, etc.) that are combined to produce a complex operation that may perform a complicated task (e.g. identifying cars in an image).

9. Parameter
Generally refers to the numbers in a neural network or Machine Learning algorithm that are changed to alter how the model behaves (sometimes also called weights). If a neural network is analogous to a radio, providing the base structure of a system, then parameters are analogous to the knobs on the radio, which are tuned to achieve a specific behavior (like tuning in to a specific frequency). Parameters are not set by the creator of the model, rather, the values are determined by the training process automatically.

10. Hyperparameter
A value that takes part in defining the overall structure of a model or behavior of an algorithm. Hyperparameters are not altered by the model training process and are set ahead of time before training. Many potential values for hyperparameters are generally tested to find those that optimize the training process. E.g, in a neural network, the number of layers is a hyperparameter (not altered by training), whereas the values within the layers (“weights”) themselves are parameters (altered by training).If the model is a radio, then a hyperparameter would be the number of knobs on the radio, while the values of these knobs would be parameters.

#آموزش

✳️ @AI_Python
اصطلاح data-drift یا dataset drift که در فارسی به جابجایی داده یا رانش داده ترجمه می‌شود زمانی اتفاق می‌افتد که مجموعه داده مورد استفاده در آموزش مدل تفاوت زیادی با داده‌هایی که در زمان استقرار یا محیط عملیاتی ( اصطلاحا deploy یا production) مشاهده خواهد شد دارد و در نتیجه مدل شما نتایج نامطلوب و عجیب ایجاد کرده و عملکرد ضعیفی دربرخواهد داشت.

در مقاله‌ای جدید، تیمی از محققان روش خاصی را برای برخورد با این مشکل در زمینه داده‌های تصویری ارائه کردند:

"Data Models for Dataset Drift Controls in Machine Learning With Images"

Paper: https://arxiv.org/abs/2211.02578
Code: https://github.com/aiaudit-org/raw2logit
Dataset: https://paperswithcode.com/dataset/raw-microscopy-and-raw-drone

#MachineLearning #DeepLearning #ArtificialIntelligence #ML #DL #AI


#یادگیری_ماشین #مقاله

✳️ @AI_Python
👉 Foundations Of Machine Learning

Using this repo, you can learn the foundations of ML through intuitive explanations, clean code and visuals. Also, you can learn how to apply ML to build a production grade product to deliver value.

🔗 https://github.com/GokuMohandas/Made-With-ML



#یادگیری_ماشین #منابع

✳️ @AI_Python
دوره کلاسی جدید دانشگاه کارنگی ملون
Advanced NLP - Carnegie Mellon 2022

https://m.youtube.com/playlist?list=PL8PYTP1V4I8D0UkqW2fEhgLrnlDW9QK7z

#پردازش_زبان_طبیعی #منابع #فیلم #کلاس_آموزشی
#NLP


✳️ @AI_Python
🎞 DOVER: the Disentangled Objective Video Quality Evaluator

🖥 Github: https://github.com/teowu/dover

🗒 Paper: https://arxiv.org/abs/2211.04894v1

➡️ Dataset: https://paperswithcode.com/dataset/youtube-ugc

#مقاله

✳️ @AI_Python
Forwarded from Meysam
اسم من میثمه،
در این کانال فقط چیزهایی که به نظر خودم جالب هستند رو پست میکنم.
هوش مصنوعی یکی از موضوعاتی هست که در موردش می‌نویسم.
دوست داشتید دنبال کنید دوست نداشتید میوت نکنید لفت بدید.
مرسی.

@ai_person
https://www.linkedin.com/jobs/view/3346079750

تویتر یک موقعیت شغلی ریموت داده که اصلا خوراک ایرانیاس که تویتهایی رو شناسایی و ازش حذف کنن و افرادی ک نباید باشند رو محدود کنن😂

پ.ن: ببینم چکار میکنین 😂😜
پ.ن: تویتر موقعیتهای شغلی Remote زیادی رو اعلام کرده که میتونین براساس چیزی ک پژوهش کردید و سابقه دارین اقدام کنید
This week was dominated by molecular ML and drug discovery events:

- Broad Institute published a YouTube playlist of talks from the recent Machine Learning and Drug Discovery Symposium and leading drug discovery researchers.

- ELLIS Machine Learning for Drug Discovery Workshop will take place online in Zoom and GatherTown on Nov 28th, registration is free!

- Valence Discovery launched a blog platform for Drug Discovery related posts, the inaugural post by Clemens Isert talks about Quantum ML for drug-like molecules.

And a new work from GemNet authors: How robust are modern graph neural network potentials in long and hot molecular dynamics simulations? The work supports a recent line of works on ML force fields particularly focusing on hot dynamics where low test MAE does not necessarily correspond to good simulations. The most important robustness factor seems to be more training data!
Molecular dynamics is one of the booming Geometric DL areas where equivariant models show the best qualities. The two cool recent papers on that topic:

⚛️ Forces are not Enough: Benchmark and Critical Evaluation for Machine Learning Force Fields with Molecular Simulations by Fu et al. introduces a new benchmark for molecular dynamics - in addition to MD17, the authors add datasets on modeling liquids (Water), peptides (Alanine dipeptide), and solid-state materials (LiPS). More importantly, apart from Energy as the main metric, the authors consider a wide range of physical properties like Stability, Diffusivity, and Radial Distribution Functions. Most SOTA molecular dynamics models were probed including SchNet, ForceNet, DimeNet, GemNet (-T and -dT), NequIP.

Density Functional Theory (DFT) calculations are one of the main workhorses of molecular dynamics (and account for a great deal of computing time in big clusters). DFT is O(n^3) to the input size though, so can ML help here? Learned Force Fields Are Ready For Ground State Catalyst Discovery by Schaarschmidt et al. present the experimental study of models of learned potentials - turns out GNNs can do a very good job in O(n) time!
🪐 For astrophysics aficionados: Mangrove: Learning Galaxy Properties from Merger Trees by Jespersen et al. apply GraphSAGE to merger trees of dark matter to predict a variety of galactic properties like stellar mass, cold gas mass, star formation rate, and even black hole mass. The paper is heavy on the terminology of astrophysics but pretty easy in terms of GNN parameterization and training. Mangrove works 4-9 orders of magnitude faster than standard models (that is, 10 000 - 1 000 000 000 times faster). Experimental charts are pieces of art that you can hang on a wall.

🤖 Compositional Semantic Parsing with Large Language Models by Drozdov, Schärli et al. pretty much solve the compositional semantic parsing task (natural language query - structured query like SPARQL) using only code-davinci-002 language model from OpenAI (which is InstructGPT fine-tuned on code). No need for hefty tailored semantic parsing models - turns out a smart extension of the Chain-of-thought prompting (aka "let's think step by step") devised as Least-to-Most prompting (where we first answer easy subproblems before generating a full query) yields whopping 95% accuracy even on hardest Compositional Freebase Questions (CFQ) dataset. CFQ was introduced at ICLR 2020, and just after two years LMs cracked this task - looks like it's time for the new, even more complex dataset.

#مقاله
Forwarded from Meysam
نگارش نسخه دوم کتاب:
Mastering Transformers
رو شروع کردیم.
دوستانی که نسخه قبلی رو خوندن و اگر موضوعی یا مطلبی به نظرتون جذاب میاد و فکر میکنید که باید اضافه بشه بفرمایید خوشحال میشم نظراتتون رو بدونم.

@ai_person
سلام به همه دوستان یک سایتی با عنوان 1500 در حال انتشار در کانالهای تلگرامی و اینستاگرامی هست به هیچ عنوان روی این سایت کلیک نکنید و تاکید میشود که اصلا روی این سایت کلیک نکنید که عواقب جبران ناپذیر دارد.
Apply for Ph.D. positions starting this year!

We have room for several outstanding machine learning Ph.D. students to join our elite group at the University of Cambridge.

We are looking for:

leaders who want to realise their full potential and shape the future of machine learning
explorers who think boldly and prefer pushing boundaries to sitting still
builders who see projects through to completion and want to make a real-world impact
If you’re interested, please click the URL below. Note that we have already received a lot of applications. If you’re considering applying, earlier is definitely better!

Join the van der Schaar Lab:
https://eur03.safelinks.protection.outlook.com/?url=https%3A%2F%2Fwww.vanderschaar-lab.com%2Fjoin&data=05%7C01%7Ctl522%40universityofcambridgecloud.onmicrosoft.com%7C25eb48cdd35849dde6d208dac1c40d86%7C49a50445bdfa4b79ade3547b4f3986e9%7C1%7C0%7C638035345048429672%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=koklpmvj4TDjLmx8OXWKFUyWVlKRA%2B%2FzuRstdljgIbI%3D&reserved=0

#اپلای
This media is not supported in your browser
VIEW IN TELEGRAM
مدل زبانی خیلی بزرگ Galactica که می‌تواند:
Can summarize academic literature, solve math problems, generate Wiki articles, write scientific code, annotate molecules and proteins, and more.

To accelerate science, we open source all models including the 120 billion model with no friction. You can access them here.

galactica.org/paper.pdf

Explore and get weights: galactica.org

#مقاله

✳️ @AI_Python