Machine Learning And AI
1.65K subscribers
198 photos
1 video
19 files
351 links
Hi All and Welcome Join our channel for Jobs,latest Programming Blogs, machine learning blogs.
In case any doubt regarding ML/Data Science please reach out to me @ved1104 subscribe my channel
https://youtube.com/@geekycodesin?si=JzJo3WS5E_VFmD1k
Download Telegram
Fundamentals of a ๐—ฉ๐—ฒ๐—ฐ๐˜๐—ผ๐—ฟ ๐——๐—ฎ๐˜๐—ฎ๐—ฏ๐—ฎ๐˜€๐—ฒ.

With the rise of GenAI, Vector Databases skyrocketed in popularity. The truth is that a Vector Database is also useful for different kinds of AI Systems outside of a Large Language Model context.

When it comes to Machine Learning, we often deal with Vector Embeddings. Vector Databases were created to perform specifically well when working with them:

โžก๏ธ Storing.
โžก๏ธ Updating.
โžก๏ธ Retrieving.

When we talk about retrieval, we refer to retrieving set of vectors that are most similar to a query in a form of a vector that is embedded in the same Latent space. This retrieval procedure is called Approximate Nearest Neighbour (ANN) search.

A query here could be in a form of an object like an image for which we would like to find similar images. Or it could be a question for which we want to retrieve relevant context that could later be transformed into an answer via a LLM.

Letโ€™s look into how one would interact with a Vector Database:

๐—ช๐—ฟ๐—ถ๐˜๐—ถ๐—ป๐—ด/๐—จ๐—ฝ๐—ฑ๐—ฎ๐˜๐—ถ๐—ป๐—ด ๐——๐—ฎ๐˜๐—ฎ.

1. Choose a ML model to be used to generate Vector Embeddings.
2. Embed any type of information: text, images, audio, tabular. Choice of ML model used for embedding will depend on the type of data.
3. Get a Vector representation of your data by running it through the Embedding Model.
4. Store additional metadata together with the Vector Embedding. This data would later be used to pre-filter or post-filter ANN search results.
5. Vector DB indexes Vector Embedding and metadata separately. There are multiple methods that can be used for creating vector indexes, some of them: Random Projection, Product Quantization, Locality-sensitive Hashing.
6. Vector data is stored together with indexes for Vector Embeddings and metadata connected to the Embedded objects.

๐—ฅ๐—ฒ๐—ฎ๐—ฑ๐—ถ๐—ป๐—ด ๐——๐—ฎ๐˜๐—ฎ.

7. A query to be executed against a Vector Database will usually consist of two parts:

โžก๏ธ Data that will be used for ANN search. e.g. an image for which you want to find similar ones.
โžก๏ธ Metadata query to exclude Vectors that hold specific qualities known beforehand. E.g. given that you are looking for similar images of apartments - exclude apartments in a specific location.

8. You execute Metadata Query against the metadata index. It could be done before or after the ANN search procedure.
9. You embed the data into the Latent space with the same model that was used for writing the data to the Vector DB.
10. ANN search procedure is applied and a set of Vector embeddings are retrieved. Popular similarity measures for ANN search include: Cosine Similarity, Euclidean Distance, Dot Product.
๐Ÿ‘1
Three different learning styles in machine learning algorithms:

1. Supervised Learning

Input data is called training data and has a known label or result such as spam/not-spam or a stock price at a time.

A model is prepared through a training process in which it is required to make predictions and is corrected when those predictions are wrong. The training process continues until the model achieves a desired level of accuracy on the training data.

Example problems are classification and regression.

Example algorithms include: Logistic Regression and the Back Propagation Neural Network.

2. Unsupervised Learning

Input data is not labeled and does not have a known result.

A model is prepared by deducing structures present in the input data. This may be to extract general rules. It may be through a mathematical process to systematically reduce redundancy, or it may be to organize data by similarity.

Example problems are clustering, dimensionality reduction and association rule learning.

Example algorithms include: the Apriori algorithm and K-Means.

3. Semi-Supervised Learning

Input data is a mixture of labeled and unlabelled examples.

There is a desired prediction problem but the model must learn the structures to organize the data as well as make predictions.

Example problems are classification and regression.

Example algorithms are extensions to other flexible methods that make assumptions about how to model the unlabeled data.
๐Ÿ‘2
๐ŸšฆTop 10 Data Science Tools๐Ÿšฆ

Here we will examine the top best Data Science tools that are utilized generally by data researchers and analysts. But prior to beginning let us discuss about what is Data Science.

๐Ÿ›ฐWhat is Data Science ?

Data science is a quickly developing field that includes the utilization of logical strategies, calculations, and frameworks to extract experiences and information from organized and unstructured data .

๐Ÿ—ฝTop Data Science Tools that are normally utilized :

1.) Jupyter Notebook : Jupyter Notebook is an open-source web application that permits clients to make and share archives that contain live code, conditions, representations, and narrative text .

2.) Keras : Keras is a famous open-source brain network library utilized in data science. It is known for its usability and adaptability.
Keras provides a range of tools and techniques for dealing with common data science problems, such as overfitting, underfitting, and regularization.

3.) PyTorch : PyTorch is one more famous open-source AI library utilized in information science. PyTorch also offers easy-to-use interfaces for various tasks such as data loading, model building, training, and deployment, making it accessible to beginners as well as experts in the field of machine learning.

4.) TensorFlow : TensorFlow allows data researchers to play out an extensive variety of AI errands, for example, image recognition , natural language processing , and deep learning.

5.) Spark : Spark allows data researchers to perform data processing tasks like data control, investigation, and machine learning , rapidly and effectively.

6.) Hadoop : Hadoop provides a distributed file system (HDFS) and a distributed processing framework (MapReduce) that permits data researchers to handle enormous datasets rapidly.

7.) Tableau : Tableau is a strong data representation tool that permits data researchers to make intuitive dashboards and perceptions. Tableau allows users to combine multiple charts.

8.) SQL : SQL (Structured Query Language) SQL permits data researchers to perform complex queries , join tables, and aggregate data, making it simple to extricate bits of knowledge from enormous datasets. It is a powerful tool for data management, especially for large datasets.

9.) Power BI : Power BI is a business examination tool that conveys experiences and permits clients to make intuitive representations and reports without any problem.

10.) Excel : Excel is a spreadsheet program that broadly utilized in data science. It is an amazing asset for information the board, examination, and visualization .Excel can be used to explore the data by creating pivot tables, histograms, scatterplots, and other types of visualizations.
Walmart built an AI semantic search system processing millions of queries with 99% recall.
- When a user searches for a product, the query goes through a Siamese network with pre-trained tokenisers. This architecture allows the model to use the context of the input queries effectively.

- Different attributes are concatenated to the query title using a special token. This ensures that the model can distinguish among different product characteristics, like brand or colour, when processing a query.

- During training, the model employs a sampled softmax loss function, where both relevant and irrelevant products are considered for each query -- this helps improve the accuracy in distinguishing between different product matches.

- The architecture combines multiple embeddings for both queries and products. This lets the system capture the varying meanings of common queries, improving the model's flexibility and interpretation.
Data Science Interview Question

Q. What is the difference between bagging and boosting in decision trees?

Answer:

Bagging or bootstrap aggregation, is the process of randomly sampling with replacement multiple times from your original dataset and fitting a decision tree to each of the datasets. Then, you average the predictions made across multiple datasets for regression, or take the majority vote for classification. This averaging approach reduces the variance of the ensemble, and performs better than a single decision tree which fits the data hard and is likely to overfit.

Boosting, on the other hand, is a sequential learning approach. Given the original dataset, the boosting approach does not attempt to fit the data hard, but learns slowly. In boosting, given the current model, the algorithm fits a small, shrunken tree to the residuals of the model. Then, it adds this shrunken tree to the original tree to update the residuals. It continues this process as more small, shrunken trees are fit to the residuals of the model. By focusing on improving residual error, and using this stepwise, sequential approach, the function improves in areas that it usually does not perform well in.
Walk away from people and situations where you do not get respect. When someone does not value you, just ignore them and move on. The more we walk away from such situations, the more we open ourselves up to future possibilities in life.
Machine Learning And AI
https://youtu.be/8r_yY23mpV0
Hi guys a lot of you have not subscribed my channel yet. If you're reading this message then don't forget to subscribe my channel and comment your views.  At least half of you go and subscribe my channel.
Thank you in advance
Finally first 200 subscribers completed. Thank You to everyone who has subscribed my channel till now. And whoever hasn't subscribed, please do consider subscribing it
โค2
Data Scientist Roadmap
|
|-- 1. Basic Foundations
|   |-- a. Mathematics
|   |   |-- i. Linear Algebra
|   |   |-- ii. Calculus
|   |   |-- iii. Probability
|   |   -- iv. Statistics
|   |
|   |-- b. Programming
|   |   |-- i. Python
|   |   |   |-- 1. Syntax and Basic Concepts
|   |   |   |-- 2. Data Structures
|   |   |   |-- 3. Control Structures
|   |   |   |-- 4. Functions
|   |   |  
-- 5. Object-Oriented Programming
|   |   |
|   |   -- ii. R (optional, based on preference)
|   |
|   |-- c. Data Manipulation
|   |   |-- i. Numpy (Python)
|   |   |-- ii. Pandas (Python)
|   |  
-- iii. Dplyr (R)
|   |
|   -- d. Data Visualization
|       |-- i. Matplotlib (Python)
|       |-- ii. Seaborn (Python)
|      
-- iii. ggplot2 (R)
|
|-- 2. Data Exploration and Preprocessing
|   |-- a. Exploratory Data Analysis (EDA)
|   |-- b. Feature Engineering
|   |-- c. Data Cleaning
|   |-- d. Handling Missing Data
|   -- e. Data Scaling and Normalization
|
|-- 3. Machine Learning
|   |-- a. Supervised Learning
|   |   |-- i. Regression
|   |   |   |-- 1. Linear Regression
|   |   |  
-- 2. Polynomial Regression
|   |   |
|   |   -- ii. Classification
|   |       |-- 1. Logistic Regression
|   |       |-- 2. k-Nearest Neighbors
|   |       |-- 3. Support Vector Machines
|   |       |-- 4. Decision Trees
|   |      
-- 5. Random Forest
|   |
|   |-- b. Unsupervised Learning
|   |   |-- i. Clustering
|   |   |   |-- 1. K-means
|   |   |   |-- 2. DBSCAN
|   |   |   -- 3. Hierarchical Clustering
|   |   |
|   |  
-- ii. Dimensionality Reduction
|   |       |-- 1. Principal Component Analysis (PCA)
|   |       |-- 2. t-Distributed Stochastic Neighbor Embedding (t-SNE)
|   |       -- 3. Linear Discriminant Analysis (LDA)
|   |
|   |-- c. Reinforcement Learning
|   |-- d. Model Evaluation and Validation
|   |   |-- i. Cross-validation
|   |   |-- ii. Hyperparameter Tuning
|   |  
-- iii. Model Selection
|   |
|   -- e. ML Libraries and Frameworks
|       |-- i. Scikit-learn (Python)
|       |-- ii. TensorFlow (Python)
|       |-- iii. Keras (Python)
|      
-- iv. PyTorch (Python)
|
|-- 4. Deep Learning
|   |-- a. Neural Networks
|   |   |-- i. Perceptron
|   |   -- ii. Multi-Layer Perceptron
|   |
|   |-- b. Convolutional Neural Networks (CNNs)
|   |   |-- i. Image Classification
|   |   |-- ii. Object Detection
|   |  
-- iii. Image Segmentation
|   |
|   |-- c. Recurrent Neural Networks (RNNs)
|   |   |-- i. Sequence-to-Sequence Models
|   |   |-- ii. Text Classification
|   |   -- iii. Sentiment Analysis
|   |
|   |-- d. Long Short-Term Memory (LSTM) and Gated Recurrent Units (GRU)
|   |   |-- i. Time Series Forecasting
|   |  
-- ii. Language Modeling
|   |
|   -- e. Generative Adversarial Networks (GANs)
|       |-- i. Image Synthesis
|       |-- ii. Style Transfer
|      
-- iii. Data Augmentation
|
|-- 5. Big Data Technologies
|   |-- a. Hadoop
|   |   |-- i. HDFS
|   |   -- ii. MapReduce
|   |
|   |-- b. Spark
|   |   |-- i. RDDs
|   |   |-- ii. DataFrames
|   |  
-- iii. MLlib
|   |
|   -- c. NoSQL Databases
|       |-- i. MongoDB
|       |-- ii. Cassandra
|       |-- iii. HBase
|      
-- iv. Couchbase
|
|-- 6. Data Visualization and Reporting
|   |-- a. Dashboarding Tools
|   |   |-- i. Tableau
|   |   |-- ii. Power BI
|   |   |-- iii. Dash (Python)
|   |   -- iv. Shiny (R)
|   |
|   |-- b. Storytelling with Data
|  
-- c. Effective Communication
|
|-- 7. Domain Knowledge and Soft Skills
|   |-- a. Industry-specific Knowledge
|   |-- b. Problem-solving
|   |-- c. Communication Skills
|   |-- d. Time Management
|   -- e. Teamwork
|
-- 8. Staying Updated and Continuous Learning
    |-- a. Online Courses
    |-- b. Books and Research Papers
    |-- c. Blogs and Podcasts
    |-- d. Conferences and Workshops
    `-- e. Networking and Community Engagement
โค7๐Ÿ‘1