Python Data Science Jobs & Interviews
17.9K subscribers
138 photos
3 videos
5 files
250 links
Your go-to hub for Python and Data Science—featuring questions, answers, quizzes, and interview tips to sharpen your skills and boost your career in the data-driven world.

Admin: @Hussein_Sheikho
Download Telegram
🙏💸 500$ FOR THE FIRST 500 WHO JOIN THE CHANNEL! 🙏💸

Join our channel today for free! Tomorrow it will cost 500$!

https://t.me/+QHlfCJcO2lRjZWVl

You can join at this link! 👆👇

https://t.me/+QHlfCJcO2lRjZWVl
This channels is for Programmers, Coders, Software Engineers.

0️⃣ Python
1️⃣ Data Science
2️⃣ Machine Learning
3️⃣ Data Visualization
4️⃣ Artificial Intelligence
5️⃣ Data Analysis
6️⃣ Statistics
7️⃣ Deep Learning
8️⃣ programming Languages

https://t.me/addlist/8_rRW2scgfRhOTc0

https://t.me/Codeprogrammer
Please open Telegram to view this post
VIEW IN TELEGRAM
2
Question 30 (Intermediate - PyTorch):
What is the purpose of torch.no_grad() context manager in PyTorch?

A) Disables model training
B) Speeds up computations by disabling gradient tracking
C) Forces GPU memory cleanup
D) Enables distributed training

#Python #PyTorch #DeepLearning #NeuralNetworks

By: https://t.me/DataScienceQ
🔥1
Question 31 (Intermediate - Django ORM):
When using Django ORM's select_related() and prefetch_related() for query optimization, which statement is correct?

A) select_related uses JOINs (1 SQL query) while prefetch_related uses 2+ queries
B) Both methods generate exactly one SQL query
C) prefetch_related works only with ForeignKey relationships
D) select_related is better for many-to-many relationships

#Python #Django #ORM #Database

By: https://t.me/DataScienceQ
🔥1
Question 32 (Advanced - NLP & RNNs):
What is the key limitation of vanilla RNNs for NLP tasks that led to the development of LSTMs and GRUs?

A) Vanishing gradients in long sequences
B) High GPU memory usage
C) Inability to handle embeddings
D) Single-direction processing only

#Python #NLP #RNN #DeepLearning

By: https://t.me/DataScienceQ
2
Please open Telegram to view this post
VIEW IN TELEGRAM
1
Please open Telegram to view this post
VIEW IN TELEGRAM
1
Please open Telegram to view this post
VIEW IN TELEGRAM
2🔥1
Python Data Science Jobs & Interviews
Question 30 (Intermediate - PyTorch): What is the purpose of torch.no_grad() context manager in PyTorch? A) Disables model training B) Speeds up computations by disabling gradient tracking C) Forces GPU memory cleanup D) Enables distributed training…
Correct answer: B) Speeds up computations by disabling gradient tracking

import torch

model = torch.nn.Linear(10, 1)
x = torch.randn(5, 10)

# Inference without gradient tracking
with torch.no_grad():
prediction = model(x) # 30-50% faster than regular forward()
print(prediction.requires_grad) # False


### Key Use Cases:
1. Model Inference:
- Reduces memory overhead by ~40%
- Prevents accidental weight updates

2. Validation/Testing:
   for data in val_loader:
with torch.no_grad():
outputs = model(data) # No backprop needed


3. Weight Freezing:
   for param in model.layer.parameters():
param.requires_grad = False # Often used with no_grad()


### Performance Impact:
| Operation | Time (ms) | Memory (MB) |
|--------------------|-----------|-------------|
| Regular Forward | 15.2 | 1200 |
| no_grad() Forward| 9.8 | 720 |

*Note: Critical for deployment where every millisecond matters*
Python Data Science Jobs & Interviews
Question 31 (Intermediate - Django ORM): When using Django ORM's select_related() and prefetch_related() for query optimization, which statement is correct? A) select_related uses JOINs (1 SQL query) while prefetch_related uses 2+ queries B) Both methods…
Correct answer: A) `select_related` uses JOINs (1 SQL query) while `prefetch_related` uses 2+ queries

# Example models
class Author(models.Model):
name = models.CharField(max_length=100)

class Book(models.Model):
title = models.CharField(max_length=100)
author = models.ForeignKey(Author, on_delete=models.CASCADE)
genres = models.ManyToManyField('Genre')

# Optimized queries
books = Book.objects.select_related('author') # Single JOIN query
books = Book.objects.prefetch_related('genres') # 2 queries: books + genres


### Key Differences:
| Method | SQL Queries | Best For | Underlying Mechanism |
|----------------------|-------------|------------------------|----------------------|
| select_related() | 1 | ForeignKey, OneToOne | SQL JOIN |
| prefetch_related() | 2+ | ManyToMany, Reverse FK | Python-level caching |

### Performance Benchmark:
# Without optimization (N+1 problem)
for book in Book.objects.all():
print(book.author.name) # 1 query per book!

# With select_related (1 query total)
for book in Book.objects.select_related('author').all():
print(book.author.name) # Data already loaded


*Pro Tip*: Use Django Debug Toolbar to verify query counts!
Python Data Science Jobs & Interviews
Question 32 (Advanced - NLP & RNNs): What is the key limitation of vanilla RNNs for NLP tasks that led to the development of LSTMs and GRUs? A) Vanishing gradients in long sequences B) High GPU memory usage C) Inability to handle embeddings D) Single…
Correct answer: A) Vanishing gradients in long sequences

# Vanilla RNN vs LSTM comparison
import torch.nn as nn

rnn = nn.RNN(input_size=100, hidden_size=50, num_layers=1)
lstm = nn.LSTM(input_size=100, hidden_size=50, num_layers=1)

# Forward pass for 10 timesteps
inputs = torch.randn(10, 1, 100) # (seq_len, batch, input_size)
h_rnn = torch.zeros(1, 1, 50) # Initial hidden state
h_lstm = (torch.zeros(1, 1, 50), torch.zeros(1, 1, 50)) # LSTM state

out_rnn, _ = rnn(inputs, h_rnn) # Prone to vanishing gradients
out_lstm, _ = lstm(inputs, h_lstm) # Better long-term memory


### Key Problems with Vanilla RNNs:
1. Gradient Issues:
- Error signals decay exponentially over timesteps
- Tanh/Sigmoid activations compound the problem

2. LSTM/GRU Solutions:
| Mechanism | Purpose |
|-----------------|----------------------------------|
| Forget Gate | Controls what to remember |
| Input Gate | Regulates new information |
| Cell State | Highway for long-term gradients |

### Practical Impact:
# Training a sentiment analyzer
rnn_model = nn.RNN(embed_dim, hidden_dim) # Fails beyond 50 words
lstm_model = nn.LSTM(embed_dim, hidden_dim) # Handles 500+ words


*Modern Alternative*: Transformers (no recurrent connections at all)
Please open Telegram to view this post
VIEW IN TELEGRAM
1
🔥 Master Vision Transformers with 65+ MCQs! 🔥

Are you preparing for AI interviews or want to test your knowledge in Vision Transformers (ViT)?

🧠 Dive into 65+ curated Multiple Choice Questions covering the fundamentals, architecture, training, and applications of ViT — all with answers!

🌐 Explore Now: https://hackmd.io/@husseinsheikho/vit-mcq

🔹 Table of Contents
Basic Concepts (Q1–Q15)
Architecture & Components (Q16–Q30)
Attention & Transformers (Q31–Q45)
Training & Optimization (Q46–Q55)
Advanced & Real-World Applications (Q56–Q65)
Answer Key & Explanations

#VisionTransformer #ViT #DeepLearning #ComputerVision #Transformers #AI #MachineLearning #MCQ #InterviewPrep


✉️ Our Telegram channels: https://t.me/addlist/0f6vfFbEMdAwODBk

📱 Our WhatsApp channel: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
Please open Telegram to view this post
VIEW IN TELEGRAM
2
🚀 Comprehensive Guide: How to Prepare for a Python Job Interview – 200 Most Common Interview Questions

Are you ready: https://hackmd.io/@husseinsheikho/Python-interviews

#PythonInterview #JobPrep #PythonQuestions #CodingInterview #DataStructures #Algorithms #OOP #WebDevelopment #MachineLearning #DevOps

✉️ Our Telegram channels: https://t.me/addlist/0f6vfFbEMdAwODBk

📱 Our WhatsApp channel: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
Please open Telegram to view this post
VIEW IN TELEGRAM
3
🚀 Comprehensive Guide: How to Prepare for a Data Analyst Python Interview – 350 Most Common Interview Questions

Are you ready: https://hackmd.io/@husseinsheikho/pandas-interview

#DataAnalysis #PythonInterview #DataAnalyst #Pandas #NumPy #Matplotlib #Seaborn #SQL #DataCleaning #Visualization #MachineLearning #Statistics #InterviewPrep


✉️ Our Telegram channels: https://t.me/addlist/0f6vfFbEMdAwODBk

📱 Our WhatsApp channel: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
Please open Telegram to view this post
VIEW IN TELEGRAM
2
🚀 Comprehensive Guide: How to Prepare for an Image Processing Job Interview – 500 Most Common Interview Questions

Let's start: https://hackmd.io/@husseinsheikho/IP

#ImageProcessing #ComputerVision #OpenCV #Python #InterviewPrep #DigitalImageProcessing #MachineLearning #AI #SignalProcessing #ComputerGraphics

✉️ Our Telegram channels: https://t.me/addlist/0f6vfFbEMdAwODBk

📱 Our WhatsApp channel: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
Please open Telegram to view this post
VIEW IN TELEGRAM
1
5 minutes of work - 127,000$ profit!

Opened access to the Jay Welcome Club where the AI bot does all the work itself💻

Usually you pay crazy money to get into this club, but today access is free for everyone!

23,432% on deposit earned by club members in the last 6 months📈

Just follow Jay's trades and earn! 👇

https://t.me/+mONXtEgVxtU5NmZl
No skills? No problem. Just copy-paste and GET PAID.
➡️ 22,000+ already started… YOU'RE NEXT! Click here @NPFXSignals

#إعلان InsideAds
👍1
🚀 Comprehensive Guide: How to Prepare for a Graph Neural Networks (GNN) Job Interview – 350 Most Common Interview Questions

Read: https://hackmd.io/@husseinsheikho/GNN-interview

#GNN #GraphNeuralNetworks #MachineLearning #DeepLearning #AI #DataScience #PyTorchGeometric #DGL #NodeClassification #LinkPrediction #GraphML

✉️ Our Telegram channels: https://t.me/addlist/0f6vfFbEMdAwODBk

📱 Our WhatsApp channel: https://whatsapp.com/channel/0029VaC7Weq29753hpcggW2A
Please open Telegram to view this post
VIEW IN TELEGRAM
2