ML Research Hub
32.7K subscribers
4.09K photos
237 videos
23 files
4.41K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
πŸš€ Master Data Science & Programming!

Unlock your potential with this curated list of Telegram channels. Whether you need books, datasets, interview prep, or project ideas, we have the perfect resource for you. Join the community today!


πŸ”° Machine Learning with Python
Learn Machine Learning with hands-on Python tutorials, real-world code examples, and clear explanations for researchers and developers.
https://t.me/CodeProgrammer

πŸ”– Machine Learning
Machine learning insights, practical tutorials, and clear explanations for beginners and aspiring data scientists. Follow the channel for models, algorithms, coding guides, and real-world ML applications.
https://t.me/DataScienceM

🧠 Code With Python
This channel delivers clear, practical content for developers, covering Python, Django, Data Structures, Algorithms, and DSA – perfect for learning, coding, and mastering key programming skills.
https://t.me/DataScience4

🎯 PyData Careers | Quiz
Python Data Science jobs, interview tips, and career insights for aspiring professionals.
https://t.me/DataScienceQ

πŸ’Ύ Kaggle Data Hub
Your go-to hub for Kaggle datasets – explore, analyze, and leverage data for Machine Learning and Data Science projects.
https://t.me/datasets1

πŸ§‘β€πŸŽ“ Udemy Coupons | Courses
The first channel in Telegram that offers free Udemy coupons
https://t.me/DataScienceC

πŸ˜€ ML Research Hub
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.
https://t.me/DataScienceT

πŸ’¬ Data Science Chat
An active community group for discussing data challenges and networking with peers.
https://t.me/DataScience9

🐍 Python Arab| Ψ¨Ψ§ΩŠΨ«ΩˆΩ† عربي
The largest Arabic-speaking group for Python developers to share knowledge and help.
https://t.me/PythonArab

πŸ–Š Data Science Jupyter Notebooks
Explore the world of Data Science through Jupyter Notebooksβ€”insights, tutorials, and tools to boost your data journey. Code, analyze, and visualize smarter with every post.
https://t.me/DataScienceN

πŸ“Ί Free Online Courses | Videos
Free online courses covering data science, machine learning, analytics, programming, and essential skills for learners.
https://t.me/DataScienceV

πŸ“ˆ Data Analytics
Dive into the world of Data Analytics – uncover insights, explore trends, and master data-driven decision making.
https://t.me/DataAnalyticsX

🎧 Learn Python Hub
Master Python with step-by-step courses – from basics to advanced projects and practical applications.
https://t.me/Python53

⭐️ Research Papers
Professional Academic Writing & Simulation Services
https://t.me/DataScienceY

━━━━━━━━━━━━━━━━━━
Admin: @HusseinSheikho
Please open Telegram to view this post
VIEW IN TELEGRAM
❀1
This media is not supported in your browser
VIEW IN TELEGRAM
✨InsertAnywhere: Bridging 4D Scene Geometry and Diffusion Models for Realistic Video Object Insertion

πŸ“ Summary:
InsertAnywhere is a framework for realistic video object insertion. It uses 4D aware mask generation for geometric consistency and an extended diffusion model for appearance-faithful synthesis, outperforming existing methods.

πŸ”Ή Publication Date: Published on Dec 19

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.17504
β€’ PDF: https://arxiv.org/pdf/2512.17504
β€’ Project Page: https://myyzzzoooo.github.io/InsertAnywhere/
β€’ Github: https://github.com/myyzzzoooo/InsertAnywhere

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#VideoEditing #DiffusionModels #ComputerVision #DeepLearning #GenerativeAI
❀1
✨MAI-UI Technical Report: Real-World Centric Foundation GUI Agents

πŸ“ Summary:
MAI-UI introduces a family of foundation GUI agents tackling real-world deployment challenges. It uses a self-evolving data pipeline, device-cloud collaboration, and online RL to set new state-of-the-art in GUI grounding and mobile navigation, significantly boosting performance and privacy.

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22047
β€’ PDF: https://arxiv.org/pdf/2512.22047

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#GUIAgents #AI #ReinforcementLearning #MobileTech #HCI
❀2
✨See Less, See Right: Bi-directional Perceptual Shaping For Multimodal Reasoning

πŸ“ Summary:
Bi-directional Perceptual Shaping BiPS improves vision-language models by using question-conditioned masked views to shape perception during training. It employs two constraints to ensure complete coverage of relevant pixels and enforce fine-grained visual reliance, preventing text-only shortcuts...

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22120
β€’ PDF: https://arxiv.org/pdf/2512.22120
β€’ Github: https://github.com/zss02/BiPS

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#MultimodalAI #VisionLanguageModels #MachineLearning #AIResearch #DeepLearning
❀1
✨ProEdit: Inversion-based Editing From Prompts Done Right

πŸ“ Summary:
Inversion-based visual editing provides an effective and training-free way to edit an image or a video based on user instructions. Existing methods typically inject source image information during the...

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22118
β€’ PDF: https://arxiv.org/pdf/2512.22118
β€’ Project Page: https://isee-laboratory.github.io/ProEdit/
β€’ Github: https://isee-laboratory.github.io/ProEdit

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#AI #DataScience #MachineLearning #HuggingFace #Research
❀1
✨SVBench: Evaluation of Video Generation Models on Social Reasoning

πŸ“ Summary:
Recent text-to-video generation models exhibit remarkable progress in visual realism, motion fidelity, and text-video alignment, yet they remain fundamentally limited in their ability to generate soci...

πŸ”Ή Publication Date: Published on Dec 25

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.21507
β€’ PDF: https://arxiv.org/pdf/2512.21507
β€’ Github: https://github.com/Gloria2tt/SVBench-Evaluation

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#AI #DataScience #MachineLearning #HuggingFace #Research
❀1
✨SWE-RM: Execution-free Feedback For Software Engineering Agents

πŸ“ Summary:
This paper introduces SWE-RM, a robust, execution-free reward model for software engineering agents. It overcomes limitations of execution-based feedback, improving coding agent performance in both test-time scaling and reinforcement learning. SWE-RM achieves new state-of-the-art results for open...

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.21919
β€’ PDF: https://arxiv.org/pdf/2512.21919

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#SoftwareEngineering #AI #ReinforcementLearning #CodingAgents #RewardModels
❀1
✨Mindscape-Aware Retrieval Augmented Generation for Improved Long Context Understanding

πŸ“ Summary:
MiA-RAG enhances RAG systems with global context awareness, inspired by human understanding. It uses hierarchical summarization to build a 'mindscape,' improving long-context retrieval and generation for better evidence-based understanding.

πŸ”Ή Publication Date: Published on Dec 19

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.17220
β€’ PDF: https://arxiv.org/pdf/2512.17220

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/MindscapeRAG/MiA-Emb-8B
β€’ https://huggingface.co/MindscapeRAG/MiA-Emb-4B
β€’ https://huggingface.co/MindscapeRAG/MiA-Emb-0.6B

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#RAG #LLM #NLP #GenerativeAI #ContextUnderstanding
❀1
✨TimeBill: Time-Budgeted Inference for Large Language Models

πŸ“ Summary:
TimeBill is a framework for LLMs in time-critical systems. It predicts execution time and adaptively adjusts KV cache eviction to balance inference efficiency and response performance within given time budgets, improving task completion rates.

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.21859
β€’ PDF: https://arxiv.org/pdf/2512.21859

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#LLM #AI #RealTimeAI #InferenceOptimization #DeepLearning
❀1
✨UniPercept: Towards Unified Perceptual-Level Image Understanding across Aesthetics, Quality, Structure, and Texture

πŸ“ Summary:
UniPercept-Bench provides a unified framework and datasets for perceptual image understanding aesthetics, quality, structure, texture. The UniPercept model, trained with DAPT and T-ARL, outperforms MLLMs, generalizes across VR and VQA, and acts as a text-to-image reward model.

πŸ”Ή Publication Date: Published on Dec 25

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.21675
β€’ PDF: https://arxiv.org/pdf/2512.21675
β€’ Project Page: https://thunderbolt215.github.io/Unipercept-project/
β€’ Github: https://github.com/thunderbolt215/UniPercept

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/Thunderbolt215215/UniPercept

✨ Datasets citing this paper:
β€’ https://huggingface.co/datasets/Thunderbolt215215/UniPercept-Bench

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#ImageUnderstanding #ComputerVision #AIResearch #PerceptualAI #DeepLearning
❀1
✨Omni-Weather: Unified Multimodal Foundation Model for Weather Generation and Understanding

πŸ“ Summary:
Omni-Weather is a new multimodal foundation model that unifies weather generation and understanding in a single architecture. It uses shared self-attention and a Chain-of-Thought dataset for interpretable, high-quality outputs, achieving state-of-the-art performance.

πŸ”Ή Publication Date: Published on Dec 25

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.21643
β€’ PDF: https://arxiv.org/pdf/2512.21643

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#WeatherGeneration #FoundationModels #MultimodalAI #AIResearch #DeepLearning
❀1
✨SlideTailor: Personalized Presentation Slide Generation for Scientific Papers

πŸ“ Summary:
SlideTailor generates personalized presentation slides for scientific papers by learning user preferences implicitly from example pairs and visual templates. It uses a chain-of-speech mechanism to align content with oral narration.

πŸ”Ή Publication Date: Published on Dec 23

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.20292
β€’ PDF: https://arxiv.org/pdf/2512.20292
β€’ Project Page: https://github.com/nusnlp/SlideTailor
β€’ Github: https://github.com/nusnlp/SlideTailor

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#SlideGeneration #ScientificCommunication #AI #NLP #ResearchTools
❀2
✨Coupling Experts and Routers in Mixture-of-Experts via an Auxiliary Loss

πŸ“ Summary:
Expert-Router Coupling ERC loss aligns MoE router decisions with expert capabilities. It uses proxy tokens and activation constraints to ensure experts specialize, improving performance and computational efficiency. ERC also allows tracking expert specialization during training.

πŸ”Ή Publication Date: Published on Dec 29

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.23447
β€’ PDF: https://arxiv.org/pdf/2512.23447

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#MixtureOfExperts #DeepLearning #MachineLearning #AI #NeuralNetworks
This media is not supported in your browser
VIEW IN TELEGRAM
✨LiveTalk: Real-Time Multimodal Interactive Video Diffusion via Improved On-Policy Distillation

πŸ“ Summary:
LiveTalk enables real-time multimodal interactive video generation from text, image, and audio by improving on-policy diffusion distillation. It reduces inference latency by 20x while maintaining quality, allowing seamless human-AI interaction.

πŸ”Ή Publication Date: Published on Dec 29

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.23576
β€’ PDF: https://arxiv.org/pdf/2512.23576
β€’ Github: https://github.com/GAIR-NLP/LiveTalk

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#VideoGeneration #AI #DiffusionModels #RealTimeAI #MultimodalAI
✨SmartSnap: Proactive Evidence Seeking for Self-Verifying Agents

πŸ“ Summary:
SmartSnap introduces proactive, in-situ self-verification for autonomous agents, moving away from passive, post-hoc task verification. Self-Verifying Agents complete tasks and curate minimal snapshot evidence to prove accomplishment, boosting scalability and performance for LLM-driven agents in G...

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://huggingface.co/collections/yolay/smartsnap
β€’ PDF: https://arxiv.org/pdf/2512.22322
β€’ Project Page: https://huggingface.co/collections/yolay/smartsnap

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/yolay/SmartSnap-LLaMA3.1-8B
β€’ https://huggingface.co/yolay/SmartSnap-Qwen2.5-7B
β€’ https://huggingface.co/yolay/SmartSnap-Qwen3-8B

✨ Datasets citing this paper:
β€’ https://huggingface.co/datasets/yolay/SmartSnap-FT
β€’ https://huggingface.co/datasets/yolay/SmartSnap-RL

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#AI #LLM #AutonomousAgents #AgentVerification #AIResearch
Media is too big
VIEW IN TELEGRAM
✨Yume-1.5: A Text-Controlled Interactive World Generation Model

πŸ“ Summary:
Yume-1.5 is a novel framework that generates realistic, interactive, and continuous worlds from a single image or text prompt. It overcomes prior limitations in real-time performance and text control by using unified context compression, streaming acceleration, and text-controlled world events.

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22096
β€’ PDF: https://arxiv.org/pdf/2512.22096
β€’ Project Page: https://stdstu12.github.io/YUME-Project/
β€’ Github: https://github.com/stdstu12/YUME

πŸ”Ή Models citing this paper:
β€’ https://huggingface.co/stdstu123/Yume-5B-720P

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#AI #GenerativeAI #WorldGeneration #ComputerGraphics #DeepLearning
✨Diffusion Knows Transparency: Repurposing Video Diffusion for Transparent Object Depth and Normal Estimation

πŸ“ Summary:
Transparent objects are hard for perception. This work observes video diffusion models can synthesize transparent phenomena, so they repurpose one. Their DKT model, trained on a new dataset, achieves zero-shot SOTA for depth and normal estimation of transparent objects, proving diffusion knows tr...

πŸ”Ή Publication Date: Published on Dec 29

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.23705
β€’ PDF: https://arxiv.org/pdf/2512.23705
β€’ Project Page: https://daniellli.github.io/projects/DKT/
β€’ Github: https://github.com/Daniellli/DKT

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#ComputerVision #DiffusionModels #DepthEstimation #TransparentObjects #AIResearch
✨SpotEdit: Selective Region Editing in Diffusion Transformers

πŸ“ Summary:
SpotEdit is a training-free framework for selective image editing in diffusion transformers. It avoids reprocessing stable regions by reusing their features, combining them with edited areas. This reduces computation and preserves unchanged regions, enhancing efficiency and precision.

πŸ”Ή Publication Date: Published on Dec 26

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22323
β€’ PDF: https://arxiv.org/pdf/2512.22323
β€’ Project Page: https://biangbiang0321.github.io/SpotEdit.github.io
β€’ Github: https://biangbiang0321.github.io/SpotEdit.github.io

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#ImageEditing #DiffusionModels #ComputerVision #AIResearch #DeepLearning
✨Dream-VL & Dream-VLA: Open Vision-Language and Vision-Language-Action Models with Diffusion Language Model Backbone

πŸ“ Summary:
Dream-VL and Dream-VLA are diffusion-based vision-language and vision-language-action models. They achieve state-of-the-art performance in visual planning and robotic control, surpassing autoregressive baselines via their diffusion backbone's superior action generation.

πŸ”Ή Publication Date: Published on Dec 27

πŸ”Ή Paper Links:
β€’ arXiv Page: https://arxiv.org/abs/2512.22615
β€’ PDF: https://arxiv.org/pdf/2512.22615
β€’ Project Page: https://hkunlp.github.io/blog/2025/dream-vlx/
β€’ Github: https://github.com/DreamLM/Dream-VLX

==================================

For more data science resources:
βœ“ https://t.me/DataScienceT

#VisionLanguageModels #DiffusionModels #Robotics #AI #ComputerVision