ML Research Hub
32.8K subscribers
4.34K photos
265 videos
23 files
4.69K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
🤖🧠 Open WebUI: The Most Powerful Self-Hosted AI Platform for Local and Private LLMs

🗓️ 09 Nov 2025
📚 AI News & Trends

In the rapidly evolving landscape of artificial intelligence, the ability to run large language models securely and efficiently has become a major priority for developers, enterprises and privacy-focused users. While cloud-based AI services are convenient, they rely heavily on remote servers, internet access and third-party control. This is where Open WebUI stands out as a ...

#OpenWebUI #SelfHostedAI #PrivateLLMs #LocalAI #AISecurity #OpenSourcePlatform
Hail to the Thief: Exploring Attacks and Defenses in Decentralised GRPO

📝 Summary:
This study identifies and demonstrates adversarial attacks in decentralized GRPO for LLMs, achieving 100% success rates by injecting malicious tokens. It also proposes effective defense mechanisms that can stop these attacks completely.

🔹 Publication Date: Published on Nov 12

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.09780
• PDF: https://arxiv.org/pdf/2511.09780

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #AdversarialAttacks #AISecurity #DecentralizedAI #GRPO
1
Evolve the Method, Not the Prompts: Evolutionary Synthesis of Jailbreak Attacks on LLMs

📝 Summary:
EvoSynth is a new framework that autonomously engineers and evolves novel, code-based jailbreak methods for LLMs, moving beyond prompt refinement. It uses self-correction to create diverse and highly successful attacks, achieving 85.5% ASR against robust models.

🔹 Publication Date: Published on Nov 16

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.12710
• PDF: https://arxiv.org/pdf/2511.12710

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #JailbreakAttacks #AISecurity #EvolutionaryAlgorithms #AIResearch
1
Multi-Faceted Attack: Exposing Cross-Model Vulnerabilities in Defense-Equipped Vision-Language Models

📝 Summary:
Multi-Faceted Attack MFA reveals cross-model safety vulnerabilities in defense-equipped Vision-Language Models. It uses Attention-Transfer Attack to hide harmful instructions and bypass filters, exploiting shared visual representations for high success rates. MFA challenges the robustness of curr...

🔹 Publication Date: Published on Nov 20

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.16110
• PDF: https://arxiv.org/pdf/2511.16110

==================================

For more data science resources:
https://t.me/DataScienceT

#VisionLanguageModels #AISecurity #AdversarialAttacks #AIvulnerabilities #MachineLearning
In-Context Representation Hijacking

📝 Summary:
Doublespeak is an in-context attack that hijacks LLM representations. It replaces harmful keywords with benign ones in examples, making LLMs interpret innocuous prompts as harmful, bypassing safety. This highlights a need for representation-level alignment.

🔹 Publication Date: Published on Dec 3

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.03771
• PDF: https://arxiv.org/pdf/2512.03771

==================================

For more data science resources:
https://t.me/DataScienceT

#LLM #AISafety #AIsecurity #InContextLearning #RepresentationLearning
1
Adversarial Confusion Attack: Disrupting Multimodal Large Language Models

📝 Summary:
The Adversarial Confusion Attack systematically disrupts multimodal LLMs, causing incoherent or confidently incorrect outputs. This basic adversarial technique transfers to diverse models, including proprietary ones, potentially hindering AI Agent reliability.

🔹 Publication Date: Published on Nov 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.20494
• PDF: https://arxiv.org/pdf/2511.20494

==================================

For more data science resources:
https://t.me/DataScienceT

#AdversarialAttack #MultimodalAI #LLMs #AISecurity #AIResearch
1👍1🔥1
OmniSafeBench-MM: A Unified Benchmark and Toolbox for Multimodal Jailbreak Attack-Defense Evaluation

📝 Summary:
OmniSafeBench-MM is a unified toolbox for evaluating multi-modal jailbreak attacks and defenses in MLLMs. It integrates various attacks, defense strategies, and a diverse dataset to provide a comprehensive, standardized, and reproducible platform for research.

🔹 Publication Date: Published on Dec 6

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.06589
• PDF: https://arxiv.org/pdf/2512.06589

==================================

For more data science resources:
https://t.me/DataScienceT

#MLLMs #AISafety #AIsecurity #Benchmark #DeepLearning
1
GateBreaker: Gate-Guided Attacks on Mixture-of-Expert LLMs

📝 Summary:
GateBreaker is the first framework to compromise MoE LLM safety by identifying and disabling ~3% of safety neurons in expert layers. This raises attack success rates from 7.4% to 64.9% across eight LLMs and generalizes to VLMs, showing concentrated and transferable safety vulnerabilities.

🔹 Publication Date: Published on Dec 24

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.21008
• PDF: https://arxiv.org/pdf/2512.21008

==================================

For more data science resources:
https://t.me/DataScienceT

#LLM #AIsecurity #MoELLMs #AIvulnerability #GateBreaker
Few Tokens Matter: Entropy Guided Attacks on Vision-Language Models

📝 Summary:
Targeting high-entropy tokens in vision-language models causes significant semantic degradation with reduced budgets. This attack strategy reveals critical transferable safety risks across different VLM architectures.

🔹 Publication Date: Published on Dec 26, 2025

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.21815
• PDF: https://arxiv.org/pdf/2512.21815

==================================

For more data science resources:
https://t.me/DataScienceT

#VisionLanguageModels #AdversarialAI #AIsecurity #MachineLearning #DeepLearning