ML Research Hub
32.8K subscribers
4.09K photos
237 videos
23 files
4.41K links
Advancing research in Machine Learning – practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
The Curious Case of Analogies: Investigating Analogical Reasoning in Large Language Models

📝 Summary:
LLMs can encode high-level relational concepts for analogies but struggle with missing relational information and transfer to new entities. Success depends on strong structural alignment. Their analogical reasoning is emerging but limited compared to humans.

🔹 Publication Date: Published on Nov 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.20344
• PDF: https://arxiv.org/pdf/2511.20344

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #AnalogicalReasoning #AIResearch #NaturalLanguageProcessing #CognitiveAI
UniQL: Unified Quantization and Low-rank Compression for Adaptive Edge LLMs

📝 Summary:
UniQL unifies quantization and low-rank compression to deploy LLMs on mobile devices. It reduces memory by 4x-5.7x and improves token throughput by 2.7x-3.4x, maintaining accuracy across various model types.

🔹 Publication Date: Published on Dec 3

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.03383
• PDF: https://arxiv.org/pdf/2512.03383
• Project Page: https://hychiang.info/projects/uniql/
• Github: https://github.com/enyac-group/UniQL

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #EdgeAI #Quantization #ModelCompression #DeepLearning
ToolOrchestra: Elevating Intelligence via Efficient Model and Tool Orchestration

📝 Summary:
ToolOrchestra uses reinforcement learning to train small orchestrators that coordinate intelligent tools. This method enables an 8B model to outperform GPT-5 on complex tasks like Humanitys Last Exam, achieving higher accuracy at significantly lower cost and improving efficiency.

🔹 Publication Date: Published on Nov 26

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.21689
• PDF: https://arxiv.org/pdf/2511.21689
• Project Page: https://research.nvidia.com/labs/lpr/ToolOrchestra/
• Github: https://github.com/NVlabs/ToolOrchestra/

🔹 Models citing this paper:
https://huggingface.co/nvidia/Orchestrator-8B
https://huggingface.co/Mungert/Orchestrator-8B-GGUF
https://huggingface.co/cyankiwi/Orchestrator-8B-AWQ-4bit

Datasets citing this paper:
https://huggingface.co/datasets/nvidia/ToolScale
https://huggingface.co/datasets/victor/ToolScale
https://huggingface.co/datasets/FranckAbgrall/ToolScale

==================================

For more data science resources:
https://t.me/DataScienceT

#ToolOrchestra #ModelOrchestration #ReinforcementLearning #LLMs #AI
Deep Research: A Systematic Survey

📝 Summary:
This survey systematically reviews Deep Research systems that integrate LLMs with external tools to enhance complex problem-solving. It provides a roadmap, key components, optimization techniques, and challenges for these advanced research agents.

🔹 Publication Date: Published on Nov 24

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.02038
• PDF: https://arxiv.org/pdf/2512.02038
• Project Page: https://deep-research-survey.github.io/
• Github: https://github.com/mangopy/Deep-Research-Survey

==================================

For more data science resources:
https://t.me/DataScienceT

#DeepResearch #LLMs #AI #ResearchAgents #SystematicSurvey
Adversarial Confusion Attack: Disrupting Multimodal Large Language Models

📝 Summary:
The Adversarial Confusion Attack systematically disrupts multimodal LLMs, causing incoherent or confidently incorrect outputs. This basic adversarial technique transfers to diverse models, including proprietary ones, potentially hindering AI Agent reliability.

🔹 Publication Date: Published on Nov 25

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2511.20494
• PDF: https://arxiv.org/pdf/2511.20494

==================================

For more data science resources:
https://t.me/DataScienceT

#AdversarialAttack #MultimodalAI #LLMs #AISecurity #AIResearch
1👍1🔥1
SignRoundV2: Closing the Performance Gap in Extremely Low-Bit Post-Training Quantization for LLMs

📝 Summary:
SignRoundV2 is a post-training quantization framework for LLMs. It uses a sensitivity metric for bit allocation and pre-tuning for scales to achieve competitive accuracy even at 2-bit quantization, closing the gap with full-precision models.

🔹 Publication Date: Published on Dec 4

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.04746
• PDF: https://arxiv.org/pdf/2512.04746

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #Quantization #DeepLearning #AI #MachineLearning
SQ-format: A Unified Sparse-Quantized Hardware-friendly Data Format for LLMs

📝 Summary:
The SQ-format is a unified sparse-quantized data format for LLM post-training quantization. It improves accuracy and efficiency balance by combining sparse and low-precision matrix multiplications. This enables better performance and throughput, especially for outlier activations, supporting next...

🔹 Publication Date: Published on Dec 5

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.05409
• PDF: https://arxiv.org/pdf/2512.05409

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #Quantization #SparseML #HardwareAcceleration #AIResearch
1
MemLoRA: Distilling Expert Adapters for On-Device Memory Systems

📝 Summary:
MemLoRA and MemLoRA-V enable efficient on-device memory-augmented AI by equipping small language and vision-language models with specialized, distilled memory adapters. This allows accurate local memory operations and native visual understanding, outperforming larger baselines in text and visual ...

🔹 Publication Date: Published on Dec 4

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.04763
• PDF: https://arxiv.org/pdf/2512.04763

==================================

For more data science resources:
https://t.me/DataScienceT

#OnDeviceAI #LLMs #VLMs #AIAdapters #MemoryAugmentedAI
1
🤖🧠 How to Run and Fine-Tune Kimi K2 Thinking Locally with Unsloth

🗓️ 11 Dec 2025
📚 AI News & Trends

The demand for efficient and powerful large language models (LLMs) continues to rise as developers and researchers seek new ways to optimize reasoning, coding, and conversational AI performance. One of the most impressive open-source AI systems available today is Kimi K2 Thinking, created by Moonshot AI. Through collaboration with Unsloth, users can now fine-tune and ...

#KimiK2Thinking #Unsloth #LLMs #LargeLanguageModels #AI #FineTuning
1
Thinking with Images via Self-Calling Agent

📝 Summary:
sCoT is a novel visual reasoning paradigm that reformulates interleaved multimodal CoT as a language-only CoT with self-calling subagents. It improves reasoning performance and efficiency by avoiding explicit multimodal interleaving and using group-relative policy optimization.

🔹 Publication Date: Published on Dec 9

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.08511
• PDF: https://arxiv.org/pdf/2512.08511
• Github: https://github.com/YWenxi/think-with-images-through-self-calling

==================================

For more data science resources:
https://t.me/DataScienceT

#VisualReasoning #MultimodalAI #LLMs #AIagents #AIResearch
Sliding Window Attention Adaptation

📝 Summary:
Sliding Window Attention Adaptation SWAA allows pretrained LLMs to use efficient sliding window attention for long contexts without retraining. SWAA combines five adaptation methods, with specific synergistic combinations effectively recovering original long-context performance.

🔹 Publication Date: Published on Dec 11

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.10411
• PDF: https://arxiv.org/pdf/2512.10411

🔹 Models citing this paper:
https://huggingface.co/yuyijiong/Qwen3-SWA-adaptation

Datasets citing this paper:
https://huggingface.co/datasets/yuyijiong/LongMemEval_24k

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #SlidingWindowAttention #LongContextAI #NLP #AIResearch
2
Causal Judge Evaluation: Calibrated Surrogate Metrics for LLM Systems

📝 Summary:
CJE improves LLM-as-judge evaluation by fixing statistical issues like uncalibrated scores and poor confidence intervals. It achieves 99% ranking accuracy at 14x lower cost by calibrating a cheaper judge with 5% oracle labels.

🔹 Publication Date: Published on Dec 11

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.11150
• PDF: https://arxiv.org/pdf/2512.11150
• Project Page: https://www.cimolabs.com/cje
• Github: https://github.com/cimo-labs/cje

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #AIEvaluation #MachineLearning #DataScience #NLP
VOYAGER: A Training Free Approach for Generating Diverse Datasets using LLMs

📝 Summary:
Voyager is a novel, training-free method that iteratively generates diverse synthetic datasets from LLMs. It uses determinantal point processes to optimize diversity, significantly outperforming baselines with a 1.5-3x improvement.

🔹 Publication Date: Published on Dec 12

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.12072
• PDF: https://arxiv.org/pdf/2512.12072

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #SyntheticData #DataScience #MachineLearning #AI
2
FiNERweb: Datasets and Artifacts for Scalable Multilingual Named Entity Recognition

📝 Summary:
FiNERweb is a new pipeline that scales multilingual Named Entity Recognition dataset creation to 91 languages using LLMs. It produces 225k high-quality passages, enabling models to achieve comparable or improved zero-shot performance with 19x less data.

🔹 Publication Date: Published on Dec 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.13884
• PDF: https://arxiv.org/pdf/2512.13884
• Github: https://github.com/whoisjones/FiNERweb

==================================

For more data science resources:
https://t.me/DataScienceT

#NER #NLP #LLMs #MultilingualAI #Datasets
1
JustRL: Scaling a 1.5B LLM with a Simple RL Recipe

📝 Summary:
JustRL uses a minimal single-stage RL approach with fixed hyperparameters to achieve state-of-the-art performance on 1.5B reasoning models. It uses less compute and shows stable training, suggesting that complex RL methods for LLMs may be unnecessary and can even hinder exploration.

🔹 Publication Date: Published on Dec 18

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.16649
• PDF: https://arxiv.org/pdf/2512.16649

==================================

For more data science resources:
https://t.me/DataScienceT

#ReinforcementLearning #LLMs #DeepLearning #AIResearch #ModelScaling
1
Hearing to Translate: The Effectiveness of Speech Modality Integration into LLMs

📝 Summary:
This paper benchmarks SpeechLLMs against cascaded systems for speech-to-text translation. It finds cascaded systems are more reliable overall, while SpeechLLMs match them only in select cases. Integrating an LLM is essential for high quality speech translation.

🔹 Publication Date: Published on Dec 18

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.16378
• PDF: https://arxiv.org/pdf/2512.16378
• Github: https://github.com/sarapapi/hearing2translate

==================================

For more data science resources:
https://t.me/DataScienceT

#SpeechTranslation #LLMs #NLP #AIResearch #DeepLearning
1
Understanding Syllogistic Reasoning in LLMs from Formal and Natural Language Perspectives

📝 Summary:
This study explores syllogistic reasoning in LLMs, examining both symbolic inference and natural language understanding. Some models achieve perfect symbolic performance, leading to questions about whether LLMs are becoming more formal reasoning mechanisms.

🔹 Publication Date: Published on Dec 14

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.12620
• PDF: https://arxiv.org/pdf/2512.12620
• Github: https://github.com/XAheli/Logic-in-LLMs

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #SyllogisticReasoning #NaturalLanguageProcessing #AIResearch #FormalLogic
Scaling Laws for Code: Every Programming Language Matters

📝 Summary:
This paper explores scaling laws for multilingual code pre-training, finding interpreted languages benefit more from scaling. It proposes an optimal token allocation strategy for programming languages based on utility and synergy, outperforming uniform distribution.

🔹 Publication Date: Published on Dec 15

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.13472
• PDF: https://arxiv.org/pdf/2512.13472

==================================

For more data science resources:
https://t.me/DataScienceT

#CodeAI #MachineLearning #ProgrammingLanguages #ScalingLaws #LLMs
SWE-EVO: Benchmarking Coding Agents in Long-Horizon Software Evolution Scenarios

📝 Summary:
SWE-EVO is a new benchmark for AI coding agents that evaluates them on long-horizon, multi-step software evolution tasks across many files. It reveals a significant gap in current models abilities, with even top models achieving only 21 percent resolution. This highlights their struggle with sust...

🔹 Publication Date: Published on Dec 20

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.18470
• PDF: https://arxiv.org/pdf/2512.18470

Datasets citing this paper:
https://huggingface.co/datasets/Fsoft-AIC/SWE-EVO

==================================

For more data science resources:
https://t.me/DataScienceT

#AICoding #SoftwareEvolution #Benchmarking #LLMs #AIResearch
2
Scaling Open-Ended Reasoning to Predict the Future

📝 Summary:
This work trains language models for open-ended future prediction using a new dataset synthesized from news. Their OpenForecaster 8B model matches larger proprietary models in accuracy, calibration, and consistency. All resources are open-sourced.

🔹 Publication Date: Published on Dec 31, 2025

🔹 Paper Links:
• arXiv Page: https://arxiv.org/abs/2512.25070
• PDF: https://arxiv.org/pdf/2512.25070
• Project Page: https://www.openforecaster.github.io
• Github: https://github.com/OpenForecaster/scaling-forecasting-training

==================================

For more data science resources:
https://t.me/DataScienceT

#LLMs #FuturePrediction #AI #OpenSourceAI #MachineLearning