Github Top Repositories
12.9K subscribers
379 photos
57 videos
9 files
1.33K links
Top GitHub repositories in one place πŸš€
Explore the best projects in programming, AI, data science, and more.
Download Telegram
This media is not supported in your browser
VIEW IN TELEGRAM
πŸˆβ€β¬› TTT Long Video Generation πŸˆβ€β¬›

▢️ A novel architecture for video generation, adapting the #CogVideoX 5B model by incorporating #TestTimeTraining (TTT) layers.
Adding TTT layers into a pre-trained Transformer enables generating a one-minute clip from text storyboards.
Videos, code & annotations released πŸ’™

πŸ”— Review: https://t.ly/mhlTN
πŸ“„ Paper: arxiv.org/pdf/2504.05298
🌐 Project: test-time-training.github.io/video-dit
πŸ§‘β€πŸ’» Repo: github.com/test-time-training/ttt-video-dit

#AI #VideoGeneration #MachineLearning #DeepLearning #Transformers #TTT #GenerativeAI

πŸ” By: https://t.me/DataScienceN5
Please open Telegram to view this post
VIEW IN TELEGRAM
πŸ‘3πŸ₯°2
πŸš€ The new HQ-SAM (High-Quality Segment Anything Model) has just been added to the Hugging Face Transformers library!

This is an enhanced version of the original SAM (Segment Anything Model) introduced by Meta in 2023. HQ-SAM significantly improves the segmentation of fine and detailed objects, while preserving all the powerful features of SAM β€” including prompt-based interaction, fast inference, and strong zero-shot performance. That means you can easily switch to HQ-SAM wherever you used SAM!

The improvements come from just a few additional learnable parameters. The authors collected a high-quality dataset with 44,000 fine-grained masks from various sources, and impressively trained the model in just 4 hours using 8 GPUs β€” all while keeping the core SAM weights frozen.

The newly introduced parameters include:

* A High-Quality Token
* A Global-Local Feature Fusion mechanism

This work was presented at NeurIPS 2023 and still holds state-of-the-art performance in zero-shot segmentation on the SGinW benchmark.

πŸ“„ Documentation: https://lnkd.in/e5iDT6Tf
🧠 Model Access: https://lnkd.in/ehS6ZUyv
πŸ’» Source Code: https://lnkd.in/eg5qiKC2



#ArtificialIntelligence #ComputerVision #Transformers #Segmentation #DeepLearning #PretrainedModels #ResearchAndDevelopment #AdvancedModels #ImageAnalysis #HQ_SAM #SegmentAnything #SAMmodel #ZeroShotSegmentation #NeurIPS2023 #AIresearch #FoundationModels #OpenSourceAI #SOTA

🌟
https://t.me/DataScienceN
❀2πŸ‘2πŸ”₯1
πŸ”₯ Trending Repository: haystack

πŸ“ Description: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.

πŸ”— Repository URL: https://github.com/deepset-ai/haystack

🌐 Website: https://haystack.deepset.ai

πŸ“– Readme: https://github.com/deepset-ai/haystack#readme

πŸ“Š Statistics:
🌟 Stars: 22.3K stars
πŸ‘€ Watchers: 158
🍴 Forks: 2.3K forks

πŸ’» Programming Languages: Python - HTML

🏷️ Related Topics:
#python #nlp #agent #machine_learning #information_retrieval #ai #transformers #orchestration #pytorch #gemini #question_answering #summarization #agents #semantic_search #rag #gpt_4 #large_language_models #llm #generative_ai #retrieval_augmented_generation


==================================
🧠 By: https://t.me/DataScienceM
πŸ”₯ Trending Repository: LLaMA-Factory

πŸ“ Description: Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)

πŸ”— Repository URL: https://github.com/hiyouga/LLaMA-Factory

🌐 Website: https://llamafactory.readthedocs.io

πŸ“– Readme: https://github.com/hiyouga/LLaMA-Factory#readme

πŸ“Š Statistics:
🌟 Stars: 61.3K stars
πŸ‘€ Watchers: 295
🍴 Forks: 7.4K forks

πŸ’» Programming Languages: Python

🏷️ Related Topics:
#nlp #agent #ai #transformers #moe #llama #gpt #lora #quantization #gemma #fine_tuning #peft #large_language_models #llm #rlhf #instruction_tuning #qlora #qwen #deepseek #llama3


==================================
🧠 By: https://t.me/DataScienceM
πŸ”₯ Trending Repository: mlx-audio

πŸ“ Description: A text-to-speech (TTS), speech-to-text (STT) and speech-to-speech (STS) library built on Apple's MLX framework, providing efficient speech analysis on Apple Silicon.

πŸ”— Repository URL: https://github.com/Blaizzy/mlx-audio

πŸ“– Readme: https://github.com/Blaizzy/mlx-audio#readme

πŸ“Š Statistics:
🌟 Stars: 3.4K stars
πŸ‘€ Watchers: 32
🍴 Forks: 285 forks

πŸ’» Programming Languages: Python - TypeScript

🏷️ Related Topics:
#text_to_speech #transformers #speech_synthesis #speech_recognition #speech_to_text #audio_processing #mlx #multimodal #apple_silicon


==================================
🧠 By: https://t.me/DataScienceM
πŸ”₯ Trending Repository: Megatron-LM

πŸ“ Description: Ongoing research training transformer models at scale

πŸ”— Repository URL: https://github.com/NVIDIA/Megatron-LM

🌐 Website: https://docs.nvidia.com/megatron-core/developer-guide/latest/get-started/quickstart.html

πŸ“– Readme: https://github.com/NVIDIA/Megatron-LM#readme

πŸ“Š Statistics:
🌟 Stars: 15.3K stars
πŸ‘€ Watchers: 174
🍴 Forks: 3.6K forks

πŸ’» Programming Languages: Python

🏷️ Related Topics:
#transformers #model_para #large_language_models


==================================
🧠 By: https://t.me/DataScienceM