ML Research Hub
32.8K subscribers
4.44K photos
272 videos
23 files
4.8K links
Advancing research in Machine Learning โ€“ practical insights, tools, and techniques for researchers.

Admin: @HusseinSheikho || @Hussein_Sheikho
Download Telegram
๐ŸŒŸ MInference 1.0 by Microsoft pre-release

In anticipation of the upcoming ICML 2024 (Vienna, July 21-27, 2024), Microsoft has published the results of a study from the MInference project. This method allows you to speed up the processing of long sequences due to sparse calculations and the use of unique templates in matrices.
The MInference technique does not require changes in pre-training settings.

Microsoft researchers' synthetic tests of the method on the LLaMA-3-1M, GLM4-1M, Yi-200K, Phi-3-128K, and Qwen2-128K models show up to a 10x reduction in latency and prefill errors on the A100 while maintaining accuracy.

๐ŸŸก Discuss at Huggingface
๐Ÿ–ฅ GitHub
๐ŸŸก Arxiv
๐ŸŸก MInference 1.0 project page

https://t.me/DataScienceT โœ…
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘2
๐ŸŒŸ Arcee Agent 7B - a new model based on Qwen2-7B

Arcee Agent 7B is superior to GPT-3.5-Turbo, and many other models in writing and interpreting code.
Arcee Agent 7B is especially suitable for those wishing to implement complex AI solutions without the computational expense of large language models.

And yes, there are also quantized GGUF versions of Arcee Agent 7B.

๐Ÿค— Hugging Face

https://t.me/DataScienceT โœ…
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘2
๐Ÿ†• Meet Kolors - a diffusion model for image generation with a focus on photorealism

Kolors is a large diffusion model published recently by the Kuaishou Kolors team.

Kolors has been trained on billions of text-to-image pairs and shows excellent results in generating complex photorealistic images.

As evaluated by 50 independent experts, the Kolors model generates more realistic and beautiful images than Midjourney-v6, Stable Diffusion 3, DALL-E 3 and other models

๐ŸŸก Kolors page
๐ŸŸก Try
๐Ÿ–ฅ GitHub

https://t.me/DataScienceT โœ…
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘3โค1
https://t.me/major/start?startapp=418788114

๐Ÿ‘‘ Telegram's official bot to win Telegram stars  โญ๏ธ

You can convert stars into money or buy ads or services in Telegram
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘1
โšก๏ธ Test-Time Training RNN (TTT) is a fundamentally new method of machine learning.

TTT is a technique that allows artificial intelligence models to adapt and learn while in use, rather than just during pre-training.
The main advantage of TTT is that it can efficiently process long contexts (large amounts of input data) without significantly increasing the computational cost.

The researchers conducted experiments on various datasets, including books, and found that TTT often outperformed traditional methods.
In comparative benchmarks with other popular machine learning methods such as transformers and recurrent neural networks, TTT was found to perform better on some tasks.

This revolutionary method will bring us closer to creating more flexible and efficient artificial intelligence models that can better adapt to new data in real time.

Adaptations of the method have been published on Github:

- adaptation for Pytorch
- adaptation to JAX

๐ŸŸก Arxiv
๐Ÿ–ฅ GitHub for Pytorch [Stars: 277 | Issues: 3 | Forks: 12 ]
๐Ÿ–ฅ GitHub for Jax [ Stars: 129 | Issues: 1 | Forks: 6 ]


#Pytorch #Jax #TTT #LLM #Training

https://t.me/DataScienceT โšซ๏ธ
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘2
Forwarded from Learn Python Hub
https://t.me/major/start?startapp=418788114

๐Ÿ‘‘ Telegram's official bot to win Telegram stars  โญ๏ธ

You can convert stars into money or buy ads or services in Telegram
Please open Telegram to view this post
VIEW IN TELEGRAM
This media is not supported in your browser
VIEW IN TELEGRAM
๐ŸŒŸ Vico is an implementation of a technique that allows you to achieve greater accuracy in generating composite videos.

Vico is a no-training framework that analyzes how individual tokens from prompt input tokens affect the generated video, and adjusts the model to prevent dominance by considering all prompt words equally.

To do this, Vico builds a spatio-temporal attention graph, with which it evaluates and adjusts the representation of all input concepts in the video.

๐Ÿ–ฅ Local launch of inference without UI (with Videocrafterv2)

git clone https://github.com/Adamdad/vico.git
pip install diffusers==0.26.3
git lfs install
git clone https://huggingface.co/adamdad/videocrafterv2_diffusers
export PYTHONPATH="$PWD"
python videocrafterv2_vico.py \
--prompts XXX \
--unet_path $PATH_TO_VIDEOCRAFTERV2 \
--attribution_mode "latent_attention_flow_st_soft"


๐Ÿ–ฅ GitHub [Stars: 19 | Issues: 0 | Forks: 0 ]
๐ŸŸก Project page
๐ŸŸก Arxiv

#T2V #Framework #ML

https://t.me/DataScienceT โœ…
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘1
โšก๏ธ NOW! Only today-tomorrow I am looking for 10 people for the experiment!

By the end of the month I will try to bring them to a good income. No cheating!

๐Ÿ’ธI bet a large sum of money that by the end of the month will make 10 unknown people teach to make money on trading. I will not take money. I will not sell anything.

I just sincerely want to help you become more successful and win your bet!

If you want to take part - write me in DM!

๐Ÿ‘‰CLICK HERE๐Ÿ‘ˆ
๐Ÿ‘‰CLICK HERE๐Ÿ‘ˆ
๐Ÿ‘‰CLICK HERE๐Ÿ‘ˆ
๐ŸŒŸ MiraData: Large, long-duration video dataset with structured annotations.

When training generative models, the training dataset plays an important role in the quality of reference of ready-made models.
One of the good sources can be MiraData from Tencent - a ready-made dataset with a total video duration of 16 thousand hours, designed for training models for generating text in videos. It includes long videos (average 72.1 seconds) with high motion intensity and detailed structured annotations (average 318 words per video).

To assess the quality of the dataset, a system of MiraBench benchmarks was even specially created, consisting of 17 metrics that evaluate temporal consistency, movement in the frame, video quality, and other parameters. According to their results, MiroData outperforms other well-known datasets available in open sources, which mainly consist of short videos with floating quality and short descriptions.

๐ŸŸก Project page
๐ŸŸก Arxiv
๐Ÿค— Hugging Face
๐Ÿ–ฅ GitHub

#Text2Video #Dataset #ML

https://t.me/DataScienceT โญ๏ธ
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘2โค1
๐ŸŒŸ Mamba Vision: An effective alternative to transformers for computer vision

Mamba Vision is an implementation of the Mamba architecture using Selective State Space Models (SSM) in image processing from Nvidia Lab.

MambaVision demonstrates more efficient use of computing resources compared to traditional transformer-based architectures (VIT and Swin), and the use of SSM opens up new ways of extracting and processing visual features. The proposed architecture shows good scalability, maintaining efficiency as the model size increases.
MambaVision is applicable to a variety of computer vision tasks, including image classification and semantic segmentation.

The project is in its early stages and its effectiveness on real-world computer vision tasks has yet to be fully assessed.
At the moment, it has only been used in the image classification task.

๐Ÿ–ผ Family of MambaVision Pretrained (ImageNet-1K) models (direct download from Google Drive):

MambaVision-T (32M)
MambaVision-T2 (35M)
MambaVision-S (50M)
MambaVision-B (98M)
MambaVision-L (228M)
MambaVision-L2 (241M)

โš ๏ธ Licensing:

For non-commercial projects: CC-BY-NC-SA-4.0
For commercial use: request via form

๐Ÿ–ฅ Github
๐ŸŸก Arxiv

#MambaVision #ML

https://t.me/DataScienceT โญ๏ธ
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
โค2๐Ÿ‘2๐Ÿ†2
Multimodal contrastive learning for spatial gene expression prediction using histology images

๐Ÿ–ฅ Github: https://github.com/modelscope/data-juicer

๐Ÿ“• Paper: https://arxiv.org/abs/2407.08583v1

๐Ÿš€ Dataset: https://paperswithcode.com/dataset/coco

https://t.me/DataScienceT โญ๏ธ
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘2
Forwarded from ๐Ÿณ
This media is not supported in your browser
VIEW IN TELEGRAM
Big Wins happening here at @whale!

๐ŸŽฎ 1000+ games and ัrazy multiplayer in the new CrashX Game!

Do you want to try it?
๐Ÿ‘€

1๏ธโƒฃSign up @whale today!
2๏ธโƒฃDeposit on BTC, USDT, TON, CELO and NOT.
3๏ธโƒฃSpin, bet, and triumph ๐Ÿ’ฐ

Become a winner easily and quickly with @whale! ๐Ÿณ
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘6๐Ÿ˜1๐Ÿ†1
๐ŸŒŸ DG-Mesh: Constructing high-quality polygonal meshes from monocular video.

DG-Mesh reconstructs a high-quality dynamic 3D vertex-matched mesh from monocular video. The pipeline uses 3D Gaussian wavelets to represent dynamic scenes and differentiable algorithms to construct polygons.

DG-Mesh allows you to track the movement of vertices, simplifying the texturing of dynamic objects.
The method is memory efficient and fully differentiable, allowing optimization of the target object's 3D mesh directly.

The Github repository contains code for local training using datasets:

- D-NeRF
- DG-Mesh
- NeuralActor
- Custom dataset , shot on Iphone 14 Pro and processed in Record3D, RealityCheck and masked in DEVA.

๐Ÿ–ฅ Local launch:

conda create -n dg-mesh python=3.9
conda activate dg-mesh
conda install pytorch torchvision torcaudio pytorch-cuda=11.8 -c pytorch -c nvidia

#Install nvdiffrast
pip install git+https://github.com/NVlabs/tiny-cuda-nn#subdirectory=bindings/torch
pip install git+https://github.com/NVlabs/nvdiffrast/

# Install pytorch3d
export FORCE_CUDA=1
conda install -c fvcore -c iopath -c conda-forge fvcore iopath -y
pip install "git+https://github.com/facebookresearch/pytorch3d.git"

# Clone this repository
git clone https://github.com/Isabella98Liu/DG-Mesh.git
cd DG-Mesh

# Install submodules
pip install dgmesh/submodules/diff-gaussian-rasterization
pip install dgmesh/submodules/simple-knn

# Install other dependencies
pip install -r requirements.txt


๐ŸŸก Project page
๐Ÿ–ฅ GitHub
๐ŸŸก Arxiv

#Video2Mesh #3D #ML #NeRF

https://t.me/DataScienceT โญ๏ธ
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
๐Ÿ‘4โค1