MiniMax-AI/MiniMax-M1
MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model.
Language: Python
#large_language_models #llm #minimax_m1 #reasoning_models
Stars: 328 Issues: 3 Forks: 9
https://github.com/MiniMax-AI/MiniMax-M1
MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model.
Language: Python
#large_language_models #llm #minimax_m1 #reasoning_models
Stars: 328 Issues: 3 Forks: 9
https://github.com/MiniMax-AI/MiniMax-M1
GitHub
GitHub - MiniMax-AI/MiniMax-M1: MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model.
MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model. - MiniMax-AI/MiniMax-M1
β€2
NVlabs/Long-RL
Long-RL: Scaling RL to Long Sequences
Language: Python
#efficient_ai #large_language_models #long_sequence #multi_modality #reinforcement_learning #sequence_parallelism
Stars: 301 Issues: 2 Forks: 3
https://github.com/NVlabs/Long-RL
Long-RL: Scaling RL to Long Sequences
Language: Python
#efficient_ai #large_language_models #long_sequence #multi_modality #reinforcement_learning #sequence_parallelism
Stars: 301 Issues: 2 Forks: 3
https://github.com/NVlabs/Long-RL
GitHub
GitHub - NVlabs/Long-RL: Long-RL: Scaling RL to Long Sequences
Long-RL: Scaling RL to Long Sequences. Contribute to NVlabs/Long-RL development by creating an account on GitHub.
Relaxed-System-Lab/Flash-Sparse-Attention
ππ Efficient implementations of Native Sparse Attention
Language: Python
#kernels #large_language_models #machine_learning_systems
Stars: 237 Issues: 1 Forks: 3
https://github.com/Relaxed-System-Lab/Flash-Sparse-Attention
ππ Efficient implementations of Native Sparse Attention
Language: Python
#kernels #large_language_models #machine_learning_systems
Stars: 237 Issues: 1 Forks: 3
https://github.com/Relaxed-System-Lab/Flash-Sparse-Attention
GitHub
GitHub - Relaxed-System-Lab/Flash-Sparse-Attention: ππ Efficient implementations of Native Sparse Attention
ππ Efficient implementations of Native Sparse Attention - Relaxed-System-Lab/Flash-Sparse-Attention
pengzhangzhi/Open-dLLM
The most open diffusion language model for code generation β releasing pretraining, evaluation, inference, and checkpoints.
Language: Python
#diffusion_models #large_language_models
Stars: 159 Issues: 3 Forks: 5
https://github.com/pengzhangzhi/Open-dLLM
The most open diffusion language model for code generation β releasing pretraining, evaluation, inference, and checkpoints.
Language: Python
#diffusion_models #large_language_models
Stars: 159 Issues: 3 Forks: 5
https://github.com/pengzhangzhi/Open-dLLM
GitHub
GitHub - pengzhangzhi/Open-dLLM: The most open diffusion language model for code generation β releasing pretraining, evaluationβ¦
The most open diffusion language model for code generation β releasing pretraining, evaluation, inference, and checkpoints. - pengzhangzhi/Open-dLLM