#python #class_activation_maps #deep_learning #grad_cam #interpretability #interpretable_ai #interpretable_deep_learning #pytorch #score_cam #vision_transformers #visualizations
https://github.com/jacobgil/pytorch-grad-cam
https://github.com/jacobgil/pytorch-grad-cam
GitHub
GitHub - jacobgil/pytorch-grad-cam: Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification…
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more. - jacobgil/pytorch-grad-cam
#other #attention_mechanism #attention_mechanisms #awesome_list #computer_vision #deep_learning #detr #papers #self_attention #transformer #transformer_architecture #transformer_awesome #transformer_cv #transformer_models #transformer_with_cv #transformers #vision_transformer #visual_transformer #vit
https://github.com/cmhungsteve/Awesome-Transformer-Attention
https://github.com/cmhungsteve/Awesome-Transformer-Attention
GitHub
GitHub - cmhungsteve/Awesome-Transformer-Attention: An ultimately comprehensive paper list of Vision Transformer/Attention, including…
An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites - cmhungsteve/Awesome-Transformer-Attention
#python #classification #computer_vision #object_detection #pytorch #self_supervised_learning #transformers #vision_transformer
https://github.com/alibaba/EasyCV
https://github.com/alibaba/EasyCV
GitHub
GitHub - alibaba/EasyCV: An all-in-one toolkit for computer vision
An all-in-one toolkit for computer vision. Contribute to alibaba/EasyCV development by creating an account on GitHub.
#python #deep_learning #deep_learning_library #image_captioning #multimodal_datasets #multimodal_deep_learning #salesforce #vision_and_language #vision_framework #vision_language_pretraining #vision_language_transformer #visual_question_anwsering
https://github.com/salesforce/LAVIS
https://github.com/salesforce/LAVIS
GitHub
GitHub - salesforce/LAVIS: LAVIS - A One-stop Library for Language-Vision Intelligence
LAVIS - A One-stop Library for Language-Vision Intelligence - salesforce/LAVIS
#python #computer_vision #convolutional_networks #embedding_vectors #embeddings #feature_extraction #feature_vector #image_processing #image_retrieval #machine_learning #milvus #pipeline #towhee #transformer #unstructured_data #video_processing #vision_transformer #vit
https://github.com/towhee-io/towhee
https://github.com/towhee-io/towhee
GitHub
GitHub - towhee-io/towhee: Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast.
Towhee is a framework that is dedicated to making neural data processing pipelines simple and fast. - towhee-io/towhee
#python #chatgpt #clip #deep_learning #gpt #hacktoberfest #hnsw #information_retrieval #knn #large_language_models #machine_learning #machinelearning #multi_modal #natural_language_processing #search_engine #semantic_search #tensor_search #transformers #vector_search #vision_language #visual_search
https://github.com/marqo-ai/marqo
https://github.com/marqo-ai/marqo
GitHub
GitHub - marqo-ai/marqo: Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai - marqo-ai/marqo
#cplusplus #artificial_intelligence #computer_vision #document #document_analysis #document_intelligence #document_recognition #document_understanding #documentai #end_to_end_ocr #multimodal #multimodal_deep_learning #ocr #scene_text_detection #scene_text_detection_recognition #scene_text_recognition #text_detection #text_recognition #vision_language #vision_language_model #vision_language_transformer
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
https://github.com/AlibabaResearch/AdvancedLiterateMachinery
GitHub
GitHub - AlibabaResearch/AdvancedLiterateMachinery: A collection of original, innovative ideas and algorithms towards Advanced…
A collection of original, innovative ideas and algorithms towards Advanced Literate Machinery. This project is maintained by the OCR Team in the Language Technology Lab, Tongyi Lab, Alibaba Group. ...
👍1
#python #augmix #convnext #distributed_training #dual_path_networks #efficientnet #image_classification #imagenet #maxvit #mixnet #mobile_deep_learning #mobilenet_v2 #mobilenetv3 #nfnets #normalization_free_training #pretrained_models #pretrained_weights #pytorch #randaugment #resnet #vision_transformer_models
PyTorch Image Models (`timm`) is a comprehensive library that includes a wide range of state-of-the-art image models, layers, utilities, optimizers, and training scripts. Here are the key benefits `timm` offers over 300 pre-trained models from various families like Vision Transformers, ResNets, EfficientNets, and more, allowing you to choose the best model for your task.
- **Pre-trained Weights** You can easily extract features at different levels of the network using `features_only=True` and `out_indices`, making it versatile for various applications.
- **Optimizers and Schedulers** It provides several augmentation techniques like AutoAugment, RandAugment, and regularization methods like DropPath and DropBlock to enhance model performance.
- **Reference Training Scripts**: Included are high-performance training, validation, and inference scripts that support multiple GPUs and mixed-precision training.
Overall, `timm` simplifies the process of working with deep learning models for image tasks by providing a unified interface and extensive tools for training and evaluation.
https://github.com/huggingface/pytorch-image-models
PyTorch Image Models (`timm`) is a comprehensive library that includes a wide range of state-of-the-art image models, layers, utilities, optimizers, and training scripts. Here are the key benefits `timm` offers over 300 pre-trained models from various families like Vision Transformers, ResNets, EfficientNets, and more, allowing you to choose the best model for your task.
- **Pre-trained Weights** You can easily extract features at different levels of the network using `features_only=True` and `out_indices`, making it versatile for various applications.
- **Optimizers and Schedulers** It provides several augmentation techniques like AutoAugment, RandAugment, and regularization methods like DropPath and DropBlock to enhance model performance.
- **Reference Training Scripts**: Included are high-performance training, validation, and inference scripts that support multiple GPUs and mixed-precision training.
Overall, `timm` simplifies the process of working with deep learning models for image tasks by providing a unified interface and extensive tools for training and evaluation.
https://github.com/huggingface/pytorch-image-models
GitHub
GitHub - huggingface/pytorch-image-models: The largest collection of PyTorch image encoders / backbones. Including train, eval…
The largest collection of PyTorch image encoders / backbones. Including train, eval, inference, export scripts, and pretrained weights -- ResNet, ResNeXT, EfficientNet, NFNet, Vision Transformer (V...
#python #chinese #clip #computer_vision #contrastive_loss #coreml_models #deep_learning #image_text_retrieval #multi_modal #multi_modal_learning #nlp #pretrained_models #pytorch #transformers #vision_and_language_pre_training #vision_language
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
This project is about a Chinese version of the CLIP (Contrastive Language-Image Pretraining) model, trained on a large dataset of Chinese text and images. Here’s what you need to know This model helps you quickly perform tasks like calculating text and image features, cross-modal retrieval (finding images based on text or vice versa), and zero-shot image classification (classifying images without any labeled examples).
- **Ease of Use** The model has been tested on various datasets and shows strong performance in zero-shot image classification and cross-modal retrieval tasks.
- **Resources**: The project includes pre-trained models, training and testing codes, and detailed tutorials on how to use the model for different tasks.
Overall, this project makes it easy to work with Chinese text and images using advanced AI techniques, saving you time and effort.
https://github.com/OFA-Sys/Chinese-CLIP
GitHub
GitHub - OFA-Sys/Chinese-CLIP: Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation. - OFA-Sys/Chinese-CLIP