This media is not supported in your browser
VIEW IN TELEGRAM
ðĨĶ GAN-generated CryptoPunks ðĨĶ
ðA simple (and funny) SN-GAN to generate cryptopunks
ððĒð ðĄðĨðĒð ðĄððŽ:
â Spectral normalization (2018)
â Easy to incorporate into training
â A project by Teddy Koker ðĐ
More: https://bit.ly/35C1rQI
ðA simple (and funny) SN-GAN to generate cryptopunks
ððĒð ðĄðĨðĒð ðĄððŽ:
â Spectral normalization (2018)
â Easy to incorporate into training
â A project by Teddy Koker ðĐ
More: https://bit.ly/35C1rQI
âĪ3ð3ð1ð1
This media is not supported in your browser
VIEW IN TELEGRAM
ðĪŠSEER: self-AI from BILLIONS picðĪŠ
ðMETA + INRIA trained models on billions of random images without any pre-processing or assumptions
ððĒð ðĄðĨðĒð ðĄððŽ:
â Self-supervised on pics from web
â Discovering properties in datasets
â More fair, less biased & less harmful
â Better OOD generalization
â Source code available!
More: https://bit.ly/3vy69dd
ðMETA + INRIA trained models on billions of random images without any pre-processing or assumptions
ððĒð ðĄðĨðĒð ðĄððŽ:
â Self-supervised on pics from web
â Discovering properties in datasets
â More fair, less biased & less harmful
â Better OOD generalization
â Source code available!
More: https://bit.ly/3vy69dd
ðĨ4ð3ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ðēA novel AI-controllable synthesisðē
ðModeling local semantic parts separately and synthesizing images in a compositional way
ððĒð ðĄðĨðĒð ðĄððŽ:
â Structure & texture locally controlled
â Disentanglement between areas
â Fine-grained editing of images
â Extendible via transfer learning
â Just accepted to #CVPR2022
More: https://bit.ly/3IBgkBy
ðModeling local semantic parts separately and synthesizing images in a compositional way
ððĒð ðĄðĨðĒð ðĄððŽ:
â Structure & texture locally controlled
â Disentanglement between areas
â Fine-grained editing of images
â Extendible via transfer learning
â Just accepted to #CVPR2022
More: https://bit.ly/3IBgkBy
ðą3ðĪŊ2âĪ1
This media is not supported in your browser
VIEW IN TELEGRAM
ðĨĢ #AI-Generation with Dream Fields ðĨĢ
ðNeural rendering with multi-modal image and text representations
ððĒð ðĄðĨðĒð ðĄððŽ:
â Aligned image & text models
â 3D from natural language
â No additional data
â D.F. neural-scene
More: https://bit.ly/3Mhwm5D
ðNeural rendering with multi-modal image and text representations
ððĒð ðĄðĨðĒð ðĄððŽ:
â Aligned image & text models
â 3D from natural language
â No additional data
â D.F. neural-scene
More: https://bit.ly/3Mhwm5D
ð10ð1
This media is not supported in your browser
VIEW IN TELEGRAM
ðŠ Mip-NeRF 360 for unbounded scenes ðŠ
ðAn extension of NeRF to overcome the challenges presented by unbounded scenes
ððĒð ðĄðĨðĒð ðĄððŽ:
â Realistic synthesized views
â Intricate/unbounded scenes
â Detailed depth maps
â Mean-squared error -54%
â No code provided ðĨ
More: https://bit.ly/36ZxsD4
ðAn extension of NeRF to overcome the challenges presented by unbounded scenes
ððĒð ðĄðĨðĒð ðĄððŽ:
â Realistic synthesized views
â Intricate/unbounded scenes
â Detailed depth maps
â Mean-squared error -54%
â No code provided ðĨ
More: https://bit.ly/36ZxsD4
ðĪŊ4âĪ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð PINA: personal Neural Avatar ð
ðA novel method to acquire neural avatars from RGB-D videos
ððĒð ðĄðĨðĒð ðĄððŽ:
â A virtual copy of themselves
â Realistic clothing deformations
â Shape & non-rigid deformation
â Avatars from RGB-D sequences
â Creative Commons Zero v1.0
More: https://bit.ly/3HAtRIh
ðA novel method to acquire neural avatars from RGB-D videos
ððĒð ðĄðĨðĒð ðĄððŽ:
â A virtual copy of themselves
â Realistic clothing deformations
â Shape & non-rigid deformation
â Avatars from RGB-D sequences
â Creative Commons Zero v1.0
More: https://bit.ly/3HAtRIh
ð4âĪ1ð1ð1
This media is not supported in your browser
VIEW IN TELEGRAM
ðĶ EfficientVIS: new SOTA for VIS ðĶ
ðSimultaneous classification, segmentation, and tracking multiple object instances in videos
ððĒð ðĄðĨðĒð ðĄððŽ:
â Efficient and fully end-to-end
â Iterative query-video interaction
â First RoI-wise clip-level RT-VIS
â Requires 15Ã fewer epochs
More: https://bit.ly/3KfqurN
ðSimultaneous classification, segmentation, and tracking multiple object instances in videos
ððĒð ðĄðĨðĒð ðĄððŽ:
â Efficient and fully end-to-end
â Iterative query-video interaction
â First RoI-wise clip-level RT-VIS
â Requires 15Ã fewer epochs
More: https://bit.ly/3KfqurN
ð10ðĨ3ð1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð #AI-clips from single frameð
ðMoving objects in #3D while generating a video by a sequence of desired actions
ððĒð ðĄðĨðĒð ðĄððŽ:
â A playable environments
â A single starting imageðĪŊ
â Controllable camera
â Unsupervised learning
More: https://bit.ly/35VDrYO
ðMoving objects in #3D while generating a video by a sequence of desired actions
ððĒð ðĄðĨðĒð ðĄððŽ:
â A playable environments
â A single starting imageðĪŊ
â Controllable camera
â Unsupervised learning
More: https://bit.ly/35VDrYO
âĪ3ð1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð§Kubric: AI dataset generatorð§
ðOpen-source #Python framework for photo-realistic scenes: full control, rich annotations, TBs of fresh data ðĪŊ
ððĒð ðĄðĨðĒð ðĄððŽ:
â Synthetic datasets with GT
â From NeRF to optical flow
â Full control over data
â Ok privacy & licensing
â Apache License 2.0
More: https://bit.ly/3hQCaFs
ðOpen-source #Python framework for photo-realistic scenes: full control, rich annotations, TBs of fresh data ðĪŊ
ððĒð ðĄðĨðĒð ðĄððŽ:
â Synthetic datasets with GT
â From NeRF to optical flow
â Full control over data
â Ok privacy & licensing
â Apache License 2.0
More: https://bit.ly/3hQCaFs
ðĨ6ð1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ðŠÂĩTransfer for enormous NNs ðŠ
ðMicrosoft unveils how to tune enormous neural networks
ððĒð ðĄðĨðĒð ðĄððŽ:
â New HP tuning: ÂĩTransfer
â Zero-shot transfer to full-model
â Outperforming BERT-large
â Outperforming 6.7B GPT-3
â Code under MIT license
More: https://bit.ly/3qc37Ij
ðMicrosoft unveils how to tune enormous neural networks
ððĒð ðĄðĨðĒð ðĄððŽ:
â New HP tuning: ÂĩTransfer
â Zero-shot transfer to full-model
â Outperforming BERT-large
â Outperforming 6.7B GPT-3
â Code under MIT license
More: https://bit.ly/3qc37Ij
ðĨ2ðĪŊ2âĪ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð§Semantic via only text supervisionð§
ðGroupViT with a text encoder on a large-scale image-text dataset: semantic with any pixel-level annotations in training!
ððĒð ðĄðĨðĒð ðĄððŽ:
â Hierarc. Grouping Vision Transf.
â Additional text encoder
â NO pixel-level annotations
â Semantic-seg task via zero-shot
â Source code available soon
More:https://bit.ly/3hPGeWr
ðGroupViT with a text encoder on a large-scale image-text dataset: semantic with any pixel-level annotations in training!
ððĒð ðĄðĨðĒð ðĄððŽ:
â Hierarc. Grouping Vision Transf.
â Additional text encoder
â NO pixel-level annotations
â Semantic-seg task via zero-shot
â Source code available soon
More:https://bit.ly/3hPGeWr
ð6ðĨ°1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
â4D-Net: Lidar + RGB synchronizationâ
ðGoogle unveils 4D-Net to combine 3D LiDAR and onboard RGB camera
ððĒð ðĄðĨðĒð ðĄððŽ:
â Point clouds/images in time
â Fusing multiple modalities in 4D
â Novel sampling for 3D P.C. in time
â New SOTA for 3D detection
More: https://bit.ly/3hZCFwN
ðGoogle unveils 4D-Net to combine 3D LiDAR and onboard RGB camera
ððĒð ðĄðĨðĒð ðĄððŽ:
â Point clouds/images in time
â Fusing multiple modalities in 4D
â Novel sampling for 3D P.C. in time
â New SOTA for 3D detection
More: https://bit.ly/3hZCFwN
ð12ðĨ2ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð New SOTA in video synthesis! ð
ðSnap unveils a novel multimodal video generation framework via text/images
ððĒð ðĄðĨðĒð ðĄððŽ:
â Multimodal video generation
â Bidirectional transformer
â Video token with self-learn.
â Text augmentation for robustness
â Longer sequence synthesis
More: https://bit.ly/3hZLXsG
ðSnap unveils a novel multimodal video generation framework via text/images
ððĒð ðĄðĨðĒð ðĄððŽ:
â Multimodal video generation
â Bidirectional transformer
â Video token with self-learn.
â Text augmentation for robustness
â Longer sequence synthesis
More: https://bit.ly/3hZLXsG
ðĪŊ4ð1ðĨ1ð1
This media is not supported in your browser
VIEW IN TELEGRAM
ð StyelNeRF source code is out ð
ð3D consistent photo-realistic image synthesis
ððĒð ðĄðĨðĒð ðĄððŽ:
â NeRF + style generator
â 3D consistency for HD image
â Novel regularization loss
â Camera control on styles
More: https://bit.ly/3t5xC49
ð3D consistent photo-realistic image synthesis
ððĒð ðĄðĨðĒð ðĄððŽ:
â NeRF + style generator
â 3D consistency for HD image
â Novel regularization loss
â Camera control on styles
More: https://bit.ly/3t5xC49
ðĨ4ðĨ°1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ðĶCLD-based generative #AI by #NvidiaðĶ
ðNvidia unveils a novel critically-damped Langevin diffusion (CLD) for synthetic data
ððĒð ðĄðĨðĒð ðĄððŽ:
â A novel diffusion process for SGMs
â Novel score matching obj. for CLD
â Hybrid denoising score matching
â Efficient sampling from CLD model
â Source code under a specific license
More: https://bit.ly/35MToBe
ðNvidia unveils a novel critically-damped Langevin diffusion (CLD) for synthetic data
ððĒð ðĄðĨðĒð ðĄððŽ:
â A novel diffusion process for SGMs
â Novel score matching obj. for CLD
â Hybrid denoising score matching
â Efficient sampling from CLD model
â Source code under a specific license
More: https://bit.ly/35MToBe
ðĨ2ðĪĐ2ð1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ðļUFO: segmentation @140+ FPSðļ
ðUnified Transformer Framework for Co-Segmentation, Co-Saliency & Salient Object Detection. All in one!
ððĒð ðĄðĨðĒð ðĄððŽ:
â Unified framework for co-segmentation
â Co-segmentation, co-saliency, saliency
â Block for long-range dependencies
â Able to reach for 140 FPS in inference
â The new SOTA on multiple datasets
â Source code under MIT License
More: https://bit.ly/3KLd9b9
ðUnified Transformer Framework for Co-Segmentation, Co-Saliency & Salient Object Detection. All in one!
ððĒð ðĄðĨðĒð ðĄððŽ:
â Unified framework for co-segmentation
â Co-segmentation, co-saliency, saliency
â Block for long-range dependencies
â Able to reach for 140 FPS in inference
â The new SOTA on multiple datasets
â Source code under MIT License
More: https://bit.ly/3KLd9b9
ðĨ6ð1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð Multi-GANs fashion ð
ðGlobal GAN blended with other GANs for faces, shoes, etc.
ððĒð ðĄðĨðĒð ðĄððŽ:
â Multi-GAN framework
â Several generators
â Free of artifacts
â Full-body generation
â Humans, 1024x1024
More: https://bit.ly/37mfOte
ðGlobal GAN blended with other GANs for faces, shoes, etc.
ððĒð ðĄðĨðĒð ðĄððŽ:
â Multi-GAN framework
â Several generators
â Free of artifacts
â Full-body generation
â Humans, 1024x1024
More: https://bit.ly/37mfOte
ðĨ2ð2âĪ1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð§ FLAG: #3D Avatar Generation ð§
ðA flow-based generative model of the 3D human body from sparse observations.
ððĒð ðĄðĨðĒð ðĄððŽ:
â FLow-based Avatar Generative
â Conditional distro of body pose
â Exact pose likelihood process
â Invertibility -> oracle latent code
More: https://bit.ly/3CQpk3p
ðA flow-based generative model of the 3D human body from sparse observations.
ððĒð ðĄðĨðĒð ðĄððŽ:
â FLow-based Avatar Generative
â Conditional distro of body pose
â Exact pose likelihood process
â Invertibility -> oracle latent code
More: https://bit.ly/3CQpk3p
ð2ðĨ1ðĪŊ1
This media is not supported in your browser
VIEW IN TELEGRAM
ð Dancing in the wild with StyleGAN ð
ðStyleGAN-based animations for AR/VR apps
ððĒð ðĄðĨðĒð ðĄððŽ:
â Video based motion retargeting
â A StyleGAN architecture based
â Novel explicit motion representation
â SOTA qualitatively & quantitatively
More: https://bit.ly/3CZbL1W
ðStyleGAN-based animations for AR/VR apps
ððĒð ðĄðĨðĒð ðĄððŽ:
â Video based motion retargeting
â A StyleGAN architecture based
â Novel explicit motion representation
â SOTA qualitatively & quantitatively
More: https://bit.ly/3CZbL1W
ð6ðĪŊ3ðĨ°2
This media is not supported in your browser
VIEW IN TELEGRAM
ðŠTensoRF: the 4D evolution of NeRF ðŠ
ðTensoRF, a novel radiance fields via 4D-tensor: 3D voxel grid with per-voxel multi-channel feats.
ððĒð ðĄðĨðĒð ðĄððŽ:
â VM decomposition technique
â Low-rank tensor factorization
â Lower memory footprint (speed)
â TensoRF is the new SOTA in R.F.
â Code under the MIT License
More: https://bit.ly/3qffZgI
ðTensoRF, a novel radiance fields via 4D-tensor: 3D voxel grid with per-voxel multi-channel feats.
ððĒð ðĄðĨðĒð ðĄððŽ:
â VM decomposition technique
â Low-rank tensor factorization
â Lower memory footprint (speed)
â TensoRF is the new SOTA in R.F.
â Code under the MIT License
More: https://bit.ly/3qffZgI
ð2ðĨ1