👉 A proof I'm not a bot...
My (short) interview to one of the biggest Italian media: AI in 2016, HPC / Quantum and how I created my startup: https://www.linkedin.com/posts/visionarynet_ai-itw25-ai-activity-7381215486115643392-t7an
Thanks for the support (and of course a new paper coming in a few hours)
My (short) interview to one of the biggest Italian media: AI in 2016, HPC / Quantum and how I created my startup: https://www.linkedin.com/posts/visionarynet_ai-itw25-ai-activity-7381215486115643392-t7an
Thanks for the support (and of course a new paper coming in a few hours)
❤17🔥7👏4😍3⚡1
This media is not supported in your browser
VIEW IN TELEGRAM
🎺Visual Grounding RVOS🎺
👉ReferDINO is a strong RVOS model that inherits region-level vision-language alignment from foundational visual grounding models, and is further endowed with pixel-level dense perception & cross-modal spatio-temporal reasoning. Code, Demo & checkpoints💙
👉Review https://t.ly/rOdkP
👉Paper https://lnkd.in/efuAFQdE
👉Project https://lnkd.in/dK3wMZqv
👉Repo https://lnkd.in/d3i2PsNF
👉ReferDINO is a strong RVOS model that inherits region-level vision-language alignment from foundational visual grounding models, and is further endowed with pixel-level dense perception & cross-modal spatio-temporal reasoning. Code, Demo & checkpoints💙
👉Review https://t.ly/rOdkP
👉Paper https://lnkd.in/efuAFQdE
👉Project https://lnkd.in/dK3wMZqv
👉Repo https://lnkd.in/d3i2PsNF
🔥8❤1👏1
This media is not supported in your browser
VIEW IN TELEGRAM
💄Pixel-Perfect Depth (SOTA)💄
👉Pixel-Perfect Depth is a mono-depth estimation model with pixel-space diffusion transformers. New SOTA. Repo under Apache 2.0💙
👉Review https://t.ly/75PGo
👉Paper https://lnkd.in/d8wxFpyY
👉Project https://lnkd.in/dV5HhsqH
👉Repo https://lnkd.in/d9JKFBJq
👉Demo https://lnkd.in/d3wBkKJ9
👉Pixel-Perfect Depth is a mono-depth estimation model with pixel-space diffusion transformers. New SOTA. Repo under Apache 2.0💙
👉Review https://t.ly/75PGo
👉Paper https://lnkd.in/d8wxFpyY
👉Project https://lnkd.in/dV5HhsqH
👉Repo https://lnkd.in/d9JKFBJq
👉Demo https://lnkd.in/d3wBkKJ9
🔥16🤯5❤3
This media is not supported in your browser
VIEW IN TELEGRAM
↗️ TrackVLA++ Visual Tracking↘️
👉TrackVLA++ is a novel Vision-Language-Action model that incorporates spatial reasoning and target identification memory, enabling SOTA performance in both long-horizon and highly crowded tracking scenarios. Model announced💙
👉Review https://t.ly/ruYzc
👉Paper https://arxiv.org/pdf/2510.07134
👉Project pku-epic.github.io/TrackVLA-plus-plus-Web/
👉Repo TBA
👉TrackVLA++ is a novel Vision-Language-Action model that incorporates spatial reasoning and target identification memory, enabling SOTA performance in both long-horizon and highly crowded tracking scenarios. Model announced💙
👉Review https://t.ly/ruYzc
👉Paper https://arxiv.org/pdf/2510.07134
👉Project pku-epic.github.io/TrackVLA-plus-plus-Web/
👉Repo TBA
🔥5🤣1