🚨🚨🚨Take part in the AI Coach: Fitness challenge and the Low Power Computer Vision Challenge @ CVPR 2026
🎯Both challenges use the Qualcomm Exercise Video Dataset (QEVD) dataset.
👉Quick start guides and sample solutions: apratimbh.github.io/whatandwhen/
@cvprconference.bsky.social
Posts by Apratim Bhattacharyya
📣📣📣Our team at Qualcomm AI Research is hiring Research Interns for Summer 2026 in Toronto to work on multi-modal LLMs and embodied AI.
👉Apply here:
1) Embodied AI:
qualcomm.wd12.myworkdayjobs.com/External/job...
2) Multi-modal LLMs:
🚨Submit by 1st May @cvprconference.bsky.social: extended abstracts on streaming vision-language models, real-time activity understanding, grounding, ego-centric video understanding, language and robot learning. Contributions are encouraged to include a demo!
👉Details: varworkshop.github.io/calls/
🚨🚨🚨 We are now accepting submissions!
Call for Participation @cvprconference.bsky.social: Multi-Modal LLMs - prepare to engage in a dynamic, face-to-face conversation with a real human user!
Details: varworkshop.github.io/challenges/
🚨🚨🚨The winning teams will receive a prize and a contributed talk.
P.S. GPT-4o does not do too well.
Call for Participation: We're excited to announce a challenge focused on developing AI assistants that can guide users through workout sessions with intelligent feedback!
🚨The winning teams will receive a prize along with a contributed talk. 🚨
Website: varworkshop.github.io/challenges/
🚨Submission are now open!
Call for Papers and Demos @cvprconference.bsky.social: on topics such as streaming vision-language models, real-time activity understanding, grounding, ego-centric video understanding, language and robot learning. Contributions are encouraged to include a demo!
Link: varworkshop.github.io/calls/
Join us at the @cvprconference.bsky.social Workshop on Vision-based Assistants in the Real-world (VAR) and tackle one of AI's biggest challenges: building systems that can comprehend and reason about dynamic, real-world scenes.
Workshop Page: varworkshop.github.io
By popular demand, we are extending #CVPR2025 coverage to Bluesky. Stay tuned!
Accepted to CVPR2025 🥳🥳
#CVPR2025
🚨We present in "Enhancing Hallucination Detection through Noise Injection" [https://arxiv.org/pdf/2502.03799] an efficient approach to detect hallucinations in LLMs, within a Bayesian framework.
TL; DR - We use noise injection to capture both epistemic and aleatoric uncertainty!
Join us at the CVPR 2025 Workshop on Vision-based Assistants in the Real-world (VAR) and tackle one of AI's biggest challenges: building systems that can comprehend and reason about dynamic, real-world scenes.
Workshop Page: varworkshop.github.io
🚨Check out our new work on distilling reasoning skills from LLMs into efficient driving policies, to deal with critical "long-tail" scenarios.
arXiv: arxiv.org/abs/2501.09757
🚨 The code for our NeurIPS 2024 (D&B track) paper: ClevrSkills: Compositional Language And Visual Understanding in Robotics (arxiv.org/abs/2411.09052), is now available.
GitHub Repo: github.com/Qualcomm-AI-...
Dataset Page: www.qualcomm.com/developer/so...