Brave New Ideas for Video Understanding

CVPR'18 Workshop:
Brave New Ideas for Video Understanding

Together with the Computer Vision and Pattern Recognition (CVPR) 2018.

Description of the workshop and its relevance

In the late years Deep Learning has been a great force of change on most computer vision tasks. In video analysis problems, however, such as action recognition and detection, motion analysis and tracking, shallow architectures remain surprisingly competitive. What is the reason for this conundrum? Larger datasets are part of the solution. The recently proposed Sports1M and Kinetics helped recently in the realistic training of large motion networks. Still, the breakthrough has not yet arrived.

Assuming that the recently proposed video datasets are large enough for training deep networks for video, another likely culprit for the standstill in video analysis is the capacity of the existing deep models. More specifically, the existing deep networks for video analysis might not be sophisticated enough to address the complexity of motion information. This makes sense, as videos introduce an exponential complexity as compared to static images. Unfortunately, state-of-the-art motion representation models are extensions of existing image representations rather than motion dedicated ones. Brave, new and motion-specific representations are likely to be needed for a breakthrough in video analysis.

The goal of this workshop is to bring together researchers from the broad area of video analysis to discuss problem statements, evaluation metrics, and benchmarks that will spur disruptive progress in the field of video understanding. The workshop will include a series of invited talks by leading researchers in this area as well as oral and poster presentations of accepted papers.

Calling papers for brave new ideas

Submissions will be in the form of short non- anonymous papers and will consist of a maximum of 4 pages (excluding references). Submissions must represent new work, i.e., work that has not been previously published or accepted for publication. However, papers that expand previous related work by the authors and papers that have appeared on non peer-reviewed websites (such as arXiv) or that have been presented at workshops (i.e., venues that do not have a publication proceedings) are acceptable. Accepted papers will be presented as posters or contributed talks. Authors of accepted papers will be asked to post their submissions on arXiv. The workshop website will provide links to the accepted papers on arXiv. Accepted papers will be considered non- archival, and may be submitted elsewhere (modified or not).

Expert speakers

To kickstart the discussion we will have several influential speakers.


The workshop focuses on video representations related, but not limited, to the following topics:

- Influence of motion in object recognition, object affordance, scene understanding.
- Object and optical flow
- Motion prediction, causal reasoning and forecasting
- Event and action recognition
- Spatio-temporal action localization
- Modeling human motion in videos and video streams
- Motion segmentation and saliency
- Tracking of objects in space and time
- Unsupervised action, actom discovery using ego motion
- Applications of motion understanding and video dynamics in sports, healthcare, autonomous driving, driver assistance and robotics


Date: TBD.

Time Event Description
8.45 - 9.00 Welcome to the workshop Information
9.00 - 10.00 Invited speaker 1
To be announced
Talk 1
10.00 - 10.20 Break Coffee
10.20 - 11.20 Invited speaker 2
To be announced
Talk 2
11.25 - 11.45 To be announced Oral Presentation 1
11.45 - 13.30 Lunch (on your own) Poster session
13.30 - 14.15 Invited speaker 3
To be announced
Talk 3
14.15 - 14.30 To be announced Oral Presentation 2
14.30 - 14.45 Break Coffee
14.45 - 15.30 Invited speaker 4
To be announced
Talk 4
15.30 - 15.45 To be announced Oral Presentation 3
15.45 - 16.00 Oral Presentation 4
16.00 - 16.45 Invited speaker 5
To be announced
Talk 5
16.45 - 17.30 Poster session Poster session

Important Dates

Together with the Computer Vision and Pattern Recognition (CVPR) 2018.

Date of the workshop: June 18, 2018


Constructive discussion

The workshop's goal is a constructive, creative and open conversation. In principle we accept all papers with interesting ideas.


Authors can visit the CMT submit 4 Page papers which will be peer reviewed. However, they will not be include in the proceedings. Please follow the CVPR 2018 camera ready format as per the instructions given here but limit your paper to 4 pages excluding references.





Registration & venue

The workshop is together with the Computer Vision and Pattern Recognition (CVPR) 2018.

Accepted papers must have at least one registered author (this can be a student).

Venue TBD.