Visual Perception via Learning in an Open World

The 3rd workshop on Open World Vision

June 18, held in conjunction with CVPR 2023, Vancouver, Canada


Visual perception is indispensable for numerous applications, spanning transportation, healthcare, security, commerce, entertainment, and interdisciplinary research. Currently, visual perception algorithms are often developed in a closed-world paradigm, which assumes the data distribution and categorical labels are fixed a priori. This is unrealistic in the real open world, which contains situations that are dynamic, vast, and unpredictable. Algorithms developed in a closed world appear brittle once exposed to the complexity of the open world, where they are unable to properly adapt or robustly generalize to new scenarios. We are motivated to invite researchers to the workshop on Visual Perception via Learning in an Open World (VPLOW) where we have multiple speakers and three challenge competitions to cover a variety of topics. We hope our workshop stimulates fruitful discussions on the open-world research.

You might be interested in our previous workshops:


Topics of interest include, but are not limited to:

  • Open-world data: long-tailed distribution, open-set, unknowns, streaming data, biased data, unlabeled data, anomaly, multi-modality, etc.
  • Learning/problems: X-shot learning, Y-supervised learning, lifelong/continual learning, domain adaptation/generalization, open-world learning, etc.
  • Social Impact: safety, fairness, real-world applications, inter-disciplinary research, etc.
  • Misc: datasets, benchmarks, interpretability, robustness, generalization, etc.


Let's consider the following motivational examples.

  • Open-world data follows a long-tail distribution. Data tends to follow a long-tailed distribution and real-world tasks often emphasize the rarely-seen data. A model trained on such long-tailed data can perform poorly on rare or underrepresentative data. For example, a visual recognition model can misclassify underrepresented minorities and make unethical predictions (ref. case1, case2).
  • Open-world contains unknown examples. Largely due to the long-tail nature of data distribution, visual perception models are invariably confronted by unknown examples in the open world. Failing to detecting the unknowns can cause serious issues. For example, a Tesla Model 3 did not identify an unknown overturned truck and crashed into the truck (ref. case).
  • Open-world requires learning with evolving data, and labels. The world of interest is changing over time, e.g., driving scenes (in different cities and under different weather), the search engine ("apple" means different things today and 20 years ago). This says that the data distribution and semantics are continually changing and evolving. How to address distribution shifts and concept drifts?


More speakers are on the way!

Shu Kong
Texas A&M

Subhransu Maji

Serge Belongie
University of Copenhagen

Lior Wolf
Tel Aviv University

Zeynep Akata
University of Tübingen / MPI-INF


Please contact Shu Kong with any questions: shu [at] tamu [dot] edu

Shu Kong
Texas A&M

Yu-Xiong Wang
University of Illinois at Urbana-Champaign
Andrew Owens
University of Michigan

Deepak Pathak
Carnegie Mellon University

Carl Vondrick
Columbia University

Abhinav Shrivastava
University of Maryland

Deva Ramanan
Carnegie Mellon University

Terrance Boult
University of Colorado Colorado Springs

Challenge Organizers

Tanmay Gupta
Allen Institute for AI (AI2)

Derek Hoiem
University of Illinois at Urbana-Champaign
Aniruddha (Ani) Kembhavi
Allen Institute for AI (AI2)

Amita Kamath
Allen Institute for AI (AI2)

Pulkit Kumar
University of Maryland

University of Maryland

Onsite Coordinators

Important Dates and Details

Please go to each challenge's website for its exact dates!
  • Submission deadline for Challenge-1: GRIT: June 10, 2023 at 11:59pm PST.
  • Submission deadline for Challenge-2: ObjCLsDisc: June 10, 2023 at 11:59pm PST.
  • Workshop date: June 18, 2023

Oral presentation will be selected from challenge participants, e.g., winners and those having innovative ideas.


This part will be updated shortly!

We organize three challenge competitions this year.

Program Schedule

This part will be updated shortly!

YouTube livestream url: url-place-holder.placeholder
Time (Central Time, UTC-5)
09:00 - 09:30
Opening remarks
Shu Kong Texas A&M
Open-World Visual Perception
09:30 - 10:00
Invited talk #1
Person tbd, institute
title tbd
10:00 - 10:30
Invited talk #2
Person tbd institute
title tbd
10:30 - 10:45
Break / time buffer
10:45 - 11:15
Invited talk #3
Person tbd, institute
title tbd
11:15 - 11:35
Host remark of Challenge-2 GRIT
GRIT, institute
title tbd
11:35 - 12:05
invited teams for GRIT
GRIT-invited#1 institute
GRIT-invited#2 institute
GRIT-invited#3 institute
GRIT-invited#4 institute
12:05 - 13:15
Lunch break
13:15 - 13:45
Invited talk #4
Person tbd, institute
title tbd
13:45 - 14:15
Invited talk #5
Person tbd, institute
title tbd
14:15 - 14:45
Invited talk #6
Person tbd, institute
title tbd
14:45 - 15:00
Break / time buffer
15:00 - 15:30
Invited talk #7
Person tbd institute
title tbd
15:30 - 16:00
Invited talk #8
Person tbd institute
title tbd
16:00 - 16:20
Host remark of Challenge-3 ObjCLsDisc
ObjCLsDisc, institute
title tbd
16:20 - 16:50
invited teams for ObjCLsDisc
16:50 - 17:00
Closing remarks