Warning: You are not logged in.
Your IP address will be recorded in this page's edit history.
A note on the various deadlines: In general, you only need some deadlines. The submission deadline should represent the single most important deadline for an event. If there is a paper deadline, it automatically counts as a submission deadline as well. The other kinds of deadlines are just relevant for events that accept many forms of contribution (e.g. papers and demos). If an event has just one type of submission (e.g. a developer's event might only accpept demos) then use submission deadline since it is the main deadline and it is clear what kind of submissions are meant.
Call for Papers
Joint Workshops of VCL’09 and ViSU’09 (held in conjunction with IEEE CVPR 2009)
June 25, 2009, Fontainebleau Resort, Miami Beach, Florida
Deadline for paper submission: March 20, 2009
Notification of acceptance: April 8, 2009
Camera-ready copies due: April 14, 2009
VCL’09: Workshop on Visual and Contextual Learning from Annotated Images and Videos
There has been a significant interest in the computer vision community on utilizing visual and contextual models for high level semantic reasoning. Large datasets of weakly annotated images and videos, along with other rich sources of information such as dictionaries, can be used to learn visual and contextual models for recognition.
The goal of this workshop is to investigate how linguistic information available in the form of captions and other sources can be used to aid in visual and contextual learning. For example, nouns in captions can help train object detectors; adjectives in captions could be used to train material recognizers; or written descriptions of objects could be used to train object recognizers. This workshop aims to bring together researchers in the fields of contextual modeling in computer vision, machine learning and natural language processing to explore a variety of perspectives on how these annotated datasets can be employed.
Scope: The list of possible topics includes (but is not limited to) the following:
• Contextual Relationships for Recognition
o Scene, Object, Action and Event Recognition using context models
o Learning structure and parameters of contextual models
• Linguistic annotations to Assist Learning and Incremental Labeling
o Annotations for learning visual and contextual models
o Rich language models of annotations.
o Learning to recognize by reading
o Utilizing other sources such as dictionaries
o Modeling annotations with errors
o Biologically motivated semantic models
o Scalable learning from large datasets
ViSU’09: 1st International Workshop on Visual Scene Understanding
One of the holy grails of computer vision research is achieving the total understanding of a visual scene, in a similar way that humans can do effortlessly. Great progress has been made in tackling one of the most critical components of visual scene understanding: object recognition and categorization. Recently, a few pioneering works have begun to look into the interactions among objects and their environments towards the goal of a more holistic representation and explanation of the visual scene. This workshop offers an opportunity to bring together experts working on different aspects of scene understanding and image interpretation and provides a common playground for a stimulating debate.
The Representation and Recognition Aspect
• What is total scene understanding? What defines objects, things, stuff, context, scenes and other high-level visual concepts (e.g., activities and events)?
• How is this problem related to the ‘image annotation’ problem from the CBIR/TRECVID community?
• Can we break down the task into individual component recognition? If yes, how? If no, why?
The Learning Aspect
• What are the challenges faced in learning for these problems? Are they the same or different from isolated object recognition?
• What can we do to exploit the huge amount of data on the web?
New Methods for Evaluation and Benchmarking
• Are the object recognition datasets still suitable for this problem? If not, what are the new datasets? What are the appropriate benchmark tasks and evaluation metrics?
Paper topics may include total scene understanding, visual recognition and annotation in complex real-world images, high-level understanding (events, activities) in single images, and all their related learning, representational, and dataset issues.
Submission and Reviews
a) General instruction
Submitted papers must have a maximum length of 8 pages and must adhere to the same CVPR 2009 layout specifications as papers submitted to the main conference. All reviewing will be carried out double-blind by the Program Committee. Please refer to CVPR 2009 website (http://www.cvpr2009.org/) for instructions on the submission format. In submitting a manuscript, authors acknowledge that no paper substantially similar in content has been or will be submitted to another conference or workshop during the review period.
b) Submitting to VCL vs. ViSU
In the submission form, the authors will be asked to indicate which workshop (VCL or ViSU) the authors prefer to submit to. If no preference is indicated, the VCL-ViSU program chairs will make the decision. Double submission to ViSU’09 workshop and the main CVPR’09 conference are allowed.
Abhinav Gupta (Maryland), Jianbo Shi (Penn), David Forsyth (UIUC)
Fei-Fei Li (Princeton), Serge Belongie (UCSD)
Kobus Barnard (Arizona), Serge Belongie (UCSD), Alex Berg (Yahoo Research), Tamara Berg (SUNY Stony Brook), David Blei (Princeton), Larry Davis (Maryland), Sven Dickinson (Toronto), Pinar Duygulu (Bilkent U.), Alexei A. Efros (CMU), Mark Everingham (Leeds), Kristen Grauman (UT Austin), James Hayes (CMU), Derek Hoiem (UIUC), Julia Hockenmaier (UIUC), Sanjiv Kumar (Google), Svetlana Lazebnik (UNC), Fei-Fei Li (Princeton), Ivan Laptev (INRIA), R. Manmatha (U Mass), Fernando Pereira (U Penn), Cordelia Schmid (INRIA), Ben Taskar (U Penn), Antonio Torralba (MIT), Tinne Tuytelaars (KU Leuven), Nuno Vasconcelos (UCSD), James Z. Wang (Penn State), Andrew Zisserman (Oxford)
Kobus Barnard (Arizona), Miguel Á. Carreira-Perpiñán (UC Merced), Tsuhan Chen (CMU), Trevor Darrell (Berkeley), Larry Davis (UMaryland), Pinar Duygulu (Bilkent U.), Alyosha Efros (CMU), Mark Everingham (Leeds), Pedro Felzenszwalb (U.Chicago), Vittorio Ferrari (ETHZ), David Forsyth (UIUC), Bill Freeman (MIT), Kristen Grauman (U.T. Austin), Abhinav Gupta (UMaryland), Geremy Heitz (Stanford), Derek Hoiem (UIUC), Xian-Sheng Hua (MSRA), Dan Huttenlocher (Cornell), Frédéric Jurie (INRIA), Daphne Koller (Stanford), Lana Lazebnik (UNC), Bastian Leibe (Aachen), Jia Li (Princeton), Jitendra Malik (Berkeley), Greg Mori (SFU), Pietro Perona (Caltech), Andrew Rabinovich (Google), Deva Ramanan (UCI), Brian Russell (ENS), Silvio Savarese (UMich), Cordelia Schmid (INRIA), Erik Sudderth (Brown), Sinisa Todorovic (Oregon State), Antonio Torralba (MIT), Zhuowen Tu (UCLA), Nuno Vasconcelos (UCSD), John Winn (MSRC), Song-Chun Zhu (UCLA), Andrew Zisserman (Oxford)
</pre>This CfP was obtained from [http://www.wikicfp.com/cfp/servlet/event.showcfp?eventid=4798&copyownerid=3424 WikiCFP][[Category:Computer graphics]]
This is a minor edit
Watch this page