Welcome Topic Programme Dates Submission Organizers Return To Top


Videos are being generated in large numbers and volumes - by the public, video conferencing tools (e.g. Teams, Zoom, Webex), and TV broadcasters such as the BBC. The videos may be stored in public archives such as Youtube or proprietary archives such as the BBC Rewind Archive. How to search video archives without using pre-defined metadata such as titles, tags, and viewer notes is a challenge.

The Workshop proposers are undertaking an EPSRC funded research project, Multimodal Video Search by Examples (MVSE), to tackle this challenge. This is motivated by a use case from the BBC, “locate clips with person X in setting Y talking about subject Z”. This use case is difficult to answer by keyword-based search. The MVSE project conducts research on the image and video analysis techniques that enable search with a face image and an audio of X, a scene image of Y and a text phrase of subject Z, where the modalities are person (face or voice), context and topic.

Insights have been gained, a demonstrator has been built, and a video archive has been well understood. It is the opinion of the proposers it is time to share with the research community our vision, our findings and techniques in order to accelerate the research and impact of an important research area, multimodel video search by examples.


  • Image and video retrieval
  • Image and video segmentation
  • Image and video embedding
  • Multimodal information retrieval
  • Multimodal language modelling
  • Human centred AI for multimedia search


Invited talks (90mins)

To start the workshop we will hear from 3 invited keynote speakers who will outline core information across the scope of the workshop areas.

Contributed talks

Approximately 5-6 papers will be presented with each presentation lasting approximately 15mins. Researchers will be able to showcase their innovative contributions or thoughts on one (or more) of the workshop areas.

Posters and Video demonstrations

Throughout the workshop, poster presentation will be on relevant research for the workshop. Additionally, some video search demonstrations will be available for participants to engage with.


  • Paper Submission: May 1, 2024 (Final - no further extension)
  • Paper Notification: May 6, 2024
  • Camera-Ready Submission: May 8, 2024


Paper format

All papers must be formatted according to the ACM proceedings style. Click here to access LaTeX and Microsoft Word templates for this format. If you use LaTeX, please use sigconf.tex as the template.

Complying with double-blind review

In a “Double-blind Review” process, authors should not know the names of the reviewers of their papers, and reviewers should not know the names of the authors. Please prepare your paper in a way that preserves the anonymity of the authors, namely:

  • Do not put your names under the title
  • Avoid using phrases such as “our previous work” when referring to earlier publications by the authors,
  • Remove information that may identify the authors in the acknowledgements (e.g., co-workers and grant IDs),
  • Check supplemental material for information that may identify the authors’ identity,
  • Avoid providing links to Websites that identify the authors.

Length of the paper

Please ensure the appropriate length of your submission. The length of both short and long papers should not exceed 6 pages including references.

All submissions can be made via EasyChair



  • Prof Hui Wang, Queen’s University Belfast, mvrmlm2024@qub.ac.uk
  • Prof Josef Kittler, University of Surrey
  • Prof Mark Gales, University of Cambridge
  • Dr Rob Cooper, the BBC
  • Prof Maurice Mulvenna, Ulster University
  • Prof Wing Ng, South China University of Technology
  • Dr Yang Hua, Queen’s University Belfast
  • Dr Richard Gault, Queen’s University Belfast