LiRI Wiki

Linguistic Research Infrastructure - University of Zurich

User Tools

Site Tools


langtech:vian-dh

This is an old revision of the document!


VIAN-DH

VIAN-DH is web-based software for multimodal corpora. It is a part of LiRI Corpus Platform (LCP), specializing in audio-visual language resources.

Corpora stored in VIAN-DH containing different types of annotations. To name some of them, annotations can include speech transcript, part-of-speech annotation and other kinds of labels marking embodied communication, such as gestures or facial epressions, or document segmentation such as scene for film, turn for conversations, or experiment phase for various experimental settings. This is not an exghaustive list, and the scope of annotations varies from one corpus of another.

VIAN-DH Data structure

VIAN-DH allows for different types of alignment. Primary types are:

  • time-frame alignment for any annotations linked to the time or frame stream
  • text-alignment referring to continuous stream of characters for annotations referring to textual content, such as part-of-speech, named-entity-recognition, etc.

Any annotation can have multiple alignment to both of these streams by marking its start and end position/index, e.g., a sentence can have a start and end time and a start and end character. See the image below for a graphical representaiton of data being linked to multiple alignement streams. Note how each annotation track (gestures, utterances, tokens, named entities and constituents) are linked to both time and text stream (utterances), or either text stream (tokens, named entities and constituents), or time stream (gestures). In case of video, time stream also refers to frame stream - since each video has a specific frame rate.

Data representation on virtual alignment streams

langtech/vian-dh.1689253400.txt.gz · Last modified: by teodora.vukovic2.uzh.ch

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki