IIIF / iiif-stories

Community repository for documenting stories and use cases related to uses of the International Image Interoperability Framework.
21 stars 0 forks source link

I want to allow the client/user to identify captions for audio and video items that include representation of all sounds, not just dialogue #143

Open elynema opened 5 months ago

elynema commented 5 months ago

Description

In the US, historically captions are intended for users who cannot hear audio. Therefore, they should contain a written form of every sound that occurs in a video or audio item. They include sound effects and music in addition to dialogue. Subtitles were originally intended for users who could not understand the spoken language, therefore only contain text that represents speech. Subtitles that contain representation of all audio may be referred to as SDH (subtitles for the deaf/hard of hearing).

For a user who is unable to hear the audio for a video item, I want to be able to select a closed caption that includes the written representation of all audio (as opposed to subtitles only).

Variation(s)

While this is appropriate for users with hearing impairments, it is also relevant for users in situations where video is being played muted or where users are in a noisy environment. In this scenario, either the user or the IIIF viewer itself may want to enable use of captions rather than subtitles (where both exist) so that maximum information is available.

Proposed Solutions

Introduce annotation motivations of captioning and subtitling to distinguish between subtitles that represent spoken words and captions that include all audio.

Additional Background

(more about your perspective, existing work, etc. goes here.)

jcoyne commented 5 months ago

What if you have data that are either captions/subtitles but you don't know the underlying motivation for which they were created? Is there a way to distinguish between the two?