Deep Watching: Towards New Methods of Analyzing Visual Media in Cultural Studies
Type
conference paper
Date Issued
2019-07
Author(s)
Gassner, Sebastian
Howanitz, Gernot
Radisch, Erik
Rehbein, Malte
Abstract
A large number of digital humanities projects focuses on text. This medial limitation may be attributed to the abundance of well-established quantitative methods applicable to text. Cultural Studies, however, analyse cultural expressions in a broad sense, including different non-textual media, physical artefacts, and performative actions. It is, to a certain extent, possible to transcribe these multi-medial phenomena in textual form; however, this transcription is difficult to automate and some information may be lost. Thus, quantitative approaches which directly access media-specific information are a desideratum for Cultural Studies.
Visual media constitute a significant part of cultural production. In our paper, we propose Deep Watching as a way to analyze visual media (films, photographs, and video clips) using cutting-edge machine learning and computer vision algorithms. Unlike previous approaches, which were based on generic information such as frame differences (Howanitz 2015), color distribution (Burghardt/Wolff 2016) or used manual annotation altogether (Dunst/Hartel 2016), Deep Watching allows to automatically identify visual information (symbols, objects, persons, body language, visual configuration of the scene) in large image and video corpora. To a certain extent, Tilton and Arnold’s Distant-Viewing Toolkit uses a comparable approach (Tilton/Arnold 2018). However, by means of our customized training of state-of-the-art convolutional neural networks for object detection and face recognition we can, in comparison to this toolkit, automatically extract more information about individual frames and their contexts.
Visual media constitute a significant part of cultural production. In our paper, we propose Deep Watching as a way to analyze visual media (films, photographs, and video clips) using cutting-edge machine learning and computer vision algorithms. Unlike previous approaches, which were based on generic information such as frame differences (Howanitz 2015), color distribution (Burghardt/Wolff 2016) or used manual annotation altogether (Dunst/Hartel 2016), Deep Watching allows to automatically identify visual information (symbols, objects, persons, body language, visual configuration of the scene) in large image and video corpora. To a certain extent, Tilton and Arnold’s Distant-Viewing Toolkit uses a comparable approach (Tilton/Arnold 2018). However, by means of our customized training of state-of-the-art convolutional neural networks for object detection and face recognition we can, in comparison to this toolkit, automatically extract more information about individual frames and their contexts.
Language
English
HSG Classification
contribution to scientific community
Refereed
Yes
Division(s)
Contact Email Address
bernhard.bermeitinger@unisg.ch
Eprints ID
258103
File(s)![Thumbnail Image]()
![Thumbnail Image]()
Loading...
open.access
Name
Deep Watching_ Towards New Methods of Analyzing Visual Media in Cultural Studies.pdf
Size
228.05 KB
Format
Adobe PDF
Checksum (MD5)
fcc8b87817232b8a17c9d9a563742c64
Loading...
open.access
Name
deep-watching_dh2019.pdf
Size
5.73 MB
Format
Adobe PDF
Checksum (MD5)
6fec4fbfedf42f0b2f87c13eb682b56d