Conference paper Open Access
Damianos Galanopoulos; Vasileios Mezaris
{ "files": [ { "links": { "self": "https://zenodo.org/api/files/0f05cc5f-6eec-44c1-a9a6-d657e62fb042/mmm19_lncs11296_1_preprint.pdf" }, "checksum": "md5:a35e6b610e7e6c2ef81cb8d449f0be1d", "bucket": "0f05cc5f-6eec-44c1-a9a6-d657e62fb042", "key": "mmm19_lncs11296_1_preprint.pdf", "type": "pdf", "size": 2840367 } ], "owners": [ 22754 ], "doi": "10.5281/zenodo.2539272", "stats": { "version_unique_downloads": 19.0, "unique_views": 25.0, "views": 28.0, "downloads": 21.0, "unique_downloads": 19.0, "version_unique_views": 25.0, "volume": 59647707.0, "version_downloads": 21.0, "version_views": 28.0, "version_volume": 59647707.0 }, "links": { "doi": "https://doi.org/10.5281/zenodo.2539272", "conceptdoi": "https://doi.org/10.5281/zenodo.2539271", "bucket": "https://zenodo.org/api/files/0f05cc5f-6eec-44c1-a9a6-d657e62fb042", "conceptbadge": "https://zenodo.org/badge/doi/10.5281/zenodo.2539271.svg", "html": "https://zenodo.org/record/2539272", "latest_html": "https://zenodo.org/record/2539272", "badge": "https://zenodo.org/badge/doi/10.5281/zenodo.2539272.svg", "latest": "https://zenodo.org/api/records/2539272" }, "conceptdoi": "10.5281/zenodo.2539271", "created": "2019-01-14T13:21:59.365047+00:00", "updated": "2019-11-14T19:10:28.653877+00:00", "conceptrecid": "2539271", "revision": 6, "id": 2539272, "metadata": { "access_right_category": "success", "doi": "10.5281/zenodo.2539272", "description": "<p>In this work the problem of temporal video lecture fragmentation in meaningful parts is addressed. The visual content of lecture video can not be effectively used for this task due to its extremely homogeneous content. A new method for lecture video fragmentation in which only automatically generated speech transcripts of a video are exploited, is proposed. Contrary to previously proposed works that employ visual, audio and textual features and use time-consuming supervised methods which require annotated training data, we present a method that analyses the transcripts’ text with the help of word embeddings that are generated from pre-trained state-of-the-art neural networks. Furthermore,we address a major problem of video lecture fragmentation research, which is the lack of large-scale datasets for evaluation, by presenting a new artificially- generated dataset of synthetic video lecture transcripts that we make publicly available. Experimental comparisons document the merit of the proposed approach.</p>", "license": { "id": "CC-BY-4.0" }, "title": "Temporal Lecture Video Fragmentation using Word Embeddings", "relations": { "version": [ { "count": 1, "index": 0, "parent": { "pid_type": "recid", "pid_value": "2539271" }, "is_last": true, "last_child": { "pid_type": "recid", "pid_value": "2539272" } } ] }, "communities": [ { "id": "moving-h2020" } ], "keywords": [ "Lecture Video Fragmentation,", "Word Embeddings", "Video Segmentation" ], "publication_date": "2019-01-08", "creators": [ { "affiliation": "Information Technologies Institute/CERTH", "name": "Damianos Galanopoulos" }, { "affiliation": "Information Technologies Institute/CERTH", "name": "Vasileios Mezaris" } ], "meeting": { "acronym": "MMM 2019", "url": "http://mmm2019.iti.gr/", "dates": "08-11 January 2019", "place": "Thessaloniki, Greece", "title": "25th International Conference on MultiMedia Modeling" }, "access_right": "open", "resource_type": { "subtype": "conferencepaper", "type": "publication", "title": "Conference paper" }, "related_identifiers": [ { "scheme": "doi", "identifier": "10.5281/zenodo.2539271", "relation": "isVersionOf" } ] } }
All versions | This version | |
---|---|---|
Views | 28 | 28 |
Downloads | 21 | 21 |
Data volume | 59.6 MB | 59.6 MB |
Unique views | 25 | 25 |
Unique downloads | 19 | 19 |