Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content

dc.contributor.author Paula Viana en
dc.contributor.author Maria Teresa Andrade en
dc.contributor.author Pedro Miguel Carvalho en
dc.contributor.author Luís Miguel Salgado en
dc.contributor.author Inês Filipa Teixeira en
dc.contributor.author Tiago André Costa en
dc.contributor.author Jonker,P en
dc.contributor.other 400 en
dc.contributor.other 1107 en
dc.contributor.other 4358 en
dc.contributor.other 5363 en
dc.contributor.other 7420 en
dc.contributor.other 7514 en
dc.date.accessioned 2023-05-10T15:10:28Z
dc.date.available 2023-05-10T15:10:28Z
dc.date.issued 2022 en
dc.description.abstract <jats:p>Applying machine learning (ML), and especially deep learning, to understand visual content is becoming common practice in many application areas. However, little attention has been given to its use within the multimedia creative domain. It is true that ML is already popular for content creation, but the progress achieved so far addresses essentially textual content or the identification and selection of specific types of content. A wealth of possibilities are yet to be explored by bringing the use of ML into the multimedia creative process, allowing the knowledge inferred by the former to influence automatically how new multimedia content is created. The work presented in this article provides contributions in three distinct ways towards this goal: firstly, it proposes a methodology to re-train popular neural network models in identifying new thematic concepts in static visual content and attaching meaningful annotations to the detected regions of interest; secondly, it presents varied visual digital effects and corresponding tools that can be automatically called upon to apply such effects in a previously analyzed photo; thirdly, it defines a complete automated creative workflow, from the acquisition of a photograph and corresponding contextual data, through the ML region-based annotation, to the automatic application of digital effects and generation of a semantically aware multimedia story driven by the previously derived situational and visual contextual data. Additionally, it presents a variant of this automated workflow by offering to the user the possibility of manipulating the automatic annotations in an assisted manner. The final aim is to transform a static digital photo into a short video clip, taking into account the information acquired. The final result strongly contrasts with current standard approaches of creating random movements, by implementing an intelligent content- and context-aware video.</jats:p> en
dc.identifier P-00W-9AC en
dc.identifier.uri http://dx.doi.org/10.3390/jimaging8030068 en
dc.identifier.uri https://repositorio.inesctec.pt/handle/123456789/14009
dc.language eng en
dc.rights info:eu-repo/semantics/openAccess en
dc.title Photo2Video: Semantic-Aware Deep Learning-Based Video Generation from Still Content en
dc.type en
dc.type Publication en
Files
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
P-00W-9AC.pdf
Size:
6.86 MB
Format:
Adobe Portable Document Format
Description: