Deep Learning Based Video Captioning through Encoder-Decoder Based Long Short-Term Memory (LSTM)

International Journal of Advance Computer Science and Application (forthcoming)
  Copy   BIBTEX

Abstract

This work demonstrates the implementation and use of an encoder-decoder model to perform a many-to-many mapping of video data to text captions. The many-to-many mapping occurs via an input temporal sequence of video frames to an output sequence of words to form a caption sentence. Data preprocessing, model construction, and model training are discussed. Caption correctness is evaluated using 2-gram BLEU scores across the different splits of the dataset. Specific examples of output captions were shown to demonstrate model generality over the video temporal dimension. Predicted captions were shown to generalize over video action, even in instances where the video scene changed dramatically. Model architecture changes are discussed to improve sentence grammar and correctness.

Links

PhilArchive

External links

  • This entry has no external links. Add one.
Setup an account with your affiliations in order to access resources via your University's proxy server

Through your library

Similar books and articles

Deep Learning Based Video Captioning through Encoder-Decoder Based Long Short-Term Memory (LSTM).Grimsby Chelsea - forthcoming - International Journal of Advanced Computer Science and Applications:1-6.
Short-term prediction of parking availability in an open parking lot.Vijay Paidi - 2022 - Journal of Intelligent Systems 31 (1):541-554.
A DEEP LEARNING APPROACH FOR LSTM BASED COVID-19 FORECASTING SYSTEM.K. Jothimani - 2022 - Journal of Science Technology and Research (JSTAR) 3 (1):28-38.
The short-term/long-term memory distinction: Back to the past?Giuseppe Vallar - 2003 - Behavioral and Brain Sciences 26 (6):757-758.

Analytics

Added to PP
2024-03-08

Downloads
44 (#363,319)

6 months
44 (#94,399)

Historical graph of downloads
How can I increase my downloads?

Author's Profile

Chelsea Grimsby
ESPAM FORMATION UNIVERSITY

Citations of this work

No citations found.

Add more citations

References found in this work

No references found.

Add more references