Automatic personalized story generation for visual media

Exemplary embodiments relate to the automatic generation of captions for visual media, including photos, photo albums, non-live video, and live video. The visual media may be analyzed to determine contextual information (such as location information, people and objects in the video, time, etc.). A s...

Full description

Saved in:
Bibliographic Details
Main Authors Zhang, Ying, Guo, Shengbo
Format Patent
LanguageEnglish
Published 11.06.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Exemplary embodiments relate to the automatic generation of captions for visual media, including photos, photo albums, non-live video, and live video. The visual media may be analyzed to determine contextual information (such as location information, people and objects in the video, time, etc.). A system may integrate this information with information from the user's social network and a personalized language model built using public-facing language from the user. The personalized language model captures the user's way of speaking to make the generated captions more detailed and personalized. The language model may account for the context in which the video was generated. The captions maybe used to simplify and encourage content generation, and may also be used to index visual media, rank the media, and recommend the media to users likely to engage with the media.
Bibliography:Application Number: US202017069263