A Reinforcement Learning-Based Automatic Video Editing Method Using Pre-trained Vision-Language Model

被引:0
|
作者
Hu, Panwen [1 ]
Xiao, Nan [1 ]
Li, Feifei [1 ]
Chen, Yongquan [2 ]
Huang, Rui [1 ]
机构
[1] Chinese Univ Hong Kong, SSE, Shenzhen, Peoples R China
[2] Chinese Univ Hong Kong, AIRS, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023 | 2023年
关键词
video editing; video representation; reinforcement learning; BROADCAST; CAPTURE; FILM;
D O I
10.1145/3581783.3611878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this era of videos, automatic video editing techniques attract more and more attention from industry and academia since they can reduce workloads and lower the requirements for human editors. Existing automatic editing systems are mainly scene- or event-specific, e.g., soccer game broadcasting, yet the automatic systems for general editing, e.g., movie or vlog editing which covers various scenes and events, were rarely studied before, and converting the event-driven editing method to a general scene is nontrivial. In this paper, we propose a two-stage scheme for general editing. Firstly, unlike previous works that extract scene-specific features, we leverage the pre-trained Vision-Language Model (VLM) to extract the editing-relevant representations as editing context. Moreover, to close the gap between the professional-looking videos and the automatic productions generated with simple guidelines, we propose a Reinforcement Learning (RL)-based editing framework to formulate the editing problem and train the virtual editor to make better sequential editing decisions. Finally, we evaluate the proposed method on a more general editing task with a real movie dataset. Experimental results demonstrate the effectiveness and benefits of the proposed context representation and the learning ability of our RL-based editing framework.
引用
收藏
页码:6441 / 6450
页数:10
相关论文
共 18 条
  • [1] Model Based Reinforcement Learning Pre-Trained with Various State Data
    Ono, Masaaki
    Ichise, Ryutaro
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 918 - 925
  • [2] Reinforcement Learning Friendly Vision-Language Model for Minecraft
    Jiang, Haobin
    Yue, Junpeng
    Luo, Hao
    Ding, Ziluo
    Lu, Zongqing
    COMPUTER VISION - ECCV 2024, PT XXXVII, 2025, 15095 : 1 - 17
  • [3] Pre-trained Bert for Natural Language Guided Reinforcement Learning in Atari Game
    Li, Xin
    Zhang, Yu
    Luo, Junren
    Liu, Yifeng
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 5119 - 5124
  • [4] A reinforcement learning intelligent deductive model with pre-trained sequence information
    Han, Xinyu
    Xu, Huosheng
    Yu, Hao
    Li, Sizhao
    INTERNATIONAL JOURNAL OF BIO-INSPIRED COMPUTATION, 2023, 22 (04) : 195 - 205
  • [5] Solar irradiance prediction using reinforcement learning pre-trained with limited historical data
    Jeon, Byung-Ki
    Kim, Eui-Jong
    ENERGY REPORTS, 2023, 10 : 2513 - 2524
  • [6] Generative Pre-Trained Transformer-Based Reinforcement Learning for Testing Web Application Firewalls
    Liang, Hongliang
    Li, Xiangyu
    Xiao, Da
    Liu, Jie
    Zhou, Yanjie
    Wang, Aibo
    Li, Jin
    IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, 2024, 21 (01) : 309 - 324
  • [7] A Pre-Trained Fuzzy Reinforcement Learning Method for the Pursuing Satellite in a One-to-One Game in Space
    Wang, Xiao
    Shi, Peng
    Zhao, Yushan
    Sun, Yue
    SENSORS, 2020, 20 (08)
  • [8] Deep Reinforcement Learning-Based Operation of Distribution Systems Using Surrogate Model
    Bu, Van -Hai
    Zarrabian, Sina
    Su, Wencong
    2023 IEEE POWER & ENERGY SOCIETY GENERAL MEETING, PESGM, 2023,
  • [9] Sample Trajectory Selection Method Based on Large Language Model in Reinforcement Learning
    Lai, Jinbang
    Zang, Zhaoxiang
    IEEE ACCESS, 2024, 12 : 61877 - 61885
  • [10] RLVS: A Reinforcement Learning-Based Sparse Adversarial Attack Method for Black-Box Video Recognition
    Song, Jianxin
    Yu, Dan
    Teng, Hongfei
    Chen, Yongle
    ELECTRONICS, 2025, 14 (02):