AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition
保存先:
| 出版年: | arXiv.org (Aug 21, 2024), p. n/a |
|---|---|
| 第一著者: | |
| その他の著者: | , , , , , , , |
| 出版事項: |
Cornell University Library, arXiv.org
|
| 主題: | |
| オンライン・アクセス: | Citation/Abstract Full text outside of ProQuest |
| タグ: |
タグなし, このレコードへの初めてのタグを付けませんか!
|
MARC
| LEADER | 00000nab a2200000uu 4500 | ||
|---|---|---|---|
| 001 | 3095811060 | ||
| 003 | UK-CbPIL | ||
| 022 | |a 2331-8422 | ||
| 035 | |a 3095811060 | ||
| 045 | 0 | |b d20240821 | |
| 100 | 1 | |a Ni, Minheng | |
| 245 | 1 | |a AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition | |
| 260 | |b Cornell University Library, arXiv.org |c Aug 21, 2024 | ||
| 513 | |a Working Paper | ||
| 520 | 3 | |a With the advancement of generative models, the synthesis of different sensory elements such as music, visuals, and speech has achieved significant realism. However, the approach to generate multi-sensory outputs has not been fully explored, limiting the application on high-value scenarios such as of directing a film. Developing a movie director agent faces two major challenges: (1) Lack of parallelism and online scheduling with production steps: In the production of multi-sensory films, there are complex dependencies between different sensory elements, and the production time for each element varies. (2) Diverse needs and clear communication demands with users: Users often cannot clearly express their needs until they see a draft, which requires human-computer interaction and iteration to continually adjust and optimize the film content based on user feedback. To address these issues, we introduce AutoDirector, an interactive multi-sensory composition framework that supports long shots, special effects, music scoring, dubbing, and lip-syncing. This framework improves the efficiency of multi-sensory film production through automatic scheduling and supports the modification and improvement of interactive tasks to meet user needs. AutoDirector not only expands the application scope of human-machine collaboration but also demonstrates the potential of AI in collaborating with humans in the role of a film director to complete multi-sensory films. | |
| 653 | |a Motion pictures | ||
| 653 | |a Task scheduling | ||
| 653 | |a Composition effects | ||
| 653 | |a User needs | ||
| 653 | |a Computer aided scheduling | ||
| 653 | |a Motion picture directors & producers | ||
| 653 | |a Production scheduling | ||
| 653 | |a Dubbing | ||
| 653 | |a Special effects | ||
| 700 | 1 | |a Wu, Chenfei | |
| 700 | 1 | |a Huaying Yuan | |
| 700 | 1 | |a Yang, Zhengyuan | |
| 700 | 1 | |a Gong, Ming | |
| 700 | 1 | |a Wang, Lijuan | |
| 700 | 1 | |a Liu, Zicheng | |
| 700 | 1 | |a Zuo, Wangmeng | |
| 700 | 1 | |a Duan, Nan | |
| 773 | 0 | |t arXiv.org |g (Aug 21, 2024), p. n/a | |
| 786 | 0 | |d ProQuest |t Engineering Database | |
| 856 | 4 | 1 | |3 Citation/Abstract |u https://www.proquest.com/docview/3095811060/abstract/embedded/7BTGNMKEMPT1V9Z2?source=fedsrch |
| 856 | 4 | 0 | |3 Full text outside of ProQuest |u http://arxiv.org/abs/2408.11564 |