CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition
We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, Cine...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on robotics 2024, Vol.40, p.1740-1757 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1757 |
---|---|
container_issue | |
container_start_page | 1740 |
container_title | IEEE transactions on robotics |
container_volume | 40 |
creator | Pueyo, Pablo Dendarieta, Juan Montijano, Eduardo Murillo, Ana Cristina Schwager, Mac |
description | We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community. |
doi_str_mv | 10.1109/TRO.2024.3353550 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TRO_2024_3353550</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10398502</ieee_id><sourcerecordid>2929260099</sourcerecordid><originalsourceid>FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</originalsourceid><addsrcrecordid>eNpNkD1PwzAQhiMEElDYGRgssTblbMdNjq0qFJCKWvGxsETGcSCo8QU7GfLvSVQGdMPd8LzvSU8UXXCYcQ54_fq8mQkQyUxKJZWCg-iEY8JjSObZ4XArJWIJmB1HpyF8w0AiyJPILStnn7bLG7Zgq26369mia8lRTV1gt56cZSNR65Y-vW6-evbSh9bW7NEZ8g153Vbuk70T1VO2ItOFKdtSsFOmXcFejB0LqG4oVG1F7iw6KvUu2PO_PYneVnevy4d4vbl_XC7WsRFZ2sYi_ZAfKEudGkwRU1Am0zbBDIUoALRKiwTN3IIRymrkStpC80xwTI0uslJOoqt9b-Ppp7Ohzb-p8254mQscZg6AOFCwp4ynELwt88ZXtfZ9ziEfreaD1Xy0mv9ZHSKX-0hlrf2HS8wUCPkLwzNzEw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2929260099</pqid></control><display><type>article</type><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><source>IEEE Electronic Library (IEL)</source><creator>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</creator><creatorcontrib>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</creatorcontrib><description>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</description><identifier>ISSN: 1552-3098</identifier><identifier>EISSN: 1941-0468</identifier><identifier>DOI: 10.1109/TRO.2024.3353550</identifier><identifier>CODEN: ITREAE</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Aerial robotics applications ; autonomous drone cinematography ; camera intrinsics ; Cameras ; Cinematography ; Depth of field ; Drones ; model predicted control (MPC) ; Nonlinear control ; Parameters ; Predictive control ; Recording ; Robots ; Target tracking ; Tracking ; Trajectory</subject><ispartof>IEEE transactions on robotics, 2024, Vol.40, p.1740-1757</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</cites><orcidid>0000-0002-5176-3767 ; 0000-0002-7871-3663 ; 0000-0001-6482-009X ; 0000-0002-7580-9037</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10398502$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,778,782,794,4012,27910,27911,27912,54745</link.rule.ids></links><search><creatorcontrib>Pueyo, Pablo</creatorcontrib><creatorcontrib>Dendarieta, Juan</creatorcontrib><creatorcontrib>Montijano, Eduardo</creatorcontrib><creatorcontrib>Murillo, Ana Cristina</creatorcontrib><creatorcontrib>Schwager, Mac</creatorcontrib><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><title>IEEE transactions on robotics</title><addtitle>TRO</addtitle><description>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</description><subject>Aerial robotics applications</subject><subject>autonomous drone cinematography</subject><subject>camera intrinsics</subject><subject>Cameras</subject><subject>Cinematography</subject><subject>Depth of field</subject><subject>Drones</subject><subject>model predicted control (MPC)</subject><subject>Nonlinear control</subject><subject>Parameters</subject><subject>Predictive control</subject><subject>Recording</subject><subject>Robots</subject><subject>Target tracking</subject><subject>Tracking</subject><subject>Trajectory</subject><issn>1552-3098</issn><issn>1941-0468</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><recordid>eNpNkD1PwzAQhiMEElDYGRgssTblbMdNjq0qFJCKWvGxsETGcSCo8QU7GfLvSVQGdMPd8LzvSU8UXXCYcQ54_fq8mQkQyUxKJZWCg-iEY8JjSObZ4XArJWIJmB1HpyF8w0AiyJPILStnn7bLG7Zgq26369mia8lRTV1gt56cZSNR65Y-vW6-evbSh9bW7NEZ8g153Vbuk70T1VO2ItOFKdtSsFOmXcFejB0LqG4oVG1F7iw6KvUu2PO_PYneVnevy4d4vbl_XC7WsRFZ2sYi_ZAfKEudGkwRU1Am0zbBDIUoALRKiwTN3IIRymrkStpC80xwTI0uslJOoqt9b-Ppp7Ohzb-p8254mQscZg6AOFCwp4ynELwt88ZXtfZ9ziEfreaD1Xy0mv9ZHSKX-0hlrf2HS8wUCPkLwzNzEw</recordid><startdate>2024</startdate><enddate>2024</enddate><creator>Pueyo, Pablo</creator><creator>Dendarieta, Juan</creator><creator>Montijano, Eduardo</creator><creator>Murillo, Ana Cristina</creator><creator>Schwager, Mac</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-5176-3767</orcidid><orcidid>https://orcid.org/0000-0002-7871-3663</orcidid><orcidid>https://orcid.org/0000-0001-6482-009X</orcidid><orcidid>https://orcid.org/0000-0002-7580-9037</orcidid></search><sort><creationdate>2024</creationdate><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><author>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Aerial robotics applications</topic><topic>autonomous drone cinematography</topic><topic>camera intrinsics</topic><topic>Cameras</topic><topic>Cinematography</topic><topic>Depth of field</topic><topic>Drones</topic><topic>model predicted control (MPC)</topic><topic>Nonlinear control</topic><topic>Parameters</topic><topic>Predictive control</topic><topic>Recording</topic><topic>Robots</topic><topic>Target tracking</topic><topic>Tracking</topic><topic>Trajectory</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Pueyo, Pablo</creatorcontrib><creatorcontrib>Dendarieta, Juan</creatorcontrib><creatorcontrib>Montijano, Eduardo</creatorcontrib><creatorcontrib>Murillo, Ana Cristina</creatorcontrib><creatorcontrib>Schwager, Mac</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Mechanical & Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on robotics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Pueyo, Pablo</au><au>Dendarieta, Juan</au><au>Montijano, Eduardo</au><au>Murillo, Ana Cristina</au><au>Schwager, Mac</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</atitle><jtitle>IEEE transactions on robotics</jtitle><stitle>TRO</stitle><date>2024</date><risdate>2024</risdate><volume>40</volume><spage>1740</spage><epage>1757</epage><pages>1740-1757</pages><issn>1552-3098</issn><eissn>1941-0468</eissn><coden>ITREAE</coden><abstract>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TRO.2024.3353550</doi><tpages>18</tpages><orcidid>https://orcid.org/0000-0002-5176-3767</orcidid><orcidid>https://orcid.org/0000-0002-7871-3663</orcidid><orcidid>https://orcid.org/0000-0001-6482-009X</orcidid><orcidid>https://orcid.org/0000-0002-7580-9037</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1552-3098 |
ispartof | IEEE transactions on robotics, 2024, Vol.40, p.1740-1757 |
issn | 1552-3098 1941-0468 |
language | eng |
recordid | cdi_crossref_primary_10_1109_TRO_2024_3353550 |
source | IEEE Electronic Library (IEL) |
subjects | Aerial robotics applications autonomous drone cinematography camera intrinsics Cameras Cinematography Depth of field Drones model predicted control (MPC) Nonlinear control Parameters Predictive control Recording Robots Target tracking Tracking Trajectory |
title | CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T18%3A12%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=CineMPC:%20A%20Fully%20Autonomous%20Drone%20Cinematography%20System%20Incorporating%20Zoom,%20Focus,%20Pose,%20and%20Scene%20Composition&rft.jtitle=IEEE%20transactions%20on%20robotics&rft.au=Pueyo,%20Pablo&rft.date=2024&rft.volume=40&rft.spage=1740&rft.epage=1757&rft.pages=1740-1757&rft.issn=1552-3098&rft.eissn=1941-0468&rft.coden=ITREAE&rft_id=info:doi/10.1109/TRO.2024.3353550&rft_dat=%3Cproquest_cross%3E2929260099%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2929260099&rft_id=info:pmid/&rft_ieee_id=10398502&rfr_iscdi=true |