CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition

We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, Cine...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on robotics 2024, Vol.40, p.1740-1757
Hauptverfasser: Pueyo, Pablo, Dendarieta, Juan, Montijano, Eduardo, Murillo, Ana Cristina, Schwager, Mac
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1757
container_issue
container_start_page 1740
container_title IEEE transactions on robotics
container_volume 40
creator Pueyo, Pablo
Dendarieta, Juan
Montijano, Eduardo
Murillo, Ana Cristina
Schwager, Mac
description We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.
doi_str_mv 10.1109/TRO.2024.3353550
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1109_TRO_2024_3353550</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10398502</ieee_id><sourcerecordid>2929260099</sourcerecordid><originalsourceid>FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</originalsourceid><addsrcrecordid>eNpNkD1PwzAQhiMEElDYGRgssTblbMdNjq0qFJCKWvGxsETGcSCo8QU7GfLvSVQGdMPd8LzvSU8UXXCYcQ54_fq8mQkQyUxKJZWCg-iEY8JjSObZ4XArJWIJmB1HpyF8w0AiyJPILStnn7bLG7Zgq26369mia8lRTV1gt56cZSNR65Y-vW6-evbSh9bW7NEZ8g153Vbuk70T1VO2ItOFKdtSsFOmXcFejB0LqG4oVG1F7iw6KvUu2PO_PYneVnevy4d4vbl_XC7WsRFZ2sYi_ZAfKEudGkwRU1Am0zbBDIUoALRKiwTN3IIRymrkStpC80xwTI0uslJOoqt9b-Ppp7Ohzb-p8254mQscZg6AOFCwp4ynELwt88ZXtfZ9ziEfreaD1Xy0mv9ZHSKX-0hlrf2HS8wUCPkLwzNzEw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2929260099</pqid></control><display><type>article</type><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><source>IEEE Electronic Library (IEL)</source><creator>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</creator><creatorcontrib>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</creatorcontrib><description>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</description><identifier>ISSN: 1552-3098</identifier><identifier>EISSN: 1941-0468</identifier><identifier>DOI: 10.1109/TRO.2024.3353550</identifier><identifier>CODEN: ITREAE</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Aerial robotics applications ; autonomous drone cinematography ; camera intrinsics ; Cameras ; Cinematography ; Depth of field ; Drones ; model predicted control (MPC) ; Nonlinear control ; Parameters ; Predictive control ; Recording ; Robots ; Target tracking ; Tracking ; Trajectory</subject><ispartof>IEEE transactions on robotics, 2024, Vol.40, p.1740-1757</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2024</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</cites><orcidid>0000-0002-5176-3767 ; 0000-0002-7871-3663 ; 0000-0001-6482-009X ; 0000-0002-7580-9037</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10398502$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,778,782,794,4012,27910,27911,27912,54745</link.rule.ids></links><search><creatorcontrib>Pueyo, Pablo</creatorcontrib><creatorcontrib>Dendarieta, Juan</creatorcontrib><creatorcontrib>Montijano, Eduardo</creatorcontrib><creatorcontrib>Murillo, Ana Cristina</creatorcontrib><creatorcontrib>Schwager, Mac</creatorcontrib><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><title>IEEE transactions on robotics</title><addtitle>TRO</addtitle><description>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</description><subject>Aerial robotics applications</subject><subject>autonomous drone cinematography</subject><subject>camera intrinsics</subject><subject>Cameras</subject><subject>Cinematography</subject><subject>Depth of field</subject><subject>Drones</subject><subject>model predicted control (MPC)</subject><subject>Nonlinear control</subject><subject>Parameters</subject><subject>Predictive control</subject><subject>Recording</subject><subject>Robots</subject><subject>Target tracking</subject><subject>Tracking</subject><subject>Trajectory</subject><issn>1552-3098</issn><issn>1941-0468</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><recordid>eNpNkD1PwzAQhiMEElDYGRgssTblbMdNjq0qFJCKWvGxsETGcSCo8QU7GfLvSVQGdMPd8LzvSU8UXXCYcQ54_fq8mQkQyUxKJZWCg-iEY8JjSObZ4XArJWIJmB1HpyF8w0AiyJPILStnn7bLG7Zgq26369mia8lRTV1gt56cZSNR65Y-vW6-evbSh9bW7NEZ8g153Vbuk70T1VO2ItOFKdtSsFOmXcFejB0LqG4oVG1F7iw6KvUu2PO_PYneVnevy4d4vbl_XC7WsRFZ2sYi_ZAfKEudGkwRU1Am0zbBDIUoALRKiwTN3IIRymrkStpC80xwTI0uslJOoqt9b-Ppp7Ohzb-p8254mQscZg6AOFCwp4ynELwt88ZXtfZ9ziEfreaD1Xy0mv9ZHSKX-0hlrf2HS8wUCPkLwzNzEw</recordid><startdate>2024</startdate><enddate>2024</enddate><creator>Pueyo, Pablo</creator><creator>Dendarieta, Juan</creator><creator>Montijano, Eduardo</creator><creator>Murillo, Ana Cristina</creator><creator>Schwager, Mac</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7TB</scope><scope>8FD</scope><scope>FR3</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-5176-3767</orcidid><orcidid>https://orcid.org/0000-0002-7871-3663</orcidid><orcidid>https://orcid.org/0000-0001-6482-009X</orcidid><orcidid>https://orcid.org/0000-0002-7580-9037</orcidid></search><sort><creationdate>2024</creationdate><title>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</title><author>Pueyo, Pablo ; Dendarieta, Juan ; Montijano, Eduardo ; Murillo, Ana Cristina ; Schwager, Mac</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c287t-27b3b93fa7c9799705c8ae498922d00a57d49c6e0c25ea9153eda182197cad8f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Aerial robotics applications</topic><topic>autonomous drone cinematography</topic><topic>camera intrinsics</topic><topic>Cameras</topic><topic>Cinematography</topic><topic>Depth of field</topic><topic>Drones</topic><topic>model predicted control (MPC)</topic><topic>Nonlinear control</topic><topic>Parameters</topic><topic>Predictive control</topic><topic>Recording</topic><topic>Robots</topic><topic>Target tracking</topic><topic>Tracking</topic><topic>Trajectory</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Pueyo, Pablo</creatorcontrib><creatorcontrib>Dendarieta, Juan</creatorcontrib><creatorcontrib>Montijano, Eduardo</creatorcontrib><creatorcontrib>Murillo, Ana Cristina</creatorcontrib><creatorcontrib>Schwager, Mac</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Mechanical &amp; Transportation Engineering Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on robotics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Pueyo, Pablo</au><au>Dendarieta, Juan</au><au>Montijano, Eduardo</au><au>Murillo, Ana Cristina</au><au>Schwager, Mac</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition</atitle><jtitle>IEEE transactions on robotics</jtitle><stitle>TRO</stitle><date>2024</date><risdate>2024</risdate><volume>40</volume><spage>1740</spage><epage>1757</epage><pages>1740-1757</pages><issn>1552-3098</issn><eissn>1941-0468</eissn><coden>ITREAE</coden><abstract>We present CineMPC, a complete cinematographic system that autonomously controls a drone to film multiple targets recording user-specified aesthetic objectives. Existing solutions in autonomous cinematography control only the camera extrinsics, namely, its position and orientation. In contrast, CineMPC is the first solution that includes the camera intrinsic parameters in the control loop, which are essential tools for controlling cinematographic effects such as focus, depth of field, and zoom. The system estimates the relative poses between the targets and the camera from an RGB-D image and optimizes a trajectory for the extrinsic and intrinsic camera parameters to film the artistic and technical requirements specified by the user. The drone and the camera are controlled in a nonlinear model predicted control (MPC) loop by reoptimizing the trajectory at each time step in response to current conditions in the scene. The perception system of CineMPC can track the targets' position and orientation despite the camera effects. Experiments in a photo-realistic simulation and with a real platform demonstrate the capabilities of the system to achieve a full array of cinematographic effects that are not possible without the control of the intrinsics of the camera. Code for CineMPC is implemented following a modular architecture in ROS and released to the community.</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TRO.2024.3353550</doi><tpages>18</tpages><orcidid>https://orcid.org/0000-0002-5176-3767</orcidid><orcidid>https://orcid.org/0000-0002-7871-3663</orcidid><orcidid>https://orcid.org/0000-0001-6482-009X</orcidid><orcidid>https://orcid.org/0000-0002-7580-9037</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1552-3098
ispartof IEEE transactions on robotics, 2024, Vol.40, p.1740-1757
issn 1552-3098
1941-0468
language eng
recordid cdi_crossref_primary_10_1109_TRO_2024_3353550
source IEEE Electronic Library (IEL)
subjects Aerial robotics applications
autonomous drone cinematography
camera intrinsics
Cameras
Cinematography
Depth of field
Drones
model predicted control (MPC)
Nonlinear control
Parameters
Predictive control
Recording
Robots
Target tracking
Tracking
Trajectory
title CineMPC: A Fully Autonomous Drone Cinematography System Incorporating Zoom, Focus, Pose, and Scene Composition
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-15T18%3A12%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=CineMPC:%20A%20Fully%20Autonomous%20Drone%20Cinematography%20System%20Incorporating%20Zoom,%20Focus,%20Pose,%20and%20Scene%20Composition&rft.jtitle=IEEE%20transactions%20on%20robotics&rft.au=Pueyo,%20Pablo&rft.date=2024&rft.volume=40&rft.spage=1740&rft.epage=1757&rft.pages=1740-1757&rft.issn=1552-3098&rft.eissn=1941-0468&rft.coden=ITREAE&rft_id=info:doi/10.1109/TRO.2024.3353550&rft_dat=%3Cproquest_cross%3E2929260099%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2929260099&rft_id=info:pmid/&rft_ieee_id=10398502&rfr_iscdi=true