Adaptive Voxel Matching for Temporal CT Subtraction

Temporal subtraction (TS) technique calculates a subtraction image between a pair of registered images acquired from the same patient at different times. Previous studies have shown that TS is effective for visualizing pathological changes over time; therefore, TS should be a useful tool for radiolo...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of digital imaging 2020-12, Vol.33 (6), p.1543-1553
Hauptverfasser: Tanaka, Toru, Ishikawa, Ryo, Nakagomi, Keita, Miyasa, Kazuhiro, Satoh, Kiyohide, Yakami, Masahiro, Akasaka, Thai, Onoue, Koji, Kubo, Takeshi, Nishio, Mizuho, Emoto, Yutaka, Togashi, Kaori
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1553
container_issue 6
container_start_page 1543
container_title Journal of digital imaging
container_volume 33
creator Tanaka, Toru
Ishikawa, Ryo
Nakagomi, Keita
Miyasa, Kazuhiro
Satoh, Kiyohide
Yakami, Masahiro
Akasaka, Thai
Onoue, Koji
Kubo, Takeshi
Nishio, Mizuho
Emoto, Yutaka
Togashi, Kaori
description Temporal subtraction (TS) technique calculates a subtraction image between a pair of registered images acquired from the same patient at different times. Previous studies have shown that TS is effective for visualizing pathological changes over time; therefore, TS should be a useful tool for radiologists. However, artifacts caused by partial volume effects degrade the quality of thick-slice subtraction images, even with accurate image registration. Here, we propose a subtraction method for reducing artifacts in thick-slice images and discuss its implementation in high-speed processing. The proposed method is based on voxel matching, which reduces artifacts by considering gaps in discretized positions of two images in subtraction calculations. There are two different features between the proposed method and conventional voxel matching: (1) the size of a searching region to reduce artifacts is determined based on discretized position gaps between images and (2) the searching region is set on both images for symmetrical subtraction. The proposed method is implemented by adopting an accelerated subtraction calculation method that exploit the nature of liner interpolation for calculating the signal value at a point among discretized positions. We quantitatively evaluated the proposed method using synthetic data and qualitatively using clinical data interpreted by radiologists. The evaluation showed that the proposed method was superior to conventional methods. Moreover, the processing speed using the proposed method was almost unchanged from that of the conventional methods. The results indicate that the proposed method can improve the quality of subtraction images acquired from thick-slice images.
doi_str_mv 10.1007/s10278-020-00376-4
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_7728871</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2473221537</sourcerecordid><originalsourceid>FETCH-LOGICAL-c474t-e0b68aa987a21fd3b44a362b87197bb93865b02be5d5940113e71927af059db53</originalsourceid><addsrcrecordid>eNp9kTlPxDAQhS0EguX4AxQoEg1NYHwk4zRIaMUlgShYEJ1lJ84SlI0XO0Hw7zEsd0E1xfvmzfEI2aawTwHwIFBgKFNgkAJwzFOxREY0pzJFhnfLZASywJRKWayR9RAeAChmKFbJGufAMprnI8KPKj3vmyeb3Lpn2yaXui_vm26a1M4nEzubO6_bZDxJrgfTe132jes2yUqt22C3PuoGuTk5nozP0our0_Px0UVaChR9asHkUutComa0rrgRQvOcGYm0QGMKLvPMADM2q7JCAKXcRoWhriErKpPxDXK48J0PZmar0nZxg1bNfTPT_kU53ajfStfcq6l7UohMxinRYO_DwLvHwYZezZpQ2rbVnXVDUEyIgqKMT4zo7h_0wQ2-i-dFCjljNOMYKbagSu9C8Lb-WoaCestELTJRMRP1nokSsWnn5xlfLZ8hRIAvgBClbmr99-x_bF8BeCCV2g</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2473221537</pqid></control><display><type>article</type><title>Adaptive Voxel Matching for Temporal CT Subtraction</title><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><source>PubMed Central</source><creator>Tanaka, Toru ; Ishikawa, Ryo ; Nakagomi, Keita ; Miyasa, Kazuhiro ; Satoh, Kiyohide ; Yakami, Masahiro ; Akasaka, Thai ; Onoue, Koji ; Kubo, Takeshi ; Nishio, Mizuho ; Emoto, Yutaka ; Togashi, Kaori</creator><creatorcontrib>Tanaka, Toru ; Ishikawa, Ryo ; Nakagomi, Keita ; Miyasa, Kazuhiro ; Satoh, Kiyohide ; Yakami, Masahiro ; Akasaka, Thai ; Onoue, Koji ; Kubo, Takeshi ; Nishio, Mizuho ; Emoto, Yutaka ; Togashi, Kaori</creatorcontrib><description>Temporal subtraction (TS) technique calculates a subtraction image between a pair of registered images acquired from the same patient at different times. Previous studies have shown that TS is effective for visualizing pathological changes over time; therefore, TS should be a useful tool for radiologists. However, artifacts caused by partial volume effects degrade the quality of thick-slice subtraction images, even with accurate image registration. Here, we propose a subtraction method for reducing artifacts in thick-slice images and discuss its implementation in high-speed processing. The proposed method is based on voxel matching, which reduces artifacts by considering gaps in discretized positions of two images in subtraction calculations. There are two different features between the proposed method and conventional voxel matching: (1) the size of a searching region to reduce artifacts is determined based on discretized position gaps between images and (2) the searching region is set on both images for symmetrical subtraction. The proposed method is implemented by adopting an accelerated subtraction calculation method that exploit the nature of liner interpolation for calculating the signal value at a point among discretized positions. We quantitatively evaluated the proposed method using synthetic data and qualitatively using clinical data interpreted by radiologists. The evaluation showed that the proposed method was superior to conventional methods. Moreover, the processing speed using the proposed method was almost unchanged from that of the conventional methods. The results indicate that the proposed method can improve the quality of subtraction images acquired from thick-slice images.</description><identifier>ISSN: 0897-1889</identifier><identifier>EISSN: 1618-727X</identifier><identifier>DOI: 10.1007/s10278-020-00376-4</identifier><identifier>PMID: 33025166</identifier><language>eng</language><publisher>Cham: Springer International Publishing</publisher><subject>Discretization ; Image acquisition ; Image quality ; Image registration ; Imaging ; Interpolation ; Matching ; Mathematical analysis ; Medicine ; Medicine &amp; Public Health ; Original Paper ; Radiology ; Searching</subject><ispartof>Journal of digital imaging, 2020-12, Vol.33 (6), p.1543-1553</ispartof><rights>Society for Imaging Informatics in Medicine 2020</rights><rights>Society for Imaging Informatics in Medicine 2020.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c474t-e0b68aa987a21fd3b44a362b87197bb93865b02be5d5940113e71927af059db53</citedby><cites>FETCH-LOGICAL-c474t-e0b68aa987a21fd3b44a362b87197bb93865b02be5d5940113e71927af059db53</cites><orcidid>0000-0003-2498-4289</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC7728871/pdf/$$EPDF$$P50$$Gpubmedcentral$$H</linktopdf><linktohtml>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC7728871/$$EHTML$$P50$$Gpubmedcentral$$H</linktohtml><link.rule.ids>230,314,723,776,780,881,27901,27902,53766,53768</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/33025166$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Tanaka, Toru</creatorcontrib><creatorcontrib>Ishikawa, Ryo</creatorcontrib><creatorcontrib>Nakagomi, Keita</creatorcontrib><creatorcontrib>Miyasa, Kazuhiro</creatorcontrib><creatorcontrib>Satoh, Kiyohide</creatorcontrib><creatorcontrib>Yakami, Masahiro</creatorcontrib><creatorcontrib>Akasaka, Thai</creatorcontrib><creatorcontrib>Onoue, Koji</creatorcontrib><creatorcontrib>Kubo, Takeshi</creatorcontrib><creatorcontrib>Nishio, Mizuho</creatorcontrib><creatorcontrib>Emoto, Yutaka</creatorcontrib><creatorcontrib>Togashi, Kaori</creatorcontrib><title>Adaptive Voxel Matching for Temporal CT Subtraction</title><title>Journal of digital imaging</title><addtitle>J Digit Imaging</addtitle><addtitle>J Digit Imaging</addtitle><description>Temporal subtraction (TS) technique calculates a subtraction image between a pair of registered images acquired from the same patient at different times. Previous studies have shown that TS is effective for visualizing pathological changes over time; therefore, TS should be a useful tool for radiologists. However, artifacts caused by partial volume effects degrade the quality of thick-slice subtraction images, even with accurate image registration. Here, we propose a subtraction method for reducing artifacts in thick-slice images and discuss its implementation in high-speed processing. The proposed method is based on voxel matching, which reduces artifacts by considering gaps in discretized positions of two images in subtraction calculations. There are two different features between the proposed method and conventional voxel matching: (1) the size of a searching region to reduce artifacts is determined based on discretized position gaps between images and (2) the searching region is set on both images for symmetrical subtraction. The proposed method is implemented by adopting an accelerated subtraction calculation method that exploit the nature of liner interpolation for calculating the signal value at a point among discretized positions. We quantitatively evaluated the proposed method using synthetic data and qualitatively using clinical data interpreted by radiologists. The evaluation showed that the proposed method was superior to conventional methods. Moreover, the processing speed using the proposed method was almost unchanged from that of the conventional methods. The results indicate that the proposed method can improve the quality of subtraction images acquired from thick-slice images.</description><subject>Discretization</subject><subject>Image acquisition</subject><subject>Image quality</subject><subject>Image registration</subject><subject>Imaging</subject><subject>Interpolation</subject><subject>Matching</subject><subject>Mathematical analysis</subject><subject>Medicine</subject><subject>Medicine &amp; Public Health</subject><subject>Original Paper</subject><subject>Radiology</subject><subject>Searching</subject><issn>0897-1889</issn><issn>1618-727X</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9kTlPxDAQhS0EguX4AxQoEg1NYHwk4zRIaMUlgShYEJ1lJ84SlI0XO0Hw7zEsd0E1xfvmzfEI2aawTwHwIFBgKFNgkAJwzFOxREY0pzJFhnfLZASywJRKWayR9RAeAChmKFbJGufAMprnI8KPKj3vmyeb3Lpn2yaXui_vm26a1M4nEzubO6_bZDxJrgfTe132jes2yUqt22C3PuoGuTk5nozP0our0_Px0UVaChR9asHkUutComa0rrgRQvOcGYm0QGMKLvPMADM2q7JCAKXcRoWhriErKpPxDXK48J0PZmar0nZxg1bNfTPT_kU53ajfStfcq6l7UohMxinRYO_DwLvHwYZezZpQ2rbVnXVDUEyIgqKMT4zo7h_0wQ2-i-dFCjljNOMYKbagSu9C8Lb-WoaCestELTJRMRP1nokSsWnn5xlfLZ8hRIAvgBClbmr99-x_bF8BeCCV2g</recordid><startdate>20201201</startdate><enddate>20201201</enddate><creator>Tanaka, Toru</creator><creator>Ishikawa, Ryo</creator><creator>Nakagomi, Keita</creator><creator>Miyasa, Kazuhiro</creator><creator>Satoh, Kiyohide</creator><creator>Yakami, Masahiro</creator><creator>Akasaka, Thai</creator><creator>Onoue, Koji</creator><creator>Kubo, Takeshi</creator><creator>Nishio, Mizuho</creator><creator>Emoto, Yutaka</creator><creator>Togashi, Kaori</creator><general>Springer International Publishing</general><general>Springer Nature B.V</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7QO</scope><scope>7RV</scope><scope>7SC</scope><scope>7TK</scope><scope>7X7</scope><scope>7XB</scope><scope>88E</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FH</scope><scope>8FI</scope><scope>8FJ</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BBNVY</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>BHPHI</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FR3</scope><scope>FYUFA</scope><scope>GHDGH</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K9.</scope><scope>KB0</scope><scope>L7M</scope><scope>LK8</scope><scope>L~C</scope><scope>L~D</scope><scope>M0S</scope><scope>M1P</scope><scope>M7P</scope><scope>NAPCQ</scope><scope>P5Z</scope><scope>P62</scope><scope>P64</scope><scope>PHGZM</scope><scope>PHGZT</scope><scope>PJZUB</scope><scope>PKEHL</scope><scope>PPXIY</scope><scope>PQEST</scope><scope>PQGLB</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0003-2498-4289</orcidid></search><sort><creationdate>20201201</creationdate><title>Adaptive Voxel Matching for Temporal CT Subtraction</title><author>Tanaka, Toru ; Ishikawa, Ryo ; Nakagomi, Keita ; Miyasa, Kazuhiro ; Satoh, Kiyohide ; Yakami, Masahiro ; Akasaka, Thai ; Onoue, Koji ; Kubo, Takeshi ; Nishio, Mizuho ; Emoto, Yutaka ; Togashi, Kaori</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c474t-e0b68aa987a21fd3b44a362b87197bb93865b02be5d5940113e71927af059db53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Discretization</topic><topic>Image acquisition</topic><topic>Image quality</topic><topic>Image registration</topic><topic>Imaging</topic><topic>Interpolation</topic><topic>Matching</topic><topic>Mathematical analysis</topic><topic>Medicine</topic><topic>Medicine &amp; Public Health</topic><topic>Original Paper</topic><topic>Radiology</topic><topic>Searching</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Tanaka, Toru</creatorcontrib><creatorcontrib>Ishikawa, Ryo</creatorcontrib><creatorcontrib>Nakagomi, Keita</creatorcontrib><creatorcontrib>Miyasa, Kazuhiro</creatorcontrib><creatorcontrib>Satoh, Kiyohide</creatorcontrib><creatorcontrib>Yakami, Masahiro</creatorcontrib><creatorcontrib>Akasaka, Thai</creatorcontrib><creatorcontrib>Onoue, Koji</creatorcontrib><creatorcontrib>Kubo, Takeshi</creatorcontrib><creatorcontrib>Nishio, Mizuho</creatorcontrib><creatorcontrib>Emoto, Yutaka</creatorcontrib><creatorcontrib>Togashi, Kaori</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Biotechnology Research Abstracts</collection><collection>Nursing &amp; Allied Health Database</collection><collection>Computer and Information Systems Abstracts</collection><collection>Neurosciences Abstracts</collection><collection>Health &amp; Medical Collection</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>Medical Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Natural Science Collection</collection><collection>Hospital Premium Collection</collection><collection>Hospital Premium Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>Biological Science Collection</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>Natural Science Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Engineering Research Database</collection><collection>Health Research Premium Collection</collection><collection>Health Research Premium Collection (Alumni)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>Nursing &amp; Allied Health Database (Alumni Edition)</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>ProQuest Biological Science Collection</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>Health &amp; Medical Collection (Alumni Edition)</collection><collection>Medical Database</collection><collection>Biological Science Database</collection><collection>Nursing &amp; Allied Health Premium</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>ProQuest Central (New)</collection><collection>ProQuest One Academic (New)</collection><collection>ProQuest Health &amp; Medical Research Collection</collection><collection>ProQuest One Academic Middle East (New)</collection><collection>ProQuest One Health &amp; Nursing</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Applied &amp; Life Sciences</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Journal of digital imaging</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Tanaka, Toru</au><au>Ishikawa, Ryo</au><au>Nakagomi, Keita</au><au>Miyasa, Kazuhiro</au><au>Satoh, Kiyohide</au><au>Yakami, Masahiro</au><au>Akasaka, Thai</au><au>Onoue, Koji</au><au>Kubo, Takeshi</au><au>Nishio, Mizuho</au><au>Emoto, Yutaka</au><au>Togashi, Kaori</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Adaptive Voxel Matching for Temporal CT Subtraction</atitle><jtitle>Journal of digital imaging</jtitle><stitle>J Digit Imaging</stitle><addtitle>J Digit Imaging</addtitle><date>2020-12-01</date><risdate>2020</risdate><volume>33</volume><issue>6</issue><spage>1543</spage><epage>1553</epage><pages>1543-1553</pages><issn>0897-1889</issn><eissn>1618-727X</eissn><abstract>Temporal subtraction (TS) technique calculates a subtraction image between a pair of registered images acquired from the same patient at different times. Previous studies have shown that TS is effective for visualizing pathological changes over time; therefore, TS should be a useful tool for radiologists. However, artifacts caused by partial volume effects degrade the quality of thick-slice subtraction images, even with accurate image registration. Here, we propose a subtraction method for reducing artifacts in thick-slice images and discuss its implementation in high-speed processing. The proposed method is based on voxel matching, which reduces artifacts by considering gaps in discretized positions of two images in subtraction calculations. There are two different features between the proposed method and conventional voxel matching: (1) the size of a searching region to reduce artifacts is determined based on discretized position gaps between images and (2) the searching region is set on both images for symmetrical subtraction. The proposed method is implemented by adopting an accelerated subtraction calculation method that exploit the nature of liner interpolation for calculating the signal value at a point among discretized positions. We quantitatively evaluated the proposed method using synthetic data and qualitatively using clinical data interpreted by radiologists. The evaluation showed that the proposed method was superior to conventional methods. Moreover, the processing speed using the proposed method was almost unchanged from that of the conventional methods. The results indicate that the proposed method can improve the quality of subtraction images acquired from thick-slice images.</abstract><cop>Cham</cop><pub>Springer International Publishing</pub><pmid>33025166</pmid><doi>10.1007/s10278-020-00376-4</doi><tpages>11</tpages><orcidid>https://orcid.org/0000-0003-2498-4289</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0897-1889
ispartof Journal of digital imaging, 2020-12, Vol.33 (6), p.1543-1553
issn 0897-1889
1618-727X
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_7728871
source Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals; PubMed Central
subjects Discretization
Image acquisition
Image quality
Image registration
Imaging
Interpolation
Matching
Mathematical analysis
Medicine
Medicine & Public Health
Original Paper
Radiology
Searching
title Adaptive Voxel Matching for Temporal CT Subtraction
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-21T17%3A36%3A10IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Adaptive%20Voxel%20Matching%20for%20Temporal%20CT%20Subtraction&rft.jtitle=Journal%20of%20digital%20imaging&rft.au=Tanaka,%20Toru&rft.date=2020-12-01&rft.volume=33&rft.issue=6&rft.spage=1543&rft.epage=1553&rft.pages=1543-1553&rft.issn=0897-1889&rft.eissn=1618-727X&rft_id=info:doi/10.1007/s10278-020-00376-4&rft_dat=%3Cproquest_pubme%3E2473221537%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2473221537&rft_id=info:pmid/33025166&rfr_iscdi=true