Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation
3D reconstruction of medical imaging from 2D images has become an increasingly interesting topic with the development of deep learning models in recent years. Previous studies in 3D reconstruction from limited X-ray images mainly rely on learning from paired 2D and 3D images, where the reconstructio...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Song, Weinan Zheng, Haoxin Tu, Dezhan Liang, Chengwen He, Lei |
description | 3D reconstruction of medical imaging from 2D images has become an
increasingly interesting topic with the development of deep learning models in
recent years. Previous studies in 3D reconstruction from limited X-ray images
mainly rely on learning from paired 2D and 3D images, where the reconstruction
quality relies on the scale and variation of collected data. This has brought
significant challenges in the collection of training data, as only a tiny
fraction of patients take two types of radiation examinations in the same
period. Although simulation from higher-dimension images could solve this
problem, the variance between real and simulated data could bring great
uncertainty at the same time. In oral reconstruction, the situation becomes
more challenging as only a single panoramic X-ray image is available, where
models need to infer the curved shape by prior individual knowledge. To
overcome these limitations, we propose Oral-3Dv2 to solve this cross-dimension
translation problem in dental healthcare by learning solely on projection
information, i.e., the projection image and trajectory of the X-ray tube. Our
model learns to represent the 3D oral structure in an implicit way by mapping
2D coordinates into density values of voxels in the 3D space. To improve
efficiency and effectiveness, we utilize a multi-head model that predicts a
bunch of voxel values in 3D space simultaneously from a 2D coordinate in the
axial plane and the dynamic sampling strategy to refine details of the density
distribution in the reconstruction result. Extensive experiments in simulated
and real data show that our model significantly outperforms existing
state-of-the-art models without learning from paired images or prior individual
knowledge. To the best of our knowledge, this is the first work of a
non-adversarial-learning-based model in 3D radiology reconstruction from a
single panoramic X-ray image. |
doi_str_mv | 10.48550/arxiv.2303.12123 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2303_12123</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2303_12123</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-1e9e26ea88b39eaf40a18e8795d419f7901d1b17bfe347a139013e7062f228d53</originalsourceid><addsrcrecordid>eNotj71ugzAURr10qNI-QKf6BaC2L2DTrUr6EylqqihDN3SB69QSGGRI2rx9Q5Lp0_mGIx3GHqSIE5Om4gnDnzvECgTEUkkFt6xcB2wiWBzUM4cFn4hvqOr8MIZ9NbrOcxu6ln-h7wK2ruLf0QaPfNnizvkd_3Xjzwn6xlVu5J-0vwj6QAP5ESfBHbux2Ax0f90Z2769bucf0Wr9vpy_rCLMNESSclIZoTEl5IQ2ESgNGZ2ndSJzq3Mha1lKXVqCRKOE0wGkRaasUqZOYcYeL9pzZNEH12I4FlNscY6Ff4MLTo8</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation</title><source>arXiv.org</source><creator>Song, Weinan ; Zheng, Haoxin ; Tu, Dezhan ; Liang, Chengwen ; He, Lei</creator><creatorcontrib>Song, Weinan ; Zheng, Haoxin ; Tu, Dezhan ; Liang, Chengwen ; He, Lei</creatorcontrib><description>3D reconstruction of medical imaging from 2D images has become an
increasingly interesting topic with the development of deep learning models in
recent years. Previous studies in 3D reconstruction from limited X-ray images
mainly rely on learning from paired 2D and 3D images, where the reconstruction
quality relies on the scale and variation of collected data. This has brought
significant challenges in the collection of training data, as only a tiny
fraction of patients take two types of radiation examinations in the same
period. Although simulation from higher-dimension images could solve this
problem, the variance between real and simulated data could bring great
uncertainty at the same time. In oral reconstruction, the situation becomes
more challenging as only a single panoramic X-ray image is available, where
models need to infer the curved shape by prior individual knowledge. To
overcome these limitations, we propose Oral-3Dv2 to solve this cross-dimension
translation problem in dental healthcare by learning solely on projection
information, i.e., the projection image and trajectory of the X-ray tube. Our
model learns to represent the 3D oral structure in an implicit way by mapping
2D coordinates into density values of voxels in the 3D space. To improve
efficiency and effectiveness, we utilize a multi-head model that predicts a
bunch of voxel values in 3D space simultaneously from a 2D coordinate in the
axial plane and the dynamic sampling strategy to refine details of the density
distribution in the reconstruction result. Extensive experiments in simulated
and real data show that our model significantly outperforms existing
state-of-the-art models without learning from paired images or prior individual
knowledge. To the best of our knowledge, this is the first work of a
non-adversarial-learning-based model in 3D radiology reconstruction from a
single panoramic X-ray image.</description><identifier>DOI: 10.48550/arxiv.2303.12123</identifier><language>eng</language><subject>Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2023-03</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2303.12123$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2303.12123$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Song, Weinan</creatorcontrib><creatorcontrib>Zheng, Haoxin</creatorcontrib><creatorcontrib>Tu, Dezhan</creatorcontrib><creatorcontrib>Liang, Chengwen</creatorcontrib><creatorcontrib>He, Lei</creatorcontrib><title>Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation</title><description>3D reconstruction of medical imaging from 2D images has become an
increasingly interesting topic with the development of deep learning models in
recent years. Previous studies in 3D reconstruction from limited X-ray images
mainly rely on learning from paired 2D and 3D images, where the reconstruction
quality relies on the scale and variation of collected data. This has brought
significant challenges in the collection of training data, as only a tiny
fraction of patients take two types of radiation examinations in the same
period. Although simulation from higher-dimension images could solve this
problem, the variance between real and simulated data could bring great
uncertainty at the same time. In oral reconstruction, the situation becomes
more challenging as only a single panoramic X-ray image is available, where
models need to infer the curved shape by prior individual knowledge. To
overcome these limitations, we propose Oral-3Dv2 to solve this cross-dimension
translation problem in dental healthcare by learning solely on projection
information, i.e., the projection image and trajectory of the X-ray tube. Our
model learns to represent the 3D oral structure in an implicit way by mapping
2D coordinates into density values of voxels in the 3D space. To improve
efficiency and effectiveness, we utilize a multi-head model that predicts a
bunch of voxel values in 3D space simultaneously from a 2D coordinate in the
axial plane and the dynamic sampling strategy to refine details of the density
distribution in the reconstruction result. Extensive experiments in simulated
and real data show that our model significantly outperforms existing
state-of-the-art models without learning from paired images or prior individual
knowledge. To the best of our knowledge, this is the first work of a
non-adversarial-learning-based model in 3D radiology reconstruction from a
single panoramic X-ray image.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71ugzAURr10qNI-QKf6BaC2L2DTrUr6EylqqihDN3SB69QSGGRI2rx9Q5Lp0_mGIx3GHqSIE5Om4gnDnzvECgTEUkkFt6xcB2wiWBzUM4cFn4hvqOr8MIZ9NbrOcxu6ln-h7wK2ruLf0QaPfNnizvkd_3Xjzwn6xlVu5J-0vwj6QAP5ESfBHbux2Ax0f90Z2769bucf0Wr9vpy_rCLMNESSclIZoTEl5IQ2ESgNGZ2ndSJzq3Mha1lKXVqCRKOE0wGkRaasUqZOYcYeL9pzZNEH12I4FlNscY6Ff4MLTo8</recordid><startdate>20230321</startdate><enddate>20230321</enddate><creator>Song, Weinan</creator><creator>Zheng, Haoxin</creator><creator>Tu, Dezhan</creator><creator>Liang, Chengwen</creator><creator>He, Lei</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230321</creationdate><title>Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation</title><author>Song, Weinan ; Zheng, Haoxin ; Tu, Dezhan ; Liang, Chengwen ; He, Lei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-1e9e26ea88b39eaf40a18e8795d419f7901d1b17bfe347a139013e7062f228d53</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Song, Weinan</creatorcontrib><creatorcontrib>Zheng, Haoxin</creatorcontrib><creatorcontrib>Tu, Dezhan</creatorcontrib><creatorcontrib>Liang, Chengwen</creatorcontrib><creatorcontrib>He, Lei</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Song, Weinan</au><au>Zheng, Haoxin</au><au>Tu, Dezhan</au><au>Liang, Chengwen</au><au>He, Lei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation</atitle><date>2023-03-21</date><risdate>2023</risdate><abstract>3D reconstruction of medical imaging from 2D images has become an
increasingly interesting topic with the development of deep learning models in
recent years. Previous studies in 3D reconstruction from limited X-ray images
mainly rely on learning from paired 2D and 3D images, where the reconstruction
quality relies on the scale and variation of collected data. This has brought
significant challenges in the collection of training data, as only a tiny
fraction of patients take two types of radiation examinations in the same
period. Although simulation from higher-dimension images could solve this
problem, the variance between real and simulated data could bring great
uncertainty at the same time. In oral reconstruction, the situation becomes
more challenging as only a single panoramic X-ray image is available, where
models need to infer the curved shape by prior individual knowledge. To
overcome these limitations, we propose Oral-3Dv2 to solve this cross-dimension
translation problem in dental healthcare by learning solely on projection
information, i.e., the projection image and trajectory of the X-ray tube. Our
model learns to represent the 3D oral structure in an implicit way by mapping
2D coordinates into density values of voxels in the 3D space. To improve
efficiency and effectiveness, we utilize a multi-head model that predicts a
bunch of voxel values in 3D space simultaneously from a 2D coordinate in the
axial plane and the dynamic sampling strategy to refine details of the density
distribution in the reconstruction result. Extensive experiments in simulated
and real data show that our model significantly outperforms existing
state-of-the-art models without learning from paired images or prior individual
knowledge. To the best of our knowledge, this is the first work of a
non-adversarial-learning-based model in 3D radiology reconstruction from a
single panoramic X-ray image.</abstract><doi>10.48550/arxiv.2303.12123</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2303.12123 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2303_12123 |
source | arXiv.org |
subjects | Computer Science - Computer Vision and Pattern Recognition |
title | Oral-3Dv2: 3D Oral Reconstruction from Panoramic X-Ray Imaging with Implicit Neural Representation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-24T20%3A44%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Oral-3Dv2:%203D%20Oral%20Reconstruction%20from%20Panoramic%20X-Ray%20Imaging%20with%20Implicit%20Neural%20Representation&rft.au=Song,%20Weinan&rft.date=2023-03-21&rft_id=info:doi/10.48550/arxiv.2303.12123&rft_dat=%3Carxiv_GOX%3E2303_12123%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |