Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models

Query by Image Content (QBIC), subsequently known as Content-Based Image Retrieval (CBIR) systems, may offer a more advantageous solution in a variety of applications, including medical, meteorological, search by image, and others. Such systems primarily use similarity matching algorithms to compare...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2023-01, Vol.11, p.1-1
Hauptverfasser: Sabry, Eman S., Elagooz, Salah, Abd El-Samie, Fathi E., El-Shafai, Walid, El-Bahnasawy, Nirmeen A., Banby, Ghada El, Algarni, Abeer D., Soliman, Naglaa F., Ramadan, Rabie A.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue
container_start_page 1
container_title IEEE access
container_volume 11
creator Sabry, Eman S.
Elagooz, Salah
Abd El-Samie, Fathi E.
El-Shafai, Walid
El-Bahnasawy, Nirmeen A.
Banby, Ghada El
Algarni, Abeer D.
Soliman, Naglaa F.
Ramadan, Rabie A.
description Query by Image Content (QBIC), subsequently known as Content-Based Image Retrieval (CBIR) systems, may offer a more advantageous solution in a variety of applications, including medical, meteorological, search by image, and others. Such systems primarily use similarity matching algorithms to compare image content to get their relevance from databases. They are essentially measuring the spatial distance between extracted visual features from a query image and their correspondence in the dataset. One of the most challenging query retrieval problems is Facial Sketched-Real Image Retrieval (FSRIR), which is content similarity matching based. These facial retrieving systems are employed in a variety of contexts, including criminal justice. The difficulties of retrieving such sorts come from the composition of the human face and its distinctive parts. In addition, the comparison between these images is made from two different domains. Besides, to our knowledge, there is a rare existence of large-scale facial datasets that can be used to evolve the performance of the retrieving system. The success of the retrieval process is governed by the method used to calculate similarity and the efficient representation of compared images. However, by effectively representing visual features, the main challenge-posing component of such approaches might be resolved. Hence, this paper has several contributions that fill the research gap in content-based similarity matching and retrieving as follows: 1) The first contribution is extending the Chinese University Face Sketch (CUFS) dataset by including augmented images, introducing to the community a novel dataset named Extended Sketched-Real Image Retrieval (ESRIR). The CUFS dataset has been extended from 100 images to include 53,000 facial sketches and 53,000 real facial images. 2) The paper's second contribution is proposing three new algorithms for sketched-real image retrieving based on convolutional autoencoder, Infogan, and Vision Transformer unsupervised models for large datasets. 3) Furthermore, to meet the subjective demands of the users because of the prevalence of multiple query formats. The third contribution of this paper is to train and assess the proposed algorithms across two additional facial datasets of various image sorts. 4) Recently, the majority of people have preferred searching for brand logo images, but it may be tricky to separate certain brand logo features from their alternatives and even from other f
doi_str_mv 10.1109/ACCESS.2023.3241858
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_ieee_primary_10035290</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10035290</ieee_id><doaj_id>oai_doaj_org_article_2892b5285f594989997da146b0593552</doaj_id><sourcerecordid>2781848563</sourcerecordid><originalsourceid>FETCH-LOGICAL-c409t-8a3a951053e5ac2ca3166088246680d1049992011ed7c5916a19b00b80e377213</originalsourceid><addsrcrecordid>eNpNkd1LwzAUxYsoKOpfoA8BX928SZo0eRzFj4Ef4KavIWtvR0bXzKQd-N8brYj3JZfDOb8LOVl2QWFKKeibWVneLhZTBoxPOcupEuogO2FU6gkXXB7-24-z8xg3kEYlSRQnWT3f2jWSV-yDw71tyRBdtyal7_a-HXrnu6TNht5jV_kawzWZd42_nz1fE9vV5N3FZCHLYLvY-LDFQN66OOww7F3EmjylTBvPsqPGthHPf9_T7O3udlk-TB5f7ufl7HFS5aD7ibLcakFBcBS2YpXlVEpQiuVSKqgp5FprBpRiXVRCU2mpXgGsFCAvCkb5aTYfubW3G7MLbmvDp_HWmR_Bh7WxoXdVi4YpzVaCKdEInWuVwEVtaS5XIDQXgiXW1cjaBf8xYOzNxg8h_UY0rFBU5UpInlx8dFXBxxiw-btKwXy3Y8Z2zHc75redlLocUw4R_yWAC6aBfwE9PYiY</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2781848563</pqid></control><display><type>article</type><title>Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models</title><source>IEEE Open Access Journals</source><source>DOAJ Directory of Open Access Journals</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Sabry, Eman S. ; Elagooz, Salah ; Abd El-Samie, Fathi E. ; El-Shafai, Walid ; El-Bahnasawy, Nirmeen A. ; Banby, Ghada El ; Algarni, Abeer D. ; Soliman, Naglaa F. ; Ramadan, Rabie A.</creator><creatorcontrib>Sabry, Eman S. ; Elagooz, Salah ; Abd El-Samie, Fathi E. ; El-Shafai, Walid ; El-Bahnasawy, Nirmeen A. ; Banby, Ghada El ; Algarni, Abeer D. ; Soliman, Naglaa F. ; Ramadan, Rabie A.</creatorcontrib><description>Query by Image Content (QBIC), subsequently known as Content-Based Image Retrieval (CBIR) systems, may offer a more advantageous solution in a variety of applications, including medical, meteorological, search by image, and others. Such systems primarily use similarity matching algorithms to compare image content to get their relevance from databases. They are essentially measuring the spatial distance between extracted visual features from a query image and their correspondence in the dataset. One of the most challenging query retrieval problems is Facial Sketched-Real Image Retrieval (FSRIR), which is content similarity matching based. These facial retrieving systems are employed in a variety of contexts, including criminal justice. The difficulties of retrieving such sorts come from the composition of the human face and its distinctive parts. In addition, the comparison between these images is made from two different domains. Besides, to our knowledge, there is a rare existence of large-scale facial datasets that can be used to evolve the performance of the retrieving system. The success of the retrieval process is governed by the method used to calculate similarity and the efficient representation of compared images. However, by effectively representing visual features, the main challenge-posing component of such approaches might be resolved. Hence, this paper has several contributions that fill the research gap in content-based similarity matching and retrieving as follows: 1) The first contribution is extending the Chinese University Face Sketch (CUFS) dataset by including augmented images, introducing to the community a novel dataset named Extended Sketched-Real Image Retrieval (ESRIR). The CUFS dataset has been extended from 100 images to include 53,000 facial sketches and 53,000 real facial images. 2) The paper's second contribution is proposing three new algorithms for sketched-real image retrieving based on convolutional autoencoder, Infogan, and Vision Transformer unsupervised models for large datasets. 3) Furthermore, to meet the subjective demands of the users because of the prevalence of multiple query formats. The third contribution of this paper is to train and assess the proposed algorithms across two additional facial datasets of various image sorts. 4) Recently, the majority of people have preferred searching for brand logo images, but it may be tricky to separate certain brand logo features from their alternatives and even from other features in an image. Thus, the fourth contribution is to compare logo image retrieval performance based on visual features derived from each of the three suggested retrieving systems. 5) The paper also proposes cloud-based energy-saving and computational complexity approaches in large-scale datasets. Due to the ubiquity of touchscreen devices, users often make drawings based on their fantasies for certain object image searches. Thus, the proposed algorithms were tested and assessed on a tough dataset of doodle-scratched human artworks. They are also studied for a multi-category dataset to cover practically all possible image types and situations. The results are compared with the most recent algorithms found in the literature. The results show that the proposed approaches are outperforming the recent algorithms.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2023.3241858</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Algorithms ; Convolutional neural networks ; Datasets ; Face recognition ; Feature extraction ; Image retrieval ; InfoGAN ; Matching ; object matching ; objects matching ; Performance evaluation ; Queries ; Retrieval performance measures ; Similarity ; Sketched-real image retrieval ; Sketches ; Spatial distance measure ; spatial distance measurement ; System effectiveness ; Touch screens ; Transformers ; Vision transformer ; Visualization</subject><ispartof>IEEE access, 2023-01, Vol.11, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c409t-8a3a951053e5ac2ca3166088246680d1049992011ed7c5916a19b00b80e377213</citedby><cites>FETCH-LOGICAL-c409t-8a3a951053e5ac2ca3166088246680d1049992011ed7c5916a19b00b80e377213</cites><orcidid>0000-0002-7089-8851 ; 0000-0002-0281-9381 ; 0000-0001-8749-9518 ; 0000-0001-7322-1857 ; 0000-0001-7509-2120 ; 0000-0002-4542-323X</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10035290$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2102,27633,27924,27925,54933</link.rule.ids></links><search><creatorcontrib>Sabry, Eman S.</creatorcontrib><creatorcontrib>Elagooz, Salah</creatorcontrib><creatorcontrib>Abd El-Samie, Fathi E.</creatorcontrib><creatorcontrib>El-Shafai, Walid</creatorcontrib><creatorcontrib>El-Bahnasawy, Nirmeen A.</creatorcontrib><creatorcontrib>Banby, Ghada El</creatorcontrib><creatorcontrib>Algarni, Abeer D.</creatorcontrib><creatorcontrib>Soliman, Naglaa F.</creatorcontrib><creatorcontrib>Ramadan, Rabie A.</creatorcontrib><title>Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models</title><title>IEEE access</title><addtitle>Access</addtitle><description>Query by Image Content (QBIC), subsequently known as Content-Based Image Retrieval (CBIR) systems, may offer a more advantageous solution in a variety of applications, including medical, meteorological, search by image, and others. Such systems primarily use similarity matching algorithms to compare image content to get their relevance from databases. They are essentially measuring the spatial distance between extracted visual features from a query image and their correspondence in the dataset. One of the most challenging query retrieval problems is Facial Sketched-Real Image Retrieval (FSRIR), which is content similarity matching based. These facial retrieving systems are employed in a variety of contexts, including criminal justice. The difficulties of retrieving such sorts come from the composition of the human face and its distinctive parts. In addition, the comparison between these images is made from two different domains. Besides, to our knowledge, there is a rare existence of large-scale facial datasets that can be used to evolve the performance of the retrieving system. The success of the retrieval process is governed by the method used to calculate similarity and the efficient representation of compared images. However, by effectively representing visual features, the main challenge-posing component of such approaches might be resolved. Hence, this paper has several contributions that fill the research gap in content-based similarity matching and retrieving as follows: 1) The first contribution is extending the Chinese University Face Sketch (CUFS) dataset by including augmented images, introducing to the community a novel dataset named Extended Sketched-Real Image Retrieval (ESRIR). The CUFS dataset has been extended from 100 images to include 53,000 facial sketches and 53,000 real facial images. 2) The paper's second contribution is proposing three new algorithms for sketched-real image retrieving based on convolutional autoencoder, Infogan, and Vision Transformer unsupervised models for large datasets. 3) Furthermore, to meet the subjective demands of the users because of the prevalence of multiple query formats. The third contribution of this paper is to train and assess the proposed algorithms across two additional facial datasets of various image sorts. 4) Recently, the majority of people have preferred searching for brand logo images, but it may be tricky to separate certain brand logo features from their alternatives and even from other features in an image. Thus, the fourth contribution is to compare logo image retrieval performance based on visual features derived from each of the three suggested retrieving systems. 5) The paper also proposes cloud-based energy-saving and computational complexity approaches in large-scale datasets. Due to the ubiquity of touchscreen devices, users often make drawings based on their fantasies for certain object image searches. Thus, the proposed algorithms were tested and assessed on a tough dataset of doodle-scratched human artworks. They are also studied for a multi-category dataset to cover practically all possible image types and situations. The results are compared with the most recent algorithms found in the literature. The results show that the proposed approaches are outperforming the recent algorithms.</description><subject>Algorithms</subject><subject>Convolutional neural networks</subject><subject>Datasets</subject><subject>Face recognition</subject><subject>Feature extraction</subject><subject>Image retrieval</subject><subject>InfoGAN</subject><subject>Matching</subject><subject>object matching</subject><subject>objects matching</subject><subject>Performance evaluation</subject><subject>Queries</subject><subject>Retrieval performance measures</subject><subject>Similarity</subject><subject>Sketched-real image retrieval</subject><subject>Sketches</subject><subject>Spatial distance measure</subject><subject>spatial distance measurement</subject><subject>System effectiveness</subject><subject>Touch screens</subject><subject>Transformers</subject><subject>Vision transformer</subject><subject>Visualization</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkd1LwzAUxYsoKOpfoA8BX928SZo0eRzFj4Ef4KavIWtvR0bXzKQd-N8brYj3JZfDOb8LOVl2QWFKKeibWVneLhZTBoxPOcupEuogO2FU6gkXXB7-24-z8xg3kEYlSRQnWT3f2jWSV-yDw71tyRBdtyal7_a-HXrnu6TNht5jV_kawzWZd42_nz1fE9vV5N3FZCHLYLvY-LDFQN66OOww7F3EmjylTBvPsqPGthHPf9_T7O3udlk-TB5f7ufl7HFS5aD7ibLcakFBcBS2YpXlVEpQiuVSKqgp5FprBpRiXVRCU2mpXgGsFCAvCkb5aTYfubW3G7MLbmvDp_HWmR_Bh7WxoXdVi4YpzVaCKdEInWuVwEVtaS5XIDQXgiXW1cjaBf8xYOzNxg8h_UY0rFBU5UpInlx8dFXBxxiw-btKwXy3Y8Z2zHc75redlLocUw4R_yWAC6aBfwE9PYiY</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Sabry, Eman S.</creator><creator>Elagooz, Salah</creator><creator>Abd El-Samie, Fathi E.</creator><creator>El-Shafai, Walid</creator><creator>El-Bahnasawy, Nirmeen A.</creator><creator>Banby, Ghada El</creator><creator>Algarni, Abeer D.</creator><creator>Soliman, Naglaa F.</creator><creator>Ramadan, Rabie A.</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7089-8851</orcidid><orcidid>https://orcid.org/0000-0002-0281-9381</orcidid><orcidid>https://orcid.org/0000-0001-8749-9518</orcidid><orcidid>https://orcid.org/0000-0001-7322-1857</orcidid><orcidid>https://orcid.org/0000-0001-7509-2120</orcidid><orcidid>https://orcid.org/0000-0002-4542-323X</orcidid></search><sort><creationdate>20230101</creationdate><title>Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models</title><author>Sabry, Eman S. ; Elagooz, Salah ; Abd El-Samie, Fathi E. ; El-Shafai, Walid ; El-Bahnasawy, Nirmeen A. ; Banby, Ghada El ; Algarni, Abeer D. ; Soliman, Naglaa F. ; Ramadan, Rabie A.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c409t-8a3a951053e5ac2ca3166088246680d1049992011ed7c5916a19b00b80e377213</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Algorithms</topic><topic>Convolutional neural networks</topic><topic>Datasets</topic><topic>Face recognition</topic><topic>Feature extraction</topic><topic>Image retrieval</topic><topic>InfoGAN</topic><topic>Matching</topic><topic>object matching</topic><topic>objects matching</topic><topic>Performance evaluation</topic><topic>Queries</topic><topic>Retrieval performance measures</topic><topic>Similarity</topic><topic>Sketched-real image retrieval</topic><topic>Sketches</topic><topic>Spatial distance measure</topic><topic>spatial distance measurement</topic><topic>System effectiveness</topic><topic>Touch screens</topic><topic>Transformers</topic><topic>Vision transformer</topic><topic>Visualization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Sabry, Eman S.</creatorcontrib><creatorcontrib>Elagooz, Salah</creatorcontrib><creatorcontrib>Abd El-Samie, Fathi E.</creatorcontrib><creatorcontrib>El-Shafai, Walid</creatorcontrib><creatorcontrib>El-Bahnasawy, Nirmeen A.</creatorcontrib><creatorcontrib>Banby, Ghada El</creatorcontrib><creatorcontrib>Algarni, Abeer D.</creatorcontrib><creatorcontrib>Soliman, Naglaa F.</creatorcontrib><creatorcontrib>Ramadan, Rabie A.</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Sabry, Eman S.</au><au>Elagooz, Salah</au><au>Abd El-Samie, Fathi E.</au><au>El-Shafai, Walid</au><au>El-Bahnasawy, Nirmeen A.</au><au>Banby, Ghada El</au><au>Algarni, Abeer D.</au><au>Soliman, Naglaa F.</au><au>Ramadan, Rabie A.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2023-01-01</date><risdate>2023</risdate><volume>11</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>Query by Image Content (QBIC), subsequently known as Content-Based Image Retrieval (CBIR) systems, may offer a more advantageous solution in a variety of applications, including medical, meteorological, search by image, and others. Such systems primarily use similarity matching algorithms to compare image content to get their relevance from databases. They are essentially measuring the spatial distance between extracted visual features from a query image and their correspondence in the dataset. One of the most challenging query retrieval problems is Facial Sketched-Real Image Retrieval (FSRIR), which is content similarity matching based. These facial retrieving systems are employed in a variety of contexts, including criminal justice. The difficulties of retrieving such sorts come from the composition of the human face and its distinctive parts. In addition, the comparison between these images is made from two different domains. Besides, to our knowledge, there is a rare existence of large-scale facial datasets that can be used to evolve the performance of the retrieving system. The success of the retrieval process is governed by the method used to calculate similarity and the efficient representation of compared images. However, by effectively representing visual features, the main challenge-posing component of such approaches might be resolved. Hence, this paper has several contributions that fill the research gap in content-based similarity matching and retrieving as follows: 1) The first contribution is extending the Chinese University Face Sketch (CUFS) dataset by including augmented images, introducing to the community a novel dataset named Extended Sketched-Real Image Retrieval (ESRIR). The CUFS dataset has been extended from 100 images to include 53,000 facial sketches and 53,000 real facial images. 2) The paper's second contribution is proposing three new algorithms for sketched-real image retrieving based on convolutional autoencoder, Infogan, and Vision Transformer unsupervised models for large datasets. 3) Furthermore, to meet the subjective demands of the users because of the prevalence of multiple query formats. The third contribution of this paper is to train and assess the proposed algorithms across two additional facial datasets of various image sorts. 4) Recently, the majority of people have preferred searching for brand logo images, but it may be tricky to separate certain brand logo features from their alternatives and even from other features in an image. Thus, the fourth contribution is to compare logo image retrieval performance based on visual features derived from each of the three suggested retrieving systems. 5) The paper also proposes cloud-based energy-saving and computational complexity approaches in large-scale datasets. Due to the ubiquity of touchscreen devices, users often make drawings based on their fantasies for certain object image searches. Thus, the proposed algorithms were tested and assessed on a tough dataset of doodle-scratched human artworks. They are also studied for a multi-category dataset to cover practically all possible image types and situations. The results are compared with the most recent algorithms found in the literature. The results show that the proposed approaches are outperforming the recent algorithms.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2023.3241858</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0002-7089-8851</orcidid><orcidid>https://orcid.org/0000-0002-0281-9381</orcidid><orcidid>https://orcid.org/0000-0001-8749-9518</orcidid><orcidid>https://orcid.org/0000-0001-7322-1857</orcidid><orcidid>https://orcid.org/0000-0001-7509-2120</orcidid><orcidid>https://orcid.org/0000-0002-4542-323X</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2023-01, Vol.11, p.1-1
issn 2169-3536
2169-3536
language eng
recordid cdi_ieee_primary_10035290
source IEEE Open Access Journals; DOAJ Directory of Open Access Journals; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Algorithms
Convolutional neural networks
Datasets
Face recognition
Feature extraction
Image retrieval
InfoGAN
Matching
object matching
objects matching
Performance evaluation
Queries
Retrieval performance measures
Similarity
Sketched-real image retrieval
Sketches
Spatial distance measure
spatial distance measurement
System effectiveness
Touch screens
Transformers
Vision transformer
Visualization
title Image Retrieval using Convolutional Autoencoder, InfoGAN, and Vision Transformer Unsupervised Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-04T12%3A23%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Image%20Retrieval%20using%20Convolutional%20Autoencoder,%20InfoGAN,%20and%20Vision%20Transformer%20Unsupervised%20Models&rft.jtitle=IEEE%20access&rft.au=Sabry,%20Eman%20S.&rft.date=2023-01-01&rft.volume=11&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2023.3241858&rft_dat=%3Cproquest_ieee_%3E2781848563%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2781848563&rft_id=info:pmid/&rft_ieee_id=10035290&rft_doaj_id=oai_doaj_org_article_2892b5285f594989997da146b0593552&rfr_iscdi=true