GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO

Technical limitation of conventional Gradient-Based Meta Learners is their inability to adapt to scenarios where input tasks are sampled from multiple distributions. Training multiple models, with one model per distribution adds to the training time owing to increased compute. A method and system fo...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: SINGHAL, Rekha, KUNDE, Shruti Kunal, PIMPALKHUTE, Varad Anant
Format: Patent
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator SINGHAL, Rekha
KUNDE, Shruti Kunal
PIMPALKHUTE, Varad Anant
description Technical limitation of conventional Gradient-Based Meta Learners is their inability to adapt to scenarios where input tasks are sampled from multiple distributions. Training multiple models, with one model per distribution adds to the training time owing to increased compute. A method and system for generating meta-subnets for efficient model generalization in a multi-distribution scenario using Binary Mask Perceptron (BMP) technique or a Multi-modal Meta Supermasks (MMSUP) technique is provided. The BMP utilizes an adaptor which determines a binary mask, thus training only those layers which are relevant for given input distribution, leading to improved training accuracy in a cross-domain scenario. The MMSUP, further determines relevant subnets for each input distribution, thus, generalizing well as compared to standard MAML. The BMP and MMSUP, beat Multi-MAML in terms of training time as they train a single model on multiple distributions as opposed to Multi-MAML which trains multiple models.
format Patent
fullrecord <record><control><sourceid>epo_EVB</sourceid><recordid>TN_cdi_epo_espacenet_US2024160949A1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>US2024160949A1</sourcerecordid><originalsourceid>FETCH-epo_espacenet_US2024160949A13</originalsourceid><addsrcrecordid>eNqNysEKgkAQgGEvHaJ6h4HOgpoEHtd11gZ0FnZnL3YQie0UJdj7U1QP0OmHn2-dnFtkdEqIW-hRVOpDzSgejHWAxpAmZIHeNtjB13Y0vL1lIAYFfeiE0oa8OKrD53uNrBzZbbK6Trcl7n7dJHuDok9pnB9jXObpEu_xOQZfZEWZH7OqrFR--E-9AGZrM0E</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>patent</recordtype></control><display><type>patent</type><title>GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO</title><source>esp@cenet</source><creator>SINGHAL, Rekha ; KUNDE, Shruti Kunal ; PIMPALKHUTE, Varad Anant</creator><creatorcontrib>SINGHAL, Rekha ; KUNDE, Shruti Kunal ; PIMPALKHUTE, Varad Anant</creatorcontrib><description>Technical limitation of conventional Gradient-Based Meta Learners is their inability to adapt to scenarios where input tasks are sampled from multiple distributions. Training multiple models, with one model per distribution adds to the training time owing to increased compute. A method and system for generating meta-subnets for efficient model generalization in a multi-distribution scenario using Binary Mask Perceptron (BMP) technique or a Multi-modal Meta Supermasks (MMSUP) technique is provided. The BMP utilizes an adaptor which determines a binary mask, thus training only those layers which are relevant for given input distribution, leading to improved training accuracy in a cross-domain scenario. The MMSUP, further determines relevant subnets for each input distribution, thus, generalizing well as compared to standard MAML. The BMP and MMSUP, beat Multi-MAML in terms of training time as they train a single model on multiple distributions as opposed to Multi-MAML which trains multiple models.</description><language>eng</language><subject>CALCULATING ; COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS ; COMPUTING ; COUNTING ; PHYSICS</subject><creationdate>2024</creationdate><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240516&amp;DB=EPODOC&amp;CC=US&amp;NR=2024160949A1$$EHTML$$P50$$Gepo$$Hfree_for_read</linktohtml><link.rule.ids>230,308,776,881,25542,76290</link.rule.ids><linktorsrc>$$Uhttps://worldwide.espacenet.com/publicationDetails/biblio?FT=D&amp;date=20240516&amp;DB=EPODOC&amp;CC=US&amp;NR=2024160949A1$$EView_record_in_European_Patent_Office$$FView_record_in_$$GEuropean_Patent_Office$$Hfree_for_read</linktorsrc></links><search><creatorcontrib>SINGHAL, Rekha</creatorcontrib><creatorcontrib>KUNDE, Shruti Kunal</creatorcontrib><creatorcontrib>PIMPALKHUTE, Varad Anant</creatorcontrib><title>GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO</title><description>Technical limitation of conventional Gradient-Based Meta Learners is their inability to adapt to scenarios where input tasks are sampled from multiple distributions. Training multiple models, with one model per distribution adds to the training time owing to increased compute. A method and system for generating meta-subnets for efficient model generalization in a multi-distribution scenario using Binary Mask Perceptron (BMP) technique or a Multi-modal Meta Supermasks (MMSUP) technique is provided. The BMP utilizes an adaptor which determines a binary mask, thus training only those layers which are relevant for given input distribution, leading to improved training accuracy in a cross-domain scenario. The MMSUP, further determines relevant subnets for each input distribution, thus, generalizing well as compared to standard MAML. The BMP and MMSUP, beat Multi-MAML in terms of training time as they train a single model on multiple distributions as opposed to Multi-MAML which trains multiple models.</description><subject>CALCULATING</subject><subject>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</subject><subject>COMPUTING</subject><subject>COUNTING</subject><subject>PHYSICS</subject><fulltext>true</fulltext><rsrctype>patent</rsrctype><creationdate>2024</creationdate><recordtype>patent</recordtype><sourceid>EVB</sourceid><recordid>eNqNysEKgkAQgGEvHaJ6h4HOgpoEHtd11gZ0FnZnL3YQie0UJdj7U1QP0OmHn2-dnFtkdEqIW-hRVOpDzSgejHWAxpAmZIHeNtjB13Y0vL1lIAYFfeiE0oa8OKrD53uNrBzZbbK6Trcl7n7dJHuDok9pnB9jXObpEu_xOQZfZEWZH7OqrFR--E-9AGZrM0E</recordid><startdate>20240516</startdate><enddate>20240516</enddate><creator>SINGHAL, Rekha</creator><creator>KUNDE, Shruti Kunal</creator><creator>PIMPALKHUTE, Varad Anant</creator><scope>EVB</scope></search><sort><creationdate>20240516</creationdate><title>GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO</title><author>SINGHAL, Rekha ; KUNDE, Shruti Kunal ; PIMPALKHUTE, Varad Anant</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-epo_espacenet_US2024160949A13</frbrgroupid><rsrctype>patents</rsrctype><prefilter>patents</prefilter><language>eng</language><creationdate>2024</creationdate><topic>CALCULATING</topic><topic>COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS</topic><topic>COMPUTING</topic><topic>COUNTING</topic><topic>PHYSICS</topic><toplevel>online_resources</toplevel><creatorcontrib>SINGHAL, Rekha</creatorcontrib><creatorcontrib>KUNDE, Shruti Kunal</creatorcontrib><creatorcontrib>PIMPALKHUTE, Varad Anant</creatorcontrib><collection>esp@cenet</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>SINGHAL, Rekha</au><au>KUNDE, Shruti Kunal</au><au>PIMPALKHUTE, Varad Anant</au><format>patent</format><genre>patent</genre><ristype>GEN</ristype><title>GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO</title><date>2024-05-16</date><risdate>2024</risdate><abstract>Technical limitation of conventional Gradient-Based Meta Learners is their inability to adapt to scenarios where input tasks are sampled from multiple distributions. Training multiple models, with one model per distribution adds to the training time owing to increased compute. A method and system for generating meta-subnets for efficient model generalization in a multi-distribution scenario using Binary Mask Perceptron (BMP) technique or a Multi-modal Meta Supermasks (MMSUP) technique is provided. The BMP utilizes an adaptor which determines a binary mask, thus training only those layers which are relevant for given input distribution, leading to improved training accuracy in a cross-domain scenario. The MMSUP, further determines relevant subnets for each input distribution, thus, generalizing well as compared to standard MAML. The BMP and MMSUP, beat Multi-MAML in terms of training time as they train a single model on multiple distributions as opposed to Multi-MAML which trains multiple models.</abstract><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier
ispartof
issn
language eng
recordid cdi_epo_espacenet_US2024160949A1
source esp@cenet
subjects CALCULATING
COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
COMPUTING
COUNTING
PHYSICS
title GENERATING META-SUBNETS FOR EFFICIENT MODEL GENERALIZATION IN A MULTI-DISTRIBUTION SCENARIO
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-28T18%3A21%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-epo_EVB&rft_val_fmt=info:ofi/fmt:kev:mtx:patent&rft.genre=patent&rft.au=SINGHAL,%20Rekha&rft.date=2024-05-16&rft_id=info:doi/&rft_dat=%3Cepo_EVB%3EUS2024160949A1%3C/epo_EVB%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true