AutoML for Large Capacity Modeling of Meta's Ranking Systems
Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to m...
Gespeichert in:
Hauptverfasser: | , , , , , , , , , , , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Yin, Hang Liu, Kuang-Hung Sun, Mengying Chen, Yuxin Zhang, Buyun Liu, Jiang Sehgal, Vivek Panchal, Rudresh Rajnikant Hotaj, Eugen Liu, Xi Guo, Daifeng Zhang, Jamey Wang, Zhou Jiang, Shali Li, Huayu Chen, Zhengxing Chen, Wen-Yen Yang, Jiyan Wen, Wei |
description | Web-scale ranking systems at Meta serving billions of users is complex.
Improving ranking models is essential but engineering heavy. Automated Machine
Learning (AutoML) can release engineers from labor intensive work of tuning
ranking models; however, it is unknown if AutoML is efficient enough to meet
tight production timeline in real-world and, at the same time, bring additional
improvements to the strong baselines. Moreover, to achieve higher ranking
performance, there is an ever-increasing demand to scale up ranking models to
even larger capacity, which imposes more challenges on the efficiency. The
large scale of models and tight production schedule requires AutoML to
outperform human baselines by only using a small number of model evaluation
trials (around 100). We presents a sampling-based AutoML method, focusing on
neural architecture search and hyperparameter optimization, addressing these
challenges in Meta-scale production when building large capacity models. Our
approach efficiently handles large-scale data demands. It leverages a
lightweight predictor-based searcher and reinforcement learning to explore vast
search spaces, significantly reducing the number of model evaluations. Through
experiments in large capacity modeling for CTR and CVR applications, we show
that our method achieves outstanding Return on Investment (ROI) versus human
tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or
$25\%$ Query per Second (QPS) increase by only sampling one hundred models on
average from a curated search space. The proposed AutoML method has already
made real-world impact where a discovered Instagram CTR model with up to -0.36%
NE gain (over existing production baseline) was selected for large-scale online
A/B test and show statistically significant gain. These production results
proved AutoML efficacy and accelerated its adoption in ranking systems at Meta. |
doi_str_mv | 10.48550/arxiv.2311.07870 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_07870</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_07870</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-f791e32fb3d3065c7c77cbf924298bdcef3fd915590ee031f715f878d77e36343</originalsourceid><addsrcrecordid>eNotj71OwzAURr0woMIDMOGNKcHOjXNtiaWK-JMSIUH3yLHvrSLapnICIm-PWpiO9A1H3xHiRqu8tMaoe59-hu-8AK1zhRbVpXhYf81j20gek2x82pKs_dGHYV5kO0baDYetHFm2NPu7Sb77w-dp-VimmfbTlbhgv5vo-p8rsXl63NQvWfP2_Fqvm8xXqDJGpwkK7iGCqkzAgBh6dkVZONvHQAwcnTbGKSIFmlEbtmgjIkEFJazE7Z_2fL87pmHv09KdMrpzBvwCIaVBKQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><source>arXiv.org</source><creator>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</creator><creatorcontrib>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</creatorcontrib><description>Web-scale ranking systems at Meta serving billions of users is complex.
Improving ranking models is essential but engineering heavy. Automated Machine
Learning (AutoML) can release engineers from labor intensive work of tuning
ranking models; however, it is unknown if AutoML is efficient enough to meet
tight production timeline in real-world and, at the same time, bring additional
improvements to the strong baselines. Moreover, to achieve higher ranking
performance, there is an ever-increasing demand to scale up ranking models to
even larger capacity, which imposes more challenges on the efficiency. The
large scale of models and tight production schedule requires AutoML to
outperform human baselines by only using a small number of model evaluation
trials (around 100). We presents a sampling-based AutoML method, focusing on
neural architecture search and hyperparameter optimization, addressing these
challenges in Meta-scale production when building large capacity models. Our
approach efficiently handles large-scale data demands. It leverages a
lightweight predictor-based searcher and reinforcement learning to explore vast
search spaces, significantly reducing the number of model evaluations. Through
experiments in large capacity modeling for CTR and CVR applications, we show
that our method achieves outstanding Return on Investment (ROI) versus human
tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or
$25\%$ Query per Second (QPS) increase by only sampling one hundred models on
average from a curated search space. The proposed AutoML method has already
made real-world impact where a discovered Instagram CTR model with up to -0.36%
NE gain (over existing production baseline) was selected for large-scale online
A/B test and show statistically significant gain. These production results
proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</description><identifier>DOI: 10.48550/arxiv.2311.07870</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Information Retrieval</subject><creationdate>2023-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.07870$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.07870$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Yin, Hang</creatorcontrib><creatorcontrib>Liu, Kuang-Hung</creatorcontrib><creatorcontrib>Sun, Mengying</creatorcontrib><creatorcontrib>Chen, Yuxin</creatorcontrib><creatorcontrib>Zhang, Buyun</creatorcontrib><creatorcontrib>Liu, Jiang</creatorcontrib><creatorcontrib>Sehgal, Vivek</creatorcontrib><creatorcontrib>Panchal, Rudresh Rajnikant</creatorcontrib><creatorcontrib>Hotaj, Eugen</creatorcontrib><creatorcontrib>Liu, Xi</creatorcontrib><creatorcontrib>Guo, Daifeng</creatorcontrib><creatorcontrib>Zhang, Jamey</creatorcontrib><creatorcontrib>Wang, Zhou</creatorcontrib><creatorcontrib>Jiang, Shali</creatorcontrib><creatorcontrib>Li, Huayu</creatorcontrib><creatorcontrib>Chen, Zhengxing</creatorcontrib><creatorcontrib>Chen, Wen-Yen</creatorcontrib><creatorcontrib>Yang, Jiyan</creatorcontrib><creatorcontrib>Wen, Wei</creatorcontrib><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><description>Web-scale ranking systems at Meta serving billions of users is complex.
Improving ranking models is essential but engineering heavy. Automated Machine
Learning (AutoML) can release engineers from labor intensive work of tuning
ranking models; however, it is unknown if AutoML is efficient enough to meet
tight production timeline in real-world and, at the same time, bring additional
improvements to the strong baselines. Moreover, to achieve higher ranking
performance, there is an ever-increasing demand to scale up ranking models to
even larger capacity, which imposes more challenges on the efficiency. The
large scale of models and tight production schedule requires AutoML to
outperform human baselines by only using a small number of model evaluation
trials (around 100). We presents a sampling-based AutoML method, focusing on
neural architecture search and hyperparameter optimization, addressing these
challenges in Meta-scale production when building large capacity models. Our
approach efficiently handles large-scale data demands. It leverages a
lightweight predictor-based searcher and reinforcement learning to explore vast
search spaces, significantly reducing the number of model evaluations. Through
experiments in large capacity modeling for CTR and CVR applications, we show
that our method achieves outstanding Return on Investment (ROI) versus human
tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or
$25\%$ Query per Second (QPS) increase by only sampling one hundred models on
average from a curated search space. The proposed AutoML method has already
made real-world impact where a discovered Instagram CTR model with up to -0.36%
NE gain (over existing production baseline) was selected for large-scale online
A/B test and show statistically significant gain. These production results
proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Information Retrieval</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAURr0woMIDMOGNKcHOjXNtiaWK-JMSIUH3yLHvrSLapnICIm-PWpiO9A1H3xHiRqu8tMaoe59-hu-8AK1zhRbVpXhYf81j20gek2x82pKs_dGHYV5kO0baDYetHFm2NPu7Sb77w-dp-VimmfbTlbhgv5vo-p8rsXl63NQvWfP2_Fqvm8xXqDJGpwkK7iGCqkzAgBh6dkVZONvHQAwcnTbGKSIFmlEbtmgjIkEFJazE7Z_2fL87pmHv09KdMrpzBvwCIaVBKQ</recordid><startdate>20231113</startdate><enddate>20231113</enddate><creator>Yin, Hang</creator><creator>Liu, Kuang-Hung</creator><creator>Sun, Mengying</creator><creator>Chen, Yuxin</creator><creator>Zhang, Buyun</creator><creator>Liu, Jiang</creator><creator>Sehgal, Vivek</creator><creator>Panchal, Rudresh Rajnikant</creator><creator>Hotaj, Eugen</creator><creator>Liu, Xi</creator><creator>Guo, Daifeng</creator><creator>Zhang, Jamey</creator><creator>Wang, Zhou</creator><creator>Jiang, Shali</creator><creator>Li, Huayu</creator><creator>Chen, Zhengxing</creator><creator>Chen, Wen-Yen</creator><creator>Yang, Jiyan</creator><creator>Wen, Wei</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231113</creationdate><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><author>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-f791e32fb3d3065c7c77cbf924298bdcef3fd915590ee031f715f878d77e36343</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Information Retrieval</topic><toplevel>online_resources</toplevel><creatorcontrib>Yin, Hang</creatorcontrib><creatorcontrib>Liu, Kuang-Hung</creatorcontrib><creatorcontrib>Sun, Mengying</creatorcontrib><creatorcontrib>Chen, Yuxin</creatorcontrib><creatorcontrib>Zhang, Buyun</creatorcontrib><creatorcontrib>Liu, Jiang</creatorcontrib><creatorcontrib>Sehgal, Vivek</creatorcontrib><creatorcontrib>Panchal, Rudresh Rajnikant</creatorcontrib><creatorcontrib>Hotaj, Eugen</creatorcontrib><creatorcontrib>Liu, Xi</creatorcontrib><creatorcontrib>Guo, Daifeng</creatorcontrib><creatorcontrib>Zhang, Jamey</creatorcontrib><creatorcontrib>Wang, Zhou</creatorcontrib><creatorcontrib>Jiang, Shali</creatorcontrib><creatorcontrib>Li, Huayu</creatorcontrib><creatorcontrib>Chen, Zhengxing</creatorcontrib><creatorcontrib>Chen, Wen-Yen</creatorcontrib><creatorcontrib>Yang, Jiyan</creatorcontrib><creatorcontrib>Wen, Wei</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Yin, Hang</au><au>Liu, Kuang-Hung</au><au>Sun, Mengying</au><au>Chen, Yuxin</au><au>Zhang, Buyun</au><au>Liu, Jiang</au><au>Sehgal, Vivek</au><au>Panchal, Rudresh Rajnikant</au><au>Hotaj, Eugen</au><au>Liu, Xi</au><au>Guo, Daifeng</au><au>Zhang, Jamey</au><au>Wang, Zhou</au><au>Jiang, Shali</au><au>Li, Huayu</au><au>Chen, Zhengxing</au><au>Chen, Wen-Yen</au><au>Yang, Jiyan</au><au>Wen, Wei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>AutoML for Large Capacity Modeling of Meta's Ranking Systems</atitle><date>2023-11-13</date><risdate>2023</risdate><abstract>Web-scale ranking systems at Meta serving billions of users is complex.
Improving ranking models is essential but engineering heavy. Automated Machine
Learning (AutoML) can release engineers from labor intensive work of tuning
ranking models; however, it is unknown if AutoML is efficient enough to meet
tight production timeline in real-world and, at the same time, bring additional
improvements to the strong baselines. Moreover, to achieve higher ranking
performance, there is an ever-increasing demand to scale up ranking models to
even larger capacity, which imposes more challenges on the efficiency. The
large scale of models and tight production schedule requires AutoML to
outperform human baselines by only using a small number of model evaluation
trials (around 100). We presents a sampling-based AutoML method, focusing on
neural architecture search and hyperparameter optimization, addressing these
challenges in Meta-scale production when building large capacity models. Our
approach efficiently handles large-scale data demands. It leverages a
lightweight predictor-based searcher and reinforcement learning to explore vast
search spaces, significantly reducing the number of model evaluations. Through
experiments in large capacity modeling for CTR and CVR applications, we show
that our method achieves outstanding Return on Investment (ROI) versus human
tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or
$25\%$ Query per Second (QPS) increase by only sampling one hundred models on
average from a curated search space. The proposed AutoML method has already
made real-world impact where a discovered Instagram CTR model with up to -0.36%
NE gain (over existing production baseline) was selected for large-scale online
A/B test and show statistically significant gain. These production results
proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</abstract><doi>10.48550/arxiv.2311.07870</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2311.07870 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2311_07870 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Information Retrieval |
title | AutoML for Large Capacity Modeling of Meta's Ranking Systems |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T01%3A48%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=AutoML%20for%20Large%20Capacity%20Modeling%20of%20Meta's%20Ranking%20Systems&rft.au=Yin,%20Hang&rft.date=2023-11-13&rft_id=info:doi/10.48550/arxiv.2311.07870&rft_dat=%3Carxiv_GOX%3E2311_07870%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |