AutoML for Large Capacity Modeling of Meta's Ranking Systems

Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to m...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Yin, Hang, Liu, Kuang-Hung, Sun, Mengying, Chen, Yuxin, Zhang, Buyun, Liu, Jiang, Sehgal, Vivek, Panchal, Rudresh Rajnikant, Hotaj, Eugen, Liu, Xi, Guo, Daifeng, Zhang, Jamey, Wang, Zhou, Jiang, Shali, Li, Huayu, Chen, Zhengxing, Chen, Wen-Yen, Yang, Jiyan, Wen, Wei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Yin, Hang
Liu, Kuang-Hung
Sun, Mengying
Chen, Yuxin
Zhang, Buyun
Liu, Jiang
Sehgal, Vivek
Panchal, Rudresh Rajnikant
Hotaj, Eugen
Liu, Xi
Guo, Daifeng
Zhang, Jamey
Wang, Zhou
Jiang, Shali
Li, Huayu
Chen, Zhengxing
Chen, Wen-Yen
Yang, Jiyan
Wen, Wei
description Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to meet tight production timeline in real-world and, at the same time, bring additional improvements to the strong baselines. Moreover, to achieve higher ranking performance, there is an ever-increasing demand to scale up ranking models to even larger capacity, which imposes more challenges on the efficiency. The large scale of models and tight production schedule requires AutoML to outperform human baselines by only using a small number of model evaluation trials (around 100). We presents a sampling-based AutoML method, focusing on neural architecture search and hyperparameter optimization, addressing these challenges in Meta-scale production when building large capacity models. Our approach efficiently handles large-scale data demands. It leverages a lightweight predictor-based searcher and reinforcement learning to explore vast search spaces, significantly reducing the number of model evaluations. Through experiments in large capacity modeling for CTR and CVR applications, we show that our method achieves outstanding Return on Investment (ROI) versus human tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or $25\%$ Query per Second (QPS) increase by only sampling one hundred models on average from a curated search space. The proposed AutoML method has already made real-world impact where a discovered Instagram CTR model with up to -0.36% NE gain (over existing production baseline) was selected for large-scale online A/B test and show statistically significant gain. These production results proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.
doi_str_mv 10.48550/arxiv.2311.07870
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_07870</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_07870</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-f791e32fb3d3065c7c77cbf924298bdcef3fd915590ee031f715f878d77e36343</originalsourceid><addsrcrecordid>eNotj71OwzAURr0woMIDMOGNKcHOjXNtiaWK-JMSIUH3yLHvrSLapnICIm-PWpiO9A1H3xHiRqu8tMaoe59-hu-8AK1zhRbVpXhYf81j20gek2x82pKs_dGHYV5kO0baDYetHFm2NPu7Sb77w-dp-VimmfbTlbhgv5vo-p8rsXl63NQvWfP2_Fqvm8xXqDJGpwkK7iGCqkzAgBh6dkVZONvHQAwcnTbGKSIFmlEbtmgjIkEFJazE7Z_2fL87pmHv09KdMrpzBvwCIaVBKQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><source>arXiv.org</source><creator>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</creator><creatorcontrib>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</creatorcontrib><description>Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to meet tight production timeline in real-world and, at the same time, bring additional improvements to the strong baselines. Moreover, to achieve higher ranking performance, there is an ever-increasing demand to scale up ranking models to even larger capacity, which imposes more challenges on the efficiency. The large scale of models and tight production schedule requires AutoML to outperform human baselines by only using a small number of model evaluation trials (around 100). We presents a sampling-based AutoML method, focusing on neural architecture search and hyperparameter optimization, addressing these challenges in Meta-scale production when building large capacity models. Our approach efficiently handles large-scale data demands. It leverages a lightweight predictor-based searcher and reinforcement learning to explore vast search spaces, significantly reducing the number of model evaluations. Through experiments in large capacity modeling for CTR and CVR applications, we show that our method achieves outstanding Return on Investment (ROI) versus human tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or $25\%$ Query per Second (QPS) increase by only sampling one hundred models on average from a curated search space. The proposed AutoML method has already made real-world impact where a discovered Instagram CTR model with up to -0.36% NE gain (over existing production baseline) was selected for large-scale online A/B test and show statistically significant gain. These production results proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</description><identifier>DOI: 10.48550/arxiv.2311.07870</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Information Retrieval</subject><creationdate>2023-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.07870$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.07870$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Yin, Hang</creatorcontrib><creatorcontrib>Liu, Kuang-Hung</creatorcontrib><creatorcontrib>Sun, Mengying</creatorcontrib><creatorcontrib>Chen, Yuxin</creatorcontrib><creatorcontrib>Zhang, Buyun</creatorcontrib><creatorcontrib>Liu, Jiang</creatorcontrib><creatorcontrib>Sehgal, Vivek</creatorcontrib><creatorcontrib>Panchal, Rudresh Rajnikant</creatorcontrib><creatorcontrib>Hotaj, Eugen</creatorcontrib><creatorcontrib>Liu, Xi</creatorcontrib><creatorcontrib>Guo, Daifeng</creatorcontrib><creatorcontrib>Zhang, Jamey</creatorcontrib><creatorcontrib>Wang, Zhou</creatorcontrib><creatorcontrib>Jiang, Shali</creatorcontrib><creatorcontrib>Li, Huayu</creatorcontrib><creatorcontrib>Chen, Zhengxing</creatorcontrib><creatorcontrib>Chen, Wen-Yen</creatorcontrib><creatorcontrib>Yang, Jiyan</creatorcontrib><creatorcontrib>Wen, Wei</creatorcontrib><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><description>Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to meet tight production timeline in real-world and, at the same time, bring additional improvements to the strong baselines. Moreover, to achieve higher ranking performance, there is an ever-increasing demand to scale up ranking models to even larger capacity, which imposes more challenges on the efficiency. The large scale of models and tight production schedule requires AutoML to outperform human baselines by only using a small number of model evaluation trials (around 100). We presents a sampling-based AutoML method, focusing on neural architecture search and hyperparameter optimization, addressing these challenges in Meta-scale production when building large capacity models. Our approach efficiently handles large-scale data demands. It leverages a lightweight predictor-based searcher and reinforcement learning to explore vast search spaces, significantly reducing the number of model evaluations. Through experiments in large capacity modeling for CTR and CVR applications, we show that our method achieves outstanding Return on Investment (ROI) versus human tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or $25\%$ Query per Second (QPS) increase by only sampling one hundred models on average from a curated search space. The proposed AutoML method has already made real-world impact where a discovered Instagram CTR model with up to -0.36% NE gain (over existing production baseline) was selected for large-scale online A/B test and show statistically significant gain. These production results proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Information Retrieval</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAURr0woMIDMOGNKcHOjXNtiaWK-JMSIUH3yLHvrSLapnICIm-PWpiO9A1H3xHiRqu8tMaoe59-hu-8AK1zhRbVpXhYf81j20gek2x82pKs_dGHYV5kO0baDYetHFm2NPu7Sb77w-dp-VimmfbTlbhgv5vo-p8rsXl63NQvWfP2_Fqvm8xXqDJGpwkK7iGCqkzAgBh6dkVZONvHQAwcnTbGKSIFmlEbtmgjIkEFJazE7Z_2fL87pmHv09KdMrpzBvwCIaVBKQ</recordid><startdate>20231113</startdate><enddate>20231113</enddate><creator>Yin, Hang</creator><creator>Liu, Kuang-Hung</creator><creator>Sun, Mengying</creator><creator>Chen, Yuxin</creator><creator>Zhang, Buyun</creator><creator>Liu, Jiang</creator><creator>Sehgal, Vivek</creator><creator>Panchal, Rudresh Rajnikant</creator><creator>Hotaj, Eugen</creator><creator>Liu, Xi</creator><creator>Guo, Daifeng</creator><creator>Zhang, Jamey</creator><creator>Wang, Zhou</creator><creator>Jiang, Shali</creator><creator>Li, Huayu</creator><creator>Chen, Zhengxing</creator><creator>Chen, Wen-Yen</creator><creator>Yang, Jiyan</creator><creator>Wen, Wei</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231113</creationdate><title>AutoML for Large Capacity Modeling of Meta's Ranking Systems</title><author>Yin, Hang ; Liu, Kuang-Hung ; Sun, Mengying ; Chen, Yuxin ; Zhang, Buyun ; Liu, Jiang ; Sehgal, Vivek ; Panchal, Rudresh Rajnikant ; Hotaj, Eugen ; Liu, Xi ; Guo, Daifeng ; Zhang, Jamey ; Wang, Zhou ; Jiang, Shali ; Li, Huayu ; Chen, Zhengxing ; Chen, Wen-Yen ; Yang, Jiyan ; Wen, Wei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-f791e32fb3d3065c7c77cbf924298bdcef3fd915590ee031f715f878d77e36343</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Information Retrieval</topic><toplevel>online_resources</toplevel><creatorcontrib>Yin, Hang</creatorcontrib><creatorcontrib>Liu, Kuang-Hung</creatorcontrib><creatorcontrib>Sun, Mengying</creatorcontrib><creatorcontrib>Chen, Yuxin</creatorcontrib><creatorcontrib>Zhang, Buyun</creatorcontrib><creatorcontrib>Liu, Jiang</creatorcontrib><creatorcontrib>Sehgal, Vivek</creatorcontrib><creatorcontrib>Panchal, Rudresh Rajnikant</creatorcontrib><creatorcontrib>Hotaj, Eugen</creatorcontrib><creatorcontrib>Liu, Xi</creatorcontrib><creatorcontrib>Guo, Daifeng</creatorcontrib><creatorcontrib>Zhang, Jamey</creatorcontrib><creatorcontrib>Wang, Zhou</creatorcontrib><creatorcontrib>Jiang, Shali</creatorcontrib><creatorcontrib>Li, Huayu</creatorcontrib><creatorcontrib>Chen, Zhengxing</creatorcontrib><creatorcontrib>Chen, Wen-Yen</creatorcontrib><creatorcontrib>Yang, Jiyan</creatorcontrib><creatorcontrib>Wen, Wei</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Yin, Hang</au><au>Liu, Kuang-Hung</au><au>Sun, Mengying</au><au>Chen, Yuxin</au><au>Zhang, Buyun</au><au>Liu, Jiang</au><au>Sehgal, Vivek</au><au>Panchal, Rudresh Rajnikant</au><au>Hotaj, Eugen</au><au>Liu, Xi</au><au>Guo, Daifeng</au><au>Zhang, Jamey</au><au>Wang, Zhou</au><au>Jiang, Shali</au><au>Li, Huayu</au><au>Chen, Zhengxing</au><au>Chen, Wen-Yen</au><au>Yang, Jiyan</au><au>Wen, Wei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>AutoML for Large Capacity Modeling of Meta's Ranking Systems</atitle><date>2023-11-13</date><risdate>2023</risdate><abstract>Web-scale ranking systems at Meta serving billions of users is complex. Improving ranking models is essential but engineering heavy. Automated Machine Learning (AutoML) can release engineers from labor intensive work of tuning ranking models; however, it is unknown if AutoML is efficient enough to meet tight production timeline in real-world and, at the same time, bring additional improvements to the strong baselines. Moreover, to achieve higher ranking performance, there is an ever-increasing demand to scale up ranking models to even larger capacity, which imposes more challenges on the efficiency. The large scale of models and tight production schedule requires AutoML to outperform human baselines by only using a small number of model evaluation trials (around 100). We presents a sampling-based AutoML method, focusing on neural architecture search and hyperparameter optimization, addressing these challenges in Meta-scale production when building large capacity models. Our approach efficiently handles large-scale data demands. It leverages a lightweight predictor-based searcher and reinforcement learning to explore vast search spaces, significantly reducing the number of model evaluations. Through experiments in large capacity modeling for CTR and CVR applications, we show that our method achieves outstanding Return on Investment (ROI) versus human tuned baselines, with up to 0.09% Normalized Entropy (NE) loss reduction or $25\%$ Query per Second (QPS) increase by only sampling one hundred models on average from a curated search space. The proposed AutoML method has already made real-world impact where a discovered Instagram CTR model with up to -0.36% NE gain (over existing production baseline) was selected for large-scale online A/B test and show statistically significant gain. These production results proved AutoML efficacy and accelerated its adoption in ranking systems at Meta.</abstract><doi>10.48550/arxiv.2311.07870</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2311.07870
ispartof
issn
language eng
recordid cdi_arxiv_primary_2311_07870
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Information Retrieval
title AutoML for Large Capacity Modeling of Meta's Ranking Systems
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T01%3A48%3A46IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=AutoML%20for%20Large%20Capacity%20Modeling%20of%20Meta's%20Ranking%20Systems&rft.au=Yin,%20Hang&rft.date=2023-11-13&rft_id=info:doi/10.48550/arxiv.2311.07870&rft_dat=%3Carxiv_GOX%3E2311_07870%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true