{"id":6501,"date":"2019-10-07T12:46:59","date_gmt":"2019-10-07T12:46:59","guid":{"rendered":"https:\/\/gtechbooster.com\/?p=6501"},"modified":"2023-04-01T01:36:48","modified_gmt":"2023-04-01T01:36:48","slug":"facebook-open-sources-roberta-nlp-model","status":"publish","type":"post","link":"https:\/\/gtechbooster.com\/facebook-open-sources-roberta-nlp-model\/","title":{"rendered":"Facebook Open Sources RoBERTA NLP Model"},"content":{"rendered":"\n<p>Facebook has made a new natural language processing model called \nRoBERTA available as open source. The model is an optimized version of \nGoogle&#8217;s BERT model.<\/p>\n\n\n\n<div class=\"gtech-migrated-from-ad-inserter-placement-2\" style=\"text-align: center;\" id=\"gtech-1358714057\"><div style=\"margin-right: auto;margin-left: auto;text-align: center;\" id=\"gtech-465231321\"><a data-bid=\"1\" data-no-instant=\"1\" href=\"https:\/\/gtechbooster.com\/linkout\/17207\" rel=\"noopener\" class=\"notrack\" aria-label=\"26001\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/gtechbooster.com\/media\/2023\/01\/26001.jpeg\" alt=\"\"  srcset=\"https:\/\/gtechbooster.com\/media\/2023\/01\/26001.jpeg 1024w, https:\/\/gtechbooster.com\/media\/2023\/01\/26001-768x960.jpeg 768w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" width=\"500\" height=\"625\"  style=\"display: inline-block;\" \/><\/a><\/div><\/div><p>The Facebook researchers describe their model as a robustly optimized  method for pretraining natural language processing (NLP) systems that  improves on Bidirectional Encoder Representations from Transformers, or  BERT, the self-supervised method released by Google in 2018.<\/p>\n\n\n\n<p>BERT has become know for the impressive results the technique has \nachieved on a range of NLP tasks while relying on un-annotated text \ndrawn from the web. Most similar NLP systems are based on text that has \nbeen labeled specifically for a given task.<\/p>\n\n\n\n<p>Facebook&#8217;s new optimized method, RoBERTa, produces state-of-the-art \nresults on the widely used NLP benchmark, General Language Understanding\n Evaluation (GLUE).<\/p>\n\n\n\n<p>RoBERTa has been implemented in PyTorch, and the team modified key \nhyperparameters in BERT, including removing BERT\u2019s next-sentence \npretraining objective. RoBERTa was also trained with much larger \nmini-batches and learning rates. The developers say this allows RoBERTa \nto improve on the masked language modeling objective compared with BERT \nand leads to better downstream task performance.<\/p>\n\n\n\n<p>After implementing these design changes, the Facebook model showed \nnotably better performance on the MNLI, QNLI, RTE, STS-B, and RACE tasks\n and a sizable performance improvement on the GLUE benchmark. With a \nscore of 88.5, RoBERTa reached the top position on the GLUE leaderboard,\n matching the performance of the previous leader, XLNet-Large. The team \nsays these results highlight the importance of previously unexplored \ndesign choices in BERT training and help disentangle the relative \ncontributions of data size, training time, and pretraining objectives.<\/p>\n\n\n\n<p>There&#8217;s a full description of RoBERTA and the research carried out in a paper published on arXiv.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">More Information<\/h2>\n\n\n\n<ul class=\"wp-block-list\"><li><a href=\"https:\/\/github.com\/pytorch\/fairseq\/tree\/master\/examples\/roberta\">RoBERTa On GitHub<\/a><\/li><li><a href=\"https:\/\/arxiv.org\/abs\/1907.11692\">RoBERTa&#8217;s technical details<\/a><\/li><\/ul>\n<div class=\"gtech-end-cont\" id=\"gtech-1510192723\"><div style=\"margin-right: auto;margin-left: auto;text-align: center;\" id=\"gtech-2870699405\"><a data-bid=\"1\" data-no-instant=\"1\" href=\"https:\/\/gtechbooster.com\/linkout\/76065\" rel=\"noopener\" class=\"notrack\" aria-label=\"26002\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/gtechbooster.com\/media\/2025\/10\/26002.jpg\" alt=\"\"  srcset=\"https:\/\/gtechbooster.com\/media\/2025\/10\/26002.jpg 1200w, https:\/\/gtechbooster.com\/media\/2025\/10\/26002-768x768.jpg 768w\" sizes=\"(max-width: 1200px) 100vw, 1200px\" width=\"500\" height=\"500\"  style=\"display: inline-block;\" \/><\/a><\/div><\/div>","protected":false},"excerpt":{"rendered":"<p>Facebook has made a new natural language processing model called RoBERTA available as open source. The model is an optimized version of Google&#8217;s BERT model.<\/p>\n","protected":false},"author":7,"featured_media":6502,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1915],"tags":[304,1043,6],"class_list":["post-6501","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-ndocs","tag-facebook","tag-natural-language-processing","tag-programming"],"blocksy_meta":{"styles_descriptor":{"styles":{"desktop":"","tablet":"","mobile":""},"google_fonts":[],"version":6}},"_links":{"self":[{"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/posts\/6501","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/users\/7"}],"replies":[{"embeddable":true,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/comments?post=6501"}],"version-history":[{"count":0,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/posts\/6501\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/media\/6502"}],"wp:attachment":[{"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/media?parent=6501"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/categories?post=6501"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/gtechbooster.com\/api-json\/wp\/v2\/tags?post=6501"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}