{"id":74676,"date":"2023-02-27T13:28:52","date_gmt":"2023-02-27T11:28:52","guid":{"rendered":"https:\/\/forklog.com\/en\/?p=74676"},"modified":"2025-09-09T21:36:58","modified_gmt":"2025-09-09T18:36:58","slug":"meta-unveils-the-llama-language-model","status":"publish","type":"post","link":"https:\/\/u1f987.com\/en\/meta-unveils-the-llama-language-model\/","title":{"rendered":"Meta unveils the LLaMA language model"},"content":{"rendered":"<p>Meta has released the large language model LLaMA for AI researchers with 13 billion and 65 billion parameters.<\/p>\n<blockquote class=\"twitter-tweet\">\n<p lang=\"en\" dir=\"ltr\">Today we&#8217;re publicly releasing LLaMA, a state-of-the-art foundational LLM, as part of our ongoing commitment to open science, transparency and democratized access to new research.<\/p>\n<p>Learn more &#038; request access \u27a1\ufe0f <a href=\"https:\/\/t.co\/8AeLVhMWkq\">https:\/\/t.co\/8AeLVhMWkq<\/a> <a href=\"https:\/\/t.co\/1BEkTngtnM\">pic.twitter.com\/1BEkTngtnM<\/a><\/p>\n<p>\u2014 Meta AI (@MetaAI) <a href=\"https:\/\/twitter.com\/MetaAI\/status\/1629156720483405824?ref_src=twsrc%5Etfw\">February 24, 2023<\/a><\/p><\/blockquote>\n<p> <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n<p>According to the developers, the smaller LLaMA-13B version demonstrated better results in most tests than OpenAI&#8217;s GPT-3. The larger LLaMA-65B system is &#8220;competitive with advanced models,&#8221; such as DeepMind&#8217;s Chinchilla-70B and Google&#8217;s PaLM-540B.<\/p>\n<p>Numbers in the model names refer to billions of parameters in each. The criterion is commonly used to gauge model size, yet these two attributes do not necessarily scale in lockstep.<\/p>\n<p>After training, LLaMA-13B can be run on a single Nvidia Tesla V100 GPU. According to the developers, this &#8220;democratizes&#8221; compute for small institutions that lack powerful hardware.<\/p>\n<p>Meta believes LLaMA will help AI researchers identify issues in language models related to bias, toxicity and tendency to hallucinate. To this end they released the algorithm under a non-commercial license.<\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;We believe that the entire community [\u2026] should work together to establish clear guidelines for responsible AI in general and responsible large language models in particular,&#8221; they said.<\/p>\n<\/blockquote>\n<p>According to Meta&#8217;s CEO Mark Zuckerberg, language models have shown promising capabilities in text generation, conversation and predicting protein structure.<\/p>\n<p><iframe loading=\"lazy\" src=\"https:\/\/www.facebook.com\/plugins\/post.php?href=https%3A%2F%2Fwww.facebook.com%2Fzuck%2Fposts%2Fpfbid0tNuVb1WiJW9rMBGgZiLq999ecmqwGgRuAJ6oKAVuyc3WKxfyGCwvP2vtwmSVk1Ksl&#038;show_text=true&#038;width=500\" width=\"500\" height=\"265\" style=\"border:none;overflow:hidden\" scrolling=\"no\" frameborder=\"0\" allowfullscreen=\"true\" allow=\"autoplay; clipboard-write; encrypted-media; picture-in-picture; web-share\"><\/iframe><\/p>\n<blockquote class=\"wp-block-quote is-layout-flow wp-block-quote-is-layout-flow\">\n<p>&#8220;Meta is committed to this open model of research, and we will make our new model accessible to the AI research community,&#8221; he added.<\/p>\n<\/blockquote>\n<p>Previously the tech giant released its own language models, but they were often criticised. In August 2022, Meta <a href=\"https:\/\/www.engadget.com\/meta-unleashes-blender-bot-3-upon-the-internet-its-most-competent-chat-ai-to-date-150021062.html\" target=\"_blank\" rel=\"noopener nofollow\" title=\"\">launched<\/a> public Blenderbot 3 with 175 billion parameters. The system was later accused of antisemitism and dissatisfaction with Facebook.<\/p>\n<p><script async src=\"https:\/\/telegram.org\/js\/telegram-widget.js?21\" data-telegram-post=\"forklogAI\/2609\" data-width=\"100%\"><\/script><\/p>\n<p>Another chatbot named Galactica <a href=\"https:\/\/u1f987.com\/en\/news\/meta-shuts-down-ai-for-science-three-days-after-launch\">shut down<\/a> just three days after launch. The system intended to summarize scientific papers was accused of creating fake information and misinformation.<\/p>\n<p>Earlier in November 2022, Meta unveiled the AI agent Cicero, which <a href=\"https:\/\/u1f987.com\/en\/news\/metas-ai-outperforms-humans-in-the-classic-board-game-diplomacy\">plays the board game Diplomacy<\/a> at a human level.<\/p>\n<p>In the same month, the tech giant&#8217;s AI lab <a href=\"https:\/\/u1f987.com\/en\/news\/meta-ai-builds-a-competitor-to-deepminds-alphafold\">discussed the transformer neural network<\/a> ESM-2 with 15 billion parameters for predicting protein structure.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Meta has released the large language model LLaMA for AI researchers with 13 billion and 65 billion parameters.<\/p>\n","protected":false},"author":1,"featured_media":74677,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"select":"1","news_style_id":"1","cryptorium_level":"","_short_excerpt_text":"","creation_source":"","_metatest_mainpost_news_update":false,"footnotes":""},"categories":[3],"tags":[438,1293],"class_list":["post-74676","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news-and-analysis","tag-artificial-intelligence","tag-meta"],"aioseo_notices":[],"amp_enabled":true,"views":"13","promo_type":"1","layout_type":"1","short_excerpt":"","is_update":"","_links":{"self":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts\/74676","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/comments?post=74676"}],"version-history":[{"count":1,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts\/74676\/revisions"}],"predecessor-version":[{"id":74678,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/posts\/74676\/revisions\/74678"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/media\/74677"}],"wp:attachment":[{"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/media?parent=74676"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/categories?post=74676"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/u1f987.com\/en\/wp-json\/wp\/v2\/tags?post=74676"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}