{"id":78506,"date":"2024-03-18T19:45:48","date_gmt":"2024-03-18T19:45:48","guid":{"rendered":"https:\/\/neclink.com\/index.php\/2024\/03\/18\/why-elon-musks-ai-company-open-sourcing-grok-matters-and-why-it-doesnt\/"},"modified":"2024-03-18T19:45:48","modified_gmt":"2024-03-18T19:45:48","slug":"why-elon-musks-ai-company-open-sourcing-grok-matters-and-why-it-doesnt","status":"publish","type":"post","link":"https:\/\/neclink.com\/index.php\/2024\/03\/18\/why-elon-musks-ai-company-open-sourcing-grok-matters-and-why-it-doesnt\/","title":{"rendered":"Why Elon Musk&#8217;s AI company &#8216;open-sourcing&#8217; Grok matters \u2014 and why it doesn&#8217;t"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p id=\"speakable-summary\">Elon Musk\u2019s xAI <a href=\"https:\/\/techcrunch.com\/2024\/03\/17\/xai-open-sources-base-model-of-grok-but-without-any-training-code\/\">released<\/a> its <a href=\"https:\/\/x.ai\/blog\/grok-os\">Grok<\/a> large language model as \u201copen source\u201d over the weekend. The billionaire clearly hopes to set his company at odds <a href=\"https:\/\/techcrunch.com\/2024\/03\/05\/openai-response-elon-musk-lawsuit\/\">with rival OpenAI<\/a>, which despite its name is not particularly open. But does releasing the code for something like Grok actually contribute to the AI development community? Yes and no.<\/p>\n<p><a href=\"https:\/\/techcrunch.com\/2024\/03\/13\/what-is-elon-musks-grok-chatbot-and-how-does-it-work\/\">Grok<\/a> is a chatbot trained by xAI to fill the same vaguely defined role as something like ChatGPT or Claude: you ask it, it answers. This LLM, however, was given a sassy tone and extra access to Twitter data as a way of differentiating it from the rest.<\/p>\n<p>As always, these systems are nearly impossible to evaluate, but the general consensus seems to be that it\u2019s competitive with last-generation medium-size models like GPT-3.5. (Whether you decide this is impressive given the short development time frame or disappointing given the budget and bombast surrounding xAI is entirely up to you.)<\/p>\n<p>At any rate, Grok is a modern and functional LLM of significant size and capability, and the more access the dev community has to the guts of such things, the better. The problem is in defining \u201copen\u201d in a way that does more than let a company (or billionaire) claim the moral high ground.<\/p>\n<p>This isn\u2019t the first time the terms \u201copen\u201d and \u201copen source\u201d have been questioned or abused in the AI world. And we aren\u2019t just talking about a technical quibble, such as picking a usage license that\u2019s not as open as another (Grok is Apache 2.0, if you\u2019re wondering).<\/p>\n<p>To begin with, AI models are unlike other software when it comes to making them \u201copen source.\u201d<\/p>\n<p>If you\u2019re making, say, a word processor, it\u2019s relatively simple to make it open source: you publish all your code publicly and let community to propose improvements or make their own version. Part of what makes open source as a concept valuable is that every aspect of the application is original or credited to its original creator \u2014 this transparency and adherence to correct attribution is not just a byproduct, but is core to the very concept of openness.<\/p>\n<p>With AI, this is arguably not possible at all, because the way machine learning models are created involves a largely unknowable process whereby a tremendous amount of training data is distilled into a complex statistical representation the structure of which no human really directed, or even understands. This process cannot be inspected, audited, and improved the way traditional code can \u2014 so while it still has immense value in one sense, it can\u2019t ever really be open. (The standards community hasn\u2019t even <a href=\"https:\/\/opensource.org\/blog\/open-source-ai-definition-where-it-stands-and-whats-ahead\">defined what open will be<\/a> in this context, but are <a href=\"https:\/\/blog.mozilla.org\/en\/mozilla\/ai\/introducing-columbia-convening-openness-and-ai\/\">actively discussing it<\/a>.)<\/p>\n<p>That hasn\u2019t stopped AI developers and companies from designing and claiming their models as \u201copen,\u201d a term that has lost much of its meaning in this context. Some call their model \u201copen\u201d if there is a public-facing interface or API. Some call it \u201copen\u201d if they release a paper describing the development process.<\/p>\n<p>Arguably the closest to \u201copen source\u201d an AI model can be is when its developers release its <em>weights<\/em>, which is to say the exact attributes of the countless nodes of its neural networks, which perform vector mathematics operations in precise order to complete the pattern started by a user\u2019s input. But even \u201copen-weights\u201d models like LLaMa-2 exclude other important data, like the training dataset and process \u2014 which would be necessary to recreate it from scratch. (<a href=\"https:\/\/techcrunch.com\/2024\/02\/01\/ai2-open-sources-text-generating-ai-models-and-the-data-used-to-train-them\/\">Some projects go further<\/a>, of course.)<\/p>\n<p>All this is before even mentioning the fact that it takes millions of dollars in computing and engineering resources to create or replicate these models, effectively restricting who can create and replicate them to companies with considerable resources.<\/p>\n<p>So where does xAI\u2019s Grok release fall on this spectrum?<\/p>\n<p>As an open-weights model, it\u2019s ready for anyone to <a href=\"https:\/\/github.com\/xai-org\/grok-1?tab=readme-ov-file\">download, use, modify, fine tine, or distill<\/a>. That\u2019s good! It appears to be among the largest models anyone can access freely this way, in terms of parameters \u2014 314 billion \u2014 which gives curious engineers a lot to work with if they want to test how it performs after various modifications.<\/p>\n<p>The size of the model comes with serious drawbacks, though: you\u2019ll need hundreds of gigabytes of high-speed RAM to use it in this raw form. If you\u2019re not already in possession of, say, a dozen Nvidia H100s in a six-figure AI inference rig, don\u2019t bother clicking that download link.<\/p>\n<p>And although Grok is arguably competitive with some other modern models, it\u2019s also far, far larger than them, meaning it requires more resources to accomplish the same thing. There\u2019s always a hierarchy of size, efficiency, and other metrics, and it\u2019s still valuable, but this is more raw material than final product. It\u2019s also not clear whether this is the latest and best version of Grok, like the clearly tuned version some have access to via X.<\/p>\n<p>Overall, it\u2019s a good thing to release this data, but it\u2019s not a game-changer the way some hoped it might be.<\/p>\n<p>It\u2019s also hard not to wonder why Musk is doing this. Is his nascent AI company really dedicated to open source development? Or is this just mud in the eye of OpenAI, with which Musk is currently <a href=\"https:\/\/techcrunch.com\/2024\/03\/01\/elon-musk-openai-sam-altman-court\/\">pursuing a billionaire-level beef<\/a>?<\/p>\n<p>If they are really dedicated to open source development, this will be the first of many releases, and they will hopefully take the feedback of the community into account, release other crucial information, characterize the training data process, and further explain their approach. If they aren\u2019t, and this is only done so Musk can point to it in online arguments, it\u2019s still valuable \u2014 just not something anyone in the AI world will rely on or pay much attention to after the next few months as they play with the model.<\/p>\n<\/p><\/div>\n<p><br \/>\n<br \/><a href=\"https:\/\/techcrunch.com\/2024\/03\/18\/why-elon-musks-ai-company-open-sourcing-grok-matters-and-why-it-doesnt\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Elon Musk\u2019s xAI released its Grok large language model as \u201copen source\u201d over the weekend. The billionaire clearly hopes to set his company at odds<\/p>\n","protected":false},"author":1,"featured_media":78507,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_monsterinsights_skip_tracking":false,"_monsterinsights_sitenote_active":false,"_monsterinsights_sitenote_note":"","_monsterinsights_sitenote_category":0,"footnotes":""},"categories":[178],"tags":[],"class_list":["post-78506","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-tech"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts\/78506","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/comments?post=78506"}],"version-history":[{"count":0,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/posts\/78506\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/media\/78507"}],"wp:attachment":[{"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/media?parent=78506"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/categories?post=78506"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/neclink.com\/index.php\/wp-json\/wp\/v2\/tags?post=78506"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}