Software

AI + ML

Grok-1 chatbot model released – open source or open Pandora's box?

What hasn't been released by Elon Musk's xAI is also of note


As promised, Elon Musk has released the model behind the xAI chatbot, Grok-1.

Released under the Apache 2.0 license, the base model weights and network architecture are now available. The model has 314 billion parameters and needs hardware with enough GPU memory to run it. It is fine-tuned for applications such as natural language dialog, and represents the raw base model checkpoint from the pre-training phase, which concluded in October 2023.

Critics have pointed to relatively lackluster performance in benchmarks; while Grok is a big model, it does poorly from what some engineers have seen. "Considering how poor it is compared to other models, it really emphasises how important fine tuning is. Models with MUCH smaller parameter counts are outperforming it in many metrics," said one poster on the Hacker News forums last night.

You might find that the smaller Mistral performs just as well as Grok-1, for instance.

To put its size in perspective, even at 314 billion parameters, it still has some catching up to do with OpenAI's GPT-4, which had 1.76 trillion params at last count.

Notably, unlike existing LLMs, which are trained on data with a cutoff point in time, Grok has access to real time data corpus of everyone's Xeets via X.com, which should make for some interesting experiments in the days to come, although as another commenter noted: "Twitter tweet data in itself is both highly idiosyncratic and short by design, which alone is not conducive towards training a LLM."

Grok will be familiar to users of Musk's social media platform, X, and subscribers have been able to ask the chatbot questions and receive answers. According to xAI, Grok was modeled after The Hitchhiker's Guide to the Galaxy. "It is intended to answer almost anything and, far harder, even suggest what questions to ask."

If a user flicks through a dog-eared copy of The Hitchhiker's Guide to the Galaxy radio scripts, the following definition can be found lurking in Fit the Tenth: "The Hitchhiker's Guide to the Galaxy is an indispensable companion to all those who are keen to make sense of life in an infinitely complex and confusing universe, for though it cannot hope to be useful or informative on all matters, it does make the reassuring claim that where it is inaccurate, it is at least definitively inaccurate.

"In case of major discrepancy it is always reality that's got it wrong."

The release comes on the first anniversary of the launch of OpenAI's GPT-4 model, and Musk's legal spat with his former AI pals remains in the background. At the beginning of this month, Musk sued OpenAI, alleging there was little open about the company, despite its name. OpenAI responded by releasing a trove of emails, claiming Musk was fully aware of its plans and wanted it folded into Tesla.

Patrik Backman, general partner at OpenOcean, said of Grok-1 being released: "For once, Elon Musk is putting his principles into action. If you sue OpenAI for transforming into a profit-driven organization, you must be prepared to adhere to the same ideals."

What hasn't been released by xAI is also of note. The Grok-1 weights are out there, yet the data used for training is not available under the same license, prompting AI expert Gary Marcus to quip: "PartlyOpenAI."

Open sourcing generative AI tools has proven controversial. Some developers worry that making the technology available risks abuse and others point to the inherent benefits of transparency.

Meta sharedsort of – its Llama 2 models last year, and other companies have followed suit. OpenAI, on the other hand, has most definitely not.

By opening up the weights behind Grok-1, Musk is attempting to plant a flag in the opposite camp to the proprietary world of OpenAI.

As for its ultimate performance, like everything Musk touches, it could go either way. ®

Send us news
16 Comments

OpenAI slams 'sham' takeover bid by wannabe 'AGI dictator' Musk in countersuit

Billionaire 'tried every tool to harm us', says super lab, and it wants judge to end 'harassment'

Microsoft’s AI masterplan: Let OpenAI burn cash, then build on their successes

Redmond’s not alone: AWS, Alibaba, DeepSeek also rely on others blazing the trail

It's fun making Studio Ghibli-style images with ChatGPT – but intellectual property is no laughing matter

Miyazaki, copyright protection and the 'insult to life itself' of AI images

Dems fret over DOGE feeding sensitive data into random AI

Using LLMs to pick programs, people, contracts to cut is bad enough – but doing it with Musk's Grok? Yikes

Ireland opens probe into Musk’s X over Grok’s AI data slurp

Watchdog wants to know whether EU posts were used without consent under GDPR

OpenAI wants to bend copyright rules. Study suggests it isn’t waiting for permission

GPT-4o likely trained on O’Reilly books without permission, figures appear to show

Billions pour into AI as emissions rise, returns stay pitiful, say Stanford boffins

Models get bulkier, burnier, bank-breakier

Meta to feed Europe's public posts into AI brains again

Who said this opt-out approach is OK with GDPR, was it Llama 4, hmm?

AI entrepreneur sent avatar to argue in court – and the judge shut it down fast

We hear from court-scolded Jerome Dewald, who insists lawyer-bots have a future

Dot com era crash on the cards for AI datacenter spending? It's a 'risk'

Analysts say the bubble won't burst, but it is possible, admits world's largest colo provider

Global datacenter electricity use to double by 2030, say policy wonks. Yup, it's AI

No worries, just use neural networks to optimize systems powering neural networks

Apps-from-prompts Firebase Studio is a great example – of why AI can't replace devs

Big G reckons this agentic IDE speeds up or simplifies coding. Developers who've used it aren't so sure