

Stability AI announces new open-source large language model
source link: https://www.theverge.com/2023/4/19/23689883/stability-ai-open-source-large-language-model-stablelm
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

Stability AI announces new open-source large language model
/Stability AI, the same company behind the AI image generator Stable Diffusion, is now open-sourcing its language model, StableLM.
:format(webp)/cdn.vox-cdn.com/uploads/chorus_asset/file/20790695/acastro_200730_1777_ai_0001.jpg)
Stability AI, the company behind the AI-powered Stable Diffusion image generator, has released a suite of open-source large language models (LLMs) collectively called StableLM. In a post shared on Wednesday, the company announced that its models are now available for developers to use and adapt on GitHub.
Like its ChatGPT rival, StableLM is designed to efficiently generate text and code. It’s trained on a larger version of the open-source dataset known as the Pile, which encompasses information from a range of sources, including Wikipedia, Stack Exchange, and PubMed. Stability AI says StableLM models are currently available between 3 billion and 7 billion parameters, with 15 to 65 billion parameter models arriving later.
While StableLM expands on the open-source language models that Stability AI has already worked on in collaboration with the nonprofit EleutherAI, it also builds on its mission to make AI tools more accessible, as it has done with Stable Diffusion. The company made its text-to-image AI available in several ways, including a public demo, a software beta, and a full download of the model, allowing developers to toy with the tool and come up with various integrations.
We might even see the same happen with StableLM, along with Meta’s open-source LLaMa language model that leaked online last month. As pointed out by my colleague James Vincent, the release of Stable Diffusion has led “to both more good stuff and more bad stuff happening,” and “we’ll likely see a similar dynamic play out once more with AI text generation: more stuff, more of the time.”
You can try out a demo of StableLM’s fine-tuned chat model hosted on Hugging Face, which gave me a very complex and somewhat nonsensical recipe when I tried asking it how to make a peanut butter sandwich. It also suggested that I add a “funny drawing” to a sympathy card. Stability AI warns that while the datasets it uses should help “steer the base language models into ‘safer’ distributions of text, not all biases and toxicity can be mitigated through fine-tuning.”
Recommend
-
5
Why Meta’s large language model does not work for researchers
-
10
The biggest bottleneck for large language model startups is UX
-
8
Guest What happens to a large language model (LLM) after it’s trained
-
11
Mark Zuckerberg Announces Meta's New Large-Language Model AI 'LLaMA' ...
-
7
Mark Zuckerberg announces Meta LLaMA large language modelKey PointsIn this article
-
9
Stable Diffusion Now Has Its Own Open Source AI Language Model
-
2
Stability AI unveils its first LLM, as open-source AI race continues
-
10
open source parrots — Stability AI launches StableLM, an open source ChatGPT alternative StableLM's 3B and 7B models are available now on GitHub under CC 4.0 license....
-
3
Introduction Ever since the launch of GPT (Generative Pre Trained) by Open AI, the world has been taken by storm by Generative AI. From that period on, many Generative Models have come into the picture. With each relea...
-
8
Llama 2 is the second version of the open source language model from Meta. It is based on a transformer architecture and has now also been released for commercial use. This article will discuss info about its benchmarks, parameters, and trai...
About Joyk
Aggregate valuable and interesting links.
Joyk means Joy of geeK