![]() ![]() Get started chatting with StableLM-Tuned-Alpha by using the following code snippet: Check out this notebook to run inference with limited GPU capabilities. SizeĪll StableLM models are hosted on the Hugging Face hub. We will be releasing these models as StableLM-Tuned-Alpha. The context length for these models is 4096 tokens.Īn upcoming technical report will document the model specifications and the training settings.Īs a proof-of-concept, we also fine-tuned the model with Stanford Alpaca's procedure using a combination of five recent datasets for conversational agents: Stanford's Alpaca, Nomic-AI's gpt4all, RyokoAI's ShareGPT52K datasets, Databricks labs' Dolly, and Anthropic's HH. ![]() These models will be trained on up to 1.5 trillion tokens. StableLM-Alpha models are trained on the new dataset that build on The Pile, which contains 1.5 trillion tokens, roughly 3x the size of The Pile. Please visit HuggingFace checkpoint for more information about how to combine our delta weights with the original model. StableVicuna's delta weights are released under ( CC BY-NC-SA-4.0). Phung leading the training effort.ĭue to the original non-commercial license of LLaMA, we can only release the weights of our model as deltas over the original model's weights. This model is developed by StabilityAI's CarperAI team, with Duy V. It is our attempt at creating an open-source RLHF LLM Chatbot. StableVicuna is an RLHF fine-tune of Vicuna-13B v0, which itself is a fine-tune of LLaMA-13B. Try to chat with our 7B model, StableLM-Tuned-Alpha-7B, on Hugging Face Spaces. ![]() Base models are released under CC BY-SA-4.0. Released initial set of StableLM-alpha models, with 3B and 7B parameters. Delta weights over the original Llama model is released under ( CC BY-NC-SA-4.0).
0 Comments
Leave a Reply. |