• 3 Posts
  • 0 Comments
Joined 1Y ago
cake
Cake day: Jun 10, 2023

help-circle
rss
Llama-2 FOSAI & LLM Roundup Series! (Summer 2023 Edition)
cross-posted from: https://lemmy.world/post/1894070 > ## **Welcome to the Llama-2 FOSAI & LLM Roundup Series!** > > **(Summer 2023 Edition)** > > Hello everyone! > > The wave of innovation I mentioned in our [Llama-2 announcement](https://lemmy.world/post/1750098) is already on its way. The first tsunami of base models and configurations are being released as you read this post. > > That being said, I'd like to take a moment to shoutout [TheBloke](https://huggingface.co/TheBloke), who is rapidly converting many of these models for the greater good of FOSS & FOSAI. > > You can support [TheBloke](https://huggingface.co/TheBloke) here. > - https://ko-fi.com/TheBlokeAI > > Below you will find all of the latest Llama-2 models that are FOSAI friendly. This means they are commercially available, ready to use, and open for development. I will be continuing this series exclusively for Llama models. I have a feeling it will continue being a popular choice for quite some time. I will consider giving other foundational models a similar series if they garner enough support and consideration. For now, enjoy this new herd of Llamas! > > All that you need to get started is capable hardware and a few moments setting up your inference platform (selected from any of your preferred software choices in the [Lemmy Crash Course for Free Open-Source AI](https://lemmy.world/post/76020) > or [FOSAI Nexus](https://lemmy.world/post/814816) resource, which is also shared at the bottom of this post). > > Keep reading to learn more about the exciting new models coming out of Llama-2! > > ### **8-bit System Requirements** > > | Model | VRAM Used | Minimum Total VRAM | Card Examples | RAM/Swap to Load* | > |-----------|-----------|--------------------|-------------------|-------------------| > | LLaMA-7B | 9.2GB | 10GB | 3060 12GB, 3080 10GB | 24 GB | > | LLaMA-13B | 16.3GB | 20GB | 3090, 3090 Ti, 4090 | 32 GB | > | LLaMA-30B | 36GB | 40GB | A6000 48GB, A100 40GB | 64 GB | > | LLaMA-65B | 74GB | 80GB | A100 80GB | 128 GB | > > ### **4-bit System Requirements** > > | Model | Minimum Total VRAM | Card Examples | RAM/Swap to Load* | > |-----------|--------------------|--------------------------------|-------------------| > | LLaMA-7B | 6GB | GTX 1660, 2060, AMD 5700 XT, RTX 3050, 3060 | 6 GB | > | LLaMA-13B | 10GB | AMD 6900 XT, RTX 2060 12GB, 3060 12GB, 3080, A2000 | 12 GB | > | LLaMA-30B | 20GB | RTX 3080 20GB, A4500, A5000, 3090, 4090, 6000, Tesla V100 | 32 GB | > | LLaMA-65B | 40GB | A100 40GB, 2x3090, 2x4090, A40, RTX A6000, 8000 | 64 GB | > > *System RAM (not VRAM), is utilized to initially load a model. You can use swap space if you do not have enough RAM to support your LLM. > > --- > > ### **The Bloke** > One of the most popular and consistent developers releasing consumer-friendly versions of LLMs. These active conversions of trending models allow for many of us to run these GPTQ or GGML variants at home on our own PCs and hardware. > > **70B** > > - [TheBloke/Llama-2-70B-chat-GPTQ](https://huggingface.co/TheBloke/Llama-2-70B-chat-GPTQ) > > - [TheBloke/Llama-2-70B-Chat-fp16](https://huggingface.co/TheBloke/Llama-2-70B-Chat-fp16) > > - [TheBloke/Llama-2-70B-GPTQ](https://huggingface.co/TheBloke/Llama-2-70B-GPTQ) > > - [TheBloke/Llama-2-70B-fp16](https://huggingface.co/TheBloke/Llama-2-70B-fp16) > > **13B** > > - [TheBloke/Llama-2-13B-chat-GPTQ](https://huggingface.co/TheBloke/Llama-2-13B-chat-GPTQ) > > - [TheBloke/Llama-2-13B-chat-GGML](https://huggingface.co/TheBloke/Llama-2-13B-chat-GGML) > > - [TheBloke/Llama-2-13B-GPTQ](https://huggingface.co/TheBloke/Llama-2-13B-GPTQ) > > - [TheBloke/Llama-2-13B-GGML](https://huggingface.co/TheBloke/Llama-2-13B-GGML) > > - [TheBloke/Llama-2-13B-fp16](https://huggingface.co/TheBloke/Llama-2-13B-fp16) > > **7B** > > - [TheBloke/Llama-2-7B-GPTQ](https://huggingface.co/TheBloke/Llama-2-7B-GPTQ) > > - [TheBloke/Llama-2-7B-GGML)](https://huggingface.co/TheBloke/Llama-2-7B-GGML) > > - [TheBloke/Llama-2-7B-fp16](https://huggingface.co/TheBloke/Llama-2-7B-fp16) > > - [TheBloke/Llama-2-7B-fp16](https://huggingface.co/TheBloke/Llama-2-7B-fp16) > > - [TheBloke/Llama-2-7b-Chat-GPTQ](https://huggingface.co/TheBloke/Llama-2-7b-Chat-GPTQ) > > ### **LLongMA** > LLongMA-2, a suite of Llama-2 models, trained at 8k context length using linear positional interpolation scaling. > > **13B** > > - [conceptofmind/LLongMA-2-13b](https://huggingface.co/conceptofmind/LLongMA-2-13b) > > **7B** > > - [conceptofmind/LLongMA-2-7b](https://huggingface.co/conceptofmind/LLongMA-2-7b) > > Also available from The Bloke in GPTQ and GGML formats: > > **7B** > > - [TheBloke/LLongMA-2-7B-GPTQ](https://huggingface.co/TheBloke/LLongMA-2-7B-GPTQ) > > - [TheBloke/LLongMA-2-7B-GGML](https://huggingface.co/TheBloke/LLongMA-2-7B-GGML) > > ### **Puffin** > The first commercially available language model released by Nous Research! Available in 13B parameters. > > **13B** > > - [NousResearch/Redmond-Puffin-13B-GGML](https://huggingface.co/NousResearch/Redmond-Puffin-13B-GGML) > > - [NousResearch/Redmond-Puffin-13B](https://huggingface.co/NousResearch/Redmond-Puffin-13B) > > Also available from The Bloke in GPTQ and GGML formats: > > **13B** > > - [TheBloke/Redmond-Puffin-13B-GPTQ](https://huggingface.co/TheBloke/Redmond-Puffin-13B-GPTQ) > > - [TheBloke/Redmond-Puffin-13B-GGML](https://huggingface.co/TheBloke/Redmond-Puffin-13B-GGML) > > ### **Other Models** > Leaving a section here for 'other' LLMs or fine tunings derivative of Llama-2 models. > > **7B** > > - [georgesung/llama2_7b_chat_uncensored](https://huggingface.co/georgesung/llama2_7b_chat_uncensored) > > --- > > ### **Getting Started w/ FOSAI!** > > Have no idea where to begin with AI/LLMs? [Try starting here with ](https://understandgpt.ai/docs/getting-started/what-is-a-llm) [UnderstandGPT](https://understandgpt.ai/) to learn the basics of LLMs before visiting our [Lemmy Crash Course for Free Open-Source AI](https://lemmy.world/post/76020) > > If you're looking to explore more resources, see our [FOSAI Nexus](https://lemmy.world/post/814816) for a list of all the major FOSS/FOSAI in the space. > > If you're looking to jump right in, visit some of the links below and stick to models that are <13B in parameter (unless you have the power and hardware to spare). > > **FOSAI Resources** > > **Fediverse / FOSAI** > - [The Internet is Healing](https://www.youtube.com/watch?v=TrNE2fSCeFo) > - [FOSAI Welcome Message](https://lemmy.world/post/67758) > - [FOSAI Crash Course](https://lemmy.world/post/76020) > - [FOSAI Nexus Resource Hub](https://lemmy.world/post/814816) > > **LLM Leaderboards** > - [HF Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) > - [LMSYS Chatbot Arena](https://chat.lmsys.org/?leaderboard) > > **LLM Search Tools** > - [LLM Explorer](https://llm.extractum.io/) > - [Open LLMs](https://github.com/eugeneyan/open-llms) > > ### **GL, HF!** > > If you found anything about this post interesting - consider subscribing to !fosai@lemmy.world where I do my best to keep you in the know about the most important updates in free open-source artificial intelligence. > > I will try to continue doing this series season by season, making this a living post for the rest of this summer. If I have missed a noteworthy model, don't hesitate to let me know in the comments so I can keep this resource up-to-date. > > Thank you for reading! I hope you find what you're looking for. Be sure to subscribe and bookmark the [main post](https://lemmy.world/post/1894070) if you want a quick one-stop shop for all of the new Llama-2 models that will be emerging the rest of this summer!
fedilink

FOSAI Nexus (v0.0.1)!
# Welcome to the FOSAI Nexus! **(v0.0.1 - Summer 2023 Edition)** The goal of this knowledge nexus is to act as a link hub for software, applications, tools, and projects that are all FOSS (free open-source software) designed for AI (FOSAI). If you haven't already, I recommend bookmarking this page ([the native one on lemmy.world](https://lemmy.world/post/814816)). It is designed to be periodically updated in new versions I release throughout the year. This is due to the rapid rate in which this field is advancing. Breakthroughs are happening weekly. I will try to keep up through the seasons while including links to each sequential nexus post - but it's best to bookmark this since it will be the start of the content series, giving you access to all future nexus posts as I release them. If you see something here missing that should be added, let me know. I don't have visibility over everything. I would love your help making this nexus better. Like I said in my [welcome message](https://lemmy.world/post/67758), I am no expert in this field, but I teach myself what I can to distill it in ways I find interesting to share with others. I hope this helps you unblock your workflow or project and empowers you to explore the wonders of emerging artificial intelligence. Consider subscribing to [/c/FOSAI](https://lemmy.world/c/fosai) if you found any of this interesting. I do my best to make sure you stay in the know with the most important updates to all things free open-source AI. [Find Us On Lemmy!](https://lemmy.world/c/fosai) [!fosai@lemmy.world](https://lemmy.world/c/fosai) --- ## Fediverse Resources **Lemmy** - [Your Lemmy Crash Course to Free Open-Source AI](https://lemmy.world/post/76020) - [!auai@programming.dev](https://programming.dev/c/auai) --- ## Large Language Model Hub [Download Models](https://huggingface.co/TheBloke) ### [oobabooga](https://github.com/oobabooga/text-generation-webui) text-generation-webui - a big community favorite gradio web UI by oobabooga designed for running almost any free open-source and large language models downloaded off of [HuggingFace](https://huggingface.co/TheBloke) which can be (but not limited to) models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and many others. Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui) of text generation. It is highly compatible with many formats. ### [Exllama](https://github.com/turboderp/exllama) A standalone Python/C++/CUDA implementation of Llama for use with 4-bit GPTQ weights, designed to be fast and memory-efficient on modern GPUs. ### [gpt4all](https://github.com/nomic-ai/gpt4all) Open-source assistant-style large language models that run locally on your CPU. GPT4All is an ecosystem to train and deploy powerful and customized large language models that run locally on consumer-grade processors. ### [TavernAI](https://github.com/TavernAI/TavernAI) The original branch of software SillyTavern was forked from. This chat interface offers very similar functionalities but has less cross-client compatibilities with other chat and API interfaces (compared to SillyTavern). ### [SillyTavern](https://github.com/SillyTavern/SillyTavern) Developer-friendly, Multi-API (KoboldAI/CPP, Horde, NovelAI, Ooba, OpenAI+proxies, Poe, WindowAI(Claude!)), Horde SD, System TTS, WorldInfo (lorebooks), customizable UI, auto-translate, and more prompt options than you'd ever want or need. Optional Extras server for more SD/TTS options + ChromaDB/Summarize. Based on a fork of TavernAI 1.2.8 ### [Koboldcpp](https://github.com/LostRuins/koboldcpp) A self-contained distributable from Concedo that exposes llama.cpp function bindings, allowing it to be used via a simulated Kobold API endpoint. What does it mean? You get llama.cpp with a fancy UI, persistent stories, editing tools, save formats, memory, world info, author's note, characters, scenarios, and everything Kobold and Kobold Lite have to offer. In a tiny package around 20 MB in size, excluding model weights. ### [KoboldAI-Client](https://github.com/KoboldAI/KoboldAI-Client) This is a browser-based front-end for AI-assisted writing with multiple local & remote AI models. It offers the standard array of tools, including Memory, Author's Note, World Info, Save & Load, adjustable AI settings, formatting options, and the ability to import existing AI Dungeon adventures. You can also turn on Adventure mode and play the game like AI Dungeon Unleashed. ### [h2oGPT](https://github.com/h2oai/h2ogpt) h2oGPT is a large language model (LLM) fine-tuning framework and chatbot UI with document(s) question-answer capabilities. Documents help to ground LLMs against hallucinations by providing them context relevant to the instruction. h2oGPT is fully permissive Apache V2 open-source project for 100% private and secure use of LLMs and document embeddings for document question-answer. --- ## Image Diffusion Hub [Download Models](https://civitai.com/) ### [StableDiffusion](https://github.com/AUTOMATIC1111/stable-diffusion-webui) Stable Diffusion is a text-to-image diffusion model capable of generating photo-realistic and stylized images. This is the free alternative to MidJourney. It is rumored that MidJourney originates from a version of Stable Diffusion that is highly modified, tuned, then made proprietary. ### [SDXL (Stable Diffusion XL)](https://clipdrop.co/stable-diffusion) With [Stable Diffusion XL](https://stability.ai/stablediffusion), you can create descriptive images with shorter prompts and generate words within images. The model is a significant advancement in image generation capabilities, offering enhanced image composition and face generation that results in stunning visuals and realistic aesthetics. ### [ComfyUI](https://github.com/comfyanonymous/ComfyUI) A powerful and modular stable diffusion GUI and backend. This new and powerful UI will let you design and execute advanced stable diffusion pipelines using a graph/nodes/flowchart-based interface. ### [ControlNet](https://github.com/lllyasviel/ControlNet) ControlNet is a neural network structure to control diffusion models by adding extra conditions. This is a very popular and powerful extension to add to AUTOMATIC111's stable-diffusion-webui. ### [TemporalKit](https://github.com/CiaraStrawberry/TemporalKit) An all-in-one solution for adding Temporal Stability to a Stable Diffusion Render via an automatic1111 extension. You must install FFMPEG to path before running this. ### [EbSynth](https://ebsynth.com/) Bring your paintings to animated life. This software can be used in conjunction with StableDiffusion + ControlNet + TemporalKit workflows. ### [WarpFusion](https://github.com/Sxela/WarpFusion) A TemporalKit alternative to produce video effects and animation styling. --- ## Training & Education ### LLMs - [Oobabooga's text-generation-webui](https://github.com/oobabooga/text-generation-webui/blob/main/docs/Training-LoRAs.md) - [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) - [OpenAI Cookbook](https://github.com/openai/openai-cookbook) - [AemonAlgiz](https://www.youtube.com/@AemonAlgiz) ### Diffusers - [enigmatic_e](https://www.youtube.com/@enigmatic_e) - [Albert Bozesan](https://www.youtube.com/@albertbozesan) - [ControlNet Tutorial](https://www.youtube.com/watch?v=dLM2Gz7GR44) - [Stable Diffusion + ControlNet + TemporalKit + EbSynth Workflow](https://www.youtube.com/watch?v=rlfhv0gRAF4) - [Stable Diffusion + Warp Fusion Workflow](https://www.youtube.com/watch?v=m8xaPnaooyg) --- ## Bonus Recommendations **[AI Business Startup Kit](https://lemmy.world/post/669438)** **LLM Learning Material from the Developer of SuperHOT** ([kaiokendev](https://kaiokendev.github.io/til)): >Here are some resources to help with learning LLMs: > >[Andrej Karpathy’s GPT from scratch](https://www.youtube.com/watch?v=kCc8FmEb1nY) > >[Huggingface’s NLP Course](https://huggingface.co/learn/nlp-course/chapter1/1) > >And for training specifically: > >[Alpaca LoRA](https://github.com/tloen/alpaca-lora#-alpaca-lora) > >[Vicuna](https://github.com/lm-sys/FastChat#fine-tuning) > >[Community training guide](https://rentry.org/llm-training) > >Of course for papers, I recommend reading anything on arXiv’s CS - Computation & Language that looks interesting to you: [https://arxiv.org/list/cs.CL/recent](https://arxiv.org/list/cs.CL/recent). --- ## Support Developers! Please consider donating, subscribing to, or buying a coffee for any of the major community developers advancing Free Open-Source Artificial Intelligence. If you're a developer in this space and would like to have your information added here (or changed), please don't hesitate to [message me](mailto:adyn@divineblade.studio)! [**TheBloke**](https://huggingface.co/TheBloke) - https://www.patreon.com/TheBlokeAI [**Oobabooga**](https://github.com/oobabooga/text-generation-webui) - https://ko-fi.com/oobabooga [**Eric Hartford**](https://erichartford.com/) - https://erichartford.com/ [**kaiokendev**](https://kaiokendev.github.io/) - https://kaiokendev.github.io/ --- **Major FOSAI News & Breakthroughs** - (June 2023) [MPT-30B: Raising the bar for open-source foundation models](https://www.mosaicml.com/blog/mpt-30b) - (May 2023) [Google "We Have No Moat, And Neither Does OpenAI"](https://www.semianalysis.com/p/google-we-have-no-moat-and-neither) - (May 2023) [Introducing MPT-7B: A New Standard for Open-Source, Commercially Usable LLMs](https://www.mosaicml.com/blog/mpt-7b) - (March 2023) [OpenAI Releases Chat-GPT 4](https://openai.com/research/gpt-4) - (November 2022) [OpenAI Releases Chat-GPT 3](https://openai.com/blog/chatgpt) - (December 2017) [Attention Is All You Need](https://arxiv.org/abs/1706.03762)
fedilink

Eric Hartford Announces OpenOrca: An open-source dataset and series of instruct-tuned language models
cross-posted from: https://lemmy.world/post/800062 > **Eric Hartford (a.k.a. faldore) has announced [OpenOrca](https://erichartford.com/openorca), an open-source dataset and series of instruct-tuned language models he plans to release alongside Microsoft's new open-source challenger, [Orca](https://www.microsoft.com/en-us/research/publication/orca-progressive-learning-from-complex-explanation-traces-of-gpt-4/).** > > You can support Eric and all of the hard work he has done for the open-source community by following his newsletter on his site [here](https://erichartford.com/). > > Eric, if you're reading this and would like to share a donation link - I would be more than happy to include it on this post and any future regarding your work. Shoot me a message anytime. > > >**Eric Hartford's Announcement** > > >*Today I'm announcing OpenOrca.* > > >https://erichartford.com/openorca > > >https://twitter.com/erhartford/status/1674214496301383680 > > >The dataset is completed. ~1mil of GPT4 augmented flanv2 instructions and ~3.5mil of GPT3.5 augmented flanv2 instructions. > > >We are currently training on LLaMA-13b. We expect completion in about 2 weeks. > > >When training is complete, we will release the dataset and the model at the same time. > > >We are seeking GPU compute sponsors for various targets, please consult the blog post and reach out if interested. > > >Thank you to our sponsors! > > >https://chirper.ai > > >https://preemo.io > > >https://latitude.sh > > A few more highlights from the full article, which you should read [here](https://erichartford.com/openorca) when you have a chance. > > >We expect to release OpenOrca-LLaMA-13b in mid-July 2023. At that time we will publish our evaluation findings and the dataset. > > >We are currently seeking GPU compute sponsors for training OpenOrca on the following platforms: > > Falcon 7b, 40b > > LLaMA 7b, 13b, 33b, 65b > > MPT-7b, 30b > > Any other targets that get a sponsor. (RWKV, OpenLLaMA) > > >**Dataset consists of:** > > - ~1 million of FLANv2 augmented with GPT-4 completions > > - ~3.5 million of FLANv2 augmented with GPT-3.5 completions If you found this post interesting, please consider subscribing to the [/c/FOSAI](https://lemmy.world/c/fosai) community at [!fosai@lemmy.world](https://lemmy.world/c/fosai) where I do my best to keep you in the know with the most important updates in free open-source artificial intelligence. Thinking about getting into AI but don't know where to start? Believe it or not, now is a great time to start! You can download local models and have them running on your PC today (as long as you have a decent GPU). For more info on exactly how, check out [Your Lemmy Crash Course to Free Open-Source AI ](https://lemmy.world/post/76020)
fedilink