I Ran ChatGPT on a Raspberry Pi Locally!

Data Slayer
Data Slayer
160.5 هزار بار بازدید - 11 ماه پیش - Full Tutorial Instructions Here:
Full Tutorial Instructions Here: Patreon: quick-start-step-by-step-guide-to-llm-on-29291

Product Links (some are affiliate links)
  - Raspberry Pi 5 👉 https://amzn.to/48Qgy4O

GitHub Repo: git clone https://github.com/antimatter15/alpac...
Model Weights: https://huggingface.co/Sosaka/Alpaca-...


Here are the instructions to run a ChatGPT-like model locally on your device:

1. Download the zip file corresponding to your operating system from the latest release (https://github.com/antimatter15/alpac.... On Windows, download `alpaca-win.zip`, on Mac (both Intel or ARM) download `alpaca-mac.zip`, and on Linux (x64) download `alpaca-linux.zip`.
2. Download ggml-alpaca-7b-q4.bin(https://huggingface.co/Sosaka/Alpaca-...) and place it in the same folder as the `chat` executable in the zip file.
3. Once you've downloaded the model weights and placed them into the same directory as the `chat` or `chat.exe` executable, run `./chat` in the terminal (for MacOS and Linux) or `.\\Release\\chat.exe` (for Windows).
4. You can now type to the AI in the terminal and it will reply.

If you prefer building from source, follow these instructions:

For MacOS and Linux:

1. Clone the repository using `git clone https://github.com/antimatter15/alpac...`
2. Navigate to the cloned repository using `cd alpaca.cpp`.
3. Run `make chat`.
4. Run `./chat` in the terminal.

1. Download the weights via any of the links in "Get started" above, and save the file as `ggml-alpaca-7b-q4.bin` in the main Alpaca directory.
2. In the terminal window, run `.\\Release\\chat.exe`.
3. You can now type to the AI in the terminal and it will reply.


As part of Meta’s commitment to open science, today we are publicly releasing LLaMA (Large Language Model Meta AI), a state-of-the-art foundational large language model designed to help researchers advance their work in this subfield of AI. Smaller, more performant models such as LLaMA enable others in the research community who don’t have access to large amounts of infrastructure to study these models, further democratizing access in this important, fast-changing field.
making LLaMA available at several sizes (7B, 13B, 33B, and 65B parameters) and also sharing a LLaMA model card that details how we built the model in keeping with our approach to Responsible AI practices.

Over the last year, large language models — natural language processing (NLP) systems with billions of parameters — have shown new capabilities to generate creative text, solve mathematical theorems, predict protein structures, answer reading comprehension questions, and more. They are one of the clearest cases of the substantial potential benefits AI can offer at scale to billions of people.

Smaller models trained on more tokens — which are pieces of words — are easier to retrain and fine-tune for specific potential product use cases. We trained LLaMA 65B and LLaMA 33B on 1.4 trillion tokens. Our smallest model, LLaMA 7B, is trained on one trillion tokens.

Like other large language models, LLaMA works by taking a sequence of words as an input and predicts a next word to recursively generate text. To train our model, we chose text from the 20 languages with the most speakers, focusing on those with Latin and Cyrillic alphabets.

There is still more research that needs to be done to address the risks of bias, toxic comments, and hallucinations in large language models. Like other models, LLaMA shares these challenges. As a foundation model, LLaMA is designed to be versatile and can be applied to many different use cases, versus a fine-tuned model that is designed for a specific task. By sharing the code for LLaMA, other researchers can more easily test new approaches to limiting or eliminating these problems in large language models. We also provide in the paper a set of evaluations on benchmarks evaluating model biases and toxicity to show the model’s limitations and to support further research in this crucial area.

To maintain integrity and prevent misuse, we are releasing our model under a noncommercial license focused on research use cases. Access to the model will be granted on a case-by-case basis to academic researchers; those affiliated with organizations in government, civil society, and academia; and industry research laboratories around the world. People interested in applying for access can find the link to the application in our research paper.

We believe that the entire AI community — academic researchers, civil society, policymakers, and industry — must work together to develop clear guidelines around responsible AI in general and responsible large language models in particular. We look forward to seeing what the community can learn — and eventually build — using LLaMA.
11 ماه پیش در تاریخ 1402/06/07 منتشر شده است.
160,524 بـار بازدید شده
... بیشتر