Things are moving at lightning speed in AI Land. On Friday, a software developer named Georgi Gerganov created a tool called "llama.cpp" that can run Meta's new GPT-3-class AI large language model, LLaMA, locally on a Mac laptop. Soon thereafter, people worked out how to run LLaMA on Windows as well. Then someone showed it running on a Pixel 6 phone, and next came a Raspberry Pi (albeit running very slowly).
If this keeps up, we may be looking at a pocket-sized ChatGPT competitor before we know it.
[...]
For example, here's a list of notable LLaMA-related events based on a timeline Willison laid out in a Hacker News comment:
- February 24, 2023: Meta AI announces LLaMA.
- March 2, 2023: Someone leaks the LLaMA models via BitTorrent.
- March 10, 2023: Georgi Gerganov creates llama.cpp, which can run on an M1 Mac.
- March 11, 2023: Artem Andreenko runs LLaMA 7B (slowly) on a Raspberry Pi 4, 4GB RAM, 10 sec/token.
- March 12, 2023: LLaMA 7B running on NPX, a node.js execution tool.
- March 13, 2023: Someone gets llama.cpp running on a Pixel 6 phone, also very slowly.
- March 13, 2023, 2023: Stanford releases Alpaca 7B, an instruction-tuned version of LLaMA 7B that "behaves similarly to OpenAI's "text-davinci-003" but runs on much less powerful hardware.
Related:
DuckDuckGo's New Wikipedia Summary Bot: "We Fully Expect It to Make Mistakes"
Robots Let ChatGPT Touch the Real World Thanks to Microsoft (Article has a bunch of other SoylentNews related links as well.)
Netflix Stirs Fears by Using AI-Assisted Background Art in Short Anime Film
Paper: Stable Diffusion "Memorizes" Some Images, Sparking Privacy Concerns
The EU's AI Act Could Have a Chilling Effect on Open Source Efforts, Experts Warn
Pixel Art Comes to Life: Fan Upgrades Classic MS-DOS Games With AI
(Score: 1) by Runaway1956 on Friday March 17 2023, @12:36AM (4 children)
Let me download and install an AI like ChatGomePhilTom. Do I get the source code? Well, personally, it doesn't really matter, because I can't read source well enough to decide if there is a back door or not. But, the point is, do I get the source code? I compile it myself? Or, is it an executable blob, and I'm just trusting the source? But, it's an AI, right? Who programmed it's ethics, exactly? Maybe calling home is the ethical thing to do?
I'll pass. I don't want my chatbot informing the FBI and NSA of the locations of my nuclear arms caches. Or how much money I'm extorting from Hunter Biden, or . . . most of you should get the idea here.
(Score: 3, Informative) by Freeman on Friday March 17 2023, @02:22PM (3 children)
https://github.com/tatsu-lab/stanford_alpaca [github.com] (Sounds like it's open-source with all the stuff you need to roll-your-own. With a possibility of them releasing the "secret sauce", if Meta says they can. Meta's "secret sauce" was already leaked and widely distributed.)
Joshua 1:9 "Be strong and of a good courage; be not afraid, neither be thou dismayed: for the Lord thy God is with thee"
(Score: 2, Informative) by guest reader on Friday March 17 2023, @06:42PM (2 children)
Alpaca is "just" a fine-tuning of a LLaMA model.
There is also another open source fine-tuning trainer ChatLLaMA [github.com].
Both are based on pre-trained LLaMA models which means that you will still need to fill in Meta's form to obtain the LLaMA’s weights. Pre-trained LLaMA models have restrictive license: do not share, do not sue, nonpermanent, non-commercial use etc.
The training of LLaMA is otherwise described in research paper LLaMA: Open and Efficient Foundation Language Models [facebook.com]. The training of 65B model took 21 days on 2048 A100 GPU cards. This article was a part of the LLaMA announcement [facebook.com].
(Score: 3, Informative) by coolgopher on Saturday March 18 2023, @01:08AM (1 child)
Actually, you do not need to fill out their form; it's available via bittorrent as well [github.com].
(Score: 3, Interesting) by guest reader on Saturday March 18 2023, @07:21AM
Danger, Will Robinson.
The BitTorrent link is just a pull request from a random guy. This pull request is not merged to the official LLaMA Meta Research repository.
The official page [github.com] still contains the following instructions: