About a week before the elections @diegop Diego Pratt published what must be the only, or one of the very few honest analysis of how can blockchains and AI collaborate.
If you have not read it I encourage you to do so here:
Diego focuses on blockchains, and not just any cloud with AI powers, which makes the discussion especially valuable. He makes three crucial points:
- AI on chain is limited by compute power (how many operations per second) and
- On chain memory available to run current open source LLM models such as Llama
- GPUs are not coming to any blockchain any time soon because of the huge challenge that is to achieve determinism while using GPUs which are inherently not determenistic.
I particularly love how he highlighted that the IC can actually run some small LLM models, provided they can run on 4 GB of RAM, and do not require much more than 2 Billion Operations per second, otherwise they will run but responses to a prompt will be quite slow.
The following are the LLMs that would run okay with those limitations according to Chat GPT:
| Model | Parameters (approx.) | Context Length | Quantization | Notes | Source |
|---------------------|----------------------|----------------|--------------|----------------------------------------------|--------|
| **BTLM-3B-8K** | 3 billion | 8,000 tokens | 4-bit | Designed for efficiency, fits within 4GB RAM | [Sci Fi Logic](https://scifilogic.com) |
| **StableLM-3B-4E1T**| 3 billion | Variable | 4-bit | General NLP, works on low-end hardware | [Sci Fi Logic](https://scifilogic.com) |
| **TinyLlama-1.1B** | 1.1 billion | Variable | 4-bit | Optimized for conversation on low-end devices | [Sci Fi Logic](https://scifilogic.com) |
| **phi-1.5** | 1.3 billion | Variable | 4-bit | Common sense, language understanding, diverse NLP tasks | [Sci Fi Logic](https://scifilogic.com) |
I hope this is useful to those building AI on the IC.
To those developers building now AI Agents or solutions, what are you guys using as an LLM?