Hi everyone, I’m excited to share DeVinci, the browser-based AI chatbot app served from the Internet Computer. You can chat with the AI model loaded into your browser so your chats remain fully on your device. If you choose to log in, you can also store your chats on the IC and reload them later.
Thank you for giving it a try and sharing this!
I’m actually facing the same issue currently when I’m trying to integrate bigger AI models into DeVinci and running them on my device (btw, currently DeVinci uses RedPajama-INCITE-Chat-3B-v1-q4f32_0, so a model with ca 3 billion parameters). My best guess at this point is that the browser/device cannot allocate enough memory to the model (which actually requires quite a lot).
Do you happen to have many other programs/processes or other tabs in the browser open? They could block needed memory.
Do you know how much RAM the device you’re running this on has? And does it have a GPU?
Yes it is a Ryzen 5 5600 6c/12t with 32 GB RAM and an intel arc 750 with 8 GB of ram, tried latest Chrome and Canary as well as Brave and Nightly. Also the same on an intel i7 -10700 with iGPU only 16 GB ram still stuck at the same step. Will try on Linux after I get home and dig a little deeper.
Sorry my bad, not sure yet if WSL2 and or Docker or other Windows11 bloat. But on a fresh Win11 with only updates nothing else, it works like a charm. Doesn’t even eat that much ram as I expected. Works on chrome and edge very smooth.
Great, happy to hear And thank you for giving it several tries!
Your machine is actually more powerful than mine, so potentially even the bigger models would run. I’ll see that I integrate the Llama2 model soon and allow users to choose between the models, so you and others with a similar device can actually try the state-of-the-art models
I am actually also looking at running Llama 2 locally before tinkering with it and try running it in a canister since Llama2.c LLM running in a canister! it has been done already. But will do another approach.
Just pushed changes which make the Llama2 model available to power the chat. Under Settings, you can now choose which model you’d like to use (the default is RedPajama 3 billion parameters).
This is the Llama2 7 billion parameters model, so quite a bit bigger than the default and thus also requires a pretty powerful device to run. Anyone who thinks their device is up to the challenge is invited to give it a try
Yes, getting models to run in a canister is amazing. I hope we can make some exponential steps to soon be able to run models like Llama2 7b and beyond in a canister. I’m not sure which improvements we’d need on the protocol/network level to achieve this.
Hi everyone, if you like you can give the new Mistral 7B model a try on DeVinci now: https://x6occ-biaaa-aaaai-acqzq-cai.icp0.io/
To use Mistral 7B, you need to log in and select it under User Settings. Then, on the DeVinci AI Assistant tab, click on Initialize to download it (will take a moment on first download) and once downloaded, you can chat with it as usual. Please let me know if you have any feedback Cheers
Happy to announce that DeVinci now also works on Android
If you want to give it a try, you’ll need the latest Chrome Canary browser on your Android device. You can then download the RedPajama 3B LLM (default, so you just need to click on Initialize) and then chat with it fully on your device as usual: https://x6occ-biaaa-aaaai-acqzq-cai.icp0.io
If you’ve got limited mobile data, please wait until you’ve got a Wifi connection as the LLM is over 1 GB big
Hi everyone, just released a new version with the following updates:
First vector database functionality: you can upload a PDF document which is turned into an in-browser vector database (so stays fully private) and can then be used by the LLM as a knowledge base (RAG)
Markdown for messages works now
Pressing the enter key will send the prompt in the message input field
Please also note that DeVinci is a Progressive Web App and you can thus install it (including using the chat functionality offline).
I’m happy to hear about any feedback or ideas you have
Thrilled to announce that @Nuno and I will redesign the DeVinci user experience and add some great features to it! Our goal is to improve DeVinci such that it can truly be your end-to-end decentralized and fully private AI Chat app as an alternative to common Web2 services The app will stay a hybrid with on-device (easily accessible through the browser) and on-chain components built with and on the Internet Computer.
With @Nuno joining this mission, we’ve got great expertise in UX/UI, design and branding on the team now, and I’m very excited to present you our redesigned app soon. It’ll be on a whole other level
We’re grateful that DFINITY’s supports this work via a dev grant As such, please let us know how DeVinci can become max valuable to you and to the ecosystem
DeVinci is also part of the shared vision with @icpp for decentralized AI where a hybrid network of on-chain and on-device AI services work together seamlessly for the user and everyone contributes (instead of only a few centralized big services providing all AI) → DeAI for the Win
Are you open to providing us feedback as a DeVinci user along the way or even interested in becoming an official tester? Please let me know, would be great to have you
On Internet Computer main site, there’s now an Ai category in the Dapps section. There are like 16 Ai dapps. This must have recently been added because I don’t see any content creators on social media covering it.
I tried your Ai chat a few weeks ago. I would say your Ai dapp is the best on ICP right now if you’re looking to chat about different topics without jumping to different Ai agents. The responses are also fast compared to one other ai chat I used where the responses were slow and limited.
I’ve also tried Elna, ICgpt, and DeAi Chat (DeAI chat not working yet).
Thank you so much for the feedback! If you’re open to sharing it (here or as a DM); which aspects about DeVinci are most important to improve from your perspective? And what would you like to do with the dapp that it currently doesn’t support? Would love to hear more about your ideas and experience, as we’re working to lift DeVinci to another level
My views would be from an average user perspective since i have no programming/developer skills/experience.
I think it would be good to add some of the same features that we find on Open Ai chat gpt, like text-to-image and even text-to-video. The Uix has a basic appearance that can be improved, but I’m guessing that you’re still developing the site. I know the important thing is the performance of the Ai.
I think we have to find more ways for ai to be decentralized and make it something that the user can take advantage of. One feature can involve having an easy process (one that doesn’t require technical skills or needing to be a developer) for users to contribute/improve on the knowledge base. For instance, Elna has an option to create an Ai companion. Maybe make an Ai bot that is managed by a DAO.