Another AI chatbot
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
I've got 3 Tesla T4s now but pricing has gone up in the last month, probably because of Bitcoin, even though the T4s would suck at mining, so I'm having a hard time getting below $500, where just last month they were selling for $400, I could kick myself! I still want 3 more, maybe even a total of 10 later. Anyways, VRing my dolls with AI is proving to be a fun challenge, better than supping up an old sports car.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
-
- Active Poster
- Posts: 70
- Joined: Tue May 14, 2024 1:36 pm
- Contact:
Re: Another AI chatbot
Was looking at - https://tykeai.com/ which is another mixed reality Qwest avatar hooked up to an LLM. Not much info and $3.50 a month, but creator mentions NSFW character prompt may not work, so I suspect he's just using ChatGPT for it all. If I could find something like that that was uncensored and connected to local LLMs, I'd happily pay for subscription. VR Hot is same situation, and virt a mate was kicked off steam. So annoying.
Might just get 24GB Mac Mini M4. You can fully offload 30B Q4 models and have room left over for other things, but won't be able to load 70B. It's only $1000 though. I've been using Qwen 2.5 32B Coder and it's been really good. I've been getting better one-shot answers than ChatGPT many times and you can set a huge context window. It's not nearly as good as claude but for 32B it's really good.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
That Mac Mini M4 is no where near a Tesla T4, if you spend the grand for 2 T4s: you get 32GB of ram, 130 TFLOPs of fp16, and 260 TOPs of INT8! The M4 gives you 38 TOPs of INT8 and all the other precision TFLOPs are way below the Tesla T4.
Oh, I got my MB and it's working fine now, so it was the MB and now I've got 4 PCIe where I can split 3 to accommodate 6 T4s, the last PCIe is for graphics.
-
- Active Poster
- Posts: 70
- Joined: Tue May 14, 2024 1:36 pm
- Contact:
Re: Another AI chatbot
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
I also picked up another T4 for $521.00! That makes 4, hoping to get better pricing for another 2.
-
- Active Poster
- Posts: 70
- Joined: Tue May 14, 2024 1:36 pm
- Contact:
Re: Another AI chatbot
64GB of VRAM should be fun! I wasn't liking the RTX 5000 series leaked specs, so I just went with a $800 used 3090 Ti. Should come in on Monday and just built a 9900X system for it. I'll probably regret it in a few months after the 5090 comes out and the 9950X3D, but at least it didn't cost 4-5k.
Re: Another AI chatbot
it's better wait for a run locally, there's somes in progress at youtube others are not open source as need to be(sadly), but yea i hope my 3060 ti 8gb can do something in the futureBarfalamule wrote: ↑Fri Nov 15, 2024 11:40 pm Second life is crazy. I haven't used it in over a decade. Just searching and found AFK sex is a thing. If you could hook up an AFK bot to an LLM, might be able to earn a ton of lidens if that's how it still works lol. Trade linden for BTC, pay for GPUs. You might be onto something =)
Was looking at - https://tykeai.com/ which is another mixed reality Qwest avatar hooked up to an LLM. Not much info and $3.50 a month, but creator mentions NSFW character prompt may not work, so I suspect he's just using ChatGPT for it all. If I could find something like that that was uncensored and connected to local LLMs, I'd happily pay for subscription. VR Hot is same situation, and virt a mate was kicked off steam. So annoying.
Might just get 24GB Mac Mini M4. You can fully offload 30B Q4 models and have room left over for other things, but won't be able to load 70B. It's only $1000 though. I've been using Qwen 2.5 32B Coder and it's been really good. I've been getting better one-shot answers than ChatGPT many times and you can set a huge context window. It's not nearly as good as claude but for 32B it's really good.
-
- Active Poster
- Posts: 70
- Joined: Tue May 14, 2024 1:36 pm
- Contact:
Re: Another AI chatbot
Sadly, I got ripped off on the ebay 3090 and was sent a HD7870. I should've known but looks like ebay is issuing a refund so not that big of a deal.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
Yeah, those too good to be true deals can suck on eBay, and it's an aspect of eBay since forever, it just makes the platform a little sleezy, but I don't think there's any real competitor, other than Alibaba, where one of my first dolls I actually did buy from them and it was exactly as described, but other stuff really looks too good to be true. Glad you're getting your money back.Barfalamule wrote: ↑Tue Nov 26, 2024 11:21 pm Hopefully one comes out that let's you swap models for performance just like you can on any video game, so downgrade graphics and LLM models\TTS based on system.
Sadly, I got ripped off on the ebay 3090 and was sent a HD7870. I should've known but looks like ebay is issuing a refund so not that big of a deal.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
OK...I might be a little too gitty. One thing though trying to run the larger models on all cards is still sluggish which I don't have problem with since my original intention was to use each gpu for a particular task using the under 16GB limit of each card were those models perform well.
I have run into an issue where the LLM thinks it can do something that it physically can't, since it doesn't have any pre-check of the hardware or software configuration it's hosted in it treats a request like: "search the internet" just like any other sentence. So, I'm gonna have to give the LLM a crude sense of body. I'm sure RAG will be able to handle that.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
-
- Active Poster
- Posts: 70
- Joined: Tue May 14, 2024 1:36 pm
- Contact:
Re: Another AI chatbot
I'm gonna be using my older GPU for slowly processing videos mostly and then use new system for bot program and gaming. I just ordered another 3090 from Amazon but not the Ti. Good reviews from seller, so should be fine.
I also got my first doll from Alibaba almost a decade ago. It melted when my AC went out for a few days
For web search\RAG, I was looking at this that lets ollama models search the web- https://www.youtube.com/watch?v=txflvGG_hIc . Once I get new machine working, might try to update program to work with something like that, ollama and vision models since LM Studio doesn't support latest vision models for now.
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
- frankinstien
- Doll Advisor
- Posts: 596
- Joined: Thu Jan 07, 2021 2:25 am
- Contact:
Re: Another AI chatbot
There are some PCIe issues where the MB distributing the bandwidth as 16x, 8x, 16x, 8x not full 16x across all slots! So, 4 T4s will have 8x bandwidth, while the other 2 will have 4x bandwidth. This is ok since each GPU will be dedicated to a specific specialized model, where feeding data that is text or audio doesn't need as much bandwidth so those will be loaded onto the T4s on the 4x+4x slot.
You might ask why not just upgrade the motherboard, well the TRX40 MBs don't support full 16x on all lanes, but the Threadripper PRO MBs do and have 7 PCIe lanes, just what I need! But I would have to buy a new CPU in addition to the MB, the cost is pretty high for the CPU and MB right now. So maybe in the future if prices come down. But I'm happy with the rig and it should work as intended.
Because of the tight spacing of the PCIe slots I had to work out a scheme to place all the gpus. In the image below you'll notice the splitter card and how I have to space the cards with top and bottom stages. Where two T4s, top and bottom, will fit adjacent to the RTX 3080 ti, and another 2 will fit to the left of the CPU fan.