|
Hammerite posted:This is a very pedestrian question, but I wanted to try out the ChatGPT demo and when I try to log in it just says "The email you provided is not supported". This occurs consistently when trying to log in in several ways: ChatGPT can be weird to sign up for. As an alternative perhaps sign up for poe.com. Its got an interface to a bunch of these models, in limited capacities. I'm rather fond of "Claude" from anthropics, its about on level with GPT4/ChatGPT+, perhaps a little behind , but its ever so well behaved and seems pretty good at explaining itself. The only problem is one it shares with GPT, which is hallucinating its brain out if you ask for citations. A lot of folk in the AI research community are fond of Claude, as it "ConstitutionalAI" (which apparently is slightly different to the usual RLHF method of politeness training) seems to work really well, so its kind of a helpful friendly little dude without Bing's freakouts. duck monster fucked around with this message at 06:12 on Apr 26, 2023 |
# ¿ Apr 26, 2023 06:10 |
|
|
# ¿ May 21, 2024 07:33 |
|
Rahu posted:I've been trying to learn some ML stuff lately and to that end I've been reading over Andrej Karpathy's nanoGPT. Transormers are probably *not* the best place to start with learning how ML actually works. It may well be the god algorithm for the modern robobrain, but its built on a bunch of simpler, but still quite useful ideas that are worth learning first.
|
# ¿ Jun 23, 2023 18:10 |
|
Mata posted:Bumping this thread with more of a practical question about LLM rather than the scientific aspect of ML... You can, but mostly just the low end ones. You really want to get the meatiest GPU you can, and with the real stickler being how much ram does it have (Theres no point having a firebreathing gpu if it cant fit the model in its brain). 3080s with 24gb seem to be the sweet spot. You *can* run them on a CPU too, and that can solve the memory issue but they are slow as balls, though there is a lot of research ongoing on fixing both the ram and the CPU power. But for reference, I can happily run GPT2 on my macbook on CPU. Its a bit slow, and kinda dumb, but its neat to gently caress about with. I can also run Llama 7b quantized to 4bit weights, but it behaves a little drunk with that much quantization.
|
# ¿ Sep 10, 2023 09:35 |
|
I'm seeing some scuttlebut around the net about AMDs new Instinct GPUs being actually...... good for AI? Aparently the A100s kick them to the curb a lot of the tensorflow metrics, but a lot of that uses 32bit operation, and AMDs offering thrashes the A100 on 64bit ops. At about $14K. Thus if the models can be adapted to be optimized for 64bit math rather than 32bit math (I'm not sure what it'd gain out that but hey) the Instincts might be a real contender at a significantly lower price.
|
# ¿ Sep 27, 2023 15:34 |