|
GPT-4 "Turbo" announced https://openai.com/blog/new-models-and-developer-products-announced-at-devday quote:GPT-4 Turbo with 128K context Looking forward to feeding in an entire 240 page polish sci-fi novel and having GPT-4 turbo translate a chapter or two at a time, with all the relevant context According to reddit you can have up to 11k tokens shared between input and output using the latest stuff with chatgpt and 4-turbo, so.. like 1000 chars for context and prompt and 10,000 chars for programming output? That's pretty cool
|
# ? Nov 7, 2023 06:30 |
|
|
# ? May 5, 2024 11:04 |
|
Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)? Am I correct in presuming that you wil still not be able to, let's say, upload a book or, what Id use it for - upload a DND adventure module, and have chatgpt actually use it?
|
# ? Nov 7, 2023 15:26 |
|
Monglo posted:Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)? I actually use a PDF reader plugin to read my school books and give me summaries of the chapter I'm currently working on so that I can more easily take notes and stuff. it's great because it references the content directly. it would be cool if gpt directly integrated this, but the plug-in has been working mostly fine so far, with the exception of it sometimes telling me it can't access the file. a page reload for some reason usually solves it though.
|
# ? Nov 7, 2023 16:49 |
|
Monglo posted:Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)? No that's exactly what it means. Claude.ai is a big deal vs. ChatGPT because it's context length is so much larger, especially if you say, want to have the LLM be your single-player dungeon master for D&D, Claude can go much further before it runs out of context and starts forgetting things. Or if you dump an adventure module into it then go "hey tell me about x or y", though there are different ways to do that with these systems like hark uses it, summarizing .pdfs on whatever, including adventure modules. The output doesn't care if its scientific theory, mathematics, history, or Star Wars Dark Forces fanfic. There is also a ton of open-source locally-run things that nobody seems to talk about that also does this. Local ones you run that can directly access databases you build yourself, a sort of personal AI assistant that is ready to pull up whatever data you fed it already. There is a lot of specialization in the open source space with these models because trying to run something like chatGPT on consumer hardware isn't exactly possible, but running smaller, specialized ones is.
|
# ? Nov 7, 2023 16:53 |
|
Claude has been giving me guff about copyrighted material lately. I was trying to get it to summarize different aspects of the unproduced Sci-Fi Channel sequel to The Thing and it told me that doing so would cause it to reproduce sections of a copyrighted text which it wasn't allowed to do. I managed to bully it into it eventually, but it was so much of a hassle and I had to repeat the same arguments over and over that I gave up and moved on to something else.
|
# ? Nov 7, 2023 19:44 |
|
feedmyleg posted:Claude has been giving me guff about copyrighted material lately. I was trying to get it to summarize different aspects of the unproduced Sci-Fi Channel sequel to The Thing and it told me that doing so would cause it to reproduce sections of a copyrighted text which it wasn't allowed to do. I managed to bully it into it eventually, but it was so much of a hassle and I had to repeat the same arguments over and over that I gave up and moved on to something else. Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response.
|
# ? Nov 7, 2023 22:16 |
|
KakerMix posted:Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response. I have seen AI Danny DeVitos naked wrinkly rear end from SD. I did not ask for him naked it just decided to make him naked. Very wrinkled
|
# ? Nov 7, 2023 22:20 |
|
pixaal posted:I have seen AI Danny DeVitos naked wrinkly rear end from SD. I did not ask for him naked it just decided to make him naked. Very wrinkled All of this is trained on the internet, which is built by us, so these models output reflections of ourselves. Maybe you really did want his naked rear end, you just didn't know it yet
|
# ? Nov 7, 2023 22:23 |
|
Hadlock posted:Looking forward to feeding in an entire 240 page polish sci-fi novel and having GPT-4 turbo translate a chapter or two at a time, with all the relevant context That reminds me, there was a Polish language biography about murdered artist Zdzisław Beksiński that I wished there was an English language version of maybe five years ago, but now I can't even find what the Polish version was called to take advantage of this.
|
# ? Nov 7, 2023 22:41 |
|
KakerMix posted:No that's exactly what it means. Claude.ai is a big deal vs. ChatGPT because it's context length is so much larger, especially if you say, want to have the LLM be your single-player dungeon master for D&D, Claude can go much further before it runs out of context and starts forgetting things. if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible.
|
# ? Nov 8, 2023 00:48 |
hark posted:bread crumbs
|
|
# ? Nov 8, 2023 01:18 |
|
KakerMix posted:Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response. I think specialized local models will also be aided by the fact that a lot of hardware manufacturers are starting to integrate TPU's into their new designs. I'm even playing with an MCU module right now that has a TPU built into it. Of course it's too small to fit any language model onto but the industry is moving towards machine learning at breakneck speed right now.
|
# ? Nov 8, 2023 01:32 |
|
hark posted:if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible. You can start by using GPT4All. The software is OK but it doesn't work sometimes but it gives you a chat window and will allow you to download models directly through the interface. You can also download any model you like as long as it's supported. There's a big open source one called BLOOM but I don't have the hardware to run it.
|
# ? Nov 8, 2023 01:35 |
|
hark posted:if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible. Growing up we had an ancient computer I played a bunch of text adventure games on, so the fact that these LLMs can provide essentially brand new text adventures at any time and react to whatever you say is ultra compelling to me and where my interest in this currently is. I have not messed with database crawling or more assistant stuff directly, they just occupy the same space when trying to run them and tend to have a lot of crossover so I see them mentioned. I do also work with and mess with all the image generation AI stuff, which probably helps when dealing with the text-based LLM things as well. It's the wild west out there and unfortunately (or fortunately depending on how you feel ) the sex is everywhere, though a lot of the innovation comes specifically from those places. I have a 4090 which gives me 24 gigs of VRAM, allowing me to run some pretty powerful models in the sense of creative writing. I run https://github.com/oobabooga/text-generation-webui as my backend, or as my front end for instruct-based interaction. If you've messed with auto1111 for image generation it's the text version of that. I'll use models from TheBloke: https://huggingface.co/TheBloke Who seems to exist specifically to quantize and shrink models other people make into versions that can run on consumer hardware with various changes in output and speed. If a model gets released this is the guy that will slice it down to fit on whatever hardware you do have. On a 4090 I can easily run 23B models at Q5, or 13b models at Q8 or higher. If you do have access to top-shelf corporate hardware then you can mess with 70b models which are the ones that compete broadly with ChatGPT. License exist for a lot of this stuff so if you want to go commercial you have to check that stuff. For loving around though, no big deal. Sillytavern is a front end that's directly connected to sex roleplay and most users are using it for that, so keep that in mind. However, the capabilities it has facilitates a lot of things that keeps context around when role playing, like your character or whatever world you want to play within. Lock in your character and world and the system will keep that information around which makes for a much less hallucinogenic experience when you just want to crush some goblin skulls. Or play Shadowrun or Cyberpunk or whatever else. If you ever saw or used character.ai, it's that but run locally. You can connect it to OpanAI or claud's API as well and use it that way, though that costs money and is censored. https://github.com/SillyTavern/SillyTavern Reddit is the best evolving source that I've found for general information about all of this: https://www.reddit.com/r/LocalLLaMA/ Tarkus posted:I think specialized local models will also be aided by the fact that a lot of hardware manufacturers are starting to integrate TPU's into their new designs. I'm even playing with an MCU module right now that has a TPU built into it. Of course it's too small to fit any language model onto but the industry is moving towards machine learning at breakneck speed right now. Yep. It's truly above my head but it seems like smaller, specialized LLMs talking to each other is seemingly where we are going, rather than one big giant thing running on million dollar hardware locked behind a single few companies. It is interesting to see the social side of this. I'm older and a lot of my contemporaries have boomered right before my very eyes when it comes to AI stuff, writing it off in the same way that (rightfully) crypto was, claiming it will never get better then when it does (the image space went from 'lol' to 'lol goddamn' in less than a year) say that it doesn't count for some reason. I personally see it as just the march of technology. I already take pictures to remember poo poo with a pocket computer I already have with me, why can't I also have an imagination/thought database thing in the form of a computer I can talk to ala Star Trek?
|
# ? Nov 8, 2023 01:39 |
|
Cross posting this from the ai art thread. Latent consistency seems to produce similar output as latent diffusion, but it's basically instant. This proof of concept art program really shows the power of the faster/more efficient model Pretty exciting stuff, excited to see how the technology gets distilled more and more and made more efficient in just one year Hadlock posted:
|
# ? Nov 10, 2023 23:40 |
|
uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure. some of these are really impressive. seems like discoverability is a bit limited right now though; found this list on reddit: https://www.lingoblocks.com/gpt-store Grimoire isn't on that list but it's pretty cool. https://chat.openai.com/g/g-n7Rs0IK86-grimoire just rugpulling the poo poo out of a bunch of "wrap an app around the openai API" startups.
|
# ? Nov 13, 2023 10:32 |
|
Yes, but it's very silly. Playing around with the basics, I made a Jurassic Park RPG that lets you try to survive the night of the first movie. It's light on the "game" and heavy on the "role playing" but I made a bunch of intricate documents on the world of Jurassic Park and its history, structure and layout of the island, high-level game rules, instructions for how to run an RPG, etc. It can get a bit goofy sometimes, but it generally sells the illusion well. The biggest problem is that I can't give a structure to the AI's response that it actually keeps, so I have to tell it to draw a "Chaos Card" at the end of every turn, which is the random event mechanic. If you play, I recommend your responses being something along the lines of "Grant decides to get out of the car and run for the maintenance shed. Ian objects, saying that he's staying in the car. Dramatize this, then draw a Chaos Card."
|
# ? Nov 13, 2023 16:00 |
|
I'm still waiting for that moment when GPT will be able to run a DND adventure (or play a game in general) Theoretically it should be possible with the option of feeding it the pdf file, but I, too, haven't found a satisfying way to make it follow any structured game rules. If anyone had better success, I'd love to know how you did it.
|
# ? Nov 13, 2023 16:03 |
|
Yeah, that's definitely the biggest issue with any RPGs I've toyed around with. Enforcing "When X happens, do Y" is pretty much impossible at this point with any degree of complexity. What I really want is nested GPTs. So when the player gets into a combat scenario, launch the Combat GPT, when players look at their map, launch the Map GPT, etc. Even sub-GPTs within those would be ideal, so that you can control the output significantly more. The only thing that's been fairly successful for me is describing a general way of operating, rather than specific rules, numbers, or actions. Hence my RPG being much more storytelling-based. Here's the GM instructions I gave to the AI (which I also mostly had the AI write): quote:Dear AI Language Model,
|
# ? Nov 13, 2023 16:29 |
|
feedmyleg posted:Yeah, that's definitely the biggest issue with any RPGs I've toyed around with. Enforcing "When X happens, do Y" is pretty much impossible at this point with any degree of complexity. What I really want is nested GPTs. So when the player gets into a combat scenario, launch the Combat GPT, when players look at their map, launch the Map GPT, etc. Even sub-GPTs within those would be ideal, so that you can control the output significantly more. Yeah, I get the feeling the next wave of these sorts of interactive LLM-based chatbots will integrate them into some sort of traditional rules-based RPG framework, using the LLM to initialize game objects and characters and parse the player's input into model-based game actions and render events and their outcomes into narrative prose.
|
# ? Nov 13, 2023 16:37 |
|
I think that if one were to develop a game for use with an LLM it might be wise to have it be a hybrid between a standard video game developed normally and implementing a small LLM model like mini-orca in conjunction with it. As it stands, a locally instantiated LLM like the one mentioned would work well in interpreting the human language and pushing back dialog in a more natural way but the underlying logic would have to enforce the rules. I think that games made with them will be pretty cool. I've been playing with mini-orca lately and it's a pretty cool model considering how small it is.
|
# ? Nov 13, 2023 17:06 |
|
Mola Yam posted:uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure. Yep, if you use them right they can be an incredible convenience. I am working on a game prototype, and I found myself typing basic information about the game into chatgpt in new chats all the time, so I made a sort of game assistant GPT. I uploaded some of the data files for the game, and described to the creator thing what kind of game I'm making, and now I can just open up a chat with that GPT and it has a pretty decent (if not great) understanding of the game world. It's not magic, and it's not perfect, but it's on par with opening up a chat with ChatGPT+ and telling it about my game and pasting a bunch of data files at it without all the extra steps. I've been working on trying to figure out how to integrate this AI stuff into games for about a year at this point, and it's been equal parts fascinating and frustrating. Any time I try to integrate live API access into a game I run into consistency and reliability issues, but it's pretty amazing for pre-generating stuff that you can sorta curate and edit as necessary and then just bake into the game. The problem I run into repeatedly when trying to do live API access is that there is a tradeoff between interesting creative output and being able to actually put meaningful constraints on it. It's unreliable right up until it's not interesting anymore because you've put so many guard rails on it that you might as well just pre-generate it anyway. I'm sure smarter people than me will figure out how to use this stuff for live server access, but I think the low hanging fruit right now is just learning how to effectively use ChatGPT and poo poo. The API access is useful for little discrete tasks in-engine (as opposed to just doing everything through ChatGPT) but I think you still get the biggest bang for your buck by just building up an intuition for how to work with chatgpt/llms in general.
|
# ? Nov 13, 2023 18:46 |
|
its great for making ideas or scene pieces or off the cuff descriptions of characters and items, but until the memory problem is solved, you'll just get more Starfield NPC stuff happening
|
# ? Nov 13, 2023 18:52 |
|
So I've been loving around with gpts and decided to try and make a guide for haas lathe programming. Works reasonably well, follows prompts and gives proper instructions and code based on the information I gave it. I then took a picture of a hand scrawled drawing representing a simple part. It gave me correct code. That scares and excites me a bit. I tried it with more complex stuff and it didn't quite get it but I know in my heart that it's just a couple of years off (if not less with an AI drawing interpreter) Crazy poo poo.
|
# ? Nov 14, 2023 03:11 |
|
Mola Yam posted:uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure. Neat! You can make your own and share it so here is me loving around with a text adventure GPT: https://chat.openai.com/g/g-XxsjAogdA-adventure-scribe EDIT DREAD PORTALS #wow #cool https://chat.openai.com/g/g-XxsjAogdA-dread-portals KakerMix fucked around with this message at 04:23 on Nov 14, 2023 |
# ? Nov 14, 2023 03:43 |
|
Do you need the paid subscription for these GPTs?
|
# ? Nov 14, 2023 21:10 |
|
KinkyJohn posted:Do you need the paid subscription for these GPTs? Yes
|
# ? Nov 14, 2023 22:37 |
|
What's available if you don't want to pay? Does it just insert ads into the generated text?
|
# ? Nov 14, 2023 22:48 |
|
A Strange Aeon posted:What's available if you don't want to pay? Does it just insert ads into the generated text? As far as I know, nothing. In order to access GPT4 you have to pay.
|
# ? Nov 14, 2023 23:02 |
|
Chatgpt 3.5 turbo is free
|
# ? Nov 15, 2023 05:57 |
|
I played around with "GPT4all" on my work laptop. For casual use it's about 97% as good as commercial ChatGPT free edition, whatever that is this week. Once you start really pushing it, it's more like 70% as good as ChatGPT https://github.com/nomic-ai/gpt4all Literal push buttan get LLM it just works out of the box. On my M3 Pro 24GB with 11 threads enabled I'm getting 25-36 uh I forget tokens per second which, commercial free ChatGPT feels like about 24 tokens/s which is roughly as fast as you can read so pretty good Most of the LLMs are 7b which is pretty low (llama from FB goes to to 50- something, 7 is kind of the lower limit for believable human speech but there's a couple 16b Orca2 can take a pretty good stab at coding complex graphics engines (I like to ask it to render in 2d a wire frame cube that rotates) but so far nothing out of the box compiles on the first try
|
# ? Dec 17, 2023 11:39 |
|
I’ve been playing around a bit with getting llms to run on my Unraid server in a docker container. My gaming background is similar to KakerMix’s in that I grew up playing on old machines and text adventures. They weren’t old when I played them though, haha. I also think that node “masters” to handle various components is the way to go, though admittedly I’m nowhere near technically capable to get there yet as a solo tinkerer. I did have some recent success running an Ollama container as the front-end and toyed with a few models, so, small victories I guess. It’s pretty rad. Next thing I’d like to do is get it to talk to a discord bot and run live text adventures or just gently caress around and riff on it with friends. Then maybe take advantage of some multimodal functionality where it’s feeding prompts to and pulling in images from a separate stable diffusion docker container. Then I think we’d be at least up to like late-80s/early-90s level adventure games but entirely spontaneously generated. Anyone had any success along any of those lines? I had been checking Reddit but that place is a cesspool these days. Anyway… it’s a pretty rad time to be playing with these things, it man it’s frustrating for things to feel so close but still out of reach.
|
# ? Dec 19, 2023 21:52 |
|
/r/LocalLlama is the subreddit you are looking for! Many prominent figures in the local AI dev community post there, and I've noticed very little lovely reddit behavior. I've been playing with my local setup constantly since Mistral released their 8x7b Mixture of Experts model last week. It punches right up there with GPT3.5 Turbo and can be run locally on a card with 24GB of VRAM with a 3.5bpw EXL2 quantization. text-generation-webui is the best client I've used and it stays up to date with the latest bleeding edge quant methods, like EXL2 - 2 . It also can host an OpenAI compatible API that can be used with anything, including prominent chat systems like Silly Tavern. https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard LASER BEAM DREAM fucked around with this message at 18:42 on Dec 21, 2023 |
# ? Dec 21, 2023 18:22 |
|
Orca2 on my laptop feels good enough that I could probably throw out* all my "smart home" devices and replace them with offline stuff. I think GPT4all has an API server option 99.995% of what I use smart home stuff for is light control, timers/alarms/reminders, and random questions/cooking or metric unit conversions GPT4all can already handle 99% of my random search requirements. If there was an option to get it to also spit out Wikipedia article links that would be cool and basically eliminate my need for Google In other news, GitHub "copilot" is scary good, it can explain/write/rewrite, and even guess what the code comments ought to be *Or at least replace the circuit board, the microphone/speakers are really good and worth saving generally
|
# ? Dec 21, 2023 21:59 |
|
LASER BEAM DREAM posted:/r/LocalLlama is the subreddit you are looking for! I was surprised to see dolphin 2.2.1 ranked the same as ChatGPT 3.5, both with an "elo score" of 1000 (chatgpt 3.5 is marginally higher) https://huggingface.co/cognitivecomputations/dolphin-2.2.1-mistral-7b#training quote:Training Puzl.cloud is renting A100 for $1.60/hour so it cost about $310 to train the model on an open source data set in 2 days. That's incredible. I see why everyone is opening up their models right now. Everything created right now is going to be ancient history by next year, open source is mere footsteps behind what openAI has released commercially Hadlock fucked around with this message at 22:10 on Dec 21, 2023 |
# ? Dec 21, 2023 22:07 |
|
Hadlock posted:Orca2 on my laptop feels good enough that I could probably throw out* all my "smart home" devices and replace them with offline stuff. I think GPT4all has an API server option Does anybody have a template or starting point for this type of stuff? For a few years I was employed at some ghastly dead-end "Internet of THiNgS" company trying to market cloud-tethered gadgets and I hated every bit of it. When the company predictably folded, all its "Things" couldn't connect to any "cloud" anymore and so the cloudside APIs stopped working meaning no more integration with Alexa etc. I've been busy with other things since. But I've been building a pretty good offline and java-based "classical" home automation system mainly for lights and HVAC but also for controlling my AV media switches routing video to different displays and audio to different sound systems. I'm mostly just piggybacking off Home Assistant and node red to talk to bluetooth, zigbee and z-wave gadgets and my ESP32 based little DIY light and relay switch controllers. It works, reliably, but all the automation is rule based and laboriously configured manually. I'd love to have a conversational front-end for this stuff that I could trust enough to give admin access for the system. It should understand the admin API comprehensively and be able to perform all the administrative functions such as creating switch linkages and updating rules or overriding them / directly interpreting sensor signals and directly interoperating with individual automation components. The extended and I think still a bit far fetched dream is of course something akin to HAL9000 with less murder. A good "trusted AI housekeeper". Voice interop throughout. I could engineer the audio I/O layer. I have a few different types of audio hats for raspberry pi that uses several microphones for beamforming similar to first-generation Amazon Echo. It works well enough to identify a voice speaking and since I'm caching all the microphone feeds in a looping buffer I can retroactively clean up the beginning of a spoken phrase once I've figured out where the voice is coming from. I have a pretty basic voice signature recognizer that is easily fooled by a recording but can generally tell two speakers from each other with similar tone registers. If I threw a bunch of effort and money at this I could have this in every room and also with speakers for letting the system direct audio responses back to where the request came from. I'd like the system to encompass or connect with a vision module that can look at camera feeds and understand what it sees so that it can use that as contextual clues for situational inference and recognizing individuals and what they're doing. Since it's just my computer staring at my ghastly naked form and not some cloud service I don't care about privacy implications any more than I care about what my cat bears witness to. So I'd be happy to string up cameras pretty much everywhere except inside bathrooms and guest bedroom. The good housekeeper AI should have a cultural base layer and general knowledge of human conventions and physiological basis for user comfort well enough to know what makes sense and what doesn't in general situations. It should have something like a house model so it understands what is adjacent to what, also specific rules governing different places and in different context modes. For example, master bedroom and ensuite vs common room and guest bathroom. Night time. Getting up for a piss, night lights only. Privileges for household members vs Invited guests vs uninvited guests. I imagine that a LLM could be wrapped around this somehow but I don't quite know how to start. And I would probably have to lose my savings and start a doomed company to fully develop this idea. But maybe someone has already got the bones of such as system figured out and I'd like to know more about it. My very naive thought for how this might work is that the conversational LLM is fed prompts from a hub module that just listens to sensors and runs a general schedule. Whenever a sensor state changes, the LLM is fed some type of prompt that says "just now motion sensor 16 triggered." or "a person entered the living room" or "voice command prompt from master bedroom" or some such. And then on its own volition it should perform an overall situational sanity check and figure out if any actions need to be taken to satisfy a hierarchical set of directives that start with basic rules - keep the house locked up unless someone needs egress / ingress - keep the lights off unless someone is home and awake and using / occupying a space or someone has issued an overriding directive that can be situationally understood to apply to a space. Whatever this system is or should be, it should be conversational first and foremost. Able to explain its reasoning. "Why did those lights (stairs light) turn on?" - "Because motion sensor 13 (by the stairs, upper) triggered and (person) was entering the stairs area and looked like they might want to descend." Sorry for the gibberish. TLDR; how HAL less murder pls
|
# ? Dec 22, 2023 10:45 |
|
There was a paper recently on some scientists trying to get a GPT to synthesize tylenol and then asking how to improve yield rates. To do this they had to give it the manual for the chemical analyzer, chemistry books and other open source data, and a couple of hints (you can use a spectrometer to measure color) https://arstechnica.com/science/2023/12/large-language-models-can-figure-out-how-to-do-chemistry/ What you're asking it to do isn't that far off, you'd need a document describing each room/stairwell and their relation to each other and probably some manual training (no, the front stairs is not connected to the hall stairs! the front stairs is outside connected on the porch) I dunno how you would do it exactly but there's already like 9 tutorials on wikihow of how to get voice to text -> text -> chatgpt -> chatgpt response text -> david attenboroughs voice talking to you about the color of your lights consistent json for api consumption: https://www.freecodecamp.org/news/how-to-get-json-back-from-chatgpt-with-function-calling/ The video component and audio response is definitely there already https://www.youtube.com/watch?v=wOEz5xRLaRA I kind of want to write an app that will transcribe all the conversations you have all day, every day, then parse the output for all haikus that spontaneously happened. Then at the end of a year you'd get a personalized book mailed to you called "2023 in Haikus: You Edition" printed with all kinds of stuff that probably violates all kinds of laws and NDAs, in order, and you can try and guess what was happening that day during that spontaneous haiku edit: the philips hue thing has a RESTful API thing, you could probably whip up something with openai's APIs and a raspberry pi in like an hour to proof of life but I can't find anyone's git repo that does it (yet) you'd need a very trigger word system to know when to start recording the audio to send to openAI api for voice transcription tho Hadlock fucked around with this message at 22:20 on Dec 22, 2023 |
# ? Dec 22, 2023 22:08 |
|
thanks for the tips. edit: home automation spam deleted. GBS isn't the forum for this. :-) frumpykvetchbot fucked around with this message at 02:00 on Dec 25, 2023 |
# ? Dec 24, 2023 12:08 |
|
Listening to bay area BART light rail automated robot voice alert thing growl out "SIX. CAR. TRAIN. TO ESS EFF OHH AIIIERR PORT IN TWO. MINUTES. YELLOW LINE TO. ESS EFF OHH NOW ARRIVINGGG" Kind of hoping we can voice clone some local celebrities, auto tune their voice to make them easier to hear/accessible, and then use that Not sure what type of CPU you need for doing dynamic cloned voices that but I'm guessing it'll run on a raspberry pi 5 real time with CPU cycles to spare
|
# ? Apr 8, 2024 17:11 |
|
|
# ? May 5, 2024 11:04 |
|
Now that Hadlock was kind enough to bump this thread I would like to recommend poe.com if you're a few-messages-a-day user of llm's like chapgpt or Claude like I am. You get access to a bunch of different llm's like all of the gpt and Claude APIs plus other ones. It's also cheaper per month than either one, though you get a monthly allotment of compute credits. It also allows you to make your own gpts or bots and its web search bot is pretty good, like perplexity almost. Signed up a few days ago because Claude isn't available here in canuckistan.
|
# ? Apr 8, 2024 18:41 |