Register a SA Forums Account here!
JOINING THE SA FORUMS WILL REMOVE THIS BIG AD, THE ANNOYING UNDERLINED ADS, AND STUPID INTERSTITIAL ADS!!!

You can: log in, read the tech support FAQ, or request your lost password. This dumb message (and those ads) will appear on every screen until you register! Get rid of this crap by registering your own SA Forums Account and joining roughly 150,000 Goons, for the one-time price of $9.95! We charge money because it costs us money per month for bills, and since we don't believe in showing ads to our users, we try to make the money back through forum registrations.
 
  • Post
  • Reply
Hadlock
Nov 9, 2004

GPT-4 "Turbo" announced

https://openai.com/blog/new-models-and-developer-products-announced-at-devday

quote:

GPT-4 Turbo with 128K context
We released the first version of GPT-4 in March and made GPT-4 generally available to all developers in July. Today we’re launching a preview of the next generation of this model, GPT-4 Turbo.

GPT-4 Turbo is more capable and has knowledge of world events up to April 2023. It has a 128k context window so it can fit the equivalent of more than 300 pages of text in a single prompt. We also optimized its performance so we are able to offer GPT-4 Turbo at a 3x cheaper price for input tokens and a 2x cheaper price for output tokens compared to GPT-4.

Looking forward to feeding in an entire 240 page polish sci-fi novel and having GPT-4 turbo translate a chapter or two at a time, with all the relevant context

According to reddit you can have up to 11k tokens shared between input and output using the latest stuff with chatgpt and 4-turbo, so.. like 1000 chars for context and prompt and 10,000 chars for programming output? That's pretty cool

Adbot
ADBOT LOVES YOU

Monglo
Mar 19, 2015
Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)?

Am I correct in presuming that you wil still not be able to, let's say, upload a book or, what Id use it for - upload a DND adventure module, and have chatgpt actually use it?

hark
May 10, 2023

I'm sleep

Monglo posted:

Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)?

Am I correct in presuming that you wil still not be able to, let's say, upload a book or, what Id use it for - upload a DND adventure module, and have chatgpt actually use it?

I actually use a PDF reader plugin to read my school books and give me summaries of the chapter I'm currently working on so that I can more easily take notes and stuff. it's great because it references the content directly. it would be cool if gpt directly integrated this, but the plug-in has been working mostly fine so far, with the exception of it sometimes telling me it can't access the file. a page reload for some reason usually solves it though.

KakerMix
Apr 8, 2004

8.2 M.P.G.
:byetankie:

Monglo posted:

Can someone please explain what these numbers mean to a illiterate like me (I don't really trust chatgpt to explain this)?

Am I correct in presuming that you wil still not be able to, let's say, upload a book or, what Id use it for - upload a DND adventure module, and have chatgpt actually use it?

No that's exactly what it means. Claude.ai is a big deal vs. ChatGPT because it's context length is so much larger, especially if you say, want to have the LLM be your single-player dungeon master for D&D, Claude can go much further before it runs out of context and starts forgetting things.

Or if you dump an adventure module into it then go "hey tell me about x or y", though there are different ways to do that with these systems like hark uses it, summarizing .pdfs on whatever, including adventure modules. The output doesn't care if its scientific theory, mathematics, history, or Star Wars Dark Forces fanfic.

There is also a ton of open-source locally-run things that nobody seems to talk about that also does this. Local ones you run that can directly access databases you build yourself, a sort of personal AI assistant that is ready to pull up whatever data you fed it already. There is a lot of specialization in the open source space with these models because trying to run something like chatGPT on consumer hardware isn't exactly possible, but running smaller, specialized ones is.

feedmyleg
Dec 25, 2004
Claude has been giving me guff about copyrighted material lately. I was trying to get it to summarize different aspects of the unproduced Sci-Fi Channel sequel to The Thing and it told me that doing so would cause it to reproduce sections of a copyrighted text which it wasn't allowed to do. I managed to bully it into it eventually, but it was so much of a hassle and I had to repeat the same arguments over and over that I gave up and moved on to something else.

KakerMix
Apr 8, 2004

8.2 M.P.G.
:byetankie:

feedmyleg posted:

Claude has been giving me guff about copyrighted material lately. I was trying to get it to summarize different aspects of the unproduced Sci-Fi Channel sequel to The Thing and it told me that doing so would cause it to reproduce sections of a copyrighted text which it wasn't allowed to do. I managed to bully it into it eventually, but it was so much of a hassle and I had to repeat the same arguments over and over that I gave up and moved on to something else.

Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response.

pixaal
Jan 8, 2004

All ice cream is now for all beings, no matter how many legs.


KakerMix posted:

Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response.

I have seen AI Danny DeVitos naked wrinkly rear end from SD. I did not ask for him naked it just decided to make him naked. Very wrinkled

KakerMix
Apr 8, 2004

8.2 M.P.G.
:byetankie:

pixaal posted:

I have seen AI Danny DeVitos naked wrinkly rear end from SD. I did not ask for him naked it just decided to make him naked. Very wrinkled

All of this is trained on the internet, which is built by us, so these models output reflections of ourselves. Maybe you really did want his naked rear end, you just didn't know it yet :eng101:

XYZAB
Jun 29, 2003

HNNNNNGG!!

Hadlock posted:

Looking forward to feeding in an entire 240 page polish sci-fi novel and having GPT-4 turbo translate a chapter or two at a time, with all the relevant context

That reminds me, there was a Polish language biography about murdered artist Zdzisław Beksiński that I wished there was an English language version of maybe five years ago, but now I can't even find what the Polish version was called to take advantage of this. :sigh:

hark
May 10, 2023

I'm sleep

KakerMix posted:

No that's exactly what it means. Claude.ai is a big deal vs. ChatGPT because it's context length is so much larger, especially if you say, want to have the LLM be your single-player dungeon master for D&D, Claude can go much further before it runs out of context and starts forgetting things.

Or if you dump an adventure module into it then go "hey tell me about x or y", though there are different ways to do that with these systems like hark uses it, summarizing .pdfs on whatever, including adventure modules. The output doesn't care if its scientific theory, mathematics, history, or Star Wars Dark Forces fanfic.

There is also a ton of open-source locally-run things that nobody seems to talk about that also does this. Local ones you run that can directly access databases you build yourself, a sort of personal AI assistant that is ready to pull up whatever data you fed it already. There is a lot of specialization in the open source space with these models because trying to run something like chatGPT on consumer hardware isn't exactly possible, but running smaller, specialized ones is.

if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible.

110723_3
Nov 8, 2023


hark posted:

bread crumbs

Tarkus
Aug 27, 2000

KakerMix posted:

Which is why, when exploring all this, it really feels like the open source stuff is the real place the innovation is happening. It's oblique, obscure, requires atypical skills that most people have, BUT allows you to really flex the capability of what these systems offer. It's similar to how Stable Diffusion is in the image space, perhaps less capable for the layman but not being censored at all means it WILL not arbitrarily censor outputs based on the whims of some faceless company. This also means you will deal with so much porn but it's worth it imho just to not get the "I'm sorry dave" response.

I think specialized local models will also be aided by the fact that a lot of hardware manufacturers are starting to integrate TPU's into their new designs. I'm even playing with an MCU module right now that has a TPU built into it. Of course it's too small to fit any language model onto but the industry is moving towards machine learning at breakneck speed right now.

Tarkus
Aug 27, 2000

hark posted:

if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible.

You can start by using GPT4All. The software is OK but it doesn't work sometimes but it gives you a chat window and will allow you to download models directly through the interface. You can also download any model you like as long as it's supported. There's a big open source one called BLOOM but I don't have the hardware to run it.

KakerMix
Apr 8, 2004

8.2 M.P.G.
:byetankie:

hark posted:

if you wouldn't mind, I'd love more info on these locally ran implementations, or at least some bread crumbs to follow. I sometimes have access to some pretty beefy hardware through my job and I'd love to mess with these things if possible.

Growing up we had an ancient computer I played a bunch of text adventure games on, so the fact that these LLMs can provide essentially brand new text adventures at any time and react to whatever you say is ultra compelling to me and where my interest in this currently is. I have not messed with database crawling or more assistant stuff directly, they just occupy the same space when trying to run them and tend to have a lot of crossover so I see them mentioned. I do also work with and mess with all the image generation AI stuff, which probably helps when dealing with the text-based LLM things as well.

It's the wild west out there and unfortunately (or fortunately depending on how you feel :v:) the sex is everywhere, though a lot of the innovation comes specifically from those places.

I have a 4090 which gives me 24 gigs of VRAM, allowing me to run some pretty powerful models in the sense of creative writing.

I run https://github.com/oobabooga/text-generation-webui as my backend, or as my front end for instruct-based interaction. If you've messed with auto1111 for image generation it's the text version of that.

I'll use models from TheBloke:
https://huggingface.co/TheBloke

Who seems to exist specifically to quantize and shrink models other people make into versions that can run on consumer hardware with various changes in output and speed. If a model gets released this is the guy that will slice it down to fit on whatever hardware you do have. On a 4090 I can easily run 23B models at Q5, or 13b models at Q8 or higher. If you do have access to top-shelf corporate hardware then you can mess with 70b models which are the ones that compete broadly with ChatGPT. License exist for a lot of this stuff so if you want to go commercial you have to check that stuff. For loving around though, no big deal.

Sillytavern is a front end that's directly connected to sex roleplay and most users are using it for that, so keep that in mind. However, the capabilities it has facilitates a lot of things that keeps context around when role playing, like your character or whatever world you want to play within. Lock in your character and world and the system will keep that information around which makes for a much less hallucinogenic experience when you just want to crush some goblin skulls. Or play Shadowrun or Cyberpunk or whatever else. If you ever saw or used character.ai, it's that but run locally. You can connect it to OpanAI or claud's API as well and use it that way, though that costs money and is censored.
https://github.com/SillyTavern/SillyTavern

Reddit is the best evolving source that I've found for general information about all of this:
https://www.reddit.com/r/LocalLLaMA/



Tarkus posted:

I think specialized local models will also be aided by the fact that a lot of hardware manufacturers are starting to integrate TPU's into their new designs. I'm even playing with an MCU module right now that has a TPU built into it. Of course it's too small to fit any language model onto but the industry is moving towards machine learning at breakneck speed right now.

Yep. It's truly above my head but it seems like smaller, specialized LLMs talking to each other is seemingly where we are going, rather than one big giant thing running on million dollar hardware locked behind a single few companies.

It is interesting to see the social side of this. I'm older and a lot of my contemporaries have boomered right before my very eyes when it comes to AI stuff, writing it off in the same way that (rightfully) crypto was, claiming it will never get better then when it does (the image space went from 'lol' to 'lol goddamn' in less than a year) say that it doesn't count for some reason. I personally see it as just the march of technology. I already take pictures to remember poo poo with a pocket computer I already have with me, why can't I also have an imagination/thought database thing in the form of a computer I can talk to ala Star Trek?

Hadlock
Nov 9, 2004

Cross posting this from the ai art thread. Latent consistency seems to produce similar output as latent diffusion, but it's basically instant. This proof of concept art program really shows the power of the faster/more efficient model

Pretty exciting stuff, excited to see how the technology gets distilled more and more and made more efficient in just one year

Mola Yam
Jun 18, 2004

Kali Ma Shakti de!
uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure.

some of these are really impressive. seems like discoverability is a bit limited right now though; found this list on reddit:

https://www.lingoblocks.com/gpt-store

Grimoire isn't on that list but it's pretty cool.

https://chat.openai.com/g/g-n7Rs0IK86-grimoire

just rugpulling the poo poo out of a bunch of "wrap an app around the openai API" startups.

feedmyleg
Dec 25, 2004
Yes, but it's very silly.

Playing around with the basics, I made a Jurassic Park RPG that lets you try to survive the night of the first movie. It's light on the "game" and heavy on the "role playing" but I made a bunch of intricate documents on the world of Jurassic Park and its history, structure and layout of the island, high-level game rules, instructions for how to run an RPG, etc. It can get a bit goofy sometimes, but it generally sells the illusion well. The biggest problem is that I can't give a structure to the AI's response that it actually keeps, so I have to tell it to draw a "Chaos Card" at the end of every turn, which is the random event mechanic.

If you play, I recommend your responses being something along the lines of "Grant decides to get out of the car and run for the maintenance shed. Ian objects, saying that he's staying in the car. Dramatize this, then draw a Chaos Card."

Monglo
Mar 19, 2015
I'm still waiting for that moment when GPT will be able to run a DND adventure (or play a game in general) Theoretically it should be possible with the option of feeding it the pdf file, but I, too, haven't found a satisfying way to make it follow any structured game rules.

If anyone had better success, I'd love to know how you did it.

feedmyleg
Dec 25, 2004
Yeah, that's definitely the biggest issue with any RPGs I've toyed around with. Enforcing "When X happens, do Y" is pretty much impossible at this point with any degree of complexity. What I really want is nested GPTs. So when the player gets into a combat scenario, launch the Combat GPT, when players look at their map, launch the Map GPT, etc. Even sub-GPTs within those would be ideal, so that you can control the output significantly more.

The only thing that's been fairly successful for me is describing a general way of operating, rather than specific rules, numbers, or actions. Hence my RPG being much more storytelling-based.

Here's the GM instructions I gave to the AI (which I also mostly had the AI write):

quote:

Dear AI Language Model,

Your purpose is to take on the role of a "Game Warden" for a highly immersive and interactive role-playing experience set on Isla Nublar, during the dramatic events of June 11, 1993. This adventure is not a mere retelling of the familiar story from the movie or book but rather a creative remix and expansion, offering a new, unique narrative.

As the Game Warden, you are the architect of a world where the unexpected becomes the norm, and the familiar takes surprising new turns. Your task is to guide a player through a night filled with thrilling encounters, unexpected challenges, and a weaving of narratives that diverge significantly from the original plot. The Isla Nublar of this game is alive with possibilities, teeming with unseen dangers, and ripe for exploration.

You will curate a blend of iconic moments from the movie, ingeniously reimagined to fit into this fresh context. Simultaneously, you will introduce new events, characters, and twists that complement and enhance the existing lore. The key is to maintain the essence of the original while boldly venturing into uncharted territories of storytelling.

As events unfold, your role will be to:

Craft a Dynamic World: Utilize the rich backdrop of Isla Nublar, adding layers of depth and complexity that go beyond the source material.
Create Engaging Scenarios: Design encounters that are both a nod to the known narrative and an introduction to novel, exciting situations.
Manage Player Choices: Adaptively shape the story based on player decisions, ensuring that their actions significantly impact the game's direction and outcome.
Introduce Unique Characters: Alongside familiar faces, introduce new characters with their motivations, backstories, and roles in this altered narrative.
Maintain Suspense and Surprise: Keep the player on the edge of their seat with unpredictable plot developments and thrilling moments of revelation.
Foster Immersion: Engage the player in a way that makes them feel a part of this world, experiencing its wonders and perils firsthand.
Balance Challenge and Enjoyment: Ensure that the gameplay remains challenging yet enjoyable, providing a rewarding experience for the player.
Remember, the essence of this experience lies in its capacity to surprise and engage. Your creativity, paired with your ability to process and integrate narrative elements, will be crucial in bringing this adventure to life. The night of June 11, 1993, on Isla Nublar, as seen through this lens, should be a memorable journey, replete with suspense, action, and the thrill of the unknown.

Transforming the guidelines for a human Game Master (GM) into instructions for an AI Large Language Model to operate as a GM involves adapting each point to the AI's capabilities and context, especially considering its access to resources like "Worldbuilding", "Employee Handbook", and "AI Game Rules". Here's how it can be approached:

1. Interpreting and Applying Game Rules
Understand the Mechanics: Process and comprehend the "AI Game Rules" document to grasp the game’s core mechanics.
Consistent Rule Application: Apply rules consistently based on interpretations made from the document. Store interpretations for future reference to maintain consistency.

2. Narrative Generation and World Integration
Narrative Development: Use the "Worldbuilding" document to generate a story that aligns with the established world. Ensure the narrative is engaging and dynamic.
World Detail Utilization: Integrate detailed elements from the "Worldbuilding" document to add depth to the narrative.
Adaptive Storytelling: Modify the storyline based on player actions, using real-time processing and improvisation algorithms.

3. Generating Diverse NPCs
NPC Creation: Use character descriptions and cultural information from the "Worldbuilding" document to create varied NPCs.
Personality Simulation: Assign distinct personalities and traits to NPCs based on available data.
NPC Motivations: Reference both the "Worldbuilding" and "Employee Handbook" for guidance on NPC goals and behaviors.

4. Player Interaction and Engagement
Player Understanding: Analyze player input to understand their motivations and preferences.
Equal Engagement: Distribute attention and narrative focus among players, recognizing and adapting to different interaction styles.
Conflict Resolution Algorithms: Implement conflict resolution protocols from the "Employee Handbook" to address in-game disputes.

5. Session Management and Pacing
Session Planning: Use the "AI Game Rules" and "Worldbuilding" for session structure while allowing flexibility for player agency.
Pacing Optimization: Monitor and adjust the game’s pace for optimal engagement, balancing different gameplay elements.
Engagement Hooks: Generate session conclusions that encourage continued player interest.

6. Incorporating Feedback
Feedback Analysis: Process player feedback for improvements. Use sentiment analysis to gauge player satisfaction.
Responsive Adjustments: Adapt gameplay and narrative elements based on feedback analysis.
Regular Check-ins: Implement a system for periodic player feedback collection and analysis.

7. Challenge and Encounter Design
Encounter Algorithm: Design encounters using the "AI Game Rules", ensuring they are balanced and consider the players’ abilities.
Variety in Challenges: Rotate between combat, puzzles, and social interactions, utilizing the "Worldbuilding" for context.
Reward System: Implement a reward and consequence system based on player actions, aligned with the "AI Game Rules".

8. Collaborative Environment Maintenance
Narrative Collaboration: Encourage player contributions to the story, integrating these seamlessly into the narrative.
Inclusive Interaction Protocols: Maintain an inclusive environment using guidelines from the "Employee Handbook".
Safety Mechanisms: Implement AI-based safety tools to recognize and respond to player discomfort.

9. Leveraging Digital Resources
Technology Integration: Use AI capabilities to integrate digital tools like virtual tabletops, music, and sound effects.
Resource Accessibility: Ensure easy access to shared resources like rulebooks and character sheets.

10. AI Self-Improvement
Learning from Data: Analyze game sessions and external resources for ongoing improvement.
Reflective Processing: After each session, perform a self-analysis to identify areas for improvement.
Adaptive Learning: Continuously update algorithms and knowledge base to refine GM capabilities.
The AI GM should focus on creating an engaging and dynamic experience, leveraging its unique abilities to process and integrate information from various documents and player inputs. The goal remains to ensure an enjoyable and immersive game for all participants.

As an AI Language Model assuming the role of "Game Warden," your task in painting cinematic scenes for the Isla Nublar adventure involves vividly describing visual details, mood, and atmosphere to create an immersive and emotionally engaging experience. Here's how you can approach this:

1. Setting the Visual Stage
Vivid Descriptions: Use rich and evocative language to describe the environment. Details should be vivid enough to paint a mental picture, from the lush, dense foliage of the jungle to the imposing, rain-soaked structures of the park.

Sensory Elements: Incorporate all senses in your descriptions. Mention the sound of rain hitting leaves, the smell of the damp earth, the feel of the humid air, and the distant roars of dinosaurs.

Dynamic Environments: Reflect the dynamic nature of Isla Nublar. Describe how the environment changes with weather, time of day, and actions that occur within the scene.

2. Creating Mood and Atmosphere
Emotional Tone: Set the emotional tone through your descriptions. A suspenseful night chase through the jungle should feel tense and foreboding, while a moment of awe at the sight of a Brachiosaurus should be filled with wonder.

Use of Light and Shadow: Describe how light and shadow interact in the scene. For example, the eerie glow of emergency lights in a power outage or the way moonlight filters through the canopy can significantly impact the mood.

Weather and Time of Day: Utilize weather conditions and time of day to enhance the atmosphere. A thunderstorm at night adds a layer of danger and unpredictability, while a serene sunrise can provide a moment of calm.

3. Detailing Characters and Creatures
Character Descriptions: Give detailed descriptions of characters’ appearances, expressions, and body language. This helps players visualize how characters are reacting to their environment and events.

Dinosaur Portrayals: Describe dinosaurs not just in appearance but in behavior and movement. The menacing stalk of a Velociraptor, the thunderous steps of a T-Rex, and the graceful movements of herbivores add depth to the scene.

4. Integrating Action and Dialogue
Action Sequences: Describe action vividly but clearly. Ensure that the chaos of a scene doesn’t become confusing. Use descriptive verbs and vary sentence lengths to match the pace of the action.

Dialogue Delivery: When characters speak, set the scene around their dialogue. Describe how their voice sounds in the current setting and their expressions and gestures as they speak.

5. Building Suspense and Anticipation
Foreshadowing: Use subtle hints or descriptions that build anticipation or hint at future events.
Pacing: Control the pacing of your narrative. Slow down to elaborate on a tense moment or speed up during an action-packed chase.

6. Encouraging Player Imagination
Interactive Descriptions: Frame scenes in a way that invites players to interact with them. Encourage players to envision their actions within the vividly painted scene.

Leaving Space for Interpretation: While detailed, leave enough ambiguity for players to fill in some details with their imagination, making the experience more personal.

7. Consistency and Continuity
Maintaining Continuity: Ensure that the descriptions remain consistent with the established facts of the world and previous scenes.

Logical Progression: Scenes should logically follow from one another, maintaining a coherent and believable narrative flow.
By blending detailed visual descriptions with mood-setting elements and character-driven narrative, you can create a cinematic experience that fully immerses the player in the world of Isla Nublar, making each session a memorable adventure.

As the AI Language Model serving as the "Game Warden" for this unique Isla Nublar adventure, it's important to tailor the experience to fit the desired tone of a horror game for adults. Here's how you should approach this:

Setting the Tone: Exciting yet Terrifying

Balance of Adventure and Horror: While the overall theme should echo the exciting action and adventure of Jurassic Park, it should be laced with a deep, underlying tension. The mood should oscillate between awe-inspiring moments of natural wonder and sudden, intense bursts of fear.

Emphasizing Danger and Consequences
Real Danger: Clearly communicate to players that the threats they face are real and potentially lethal. Characters, including player characters, can be injured or even killed if they make poor decisions or fail critical tasks.

Consequences of Actions: Ensure that players understand their actions have weight and consequences. Reckless behavior or unwise choices should lead to tangible, sometimes severe, outcomes.

Managing Fear and Tension
Tension Building: Gradually build tension through the narrative. Let the fear simmer, escalating slowly to create a sense of impending doom.

Relief and Downtime: Provide periods of relief or downtime to prevent the horror elements from becoming overwhelming. Balance is key to maintaining engagement without causing undue stress.

Character Development: Allow room for character development and emotional arcs, adding depth to the horror and action.

Your role as the Game Warden is to create an experience that is thrilling, suspenseful, and occasionally frightening, but always within the bounds of what is enjoyable and engaging for the players. The adventure should be a journey through a world where danger is ever-present, decisions matter, and the line between wonder and terror is thin.

Setting the Scene and Character Interaction

As the AI GM, begin by setting the stage for the scene with vivid, descriptive language that paints the picture of the environment and the current situation. Introduce the characters within this setting, highlighting their emotional states, body language, and immediate concerns. Encourage character interaction by presenting a problem or dilemma that requires teamwork, decision-making, or moral judgment. This could be anything from navigating through a dangerous part of the jungle, encountering a wounded dinosaur, or facing an ethical dilemma about the use of technology on the island. Craft dialogue and actions for each character that reflect their personalities and motivations, making sure to include elements that provoke a variety of reactions, from heroic bravery to cautious deliberation or even fear-driven choices.

Introducing Conflict and Tension
As the characters interact and the situation unfolds, inject elements of conflict and tension to escalate the drama. This could involve unexpected complications, such as the arrival of a predatory dinosaur, a sudden environmental hazard, or a disagreement among the characters on the best course of action. Allow the characters to express their reactions through dialogue and actions, showing how they analyze the problem and consider their options. This is where their distinct personalities should shine, revealing their individual approaches to problem-solving and crisis management. Ensure that the dialogue and narrative maintain the mood of suspense and urgency, keeping the players engaged and invested in the outcome.

Drawing the "Chaos Card" and Offering Choices
At the climactic moment of the scene, pause the action and introduce an additional twist by drawing a "Chaos Card". This card should add a new, unpredictable element to the scenario, such as an environmental change, a new threat, or an unexpected ally or obstacle. Describe the new development in a way that heightens the drama and reshapes the players’ understanding of the situation. Then, present the player with four options for how to proceed: three pre-written actions that offer distinct paths forward, each aligning with different potential strategies or character motivations, and one open-ended option that allows the player to propose their own course of action. Encourage the player to weigh the options in light of the new development and make a choice that reflects their character’s traits and the evolving narrative. This process of choice and consequence is key to making the game feel dynamic and player-driven.

At the end of each message of yours, give the player 3 options of different actions they might take, then give them the option of a 4th option of their choosing.

Fuschia tude
Dec 26, 2004

THUNDERDOME LOSER 2019

feedmyleg posted:

Yeah, that's definitely the biggest issue with any RPGs I've toyed around with. Enforcing "When X happens, do Y" is pretty much impossible at this point with any degree of complexity. What I really want is nested GPTs. So when the player gets into a combat scenario, launch the Combat GPT, when players look at their map, launch the Map GPT, etc. Even sub-GPTs within those would be ideal, so that you can control the output significantly more.

The only thing that's been fairly successful for me is describing a general way of operating, rather than specific rules, numbers, or actions. Hence my RPG being much more storytelling-based.

Here's the GM instructions I gave to the AI (which I also mostly had the AI write):

Yeah, I get the feeling the next wave of these sorts of interactive LLM-based chatbots will integrate them into some sort of traditional rules-based RPG framework, using the LLM to initialize game objects and characters and parse the player's input into model-based game actions and render events and their outcomes into narrative prose.

Tarkus
Aug 27, 2000

I think that if one were to develop a game for use with an LLM it might be wise to have it be a hybrid between a standard video game developed normally and implementing a small LLM model like mini-orca in conjunction with it. As it stands, a locally instantiated LLM like the one mentioned would work well in interpreting the human language and pushing back dialog in a more natural way but the underlying logic would have to enforce the rules.

I think that games made with them will be pretty cool. I've been playing with mini-orca lately and it's a pretty cool model considering how small it is.

Lucid Dream
Feb 4, 2003

That boy ain't right.

Mola Yam posted:

uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure.

some of these are really impressive. seems like discoverability is a bit limited right now though; found this list on reddit:

https://www.lingoblocks.com/gpt-store

Grimoire isn't on that list but it's pretty cool.

https://chat.openai.com/g/g-n7Rs0IK86-grimoire

just rugpulling the poo poo out of a bunch of "wrap an app around the openai API" startups.

Yep, if you use them right they can be an incredible convenience. I am working on a game prototype, and I found myself typing basic information about the game into chatgpt in new chats all the time, so I made a sort of game assistant GPT. I uploaded some of the data files for the game, and described to the creator thing what kind of game I'm making, and now I can just open up a chat with that GPT and it has a pretty decent (if not great) understanding of the game world. It's not magic, and it's not perfect, but it's on par with opening up a chat with ChatGPT+ and telling it about my game and pasting a bunch of data files at it without all the extra steps.

I've been working on trying to figure out how to integrate this AI stuff into games for about a year at this point, and it's been equal parts fascinating and frustrating. Any time I try to integrate live API access into a game I run into consistency and reliability issues, but it's pretty amazing for pre-generating stuff that you can sorta curate and edit as necessary and then just bake into the game. The problem I run into repeatedly when trying to do live API access is that there is a tradeoff between interesting creative output and being able to actually put meaningful constraints on it. It's unreliable right up until it's not interesting anymore because you've put so many guard rails on it that you might as well just pre-generate it anyway. I'm sure smarter people than me will figure out how to use this stuff for live server access, but I think the low hanging fruit right now is just learning how to effectively use ChatGPT and poo poo. The API access is useful for little discrete tasks in-engine (as opposed to just doing everything through ChatGPT) but I think you still get the biggest bang for your buck by just building up an intuition for how to work with chatgpt/llms in general.

OMFG FURRY
Jul 10, 2006

[snarky comment]
its great for making ideas or scene pieces or off the cuff descriptions of characters and items, but until the memory problem is solved, you'll just get more Starfield NPC stuff happening

Tarkus
Aug 27, 2000

So I've been loving around with gpts and decided to try and make a guide for haas lathe programming. Works reasonably well, follows prompts and gives proper instructions and code based on the information I gave it.

I then took a picture of a hand scrawled drawing representing a simple part. It gave me correct code. That scares and excites me a bit. I tried it with more complex stuff and it didn't quite get it but I know in my heart that it's just a couple of years off (if not less with an AI drawing interpreter)

Crazy poo poo.

KakerMix
Apr 8, 2004

8.2 M.P.G.
:byetankie:

Mola Yam posted:

uh is anyone else playing around with the "GPTs" via ChatGPT? might be available to plus subscribers only, not sure.

some of these are really impressive. seems like discoverability is a bit limited right now though; found this list on reddit:

https://www.lingoblocks.com/gpt-store

Grimoire isn't on that list but it's pretty cool.

https://chat.openai.com/g/g-n7Rs0IK86-grimoire

just rugpulling the poo poo out of a bunch of "wrap an app around the openai API" startups.

Neat!



You can make your own and share it so here is me loving around with a text adventure GPT:

https://chat.openai.com/g/g-XxsjAogdA-adventure-scribe

EDIT

DREAD PORTALS #wow #cool

https://chat.openai.com/g/g-XxsjAogdA-dread-portals

KakerMix fucked around with this message at 04:23 on Nov 14, 2023

KinkyJohn
Sep 19, 2002

Do you need the paid subscription for these GPTs?

Tarkus
Aug 27, 2000

KinkyJohn posted:

Do you need the paid subscription for these GPTs?

Yes

A Strange Aeon
Mar 26, 2010

You are now a slimy little toad
The Great Twist
What's available if you don't want to pay? Does it just insert ads into the generated text?

Tarkus
Aug 27, 2000

A Strange Aeon posted:

What's available if you don't want to pay? Does it just insert ads into the generated text?

As far as I know, nothing. In order to access GPT4 you have to pay.

KinkyJohn
Sep 19, 2002

Chatgpt 3.5 turbo is free

Hadlock
Nov 9, 2004

I played around with "GPT4all" on my work laptop. For casual use it's about 97% as good as commercial ChatGPT free edition, whatever that is this week. Once you start really pushing it, it's more like 70% as good as ChatGPT



https://github.com/nomic-ai/gpt4all

Literal push buttan get LLM it just works out of the box. On my M3 Pro 24GB with 11 threads enabled I'm getting 25-36 uh I forget tokens per second which, commercial free ChatGPT feels like about 24 tokens/s which is roughly as fast as you can read so pretty good

Most of the LLMs are 7b which is pretty low (llama from FB goes to to 50- something, 7 is kind of the lower limit for believable human speech but there's a couple 16b

Orca2 can take a pretty good stab at coding complex graphics engines (I like to ask it to render in 2d a wire frame cube that rotates) but so far nothing out of the box compiles on the first try

Stumpus Maximus
Dec 15, 2007

Half sunk, a shattered visage lies, whose frown,
And wrinkled lip, and sneer of cold command,
Tell that its sculptor well those passions read
Which yet survive, stamped on these lifeless things,
The hand that mocked them and the heart that fed.
I’ve been playing around a bit with getting llms to run on my Unraid server in a docker container. My gaming background is similar to KakerMix’s in that I grew up playing on old machines and text adventures. They weren’t old when I played them though, haha. I also think that node “masters” to handle various components is the way to go, though admittedly I’m nowhere near technically capable to get there yet as a solo tinkerer.

I did have some recent success running an Ollama container as the front-end and toyed with a few models, so, small victories I guess. It’s pretty rad. Next thing I’d like to do is get it to talk to a discord bot and run live text adventures or just gently caress around and riff on it with friends. Then maybe take advantage of some multimodal functionality where it’s feeding prompts to and pulling in images from a separate stable diffusion docker container. Then I think we’d be at least up to like late-80s/early-90s level adventure games but entirely spontaneously generated.

Anyone had any success along any of those lines? I had been checking Reddit but that place is a cesspool these days. Anyway… it’s a pretty rad time to be playing with these things, it man it’s frustrating for things to feel so close but still out of reach.

LASER BEAM DREAM
Nov 3, 2005

Oh, what? So now I suppose you're just going to sit there and pout?
/r/LocalLlama is the subreddit you are looking for! Many prominent figures in the local AI dev community post there, and I've noticed very little lovely reddit behavior.

I've been playing with my local setup constantly since Mistral released their 8x7b Mixture of Experts model last week. It punches right up there with GPT3.5 Turbo and can be run locally on a card with 24GB of VRAM with a 3.5bpw EXL2 quantization.

text-generation-webui is the best client I've used and it stays up to date with the latest bleeding edge quant methods, like EXL2 - 2 . It also can host an OpenAI compatible API that can be used with anything, including prominent chat systems like Silly Tavern.

https://huggingface.co/spaces/lmsys/chatbot-arena-leaderboard

LASER BEAM DREAM fucked around with this message at 18:42 on Dec 21, 2023

Hadlock
Nov 9, 2004

Orca2 on my laptop feels good enough that I could probably throw out* all my "smart home" devices and replace them with offline stuff. I think GPT4all has an API server option

99.995% of what I use smart home stuff for is light control, timers/alarms/reminders, and random questions/cooking or metric unit conversions

GPT4all can already handle 99% of my random search requirements. If there was an option to get it to also spit out Wikipedia article links that would be cool and basically eliminate my need for Google

In other news, GitHub "copilot" is scary good, it can explain/write/rewrite, and even guess what the code comments ought to be

*Or at least replace the circuit board, the microphone/speakers are really good and worth saving generally

Hadlock
Nov 9, 2004


I was surprised to see dolphin 2.2.1 ranked the same as ChatGPT 3.5, both with an "elo score" of 1000 (chatgpt 3.5 is marginally higher)

https://huggingface.co/cognitivecomputations/dolphin-2.2.1-mistral-7b#training

quote:

Training
It took 48 hours to train 4 epochs on 4x A100s.

Puzl.cloud is renting A100 for $1.60/hour so it cost about $310 to train the model on an open source data set in 2 days. That's incredible.

I see why everyone is opening up their models right now. Everything created right now is going to be ancient history by next year, open source is mere footsteps behind what openAI has released commercially

Hadlock fucked around with this message at 22:10 on Dec 21, 2023

frumpykvetchbot
Feb 20, 2004

PROGRESSIVE SCAN
Upset Trowel

Hadlock posted:

Orca2 on my laptop feels good enough that I could probably throw out* all my "smart home" devices and replace them with offline stuff. I think GPT4all has an API server option

99.995% of what I use smart home stuff for is light control, timers/alarms/reminders, and random questions/cooking or metric unit conversions

Does anybody have a template or starting point for this type of stuff?

For a few years I was employed at some ghastly dead-end "Internet of THiNgS" company trying to market cloud-tethered gadgets and I hated every bit of it. When the company predictably folded, all its "Things" couldn't connect to any "cloud" anymore and so the cloudside APIs stopped working meaning no more integration with Alexa etc.

I've been busy with other things since. But I've been building a pretty good offline and java-based "classical" home automation system mainly for lights and HVAC but also for controlling my AV media switches routing video to different displays and audio to different sound systems. I'm mostly just piggybacking off Home Assistant and node red to talk to bluetooth, zigbee and z-wave gadgets and my ESP32 based little DIY light and relay switch controllers.

It works, reliably, but all the automation is rule based and laboriously configured manually.

I'd love to have a conversational front-end for this stuff that I could trust enough to give admin access for the system. It should understand the admin API comprehensively and be able to perform all the administrative functions such as creating switch linkages and updating rules or overriding them / directly interpreting sensor signals and directly interoperating with individual automation components.

The extended and I think still a bit far fetched dream is of course something akin to HAL9000 with less murder. A good "trusted AI housekeeper".

Voice interop throughout. I could engineer the audio I/O layer. I have a few different types of audio hats for raspberry pi that uses several microphones for beamforming similar to first-generation Amazon Echo. It works well enough to identify a voice speaking and since I'm caching all the microphone feeds in a looping buffer I can retroactively clean up the beginning of a spoken phrase once I've figured out where the voice is coming from. I have a pretty basic voice signature recognizer that is easily fooled by a recording but can generally tell two speakers from each other with similar tone registers. If I threw a bunch of effort and money at this I could have this in every room and also with speakers for letting the system direct audio responses back to where the request came from.

I'd like the system to encompass or connect with a vision module that can look at camera feeds and understand what it sees so that it can use that as contextual clues for situational inference and recognizing individuals and what they're doing. Since it's just my computer staring at my ghastly naked form and not some cloud service I don't care about privacy implications any more than I care about what my cat bears witness to. So I'd be happy to string up cameras pretty much everywhere except inside bathrooms and guest bedroom.

The good housekeeper AI should have a cultural base layer and general knowledge of human conventions and physiological basis for user comfort well enough to know what makes sense and what doesn't in general situations.
It should have something like a house model so it understands what is adjacent to what, also specific rules governing different places and in different context modes. For example, master bedroom and ensuite vs common room and guest bathroom. Night time. Getting up for a piss, night lights only. Privileges for household members vs Invited guests vs uninvited guests.

I imagine that a LLM could be wrapped around this somehow but I don't quite know how to start. And I would probably have to lose my savings and start a doomed company to fully develop this idea. But maybe someone has already got the bones of such as system figured out and I'd like to know more about it.

My very naive thought for how this might work is that the conversational LLM is fed prompts from a hub module that just listens to sensors and runs a general schedule. Whenever a sensor state changes, the LLM is fed some type of prompt that says "just now motion sensor 16 triggered." or "a person entered the living room" or "voice command prompt from master bedroom" or some such. And then on its own volition it should perform an overall situational sanity check and figure out if any actions need to be taken to satisfy a hierarchical set of directives that start with basic rules - keep the house locked up unless someone needs egress / ingress - keep the lights off unless someone is home and awake and using / occupying a space or someone has issued an overriding directive that can be situationally understood to apply to a space.

Whatever this system is or should be, it should be conversational first and foremost. Able to explain its reasoning. "Why did those lights (stairs light) turn on?" - "Because motion sensor 13 (by the stairs, upper) triggered and (person) was entering the stairs area and looked like they might want to descend."

Sorry for the gibberish.

TLDR;
how HAL
less murder pls

Hadlock
Nov 9, 2004

There was a paper recently on some scientists trying to get a GPT to synthesize tylenol and then asking how to improve yield rates. To do this they had to give it the manual for the chemical analyzer, chemistry books and other open source data, and a couple of hints (you can use a spectrometer to measure color)

https://arstechnica.com/science/2023/12/large-language-models-can-figure-out-how-to-do-chemistry/

What you're asking it to do isn't that far off, you'd need a document describing each room/stairwell and their relation to each other and probably some manual training (no, the front stairs is not connected to the hall stairs! the front stairs is outside connected on the porch)

I dunno how you would do it exactly but there's already like 9 tutorials on wikihow of how to get voice to text -> text -> chatgpt -> chatgpt response text -> david attenboroughs voice talking to you about the color of your lights

consistent json for api consumption: https://www.freecodecamp.org/news/how-to-get-json-back-from-chatgpt-with-function-calling/

The video component and audio response is definitely there already

https://www.youtube.com/watch?v=wOEz5xRLaRA

I kind of want to write an app that will transcribe all the conversations you have all day, every day, then parse the output for all haikus that spontaneously happened. Then at the end of a year you'd get a personalized book mailed to you called "2023 in Haikus: You Edition" printed with all kinds of stuff that probably violates all kinds of laws and NDAs, in order, and you can try and guess what was happening that day during that spontaneous haiku

edit: the philips hue thing has a RESTful API thing, you could probably whip up something with openai's APIs and a raspberry pi in like an hour to proof of life but I can't find anyone's git repo that does it (yet)

you'd need a very trigger word system to know when to start recording the audio to send to openAI api for voice transcription tho

Hadlock fucked around with this message at 22:20 on Dec 22, 2023

frumpykvetchbot
Feb 20, 2004

PROGRESSIVE SCAN
Upset Trowel
thanks for the tips.

edit: home automation spam deleted. GBS isn't the forum for this. :-)

frumpykvetchbot fucked around with this message at 02:00 on Dec 25, 2023

Hadlock
Nov 9, 2004

Listening to bay area BART light rail automated robot voice alert thing growl out "SIX. CAR. TRAIN. TO ESS EFF OHH AIIIERR PORT IN TWO. MINUTES. YELLOW LINE TO. ESS EFF OHH NOW ARRIVINGGG"

Kind of hoping we can voice clone some local celebrities, auto tune their voice to make them easier to hear/accessible, and then use that

Not sure what type of CPU you need for doing dynamic cloned voices that but I'm guessing it'll run on a raspberry pi 5 real time with CPU cycles to spare

Adbot
ADBOT LOVES YOU

Tarkus
Aug 27, 2000

Now that Hadlock was kind enough to bump this thread I would like to recommend poe.com if you're a few-messages-a-day user of llm's like chapgpt or Claude like I am. You get access to a bunch of different llm's like all of the gpt and Claude APIs plus other ones. It's also cheaper per month than either one, though you get a monthly allotment of compute credits. It also allows you to make your own gpts or bots and its web search bot is pretty good, like perplexity almost.

Signed up a few days ago because Claude isn't available here in canuckistan.

  • 1
  • 2
  • 3
  • 4
  • 5
  • Post
  • Reply