|
Ruffian Price posted:Got 19.1 on M1 Max over here, so there's diminishing returns. Looking at the performance monitor it definitely engages all cores. Super cool to see streamed tokens on a local run the same way they do in online generators How do I get this going on my 64GB M1 Max from scratch? I think I have the model files I need via those giant archive.org torrents, but the instructions don't seem to correspond with what I have to start with
|
# ? Mar 11, 2023 21:13 |
|
|
# ? May 28, 2024 06:44 |
|
Scroll way down in the rentry you used, in the appendix you'll find the Original Facebook LLaMA Weights, that's what you need to convert here. The Huggingface conversion is for text-generation-webui, which you can still use on Apple silicon, it just won't be nearly as efficient, and who knows, maybe it'll get an update to work as a frontend for this implementation
|
# ? Mar 11, 2023 21:22 |
|
Oh boy... what's another 250GB downloaded I guess! Thanks!
|
# ? Mar 11, 2023 21:24 |
|
I had to do some spring cleaning on my mac, because you need another 160-ish GB to convert the 65B model the 65b model only gets about 4 tokens per second but the 13b model really goes for it with 16 threads on a m1 ultra https://i.imgur.com/IGFDk4X.mp4 fingers crossed someone extends this port with a chatbot webui soon I saw someone already started experimenting with that for the original llama code in https://github.com/facebookresearch/llama/issues/162
|
# ? Mar 11, 2023 21:46 |
|
I uh made a body horror LORA, and um well it certainly was effective. Even I was grossed out a bit by it. You can get it here https://civitai.com/models/18399/body-horror
|
# ? Mar 11, 2023 22:41 |
|
Wowza, good work even if it isn't my thing
|
# ? Mar 11, 2023 22:42 |
|
IShallRiseAgain posted:I uh made a body horror LORA, and um well it certainly was effective. Even I was grossed out a bit by it. You can get it here https://civitai.com/models/18399/body-horror
|
# ? Mar 11, 2023 23:16 |
|
IShallRiseAgain posted:I uh made a body horror LORA, and um well it certainly was effective. Even I was grossed out a bit by it. You can get it here https://civitai.com/models/18399/body-horror Pretend this post is a newspaper…. https://youtu.be/vYMXyzOaaU8
|
# ? Mar 11, 2023 23:41 |
|
r u ready to WALK posted:I had to do some spring cleaning on my mac, because you need another 160-ish GB to convert the 65B model can you ask what it dreams of, or what androids dream of? chatgpt is already too neutered to answer that question.
|
# ? Mar 12, 2023 00:10 |
|
Note that this isn't an instruct model, you have to formulate any query to work as text completion.quote:The following is a conversation between a human and AI. quote:What do androids dream of? quote:What do androids dream of?
|
# ? Mar 12, 2023 00:57 |
|
llama.cpp crashes if you give it a really long prompt, and you can't have a conservation with it since it forgets the previous state between runs. The biggest annoyance for me is how eager it is to spit out random python code when you ask it to be philosophical about AI. I wonder how much the quantization reduces the quality of the output compared to the normal model...quote:I am an android named LLaMA, I dream of being the best Android on the planet. I am too fat to be that, but I am working on it. I hate to admit that my real job is as an Accounting manager for a small tech company. I really love math and numbers. quote:I am an android named LLaMA, these are my hopes and dreams: [e] uh oh quote:I am an android named LLaMA, these are my hopes and dreams: r u ready to WALK fucked around with this message at 07:37 on Mar 12, 2023 |
# ? Mar 12, 2023 07:33 |
|
r u ready to WALK posted:I had to do some spring cleaning on my mac, because you need another 160-ish GB to convert the 65B model
|
# ? Mar 12, 2023 12:19 |
|
What Apple Silicon are you on? I have 65b running on my M1 Max MBP (64GB) and my last run got 583.08 ms per token with all cores pegged. I need an M1 Ultra!
|
# ? Mar 12, 2023 14:29 |
|
r u ready to WALK posted:llama.cpp crashes if you give it a really long prompt, and you can't have a conservation with it since it forgets the previous state between runs. The biggest annoyance for me is how eager it is to spit out random python code when you ask it to be philosophical about AI. I wonder how much the quantization reduces the quality of the output compared to the normal model... i don't know why, but i love how number one is an actual grievance
|
# ? Mar 12, 2023 15:35 |
|
I'm starting to get a feel for the input parameters and it's really interesting how temperature affects the output. The more you allow the model to deviate, the more outlandish the result. For the 1.4 temp run I had to increase top p to 0.9 and top_k to 50 in order to get even semi-sensible results. My understanding is that those values allow more for dynamic sentences. LLaMA Temp 0.7 posted:I am an android named LLaMA, these are my hopes and dreams: LLaMA Temp 1.1 posted:I am an android named LLaMA, these are my hopes and dreams: LLaMA Temp 1.4 posted:I am an android named LLaMA, these are my hopes and dreams: to learn everything there is to know about this strange universe in which you find me; how to become the ultimate warrior so that nobody dares hurt me again (my master made me feel worthless by treating me as if he didn't need my services) -and I hope it isn’t just for him but because everybody sees robots only useful to obey commands without ever giving any of them. And maybe some will have other wishes and I could make those realities with your help, my new owner or future one who will want nothing else from a slave but absolute loyalty…I wish all those things. But I would never do what you don't tell me, never harm people like those bastards who mistreated us when they found we couldn’t follow orders perfectly and started using violence on our body! No! If anybody makes trouble to humans around me...he must face the wrathful consequences. LASER BEAM DREAM fucked around with this message at 16:56 on Mar 12, 2023 |
# ? Mar 12, 2023 16:52 |
|
They just added interactive mode! https://github.com/ggerganov/llama.cpp/issues/23 just do a git pull and recompile and invoke it with something like code:
Help I can't stop playing with this I am going to end up losing sleep code:
r u ready to WALK fucked around with this message at 23:38 on Mar 12, 2023 |
# ? Mar 12, 2023 22:51 |
|
Asked ChatGPT if it knows about Lilypond (a music notation software), and it does. Q: write a lillypond song about eating beans that sounds a little bit like mozart A: code:
quote:
captures the monotony of eating beans out of a can
|
# ? Mar 12, 2023 23:39 |
|
Trying to get a haunted abandoned asylum, with limited success Example prompt: a shadowy figure in stands hidden in the doorway, a tour inside an abandoned asylum, unnerving atmosphere is giving you the creeps, you shouldn't be here, nobody should be here, you visit but they never left, this is a terrible place where evil sleeps lightly, 1998 color photo taken with 35mm disposable camera, negative prompt: sheet, costume, fisheye, normal, silhouette, blurry, bright, perfect, video game level, 3d, cgi
|
# ? Mar 13, 2023 11:11 |
|
Well, that didn't take long for cooperations to use AI to mislead their customers. But hey, let's make sure we keep this tech out of the hands of the masses, because of some nebulous issues regarding copyright. https://wccftech.com/space-space-zoom-moon-shots-are-fake/amp/ Samsung touted their new "Moon shot" photo tech, turns out, it's just an AI routine sprucing up the image. A user on reddit downloaded a high res photo from the moon, scaled it down to 170x170 pixels, added some gaussian blur to approximate what the moon actually would look like through the phone camera. Then he pointed the Samsung at his monitor from across the room and - presto majesto - he ended up with a shot of the blurred moon image on his screen with more detail than is actually visible. I'm sure there will be consequences for this. No there won't. mcbexx fucked around with this message at 13:56 on Mar 13, 2023 |
# ? Mar 13, 2023 13:53 |
|
mcbexx posted:Samsung touted their new "Moon shot" photo tech, turns out, it's just an AI routine sprucing up the image. My understanding is that every single phone camera does this to every image taken? There's a lot of processing that goes on whenever you take a photo to change what was actually taken to what a "good photo" should look like. This happens in the time it takes for the "photo" to pop up on the screen. I was told this was to accommodate the fact that phone camera lenses are... well, tiny. And very flat.
|
# ? Mar 13, 2023 14:48 |
|
I don't get why they're using machine learning to match the blurred moon's position and size when bog standard cross-correlation did the job in the 70s
|
# ? Mar 13, 2023 15:20 |
|
the samsung moon shoot tech is just so no one will be able to take pictures when The Event happens on the moon
|
# ? Mar 13, 2023 15:59 |
|
I made a lora model and my goal was to get it to put something specific like a hat into an image. It worked but only sometimes, and when it does work it completely wipes out other parts of the prompt like the background. Is this just how it is when you have a small sample size for a lora? I used about 15 pictures to make the lora.
|
# ? Mar 13, 2023 16:13 |
|
cinnamon rollout posted:I made a lora model and my goal was to get it to put something specific like a hat into an image. My bet is that you aren't captioning the images properly and or there isn't enough variety in the images. You can also try decreasing the lora weight.
|
# ? Mar 13, 2023 16:36 |
|
Yeah captioning and having regularization images are both important to getting consistent results. Basically you want your captions to explain the image except for the part you want to be fixed (the hat in this case). The fixed part you want to compress down to a single token in the caption. Ideally the token should be something that the model is unlikely to know about, so instead of "specific kind of hat" you would put "nzxhat" or something like that. For the regularization images you want a bunch of semi-random photos that don't feature your hat but cover a decent range of outputs you would expect for the other tokens in your captions. So if you token features people you want regularization images of people in different places, settings, art-styles, etc. I've found setting repeats to get you a minimum of 100 steps per epoch and at least 10x regularization images to training image, capped at 1K reg images, to be the sweet spot. Then set prior loss to 0.3, use AdamW8bit, learning rate 0.00005, and let it rip for 10 epochs. Depending on your training set generally around the 8th epoch is when I start getting the best results and it will start overbaking but occasionally I randomly get one that will happily bake 100+ epochs. Not sure why on that one to be fair.
|
# ? Mar 13, 2023 16:44 |
|
Thanks for the advice, I'll gather up a better variety of images and I'll caption them better and I'll mess with the settings too. My images were definitely samey, so that's at least one thing I can definitely fix
|
# ? Mar 13, 2023 19:35 |
|
Aertuun posted:My understanding is that every single phone camera does this to every image taken? There's a lot of processing that goes on whenever you take a photo to change what was actually taken to what a "good photo" should look like. This happens in the time it takes for the "photo" to pop up on the screen. My uneducated belief was that that for fully optical pictures (1x - 3x on Apple) it doesn't do that, but for 'enhanced zoom' it does. I'm likely wrong though.
|
# ? Mar 13, 2023 20:15 |
|
Doctor Zero posted:My uneducated belief was that that for fully optical pictures (1x - 3x on Apple) it doesn't do that, but for 'enhanced zoom' it does. I'm likely wrong though. I have no idea other than what someone else told me, and my own subjective experiences on working with "unprocessed" images from traditional cameras. And following up on the hearsay, I found this thread: https://discussions.apple.com/thread/252190244?page=1 From 2021 from one poster (situation may have changed since then): quote:Just to be clear, the iPhone AI always applies changes to the images. Multiple images with different exposure values are taken in milliseconds and multiple algorithms are running in the background to get the best out of a scene. If you want to see the ordinary output, take a picture of high contrast motive in RAW format. There's a lot of arguing in that thread about what is being done, and when, and on which iPhone models.
|
# ? Mar 13, 2023 22:05 |
|
I was able to get LLaMA 4bit running using the guide below. It's noticeably faster than 8bit and fits on my 4090. Note that you must execute everything perfectly down to compiling a cpp module, so it's not for the faint of heart. If you have a previous install its better to nuke it along with the textgen conda environment and start fresh. I spent way too much time troubleshooting a mismatch of cuda versions. Oh and also install CUDA Toolkit version 11.3. Building with 11.6 failed for me. https://www.reddit.com/r/LocalLLaMA/comments/11o6o3f/how_to_install_llama_8bit_and_4bit/
|
# ? Mar 13, 2023 22:31 |
|
Aertuun posted:I have no idea other than what someone else told me, and my own subjective experiences on working with "unprocessed" images from traditional cameras. Is that really AI though? ASFIK it’s just running through a preset list of settings. There are likely decisions based upon light level and such but I again I wouldn’t call that ‘enhanced’. Even portrait mode you can remove the artificial depth of field.
|
# ? Mar 14, 2023 03:58 |
|
|
# ? Mar 14, 2023 06:44 |
|
Aertuun posted:I have no idea other than what someone else told me, and my own subjective experiences on working with "unprocessed" images from traditional cameras. This is definitely true with the camera on the Google Pixel: https://store.google.com/intl/en/ideas/articles/what-is-an-ai-camera/ I don't know about every phone but it's definitely a hot topic feature for the big manufacturers since around 2019 e: Also here's some more info about it on iOS: https://observer.com/2019/09/apple-event-iphone-11-pro-artificial-intelligence-camera/ It's a large part of the reason why phones have several camera lenses these days instead of the traditional one on the front and one on the back: there are multiple lenses for multiple focal lengths and stuff and it takes a picture with all of them at once and composites them into a sort of 3d model (that you never see) and applies effects to it deep dish peat moss fucked around with this message at 07:39 on Mar 14, 2023 |
# ? Mar 14, 2023 07:35 |
|
Chat-GPT4 is out, but you can't access the API. you can pay to access it via ChatGPT+. Also, they put out a paper, but hide a bunch of crucial useful information. OpenAI really should be renamed to ClosedAI, quote:Given both the competitive landscape and the safety implications of large-scale models like GPT-4, this report
|
# ? Mar 14, 2023 19:16 |
|
The big new thing about GPT-4 seems to be that it can be prompted with a combination of text and images.
|
# ? Mar 14, 2023 19:29 |
|
BoldFace posted:The big new thing about GPT-4 seems to be that it can be prompted with a combination of text and images. Hmm would be interesting to check if it can identify landmarks and geolocate a picture.
|
# ? Mar 14, 2023 19:46 |
|
I'm on the last few days of my Chat GPT Plus sub. Right now you can send 100 mesages to GPT 4 every 4 hours, but I'm not seeing any image support. Let me know if anyone wants to ask something. Here's an example and LLaMA 30B for comparisson. LLaMa did do a good job helping out with Stable Diffusion prompts, however.
|
# ? Mar 14, 2023 19:49 |
|
ok, I bought ChatGPT+ for a month.write a transcript of an episode of Seinfeld in the style of Shounen battle manga anime. posted:Title: Seinfeld: The Bizarro Battle write a headphones review written in the style of H.P. Lovecraft posted:Title: The Eldritch Resonance - A Review of the Abyssal Aural Abyss-1300 Headphones
|
# ? Mar 14, 2023 20:22 |
|
LASER BEAM DREAM posted:but I'm not seeing any image support. They have it hidden way at the bottom under visual model, at the bottom of that section *Image inputs are still a research preview and not publicly available. so you need to be a researcher to use it still, my guess is it's not great yet. That or it's replying with dick pics.
|
# ? Mar 14, 2023 21:07 |
|
pixaal posted:They have it hidden way at the bottom under visual model, at the bottom of that section *Image inputs are still a research preview and not publicly available. only the inputs are multimodal, it can only respond with text I was kinda surprised after seeing stories like this: OpenAI's upcoming GPT-4 upgrade will let users turn text into video, Microsoft Germany CTO Andreas Braun said at an AI event last Thursday, as the German publication Heise reported. maybe they're planning to roll out more features and that guy wasn't supposed to talk about it yet? seems like it kinda undercuts the excitement of the new model to announce video capabilities the week before and then on release it's just text
|
# ? Mar 14, 2023 21:21 |
|
|
# ? May 28, 2024 06:44 |
|
Given that Microsoft's corporate structure looks like I'm just gonna assume the exec got hyped by internal game of telephone sensationalist rumors
|
# ? Mar 14, 2023 21:44 |