|
Mata posted:Bumping this thread with more of a practical question about LLM rather than the scientific aspect of ML... Github Copilot for Business is the way you'd want to go for an out-of-the-box coding assistant that doesn't feed prompts or suggestions back into the model (assuming you trust Microsoft): https://docs.github.com/en/enterpri...usiness-collect It does require an enterprise account, so I'm not sure how that would work if you're contracting. Copilot isn't perfect, but I'm not sure if there's anything better out there right now. I was unimpressed by the demo I got of Databricks' assistant. Apparently StackOverflow has one now, too. Haven't seen it in action.
|
# ¿ Oct 24, 2023 03:51 |
|
|
# ¿ May 17, 2024 19:08 |
|
His sister made some allegations that resurfaced recently. If I'm being cynical, though, I don't know that enough people paid attention that the board would care.
|
# ¿ Nov 17, 2023 22:17 |
|
Keisari posted:Has anyone been able to really well fiddle with the custom GPTs and more specifically, the "knowledge" part of them? You can upload all kinds of poo poo to be their knowledgebase. I have tried to make one to help me build a program to use a certain API, and uploaded a bunch of JSON files that describe the API. Another I made was inspired by the built in board game explainer, I made one that was focused on explaining and clarifying the rules, and uploaded some game manuals. Did you give it an OpenAPI spec? And were you asking it to write code that queries an API or setting that up as an action for a custom GPT? I got good results with regular ChatGPT when uploading a basic JSON file describing a service I wanted it to create using FastAPI. I had much less success trying to set up an action for my own GPT. It gave an unhelpful error message and crashed each time. I think the root of the issue was that I gave it the full JSON file that I pulled from Swagger. I want to go back with a cut down file to try adding endpoints one at a time. That might help with finding the source of the error, but also I think it should perform better if I can limit extraneous context. If anyone has found things that work well for cleaning up knowledge inputs to GPTs, I'd love to hear about that. ...also this OpenAI drama makes me feel better about the dumb politics poo poo that happens at my company. BAD AT STUFF fucked around with this message at 20:48 on Nov 22, 2023 |
# ¿ Nov 22, 2023 20:45 |
|
Insurrectionist posted:- What kinds of network designs are favored for generative NNs? Is there a focus of width (nodes per layer) over depth (# of layers) or vise versa? Do they utilise input-reducing layers like CNNs (pooling etc)? GPT stands for Generative Pre-trained Transformer, and the "transformer" bit is what to look at if you want to understand how these new generative models are different. The paper that really kicked things off is called "Attention Is All You Need". I don't think I understand attention and transformers enough to do it justice, but that's a good starting point.
|
# ¿ Mar 6, 2024 04:38 |