Im gonna watch harry potter but draco is macho man randy savage
Im gonna watch harry potter but draco is macho man randy savage
HEYY!!!
thanks for the real discussion and not memeing. i appreciate you. lemmy appreciates you.
Me. Moderate ai enthusiast and software engineer.
I think its the company’s responsibility to incorporate a technology to carry out their policy accurately. They cant just use an LLM stock from a vendor. They work to adapt it for their needs and get acceptable results. I think if an llm isnt considerably more accurate than humans then its a disservice to their customers and they should be responsible for that. There should be regulations to keep companies from using models if they dont work
I often use prompts that are simple and consistent with their results and then use additional prompts for more complicated requests. Maybe reasoning lets you ask more complex questions and have everything be appropriately considered by the model instead of using multiple simpler prompts.
Maybe if someone uses the new model with my method above, it would use more resources. Im not really sure. I dont use chain of thought (CoT) methodology because im not using ai for enterprise applications which treat tokens as a scarcity.
Was hoping to talk about it but i dont think im going to find that here.
Happy to hear about anything that supports the idea.
I agree that is a bit of an ethical minefield to employ it to make decisions that affect peoples livelihood. But my point is if a company uses it to decide if an insurance claim should be paid out, the models ability to make those decisions isnt changed by what we call the steps it takes to come to a decision.
If an insurance company can dissect any particular claim decision and agree with each step the model took, then is it really different than having someone do it? Might it be better in some ways? A real concern is the fact that ai isnt perfect and mistakes made are pretty hard to accept… seems pretty dystopian i get that. But if less mistakes are made and you can still appeal decisions then maybe its overblown?
Are you saying thats not true? Anything to substaniate your claim?
They say it uses roughly the same amount of computing resources.
Sorry but thats not an explanation of your position, thats restating what you just said.
Why does ai that has a “reasoning” step become dangerous?
100-250 per month
I appreciate the effortful response but i dont think regulators would get caught up on colloquial names when weighing benefit versus harm and deciding to do something like ban a model.
We just arent close enough to the same perspective to discuss it further. Thanks again for the good faith clarification.
Their work is making progress. What is irresponsible or dangerous? Im not understanding what you mean.
Show it.
It scores 83% on a qualifying exam for the international mathematics olympiad compared to the previous model’s 13% so…
Personally its the result that matters to me, and whether or not its entertaining regardless of how it was made.