Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
"ChatGPT sucks at something it wasn't trained to do"
🙄
This is a fucking terrible study.
They compare their results to a general diagnostic evaluation of GPT-4 which scored better and discuss it as relating to the fact it's a pediatric focus.
While largely glossing over the fact they are using GPT-3.5 instead.
GPT-3.5 sucks for any critical reasoning tasks, and this is a pretty worthless study not using the SotA or using best practices in prompting to actually reflect what a production grade deployment of a LLM for pediatric diagnostics would be.
And we really need to stop just spamming upvotes for stuff with little actual worth just because it's a negative headline about AI and that's all the jazz these days.
Why don't we stop acting like ordering words correctly can 100% replace any profesional?
Can it be used as a tool for the professionals? Hell yes. Fear of losing jobs is hindering this discussion. These LLM models are tools, which can make people more efficient and make less mistakes.
WebMD over here excited they aren’t the worse at web diagnosis anymore.
You know as someone who lives in the UK our NHS( national health service, which is basically social health care) already has a website to help you figure out if you need to see a doctor( the 111 site), and it's kinda useless. There are some things humans are simply better at, and understanding a humans physical needs is one of them.
I really think trying to replace doctors with AI is an awful idea.
I'm fine with it being used as another tool to help with the process, but that doesn't seem to be the goal of this.
The NHS website is fantastic. It's one of the best resources for getting good quality medical advice (if your not a medical professional). It ties symptoms to causes very well and provides information on the appropriate service you need if you have certain symptoms.
It's not a substitute for doctors. It a means to get people to go to the correct service depending on their immediate need. I have used it to get family members to go to a doctor where they otherwise wouldn't. It can help you be informed of any issues you are having, so you can see the possible treatment options. It tells you when a pharmacist can solve the issue rather than take time off work to go to a doctor's appointment. It also tells you when to call 999 rather than wait for a GPs appointment.
I suspect your not actually reading through the articles or have some comprehension issue. It's a fantastic tool, that is extremely useful. It's particularly useful because it's created by informed humans, not AI. It's also one of the few medical resources that is trying to sell stuff to you.
It's not about replacing. It's about supplementing.
Well of course its fucking chat GPT. I mean what did they expect? Are they doing like aperture science, throwing shitty experiments until something comes out eventually? Look at me, today I'm gonna try to see if my table is good to send SMS...
This is the best summary I could come up with:
While the chatty AI bot has previously underwhelmed with its attempts to diagnose challenging medical cases—with an accuracy rate of 39 percent in an analysis last year—a study out this week in JAMA Pediatrics suggests the fourth version of the large language model is especially bad with kids.
The medical field has generally been an early adopter of AI-powered technologies, resulting in some notable failures, such as creating algorithmic racial bias, as well as successes, such as automating administrative tasks and helping to interpret chest scans and retinal images.
But AI's potential for problem-solving has raised considerable interest in developing it into a helpful tool for complex diagnostics—no eccentric, prickly, pill-popping medical genius required.
For ChatGPT's test, the researchers pasted the relevant text of the medical cases into the prompt, and then two qualified physician-researchers scored the AI-generated answers as correct, incorrect, or "did not fully capture the diagnosis."
Though the chatbot struggled in this test, the researchers suggest it could improve by being specifically and selectively trained on accurate and trustworthy medical literature—not stuff on the Internet, which can include inaccurate information and misinformation.
"This presents an opportunity for researchers to investigate if specific medical data training and tuning can improve the diagnostic accuracy of LLM-based chatbots," the authors conclude.
The original article contains 721 words, the summary contains 211 words. Saved 71%. I'm a bot and I'm open source!