this post was submitted on 25 Jan 2024
67 points (79.6% liked)
Technology
59963 readers
3481 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Which is so silly, because AI writing still needs a human editor. I write for a living and there tons of work that involves using AI as a tool to increase productivity rather than to replace writers completely... like photoshop didnt put photographers out of business it just changed the work flow.
I work in a clinical setting where some Doctors are trying an AI program for generating their clinical notes out of the casual conversation between them and the patient. It's way off its mark for what we demand in quality. It requires significant editing from the healthcare provider, and if the note is very robust it quickly becomes more of a chore than modern voice transcription. Our review is not great so far.
That's a terrible way to be using a LLM for generating clinical notes.
Sounds more like trying to use a screwdriver to hammer in screws than an issue with the screwdriver itself.
This is what I think about AI being forced into many things these days. Feels more like an attempt to justify subscription plans than anything actually productive.
In part this is because the SotA model is by far GPT-4, but OpenAI has pigeon holed it into 'chatbot.'
The earliest versions of it pre-release when it was being incorporated into Bing were amazing. Probably the most impressive thing I've seen in tech.
But it was too human-like and freaking users out, so rather than wait for the market to adjust they did extensive fine tuning to make the large language model trained to predict human ouput be less likely to produce human-like output.
The problem is that they don't have a scalpel for this sort of thing and ended up with a model that's very good as a chatbot within a certain scope, but significantly impaired at some of the outside the box mechanics visible early on.
And because it's the SotA, everyone is now using it to fine tune their own models.
So the entire industry is being set back in practical applications outside of "kind of boring chatbot."
Right. It seemed like a reach when I first heard of it, but that's how it's advertised and the Hospital was sold on at least trying it out.