‘Nothing is changing’ — Reddit is denying a report from The Washington Post that it might force users to log in to see content if it can’t reach deals with AI companies
Reddit is apparently seeking to strike deals with big AI companies.
‘Nothing is changing’ — Reddit is denying a report from The Washington Post that it might force users to log in to see content if it can’t reach deals with AI companies::Reddit initially denied a report from The Washington Post that it might force users to log in to see content. However, the Post says it may still block search crawlers, and Reddit didn’t deny to The Verge that it may do so.
people talk about how big AI is but, It'll crash like everything else as enshittification hits. I tried to use Bing AI the other day for the first time in a few months, it didn't even let me do more then a handful of entries before locking me out saying I used too many queries in 24h. How is that supposed to be helpful to a consumer as a valid feature of you lock it down.
Yeah, hasn't anyone else noticed that there hasn't been a single profitable product to come out of it? Even copilot is biting the dust already as they try to reduce computing costs. I also haven't heard of a single person actually paying for chatgpt access either...
There are a lot of companies using ChatGPT believe me. Plus some users are also paying for getting access to GPT4. I am also using it a lot. It helps but it is not exactly the silver bullet to everything and you should always check what the result is because sometimes it is complete garbage.
Lol. I will believe it when I see it. I don't think LLMs especially will do that much good in Healthcare, and I would be particulary wary of them diagnosing patients. Aside from some very limited signal analysis for telehealth, I am very wary on the applications of "new" AI on healthcare. I believe it will be a disaster.
8f your in enterprise software, this stuff is pretty malleable. You are regularly asked to give pitches and lectures on medical projects, for sales reasons. You would be surprised - most people that work on this stuff have no idea the fist thing about medicine.
My Mom's a doctor, so I can ask her to have a bit of insight about this stuff. The challenges facing healthcare don't have that much to do with technology, at least in the US.
We're largely still working with LLMs at the moment -- Using them to immediately pull in relevant clinical information from previous encounters when a doctor sees a patient. Or using generative AI to edit doctors' messages to patients be more empathetic and... human (our pilot organizations have really loved this one so far). Using procedure codes on claims to guess if certain diagnoses were missed and to make more robust health risk profiles for populations as a whole -- these are a bit more NLP/data mining.
I'm in all the business meetings. "Use Chatham to save time generating analysis" or something like that. I think it has been proven that merely using it as a tool to generate content isn't profitable- at some level even your paid subscription is subsidized by VC money. The real test is if it provides "valuable" content. But then why does your employer even need you to make the prompts? Don't worry, I believe LLMs are fundamentally incapable of this and that your job is safe.
Those companies learned their lesson from search engines. They gave it away for free for far too long and with too few strings attached. It became impossible to realistically gate features and charge for them.
But chatbots, on the other hand, just need a little big money razzle dazzle and, boom, now it is AI and people are conditioned to accept any limits thrown at them.
I get why some limits are necessary, a company doesn't want a Microsoft Tae repeat...but after Chat GPT was made available to the public, the rapid addition of a whole range of guardrails made it nearly immediately unusable.
You ask it about anything which is controversial in the slightest regard and it shuts down, which for me at least, removes any interest in using it.