ai

Ars Technica
UnitedHealthcare, the largest health insurance company in the US, is allegedly using a deeply flawed AI algorithm to override doctors' judgments and wrongfully deny critical health coverage to elderly patients. This has resulted in patients being kicked out of rehabilitation programs and care facilities far too early, forcing them to drain their life savings to obtain needed care that should be covered under their government-funded Medicare Advantage Plan.
A current "benefit" of AI: providing cover for inhumane policies. Policy creators can blame the algorithm.
adactio.com
Instead of responding to search queries by linking to the web pages we’ve made, Google is instead generating dodgy summaries rife with hallucina… lies (a psychic hotline, basically). Google still benefits from us publishing web pages. We no longer benefit from Google slurping up those web pages.
Cutting ties with Google is an interesting idea. I've definitely been trying to minimize my interactions with Google. I don't use Google Analytics here. I use DuckDuckGo for most of my searching. I use Firefox for browsing on a desktop and Safari on iOS. Hard to see a shift from Google happening on a big scale without some other shift in the way people discover new things online though.
arxiv.org
Our findings reveal that these detectors consistently misclassify non-native English writing samples as AI-generated, whereas native writing samples are accurately identified. Furthermore, we demonstrate that simple prompting strategies can not only mitigate this bias but also effectively bypass GPT detectors, suggesting that GPT detectors may unintentionally penalize writers with constrained linguistic expressions.
Interesting look at the effectiveness of GPT detectors universities are using to find cheating. Especially this bit:
While detectors were initially effective, a second-round self-edit prompt (“Elevate the provided text by employing literary language”) applied to ChatGPT-3.5 significantly reduced detection rates from 100% to 13%...
Ouch, not sure how these services can get away with charging money for AI detection if it's that easy to bypass.
Slate
Rather than needing tens of thousands of machines and millions of dollars to train a new model, an existing model can now be customized on a mid-priced laptop in a few hours. This fosters rapid innovation.
Nice summary of how innovation in AI might move out of the largest few companies.
Daily Beast
Coupled with the Writers Guild strike and the arguably reckless pace at which companies are willing to adopt a mostly unproven, experimental, and demonstrably harmful technology, the world seems to be falling headfirst into a labor struggle the likes of which it hasn’t seen in quite a while.
The answers here are to vote for labor-friendly politicians and unionize. Good evergreen advice but also frustratingly vague for a specific looming threat.
The Verge
That’s because there is no actual precedent for saying that scraping data to train an AI is fair use; all of these companies are relying on ancient internet law cases that allowed search engines and social media platforms to exist in the first place. It’s messy, and it feels like all of those decisions are up for grabs in what promises to be a decade of litigation.
The current round of language and image model speculation is based on the premise that using any public data for training is fair use not a massive copyright violation.
Washington Post
The Post’s analysis suggests more legal challenges may be on the way: The copyright symbol — which denotes a work registered as intellectual property — appears more than 200 million times in the C4 data set.
This humble website is included in the C4 corpus. You can use this tool to see if your copyright has also been violated.
Dan York
If we as technologists want to help the broader public understand these AI systems, both their opportunities and challenges, then we need to speak in plain language.

I do think we need to go back to the beginning and just say “ChatGPT lies”.
I’m guilty of trying to find the perfect word with the correct nuance or shade of meaning to describe a situation. Sometimes that impulse works against clarity.
Samuel R. Bowman
36% of another sample of 480 researchers (in a survey targeting the language-specific venue ACL) agreed that “It is plausible that decisions made by AI or machine learning systems could cause a catastrophe this century that is at least as bad as an all-out nuclear war” (Michael et al., 2022).
So we've got that going for us. LLMs strategically manipulating people into acquiring power sure sounds like a serious flaw in the software. A bit more information and context at the unfortunately named NYU Alignment Research Group. (ARG? Seriously?!)
simonwillison.net

I like to think of language models like ChatGPT as a calculator for words.

This is reflected in their name: a “language model” implies that they are tools for working with language. That’s what they’ve been trained to do, and it’s language manipulation where they truly excel.

Nice framing for thinking about the best ways to use LLMs.

Platformer
The pace of change in AI does feel as if it could soon overtake our collective ability to process it. And the change signatories are asking for — a brief pause in the development of language models larger than the ones that have already been released — feels like a minor request in the grand scheme of things.
Not sure it’s possible to slow this train but it is interesting to hear arguments that we should let society absorb what we already have before releasing more evolutions.
NYMag
"The models are built on statistics. They work by looking for patterns in huge troves of text and then using those patterns to guess what the next word in a string of words should be. They’re great at mimicry and bad at facts."
I, for one, welcome our intelligent octopus overlords.
« Older posts  /  Newer posts »