eftychia: Me in kilt and poofy shirt, facing away, playing acoustic guitar behind head (cyhmn)
Add MemoryShare This Entry
posted by [personal profile] eftychia at 05:24am on 2022-10-19

"There's this idea I've seen presented by [twitter.com profile] mcclure111 of 'grey data', in the model of 'grey water' - that is, water that we can be 100% certain is not healthy for human consumption but is still nonetheless water and can be put to some uses like sewage and irrigation.

"There is stuff that has absolutely useful application to understanding systems and how they train things, or as useful low-end stand-in tools, like AI generative art, but which absolutely are not ethically acceptable for the majority of human-interfacing uses"

-- Talen Lee ([twitter.com profile] Talen_Lee), 2022-08-14

Searching for the phrase on Twitter, I found:

"what animals (including humans) have access to that machines lack is access to many sources of truth - humans can do arithmetic on our fingers, by counting out beans, purely in our heads, etc. AI esp ML etc don't 'know' when one of its axioms is faulty, its truth is too fragile.

"and i think despite the common defense that even when an AI is wrong about something, it's kinda right about some larger idea is largely irrelevant given how often it's being pressed into service where concrete truth is critical, eg if a car is about to run over an old lady."

-- JP ([twitter.com profile] vectorpoem), 2022-04-08

"I've been thinking lately about trying to formally develop the idea of 'gray data', named in the sense of graywater. Data that's ok for limited/frivolous uses, like aesthetics, but 'not fit for human consumption'. Anything that comes out of machine learning ('AI') is gray-data.

"The thing about machine learning is ML only works when there are no consequences for the ML model being wrong. Decisions should never be made based on ML output, and ML output should be quarantined, like graywater, to make sure it isn't fed into a decisionmaking process.

"Actually I'm not sure even THIS approach works since even tech applied for pure aesthetics can do harm (imagine an app that makes cute cartoon selfies, but because of a limited training set breaks on people of color). But if you're gonna use ML at all u have to outline its limits"

-- mcc ([twitter.com profile] mcclure111), 2022-04-09

Searching on Google, I found other references to the phrase, using it in at least two different ways, distinct from this but kinda overlapping if you look at them right, at least as far back as 2010:

  • vast piles of unstructured data archived by an organization but not structured in any especially useable way -- piles of email, automaticaly generated logs, etc.
  • un-peer-reviewed / unverified data that may include spam, disinformation, and garbage -- in contrast to verified and trustworthy "white literature", and mostly hidden internal analytics that are seldom examined directly but are used indirectly ("black data").

Of all these, I like the analogy-to-grey-water version best.

There are 5 comments on this entry. (Reply.)
extraarcha: US flag inverted - distress & alarm (Default)
posted by [personal profile] extraarcha at 03:24pm on 2022-10-19
Interesting!
The rush to ML/AI is, while predictable, going to have unintended consequences and demonstrably already has. One only need look at Tesla's "self-driving" cars. Their "accident" rate is climbing and ideas laid out in this post point directly to why.
Many more areas will show up as ML/AI invades more of our society. It's in use already in things like utility bills and answering phone calls. I've been refusing to be forced into line when i am 'answered' by an AI and keep repeating "representative" or similar until the AI gives up and connects me to a real, breathing person.
Skynet is very close ...

metahacker: A picture of my eye reflecting the camera taking the picture. I'm probably feeling introspective.  (eye)
posted by [personal profile] metahacker at 08:41pm on 2022-10-19
Long ago I likened the collection of private data to concentration of toxic waste; the grey-water analogy is awesome, capturing the idea that some uses are merited but others are poisonous...
chemicalcain: a dog with a knife. there is a red glare in its eyes (Default)
posted by [personal profile] chemicalcain at 12:58am on 2022-10-20

Oh, I really like this.

blueeowyn: (Default)
posted by [personal profile] blueeowyn at 09:04pm on 2022-10-23
"Actually I'm not sure even THIS approach works since even tech applied for pure aesthetics can do harm (imagine an app that makes cute cartoon selfies, but because of a limited training set breaks on people of color). But if you're gonna use ML at all u have to outline its limits"

You mean like has already happened? Along with a bunch of other bad results for people of color (in prticular but others as well) in the results of algorithms based on biased and/or incomplete data?
selki: (Default)
posted by [personal profile] selki at 12:50am on 2022-10-29
I've been uneasy about proliferating seemingly uncritical acceptance and promotion of ML art. Thanks for that graywater analogy.

Links

January

SunMonTueWedThuFriSat
1 2 3 4 5 6 7
8 9 10 11 12 13 14
15 16 17 18 19 20 21
22 23 24
 
25
 
26
 
27
 
28
 
29
 
30
 
31