its a function of paying their employees less for more work relatively speaking and extracting more profit from consumers through ads and enshitification in general
Comment on Microsoft insiders worry the company has become just 'IT for OpenAI'
justaderp@lemmy.world 4 months agoMonopolies don’t care about the user experience, only profit. The AI doesnt understand the former, only the latter. The continued degredation of the user experience is a likely indicator of an increase in revenue as function of successful application of AI.
umbrella@lemmy.ml 4 months ago
brianorca@lemmy.world 4 months ago
But that’s also a path for them to no longer be a monopoly, if the right competitor makes the right moves.
justaderp@lemmy.world 4 months ago
We’re living in a late stage capitalistic hellhole and you’re advocating faith in the free market.
What. The. Fuck.
rottingleaf@lemmy.zip 4 months ago
I don’t remember anything ever in history undermining faith in the free - from regulation, but not from jailing crooks, - market.
It’s not as if anything lefties claim to be that were free. And when one talks about what is needed to make it free, one can hear screeching of the “reeeeeee useful idiots for capitalism reeeeee you just want poor people to die reeeeee we should all vote for 8 hour work week and peace on Earth reeeeee what do you mean it’s not enough to vote reeeee” kind.
Even Ponzi schemes are usually about everyone being conscious it’s a scheme, but thinking they are very smart and will fool some other suckers, and those suckers think the same in turn. That is covered by the “jailing crooks” part.
And various cartels and trusts and such usually make government regulation their instrument. They benefit from it.
I mean, all this has been said and proven many times.
brianorca@lemmy.world 4 months ago
I’m saying it’s happened before. AOL. Palm. Yahoo. Blackberry. A company with an effective monopoly gets complacent and fails to serve their users. They get replaced.
thurstylark@lemm.ee 4 months ago
Do you possibly mean “The AI evangelists” or something similar?
Like, I could totally understand it in the “software will also include the biases of those who wrote it” kind of way (a la Amazon’s failed attempt at automating job candidate search). If the only incentive you’re given as a programmer is “make it make money”, then yeah, your AI is going to bias towards that end.
Just couldn’t tell on first reading
justaderp@lemmy.world 4 months ago
I’m not actually asking for good faith answers to these questions. Asking seems the best way to illustrate the concept.
Does the programmer fully control the extents of human meaning as the computation progresses, or is the value in leveraging ignorance of what the software will choose?
Shall we replace our judges with an AI?
Does the software understand the human meaning in what it does?
The problem with the majority of the AI projects I’ve seen (in rejecting many offers) is that the stakeholders believe they’ve significantly more influence over the human meaning of the results than exists in the quality and nature of the data they’ve access to. A scope of data limits a resultant scope of information, which limits a scope of meaning. Stakeholders want to break the rules with “AI voodoo”. Then, someone comes along and sells the suckers their snake oil.
justaderp@lemmy.world 4 months ago
MutilationWave@lemmy.world 4 months ago
Oooh la la