Not_mikey
@Not_mikey@lemmy.world
- Comment on AI Generated Videos Just Changed Forever (OpenAI Sora) 8 months ago:
If anything this was worse under the old system. Making art previously costed a lot of money, you had to pay the artists for their time and money, and better artists cost more. So in the past that oil company could commission 100 top quality artists to make corporate propaganda while a person who cares for the environment but has no money could only make a drawing limited by their own personal technical artistic ability, which could be just stick figures.
This is why “high quality” consumerist and capitalist “art” and branding in the form of advertising is so abundant meanwhile anti-consumerist, anti-capitalist art is rarer, no one’s paying to get it made.
Now any cause, regardless of money, can create at least mid art to get there message across. Those causes can also have way more people behind them then an oil company can reasonably hire
It’s sort of like how the gun changed how power worked. Previously a king could use there resources to pay for a smaller army of well equipped highly trained knights to subjugate a group of people. Then when the gun came training and equipment didn’t matter nearly as much and it became more of a numbers game, and to get those numbers rulers needed to give more power to the masses in order to be able to marshall them for their cause. Those rulers who didn’t got overthrown in revolutions.
- Comment on AI Generated Videos Just Changed Forever (OpenAI Sora) 8 months ago:
Why would real meaning and messages be harder to find, does AI generated art inherently have less meaning?
Let’s say I wanted to convey the message that oil companies are destroying the environment so , throwing subtlety out the window, come up with an idea of “a vampiric oil baron draining mother nature of oil”, does the picture that is generated from me putting that prompt into an AI generator have any less meaning then if I actually drew it myself?
For all the advances in AI it still lacks intentionality, and always will under these current models, that has to be supplied by the person in the form of a prompt. I’d say that intention is the source of messages and meaning in art. AI just allows people without technical abilities in art to express those intentions, feelings and messages.
- Comment on Hyperloop One is reportedly shutting down 10 months ago:
Oh definitely, I don’t see it getting to SF before the bright line opens and I don’t see it getting to la in the next decade. But the bright line is using tried and tested technology and methods in a significantly less populated area, while hsr is building from scratch through the heartland of California.
It’ll take a long time but it will eventually get done, because there is still a will, not a strong one, to get it done. Most Californians recognize the immense value it will bring and will keep pushing for it.
- Comment on Hyperloop One is reportedly shutting down 10 months ago:
Oh definitely, I don’t see it getting to SF before the bright line opens and I don’t see it getting to la in the next decade. But the bright line is using tried and tested technology and methods, while hsr is building from scratch.
It’ll take a long time but it will eventually get done, because there is still a will, not a strong one, to get it done. Most Californians recognize the immense value it will bring and will keep pushing for it.
- Comment on Hyperloop One is reportedly shutting down 10 months ago:
Hsr isn’t dead, it’s still being actively built, employing thousands of people, still has broad public support and just got a shit ton of funding from the feds. It’s over budget and delayed yes but so was the original shinkansen because bootstrapping that kind of project and industrial knowledge is hard.
- Comment on If we're going to have an effective strategy against FB/Meta, we should clear up some misconceptions around defederation 11 months ago:
degeneration might be the ONLY way to prevent Faceboom
So we should just start posting furry porn to try and scare them off?
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
That study is like giving a written test to an illiterate adult, seeing them do worse than a child and saying they aren’t intelligent or innovative. Like I said earlier intelligence is multi-faceted, and chatgpt excels at rhetorical, conversational and other types of written intelligence. It does not, as that study shows, do well in spatial manipulation, that doesn’t mean it’s not intelligent. If you gave that same test to a paralyzed blind person with little to no concept of spatial reality they’d probably do just as bad. If you asked them to compose a short story or an essay they might be good at it because that’s where they’re capabilities lye. That short story could still be innovative in its composition and characters, and could be way better than anything a child wrote.
You have to measure different types of intelligence with different tests. If you asked chatgpt and a set of adults and children to write a short story about a wholey new subject chatgpt would beat most of the children and probably some of the adults.
And if that short story is about a new subject matter completey out of its training set what/who is it plagiarizing from? You could say it’s taking common tropes, themes and story elements from other stories, but that’s fundamentally what a lot of writing and culture is. If that’s plagiarism then you should be more worried about the marvel franchise as it’s a plagiarism machine that has way more cultural impact.
- Comment on A fair trade 11 months ago:
Screw ball golf, disc golf though solves all it’s problems
-
Can play in almost any environment so no habitat destruction needed. Might have to clear a few trees or brush in dense forest but otherwise mostly keeps forests intact
-
No elitism or arrogance. It’s one of the cheapest sports there is, just requires a couple $15 discs and most courses are free and are part of parks. Not much maintenance is needed on the course.
-
Easier to pick up. Most people can at least throw a disc 10 yards or so after a couple tries. Also more forgiving if drunk or high in that way.
-
More interesting to watch /play. Courses usually have obstacles like trees and the flight path of discs has a lot of lateral movement so if your good/lucky you can weave it through the obstacles.
-
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
It can answer questions as well as any person. Just because you may need to check with another source doesn’t mean it didn’t answer the question it just means you can’t fully trust it. If I ask someone who’s the fourth u.s. president and they say Jefferson they still answered the question, they just answered it wrong. You also don’t have to check with another source in the same way you do with asking a person a question, if it sounds right. If that person answered Madison and I faintly recall it and think it sounds right I will probably not check their answer and take it as fact.
For example I asked chatgpt for a chocolate chip cookie recipe once. I make cookies pretty often so would know if the recipe seemed off but the one it provided seemed good, I followed it and made some pretty good cookies. It answered the question correctly as shown by the cookies. You could argue it plagiarized but while the ingredients and steps were pretty close to some I found later none were a perfect match which is about as good as you can get with recipes which tend to converge in the same thing. The only real difference between most of them is the dumb story they give at the beginning which thankfully chatgpt doesn’t do.
The 7th grader and plagiarism comment make me think you haven’t played with them much or really tested them. I have had it write contracts, one of which I had reviewed by a lawyer who only had some small comments, as well as other letters and documents I needed for my mortgage and buying a home. All of these were looked over by proffesionals and none of them realized it was a bot. None of them were plagiarized too because the parameters I gave it and the output it created were way too unique to be in its training set.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
I’m not saying it doesn’t encode some of its training data, I’m saying it’s not just encoding its training data. It probably does “memorize” a bunch of trivial facts from its training data and regurgitate them when asked. I’m saying that’s not all they are and that’s not what makes the intelligent, their ability to also answer questions outside their training data is.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
Auto complete is not a lossy encoding of a database either, it’s a product of a dataset, just like you are a product of your experiences, but it is not wholly representative of that dataset.
A wind tunnel is not intelligent because it doesn’t answer questions or process knowledge/data it just creates data. A wind tunnel will not answer the question “is this aerodynamic” but you can observe a wind tunnel and use your intelligence to process that and answer the question.
Temperature and randomness don’t explain hallucinations, they are a product of inference. If you turned the temperature down to 0 and asked it the question " what happened in the great Christmas fire of 1934" it will give it’s best guess of what happened then even though that question is not in it’s dataset and it can’t look up the answer. The temperature would just mean that between runs it would consistently give the same story, the one that is most statistically probable, as opposed to another one that may be less probable but was pushed up due to randomness. Hallucinations are a product of inference, of taking something at face value then trying to explain it. People will do this too, if you tell someone a lie confidently then ask them about it they will use there intelligence to rationalize a story about what happened.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
If LLMs were just lossy encodings of their database they wouldn’t be able to answer any questions outside of there training set. They can though, and quite well as shown by the fact you can give it completely made up information that it can’t possibly have “seen” and it will go along with it and give plausible answers. That is where it’s intelligence lyes and what separates it from older chatbots like Siri that cannot infer and are bound by the database they pull from.
How do you explain the hallucinations if the llm is just a complex lookup engine? You can’t lookup something you’ve never seen.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
All inference is just statistical probability. Every answer you give outside of your direct experience is just you infering what might be the answer. Even things we hold as verifiable truth that we haven’t experienced is just a guess that the person who told it to us isn’t lying or has some sort of proof to there statement.
Take some piece of knowledge like “Biden won the 2020 election” me and you would probably agree this is the truth, but we can’t possibly “know” it’s the truth or connect it to some verifiable experience, we never counted every ballot or were at every polling station. We “know” it’s the truth because more people, and more respectable people, told us it was and our brain makes a statistical guess that their answer is right based on their weight. Just like an LLM other people will hallucinate or bullshit and come on the other side of that guess and assert the opposite and even make up stuff to go along with that story.
This in essence is what reasoning is, you weigh the possibilities of either side being correct, and pick the one that has more weight. That’s why science, an epistemological application of reason, is so heavily reliant on statistics…
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
This is not how LLMs work, they are not a database nor do they have access to one. They are a trained neural net with a set of weights on matrices that we don’t fully understand. We do know that it can’t possibly have all the information from its training set since the training sets (measured in tb or pb) are orders of magnitude bigger than the models (measured in gb). The llm itself is just what it learned from reading all the training data, just like how you don’t memorize every passage in a book you read, just core concepts, relationships and lessons. So if I ask you " who was gatsbys love interest?" You don’t remember the line and page of the text that says he loves Daisy, your brain just has a strong connection of neurons between Gatsby, Daisy , love, longing etc. that produces the response “Daisy”. The same is true in an LLM, it doesn’t have the whole of the great Gatsby in its model but it too would have a strong connection somewhere between Gatsby, Daisy, love etc. to answer the question.
What your thinking of are older chatbots like Siri or Google assistant which do have a set of preset responses mixed in with some information from a structured database.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
computer scientists, neurologists, and philosophers can’t answer that either, or else we’d already have the algorithms we’d need to build human equivalent A.I.
I think your mixing up sentience / consciousness with intelligence. What is consciousness doesn’t have a good answer right now and like you said philosophers, computer scientists and neurologist can’t come to a clear answer but most think llms aren’t conscious.
Intelligence on the other hand does have more concrete definitions that at least computer scientists use that usually revolve around the ability to solve diverse problems and answer questions outside of the entities original training set / database. Yes doing an SAT test with the answer key isn’t intelligent because that’s in your “database” and is just a matter of copying over the answers. LLMs don’t do this though, it doesn’t do a lookup of past SAT questions it’s seen and answer it, it uses some process of “reasoning” to do it. If you gave an LLM an SAT question that was not in it’s original training set it would probably still answer it correctly.
That isn’t to say that LLMs are the be all and end all of intelligence, there are different types of intelligence corresponding to the set of problems that intelligence is solving. A plant identification A.I. is intelligent for being able to identify various plants in different scenarios but it completely lacks any emotional, conversational intelligence, etc. The same can be said of a botanist who also may be able to identify plants but may lack some artistic intelligence to depict them. Intelligence comes in many forms.
Different tests can measure different forms of intelligence. The SAT measures a couple like reasoning, rhetoric, scientific etc. The turing test measures conversational intelligence , and the article you showed doesn’t seem to show a quote from him saying that it doesn’t measure intelligence, but turing would probably agree it doesn’t measure some sort of general intelligence, just one facet.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
This seems like circular reasoning. SAT scores don’t measure intelligence because llm can pass it which isn’t intelligent.
Why isn’t the llm intelligent?
Because it can only pass tests that don’t measure intelligence.
You still haven’t answered what intelligence is or what an a.i. would be. Without a definition you just fall into the trap of “A.I. is whatever computers cant do” which has been going on for a while:
Computers can do arithmetic but they can’t do calculus, that requires true intelligence.
Ok computers can do calculus, but they can’t beat someone in chess, that requires true intelligence.
Ok computers can beat us in chess, but they can’t form coherent sentences and ideas, that requires true intelligence.
Ok computers can form coherent sentences but …
It’s all just moving the goal post to try and preserve some exclusively human/organic claim to intelligence.
There is one goalpost that has stayed steady, the turing test, which llm seems to have passed, at least for shorter conversation.
- Comment on Threads is making moves for Mastodon integration 11 months ago:
then they’ll find some sort of technical excuse and pull the plug on ActivityPub support
How do they do this without running a foul of regulators? People are already mad at meta and want to break them up for having instagram and Facebook, if they add the last big social media platform every politician right and left will be lining up to take them down. There’s a reason they never bought twitter despite being able to 10x over. Combine that with new EU interoperability laws and there’s no way meta could get away with that.
- Comment on Threads is making moves for Mastodon integration 11 months ago:
The platform never really took off. It was a niche messaging platform before Facebook and Google and went back to being one after they left. I have yet to see any evidence that Google or Facebook helped or hurt xmpp, just speculation and anger that it didn’t take off.
- Comment on Threads is making moves for Mastodon integration 11 months ago:
To all the people wondering about metas intentions in this it’s not the big bad corporation taking down the upstart competition. All the people saying it’s EEE can’t show any sign metas doing this or even wants to because the strategy doesn’t work, any time a company does it it either doesn’t take off or they get brought up on anti-trust laws. Show me a standard that was destroyed by EEE and I’ll show you a standard that never took off in the first place. All the usual examples given, email, java, html, remain open standards to this day.
The truth is the fediverse isn’t competition to meta, it’s a fraction of the size and is populated by users who would never use meta services in the first place. They can pretend it’s a competitor though. If twitter does actually collapse and people switch to threads meta will face anti-trust suits for owning the three largest social media platforms. If they add activity pub support though they can point to the fediverse and say it’s competition, even if it’s only 1 % of the platform. They also have to deal with EU interoperability laws that might start getting enforced.
TL;DR this is about compliance for meta, not conquest.
- Comment on Mind-reading AI can translate brainwaves into written text: Using only a sensor-filled helmet combined with artificial intelligence, a team of scientists has announced they can turn a person’s thou... 11 months ago:
Calling llm “big auto-complete” is like calling people “big bacteria” . It’s true that they act on the same goal, guess the next word for llm and auto-complete; survive and reproduce for people and bacteria, but they are vastly different in scale and complexity.
Also what would AI be to you if not an llm? Cause I’d say anything that has an SAT score higher than most Americans has some form of intelligence.
- Comment on US sets policy to seize patents of government-funded drugs if price deemed too high 11 months ago:
No, the greediest companies take all the tax payer funding and research they can before patenting it and pretending they made it, see moderna. Why pay for your research when you can get the taxpayers to do it?
- Comment on Blade Runner director Ridley Scott calls AI a "technical hydrogen bomb" | "we are all completely f**ked" 11 months ago:
they’re a particularly beefed-up auto complete
Saying this is like saying your a particularly beefed-up bacteria. In both cases they operate on the same basic objective, survive and reproduce for you and the bacteria, guess the next word for llm and auto-complete, but the former is vastly more complex in the way it achieves those goals.
- Comment on Is jogging on sidewalks harder on joints than if one jogged on dirt? 1 year ago:
Those are mostly for sprinters and short distance runners as they are higher impact and use running cleats with less cushion then softer distance running shoes. They’re also more likely to fall in events like hurdles and it’s better to land on that then on concrete. Most of the shock that could be absorbed by those tracks could also be absorbed by good running shoes. Most events above a mile take place on the streets.
- Comment on Nepal decides to ban TikTok 1 year ago:
Yeah but to be fair it’s not like children, or even adults really, are flocking to this app.
- Comment on Microsoft tries to dump S.F. office space amid tech industry cost-cutting 1 year ago:
Which will collapse the residential market with a supply glut. Which in SF might be a good thing.
- Comment on Cities Skylines 2 reportedly runs with 7-12fps on an Intel Core i9 13900KS with AMD Radeon RX 7900XTX at 4K/High Settings 1 year ago:
If you can’t actually play it with the new graphics what’s it’s advantage over original cities skylines?
- Comment on Apple considered ditching Google for DuckDuckGo in Safari’s private mode | But Apple exec argued DuckDuckGo wasn't as private as believed. 1 year ago:
Where should it be based then? I feel like you can’t trust any state to not do this sort of thing to fight “terrorism” and “crime”
- Comment on Meta wants to charge EU users $14 a month if they don't agree to personalized ads on Facebook and Instagram 1 year ago:
I think the idea of social media dividing us ignores the scale of it. All those other examples you gave were very local, and in that environment a consensus can form about certain political or ideological views. Those views could be vastly different than those a similar sized community holds 100 miles away though. Social media and it’s global scale exposes those differences and makes consensus on any sort of issue impossible.
At the same time it also allows for minority solidarity outside of the traditional local community. For example there may only be 1 or 2 LGBT+ people in a town, which can easily be marginalized, shamed and ignored. But if they’re able to communicate across geographic boundaries they’re able to create a larger stronger community that is harder to ignore. It also does the same for nazis though.
- Comment on Spotify is going to clone podcasters’ voices — and translate them to other languages 1 year ago:
Interesting and scary to think ai understands the black box of human neurology more than we understand the black box of ai.
- Comment on Microsoft estimated Valve’s revenue in 2021 at $6.5bn Interesting to see another view on the scale of Valve’s business 1 year ago:
Seems about right. They don’t have a complete monopoly and the industry seems to be making most of their money off micro-transactions these days instead of sales, something valve takes no cut of outside there own games. Meanwhile apple and google do have monopolies and take a cut of micro-transactions which accounts for there significantly larger revenue.