korazail
@korazail@lemmy.myserv.one
- Comment on YSK TikTok Is Harming Children at an Industrial Scale. We know this because we obtained messages from TikTok engineers and executives 2 hours ago:
Did you sneak around and do things you were told not to? Probably.
While doing so, did you have the context that you shouldn’t do it? Maybe. Sometimes the learning happens when you get caught, get hurt or have other consequences.
Sure, the answer is education. Tell them that they shouldn’t do <thing> and why. Hopefully, the guilt/shame/pain of doing the thing they know is incorrect will be enough of a deterrent, but adults are fallible and kids cannot be expected to be better at it than adults who also have vices they know intellectually are bad. I don’t want to “completely control” my children, but I do want to prevent harm. Same way we put guard rails at the edge of a cliff.
Just to be clear:
Oh god, they’ll get some access? Like, I can’t completely control my children and they are individuals who have the right to start making choices? Jesus Christ, I’m not going to be able to exert my will over them indefinitely?
Are you recommending that we just sit back and let kids random-walk through tiktok? At what age should algorithm-dopamine-drug-app be allowed? There are studies out there showing that this stuff is harmful to ADULTS and this thread is about known impacts on kids. We prevent kids from smoking or drinking. Why do you think preventing access to social media like this is a step too far?
There’s also a question of age. I’m talking as a parent of a pre-teen. I need these controls where I can get them because the internet is a dopamine machine. It’s a real challenge to limit access to it and my kid isn’t prepared to stop watching tiktok the same way they aren’t prepared to stop eating candy. I can physically limit the candy in the house, but guess where I find rogue candy wrappers? Maybe by the time they are 15, I’ll have taken the training-wheels off, in which case we probably agree.
Finally, there’s an additional context for parents that is cultural context: My kid has never watched squidgames, five nights at freddy’s or stranger things. Many, maybe even most, of his peers have, and that leaves him out of those conversations. There are threads up in this post that haunt me: Am I preventing my child from being able to socialize because I won’t let them play/watch <content> that I think is unacceptable? I don’t want roblox, fortnight, or predatorily-monitized games in my kid’s hands until they are ready.
I recently relented on fortnight. My kid spent about $20 of their money on skins and a battle pass. I asked them recently if it was worth it. They said, “no”. I also recently let them create a roblox account. It took about 2 hours for them to determine the whole game was dumb. I think I’m a good parent.
- Comment on CEO of Palantir Says AI Means You’ll Have to Work With Your Hands Like a Peasant 4 days ago:
Perfect. And then later, “I hope you enjoyed your glue pizza. We don’t have enough fuel to reach Paris or return to land. This plane has no emergency beacon or flotation devices and is about to “land” in the ocean. Sorry for the inconvenience!”
- Comment on CEO of Palantir Says AI Means You’ll Have to Work With Your Hands Like a Peasant 4 days ago:
I wish him to sleep in a bed made by AI. Eat a meal made by AI. And then take a flight in a plane made and pulled by AI.
- Comment on lightbulbs 5 days ago:
That’s a neat form factor.
I have something like this mounted in a torchiere floor lamp. high output (something like 15000LM) and a cool color temp (6500K I think). My office has daylight when I turn it on, but it’s aimed up, indirect and high enough it’s not generally in line of sight. Probably would have used the bulb above if I could find cob lights when I bought this one.
- Comment on lightbulbs 5 days ago:
I feel… seen? or maybe not seen.
If I need to read some tiny-ass model number printed in silver on a grey background off the back of a gadget, I need me some 6000K and 1000+ lumens, not some muddy-warm-white <800.
- Comment on France will replace Microsoft Teams, Google Meet, Zoom, Webex and others with its own sovereign video conferencing application "Visio" for public officials 1 week ago:
Does that make this better? A translated French search query would be ‘joining video call isn’t working’ and that will return results for every conference tool known to man.
Call it something like FVC (la France VisioConférence) , or some French play on the way that sounds, which would be a uniquely searchable term in this domain.
This is not a hill I’m dying on, but it’s terminally short sighted and a bad user experience to name your product the same thing as a microslop trademark. They are the worst for this already with their multiple active variants of office 365 tools like outlook and their xbox name nonsense.
Oh, I have a great idea for a new car company. Lets call it ‘Car’! Then people can have a Car Car, or maybe even a Car Car 2026… oh or a Car Truck when we branch out. (future google search: replace car truck 2028 oil filter)
- Comment on France will replace Microsoft Teams, Google Meet, Zoom, Webex and others with its own sovereign video conferencing application "Visio" for public officials 1 week ago:
Came to comment this. I know there are only so many letters, and so many combinations of 4-8 of them, but can we quit naming new things with the name of an old thing?
Finding any details about France’s Visio is going to be a cluster.
- Comment on Microsoft gave FBI a set of BitLocker encryption keys to unlock suspects' laptops: Reports | TechCrunch 2 weeks ago:
I’m certainly not a microslop supporter, but…
They designed a system that recommended that the average user use full disk encryption as part of device setup, and then provided a way that Grandma could easily recover her family photos when she set it up with their cloud.
This was built by an engineer trying to prevent a foreseeable issue. The intent was not malicious. The intent was to get more people more secure by default, since random hacker couldn’t compell ms to give them keys, while still allowing low tech literacy people to not get fucked.
It’s been a while since I installed a new Windows OS, but I’m pretty sure it prompts you to allow uploading your bitlocker key. It probably defaults to yes, but I doubt you can’t say no, or reset the key post onboarding if you want the privacy, and now it’s on you to record your key. You do have to have some technical understanding of the process, though, which is true of just about everything.
That all said, if a company has your data, it can be demanded by the government. This is a cautionary tale about keeping your secrets secret. Don’t put them in GitHub, don’t put them in Chrome, don’t put them online anywhere because the Internet never forgets.
- Comment on AI boom could falter without wider adoption, Microsoft chief Satya Nadella warns 2 weeks ago:
The big difference is that smart phones and centralized internet are somewhat useful. Smartphones at least. Centralized internet… meh, but maybe a dependency.
AI is useful in only very niche and intentional cases. A ‘generic’ LLM is pretty bad at almost everything.
If ‘AI’ had been sold more like: “Give us a year of data samples from your production line and we can use ML to optimize time and temperature based on current weather patterns…” (real world use case I was working in on 2019) etc. then they would have really made the world better. Instead, I have crappy clippy constantly reading my email and suggesting words I wasn’t going to type*.
- I don’t understand how corps accept the idea that their internal emails are no longer internal, since everything is sent to chatgpt/copilot/gemeni/etc as it’s created. Shouldn’t Legal have thrown a tantrum over this?!
- Comment on AI boom could falter without wider adoption, Microsoft chief Satya Nadella warns 2 weeks ago:
Big tech boss tells delegates at Davos that broader global use is essential if technology is to deliver lasting growth
Let me rephrase:
“Smart” entitled person says our product is not showing value, so we need to force people to use it more than we already are after years of cramming it down their throats.
- Comment on Firefox Will Ship with an "AI Kill Switch" to Completely Disable all AI Features - 9to5Linux 1 month ago:
I really like this comment. It covers a variety of use cases where an LLM/AI could help with the mundane tasks and calls out some of the issues.
The ‘accuracy’ aspect is my 2nd greatest concern: An LLM agent that I told to find me a nearby Indian restaurant, which it then hallucinated is not going to kill me. I’ll deal, but be hungry and cranky. When that LLM (which are notoriously bad at numbers) updates my spending spreadsheet with a 500 instead of a 5000, that could have a real impact on my long-term planning, especially if it’s somehow tied into my actual bank account and makes up numbers. As we/they embed AI into everything, the number of people who think they have money because the AI agent queried their bank balance, saw 15, and turned it into 1500 will be too damn high. I don’t ever foresee trusting an AI agent to do anything important for me.
“trust”/“privacy” is my greatest fear, though. There’s documentation for the major players that prompts are used to train the models. I can’t immediately find an article link because ‘chatgpt prompt train’ finds me a ton of slop about the various “super” prompts I could use. Here’s OpenAI’s ToS about how they will use your input to train their model unless you specifically opt-out: openai.com/…/how-your-data-is-used-to-improve-mod…
Note that that means when you ask for an Indian restaurant near your home address, Open AI now has that address in it’s data set and may hallucinate that address as an Indian restaurant in the future. The result being that some hungry, cranky dude may show up at your doorstep asking, “where’s my tikka masala”. This could be a net-gain, though; new bestie.
The real risk, though, is that your daily life is now collected, collated, harvested and added to the model’s data set; all without your clear explicit actions: using these tools requires accepting a ToS that most people will not really read and understand. Maaaaaany people will expose what is otherwise sensitive information to these tools without understanding that their data becomes visible as part of that action.
To get a little political, I think there’s a huge downside on the trust aspect of: These companies have your queries(prompts), and I don’t trust them to maintain my privacy. If I ask something like “where to get abortion in texas”, I can fully see OpenAI selling that prompt to law enforcement. That’s an egregious example for impact, but imagine someone could query prompts (using an AI which might make shit up) and asks “who asked about topics anti-X” or “pro-Y”.
My personal use of ai: I like the NLP paradigm for turning a verbose search query into other search queries that are more likely to find me results. I run a local 8B model that has, for example, helped me find a movie from my childhood that I couldn’t get google to identify.
There’s use-case here, but I can’t accept this as a SaaS-style offering. Any modern gaming machine can run one of these LLMs and get value without the tradeoff from privacy.
Adding agent power just opens you up to having your tool make stupid mistakes on your behalf. These kinds of tools need to have oversight at all times. They may work for 90% of the time, but they will eventually send an offensive email to your boss, delete your whole database, wire money to someone you didn’t intend, or otherwise make a mistake.
I kind of fear the day that you have a crucial confrontation with your boss and the dialog goes something like:
Why did you call me an asshole?
I didn’t the AI did and I didn’t read the response as much as I should have.
Oh, OK.
- Comment on Attitudes 1 month ago:
Similarly, my fantasy is that If I won the lottery, or otherwise became independently wealthy, I’d be doing a ton of different entry-level jobs to find one that hit as a passion.
Construction worker, stagehand, (i’ve already been retail), food service, intern for anything that requires a degree I don’t have, etc.
I like my current job but if I didn’t need the paycheck then I’m not sure I’d stay. I might stick around if I could negotiate terms and only do the parts I liked, though.
I wish I could learn a little about everything, but our culture pushes us to commit and be deep instead, and then we get stuck in a job that used to be a fun hobby.
- Comment on Do we have No Man's Sky fans here? 1 month ago:
This is my issue with NMS.
It’s fun for a while, but it’s a pretty shallow sandbox and after you’ve played in the sand for a bit, it’s all just sand.
If you’re not setting yourself a complex and/or grindy goal, like building a neat base, finding the perfect weapon or ship, filling out your reputations or lexicon, or learning all the crafting recipes to make the ultimate mcGuffin, then there is really not much to do. And, for me, once that goal is accomplished, I’m done for a while.
Each planet is just a collection of random tree/bush/rock/animal/color combinations that are mechanically identical (unless something’s changed. I haven’t played since they added VR). I’m also a gamer who likes mechanical complexity and interactions; I don’t tend to play a game for the actual ‘role playing’.
The hand-written “quests” were fun to do most of the time, but that content runs out quickly.
I have the same problems with Elite Dangerous (I have an explorer somewhere out a solid few hours away from civilized space) and unmodded Minecraft (I can only build so many houses/castles). I’ll pick all of these up every now and then, but the fun wears off more quickly each time.
- Comment on Corn or something idk 1 month ago:
On Mobile, copy paste is not working, so paraphrasing:
They are everywhere and almost nobody is racist anymore
Take a cue from the US. The racists are still racists. They are just currently quietly pissed and likely working to be openly racist again.
Fighting against hate requires constant vigilance.
- Comment on Lemmy users who say that Lemmy users are smarter than Reddit users 2 months ago:
Not antagonistically speaking here.
Do you think your input is not being used to train LLMs when posting on Lemmy? It’s publicly visible without an account.
I’d be shocked if there wasn’t either a scraper, or a whole federated instance, that was harvesting lemmy comments for the big ai companies.
The only difference is that no one is trying to make money off providing that content to them. A big part of the reddit exodus was that reddit started charging for api calls to make cash off the AI feeding frenzy, which broke tools the users liked. With lemmy, there’s no need for a rent-seeking middle man.
- Comment on Parents App'rule'ved 2 months ago:
It suggests to me that someone is measuring the length of a middle finger.
- Comment on YouTube is taking down videos on performing nonstandard Windows 11 installs 3 months ago:
It seemed somewhat topical to me. Google’s censorship is the same trend of enshittification that Yar is talking about.
There are tons of other comments talking about the censorship issue. Using this moment to plug open source software is not unreasonable.
- Comment on Apparently Palantir can access the content of social media accounts that were deleted a decade ago. 3 months ago:
I’ll admit that I should have been more clear that I was paraphrasing and interpreting instead of actually quoting you. The previous message was right above mine, though, so I though it was pretty clear.
Just as you have written me off, I’ve done the same for you. I’m just responding for anyone else who reads this far down and finds this thread, and only because I’m in a waiting room and this is more interesting than HGTV.
I said, and I quote:
I still don’t get your angle. Why are you defending this… I assume the lack of a defense is clear enough proof that you don’t have one.
Palantir scouring the internet, cloud cameras like Flock, Facebook and Google retaining your data forever to maximize profit. None of that is defensible. We should be sounding alarms like OP did and making sure people are aware. Putting others down for ‘not having caught on yet’ (interpreted, you can still correct me if I’m misunderstanding) is counterproductive. We can still resist or reverse the power these huge companies have… but there might be a point where it becomes too late.
Would you prefer to be someone who helped fight, or someone who complained it was futile until it was?
Call your Senators and Representatives. Demand privacy. Elect and support people who are against these kinds of overreach if the current ones won’t.
Love you!
- Comment on Apparently Palantir can access the content of social media accounts that were deleted a decade ago. 3 months ago:
“Anyone could already do this, so why bother being worried that it’s easier now” they said.
I still don’t get your angle. Why are you defending this, or at the very least downplaying it’s impacts? You seem to also be aggravated by this data collection and spying, so why are you so mad that other people are catching on?
“Oh, I’m so smart” they said. Enjoy your useless internet points?
- Comment on Apparently Palantir can access the content of social media accounts that were deleted a decade ago. 3 months ago:
No, that’s not what I said. Widespread data collection and searching used to be something only state actors could accomplish and there were at least theoretically guard rails. Now the barrier of entry has been seriously reduced, the data is owned by a corporation, and being fed to AI. That has a chilling effect as well as being ripe for abuse.
I don’t see an upside.
- Comment on Apparently Palantir can access the content of social media accounts that were deleted a decade ago. 3 months ago:
I’m going to say that this is actually spooky.
Not that it’s unreasonable, but that the scale of what AI can surveil is so vast that there’s no more personal security-via-obscurity.
It used to be that unless someone had a reason to start looking at you, anything you did online or off was effectively impossible to search. You might be caught on some store’s CCTV, Or your cell provider might have location pings, but that wasn’t online for anyone and needed a warrant to have the police use it to track your activities. Now cities are using Flock and similar tools to enable tracking vehicles across the country without any reason, and stores are using cloud-service AI cameras to attempt to track your mood as you move through the store. These tools can and have been abused.
Now, due to the harvesting of this data for AI, anything that’s ever been recorded (video footage, social media posts, etc) and used as training data can be correlated much more easily, long after it occurred, and without needing to be law enforcement with a warrant.
I’d call that spooky.
- Comment on soda 3 months ago:
And you are a hero to that plow driver. And others like you are heroes to the people that also had to be out on terrible weather days and holidays.
I assume a gas station could run without anyone present, leaving the convenience store part closed, but having someone on-hand to hit an e-stop if needed is pretty important.
My goal is not to devalue your work, but rather to support it. “Essential workers” are called that for a reason. We should work to ensure that they are paid their worth. Just because it’s not necessarily a “skilled” job doesn’t mean it’s not important. The bro running the local hedge fund is providing way less actual value than anyone in a service job.
- Comment on No excuses 4 months ago:
This is so me.
I live in a neighborhood with a school. Lots of children roam the streets. Presumably, they are taught to always look both ways, expect cars to misbehave, and otherwise look out for their safety. I haven’t heard of any injuries from cars.
I constantly watch cars, especially the oversized trucks, blow through stop signs, or accelerate to 35 mph on a 1/4 mile stretch of road with side street parking. It really boggles my mind when these same drivers then stop just before the school and drop their kids off to walk the rest of the way. Do you not understand that you put local neighborhood kids at risk from the driving you do, just to save a few moments of avoiding the actual car rider line? Where child safety is a priority.
“Nice stop!”
And then at night, delivery drivers blast through the area. Ignoring stop signs, driving excessively fast past parked cars. Kids live here, and they don’t always remember to look. Especially at dusk, when kids are still playing, but visibility is poor. These drivers gamble with three lives: the inexperienced child, their own, and that of the family that would grieve. For what? A few seconds?
Be nice to each other!!!
- Comment on AI Coding Is Massively Overhyped, Report Finds 4 months ago:
Thanks for your reply, and I can still see how it might work.
I’m curious if you have any resources that do some end-to-end examples. This is where I struggle. If I have an atomic piece of code I need and I can maybe get it started with a LLM and finish it by hand, but anything larger seems to just always fail. So far the best video I found to try a start-to-finish demo was this: www.youtube.com/watch?v=8AWEPx5cHWQ
He spends plenty of time describing the tools and how to use them, but when we get to the actual work, we spend 20 minutes telling the LLM that it’s doing stuff wrong. There’s eventually a prototype, but to get there he had to alternate between ‘I still can’t jump’ and ‘here’s the new error.’ He eventually modified code himself, so even getting a ‘mario clone’ running requires an actual developer and the final result was underwhelming at best.
For me, a ‘game’ is this tiny product that could be a viable unit. It doesn’t need to talk to other services, it just needs to react to user input. I want to see a speed-run of someone using LLMs to make a game that is playable. It doesn’t need to be “fun”, but the video above only got to the ‘player can jump and gets game over if hitting enemy’ stage. How much extra effort would it take to make the background not flat blue? Is there a win condition? How to refactor this so that the level is not hard-coded? Multiple enemy types? Shoot a fireball that bounces? Power Ups? And does doing any of those break jump functionality again? How much time do I have to spend telling the LLM that the fireball still goes through the floor and doesn’t kill an enemy when it hits them?
I could imagine that if the LLM was handed a well described design document and technical spec that it could do better, but I have yet to see that demonstrated. Given what it produces for people publishing tutorials online, I would never let it handle anything business critical.
The video is an hour long, and spends about 20 minutes in the middle actually working on the project. I probably couldn’t do better, but I’ve mostly forgotten my javascript and HTML canvas. If kaboom.js was my focus, though, I imagine I could knock out what he did in well under 20 minutes and have a better architected design that handled the above questions.
I’ve, luckily, not yet been mandated that I embed AI into my pseudo-developer role, but they are asking.
- Comment on AI Coding Is Massively Overhyped, Report Finds 4 months ago:
I think this is what will kill vibe coding, but not before there’s significant damage done. Junior developers will be let go and senior devs will be told they have to use these tools instead and to be twice as efficient. At some point enough major companies will have had data breaches through AI-generated code that they all go back to using people, but there will be tons of vulnerable code everywhere. And letting Cursor touch your codebase for a year, even with oversight, will make it really tricky to find all the places it subtly fucked up.
- Comment on AI Coding Is Massively Overhyped, Report Finds 4 months ago:
I have 3 questions, and I’m coming from a heavily AI-skeptic position, but am open:
-
Do you believe that providing all that context, describing the existing patterns, creating an implementation plan, etc, allows the AI to both write better code and faster than if you just did it yourself? To me, this just seems like you have to re-write your technical documentation in prose each time you want to do something. You are saying this is better than ‘Do XYZ’, but how much twiddling of your existing codebase do you need to do before an AI can understand the business context of it? I don’t currently do development on an existing codebase, but every time I try to get these tools to do something fairly simple from scratch, they just flail. Maybe I’m just not spending the hours to build my AI-parsable functional spec. Every time I’ve tried this, asking something as simple as (and paraphrased for brevity) “write an Asteroids clone using JavaScript and HTML 5 Canvas” results in a full failure, even with multiple retries chasing errors. I wrote something like that a few years ago to learn Javascript and it took me a day-ish to get something that mostly worked.
-
Speaking of that context. Are you running your models locally, or do you have some cloud service? If you give your entire codebase to a 3rd party as context, how much of your company’s secret sauce have you disclosed? I’d imagine most sane companies are doing something to make their models local, but we see regular news articles about how ChatGPT is training on user input and leaking sensitive data if you ask it nicely and I can’t imagine all the pro-AI CEOs are aware of the risks here.
-
How much pen-testing time are you spending on this code, error handling, edge cases, race conditions, data sanitation? An experienced dev understands these things innately, having fixed these kinds of issues in the past and knows the anti-patterns and how to avoid them. In all seriousness, I think this is going to be the thing that actually kills AI vibe coding, but it won’t be fast enough. There will be tons of new exploits in what used to be solidly safe places. Your new web front-end? It has a really simple SQL injection attack. Your phone app? You can tell it your username is admin’joe@google.com and it’ll let you order stuff for free since you’re an admin.
I see a place for AI-generated code, for instant functions that do something blending simple and complex. “Hey claude, write a function to take a string and split it at the end of every sentence containing an uppercase A”. I had to write weird functions like that constantly as a sysadmin, and transforming data seems like a thing an AI could help me accelerate. I just don’t see that working on a larger scale, though, or trusting an AI enough to allow it to integrate a new function like that into an existing codebase.
-
- Comment on AI Coding Is Massively Overhyped, Report Finds 4 months ago:
I’d wager that the votes are irrelevant. Stock overflow is generously <50% good code and is mostly people saying ‘this code doesn’t work – why?’ and that is the corpus these models were trained on.
I’ve yet to see something like a vibe coding livestream where something got done. I can only find a lot of ‘tutorials’ that tell how to set up tools. Anyone want to provide one?
I could… possibly… imagine a place where someone took quality code from a variety of sources and generate a model that was specific to a single language, and that model was able to generate good code, but I don’t think we have that.
Vibe coders: Even if your code works and seems to be a success, do you know why it works, how it works? Does it handle edge cases you didn’t include in your prompt? Does it expose the database to someone smarter than the LLM? Does it grant an attacker access to the computer it’s running on, if they are smarter than the LLM? Have you asked your LLM how many 'r’s are in strawberry?
At the very least, we will have a cyber-security crisis due to vibe coding; especially since there seems to be a high likelihood of HR and Finance vibe coders who think they can do the traditional IT/Dev work without understanding what they are doing and how to do it safely.
- Comment on Samsung brings ads to US fridges 4 months ago:
This is my fear. It’s still possible, barely, to buy a dumb TV. When my current fridge/dishwasher/stove/etc dies in a few years, will there even be a dumb version? Will it cost 5x the price of a spyware version? How about my thermostat. HVAC? Car? And will attempting to disable any of this spyware land me in prison?
Right now, uninformed/unaware/stupid people are affected by this. Pretty soon, everyone will be, or they will have to forego things we consider to be necessities now, like refrigeration and cell phones or be rich enough to buy the privacy-focused models.
I can’t immediately find it, but I just saw another post about a new privacy-focused cellphone with a huge price tag. The established manufacturers have a cost advantage. Samsung et al. can easily make a new fridge with fewer consumer rights, but a new company will have to spend tons of capital to make a factory to put out a comparable product; and they won’t have the advantage of selling your data to subsidize the price.
Privacy is and will become more-so a commodity unless we fight for it.
- Comment on Exactly Six Months Ago, the CEO of Anthropic Said That in Six Months AI Would Be Writing 90 Percent of Code 4 months ago:
That new hire might eat resources, but they actually learn from their mistakes and gain experience. If you can’t hold on to them once they have experience, that’s a you problem. Be more capitalist and compete for their supply of talent; if you are not willing to pay for the real human, then you can have a shitty AI that will never grow beyond a ‘new hire.’
The future problem, though, is that without the experience of being a junior dev, where do you think senior devs come from? Can’t fix crappy code if all you know how to do is engineer prompts to a new hire.
“For want of a nail,” no one knew how to do anything in 2030. Doctors were AI, Programmers were AI, Artists were AI, Teachers were AI, Students were AI, Politicians were AI. Humanity suffered and the world suffocated under the energy requirements of doing everything poorly.
- Comment on Plex got hacked. 4 months ago:
I fully agree: Companies and their leadership should be held accountable when they cut corners and disregard customer data security. The ideal solution would be that a company is required to not store any information beyond what is required to provide the service, a la GDPR, but with a much stricter limit. I would put “marketing” outside that boundary. As a youtube user, you need literally nothing, maybe a username and password to retain history and inferred preferences, but trying to collect info about me should be punished. If your company can’t survive without targeted content, your company should not survive.
In bygone days, your car’s manufacturer didn’t know anything about you and we still bought cars. Not to start a whole new thread, but this ties in to right-to-repair and subscriptions for features as well. I did not buy a license to the car, I bought the fucking car; a license to use the car is called a lease.