5C5C5C
@5C5C5C@programming.dev
- Comment on Google Was Set to Host an Israeli Military Conference. When We Asked About It, the Event Disappeared. 3 months ago:
Google is an enormous company which operates flatter than you’d expect for an organization of its size. It’s entirely possible that someone from Google was involved in organizing this (i.e. booking the venue) without having buy-in from leadership. Once leadership became aware after being asked about it, they may have shut the whole thing down because they knew the optics would be bad.
- Comment on Existing California solar customers may get blindsided with net metering cuts 4 months ago:
PG&E was literally the villain in the real life Erin Brockovich story.
- Comment on Pizza 4 months ago:
That’s why he couldn’t hold it in.
- Comment on Dune is about worms. 4 months ago:
Some might not realize that this is necessary to avoid the Australian sandworms. She was just trying to share her culture with us, and we gave her so much grief for it, smh.
- Comment on New Cobalt-Free Silicon EV Battery Is The Best Thing Ever 4 months ago:
Nothing is ever better in every conceivable way than the current state of the art.
Probabilistically, sure, but it’s not impossible that there has been some piece of knowledge or understanding that’s been missing, and that massive breakthroughs are possible once the process is figured out.
I think a fair modern example is LED light bulbs. They are better in every conceivable way than incandescent or fluorescent lightbulbs: they last longer, use less energy, shine brighter, use less toxic materials, and are easy to mass produce. But there were several decades where much of the industry believed that LEDs would never be very useful as a light source because we could only produce red and green, and it was generally believed that a blue LED would be impossible to produce.
Then one guy decided it would be his life mission to invent the blue LED, and the sonuvabitch did it. Now LEDs are the only sensible thing to use to produce light.
It’s always possible for this kind of breakthrough to happen, especially in material science where the complexity of how molecules interplay is nearly incomprehensible.
- Comment on Checkmate, Atheists 5 months ago:
My only concern is the demographic that would have been too lazy to vote but now will be frothing at the mouth to vote against a black woman.
I can only hope they’re outweighed by the demographic that was apathetic toward Biden but is willing to get off the couch to vote for Harris.
- Comment on CrowdStrike downtime apparently caused by update that replaced a file with 42kb of zeroes 5 months ago:
I don’t doubt that in this case it’s both silly and unacceptable that their driver was having this catastrophic failure, and it was probably caused by systemic failure at the company, likely driven by hubris and/or cost-cutting measures.
Although I wouldn’t take it as a given that the system should be allowed to continue if the anti-virus doesn’t load properly more generally.
For an enterprise business system, it’s entirely plausible that if a crucial anti-virus driver can’t load properly then the system itself may be compromised by malware, or at the very least the system may be unacceptably vulnerable to malware if it’s allowed to finish booting. At that point the risk of harm that may come from allowing the system to continue booting could outweigh the cost of demanding manual intervention.
In this specific case, given the scale and fallout of the failure, it probably would’ve been preferable to let the system continue booting to a point where it could receive a new update, but all I’m saying is that I’m not surprised more generally that an OS just goes ahead and treats an anti-virus driver failure at BSOD worthy.
- Comment on CrowdStrike downtime apparently caused by update that replaced a file with 42kb of zeroes 5 months ago:
When talking about the driver level, you can’t always just proceed to the next thing when an error happens.
Imagine if you went in for open heart surgery but the doctor forgot to put in the new valve while he was in there. He can’t just stitch you up and tell you to get on with it, you’ll be bleeding away inside.
In this specific case we’re talking about security for business devices and critical infrastructure. If a security driver is compromised, in a lot of cases it may legitimately be better for the computer to not run at all, because a security compromise could mean it’s open season for hackers on your sensitive device. We’ve seen hospitals held random, we’ve seen customer data swiped from major businesses. A day of downtime is arguably better than those outcomes.
The real answer here is crowdstrike needs a more reliable CI/CD pipeline. A failure of this magnitude is inexcusable and represents a major systemic failure in their development process. But the OS crashing as a result of that systemic failure may actually be the most reasonable desirable outcome compared to any other possible outcome.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
Thanks for your candid views on this.
To be clear, our interest in subsistence farming is not intended to do anything to solve the problems we’re facing. It’s an attempt to figure out how we might try to survive locally after the global supply chains collapse. We’re particularly researching what crops might be viable in a landscape that has been reshaped by the changing climate. Additionally we’re studying everything we can about community organizing and systems of self-governance that promote collaboration over individual greed.
This might all sound defeatist to someone like yourself who is still committed to fighting the good fight, but we see it as a contingency plan that our community’s ability to survive may depend on in the future.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
I really admire that you’re committed to recycling and waste reduction. Do you have any resources you’d recommend for me to learn more about what’s going on in that space and what’s being done to combat the acceleration of plastic and electronics waste?
I know it’s “not your job” to educate me, but everything I can find on the topic suggests that we don’t have a viable path to manage the accelerating growth of waste, and we don’t have very effective systems for recycling, so even recyclable waste is mostly just being dumped in landfills because it’s more “economical” to just keep churning out products from new materials. I’d be very happy for all of that to be wrong, so any credible source you can point me at to debunk that narrative would be very much appreciated.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
Let me know which part was confusing to you
The part where you left out any viable path for any of the hypothetical solutions to be realized 🤷♂️ You of all people should know that a blueprint is worthless if there’s no process available to build what it describes.
Damn here I am thinking that this is one of the most important parts of civilization.
I mean yeah, I do agree that sanitation and water works are the crowning achievement of human civilization to this very day. But I’ve gotta say it doesn’t inspire confidence if the people running those systems think that concerns about sustainability are something to have a group chuckle about.
Just because the work you do is important doesn’t mean it’s beyond the scrutiny of ecological sustainability. All your good work won’t amount to much in the long run if we can’t find a path to reducing consumption and prolonging the viability of these systems. We don’t have infinite resources, and our ability to recycle is nowhere near what it needs to be up keep up with economic demand.
Tell you what, why not be the change you want to see in the world and stop flushing your toilet, stop using tap water, stop recycling anything, and don’t set your garbage out.
My partner and I are unironically taking the time to research subsistence farming and how to maintain very basic personal water collection and waste removal/reuse systems. We’re also learning about perma-computing so that hopefully we can preserve some of the knowledge that humans have accumulated into the future.
We see it as a foregone conclusion that human civilization as we know it will entirely collapse, probably sooner than anyone cares to admit, so we’re making contingency plans. People with your dismissive attitude are a big part of why we see it as a forgone conclusion. Because as far as we can tell you’re in the 95%+ majority of people on this planet, which means hardly anyone is putting effort into solving these existential problems that we’re facing. Problems which you have offered no viable solution to, despite your insistence otherwise.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
That isn’t my job.
So then you didn’t “solve” the impossible problems.
I hope you and your colleagues have a good laugh about how the work you do is contributing to the march towards the end of human civilization as we know it.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
I never suggested these problems are impossible to solve, but you haven’t solved them in your post because you haven’t laid out how to overcome the political and economic resistance to implementing any of this, and that’s where the biggest challenge is.
Although I think it’s naive to believe that nuclear power and renewable energy can allow us to keep consuming energy recklessly. Renewable energy technology still puts a significant strain on the environment, in terms of mining rare-earth elements, pollution produced during manufacturing, and material waste from devices that have reached end of life. Nuclear energy is rife with controversy… I used to be firmly in support of it, but I’ve grown skeptical, largely because of the ecological damage from the mining and construction processes, and we don’t have a clear story of what end of life looks like for a nuclear power plant. A plant can only be expected to operate for 40-60 years at which point it needs to be demolished and rebuilt, repeating the massive costs of material waste and construction all over again.
At the end of the day the only way for humanity to survive is for everyone to be reducing their consumption, but I honestly the think the vast majority of people today would rather die and take everyone else down with them than accept more responsible consumption habits.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
I can’t tell if you’re suggesting that foundation models (which is the underpinning technology of LLMs) aren’t being used for the things that I said they’re being used for, but I can assure you they are, either in commercial R&D or in live commercial pictures l products.
The fact that they shouldn’t be used for these things is something we can certainly agree on, but the fact remains that they are.
Sources:
-
Wayve is using foundation models for driving, and I am under the impression that their neutral net extends all the way from sensor input to motor control: wayve.ai/thinking/introducing-gaia1/
-
Research recommending the use of LLMs for giving financial advice: papers.ssrn.com/sol3/papers.cfm?abstract_id=48500…
-
LLMs for therapy: blog.langchain.dev/mental-health-therapy-as-an-ll…
So this all kinds of goes back to my point that some form of accountability is needed for how these tools get used. I haven’t examined the specific legislation proposal to give any firm opinion on it, but I think it’s a good thing that the conversation is happening in a serious way.
-
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
Literally nothing you’ve said gives any indication that you actually know the current state of foundation model research. I won’t claim it’s my research specialty, but I work directly with people whose full time job is research and tuning on foundation models, and everything I’m saying is being relayed from conversations that I have with them.
“Cannot ever possibly be used like that”… Like what specifically? To drive a car? That’s being done. To give financial advice? That’s being done. To console people who are suicidal or at risk of harming themselves? That’s being done. To make kill / no kill decisions in an active warzone? It’s being considered (if not already being done in secret).
This technology is being used in extremely consequential positions despite having very weak guarantees around safety. This should give any reasonable person pause. I’m not taking any firm stance on whether this specific regulation is the right approach, but if you think there should be no accountability for the outcomes of how this technology gets used then I guess you’re someone who thinks seatbelts should be optional in cars and it’s okay for airplanes to fall out of the sky due to neglect.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
Non-deterministic algorithms such as Monte Carlo methods or simulated annealing can still be constrained to an acceptable state space. How to do this effectively for LLMs is a very open question, largely because the state space of the problems that they are applied to is incomprehensibly huge.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
That’s not what an algorithms researcher means when we talk about “understanding”. Obviously we know the mechanism by which it operates, it’s not an unknown alien technology that dropped into our laps.
Understanding an algorithm means being able to predict the characteristics of its outputs based on the characteristics of its inputs. E.g. will it give an optimal solution to a problem that we pose? Will its response satisfy certain constraints or fall within certain bounds?
Figuring this stuff out for foundation models is an active area of research, and the absence of this predictability is an enormous safety concern for any use cases where the output can be consequential.
It cannot possibly develop agency.
I don’t believe I’ve suggested anywhere that I think it will, but I’ll play around with this concern anyway… There’s a lot of discussion going on about having models feed back on themselves to learn from their own output. I don’t find it all that hard to imagine that something we could reasonably consider self awareness could be formed by a very complex neural network that is able to consume and process its own outputs. And once self awareness starts to form, it’s not that hard for me to imagine a sense of agency following. I have no idea what the model might use that agency for, but I don’t think it’s all that far fetched to consider the possibility of it happening.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
Sure, but this outcome is not at all surprising. There are plenty of smart AI people that have nuanced views of what kind of threat could be posed by recklessly unleashing tools that we don’t fully understand into the hands of people who are likely to do harmful things with them.
It’s not surprising that those valid nuanced concerns get translated into overly simplistic misrepresentations entangled with pop sci fi panic around rogue AI as they try to move into public discourse.
- Comment on AI's Future Hangs in the Balance With California Law 5 months ago:
AI person reporting in. Without saying whether or not I personally believe that the current tools will lead to the end of humanity, I’ll point out a few possibilities that I find concerning about what’s going on:
-
The hype around AI is being used to justify mass layoffs, where humans are being replaced by tools that do a questionable job and can’t really understand the things those humans could understand. Whether or not the AI can do as good of a job according to some statical measurement is less relevant than the fact that a human is less likely to make an extremely grave mistake and more likely to be able to recognize when that does happen. I’m concerned this will lead to cross-industry enshitification on an unprecedented scale.
-
The foundation models consume a huge amount of energy. The more impressive you want it to be, the more energy it needs. As long as the data centers which run them are dependent on fossil fuels, they’ll be pumping a huge amount of carbon in the air just to do replace jobs that we didn’t need to have replaced.
-
As these tools are used more and more, they’re going to end up “learning” from content created by themselves instead of something that’s closer to a ground truth. It’s hard to predict what kind of degradation of service will come from this, but the more we create systems that rely on these tools, the more harm it will do to us.
-
Given the cost and nature of these tools, they’re likely to yield the most benefit to moneyed interests that want to automate the systems that maintain their power and wealth. E.g. generating large amounts of convincing disinformation to manipulate the public into supporting politicians or policies that benefit a small number of wealthy people in the short term while locking humanity into a path towards destruction.
And none of this accounts for possible future iterations of AI tools that may be far more capable than what exists today. That future technology will most likely be controlled by powerful people who are primarily interested in using it to bolster the systems that keep them in power, to the detriment of humanity as a whole.
Personally I’m far less concerned about a malicious AI intentionally doing harm to humanity than AI being used as a weapon by unscrupulous people.
-
- Comment on Also, you have been turned into a worm. 6 months ago:
If we are now considering philosophical intellectual exercises to be memes then this description is accurate.
- Comment on Another mystery solved. 6 months ago:
You can’t prove that Godzilla’s bones aren’t hollow ballast tanks that can be emptied and filled as needed.
- Comment on Another mystery solved. 6 months ago:
Ordinary biomatter is very close to the density of water to begin with. That’s why having a little air in your lungs is enough to be the difference between sinking and floating.
If Godzilla’s biomatter under 1atm of pressure has a density close to water then being able to compress or expand an empty chamber inside his body by even just a tiny percentage of his ordinary overall volume could be the difference between floating at sea level or sinking to extreme depths.
- Comment on Another mystery solved. 6 months ago:
“Lungs aren’t really inside” is not an argument that I thought I’d be confronted with.
If you find that your lungs are not inside your body then I urge you to seek immediate medical attention.
- Comment on Another mystery solved. 6 months ago:
You’d be right if the cavity is only compressing other organs inside the body without changing the overall volume, but I don’t know why you seem to insist on making that assumption.
I thought it would be clear from my original description, via the analogy with lungs, that the cavity would not squish the internal organs but rather expand the overall volume of the body.
- Comment on Another mystery solved. 6 months ago:
My head canon for sea-based Kaiju is they have a sack of muscles somewhere inside their body that can expand a cavity, kind of like the diaphragm expands the lungs, except instead of taking in air it just creates a volume of vacuum inside of them. This makes them extremely bouyant relative to the surrounding sea pressure, so they rapidly ascend and can casually float like a boat near the surface.
But if they ever want to dive again, they just let that cavity collapse and all their bouyancy goes away.
- Comment on Connected cars’ illegal data collection and use now on FTC’s “radar” 7 months ago:
I’m disappointed to learn that she was born British so she can never run for president in America.
… Not that the FTC chair is known to be a pipeline to the presidency, but I’m ready to turn over every stone at this point.
- Comment on Risk your life with this one easy trick! 7 months ago:
I feel like this is rapidly approaching the “is water wet?” conundrum.
- Comment on Is Boeing in big trouble? World's largest aerospace firm faces 10 more whistleblowers after sudden death of two 7 months ago:
Life insurance premiums are about to skyrocket for Boeing whistleblowers.
- Comment on Are you prepared for the ramifications of windows 10 EoL? 8 months ago:
Sounds like there are going to be a lot of machines running a fresh install of Linux next year. Microsoft really does ♥️ Linux.
- Comment on Critical 'BatBadBut' Rust Vulnerability Exposes Windows Systems to Attacks 8 months ago:
Except it’s actually an “Every language and library that provides this feature” problem because literally no one was aware that this sanitization problem even existed, and Rust is among the first to actually fix it.