Amazon probably does have some programmatic way of determining how much to fit in a truck, but that’s not what this is. Instead, it’s them trying to cheap out on packaging materials in the dumbest way possible, by figuring out what the reasonably acceptable minimum threshold is for packaging durability but not taking into account size or packing multiples of items at all (as far as I can tell).
Comment on Amazon builds AI model to optimize packaging
pixxelkick@lemmy.world 8 months ago
This seems like it has pretty powerful potential for space flight.
Being able to aggressively min max packaging materials to secure materials could be critical for reducing payload sizes on shuttles, where every single individual gram counts.
Each kg of packaging is thousands of dollars to get into orbit, so that’s really appealing.
I’d be curious to see if Amazon is also working on box packing algorithms for maximizing fitting n parcels across x delivery trucks.
IE if you have 10,00 boxes to move, what’s the fewest delivery trucks you can fit those boxes into as fast as possible too, which introduces multiple complex concepts. Both packing to maximize space usage and the order you pack it in to minimize armature travel time…
I’d put money down amazon is perfecting this algorithm right now, and has been for awhile.
dual_sport_dork@lemmy.world 8 months ago
polygon6121@lemmy.world 8 months ago
This is already worked in through mathematics, it is its own mathematical field. We can optimize packaging through formulas that are very fast and accurate. No need to train a AI for that. Especially not for space flight, AI are prone to hallucinations that is not something you want anywhere near any space mission that requires precision and predictability. I believe Johannes Kepler started this field in the 1600s, it is not something new. It is definitely a complex problem, but not new and not unheard of. Amazon is not exactly inventing something new and amazing here…
Syntha@sh.itjust.works 8 months ago
AI is not prone to hallucinations, LLMs are. I doubt Amazon is building a chatbot to optimise packaging.
Ashelyn@lemmy.blahaj.zone 8 months ago
I mean, AI is used in fraud detection pretty often; when it hits a false positive (which happens frequently on a population-level basis), is that not a hallucination of some sort? Obviously LLMs can go off the rails much further because it’s readable text, but any machine learning model will occasionally spit out really bad guesses almost any person could have done better with. (To be fair, humans are highly capable of really bad guesses too).
Womble@lemmy.world 8 months ago
No, false positives and false negatives are not hallucinations. Otherwise things like a blood test not involving any ml would also be halucinating which removes all meaning from the term.
Llewellyn@lemm.ee 8 months ago
What do you consider to be an AI?
And do you consider any of the existing systems to be the one?
Syntha@sh.itjust.works 8 months ago
When I use “AI” I’m using computer science terminology. Artificial intelligence is a subfield of CS, in that sense, any model that comes of that field is, by definition, AI.
polygon6121@lemmy.world 8 months ago
AI in general is definitely prone to hallucinations. It is more commonly seen in LLMs because it is more widely used by the public. It is definitely a problem with all AI
Syntha@sh.itjust.works 8 months ago
Besides generative AI, what models can hallucinate?