They are. They record the data, stealing it. They search it, and reprint it (in whole or in part) upon request.
They search the data-space or what they’re trained on (our content, the content of human beings), and reproduce statistically defined elements of it.
They’re search engines that have stolen what they’re trained on, and reproduce it as “results”.
Searching and reproducing content they’ve already recorded, is absolutely part of what they are.
futatorius@lemm.ee 2 weeks ago
Their input sides are based on crawling, just as search is.
DarkCloud@lemmy.world 2 weeks ago
Yeah, and then they convert that to a weighted probabilities or a “data space” they then search during content generation.