inspxtr
@inspxtr@lemmy.world
- Comment on 2024 Self-Host User Survey Results 4 weeks ago:
Wonder how the survey was sent out and whether that affected sampling.
Regardless, with -3-4k responses, that’s disappointing, if not concerning.
I only have a more personal sense for Lemmy. Do you have a source for Lemmy gender diversity?
Anyway, what do you think are the underlying issues? And what would be some suggestions to the community to address them?
- Comment on Why do Americans measure everything in cups? 7 months ago:
someone should make an alternate history tv show where the ship made it. bonus if it’s of a parody kind.
- Comment on How Quickly Do Large Language Models Learn Unexpected Skills? 8 months ago:
re: your last point, AFAIK, the TLDR bot is also not AI or LLM; it uses more classical NLP methods for summarization.
- Comment on Skiff, the private email provider has been acquired by Notion. It is set to shut down its services after 6 months. 9 months ago:
Is there a database tracking companies that start out with good intentions and then eventually gets bought out or sells out their initial values? I’m wondering what the deciding factors are, and how long it takes for them to turn.
- Comment on The tyranny of the algorithm: why every coffee shop looks the same 10 months ago:
Reminds me of this article www.alexmurrell.co.uk/…/the-age-of-average where the author pulls in different examples of designs and aesthetics converging to some “average”.
I’m feeling conflicted with these trends, on one hand it seems like things are becoming more accessible, while on another, feels like a loss.
This especially may be relevant with generative AI - at least for the very few generative arts I look at, at some point they start to feel the same, impersonal.
- Comment on Comcast says hackers stole data of close to 36 million Xfinity customers 11 months ago:
They don’t seem to allow account deletions. Does it mean that this could include accounts that they still keep but people don’t use their services anymore?
- Comment on NEW COMMUNITY ALERT: Boomers Being Fools 11 months ago:
forgive my naivety, how does such a community avoid promoting ageism?
- Comment on Google proposes Project Ellmann, a chatbot that intimately knows you 11 months ago:
suggests either these people are so detached from reality, or they are appealing this to a very specific sets of people under the guise of a general appeal
- Comment on How Googlers cracked an SF rival's tech model with a single word | A research team from the tech giant got ChatGPT to spit out its private training data 11 months ago:
Something like this, unless they know the root cause (I didn’t read the paper so not sure if they do), or something close to it, may still be exploitable.
- Comment on Disclosure of sensitive credentials and configuration in containerized deployments - ownCloud 11 months ago:
what are the other alternatives to ENV that are more preferred in terms of security?
- Comment on Selfhoster “plagiarism” checker with custom sources? 1 year ago:
Thanks for the suggestions! I’m actually also looking into llamaindex for more conceptual comparison, though didn’t get to building an app yet.
But I’m also looking for more literal comparison, AFAIK, the choice of embedding model will affect how the similarity will be defined. Most of the current LLM embedding models are usually abstract and the similarity will be conceptual, like “I have 3 large dogs” and “There are three canine that I own” will probably be very similar. Do you know which choice of embedding model I should choose to have it more literal comparison?
That aside, like you indicated, there are some issues. One of it involves length. I hope to find something that can build up to find similar paragraphs iteratively from similar sentences. I can take a stab at coding it up but was just wondering if there are some similar frameworks out there already that I can model after.
- Comment on We watched 1,000 TikToks in one sitting. The algorithm served up a shocking number of ads, rivaling network TV. 1 year ago:
yeah agreed with your sentiment. I think it’s good to have an intuition about something, but it’s much better when there’s data to back it up.
Cuz then, they can do the same with others, say Youtube or other streaming services, and start to compare the numbers, like % of ads, what types of ads, how long are the ads relative to content, how many of these ads are political, how many of these ads may be harmful, …
Having these numbers can be quite handy for other researchers and regulators to look into these issues more concretely, rather than just say, “as your brothers and sisters already know, tiktok serves ads”
- Submitted 1 year ago to selfhosted@lemmy.world | 4 comments
- Comment on Mark Zuckerberg tears ACL while training for MMA fight early next year 1 year ago:
Why does this sound like an episode of Friends?
- Comment on What got you into coding ? (aside from money) 1 year ago:
got into coding cuz I found out that’s how I can automate analysis and play with research questions more easily.
- Comment on (Please see comments)Alternatives to Signal if they exit EU due to ending E2EE 1 year ago:
I’m quoting the page that I linked from privacyguides warning
These messengers do not have Forward Secrecy, and while they fulfill certain needs that our previous recommendations may not, we do not recommend them for long-term or sensitive communications. Any key compromise among message recipients would affect the confidentiality of all past communications.
- Comment on (Please see comments)Alternatives to Signal if they exit EU due to ending E2EE 1 year ago:
Based on privacyguides suggestion page itself, SimpleX chat would be the next in line you can try.
Briar is only for Android AFAIK. Matrix/Element does offer E2EE chat/vid but, based on the page, it’s not recommended for long term sensitive use.
Regardless, with the current situation against encryption, any app that stays will be subject to similar conundrum about leaving/abiding the law like Signal. The ones abiding may need more scrutiny, of course.
- Comment on Meta is paying the celebrity faces behind its AI chatbots as much as $5 million for 6 hours of work, report says 1 year ago:
This is straight out of the movie “The Congress”
- Comment on Kids and teens are inundated with phone prompts day and night 1 year ago:
I wonder whether emails, slack/discord notifications, chatrooms for work/study/play, … are the equivalent to social media notifications for adults.
- Comment on Are We Ready For This Site's Endless Feed of AI-Generated Porn? 1 year ago:
I remember reading that this may be already happening to some extent, eg people sharing tips on creating it on the deep web, maybe through prompt engineer, fine tuning or pretraining.
I don’t know how those models are made, but I do wonder the ones that need retraining/finetuning by using real csam can be classified as breaking the law.
- Comment on The Gruesome Story of How Neuralink’s Monkeys Actually Died 1 year ago:
lol I know you’re kidding, but there’s implication of those willing to get things implanted. Society seems to run on hype nowadays. Look at AI and how fast people are jumping on board with trying it, sometimes out of FOMO. Not to say there’s no merit, but if that FOMO feeling spreads real quick, without proper guardrails, Musk will eventually get what he wants.
- Comment on The Gruesome Story of How Neuralink’s Monkeys Actually Died 1 year ago:
Not at the cost of humanity. Plus, that statement can be recycled to defend all the horrible inventions and experiments in the past.
- Comment on The Gruesome Story of How Neuralink’s Monkeys Actually Died 1 year ago:
what really confuses me is how the FDA approves this without a few more years of animal testing and protocol refinement.
- Comment on How social media killed the protest — For a certain kind of activist, politics has been reduced to pure performance 1 year ago:
- Comment on 'A cavity is not a vagina': Trans woman refused healthcare in France 1 year ago:
I understand doctors are humans too, and presume that is your reason for questioning “Why?”.
But in defense of doctors being more “thick skinned”, I think patients are generally the more vulnerable side as they are usually the ones with issues. If every doctor they come to snaps at them, they start to lose trust in healthcare providers.
- Comment on Your Gmail and Instagram are training AI. There’s little you can do about it. 1 year ago:
how does this work with universities and companies that use GMail/Outlook for their emails?
- Comment on Selfhosted private/secure blog/journal 1 year ago:
Here are some options:
- crypt.ee: I tried this before, I don’t think it’s selfhostable but quite usable, and nice UI. Encryption available. Ghost folders if you want to. Multimedia available, not sure about storage
- joplin: you can use Nextcloud (or many other options like Dropbox) for sync and hence storage depends on your cloud solution. E2EE, has plugins, and simple enough to use.
- anytype.io and logseq: I’ve seen these mention in many places but I haven’t used either. But they seem to have very rich features, not sure about selfhosting though.
- Comment on Very, Very Few People Are Falling Down the YouTube Rabbit Hole | The site’s crackdown on radicalization seems to have worked. But the world will never know what was happening before that 1 year ago:
I was aware of this study when they presented it virtually (can’t remember where), and while I don’t have an issue with their approach and results, I’m more concerned about the implications of these numbers. The few percent that were exposed to extremist content may seen small. But scaling that up to population level, personally that is worrisome to me … The impact of the few bad apples can still catastrophic.
- Comment on ChatGPT generates cancer treatment plans that are full of errors — Study finds that ChatGPT provided false information when asked to design cancer treatment plans 1 year ago:
while I agree it has become more of a common knowledge that they’re unreliable, this can add on to the myriad of examples for corporations, big organizations and government to abstain from using them, or at least be informed about these various cases with their nuances to know how to integrate them.
Why? I think partly because many of these organizations are racing to adopt them, for cost-cutting purposes, to chase the hype, or too slow to regulate them, … and there are/could still be very good uses that justify it in the first place.
I don’t think it’s good enough to have a blanket conception to not trust them completely. I think we need multiple examples of the good, the bad and the questionable in different domains to inform the people in charge, the people using them, and the people who might be affected by their use.
Kinda like the recent event at DefCon trying to exploit LLMs, it’s not enough we have some intuition about their harms, the people at the event aim to demonstrate the extremes of such harms AFAIK. These efforts can help inform developers/researchers to mitigate them, as well as showing concretely to anyone trying to adopt them how harmful they could be.
Regulators also need these examples in specific domains so they may be informed on how to create policies on them, sometimes building or modifying already existing policies of such domains.
- Comment on [deleted] 1 year ago:
Except for signal, matrix is also decentralized. I understand what you meant tho. I think matrix still stores user data on servers, but they could be encrypted. Others who know please correct me.