• 2 Posts
  • 268 Comments
Joined 2 years ago
cake
Cake day: June 6th, 2023

help-circle



  • It’s not sunk cost, dude. We agreed that $120 will get them 5 years of service that meets their needs. Even if they switch to jellyfin after 5 years, they still got their money’s worth.

    It’s only sunk cost if they are worse off than if they had switched earlier. I guess if you’re arguing that they would still have $120 if they switch today, I would argue they should still pay that $120 toward jellyfin’s development. And that’s assuming they have time to switch to jellyfin AND it fits 100% of their usecases, either of which could be untrue.


  • Or Plex currently does everything they need it to, and $120 for 5+ years of keeping that going without any interruption of service is very reasonable. In the meantime, jellyfin will only get better and there might even be other options available by then.

    Stop trying to make the issue black and white, one-size-fits-all. There are perfectly legitimate reasons for people to use both Plex and Jellyfin.




  • To be clear, I’m talking about people saying things like “those people are lesser than me”, not things like “those people should be eradicated”. Inciting violence, or any crime, is not an exercise of free speech, that’s a crime.

    I guess I just don’t see any ethical difference between wielding the power of legislation to silence speech, and an angry mob of vigilantes gathering and silencing them in person. Either way, it’s the society saying “we don’t like your words, and we’re gonna punish you for that.”

    I just know that throughout history, people have used “I’m confident in my beliefs” to justify limiting speech they thought would be harmful to their society, only for us to look back in shame at their intolerance.

    I can say I’m confident that intolerance harms our society, I just don’t think it’s possible to legislate away hate. We can physically intimidate people into hiding their hate, but making hate illegal will never get rid of it. But maybe that’s the best we can ever do, I don’t know.

    Looking at history, i just don’t have any reason to believe that any sociological hurdle can be solved by moving strictly in a “positive” direction. I understand local maxima, and understand that society always has to regress before it can progress. For the same reason we can’t legislate away hate, we can’t legislate in “progress”. We might try, and it might seem like it’s working for a little while, until it doesn’t. And that’s when humanity learns a new lesson.


  • Have hard lines like this ever worked throughout history, though? It’s not like the people who originally came up with the concept of free speech didn’t think of this exact case. But they believed it was more important for the people to deal with speech they don’t like themselves (within the bounds of the law, of course) than for a government to silence speech.

    I see a problem with inauthentic behaviour online, using bots to artificially amplify hate speech to make it seem more prominent than it actually is. But I think having 100 people tolerate 1 hateful asshat’s speech is the definition of democracy. That doesn’t mean harassment is legal. That doesn’t mean assault or murder or jim crow laws should be tolerated. The worst case is the hate catches on and spreads democratically, and that sucks, but if it happens I guess that’s the society we live in for now, and hopefully it’s just a phase. But if a government artificially silences hate speech, you’re just asking for that to come back and bite you later. Now all those people who would have simply been hateful now also distrust the system they live in, and will seek to dismantle it and replace it with a hateful one.

    IMO this is exactly why Churchill said democracy is the worst form of government, except for all the others. Thinking that we can live in a society that is systematically devoid of hate is attractive, but it’s a Nirvana Fallacy and is destined to fail. This isn’t new ground we’re treading.


  • I agree that you can’t know if the AI has been deliberately trained to act nefarious given the right circumstances. But I maintain that it’s (currently) impossible to know if any AI had been inadvertently trained to do the same. So the security implications are no different. If you’ve given an AI the ability to exfiltrating data without any oversight, you’ve already messed up, no matter whether you’re using a single AI you trained yourself, a black box full of experts, or deepseek directly.

    But all this is about whether merely sharing weights is “open source”, and you’ve convinced me that it’s not. There needs to be a classification, similar to “source available”; this would be like “weights available”.



  • Is there any good LLM that fits this definition of open source, then? I thought the “training data” for good AI was always just: the entire internet, and they were all ethically dubious that way.

    What is the concern with only having weights? It’s not abritrary code exectution, so there’s no security risk or lack of computing control that are the usual goals of open source in the first place.

    To me the weights are less of a “blob” and more like an approximate solution to an NP-hard problem. Training is traversing the search space, and sharing a model is just saying “hey, this point looks useful, others should check it out”. But maybe that is a blob, since I don’t know how they got there.



  • Yeah, I agree that in the long term those two sentiments are inconsistent, but in the short term we have to deal with allegedly misguided layoffs, and worse user experiences, which I think makes both fair to criticise. Maybe firing everyone and using slop AI will make your company go bankrupt in a few years, and that’s great; in the meantime, employees everywhere can rightfully complain about the slop and the jobs.

    But yeah, I don’t think it’s fair to complain about how “inefficient” an early technology is and also call it “magic beans”.




  • Yeah, I understand that you personally choose to disagree with reality, maybe you don’t like what reality has become, but unfortunately that doesn’t make it less real.

    Twitter wasn’t profitable for its entire existence, it’s often a cesspool of ragebaiters, but clearly it has value because the second it was taken over, everyone insisted on continuing to use it, even choosing to migrate to various clones.

    Uber and Lyft have been struggling to be profitable by effectively stealing from their drivers, but millions of people get off a plane and immediately use the services every day. It clearly has value.

    Same for doordash and uber eats.

    Your personal distaste for the business practices are valid, but they’re not relevant when discussing what the current state of the technology is. For many millions of people, chatgpt has (for better and worse) replaced traditional search engines. Something like 80% of students now regularly use AI for their homework. When Deepseek released, it immediately jumped to #1 on the Apple Store.

    None of that is because they’re “magic beans” from which no value sprouts. Like it or not, people use AI all. the. time. for everything they can imagine. It objectively, undeniably has value. You can staunchly say pretend it doesn’t, but only if you are willingly blind to the voluntary usage patterns of hundreds of millions (possibly billions) of people every hour of every day.

    And for the record, I am not in that group. I do not use any LLMs for anything currently, and if anything makes me use AI against my will, I will promptly uninstall it (pun intended).




  • Lol this article is very relevant to a lot of scam industries (essential oils, Earthing, 5G protection crystals, etc), but AI is objectively not one of them.

    Regardless of how much of a bubble we’re in, regardless of how many bad ideas are being pushed to get VC funding or pump a stock, regardless of how unethical or distopian the tech is, AI objectively has value. It’s proving to be the most disruptive tech since the world wide web (which famously had a very similar bubble of bad ideas), so to call it “magic beans” is just wishful thinking at best.