• 1 Post
  • 19 Comments
Joined 1 year ago
cake
Cake day: July 6th, 2023

help-circle
  • I agree with you, however Jellyfin is not intrinsically more secure than any other piece of software. You have to be very careful how you go about deploying it if you open up external access, as you are dependent on the Jellyfin devs to fix vulnerabilities and they aren’t actually being paid to do this. If you’re paranoid about privacy, you should be paranoid about this too; the people sending subpoenas aren’t above port-scans on ISP subscribers, they did it back in the early days of torrents.

    You get control and privacy, but you also get responsibility. It’s a trade-off, and one I’d certainly make if Jellyfin were more mature. That’s just me though, I’ve been hosting my own stuff for about a decade now and I can set up an isolated environment for Jellyfin to run within. Plex is a lot more newbie-friendly and I’d still recommend it for most folks unless they for sure know what they’re doing.

    As an aside, these concerns are common to all FOSS software that don’t have deep-pocketed backers. Jellyfin is likely never getting those, unfortunately. I hope they can find some other way of sustaining themselves, they’ve not got much money for the scale of development needed and it’s all volunteer-driven today.

    https://opencollective.com/jellyfin

    I want them to keep going, and I’ve even donated to them. I still don’t think it’s at a place to replace Plex for most people yet though.



  • Switching between wasn’t seamless, it kept forgetting where I left off on the last device, which was pretty annoying. Also, mobile/remote connectivity was spotty for me. Never got to the bottom of that, but my best guess is Plex’s relay system makes up for a lot of random network issues. My best work-around was to add my phone to tailscale, but obviously that’s not a great solution and won’t work for a lot of devices.

    Overall, my impression was that Plex is a lot more polished. I also bought a lifetime membership years ago, so I have no incentive to switch to something that isn’t better. Plex isn’t perfect, but it was still better than Jellyfin as of a few months ago. I honestly hope that changes soon, I have zero faith in Plex as a company.






  • I’m sure there are plenty of Israelis that want to do this even if they won’t admit it to themselves but this isn’t the final anything. The IDF has killed around 37,000 Palestinians out of ~2.3 million. That’s horrible but nowhere near the “barely any left” stage.

    A genocide on the scale of millions takes industrial effort to accomplish. I’m not saying it couldn’t happen, but given Israel’s reliance on foreign aid, current industrial capacity, and political position, it seems unlikely. My guess is Israel will take some more territory and the conflict (kinda tough to call the IDF bombing almost exclusively civilians a war) will peter out. Foreign aid will be allowed back in and Israel will put its mask back on.

    Personally, I don’t see how this doesn’t end with half the middle east actively going to war with Israel if they don’t stop soon. The only thing really keeping them safe is the US, and Israel has burned a lot of political capital here. Their leaders are awful, power-hungry shits, but they’re not stupid. If they don’t try to rebuild some of that capital, there’s every chance that Israel loses its lifeline.

    What comes years after things die down, I don’t know. Gazan sentiment towards Israel was already overwhelmingly negative before this, but the IDF has never done anything on this scale before. I don’t think Israel can allow Gaza any type of self-governance for decades after this. This is beyond even post-WW2 Japan levels of destruction, and unlike Japan every nation around them is still on their side.


  • You’re using “machine learning” interchangeably with “AI.” We’ve been doing ML for decades, but it’s not what most people would consider AI and it’s definitely not what I’m referring to when I say “AI winter.”

    “Generative AI” is the more precise term for what most people are thinking of when they say “AI” today and it’s what is driving investments right now. It’s still very unclear what the actual value of this bubble is. There are tons of promises and a few clear use-cases, but not much proof on the ground of it being as wildly profitable as the industry is saying yet.





  • Updated to be specific, I’m using Cinnamon. Muffin is the builtin tiling window manager for Cinnamon and it does exactly what you’re describing. The problem is that it moves tiles, it doesn’t absolutely position them. You have to keep moving tiles around to get them where you want them, Rectangle just has hotkeys to immediately place and resize to fit the active window for each quadrant that it supports:

    • ctrl+cmd+left: top left quadrant
    • ctrl+cmd+right: top left quadrant
    • shift+ctrl+cmd+left: bottom left quadrant
    • shift+ctrl+cmd+right: bottom left quadrant
    • alt+cmd+left: left half
    • alt+cmd+right: right half
    • alt+cmd+up: top half
    • alt+cmd+left: bottom half
    • alt+cmd+f: full screen

    It’s hard to express how natural that feels after using it for a bit, and I’m still using a Macbook for work so the muscle memory is not going away.



  • I didn’t say it wasn’t amazing nor that it couldn’t be a component in a larger solution but I don’t think LLMs work like our brains and I think the current trend of more tokens/parameters/training LLMs is a dead-end. They’re simulating the language area of human brains, sure, but there’s no reasoning or understanding in an LLM.

    In most cases, the responses from well-trained models are great, but you can pretty easily see the cracks when you spend extended time with them on a topic. You’ll start to get oddly inconsistent answers the longer the conversation goes and the more branches you take. The best fit line (it’s a crude metaphor, but I don’t think it’s wrong) starts fitting less and less well until the conversation completely falls apart. That’s generally called “hallucination” but I’m not a fan of that because it implies a lot about the model that isn’t really true. Y

    You may have already read this, but if you haven’t: Steven Wolfram wrote a great overview of how GPT works that isn’t too technical. There’s also a great sci-fi novel from 2006 called Blindsight that explores the way facsimiles of intelligence can be had without consciousness or even understanding and I’ve found it to be a really interesting way to think about LLMs.

    It’s possible to build a really good Chinese room that can pass the Turing test, and I think LLMs are exactly that. More tokens/parameters/training aren’t going to change that, they’ll just make them better Chinese rooms.


  • Maybe this comment will age poorly, but I think AGI is a long way off. LLMs are a dead-end, IMO. They are easy to improve with the tech we have today and they can be very useful, so there’s a ton of hype around them. They’re also easy to build tools around, so everyone in tech is trying to get their piece of AI now.

    However, LLMs are chat interfaces to searching a large dataset, and that’s about it. Even the image generators are doing this, the dataset just happens to be visual. All of the results you get from a prompt are just queries into that data, even when you get a result that makes it seem intelligent. The model is finding a best-fit response based on billions of parameters, like a hyperdimensional regression analysis. In other words, it’s pattern-matching.

    A lot of people will say that’s intelligence, but it’s different; the LLM isn’t capable of understanding anything new, it can only generate a response from something in its training set. More parameters, better training, and larger context windows just refine the search results, they don’t make the LLM smarter.

    AGI needs something new, we aren’t going to get there with any of the approaches used today. RemindMe! 5 years to see if this aged like wine or milk.


  • Hyperfixating on producing performant code by using Rust (when you code in a very particular way) makes applications worse. Good API and system design are a lot easier when you aren’t constantly having to think about memory allocations and reference counting. Rust puts that dead-center of the developer experience with pointers/ownership/Arcs/Mutexes/etc and for most webapps it just doesn’t matter how memory is allocated. It’s cognitive load for no reason.

    The actual code running for the majority of webapps (including Lemmy) is not that complicated, you’re just applying some business logic and doing CRUD operations with datastores. It’s a lot more important to consider how your app interacts with your dependencies than how to get your business logic to be hyper-efficient. Your code is going to be waiting on network I/O and DB operations most of the time anyway.

    Hindsight is 20/20 and I’m not faulting anyone for not thinking through a personal project, but I don’t think Rust did Lemmy any favors. At the end of the day, it doesn’t matter how performant your code is if you make bad design and dependency choices. Rust makes it harder to see these bad choices because you have to spend so much time in the weeds.

    To be clear, I’m not shitting on Rust. I’ve used it for a few projects and great for apps where processing performance is important. It’s just not a good choice for most webapps, you’d be far better off in a higher-level language.


  • I wouldn’t shortchange how much making the barrier to entry lower can help. You have to fight Rust a lot to build anything complex, and that can have a chilling effect on contributions. This is not a dig at Rust; it has to force you to build things in a particular way because it has to guarantee memory safety at compile time. That isn’t to say that Rust’s approach is the only way to be sure your code is safe, mind you, just that Rust’s insistence on memory safety at compile time is constraining.

    To be frank, this isn’t necessary most of the time, and Rust will force you to spend ages worrying about problems that may not apply to your project. Java gets a bad rap but it’s second only to Python in ease-of-use. When you’re working on an API-driven webapp, you really don’t need Rust’s efficiency as much as you need a well-defined architecture that people can easily contribute to.

    I doubt it’ll magically fix everything on its own, but a combo of good contribution policies and a more approachable codebase might.



  • People from east and southeast Asia have been cultivating and eating soy beans as a staple food since before Babylon. I mean that literally; there is evidence of soy bean cultivation in what is now China from like 7000 BC.

    It’s tough to take a phrase like, “Soy makes men weak,” as anything other than racism when it puts down a quarter of the population of the planet. At best, it’s ignorance, but in my experience the people who hold this opinion don’t change their mind when you explain this to them.