Basically a deer with a human face. Despite probably being some sort of magical nature spirit, his interests are primarily in technology and politics and science fiction.

Spent many years on Reddit before joining the Threadiverse as well.

  • 0 Posts
  • 926 Comments
Joined 1 year ago
cake
Cake day: March 3rd, 2024

help-circle

  • FaceDeer@fedia.iotoSelfhosted@lemmy.worldIPv6 for self hosters
    link
    fedilink
    arrow-up
    41
    arrow-down
    1
    ·
    1 day ago

    You may know IPv6 is ridiculously bigger, but you don’t know it.

    There are enough IPv6 addresses that you could give 10^17 addresses to every square millimeter of Earth’s surface. Or 5×10^28 addresses for every living human being. On a more cosmic scale, you could issue 4×10^15 addresses to every star in the observable universe.

    We’re not going to run out by giving them to lightbulbs.





  • It’s a bit technical, I haven’t found any pre-packaged software to do what I’m doing yet.

    First I installed https://github.com/openai/whisper , the speech-to-text model that OpenAI released back when they were less blinded by dollar signs. I wrote a Python script that used it to go through all of the audio files in the directory tree where I’m storing this stuff and produced a transcript that I stored in a .json file alongside it.

    For the LLM, I installed https://github.com/LostRuins/koboldcpp/releases/ and used the https://huggingface.co/unsloth/Qwen3-30B-A3B-128K-GGUF model, which is just barely small enough to run smoothly on my RTX 4090. I wrote another Python script that methodically goes through those .json files that Whisper produced, takes the raw text of the transcript, and feeds it to the LLM with a couple of prompts explaining what the transcript is and what I’d like the LLM to do with it (write a summary, or write a bullet-point list of subject tags). Those get saved in the .json file too.

    Most recently I’ve been experimenting with creating an index of the transcripts using those LLM results and the Whoosh library in Python, so that I can do local searches of the transcripts based on topics. I’m building towards writing up something where I can literally tell it “Tell me about Uncle Pete” and it’ll first search for the relevant transcripts and then feed those into the LLM with a prompt to extract the relevant information from them.

    If you don’t find the idea of writing scripts for that sort of thing literally fun (like me) then you may need to wait a bit for someone more capable and more focused than I am to create a user-friendly application to do all this. In the meantime, though, hoard that data. Storage is cheap.



  • Bear in mind, though, that the technology for dealing with these things are rapidly advancing.

    I have an enormous amount of digital archives I’ve collected both from myself and from my now-deceased father. For years I just kept them stashed away. But about a year ago I downloaded the Whisper speech-to-text model from OpenAI and transcribed everything with audio into text form. I now have a Qwen3 LLM in the process of churning through all of those transcripts writing summaries of their contents and tagging them based on subject matter. I expect pretty soon I’ll have something with good enough image recognition that I can turn loose on the piles of photographs to get those sorted out by subject matter too. Eventually I’ll be able to tell my computer “give me a brief biography of Uncle Pete” and get something pretty good out of all that.

    Yeah, boo AI, hallucinations, and so forth. This project has given me first-hand experience with what they’re currently capable of and it’s quite a lot. I’d be able to do a ton more if I wasn’t restricting myself to what can run on my local GPU. Give it a few more years.






  • So how are languages lost, and what does that mean for the people who speak them?

    If the language stops being spoken then there are no people who speak them, and asking what something means for those nonexistent people is kind of weird.

    I’m thinking that the loss of distinct languages in active use is not necessarily a bad thing overall. It means more people can communicate with each other more widely. By all means document these disappearing languages as much as possible before they’re gone, but there’s likely a good reason most of them are disappearing.






  • Elon Musk decided they absolutely would not use lidar, years ago when lidar was expensive enough that a decision like that made economic sense to at least try making work. Nowadays lidar is a lot cheaper but for whatever reason Musk has drawn a line in the sand and refuses to back down on it.

    Unlike many people online these days I don’t believe that Musk is some kind of sheer-luck bought-his-way-into-success grifter, he has been genuinely involved in many of the decisions that made his companies grow. But this is one of the downsides of that (Cybertruck is another). He’s forced through ideas that turned out to be amazing, but he’s also forced through ideas that sucked. He seems to be increasingly having trouble distinguishing them.