• SomeoneSomewhere@lemmy.nz
      link
      fedilink
      English
      arrow-up
      30
      ·
      1 day ago

      RAM’s main advantage over HDDs/SSDs is fast access times.

      Needing to fetch anything over the internet would make it faster to just use HDDs.

      • PabloSexcrowbar@piefed.social
        link
        fedilink
        English
        arrow-up
        8
        ·
        1 day ago

        Theoretically, you could do whatever processing you need using the user’s CPU and RAM and then send the result back over the Internet. Not saying that’s what’s happening, of course, but it’s not completely ridiculous.

      • pelespirit@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        3
        ·
        1 day ago

        But it isn’t this idea just kind of a reverse cloud? Since running AI is so expensive, they could “borrow” other people’s ram. Just an idea.

        • Zorque@lemmy.world
          link
          fedilink
          English
          arrow-up
          11
          ·
          1 day ago

          Sounds like you’re looking for zebras when horses are a much simpler explanation.

        • FauxLiving@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          ·
          edit-2
          23 hours ago

          Conceptually what you’re describing is feasible, there’s lots of distributed computing projects that borrow compute/space/bandwidth for their ends but is unlikely to have any practical use.

          If there were a distributed system that could be used as memory in a large virtual inferencing machine, it would be incredibly slow. The model would be stored across a large number of different computers which would all have to coordinate. Each step of inferencing would be orders of magnitude slower because the latency between two different computers is orders of magnitude slower than the latency between a GPU and physical RAM.

          On the other end, if we just assume inferencing is feasible in reasonable time through some technique that isn’t public… a model that was large enough to take advantage of an Internet worth of memory doesn’t exist.

          So, assuming you had access to the largest model that we know of, you would have a model as complex as Claude Opos, but would take hours or days to respond finish inferencing and the quality would be about the same as you could get in under a second for $20/mo.

          And, going with a hypothetical ‘Internet Scale’ model.

          First, it would have to be trained which would use take an incredibly long time. Some of these frontier models take months to train on the fastest hardware available, a larger model would take even longer to train due to the increased latency.

          More importantly, there are strong diminishing returns on capability vs model size. This is why the AI companies are focusing on agentic tasks, where the AI spends a lot of time talking to itself and using tools, rather than pushing for a model with more parameters. This is referred to as the “scaling wall” (though AI companies, for obvious reasons, deny that such a thing exists and smoking companies say there’s no cancer risk for smokers).

          It’s a neat idea (Skynet may be loose on the world, hiding out as widespread ‘bugs’ that happen to consume a lot of resources and compute), but it would require a lot of things to be magic’d into existence to be remotely practical.

          You may find this funny: https://youtu.be/JcJSW7Rprio

    • JohnnyCanuck@lemmy.ca
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      2
      ·
      1 day ago

      Does a thing like crowd-sourcing ram work?

      No.

      Is it a thing?

      No.

      This would probably be the symptoms though, yeah?

      No.

      You seem very confused about what RAM is and what’s happening here. You seem to think that RAM is something you make on your computer. It’s a physical part of your computer that you load information into.

      Imagine you’re sitting at a desk in an office. The desk has little shelves where you can put documents you’re working on. You can only put a small number of files there. The office has filing cabinets where other files are kept that you’re not working on. You can store a lot in there but it takes time to go find it. You also have some special filing cabinets that are still slow but you only use it to store files temporarily that someone brings you from another office, or when you run out of space on your desk but still need to keep files handy.

      In this analogy, the shelves on the desk is RAM. You only put the stuff you’re immediately working on in those shelves because of the limited space, but it’s really fast to find stuff compared to the filling cabinets, which are your hard drive. When you go on a website, like YouTube, you’re calling someone in an office in another building and asking them for some files. They send over a bunch of files, which takes a really long time. You put a much as possible in your desk shelves to use right now, but anything that doesn’t fit you put in one of those special filing cabinets, which will call the cache, which is slow, but not nearly as slow as waiting for the files to come from the other office. When you’re ready for the extra files from YouTube, you just grab them from the cache.

      What’s happening in this problem with youtube is that you request the files from them, they send them over, along with instructions on how to use them. The instructions say something that requires putting a bunch of things in RAM. At first this is normal. But at some point the instructions start repeating and tell you to put more and more files into RAM, maybe even repeats of files you already have there, shouldn’t need again. But you just follow instructions, that’s your job. So you keep loading things into RAM, but then there’s no room left and your system falls apart and you can no longer do any work. Until you close youtube and chuck all the youtube files out of RAM.

      Hopefully that makes it clear why you can’t outsource RAM. Essentially you would be putting your little desk shelves in a different office, but we already have a better solution than that: the cache or special local filing cabinet on your hard drive.

      What we outsource normally is the hard drive (filing cabinets) and call it cloud storage (for example), and the creation and processing of information (done by the CPU, GPU, or other chips on your computer) and call it cloud computing (for example). That’s because those things are slow, and the extra time to move the files between offices isn’t necessarily the bottleneck.