Note: Unfortunately the research paper linked in the article is a dead/broken/wrong link. Perhaps the author will update it later.

From the limited coverage, it doesn’t sound like there’s an actual optical drive that utilizes this yet and that it’s just theoretical based on the properties of the material the researchers developed.

I’m not holding my breath, but I would absolutely love to be able to back up my storage system to a single optical disc (even if tens of TBs go unused).

If they could make a R/W version of that, holy crap.

  • Flumpkin@slrpnk.net
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 months ago

    Then comes the discussion if “1KB” is 1024 bytes or if 1000 bytes is a kilobyte.

    It’s the metric system and it’s standard now. 1 kilobyte is 1000 bytes, just like 1 kilometer is 1000 meters. It is much easier to convert 20.415.823 bytes into megabytes - 23.4 MB.

    Only windows insists on mislabeling the base 1024 kibibyte as kilobyte. The metric unit is much easier to use.

    • mb_@lemm.ee
      link
      fedilink
      English
      arrow-up
      3
      ·
      10 months ago

      What? Every BIOS in the world still uses the same system. Same thing for me on Linux.

      Only hard driver manufacturers used a different system to inflate their numbers and pushed a market campaign, a lot of people who didn’t even use computers said “oh that makes sense - approved”

      People who actually work with computer, memory, CPU, and other components in base 8 just ignores this non-sense of “x1000”

    • DacoTaco@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      10 months ago

      I never knew the whole thing was considered part of the metric system, makes sense though.
      I love the metric system to death because its so simple and easy, and it links different measurements together ( 1l of water = 1kg etc ).

      That said, a computer works differently and because we work in factors of 2, 1000 bytes being a kilobyte makes no sense once you start working with bits and low level stuff. Other than that, i can see why the stuff was redefined.

      Also, i think linux also works in factors of 1024, but id need to check

      • Flumpkin@slrpnk.net
        link
        fedilink
        English
        arrow-up
        1
        ·
        10 months ago

        There is nothing to keep you from using factors of 1024 (except he slightly ludicrous prefix “kibi” and “mebi”), but other than low level stuff like disc sectors or bios where you might want to use bit logic instead of division it’s rather rare. I too started in the time when division op was more costly than bit level logic.

        I’d argue that any user facing applications are better off with base 1000, except by convention. Like a majority of users don’t know or care or need to care what bits or bytes do. It’s programmers that like the beauty of the bit logic, not users. @mb_@lemm.ee

        • DacoTaco@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          I agree with what you said, and its imo why the discussion of a factor of 1000 and 1024 will always rage on. Im a developer, and do embedded stuff in my free time. Everything around me is factor 1024 because of it, and i hate the factor 1000. But from a generic user standpoint, i agree its a lot more user friendly, as they are used to the metric system of a factor of 10

          • mb_@lemm.ee
            link
            fedilink
            English
            arrow-up
            1
            ·
            10 months ago

            It is user friendly, and technically incorrect, since nothing ever lines up with reality when you use 1000 because the underlying system is base 8.

            Or you get the weird non-sense all around “my computer has 18.8gb of memory”…