I’m writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.

This has caused me to wonder what the largest storage operation you guys have done. I’ve taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.

  • Urist@lemmy.ml
    link
    fedilink
    English
    arrow-up
    48
    arrow-down
    1
    ·
    1 year ago

    I obviously downloaded a car after seeing that obnoxious anti-piracy ad.

  • Neuromancer49@midwest.social
    link
    fedilink
    English
    arrow-up
    36
    ·
    1 year ago

    In grad school I worked with MRI data (hence the username). I had to upload ~500GB to our supercomputing cluster. Somewhere around 100,000 MRI images, and wrote 20 or so different machine learning algorithms to process them. All said and done, I ended up with about 2.5TB on the supercomputer. About 500MB ended up being useful and made it into my thesis.

    Don’t stay in school, kids.

  • fuckwit_mcbumcrumble@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    29
    ·
    1 year ago

    Entire drive/array backups will probably be by far the largest file transfer anyone ever does. The biggest I’ve done was a measly 20TB over the internet which took forever.

    Outside of that the largest “file” I’ve copied was just over 1TB which was a SQL file backup for our main databases at work.

    • cbarrick@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      1 year ago

      +1

      From an order of magnitude perspective, the max is terabytes. No “normal” users are dealing with petabytes. And if you are dealing with petabytes, you’re not using some random poster’s program from reddit.

      For a concrete cap, I’d say 256 tebibytes…

  • Hugin@lemmy.world
    link
    fedilink
    arrow-up
    18
    ·
    1 year ago

    It was something around 40 TB X2 . We were doing a terrain analysis of the entire Earth. Every morning for 25 days I would install two fresh drives in the cluster doing the data crunching and migrate the filled drives to our file server rack.

    The drives were about 80% full and our primary server was mirrored to two other 50 drive servers. At the end of the month the two servers were then shipped to customer locations.

  • ramble81@lemm.ee
    link
    fedilink
    arrow-up
    18
    ·
    1 year ago

    I’ve done a 1PB sync between a pair of 8-node SAN clusters as one was being physically moved since it’d be faster to seed the data and start a delta sync rather than try to do it all over a 10Gb pipe. M

  • Trigger2_2000@sh.itjust.works
    link
    fedilink
    arrow-up
    15
    ·
    1 year ago

    I once abused an SMTP relay (my own) by emailing Novell a 400+ MB memory dump. Their FTP site kept timing out.

    After all that, and them swearing they had to have it, the OS team said “Nope, we’re not going to look at it”. Guess how I feel about Novell after that?

    This was in the mid-90’s.

  • brygphilomena@lemmy.world
    link
    fedilink
    arrow-up
    14
    ·
    1 year ago

    In the middle of something 200tb for my Plex server going from a 12 bay system to a 36 LFF system. But I’ve also literally driven servers across the desert because it was faster than trying to move data from one datacenter to another.

  • d00phy@lemmy.world
    link
    fedilink
    English
    arrow-up
    13
    ·
    1 year ago

    I’ve migrated petabytes from one GPFS file system to another. More than once, in fact. I’ve also migrated about 600TB of data from D3 tape format to 9940.

  • Decency8401@discuss.tchncs.de
    link
    fedilink
    arrow-up
    8
    ·
    1 year ago

    A few years back I worked at a home. They organised the whole data structure but needed to move to another Providor. I and my colleagues moved roughly just about 15.4 TB. I don’t know how long it took because honestly we didn’t have much to do when the data was moving so we just used the downtime for some nerd time. Nerd time in the sense that we just started gaming and doing a mini LAN party with our Raspberry and banana pi’s.

    Surprisingly the data contained information of lots of long dead people which is quiet scary because it wasn’t being deleted.

    • 🐍🩶🐢@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 year ago

      No idea about which specific type of business it is, but keeping that history long term can have some benefits, especially to outside people. Some government agencies require companies to keep records for a certain number of years. It could also help out in legal investigations many years in the future and show any auditors you keep good records. From a historical perspective, it can be matched to census, birth, and death certificates. A lot of generational history gets lost.

      Companies also just hoard data. Never know what will be useful later. shrug

  • Avid Amoeba@lemmy.ca
    link
    fedilink
    arrow-up
    8
    ·
    1 year ago

    ~15TB over the internet via 30Mbps uplink without any special considerations. Syncthing handled any and all network and power interruptions. I did a few power cable pulls myself.

    • Pasta Dental@sh.itjust.works
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      I think it’s crazy that not that long ago 30mbps was still pretty good, we now have 1gbps+ at residential addresses and it fairly common too

      • Confused_Emus@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I’ve got symmetrical gigabit in my apartment, with the option to upgrade to 5 or 8. I’d have to upgrade my equipment to use those speeds, but it’s nice to know I have the option.

      • Avid Amoeba@lemmy.ca
        link
        fedilink
        arrow-up
        1
        ·
        1 year ago

        Yeah, I also moved from 30Mb upload to 700Mb recently and it’s just insane. It’s also insane thinking I had a symmetric gigabit connection in Eastern Europe in the 2000s for fairly cheap. It was Ethernet though, not fiber. Patch cables and switches all the way to the central office. 🫠

        Most people in Canada today have 50Mb upload at the most expensive connection tiers - on DOCSIS 3.x. Only over the last few years fiber began becoming more common but it’s still fairly uncommon as it’s the most expensive connection tier if at all available.

        • Pasta Dental@sh.itjust.works
          link
          fedilink
          arrow-up
          2
          ·
          1 year ago

          We might pay some of the most expensive internet in the world in Canada but at least we can’t fault them for providing an unstable or unperformqnt service. Download llama models is where 1gbps really shines, you see a 7GB model? It’s done before you are even back from the toilet. Crazy times.

          • Avid Amoeba@lemmy.ca
            link
            fedilink
            arrow-up
            2
            ·
            1 year ago

            I should have know that the person on the internet noting 30Mbps was pretty good till recently is a fellow Canadian. 🍁 #ROBeLUS

            BTW, TekSavvy recently started offering fiber seemingly on Bell’s last mile.

    • pete_the_cat@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      How long did that take? A month or two? I’ve backfilled my NAS with about 40 TB before over a 1 gig fiber pipe in about a week or so of 24/7 downloading.

  • HarriPotero@lemmy.world
    link
    fedilink
    arrow-up
    7
    ·
    1 year ago

    I worked at a niche factory some 20 years ago. We had a tape robot with 8 tapes at some 200GB each. It’d do a full backup of everyone’s home directories and mailboxes every week, and incremental backups nightly.

    We’d keep the weekly backups on-site in a safe. Once a month I’d do a run to another plant one town over with a full backup.

    I guess at most we’d need five tapes. If they still use it, and with modern tapes, it should scale nicely. Today’s LTO-tapes are 18TB. Driving five tapes half an hour would give a nice bandwidth of 50GB/s. The bottleneck would be the write speed to tape at 400MB/s.

  • IphtashuFitz@lemmy.world
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 year ago

    Back in the late 90’s I worked for an internet search company, long before Google was a thing. We would regularly physically drive a dozen SCSI drives from a RAID array between two datacenters about 20 miles apart.