this post was submitted on 18 Aug 2024
220 points (97.4% liked)

Linux

48143 readers
837 users here now

From Wikipedia, the free encyclopedia

Linux is a family of open source Unix-like operating systems based on the Linux kernel, an operating system kernel first released on September 17, 1991 by Linus Torvalds. Linux is typically packaged in a Linux distribution (or distro for short).

Distributions include the Linux kernel and supporting system software and libraries, many of which are provided by the GNU Project. Many Linux distributions use the word "Linux" in their name, but the Free Software Foundation uses the name GNU/Linux to emphasize the importance of GNU software, causing some controversy.

Rules

Related Communities

Community icon by Alpár-Etele Méder, licensed under CC BY 3.0

founded 5 years ago
MODERATORS
 

I'm writing a program that wraps around dd to try and warn you if you are doing anything stupid. I have thus been giving the man page a good read. While doing this, I noticed that dd supported all the way up to Quettabytes, a unit orders of magnitude larger than all the data on the entire internet.

This has caused me to wonder what the largest storage operation you guys have done. I've taken a couple images of hard drives that were a single terabyte large, but I was wondering if the sysadmins among you have had to do something with e.g a giant RAID 10 array.

(page 3) 42 comments
sorted by: hot top controversial new old
[–] [email protected] 7 points 2 months ago

I think 16 terabytes? Might have been twelve. I was consolidating a bunch of old drives and data into a nas for a friend. He just didn't have the time, between working and school and brought me all the hardware and said "go" lol.

[–] [email protected] 32 points 2 months ago (1 children)

I’ve done a 1PB sync between a pair of 8-node SAN clusters as one was being physically moved since it’d be faster to seed the data and start a delta sync rather than try to do it all over a 10Gb pipe. M

[–] [email protected] 9 points 2 months ago

I'm in the same boat, just under 3PiB

[–] [email protected] 6 points 2 months ago

I transferred my entire NAS storage, which includes all of my backups, cloud files, my family’s backups, and my… Linux ISOs. That was about 12TB.

[–] [email protected] 5 points 2 months ago

I did 100TB, 100 streams of 1TB, all simultaneous with rsync

[–] [email protected] 21 points 2 months ago (2 children)

I once deleted an 800 gb log file, does that count

load more comments (2 replies)
[–] [email protected] 61 points 2 months ago

I obviously downloaded a car after seeing that obnoxious anti-piracy ad.

[–] [email protected] 4 points 2 months ago

I once robocopied 16tb of media

[–] [email protected] 7 points 2 months ago

I'll let you know... If it finishes.

[–] [email protected] 18 points 2 months ago

I’ve migrated petabytes from one GPFS file system to another. More than once, in fact. I’ve also migrated about 600TB of data from D3 tape format to 9940.

[–] [email protected] 5 points 2 months ago

Rsynced 4.2TB of data from one server to another but with multiple files

[–] [email protected] 4 points 2 months ago* (last edited 2 months ago)

I'm currently in the process of transferring about 50 TB from one zpool to another (locally), so I can destroy and recreate it.

I've downloaded a few torrents that were around 5 TB each, they're PS4 and Xbox 360 game collections.

[–] [email protected] 7 points 2 months ago

Today I've migrated my data from my old zfs pool to a new bigger one, the rsync of 13.5TiB took roughly 18 hours. It's slow spinning disks storage so that's fine.

The second and third runs of the same rsync took like 5 seconds, blazing fast.

[–] [email protected] 4 points 2 months ago* (last edited 2 months ago)

Local file transfer?

I cloned a 1TB+ system a couple of times.

As the Anaconda installer of Fedora Atomic is broken (yes, ironic) I have one system originally meant for tweaking as my "zygote" and just clone, resize, balance and rebase that for new systems.

Remote? 10GB MicroWin 11 LTSC IOT ISO, the least garbage that OS can get.

[–] [email protected] 47 points 2 months ago (1 children)

In grad school I worked with MRI data (hence the username). I had to upload ~500GB to our supercomputing cluster. Somewhere around 100,000 MRI images, and wrote 20 or so different machine learning algorithms to process them. All said and done, I ended up with about 2.5TB on the supercomputer. About 500MB ended up being useful and made it into my thesis.

Don't stay in school, kids.

[–] [email protected] 18 points 2 months ago (1 children)

You should have said no to math, it's a helluva drug

[–] [email protected] 3 points 2 months ago

I routinely do 1-4TB images of SSDs before making major changes to the disk. Run fstrim on all partitions and pipe dd output through zstd before writing to disk and they shrink to actually used size or a bit smaller. Largest ever backup was probably ~20T cloned from one array to another over 40/56GbE, the deltas after that were tiny by comparison.

[–] [email protected] 4 points 2 months ago

You should ping CERN or Fermilab about this. Or maybe the Event Horizon Telescope team but I think they used sneakernet to image the M87 black hole.

Anyway, my answer is probably just a SQL backup like everyone else.

[–] [email protected] 5 points 2 months ago

Around 15 TB migrating to a new NAS.

[–] [email protected] 7 points 2 months ago

Largest one I ever did was around 4.something TB. New off-site backup server at a friends place. Took me 4 months due to data limits and an upload speed that maxed out at 3MB/s.

[–] [email protected] 3 points 2 months ago

As a single file? Likely 20GB iso.
As a collective job, 3TB of videos between hard drives for Jellyfin.

[–] [email protected] 6 points 2 months ago

i've transferred 10's of ~300 GB files via manual rsyncs. it was a lot of binary astrophysical data, most of which was noise. eventually this was replaced by an automated service that bypassed local firewalls with internet-based transfers and aws stuff.

[–] [email protected] 82 points 2 months ago (3 children)

Not that big by today's standards, but I once downloaded the Windows 98 beta CD from a friend over dialup, 33.6k at best. Took about a week as I recall.

[–] [email protected] 31 points 2 months ago (2 children)

I remember downloading the scene on American Pie where Shannon Elizabeth strips naked over our 33.6 link and it took like an hour, at an amazing resolution of like 240p for a two minute clip 😂

[–] [email protected] 13 points 2 months ago

And then you busted after 15 seconds?

load more comments (1 replies)
[–] [email protected] 17 points 2 months ago

Yep, downloaded XP over 33.6k modem, but I'm in NZ so 33.6 was more advertising than reality, it took weeks.

load more comments (1 replies)
[–] [email protected] 3 points 2 months ago (2 children)

I mean dd claims they can handle a quettabyte but how can we but sure.

load more comments (2 replies)
[–] [email protected] 7 points 2 months ago

Currently pushing about 3-5 TB of images to AI/ML scanning per day. Max we've seen through the system is about 8 TB.

Individual file? Probably 660 GB of backups before a migration at a previous job.

[–] [email protected] 10 points 2 months ago* (last edited 2 months ago) (7 children)

~15TB over the internet via 30Mbps uplink without any special considerations. Syncthing handled any and all network and power interruptions. I did a few power cable pulls myself.

[–] [email protected] 3 points 2 months ago (1 children)

How long did that take? A month or two? I've backfilled my NAS with about 40 TB before over a 1 gig fiber pipe in about a week or so of 24/7 downloading.

[–] [email protected] 3 points 2 months ago

Yeah, something like that. I verified it it with rsync after that, no errors.

load more comments (6 replies)
[–] [email protected] 7 points 2 months ago

We have DBs in the dozens of TB at work so probably one of them

[–] [email protected] 15 points 2 months ago* (last edited 2 months ago) (1 children)

I don't remember how many files, but typically these geophysical recordings clock in at 10-30 GB. What I do remember, though, was the total transfer size: 4TB. It was kind of like a bunch of .segd, and they were stored in this server cluster that was mounted in a shipping container for easy transport and lifting onboard survey ships. Some geophysics processors needed it on the other side of the world. There were nobody physically heading in the same direction as the transfer, so we figured it would just be easier to rsync it over 4G. It took a little over a week to transfer.

Normally when we have transfers of a substantial size going far, we ship it on LTO. For short distance transfers we usually run a fiber, and I have no idea how big the largest transfer job has been that way. Must be in the hundreds of TB. The entire cluster is 1.2PB, bit I can't recall ever having to transfer everything in one go, as the receiving end usually has a lot less space.

[–] [email protected] 6 points 2 months ago (1 children)

4G?! That strikes fear into my heart!

[–] [email protected] 11 points 2 months ago (1 children)

The alternative was 5mbit/s VSAT. 4G was a luxury at that time.

load more comments (1 replies)
[–] [email protected] 41 points 2 months ago (3 children)

Entire drive/array backups will probably be by far the largest file transfer anyone ever does. The biggest I've done was a measly 20TB over the internet which took forever.

Outside of that the largest "file" I've copied was just over 1TB which was a SQL file backup for our main databases at work.

[–] [email protected] 9 points 2 months ago

+1

From an order of magnitude perspective, the max is terabytes. No "normal" users are dealing with petabytes. And if you are dealing with petabytes, you're not using some random poster's program from reddit.

For a concrete cap, I'd say 256 tebibytes...

load more comments (2 replies)
[–] [email protected] 12 points 2 months ago (1 children)

@data1701d downloading forza horizon 5 on Steam with around 120gb is the largest web-download, I can remember. In LAN, I’ve migrated my old FreeBSD NAS to my new one, which was a roughly 35TB transfer over NFS.

[–] [email protected] 5 points 2 months ago

How long did that 35TB take? 12 hours or so?

load more comments
view more: ‹ prev next ›