Resizing ext4 filesystem > 16TB

Discussion in 'SSDs & Data Storage' started by apekatt, Aug 30, 2011.

  1. apekatt

    apekatt n00bie

    Messages:
    16
    Joined:
    Jul 31, 2009
    I have file server that I've kept adding disks to for many years now, and I recently hit a wall when I reached 16TB. The problem from what I can find out is that resize2fs cant resize filesystems past 16TB using ext4. I've been googling a bit and it seems the latest version of resize2fs from 07-02-11 still doesn't have support for this, and it is a WIP ( See screenshot ).

    Now I'm wondering what my options are.

    1. Find a workaround so I can resize me existing filesystem. ( Wait for an update to resize2fs. )
    2. Convert to another filesystem.
    3. Make another LVM Volume.
    4. Reformat the filesystem with another filesystem.

    I would prefer to keep the system as one big volume, and not split it in two.
    Alternatives 1 or 2 would be preferable.

    [​IMG]
     
  2. rmdashrrootsplat

    rmdashrrootsplat Gawd

    Messages:
    924
    Joined:
    Apr 2, 2008
    I think the best option would be to look at other filesystems such as xfs or zfs... Unfortunately that means backing up everything and reformatting.

    e2fsprogs still does not support >16TB officially and any of the beta releases are buggy as hell. Not something I would trust with my data. It's been a known issue for years so I wouldn't wait for them to release a stable version. ;)
     
  3. danswartz

    danswartz 2[H]4U

    Messages:
    3,579
    Joined:
    Feb 25, 2011
    Might want to take a look at zfs. Supposedly the zfs on linux project is stable now. If not, you could maybe move to OpenIndiana or SolarisExpress11 with zfs.
     
  4. praetorian

    praetorian [H]Lite

    Messages:
    84
    Joined:
    Sep 9, 2003
    Anyway you look at it, at present you're going to need to look at option 4 at the moment unless you have the ability to wait until they have the issues with EXT4 sorted out and proven.

    If you can't I'd recommend JFS as a suitable filesystem as its often overlooked down to it not being actively developed by IBM. The reason however for that is different from the norm however, as from what I understand, it is that they don't need to add any further features but are still very much in the development game with it.

    Houkouonchi uses it on this very large arrays with no issues and I also do on my (rather small) 18tb RAID-6 array and I've never had any problems with it so far. A few others also do, the most read about example being Backblaze on their Petabyte storage platform.
     
  5. BobbyTables

    BobbyTables n00bie

    Messages:
    20
    Joined:
    Jun 6, 2011
    Another alternative to reformatting would be to convert to Btrfs.

    Although with that option you will still need to wait for the fsck tool, which should be available very soon, possibly sometime in September.
     
  6. apekatt

    apekatt n00bie

    Messages:
    16
    Joined:
    Jul 31, 2009
    Yeah, I thought about converting to Btrfs, but not sure if it's stable enough. Read some reports of people loosing data after power failures etc.

    Maybe I'll just wait and see what comes first; Btrfs fschk or updated version of e2fsprogs.
     
  7. parityboy

    parityboy Limp Gawd

    Messages:
    364
    Joined:
    Nov 13, 2010
    Another option would to create another LVM and then pool the two volumes using GreyHole. This will present a single share over the network, but obviously they will still be two separate volumes on the server itself.

    Since (I assume) your PVs are in fact RAID devices rather than individual drives, you can set GreyHole to zero redundancy and just use it to pool your storage. I'd do this as a temporary fix until e2fsprogs 1.42 is released, or you find a way to back up your data and move to another filesystem.

    I too would recommend ZFS on Linux; KQ Infotech look to have done a very good job.
     
  8. houkouonchi

    houkouonchi RIP

    Messages:
    1,624
    Joined:
    Sep 14, 2008
    only real two options IMHO for >16 TiB support is JFS or XFS. I have really been bitten badly by XFS on mutliple things in the past. I hear its better but JFS has been extremely reliable for me and it should work for file-systems up to 512 TiB.
     
  9. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
  10. drescherjm

    drescherjm [H]ardForum Junkie

    Messages:
    13,628
    Joined:
    Nov 19, 2008
    You have to wait for a 3.2.X kernel however.

     
  11. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
    I did a quick Google search after reading this. It seems the 3.2 kernel is in RC7 and they said that's the last one before release. So hopefully once the repos get it, it will be as simple as upgrading all required software and have at it. Or I hope.
     
  12. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
  13. drescherjm

    drescherjm [H]ardForum Junkie

    Messages:
    13,628
    Joined:
    Nov 19, 2008
    Did 3.2.0 give you > 16TB ext4 filesystems? I already have 3.2.1 on several of my servers here at work.
     
  14. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
    I don't have enough hard drive space to check. I am mainly doing this in preparation for once I get there.

    Another thing I did was update E2fsprogs to the latest 1.42 from directions from this site.


    http://blog.ronnyegner-consulting.de/2011/08/18/ext4-and-the-16-tb-limit-now-solved/

    Code:
    [root@Mega-Nas build]# tune2fs
    tune2fs 1.42 (29-Nov-2011)
    Usage: tune2fs [-c max_mounts_count] [-e errors_behavior] [-g group]
            [-i interval[d|m|w]] [-j] [-J journal_options] [-l]
            [-m reserved_blocks_percent] [-o [^]mount_options[,...]] [-p mmp_update_interval]
            [-r reserved_blocks_count] [-u user] [-C mount_count] [-L volume_label]
            [-M last_mounted_dir] [-O [^]feature[,...]]
            [-E extended-option[,...]] [-T last_check_time] [-U UUID]
            [ -I new_inode_size ] device
    With both of these installed, I think I should be able to create a >16tb file system, just not resize one yet.

    So my thought would be once E2fsprogs is updated to allow for resizing above 16tb, I,We should be ready to go, no?
     
  15. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
  16. apekatt

    apekatt n00bie

    Messages:
    16
    Joined:
    Jul 31, 2009
    Just to update. I have been testing a bit with new kernel, and I still cant resize it past 16TB.

    I can create a new file system bigger than 16tb, but once I try to resize it again it fails.

    This is my test system in Vmware. Just added a bunch of disks with thin provisioning to test:

    [​IMG]

    Any suggestions? Can I do something with tune2fs to make it work?
     
  17. danswartz

    danswartz 2[H]4U

    Messages:
    3,579
    Joined:
    Feb 25, 2011
    Ugh, you're already way out of my comfort zone with the bleeding edge stuff. I'd switch to a FS that supports it better, but that's just me...
     
  18. Silhouette

    Silhouette Limp Gawd

    Messages:
    205
    Joined:
    Dec 14, 2006
    I agree, the problems you are seeing is an indication that very few people are using ext4 on volumes that large. I'd be very hesitant to use resize2fs for a while after they fix the 16TB limit.
     
  19. asgards

    asgards Limp Gawd

    Messages:
    204
    Joined:
    May 8, 2008
    had this very same issue, 13x1,5 r6 expanded to 18x w/ no problem, but exty failed on me, thankfully i had 5empty drives, so it took a week but im on jfs now, problem fixed :D
     
  20. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
    Once E2fsprogs is updated it will be able to resize above 16tb.

    http://e2fsprogs.sourceforge.net/e2fsprogs-release.html#1.42
     
  21. drescherjm

    drescherjm [H]ardForum Junkie

    Messages:
    13,628
    Joined:
    Nov 19, 2008
    It does not look like the update made it into 3.2. Here is what I see from the changes related to ext4:

    Source:
    http://kernelnewbies.org/LinuxChanges#head-e3ffec0bbdab939b186cd5ec73a4eb40a5a766d0
     
  22. misterpat

    misterpat [H]Lite

    Messages:
    103
    Joined:
    Dec 24, 2011
    Ok, I finally have enough drives to put me above 16tb. I started researching it some more and found this.

    http://www.phoronix.com/scan.php?page=news_item&px=MTA0MDg

    So I updated my system, Debian Wheezy, to the 3.3 kernel. Im also running e2fsprogs 1.42.1-2 which should support online resizing.

    But when I tried to resize from 12tb to over 16tb, Im getting the famous message.

    Code:
    root@mega:~# resize2fs /dev/md0
    resize2fs 1.42.1 (17-Feb-2012)
    resize2fs: New size too large to be expressed in 32 bits
    Upon further research, I found this.

    http://ubuntuforums.org/showpost.php?p=10995759&postcount=6

    So, it seems that if I back up my data, wipe the drives, make a new file system, we should be good to go?
     
  23. robinmdh

    robinmdh n00bie

    Messages:
    2
    Joined:
    Mar 25, 2012
    I found this thread quite useful so I decided to share my experiences as well.

    I'd just like to mention here that I found that ext4 resizing to larger then 16TB did not work with kernel 3.2.0 and e2fsprogs 1.42. either online or offline.
    and while this may have been possible with the 3.3.0 kernel I'm not comfortable putting my data that close to beta versions of a kernel.

    but btrfs-convert works fine and then a
    "btrfs filesystem resize max /mnt"
    appears to have done the resizing fine.

    so far this works fine, I am leaving the ext4 snapshot in place for a bit though.
    (btrfs-convert can do a roll back, though if you use checksums(default) it'll take a long time to do the conversion and that is offline only.
    with btrfs on top of (sw) raid6 checksums seemed a bit redundant as well.

    just my 2 cents.

    I realise btrfs is still a little bit in beta but with the roll-back option i can presumably go back to what I had before, so no data loss except for the new data...

    it's like having a snapshot only of the old FS as well as of the old files...
    the alternative (for me)was getting a couple of drives out of the raid setup (re-syncing for each drive) and then running user-space ZFS on the freed drives and copying chunks there and grow the zfs as it goes on etc, I chose to be lazy...
    of course my limitations may be that I don't have a place to back up 10TB of data etc... if I'd have that option I would take it and run on ZFS from then on..

    the net result is i have my bigger FS and still have all my data without all the trouble it would have taken to get that.
     
    Last edited: Mar 26, 2012
  24. brutalizer

    brutalizer [H]ard|Gawd

    Messages:
    1,588
    Joined:
    Oct 23, 2010
    I would be careful to trust btrfs in some years. Last month, it corrupts and loose all data, new and old:
    http://phoronix.com/forums/showthre...ion-Ready-In-Oracle-Linux&p=249731#post249731

    "...After this system was completely corrupted and it is not recoverable.
    ....
    I think Oracle need to spend couple more years before [BTRFS] will work...."
     
  25. drescherjm

    drescherjm [H]ardForum Junkie

    Messages:
    13,628
    Joined:
    Nov 19, 2008