NETGEAR is aware of a growing number of phone and online scams. To learn how to stay safe click here.

Forum Discussion

HighOctane's avatar
HighOctane
Aspirant
Mar 30, 2025

Remove Inactive Volumes to sue the disk - RN316

Reading this forum, it sounds like this is not an uncommon situation... 

I recently had a hard drive fail on my RN316 - replaced it, knowing another was expected to fail soon, and about 2 hours after the resync of the first drive completed, the second drive failed.

 

When I replaced that drive, it did not automatically resync.  When I finally managed to get the resync to happen, I was presented with the attached image in the Volumes page - warnings about removing inactive volumes to use disks, and the red drives...

I have managed to assemble the arrays /dev/md127 and /dev/md126 - however btrfs is giving me the following when I attempt to mount /dev/md127

 

mount: wrong fs type, bad option, bad superblock on /dev/md127,
missing codepage or helper program, or other error

In some cases useful info is found in syslog - try
dmesg | tail or so.

 

Does this sound like it's salvageable by a rebuild of the NAS, or should I be handing the whole thing over to a data recovery expert?

 

(unfortunately, I have no backup - I was naive and trusted RAID5 with the bulk of the data on this NAS - I won't be doing that again)

19 Replies

  • StephenB's avatar
    StephenB
    Guru - Experienced User

    Data Recovery is probably the safest path.

     

    I suspect there might be a third RAID group given your mix of disks.  Check the partitions on one of the 12 TB disks.

     

     

    Did you do a btrfs device scan before you tried to mount md127

     

    Can you post mdstat.log?

    • HighOctane's avatar
      HighOctane
      Aspirant

      Sorry for not posting this one earlier StephenB - here's mdstat.log

       

      Personalities : [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
      md126 : active raid5 sdd4[0] sdf4[3]
      11718576384 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/2] [UU_]

      md127 : active raid5 sdb3[0] sda3[6] sdf3[7] sde3[3] sdd3[2] sdc3[1]
      19510827520 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6] [UUUUUU]

      md1 : active raid10 sda2[0] sdf2[5] sde2[4] sdd2[3] sdc2[2] sdb2[1]
      1566720 blocks super 1.2 512K chunks 2 near-copies [6/6] [UUUUUU]

      md0 : active raid1 sdb1[6] sdc1[5] sdd1[7] sda1[8] sde1[1]
      4190208 blocks super 1.2 [5/5] [UUUUU]

      unused devices: <none>
      /dev/md/0:
      Version : 1.2
      Creation Time : Sat Jan 3 12:35:26 2015
      Raid Level : raid1
      Array Size : 4190208 (4.00 GiB 4.29 GB)
      Used Dev Size : 4190208 (4.00 GiB 4.29 GB)
      Raid Devices : 5
      Total Devices : 5
      Persistence : Superblock is persistent

      Update Time : Thu Mar 27 06:32:40 2025
      State : clean
      Active Devices : 5
      Working Devices : 5
      Failed Devices : 0
      Spare Devices : 0

      Consistency Policy : unknown

      Name : 43f67c14:0 (local to host 43f67c14)
      UUID : e6e3fdb5:875ced54:ab63493a:b1f7bff3
      Events : 30662

      Number Major Minor RaidDevice State
      6 8 17 0 active sync /dev/sdb1
      1 8 65 1 active sync /dev/sde1
      8 8 1 2 active sync /dev/sda1
      7 8 49 3 active sync /dev/sdd1
      5 8 33 4 active sync /dev/sdc1
      /dev/md/1:
      Version : 1.2
      Creation Time : Mon Mar 24 18:05:45 2025
      Raid Level : raid10
      Array Size : 1566720 (1530.00 MiB 1604.32 MB)
      Used Dev Size : 522240 (510.00 MiB 534.77 MB)
      Raid Devices : 6
      Total Devices : 6
      Persistence : Superblock is persistent

      Update Time : Tue Mar 25 17:39:40 2025
      State : clean
      Active Devices : 6
      Working Devices : 6
      Failed Devices : 0
      Spare Devices : 0

      Layout : near=2
      Chunk Size : 512K

      Consistency Policy : unknown

      Name : 43f67c14:1 (local to host 43f67c14)
      UUID : ca988b5d:26138aec:9aff2560:000789c7
      Events : 19

      Number Major Minor RaidDevice State
      0 8 2 0 active sync set-A /dev/sda2
      1 8 18 1 active sync set-B /dev/sdb2
      2 8 34 2 active sync set-A /dev/sdc2
      3 8 50 3 active sync set-B /dev/sdd2
      4 8 66 4 active sync set-A /dev/sde2
      5 8 82 5 active sync set-B /dev/sdf2
      /dev/md/127:
      Version : 1.2
      Creation Time : Sun Mar 23 17:37:41 2025
      Raid Level : raid5
      Array Size : 19510827520 (18606.98 GiB 19979.09 GB)
      Used Dev Size : 3902165504 (3721.40 GiB 3995.82 GB)
      Raid Devices : 6
      Total Devices : 6
      Persistence : Superblock is persistent

      Update Time : Tue Mar 25 22:40:48 2025
      State : clean
      Active Devices : 6
      Working Devices : 6
      Failed Devices : 0
      Spare Devices : 0

      Layout : left-symmetric
      Chunk Size : 512K

      Consistency Policy : unknown

      Name : 43f67c14:127 (local to host 43f67c14)
      UUID : b5e891a1:66389bd0:7f2532a1:f84f92db
      Events : 375

      Number Major Minor RaidDevice State
      0 8 19 0 active sync /dev/sdb3
      1 8 35 1 active sync /dev/sdc3
      2 8 51 2 active sync /dev/sdd3
      3 8 67 3 active sync /dev/sde3
      7 8 83 4 active sync /dev/sdf3
      6 8 3 5 active sync /dev/sda3
      /dev/md/data-1:
      Version : 1.2
      Creation Time : Sun Apr 9 05:46:36 2023
      Raid Level : raid5
      Array Size : 11718576384 (11175.71 GiB 11999.82 GB)
      Used Dev Size : 5859288192 (5587.85 GiB 5999.91 GB)
      Raid Devices : 3
      Total Devices : 2
      Persistence : Superblock is persistent

      Update Time : Tue Mar 25 22:40:48 2025
      State : clean, degraded
      Active Devices : 2
      Working Devices : 2
      Failed Devices : 0
      Spare Devices : 0

      Layout : left-symmetric
      Chunk Size : 64K

      Consistency Policy : unknown

      Name : 43f67c14:data-1 (local to host 43f67c14)
      UUID : 09dd9a16:97d5ab01:52fa5df2:24195452
      Events : 19934

      Number Major Minor RaidDevice State
      0 8 52 0 active sync /dev/sdd4
      3 8 84 1 active sync /dev/sdf4
      - 0 0 2 removed

    • HighOctane's avatar
      HighOctane
      Aspirant

      This is the breakdown from lsblk..

       

      root@HighOctaneNAS:~# lsblk -f
      NAME FSTYPE LABEL UUID MOUNTPOINT
      sda
      ├─sda1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      │ └─md0 btrfs 43f67c14:root 91b6ea7f-9705-4e67-8714-8cc5a9ad584c /
      ├─sda2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      └─sda3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      └─md127
      sdb
      ├─sdb1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      │ └─md0 btrfs 43f67c14:root 91b6ea7f-9705-4e67-8714-8cc5a9ad584c /
      ├─sdb2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      └─sdb3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      └─md127
      sdc
      ├─sdc1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      │ └─md0 btrfs 43f67c14:root 91b6ea7f-9705-4e67-8714-8cc5a9ad584c /
      ├─sdc2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      └─sdc3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      └─md127
      sdd
      ├─sdd1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      │ └─md0 btrfs 43f67c14:root 91b6ea7f-9705-4e67-8714-8cc5a9ad584c /
      ├─sdd2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      ├─sdd3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      │ └─md127
      └─sdd4 linux_raid_member 43f67c14:data-1 09dd9a16-97d5-ab01-52fa-5df224195452
      └─md126 btrfs 43f67c14:data 2211f852-4973-412d-97ec-e340df756809
      sde
      ├─sde1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      │ └─md0 btrfs 43f67c14:root 91b6ea7f-9705-4e67-8714-8cc5a9ad584c /
      ├─sde2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      └─sde3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      └─md127
      sdf
      ├─sdf1 linux_raid_member 43f67c14:0 e6e3fdb5-875c-ed54-ab63-493ab1f7bff3
      ├─sdf2 linux_raid_member 43f67c14:1 ca988b5d-2613-8aec-9aff-2560000789c7
      │ └─md1 swap swap 3f098acf-1ece-46e4-920a-31e7f71ab658 [SWAP]
      ├─sdf3 linux_raid_member 43f67c14:127 b5e891a1-6638-9bd0-7f25-32a1f84f92db
      │ └─md127
      └─sdf4 linux_raid_member 43f67c14:data-1 09dd9a16-97d5-ab01-52fa-5df224195452
      └─md126 btrfs 43f67c14:data 2211f852-4973-412d-97ec-e340df756809

      • Sandshark's avatar
        Sandshark
        Sensei - Experienced User

        What were the drive sizes and positions before and after the first and second swap?  Did you re-boot after the first re-sync?

         

        What may be the issue is that the OS was actually still syncing a second RAID group when you thought it was done.  I have often seen a notice that re-sync is complete when it's only the first that is complete.

         

        And something that may be confusing the issue is that when you swap out a drive, it can cause the drives not to be in the order you think (e.g. sda may not be in bay 1) until you re-boot.

NETGEAR Academy

Boost your skills with the Netgear Academy - Get trained, certified and stay ahead with the latest Netgear technology! 

Join Us!

ProSupport for Business

Comprehensive support plans for maximum network uptime and business peace of mind.

 

Learn More