NETGEAR is aware of a growing number of phone and online scams. To learn how to stay safe click here.
Forum Discussion
Platypus69
Feb 19, 2021Luminary
Cannot copy files to RN316 although I have 22TB free...
Hi all. I have the following RN316: Firmware 6.10.4 Running 6 x 10TB IronWolf HDDs X-RAID 21.9TB free / 23.4TB used History: Last year I replaced all the 8TB Ironwold HDDs (from memory) one by...
Platypus69
Feb 21, 2021Luminary
Thanks.
I am familiar with Microsoft stack and NTFS/ReFS, and VSAN.
I know nothing about BTRFS, so I assume that it would expand existing volume not create a second. Thus my confusion through ignorance.
Yes, you are correctm they were 4TB HDDs, it was over 5 years ago probably...
Thanks for the heads up on the Snapshots. I expect it was a good idea 5 years when I set it up as it was designed to store family photos/videos.
So how do I turn of snapshots. Is it simply a matter of going into the snapshot GUI and say turn off "Smart" management? I am nervouse about doing something wrong and losing data :)
What would I be looking for in kernel.log or system.log? I tried to attach them as a ZIP file but it ssmes you can't...
I had a quick look and noithing made sense to me or stood out.
Although I am wondering if it is snapshot related. I am getting lots of errors like the following in system.log:
Feb 21 22:56:03 RN316 snapperd[545]: loading 12848 failed Feb 21 22:56:03 RN316 snapperd[545]: loading 12850 failed Feb 21 22:56:03 RN316 snapperd[545]: loading 1036 failed Feb 21 22:56:03 RN316 snapperd[545]: loading 1037 failed
A quick Google search seemed to imply that snapperd is snapshot related???
Any help appreciated.
Sorry for delay, RN316 has been unresponsive for 3 days due to Scrub operation :(
Actually I just randomly found this in VOLUME.LOG:
data disk test 2020-09-01 01:00:01 2020-09-01 15:20:27 pass data resilver 2020-09-13 15:54:14 2020-09-14 20:35:46 completed data balance 2021-02-18 21:17:16 2021-02-18 21:18:42 completed ERROR: error during balancing '/data': No space left on device T data scrub 2021-02-18 21:29:29
Is that relevant???
Should I try moving some large files off and doing Balance again?
TIA
StephenB
Feb 21, 2021Guru - Experienced User
Platypus69 wrote:
So how do I turn off snapshots. Is it simply a matter of going into the snapshot GUI and say turn off "Smart" management? I am nervous about doing something wrong and losing data :)
You can either turn them off altogether in the GUI or you can change to Custom and explicitly set retention. Just turning them off won't delete existing snapshots - you need to go into "recover", select them, and delete them manually.
Platypus69 wrote:
I assume that it would expand existing volume not create a second.
It did expand the existing volume. You are confusing "volumes" with "RAID groups". They aren't the same thing. A ReadyNAS volume consists of one or more RAID groups.
Platypus69 wrote:
What would I be looking for in kernel.log or system.log? I tried to attach them as a ZIP file but it ssmes you can't...
Actually I just randomly found this in VOLUME.LOG:
data disk test 2020-09-01 01:00:01 2020-09-01 15:20:27 pass data resilver 2020-09-13 15:54:14 2020-09-14 20:35:46 completed data balance 2021-02-18 21:17:16 2021-02-18 21:18:42 completed ERROR: error during balancing '/data': No space left on device T data scrub 2021-02-18 21:29:29Is that relevant???
Should I try moving some large files off and doing Balance again?
You can't attach zips, and you should be cautious about including links to the full logs. There is some privacy leakage.
Generally you are looking for errors that include disks or "btrfs".
Your error is certainly relevant, and I would suggest looking for errors in system and kernel.log around the time of the error in volume.log. Deleting some large files and trying to balance again is a reasonable next step.
- SandsharkFeb 21, 2021Sensei
StephenB wrote:
Deleting some large files and trying to balance again is a reasonable next step.And perhaps some of the oldest (and thus likely largest) snapshots as well before the balance. The first balance will normally take a while, so that error explains why it only took seconds. That the ReadyNAS didn't tell you it ended in error is, IMHO, an oversight in the ReadyNAS OS.
I've seen some posts on general Linux forums recommending a balance after a BTRFS expansion, and it does not appear Netgear does that automatically. Apparently, that helps properly allocate data and metadata across the volume. The MAN page says "The primary purpose of the balance feature is to spread block groups across all devices so they match constraints defined by the respective profiles". I've not found a good list of those restraints, but you may have arrived at one of them. So having added to and then expanded one of the "devices" (which is a multi-partition MDADM RAID) in your file system, a balance should help. Of course, it is sometimes hard to pick out what in the general BTRFS forums and wiki are applicable to ReadyNAS because ReadyNAS uses BTRFS on top of MDADM RAID rather than BTRFS for both file system and RAID and the general discussions typically assume you are using BTRFS for both..
- StephenBFeb 21, 2021Guru - Experienced User
Sandshark wrote:
I've seen some posts on general Linux forums recommending a balance after a BTRFS expansion, and it does not appear Netgear does that automatically. Apparently, that helps properly allocate data and metadata across the volume. The MAN page says "The primary purpose of the balance feature is to spread block groups across all devices so they match constraints defined by the respective profiles". I've not found a good list of those restraints, but you may have arrived at one of them.
I'm wondering that also. Looking at the first post, I see this.
Label: 'data' uuid: ... Total devices 2 FS bytes used 23.43TiB devid 1 size 18.17TiB used 18.17TiB path /dev/md127 devid 2 size 27.28TiB used 5.29TiB path /dev/md126
Note all the unallocated space is on md126. md127 is completely full.
- Platypus69Mar 05, 2021Luminary
Hi all....
So I have not been able to resolve the issue yet, I believe.
BTW is 6.10.4 (Hotfix 1) buggy??? Should I be downgrading to 6.10.3?I've been offline as the RN316 was horrendously unavailabe during my scrub, and even after it failed or completed after 7 days it still feels sluggish to me...
Anyway... I have moved about 500GB off the RN316. But is that enough? It have moved most of the files I have copied across this year, but also some older stuff from late last year, but probably AFTER I replaced all the 4TB HDDs with 10TB HDDs so I am again wondering if everything is going to the old md127 and not new md126???
So right now here are my stats / logs / telemetry:
BTRFS.LOG
Label: 'blah:root' uuid: blah-blah Total devices 1 FS bytes used 1.43GiB devid 1 size 4.00GiB used 3.61GiB path /dev/md0 Label: 'blah:data' uuid: blah-blah Total devices 2 FS bytes used 22.95TiB devid 1 size 18.17TiB used 18.14TiB path /dev/md127 devid 2 size 27.28TiB used 4.84TiB path /dev/md126 === filesystem /data === Data, single: total=22.95TiB, used=22.93TiB System, RAID1: total=32.00MiB, used=2.95MiB Metadata, RAID1: total=5.85GiB, used=5.32GiB Metadata, DUP: total=10.50GiB, used=10.03GiB GlobalReserve, single: total=512.00MiB, used=0.00B === subvolume /data ===
Why does Data, single: showing 22.95TiB, which suspiciously seems to be the limit of the amount of data I can store? Recall the UI is showing data 22.96TB and Free space: 22.49TB. Is this the "smoking gun"???
VOLUME.LOG
data disk test 2020-09-01 01:00:01 2020-09-01 15:20:27 pass data resilver 2020-09-13 15:54:14 2020-09-14 20:35:46 completed data balance 2021-02-18 21:17:16 2021-02-18 21:18:42 completed ERROR: error during balancing '/data': No space left on device T data scrub 2021-02-18 21:29:29 data disk test 2021-03-01 08:15:53 data balance 2021-03-01 21:03:16 2021-03-01 21:04:48 completed ERROR: error during balancing '/data': No space left on device T data balance 2021-03-03 15:34:37 2021-03-04 03:44:36 completed ERROR: error during balancing '/data': No space left on device T data balance 2021-03-05 09:34:32 2021-03-05 10:29:27 completed ERROR: error during balancing '/data': No space left on device T data balance 2021-03-05 19:39:44 2021-03-05 19:49:07 completed ERROR: error during balancing '/data': No space left on device T data balance 2021-03-05 21:09:45 2021-03-05 21:27:23 completed ERROR: error during balancing '/data': No space left on device T data balance 2021-03-05 21:28:15 2021-03-05 21:28:19 completed Done, had to relocate 1 out of 23557 chunks data balance 2021-03-05 21:45:20 2021-03-05 21:46:05 completed Done, had to relocate 29 out of 23557 chunks data balance 2021-03-05 21:57:26 2021-03-05 21:57:31 completed Done, had to relocate 1 out of 23529 chunks data balance 2021-03-05 21:59:22 2021-03-05 21:59:27 completed Done, had to relocate 1 out of 23529 chunks data balance 2021-03-05 21:59:48 2021-03-05 21:59:53 completed Done, had to relocate 1 out of 23529 chunks data balance 2021-03-05 22:25:13 2021-03-05 22:25:18 completed Done, had to relocate 1 out of 23529 chunks
Why does it keep relocate 1 out of 23529 chunks only? That chunk size does not go down? I have no idea. Do I keep doing balances?
Should I Defrag now?
I also have SMB Plus installed and have enabled Preallocate (FYI: Preallocate disk space before writing data. This can slow down write speed slightly, but should result in the file being nicely laid out on the disk, with minimal fragmentation.)
I have removed a lot of snapshots but would like to keep the ones that I have set up for OneDrive and DropBox apps. Both report 19 snapshots with 2 years protection.
I am happy to turn off the snapshots, ie: set them to manual. But can drop all snapshots if people think that's good. Just being a bit nervous...
I'm pulling my hair out... What do I do?
Has the problem been solved? Can you tell? Or should I try to find much older data and remove another 500Gb or 1 TB of older data before I try balancing again?
Any help appreciated!
For what it's worth:
KERNEL.LOG
Mar 05 22:11:19 RN316 systemd[1]: Set hostname to <RN316>. Mar 05 22:11:19 RN316 systemd[1]: systemd-journald-audit.socket: Cannot add dependency job, ignoring: Unit systemd-journald-audit.socket is masked. Mar 05 22:11:19 RN316 systemd[1]: systemd-journald-audit.socket: Cannot add dependency job, ignoring: Unit systemd-journald-audit.socket is masked. Mar 05 22:11:19 RN316 systemd[1]: Started Forward Password Requests to Wall Directory Watch. Mar 05 22:11:19 RN316 systemd[1]: Listening on Journal Socket (/dev/log). Mar 05 22:11:19 RN316 systemd[1]: Set up automount Arbitrary Executable File Formats File System Automount Point. Mar 05 22:11:19 RN316 systemd[1]: Created slice System Slice. Mar 05 22:11:19 RN316 systemd[1]: Created slice system-serial\x2dgetty.slice. Mar 05 22:11:19 RN316 systemd[1]: Created slice system-getty.slice. Mar 05 22:11:19 RN316 systemd[1]: Listening on /dev/initctl Compatibility Named Pipe. Mar 05 22:11:19 RN316 systemd[1]: Started Dispatch Password Requests to Console Directory Watch. Mar 05 22:11:19 RN316 systemd[1]: Reached target Encrypted Volumes. Mar 05 22:11:19 RN316 systemd[1]: Listening on udev Control Socket. Mar 05 22:11:19 RN316 systemd[1]: Reached target Paths. Mar 05 22:11:19 RN316 systemd[1]: Reached target Remote File Systems (Pre). Mar 05 22:11:19 RN316 systemd[1]: Reached target Remote File Systems. Mar 05 22:11:19 RN316 systemd[1]: Listening on udev Kernel Socket. Mar 05 22:11:19 RN316 systemd[1]: Listening on Journal Socket. Mar 05 22:11:19 RN316 systemd[1]: Starting Remount Root and Kernel File Systems... Mar 05 22:11:19 RN316 systemd[1]: Mounting POSIX Message Queue File System... Mar 05 22:11:19 RN316 systemd[1]: Starting Create Static Device Nodes in /dev... Mar 05 22:11:19 RN316 systemd[1]: Mounting Debug File System... Mar 05 22:11:19 RN316 systemd[1]: Created slice User and Session Slice. Mar 05 22:11:19 RN316 systemd[1]: Reached target Slices. Mar 05 22:11:19 RN316 systemd[1]: Listening on Syslog Socket. Mar 05 22:11:19 RN316 systemd[1]: Starting Journal Service... Mar 05 22:11:19 RN316 systemd[1]: Starting Load Kernel Modules... Mar 05 22:11:19 RN316 systemd[1]: Started ReadyNAS LCD splasher. Mar 05 22:11:19 RN316 systemd[1]: Starting ReadyNASOS system prep... Mar 05 22:11:19 RN316 systemd[1]: Mounted POSIX Message Queue File System. Mar 05 22:11:19 RN316 systemd[1]: Mounted Debug File System. Mar 05 22:11:19 RN316 systemd[1]: Started Remount Root and Kernel File Systems. Mar 05 22:11:19 RN316 systemd[1]: Started Create Static Device Nodes in /dev. Mar 05 22:11:19 RN316 systemd[1]: Started Load Kernel Modules. Mar 05 22:11:19 RN316 systemd[1]: Starting Apply Kernel Variables... Mar 05 22:11:19 RN316 systemd[1]: Mounting FUSE Control File System... Mar 05 22:11:19 RN316 systemd[1]: Mounting Configuration File System... Mar 05 22:11:19 RN316 systemd[1]: Starting udev Kernel Device Manager... Mar 05 22:11:19 RN316 systemd[1]: Starting Load/Save Random Seed... Mar 05 22:11:19 RN316 systemd[1]: Starting Rebuild Hardware Database... Mar 05 22:11:19 RN316 systemd[1]: Mounted Configuration File System. Mar 05 22:11:19 RN316 systemd[1]: Mounted FUSE Control File System. Mar 05 22:11:19 RN316 systemd[1]: Started Apply Kernel Variables. Mar 05 22:11:19 RN316 systemd[1]: Started ReadyNASOS system prep. Mar 05 22:11:19 RN316 systemd[1]: Started Load/Save Random Seed. Mar 05 22:11:19 RN316 systemd[1]: Started udev Kernel Device Manager. Mar 05 22:11:19 RN316 systemd[1]: Started Journal Service. Mar 05 22:11:19 RN316 kernel: md: md127 stopped. Mar 05 22:11:19 RN316 kernel: md: bind<sdb3> Mar 05 22:11:19 RN316 kernel: md: bind<sdc3> Mar 05 22:11:19 RN316 kernel: md: bind<sdd3> Mar 05 22:11:19 RN316 kernel: md: bind<sde3> Mar 05 22:11:19 RN316 kernel: md: bind<sdf3> Mar 05 22:11:19 RN316 kernel: md: bind<sda3> Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sda3 operational as raid disk 0 Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sdf3 operational as raid disk 5 Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sde3 operational as raid disk 4 Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sdd3 operational as raid disk 3 Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sdc3 operational as raid disk 2 Mar 05 22:11:19 RN316 kernel: md/raid:md127: device sdb3 operational as raid disk 1 Mar 05 22:11:19 RN316 kernel: md/raid:md127: allocated 6474kB Mar 05 22:11:19 RN316 kernel: md/raid:md127: raid level 5 active with 6 out of 6 devices, algorithm 2 Mar 05 22:11:19 RN316 kernel: RAID conf printout: Mar 05 22:11:19 RN316 kernel: --- level:5 rd:6 wd:6 Mar 05 22:11:19 RN316 kernel: disk 0, o:1, dev:sda3 Mar 05 22:11:19 RN316 kernel: disk 1, o:1, dev:sdb3 Mar 05 22:11:19 RN316 kernel: disk 2, o:1, dev:sdc3 Mar 05 22:11:19 RN316 kernel: disk 3, o:1, dev:sdd3 Mar 05 22:11:19 RN316 kernel: disk 4, o:1, dev:sde3 Mar 05 22:11:19 RN316 kernel: disk 5, o:1, dev:sdf3 Mar 05 22:11:19 RN316 kernel: created bitmap (30 pages) for device md127 Mar 05 22:11:19 RN316 kernel: md127: bitmap initialized from disk: read 2 pages, set 0 of 59543 bits Mar 05 22:11:19 RN316 kernel: md127: detected capacity change from 0 to 19979093934080 Mar 05 22:11:19 RN316 kernel: Adding 1566716k swap on /dev/md1. Priority:-1 extents:1 across:1566716k Mar 05 22:11:20 RN316 kernel: BTRFS: device label 43f5fa04:data devid 1 transid 1895561 /dev/md127 Mar 05 22:11:20 RN316 kernel: md: md126 stopped. Mar 05 22:11:20 RN316 kernel: md: bind<sdb4> Mar 05 22:11:20 RN316 kernel: md: bind<sdc4> Mar 05 22:11:20 RN316 kernel: md: bind<sdd4> Mar 05 22:11:20 RN316 kernel: md: bind<sde4> Mar 05 22:11:20 RN316 kernel: md: bind<sdf4> Mar 05 22:11:20 RN316 kernel: md: bind<sda4> Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sda4 operational as raid disk 0 Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sdf4 operational as raid disk 5 Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sde4 operational as raid disk 4 Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sdd4 operational as raid disk 3 Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sdc4 operational as raid disk 2 Mar 05 22:11:20 RN316 kernel: md/raid:md126: device sdb4 operational as raid disk 1 Mar 05 22:11:20 RN316 kernel: md/raid:md126: allocated 6474kB Mar 05 22:11:20 RN316 kernel: md/raid:md126: raid level 5 active with 6 out of 6 devices, algorithm 2 Mar 05 22:11:20 RN316 kernel: RAID conf printout: Mar 05 22:11:20 RN316 kernel: --- level:5 rd:6 wd:6 Mar 05 22:11:20 RN316 kernel: disk 0, o:1, dev:sda4 Mar 05 22:11:20 RN316 kernel: disk 1, o:1, dev:sdb4 Mar 05 22:11:20 RN316 kernel: disk 2, o:1, dev:sdc4 Mar 05 22:11:20 RN316 kernel: disk 3, o:1, dev:sdd4 Mar 05 22:11:20 RN316 kernel: disk 4, o:1, dev:sde4 Mar 05 22:11:20 RN316 kernel: disk 5, o:1, dev:sdf4 Mar 05 22:11:20 RN316 kernel: md126: detected capacity change from 0 to 29999560785920 Mar 05 22:11:20 RN316 kernel: BTRFS: device label 43f5fa04:data devid 2 transid 1895561 /dev/md126 Mar 05 22:13:08 RN316 kernel: e1000e: eth1 NIC Link is Down Mar 05 22:13:09 RN316 kernel: IPv6: ADDRCONF(NETDEV_UP): eth1: link is not ready Mar 05 22:13:09 RN316 kernel: 8021q: adding VLAN 0 to HW filter on device eth1 Mar 05 22:13:12 RN316 kernel: e1000e: eth1 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None Mar 05 22:13:12 RN316 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth1: link becomes ready Mar 05 22:13:15 RN316 kernel: Adjusting tsc more than 11% (6835455 vs 8751273) Mar 05 22:16:07 RN316 kernel: nr_pdflush_threads exported in /proc is scheduled for removal Mar 05 22:25:11 RN316 kernel: BTRFS info (device md126): relocating block group 36377627721728 flags system|raid1
Not sure if these snapperd errors are relevant:
SYSTEM.LOG
Mar 05 22:21:54 RN316 dbus[2986]: [system] Activating service name='org.opensuse.Snapper' (using servicehelper) Mar 05 22:21:54 RN316 dbus[2986]: [system] Successfully activated service 'org.opensuse.Snapper' Mar 05 22:21:54 RN316 snapperd[6838]: loading 13409 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19029 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19504 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19543 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19557 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19608 failed Mar 05 22:21:54 RN316 snapperd[6838]: loading 19614 failed ... Mar 05 22:25:25 RN316 clamd[4134]: SelfCheck: Database status OK. Mar 05 22:25:32 RN316 snapperd[6838]: loading 13409 failed Mar 05 22:25:32 RN316 snapperd[6838]: loading 19029 failed Mar 05 22:25:32 RN316 snapperd[6838]: loading 19504 failed ... Mar 05 22:25:32 RN316 snapperd[6838]: loading 12924 failed Mar 05 22:25:32 RN316 snapperd[6838]: loading 12925 failed Mar 05 22:25:32 RN316 snapperd[6838]: loading 1036 failed
Related Content
NETGEAR Academy
Boost your skills with the Netgear Academy - Get trained, certified and stay ahead with the latest Netgear technology!
Join Us!