This isn't quite right. You can't expand a vdev, but you definitely can add new vdevs to a pool. If you run the command "zpool status" on your NAS, you'll get a printout of the pools and vdevs.
@Nick2253
well i think i messed it a bit ... current setup looks like
[root@freenas] ~# zpool status
pool: freenas-boot
state: ONLINE
scan: scrub repaired 0 in 0h1m with 0 errors on Fri Apr 22 03:46:06 2022
config:
NAME STATE READ WRITE CKSUM
freenas-boot ONLINE 0 0 0
da0p2 ONLINE 0 0 0
errors: No known data errors
pool: storage
state: ONLINE
scan: scrub repaired 0 in 49h47m with 0 errors on Tue Mar 29 02:47:05 2022
config:
NAME STATE READ WRITE CKSUM
storage ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
gptid/6f9e603c-15bd-11e6-940f-0cc47ab3250a ONLINE 0 0 0
gptid/70c0215e-15bd-11e6-940f-0cc47ab3250a ONLINE 0 0 0
gptid/71d80541-15bd-11e6-940f-0cc47ab3250a ONLINE 0 0 0
gptid/72cb0955-15bd-11e6-940f-0cc47ab3250a ONLINE 0 0 0
gptid/73c3c94b-15bd-11e6-940f-0cc47ab3250a ONLINE 0 0 0
errors: No known data errors
pool: storage2
state: ONLINE
scan: scrub repaired 0 in 17h51m with 0 errors on Sun May 1 17:51:59 2022
config:
NAME STATE READ WRITE CKSUM
storage2 ONLINE 0 0 0
gptid/54e8a441-5593-11e7-b888-0cc47ab3250a.eli ONLINE 0 0 0
errors: No known data errors
pool: storage3
state: ONLINE
scan: scrub repaired 0 in 15h14m with 0 errors on Sun Mar 27 16:14:23 2022
config:
NAME STATE READ WRITE CKSUM
storage3 ONLINE 0 0 0
gptid/ae864c15-247f-11e8-acd7-0cc47ab3250a.eli ONLINE 0 0 0
errors: No known data errors
pool: storage4
state: ONLINE
scan: scrub repaired 0 in 11h53m with 0 errors on Sun Apr 3 11:53:38 2022
config:
NAME STATE READ WRITE CKSUM
storage4 ONLINE 0 0 0
gptid/373e2a6b-cdd2-11e9-b7e5-0cc47ab3250a.eli ONLINE 0 0 0
errors: No known data errors
RAID is not backup, and backup is not RAID. Both serve different purposes (though there is some overlap for sure). RAID is about resiliency in the face of hardware failure. Backup is about data protection. As such, RAID helps keep your system up and working when a hard drive fails. This can be the catastrophic loss of a disk, or as small as a bad sector (URE). If any such failure happens and you don't have any kind of RAID, then your NAS effectively goes offline, and you must restore from backup to get your system going again.
For most users, this kind of failure is common enough that it's worth protecting against.
Now got your point so I think i should be protected that way ie raidz1/2 instead of A and B machines which wont bring any benefit.
If NAS A or B can access NAS C via stored credentials, then a compromise of NAS A or B is as good as a compromise of NAS C.
Yeah, assuming here these will be secured ie AB/C will be protected.
Again, RAID is not backup, backup is not RAID.
If you're talking about NAS A and B being real-time mirrors of each other, then that could serve much the same purpose as RAID. But then NAS B really isn't a backup. Furthermore, from a home user perspective, running a NAS cluster really doesn't make sense from a performance or cost-benefit analysis. You don't need the performance benefits of a cluster (most likely), and why run a whole bunch of disks and a second system when you can put just one or two drives in NAS A?
I was thinking to have lets say media/documents etc data primarily on A, and B would hold backup of A; and data as backup of local laptops.
C would act as offsite - ie backup of A and B
But the thing is as you mentioned in case my A or B dies - there will be outage and I will have to restore from C which can take a time.
Also Why I was thinking about B is that my line is 1000/40 so basically I am not sure about offsite from A on daily basis to C; thats why i wanted to have it partially (and locally via LAN on B). Maybe that thinking is incorrect.
Really, though, what you need to analyze is your failure modes, and acceptable risks. Can you survive your NAS being down for a day? A week? What are the costs, consequences, and likelihood of: hardware failure? HDD failure? Regional catastrophe? Local fire? Etc.
For most home users, a NAS with some kind of HDD redundancy (e.g. RAIDZ2) and a single daily offsite backup meets their needs. This allows the NAS to continue working reliably in the face of a HDD failure (single most common hardware failure), and protects against data loss caused by natural disaster or fire. For users who need extra data protection or to address concerns about corruption, the adding some kind of monthly or yearly offsite/offline backup is usually the third copy needed.
What do you mean by extra data protection, or how do you want to protect against the corruption?
So in the backup pattern
3-2-1 ...
1-2 is online replica/mirror physically separated
3 is offsite?
In my experience, there is no meaingful difference. I've met people who have had great and terrible experiences with every brand, so I believe that getting a good batch is much more important than picking a good brand or model. Certain drive models have been lemons, for sure, but it's usually not obvious until the drives have been in circulation for a while.
Warranty, if you buy from an authorized seller, comes from the manufacturer.
I will have to find out somehow if they are authorized sellers.
If warranty comes from manufacturer it means in case of failure you send drive to manufacturer for a replacement?
I saw some sellers selling drive with standard warranty and then option to extend that by 2-3y... but that extension contract is between me and seller so no clue if manufacturer will be able to step-in.
I can see your "Steel" is using HGST Ultrastar drives, which you didnt mention, for a some specific reason?
You've probably seen the "RAID5 is Dead" article. The math used in that article isn't quite right, but the general point is sound: as drives get larger, rebuild times get longer, and the possibility of a URE during rebuild increases. So, the takeaway is that double redundancy is necessary in some cases. However, single redundancy configurations, like mirrors, RAID5, and RAIDZ1, still have their time and place. If you want to be able to fix UREs during a scrub, you need some kind of redundancy. If you use a checksumming filesystem like ZFS, and regularly scrub your data, then the likelihood that you will run into both a URE and a drive failure at the same time substantially decreases. Also, ZFS is pretty resilient in the face of a checksum error, and can tell you exactly which data is corrupt, so you can restore from backups.
Yeah, exactly that article.
In that case is RAIDZ1(RAID5) enough for a home use? Using ZFS and ECC, or do I have to go for 2 parity drives ie raid6(z2)?
In general, a mirror isn't great for a NAS from a cost-benefit perspective: a full 50% of your available capacity is taken up in redundancy. And since most NASes are measured in cost per usable TB, you usually see NASes with 6+ drives, and because resiliency is important, you usually see double redundancy like RAIDZ2.
Iron is currently my main NAS. Steel will replace Iron; I'm in the process of migrating the many jails I was running on Iron (which is why it's still on 9.10) to containers on my Proxmox server. Once that's done, then Iron will be used for a different purpose (likely set up for a friend or family). Aluminum is the destination for my local server and PC backups. A weekly snapshot of these backups is replicated to Iron.
Since FreeNAS 9.10 doesn't have the cloud backup tool, I'm currently using Steel for this function. Daily snapshots from Iron are replicated to Steel, and then Steel pushes these to the cloud. Once Iron is decommissioned, then all the programs that currently "talk" to Iron will go directly to Steel, and I won't have a "middleman" in this role.
Got your point, but still not clear what cloud do you mean or what kind of tool/sw/mechanism is used (in truenas v12) for such a push to the cloud, as clouds like dropbox, backblaze dont provide good interface for offsite-backups (When i did research... )
So apart of the NAS boxes, you set up a separate HW/box for Proxmox server? Just asking bc I am also running some virtual machines on bhyve but have to move/migrate all of the off the NAS if possible.
Once Steel is running everything, then I'll probably decommission Aluminum as well, and do the PC/server backups directly to Steel. Aluminum is very under-powered for TrueNAS and ZFS, and it struggles in it's current role.
Also there is another thing ... to pickup reliable tool for win/linux/osx that can backup to NAS.
More about specific setup / scenario
I have:
9x 3TB drives
3x 8TB drives
so I was thinking to do following setup
RAIDZ2: 7x 3TB total 15TB
RAIDZ1: 3x 8TB total 16 TB
to buy 1x 16TB (not important data ie backups from laptops, movies/music etc - that can be restored from offsite within 1 week and i can live without)
Z2 and Z1 will store the critical data like VM images, documents, photos ... , data from various working projects.
Here I am not sure which pool suits better (more integrity, resilient etc) for *the most critical data* ie. 15TB or 16TB volume?
Then Offsite machine (was thinking about HW without ECC memories - which mean some check to prevent mentioned data corruption has to be done, on SW level? Or maybe design is wrong and that machine has to come with ECC also)
8TB smr (have the old one) + 2x 3TB - to partialy backup 15TB
2x new 16TB - to backup RAIDZ1, and single 16TB
Again, maybe i am completely missing something here, so happy to get tips/ideas.
Appreciate so much.
Theory>
https://calomel.org/zfs_raid_speed_capacity.html
Be Safe, Not Sorry. When choosing a raid configuration you may look at raidz or RAID5 and see the speed and capacity benefits and decide it is a good choice. From real world experience we highly suggest NOT using raid5. It is simply not safe enough. We recommend using RAID1 mirroring or RAID6 (double parity) or even RAID7 (triple parity). The problem with raid5 is you only have one drive with parity. When one drive dies the raid5 array is degraded and you can not loose another drive before you lose the entire array; i.e. all your data is lost. What happens most of the time is one drive dies, you replace the drive and the array starts resilvering or rebuilding. This is when the other disks are being stressed and during this rebuild is when you have a very good chance (around 8%) that you you will lose another drive. For more information please take some time to read
NetApp Weighs In On Disks which was written in 2007, but is still valid today since drive technology has not changed much.