Talk by Ronen Narkis / @narkisr settings
“annualized failure rate for the drive models listed is 2.07% ”
3 copies of your data
2 media formats
1 off site
Originally developed at Sun, opensource since 2005
Ported to FreeBSD, Linux, OSx (Windows maybe?)
A volume manager and a Filesystem
Forked into OpenZFS in Sep 2013
COW
Volume Management
Checksummed and self healing
Compression/Dedup
Multiple Redundancy levels
Smart caching
Collection of VDEVs
VDEV is mirror/Raid/Single disk
Writes distributed evenly among VDEVs
Divided into datasets
No redundancy
Fastest performance
Pair of disks
1 Drive can die
Slow(er) write
Fast READ
Easier to expand
RAIDZ 1/2/3 (parity disks)
More CPU time (parity)
Better storage efficiency
Partitioning the pool
ZFS properties (mount point/compression/quota etc..)
Snapshot-able
Create as many as you want (per VM, etc..)
Checksum per block on write (per VDEV)
Verified on read
Restored on failure (depending on redundancy)
No need for an fsck (We can scrub)
Cheap to create
Can be mounted
Sent/Receive
Not a backup!
The process of replacing an HDD
Replacing faulty or growing a VDEV
The pool continues to function
Might moves to a degraded state
Created by IBM
Page level cache
MRU and MFU combined
Self Adjusting (shrinking/growing)
Persistent (L2ARC)
Linux (Ubuntu)
FreeBSD (Ubuntu)
FreeNAS
Omnios, SmartOS (Solaris)
ECC vs Non ECC RAM
Disk controllers (No RAID!)
Disks (NAS vs Consumer)
How much RAM?
CPU?
ZVOLs (block devices)
Boot environments
NFS,SAMBA,iSCSI
Select theme:
Black (default) -
White -
League -
Sky -
Beige -
Simple
Serif -
Blood -
Night -
Moon -
Solarized