Differences between revisions 1 and 3 (spanning 2 versions)
Revision 1 as of 2014-07-27 14:00:17
Size: 28
Editor: wireless46
Comment:
Revision 3 as of 2014-07-27 18:09:19
Size: 4371
Editor: SamMoore
Comment:
Deletions are marked like this. Additions are marked like this.
Line 2: Line 2:

== Root Filesystem ==

There are 2 SSDs with 3 partitions in RAID1.

  * /dev/md0 is /dev/sda1 and /dev/sdb1 - /boot, ext2
  * /dev/md1 is /dev/sda2 and /dev/sdb2 - swap (because)
  * /dev/md2 is /dev/sda5 and /dev/sdb5 - /, LVM ext4

In theory this will work.
In practice it did work but only after a reinstall (keeping the first install's disk layouts).

{{{
root@molmol:~# cat /proc/mdstat
Personalities : [raid1]
md2 : active raid1 sdb5[3] sda5[2]
      245520192 blocks super 1.2 [2/2] [UU]
      
md1 : active raid1 sda2[2] sdb2[1]
      3904448 blocks super 1.2 [2/2] [UU]
      
md0 : active raid1 sdb1[3] sda1[2]
      498368 blocks super 1.2 [2/2] [UU]
}}}

{{{
root@molmol:~# fdisk -l /dev/sda /dev/sdb

Disk /dev/sda: 256.1 GB, 256060514304 bytes
255 heads, 63 sectors/track, 31130 cylinders, total 500118192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0004e00d

   Device Boot Start End Blocks Id System
/dev/sda1 * 2048 999423 498688 fd Linux raid autodetect
/dev/sda2 999424 8812543 3906560 fd Linux raid autodetect
/dev/sda3 8814590 500117503 245651457 5 Extended
/dev/sda5 8814592 500117503 245651456 fd Linux raid autodetect

WARNING: GPT (GUID Partition Table) detected on '/dev/sdb'! The util fdisk doesn't support GPT. Use GNU Parted.


Disk /dev/sdb: 256.1 GB, 256060514304 bytes
255 heads, 63 sectors/track, 31130 cylinders, total 500118192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000

   Device Boot Start End Blocks Id System
/dev/sdb1 * 2048 999423 498688 fd Linux raid autodetect
/dev/sdb2 999424 8812543 3906560 fd Linux raid autodetect
/dev/sdb3 8814590 500117503 245651457 5 Extended
/dev/sdb5 8814592 500117503 245651456 fd Linux raid autodetect\
}}}


== ZFS ==

Followed this guide: [[http://bernaerts.dyndns.org/linux/75-debian/279-debian-wheezy-zfs-raidz-pool|http://bernaerts.dyndns.org/linux/75-debian/279-debian-wheezy-zfs-raidz-pool]]

Didn't setup snapshots.

{{{
root@molmol:~# zpool list
NAME SIZE ALLOC FREE CAP DEDUP HEALTH ALTROOT
over 7.25T 1.64M 7.25T 0% 1.00x ONLINE -
}}}

{{{
root@molmol:~# zpool status
  pool: over
 state: ONLINE
  scan: scrub repaired 0 in 0h0m with 0 errors on Sun Jul 27 18:02:05 2014
config:

 NAME STATE READ WRITE CKSUM
 over ONLINE 0 0 0
   raidz1-0 ONLINE 0 0 0
     sdc ONLINE 0 0 0
     sdd ONLINE 0 0 0
     sde ONLINE 0 0 0
     sdf ONLINE 0 0 0
     sdg ONLINE 0 0 0
     sdh ONLINE 0 0 0
     sdi ONLINE 0 0 0
     sdj ONLINE 0 0 0

errors: No known data errors
}}}

The `zfs` is `/there` and is currently exported to [[Motsugo]].

{{{
root@molmol:~# df -h
Filesystem Size Used Avail Use% Mounted on
rootfs 227G 1.5G 214G 1% /
udev 10M 0 10M 0% /dev
tmpfs 1.6G 540K 1.6G 1% /run
/dev/mapper/molmol-molmol--rootfs 227G 1.5G 214G 1% /
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 3.9G 0 3.9G 0% /run/shm
/dev/md0 457M 19M 414M 5% /boot
services.ucc.gu.uwa.edu.au:/space/away/home 1.9T 1.7T 26G 99% /away
home.ucc.gu.uwa.edu.au:/home 2.0T 952G 963G 50% /home
nortel.ucc.gu.uwa.edu.au:/vol/space/services 884G 674G 211G 77% /services
over/there 6.1T 256K 6.1T 1% /there
}}}

{{{
root@molmol:~# cat /etc/exports
/there motsugo(rw,sync,no_root_squash,mountpoint,no_subtree_check,secure)
}}}

GLHFDD.

Molmol is the File Server.

Root Filesystem

There are 2 SSDs with 3 partitions in RAID1.

  • /dev/md0 is /dev/sda1 and /dev/sdb1 - /boot, ext2
  • /dev/md1 is /dev/sda2 and /dev/sdb2 - swap (because)
  • /dev/md2 is /dev/sda5 and /dev/sdb5 - /, LVM ext4

In theory this will work. In practice it did work but only after a reinstall (keeping the first install's disk layouts).

root@molmol:~# cat /proc/mdstat
Personalities : [raid1] 
md2 : active raid1 sdb5[3] sda5[2]
      245520192 blocks super 1.2 [2/2] [UU]
      
md1 : active raid1 sda2[2] sdb2[1]
      3904448 blocks super 1.2 [2/2] [UU]
      
md0 : active raid1 sdb1[3] sda1[2]
      498368 blocks super 1.2 [2/2] [UU]

root@molmol:~# fdisk -l /dev/sda /dev/sdb

Disk /dev/sda: 256.1 GB, 256060514304 bytes
255 heads, 63 sectors/track, 31130 cylinders, total 500118192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0004e00d

   Device Boot      Start         End      Blocks   Id  System
/dev/sda1   *        2048      999423      498688   fd  Linux raid autodetect
/dev/sda2          999424     8812543     3906560   fd  Linux raid autodetect
/dev/sda3         8814590   500117503   245651457    5  Extended
/dev/sda5         8814592   500117503   245651456   fd  Linux raid autodetect

WARNING: GPT (GUID Partition Table) detected on '/dev/sdb'! The util fdisk doesn't support GPT. Use GNU Parted.


Disk /dev/sdb: 256.1 GB, 256060514304 bytes
255 heads, 63 sectors/track, 31130 cylinders, total 500118192 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1   *        2048      999423      498688   fd  Linux raid autodetect
/dev/sdb2          999424     8812543     3906560   fd  Linux raid autodetect
/dev/sdb3         8814590   500117503   245651457    5  Extended
/dev/sdb5         8814592   500117503   245651456   fd  Linux raid autodetect\

ZFS

Followed this guide: http://bernaerts.dyndns.org/linux/75-debian/279-debian-wheezy-zfs-raidz-pool

Didn't setup snapshots.

root@molmol:~# zpool list
NAME   SIZE  ALLOC   FREE    CAP  DEDUP  HEALTH  ALTROOT
over  7.25T  1.64M  7.25T     0%  1.00x  ONLINE  -

root@molmol:~# zpool status
  pool: over
 state: ONLINE
  scan: scrub repaired 0 in 0h0m with 0 errors on Sun Jul 27 18:02:05 2014
config:

        NAME        STATE     READ WRITE CKSUM
        over        ONLINE       0     0     0
          raidz1-0  ONLINE       0     0     0
            sdc     ONLINE       0     0     0
            sdd     ONLINE       0     0     0
            sde     ONLINE       0     0     0
            sdf     ONLINE       0     0     0
            sdg     ONLINE       0     0     0
            sdh     ONLINE       0     0     0
            sdi     ONLINE       0     0     0
            sdj     ONLINE       0     0     0

errors: No known data errors

The zfs is /there and is currently exported to Motsugo.

root@molmol:~# df -h
Filesystem                                    Size  Used Avail Use% Mounted on
rootfs                                        227G  1.5G  214G   1% /
udev                                           10M     0   10M   0% /dev
tmpfs                                         1.6G  540K  1.6G   1% /run
/dev/mapper/molmol-molmol--rootfs             227G  1.5G  214G   1% /
tmpfs                                         5.0M     0  5.0M   0% /run/lock
tmpfs                                         3.9G     0  3.9G   0% /run/shm
/dev/md0                                      457M   19M  414M   5% /boot
services.ucc.gu.uwa.edu.au:/space/away/home   1.9T  1.7T   26G  99% /away
home.ucc.gu.uwa.edu.au:/home                  2.0T  952G  963G  50% /home
nortel.ucc.gu.uwa.edu.au:/vol/space/services  884G  674G  211G  77% /services
over/there                                    6.1T  256K  6.1T   1% /there

root@molmol:~# cat /etc/exports
/there  motsugo(rw,sync,no_root_squash,mountpoint,no_subtree_check,secure)

GLHFDD.