Well, that was not my initial question, but if you like, I'd appreciate it:
It's a netcat/tar transfer via Gigabit connection. 1000 Full Duplex
Rate is 1 Terrabyte in 10 hours
There is just one netgear Switch between them, Gigabit of course
Sender has an ext3 FS connected via 3ware controller to a 16port Raid 5.
Receiver has a ZFS raidz via Areca 1260 controller in JDOB mode.
Here's iostat:
Linux 2.6.26-2-686 (Sender) 12/04/09 _i686_
Time: 13:08:34
avg-cpu: %user %nice %system %iowait %steal %idle
1.01 0.00 0.83 1.78 0.00 96.38
Device: rrqm/s wrqm/s r/s w/s rsec/s wsec/s avgrq-sz avgqu-sz await svctm %util
sda 4.08 99.04 8.88 9.38 63.53 68.07 7.21 0.09 5.07 2.69 4.91
sda1 0.02 0.00 0.00 0.00 0.02 0.00 45.63 0.00 12.33 9.13 0.00
sda2 0.01 0.00 0.00 0.00 0.01 0.00 63.98 0.00 7.24 4.53 0.00
sda3 0.01 1.94 0.04 0.40 0.91 18.75 43.76 0.01 20.34 8.72 0.39
sda4 0.00 0.00 0.00 0.00 0.00 0.00 2.00 0.00 13.33 13.33 0.00
sda5 0.01 18.21 0.12 0.73 2.30 151.48 181.64 0.08 99.30 2.70 0.23
sda6 0.00 4.54 0.06 4.54 1.19 72.67 16.05 0.25 54.75 5.22 2.40
sda7 0.00 0.00 0.00 0.00 0.00 0.01 38.96 0.00 29.93 24.06 0.00
sda8 4.02 74.35 8.65 3.71 59.10 226.37 23.09 0.15 12.02 2.68 3.31
sdb 3.12 47.52 7.11 2.63 179.03 2.62 18.66 0.19 19.13 2.96 2.88
sdb1 3.12 47.52 7.11 2.63 179.03 2.62 18.66 0.19 19.13 2.96 2.88
sdc 3.10 87.86 7.20 3.90 241.43 338.17 52.24 0.00 0.05 2.64 2.93
sdc1 3.10 87.86 7.20 3.90 241.43 338.17 52.24 0.00 0.05 2.64 2.93
sdd 0.44 8.70 1.10 0.39 229.58 73.09 202.68 0.11 73.84 3.90 0.58
sdd1 0.44 8.70 1.10 0.39 229.58 73.09 202.68 0.11 73.84 3.90 0.58
dm-0 0.00 0.00 34.93 230.48 307.94 239.04 2.06 0.34 1.23 0.31 8.10
sde 0.00 0.77 0.00 0.04 0.00 12.81 297.92 0.01 155.43 1.43 0.01
sde1 0.00 0.77 0.00 0.04 0.00 12.81 297.96 0.01 155.42 1.42 0.01
sde9 0.00 0.00 0.00 0.00 0.00 0.00 25.96 0.00 206.45 160.57 0.00
sdf 0.00 0.77 0.00 0.02 0.00 6.29 293.46 0.01 288.21 2.08 0.00
sdf1 0.00 0.77 0.00 0.02 0.00 6.29 293.54 0.01 288.24 2.07 0.00
sdf9 0.00 0.00 0.00 0.00 0.00 0.00 26.50 0.00 195.58 143.33 0.00
sdg 0.00 0.77 0.00 0.02 0.00 6.29 293.43 0.01 289.77 2.09 0.00
sdg1 0.00 0.77 0.00 0.02 0.00 6.29 293.51 0.01 289.80 2.08 0.00
sdg9 0.00 0.00 0.00 0.00 0.00 0.00 25.44 0.00 187.52 147.36 0.00
sdh 0.00 0.14 0.00 0.00 0.00 1.15 292.28 0.00 274.15 2.23 0.00
sdh1 0.00 0.14 0.00 0.00 0.00 1.15 292.70 0.00 274.26 2.17 0.00
sdh9 0.00 0.00 0.00 0.00 0.00 0.00 25.96 0.00 221.06 190.37 0.00
dm-1 0.00 0.00 0.00 2.53 0.00 26.55 10.50 0.34 133.89 0.06 0.02
Receiver:
extended device statistics tty cpu
device r/s w/s kr/s kw/s wait actv svc_t %w %b tin tout us sy wt id
fd0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 0 254 1 18 0 81
sd0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0
sd2 0.1 1.6 6.8 16.4 0.0 0.0 2.2 0 0
sd3 0.1 1.6 4.6 16.4 0.0 0.0 3.0 0 0
sd4 1.3 31.1 86.2 2178.7 0.9 0.7 51.6 5 6
sd5 1.3 31.1 86.0 2178.7 0.9 0.8 50.8 5 6
sd6 1.3 31.1 86.3 2178.7 1.0 0.7 52.0 5 6
sd7 1.3 31.1 86.0 2178.7 0.9 0.7 51.4 5 6
sd8 1.3 31.1 86.0 2178.7 1.0 0.7 52.8 5 6
sd9 1.3 31.1 86.0 2178.7 1.0 0.7 52.0 5 6
sd10 1.4 31.1 86.5 2178.7 0.9 0.8 51.1 5 6
sd11 1.3 31.1 86.2 2179.0 0.9 0.7 51.7 5 6
sd12 1.3 31.1 85.9 2179.0 1.0 0.7 52.9 5 6
sd13 1.3 31.1 86.1 2179.0 1.0 0.7 52.5 5 6
sd14 1.3 31.1 86.1 2179.0 1.0 0.7 52.0 5 6
sd15 1.3 31.1 86.3 2179.0 1.0 0.7 52.7 5 6
sd16 1.3 31.1 86.0 2179.1 0.9 0.7 51.2 5 6
sd17 1.4 31.1 86.6 2179.1 1.0 0.7 51.9 5 6
st7 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0
st8 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0
nfs1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0
This is netstat:
Sender:
Kernel Interface table
Iface MTU Met RX-OK RX-ERR RX-DRP RX-OVR TX-OK TX-ERR TX-DRP TX-OVR Flg
bond0 1500 0 2408030752 615 1238006016 0 2825662764 0 0 0 BMmRU
eth0 1500 0 3030502353 594 1570237466 0 3487222518 0 0 0 BMsRU
eth1 1500 0 1700551651 4 1594719503 0 1836835136 0 0 0 BMsRU
eth2 1500 0 2046842103 10 1572882366 0 2168071782 0 0 0 BMsRU
eth3 1500 0 1421965044 2 1749190229 0 533378842 0 0 0 BMsRU
eth4 1500 0 1776744686 4 1646652888 0 2526363523 0 0 0 BMsRU
eth5 1500 0 1021359507 1 1694258156 0 863725555 0 0 0 BMsRU
lo 16436 0 10305047 0 0 0 10305047 0 0 0 LRU
Receiver:
Name Mtu Net/Dest Address Ipkts Ierrs Opkts Oerrs Collis Queue
lo0 8232 loopback localhost 179 0 179 0 0 0
lo0 8232 loopback localhost 0 N/A 115 N/A N/A 0
aggr1 1500 obelix obelix 1601598744 0 796084178 0 0 0
aggr1 1500 obelix obelix 1601361709 N/A 795934688 N/A N/A 0
What I noticed is the sender accessing the Raid like crazy, while the receivers HDDs merely write every twenty seconds and then pause. So it needs to be either the sender or the receivers network connection.
---------- Post updated at 07:59 AM ---------- Previous update was at 07:13 AM ----------
Running a parallel copy between two raid-arrays, one internal pci-x connected, the other one scsi:
keto:~/iostat-2.2# df -m /dev/mapper/external-external; sleep 120 ; df -m /dev/mapper/external-external
Filesystem 1M-blocks Used Available Use% Mounted on
/dev/mapper/external-external
6570914 74220 6162911 2% /external_raid
Filesystem 1M-blocks Used Available Use% Mounted on
/dev/mapper/external-external
6570914 78347 6158784 2% /external_raid
An awesome 2 Gigabyte per Minute. Hell breaks loose...