Corrections, more text
This commit is contained in:
parent
ea9a5fab26
commit
ac3c5ab5cc
2 changed files with 6 additions and 2 deletions
|
@ -54,7 +54,9 @@ Compared to the figures at the time of the LSM-Tree paper, the actual delta in t
|
||||||
|
|
||||||
The availability of SDDs is not a silver bullet to disk i/o problems when cost is considered, as although they eliminate the additional costs of random page access through the removal of the disk head movement overhead (of about 6.5ms per shift), this benefit is at an order of magnitude difference in cost compared to spinning disks, and at a cost greater than half the price of DRAM. SSDs have not taken the problem of managing the overheads of disk persistence away, they've simply added another dimension to the economic profiling problem.
|
The availability of SDDs is not a silver bullet to disk i/o problems when cost is considered, as although they eliminate the additional costs of random page access through the removal of the disk head movement overhead (of about 6.5ms per shift), this benefit is at an order of magnitude difference in cost compared to spinning disks, and at a cost greater than half the price of DRAM. SSDs have not taken the problem of managing the overheads of disk persistence away, they've simply added another dimension to the economic profiling problem.
|
||||||
|
|
||||||
In physical on-premise server environments there is also commonly the cost of disk controllers. Disk controllers bend the economics of persistence through the presence of flash-backed write caches. However, disk controllers also fail - within the NHS environment disk controller failures are the second most common device failure after individual disks. Failures of disk controllers are also expensive to resolve, not being hot-pluggable like disks, and carrying greater risk of node data-loss due to either bad luck or bad process during the change. It is noticeable that EC2 does not have disk controllers and given their failure rate and cost of recovery, this appears to be a sensible trade-off.
|
In physical on-premise server environments there is also commonly the cost of disk controllers. Disk controllers bend the economics of persistence through the presence of flash-backed write caches. However, disk controllers also fail - within the NHS environment disk controller failures are the second most common device failure after individual disks. Failures of disk controllers are also expensive to resolve, not being hot-pluggable like disks, and carrying greater risk of node data-loss due to either bad luck or bad process during the change.
|
||||||
|
|
||||||
|
It is noticeable that EC2 does not have disk controllers and given their failure rate and cost of recovery, this appears to be a sensible trade-off. However, software-only RAID has drawbacks, include the time to setup RAID (24 hours on a d2.2xlarge node), recover from a disk failure and the time to run [scheduled checks](https://www.thomas-krenn.com/en/wiki/Mdadm_checkarray).
|
||||||
|
|
||||||
Making cost-driven decisions about storage design remains as relevant now as it was two decades ago when the LSM-Tree paper was published, especially as we can now directly see those costs reflected in hourly resource charges.
|
Making cost-driven decisions about storage design remains as relevant now as it was two decades ago when the LSM-Tree paper was published, especially as we can now directly see those costs reflected in hourly resource charges.
|
||||||
|
|
||||||
|
|
|
@ -60,7 +60,7 @@ An identical test was run as above, but with d2.2xlarge instances, so that perfo
|
||||||
|
|
||||||
Although append_only writes are being used, almost every write still requires a disk head movement even if the server had all reads handled by in-memory cache (as there are normally more vnodes on the server than there are disk heads). It is clear that without a Flash-Backed Write Cache, spinning disks are unusable as the sole storage mechanism.
|
Although append_only writes are being used, almost every write still requires a disk head movement even if the server had all reads handled by in-memory cache (as there are normally more vnodes on the server than there are disk heads). It is clear that without a Flash-Backed Write Cache, spinning disks are unusable as the sole storage mechanism.
|
||||||
|
|
||||||
Also tested was d2.2zlarge clusters, but without sync_on_write. Results were:
|
Also tested was this same d2.2xlarge cluster, but without sync_on_write. Results were:
|
||||||
|
|
||||||
leveled Results | eleveldb Results
|
leveled Results | eleveldb Results
|
||||||
:-------------------------:|:-------------------------:
|
:-------------------------:|:-------------------------:
|
||||||
|
@ -68,6 +68,8 @@ leveled Results | eleveldb Results
|
||||||
|
|
||||||
This test showed a <b>26.7%</b> improvement in throughput when using LevelEd. The improvement in tail latency in this test had leveled at about <b>25%</b> of the tail latency of leveldb.
|
This test showed a <b>26.7%</b> improvement in throughput when using LevelEd. The improvement in tail latency in this test had leveled at about <b>25%</b> of the tail latency of leveldb.
|
||||||
|
|
||||||
|
This indicates that without sync_on_writes, there remains potential value in using HDD drives, with either leveldb and leveled. It seems reasonably likely (although it remains unproven) that spinning drives with a hardware RAID and flash-backed write cache may perform reasonably even with sync on writes.
|
||||||
|
|
||||||
## Riak Cluster Test - 3
|
## Riak Cluster Test - 3
|
||||||
|
|
||||||
to be completed ..
|
to be completed ..
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue