Well, file systems have for all time made efforts to write things
out contiguously. If an IO system were to come along and deoptimise
that case it'd be pretty broken?
> ...
>
> There is a lot of difference between doing a raw read of data from
> the device, and copying a large directory tree. The raw read is
> purely sequential, the tree copy is basically random access since
> the kernel is being asked to read and then write a lot of small
> chunks of disk space.
hum. Yes. Copying a tree from and to the same disk won't
achieve disk bandwidth. If they're different disks then
ext2+Orlov will actually get there.
We in fact do extremely well on IDE with Orlov for this workload,
even though we're not doing inter-inode readahead. This will be because
the disk is doing the readhead for us. It's going to be highly dependent
upon the disk cache size, readahead cache partitioning algorithm, etc.
BTW, I've been trying to hunt down a suitable file system aging tool.
We're not very happy with Keith Smith's workload because the directory
infomation was lost (he was purely studying FFS algorithmic differences
- the load isn't 100% suitable for testing other filesystems / algorithms). Constantin Loizides' tools are proving to be rather complex to compile,
drive and understand. Does the XFS team have anything like this in the
kitbag?
-
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/