Mel,
There's no question that that's the case today. However, there are good
reasons for using a largish filesystem blocksize, 16K for example, once it
becomes possible to do so. With an active volume mounted using 16K blocks,
you'd see that the balance of allocations shifts towards order 2. The size
of the shift will be workload-dependent, ranging from almost no order 2
allocations, to almost all. To keep things interesting, it's quite possible
for the balance to change suddenly and/or strongly.
> Because they are so common in comparison to other orders, I
> think that putting order0 in slabs of size 2^MAX_ORDER will make
> defragmentation *so* much easier, if not plain simple, because you can
> shuffle around order0 pages in the slabs to free up one slab which frees
> up one large 2^MAX_ORDER adjacent block of pages.
But how will you shuffle those pages around?
Regards,
Daniel
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/