My fault for being imprecise. By "hit" I meant "hit readahead data in the
buffer cache". Hits on data already in the page cache will work exactly as
they do now, and page cache data that is originally accessed through the page
cache (i.e., whenever we know its mapping) will not take a side trip through
the buffer cache.
> > Aha. So we are going to do the buffer cache hash probe anyway, the thing I
> > call the the reverse lookup. Now I'm just suggesting we drop the other shoe
> > and chain all the page cache blocks to the buffer hash. The only extra cost
> > will be the buffer hash insert and delete, and in return we get complete
> > coherency.
>
> "complete"? No. The coherency is very much one-way: we'd better not have
> anything that actually dirties the buffer cache, because that dirtying
> will _not_ be seen by a virtual cache.
Then I haven't communicated what I had in mind to do. Dirtying data through
the buffer cache would be ok. There are two cases: 1) the block is in the
page cache, but we get to it through the buffer hash chain instead of the
page hash chain. Fine, we can dirty it. 2) the block is not in the page
cache. We can still dirty it, because if the page cache tries to add that
block it will first go look for it in the "buffer cache" (which is really
just a way of saying "not in the page cache") and move it into the page
cache.
> [...]
> Your argument is fundamentally flawed: remember that you cannot actually
> _use_ the data you read ahead before you actually have the linear mapping.
> And you have to get the meta-data information before you can _create_ the
> linear mapping.
You are seeing an inconsistency I'm not seeing.
> Doing physical read-ahead does not mean you can avoid reading meta-data,
> and if you claim that as a "win", you're just lying to yourself.
>
> Doing physical read-ahead only means that you can read-ahead _before_
> reading meta-data: it does not remove the need for meta-data, it only
> potentially removes a ordering dependency.
Right.
> But it potentially removes that ordering dependency through speculation
> (only if the speculation ends up being correct, of course), and
> speculation has a cost. This is nothing new - people have been doing
> things like this at other levels for a long time (doing things like data
> speculation inside CPU's, for example).
>
> Basically, you do not decrease IO - you only try to potentially make it
> more parallel. Which can be a win.
>
> But it can be a big loss too. Speculation always ends up depending ont he
> fact that you have more "throughput" than you actually take advantage of.
> By implication, you can also clearly see that speculation is bound to be
> provably _bad_ in any case where you can already saturate your resources.
Absolutely, so such a facility must shut itself down when there are other
demands for the bandwidth.
> [...]
> Marginally better is not good enough if other loads are marginally worse.
Of course I did not mean that. I meant marginally better on average.
> And I will bet that you _will_ see marginally worse numbers. Which is why
> I think you want "clearly better" just to offset the marginally worse
> numbers.
OK, I'll buy that.
> > > Basically, the things where physical read-ahead might win:
> > > - it can be done without metadata (< 1%)
> >
> > See above. It's not the actual size of the metadata that matters, it's
> > the way it serializes access to the data blocks.
>
> See above. It always will. There is NO WAY you can actually return the
> file data to the user before having read the metadata. QED.
Fine, but you may already have read the *data* by the time you read the
metadata, and so don't have to wait for it, it's already in cache. That's
where the win is.
> You ONLY remove ordering constraints, nothing more.
Agreed. But that's a big deal.
> [...]
> With 2MB of disk cache (which is what all the IDE disks _I_ have access to
> have), the disk will have more of a read-ahead buffer than you'd
> reasonably do in software. How much did you imagine you'd read ahead
> physically? Megabytes? I don't think so..
Yes, I was thinking megabytes in order to get anything close to the 5X
improvement that appears to be possible. But first I was planning to start
more modestly, just pick up a few preemptive data reads here and there that
the logical readahead mechanism would miss, in addition to servicing the
logical readahead just as we do now.
To do the more ambitious multi-megabyte readahead, I had in mind something
that would combine information from some highlevel sources - directory
operations, generic_read, etc - with information picked up at the
physical level, massaging it all into a kind of display from which we'd
be able to determine easily that we have a trend going on like "reading
the whole damm directory". Once we know that's what we're doing it's not
hard to see what to do about it.
Just to do a reality check on that: if somebody reads the first 1/10 of
the files in a directory tree in getdents order, it's a pretty safe bet
they're going to read the other 9/10ths, right?
-- Daniel - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/