That depends very much on what you're using the system for. Suppose you're
running a trivial database application on a gigantic disk array - the name of
the game is to cache as much metadata as possible, and that goes directly to
the bottom line as performance. Might as well use 90%+ of your memory for
that.
The conclusion to draw here is, the balance between file cache and process
memory should be able to slide all the way from one extreme to the other.
It's not a requirement that that be fully automatic but it's highly
desireable.
-- Daniel - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/