This only matters if your benchmark is dependent on round-trip latency
(no read-ahead or write behind and no command overlap) or if you have
saturated the bus. None of these are the case with the single drive
I/O benchmarks that have been talked about in this thread. I suppose
I should have been more specific in saying, "the command overhead is
not a factor in the issues raised by this thread". Now if you want
to use command overhead as a reason to use tagged queuing to mitigate
that overhead, by all means, go right ahead.
>> Hooks for sending ordered tags have been in the aic7xxx driver, at
>> least in FreeBSD's version, since '97. As soon as the Linux cmd
>> blocks have such information it will be trivial to have the aic7xxx
>> driver issue the appropriate tag types.
>
> They already do in 2.5, see scsi_populate_tag_msg() in scsi.h. This
> assumes you're using the generic tag queueing, which the aic7xxx
> doesn't, but you could easily key the tag type off REQ_BARRIER.
Okay.
>> But this misses the point. Andrew's original speculation was that
>> writes were "passing reads" once the read was submitted to the drive.
>
> The speculation is based on the observation that for transactions
> consisting of multiple writes and small reads, the reads take a long
> time to complete.
I've seen evidence that a series of reads takes a long time to complete,
but nothing that indicates that every read is starved beyond what you
would expect to see if a huge number of writes were issued between each
read.
> That translates to starving a read in favour of a
> bunch of contiguous writes. I'm sure we've all seen SCSI drives indulge
> in this type of unfair behaviour (it does make sense to keep servicing
> writes if they're direct follow on's from the previously serviced ones).
Actually I haven't. The closest I can come to this is a single read way
off on the far side of the disk starved by a continuous stream or reads
on the other side of the platter. This behavior was fixed by all major
drive manufacturers that I know of back in 97 or 98.
>> I would like to understand the evidence behind that assertion since
>> all drive's I've worked with automatically give a higher priority to
>> read traffic than writes since writes can be buffered but reads
>> cannot.
>
> The evidence is here:
>
> http://marc.theaimsgroup.com/?l=linux-kernel&m=103302456113997&w=1
Which unfortunately characterizes only a single symptom without breaking
it down on a transaction by transaction basis. We need to understand
how many writes were queued by the OS to the drive between each read to
know if the drive is actually allowing writes to pass reads or not.
-- Justin - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/