I have one such unmentionable benchmark running right now, current depth
is 5127 :-)
> > Please play with it. Andrew, want a version for -mm?
>
> Would be much appreciated, thanks.
Ok, getting on it.
> > */
> > static struct request *get_request_wait(request_queue_t *q, int rw)
> > {
> > - DEFINE_WAIT(wait);
> > - struct request_list *rl = &q->rq[rw];
> > struct request *rq;
> >
> > - spin_lock_prefetch(q->queue_lock);
> > -
> > generic_unplug_device(q);
> > do {
> > - int block = 0;
> > + rq = get_request(q, rw, GFP_NOIO);
> >
> > - prepare_to_wait_exclusive(&rl->wait, &wait,
> > - TASK_UNINTERRUPTIBLE);
> > - spin_lock_irq(q->queue_lock);
> > - if (!rl->count)
> > - block = 1;
> > - spin_unlock_irq(q->queue_lock);
> > -
> > - if (block)
> > + if (!rq)
> > io_schedule();
>
> hmm. I fear that if a SCHED_FIFO/SCHED_RR task hits this, it will just pick
> itself to run again in the schedule() and the box locks up.
>
> A blk_congestion_wait(WRITE, HZ/50) may be better here. It will send the
> caller to sleep until someone puts a write request back, which seems
> appropriate.
Yes good point, I'll make that change.
-- Jens Axboe- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/