|
@@ -1119,3 +1119,139 @@ retry:
|
|
|
}
|
|
|
|
|
|
struct request *blk_get_request(struct request_queue *q, int rw, gfp_t gfp_mask)
|
|
|
+{
|
|
|
+ struct request *rq;
|
|
|
+
|
|
|
+ BUG_ON(rw != READ && rw != WRITE);
|
|
|
+
|
|
|
+ /* create ioc upfront */
|
|
|
+ create_io_context(gfp_mask, q->node);
|
|
|
+
|
|
|
+ spin_lock_irq(q->queue_lock);
|
|
|
+ rq = get_request(q, rw, NULL, gfp_mask);
|
|
|
+ if (!rq)
|
|
|
+ spin_unlock_irq(q->queue_lock);
|
|
|
+ /* q->queue_lock is unlocked at this point */
|
|
|
+
|
|
|
+ return rq;
|
|
|
+}
|
|
|
+EXPORT_SYMBOL(blk_get_request);
|
|
|
+
|
|
|
+/**
|
|
|
+ * blk_make_request - given a bio, allocate a corresponding struct request.
|
|
|
+ * @q: target request queue
|
|
|
+ * @bio: The bio describing the memory mappings that will be submitted for IO.
|
|
|
+ * It may be a chained-bio properly constructed by block/bio layer.
|
|
|
+ * @gfp_mask: gfp flags to be used for memory allocation
|
|
|
+ *
|
|
|
+ * blk_make_request is the parallel of generic_make_request for BLOCK_PC
|
|
|
+ * type commands. Where the struct request needs to be farther initialized by
|
|
|
+ * the caller. It is passed a &struct bio, which describes the memory info of
|
|
|
+ * the I/O transfer.
|
|
|
+ *
|
|
|
+ * The caller of blk_make_request must make sure that bi_io_vec
|
|
|
+ * are set to describe the memory buffers. That bio_data_dir() will return
|
|
|
+ * the needed direction of the request. (And all bio's in the passed bio-chain
|
|
|
+ * are properly set accordingly)
|
|
|
+ *
|
|
|
+ * If called under none-sleepable conditions, mapped bio buffers must not
|
|
|
+ * need bouncing, by calling the appropriate masked or flagged allocator,
|
|
|
+ * suitable for the target device. Otherwise the call to blk_queue_bounce will
|
|
|
+ * BUG.
|
|
|
+ *
|
|
|
+ * WARNING: When allocating/cloning a bio-chain, careful consideration should be
|
|
|
+ * given to how you allocate bios. In particular, you cannot use __GFP_WAIT for
|
|
|
+ * anything but the first bio in the chain. Otherwise you risk waiting for IO
|
|
|
+ * completion of a bio that hasn't been submitted yet, thus resulting in a
|
|
|
+ * deadlock. Alternatively bios should be allocated using bio_kmalloc() instead
|
|
|
+ * of bio_alloc(), as that avoids the mempool deadlock.
|
|
|
+ * If possible a big IO should be split into smaller parts when allocation
|
|
|
+ * fails. Partial allocation should not be an error, or you risk a live-lock.
|
|
|
+ */
|
|
|
+struct request *blk_make_request(struct request_queue *q, struct bio *bio,
|
|
|
+ gfp_t gfp_mask)
|
|
|
+{
|
|
|
+ struct request *rq = blk_get_request(q, bio_data_dir(bio), gfp_mask);
|
|
|
+
|
|
|
+ if (unlikely(!rq))
|
|
|
+ return ERR_PTR(-ENOMEM);
|
|
|
+
|
|
|
+ for_each_bio(bio) {
|
|
|
+ struct bio *bounce_bio = bio;
|
|
|
+ int ret;
|
|
|
+
|
|
|
+ blk_queue_bounce(q, &bounce_bio);
|
|
|
+ ret = blk_rq_append_bio(q, rq, bounce_bio);
|
|
|
+ if (unlikely(ret)) {
|
|
|
+ blk_put_request(rq);
|
|
|
+ return ERR_PTR(ret);
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return rq;
|
|
|
+}
|
|
|
+EXPORT_SYMBOL(blk_make_request);
|
|
|
+
|
|
|
+/**
|
|
|
+ * blk_requeue_request - put a request back on queue
|
|
|
+ * @q: request queue where request should be inserted
|
|
|
+ * @rq: request to be inserted
|
|
|
+ *
|
|
|
+ * Description:
|
|
|
+ * Drivers often keep queueing requests until the hardware cannot accept
|
|
|
+ * more, when that condition happens we need to put the request back
|
|
|
+ * on the queue. Must be called with queue lock held.
|
|
|
+ */
|
|
|
+void blk_requeue_request(struct request_queue *q, struct request *rq)
|
|
|
+{
|
|
|
+ blk_delete_timer(rq);
|
|
|
+ blk_clear_rq_complete(rq);
|
|
|
+ trace_block_rq_requeue(q, rq);
|
|
|
+
|
|
|
+ if (blk_rq_tagged(rq))
|
|
|
+ blk_queue_end_tag(q, rq);
|
|
|
+
|
|
|
+ BUG_ON(blk_queued_rq(rq));
|
|
|
+
|
|
|
+ elv_requeue_request(q, rq);
|
|
|
+}
|
|
|
+EXPORT_SYMBOL(blk_requeue_request);
|
|
|
+
|
|
|
+static void add_acct_request(struct request_queue *q, struct request *rq,
|
|
|
+ int where)
|
|
|
+{
|
|
|
+ drive_stat_acct(rq, 1);
|
|
|
+ __elv_add_request(q, rq, where);
|
|
|
+}
|
|
|
+
|
|
|
+static void part_round_stats_single(int cpu, struct hd_struct *part,
|
|
|
+ unsigned long now)
|
|
|
+{
|
|
|
+ if (now == part->stamp)
|
|
|
+ return;
|
|
|
+
|
|
|
+ if (part_in_flight(part)) {
|
|
|
+ __part_stat_add(cpu, part, time_in_queue,
|
|
|
+ part_in_flight(part) * (now - part->stamp));
|
|
|
+ __part_stat_add(cpu, part, io_ticks, (now - part->stamp));
|
|
|
+ }
|
|
|
+ part->stamp = now;
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * part_round_stats() - Round off the performance stats on a struct disk_stats.
|
|
|
+ * @cpu: cpu number for stats access
|
|
|
+ * @part: target partition
|
|
|
+ *
|
|
|
+ * The average IO queue length and utilisation statistics are maintained
|
|
|
+ * by observing the current state of the queue length and the amount of
|
|
|
+ * time it has been in this state for.
|
|
|
+ *
|
|
|
+ * Normally, that accounting is done on IO completion, but that can result
|
|
|
+ * in more than a second's worth of IO being accounted for within any one
|
|
|
+ * second, leading to >100% utilisation. To deal with that, we call this
|
|
|
+ * function to do a round-off before returning the results when reading
|
|
|
+ * /proc/diskstats. This accounts immediately for all queue usage up to
|
|
|
+ * the current jiffies and restarts the counters again.
|
|
|
+ */
|
|
|
+void part_round_stats(int cpu, struct hd_struct *part)
|