git.openpandora.org
/
pandora-kernel.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge git://git.kernel.org/pub/scm/linux/kernel/git/jejb/scsi-rc-fixes-2.6
[pandora-kernel.git]
/
block
/
as-iosched.c
diff --git
a/block/as-iosched.c
b/block/as-iosched.c
index
3e316dd
..
8c39467
100644
(file)
--- a/
block/as-iosched.c
+++ b/
block/as-iosched.c
@@
-170,9
+170,11
@@
static void free_as_io_context(struct as_io_context *aic)
static void as_trim(struct io_context *ioc)
{
static void as_trim(struct io_context *ioc)
{
+ spin_lock_irq(&ioc->lock);
if (ioc->aic)
free_as_io_context(ioc->aic);
ioc->aic = NULL;
if (ioc->aic)
free_as_io_context(ioc->aic);
ioc->aic = NULL;
+ spin_unlock_irq(&ioc->lock);
}
/* Called when the task exits */
}
/* Called when the task exits */
@@
-233,10
+235,12
@@
static void as_put_io_context(struct request *rq)
aic = RQ_IOC(rq)->aic;
if (rq_is_sync(rq) && aic) {
aic = RQ_IOC(rq)->aic;
if (rq_is_sync(rq) && aic) {
- spin_lock(&aic->lock);
+ unsigned long flags;
+
+ spin_lock_irqsave(&aic->lock, flags);
set_bit(AS_TASK_IORUNNING, &aic->state);
aic->last_end_request = jiffies;
set_bit(AS_TASK_IORUNNING, &aic->state);
aic->last_end_request = jiffies;
- spin_unlock
(&aic->lock
);
+ spin_unlock
_irqrestore(&aic->lock, flags
);
}
put_io_context(RQ_IOC(rq));
}
put_io_context(RQ_IOC(rq));
@@
-462,7
+466,9
@@
static void as_antic_timeout(unsigned long data)
spin_lock_irqsave(q->queue_lock, flags);
if (ad->antic_status == ANTIC_WAIT_REQ
|| ad->antic_status == ANTIC_WAIT_NEXT) {
spin_lock_irqsave(q->queue_lock, flags);
if (ad->antic_status == ANTIC_WAIT_REQ
|| ad->antic_status == ANTIC_WAIT_NEXT) {
- struct as_io_context *aic = ad->io_context->aic;
+ struct as_io_context *aic;
+ spin_lock(&ad->io_context->lock);
+ aic = ad->io_context->aic;
ad->antic_status = ANTIC_FINISHED;
kblockd_schedule_work(&ad->antic_work);
ad->antic_status = ANTIC_FINISHED;
kblockd_schedule_work(&ad->antic_work);
@@
-475,6
+481,7
@@
static void as_antic_timeout(unsigned long data)
/* process not "saved" by a cooperating request */
ad->exit_no_coop = (7*ad->exit_no_coop + 256)/8;
}
/* process not "saved" by a cooperating request */
ad->exit_no_coop = (7*ad->exit_no_coop + 256)/8;
}
+ spin_unlock(&ad->io_context->lock);
}
spin_unlock_irqrestore(q->queue_lock, flags);
}
}
spin_unlock_irqrestore(q->queue_lock, flags);
}
@@
-635,9
+642,11
@@
static int as_can_break_anticipation(struct as_data *ad, struct request *rq)
ioc = ad->io_context;
BUG_ON(!ioc);
ioc = ad->io_context;
BUG_ON(!ioc);
+ spin_lock(&ioc->lock);
if (rq && ioc == RQ_IOC(rq)) {
/* request from same process */
if (rq && ioc == RQ_IOC(rq)) {
/* request from same process */
+ spin_unlock(&ioc->lock);
return 1;
}
return 1;
}
@@
-646,20
+655,25
@@
static int as_can_break_anticipation(struct as_data *ad, struct request *rq)
* In this situation status should really be FINISHED,
* however the timer hasn't had the chance to run yet.
*/
* In this situation status should really be FINISHED,
* however the timer hasn't had the chance to run yet.
*/
+ spin_unlock(&ioc->lock);
return 1;
}
aic = ioc->aic;
return 1;
}
aic = ioc->aic;
- if (!aic)
+ if (!aic) {
+ spin_unlock(&ioc->lock);
return 0;
return 0;
+ }
if (atomic_read(&aic->nr_queued) > 0) {
/* process has more requests queued */
if (atomic_read(&aic->nr_queued) > 0) {
/* process has more requests queued */
+ spin_unlock(&ioc->lock);
return 1;
}
if (atomic_read(&aic->nr_dispatched) > 0) {
/* process has more requests dispatched */
return 1;
}
if (atomic_read(&aic->nr_dispatched) > 0) {
/* process has more requests dispatched */
+ spin_unlock(&ioc->lock);
return 1;
}
return 1;
}
@@
-680,6
+694,7
@@
static int as_can_break_anticipation(struct as_data *ad, struct request *rq)
}
as_update_iohist(ad, aic, rq);
}
as_update_iohist(ad, aic, rq);
+ spin_unlock(&ioc->lock);
return 1;
}
return 1;
}
@@
-688,20
+703,27
@@
static int as_can_break_anticipation(struct as_data *ad, struct request *rq)
if (aic->ttime_samples == 0)
ad->exit_prob = (7*ad->exit_prob + 256)/8;
if (aic->ttime_samples == 0)
ad->exit_prob = (7*ad->exit_prob + 256)/8;
- if (ad->exit_no_coop > 128)
+ if (ad->exit_no_coop > 128) {
+ spin_unlock(&ioc->lock);
return 1;
return 1;
+ }
}
if (aic->ttime_samples == 0) {
}
if (aic->ttime_samples == 0) {
- if (ad->new_ttime_mean > ad->antic_expire)
+ if (ad->new_ttime_mean > ad->antic_expire) {
+ spin_unlock(&ioc->lock);
return 1;
return 1;
- if (ad->exit_prob * ad->exit_no_coop > 128*256)
+ }
+ if (ad->exit_prob * ad->exit_no_coop > 128*256) {
+ spin_unlock(&ioc->lock);
return 1;
return 1;
+ }
} else if (aic->ttime_mean > ad->antic_expire) {
/* the process thinks too much between requests */
} else if (aic->ttime_mean > ad->antic_expire) {
/* the process thinks too much between requests */
+ spin_unlock(&ioc->lock);
return 1;
}
return 1;
}
-
+ spin_unlock(&ioc->lock);
return 0;
}
return 0;
}
@@
-796,7
+818,7
@@
static void update_write_batch(struct as_data *ad)
* as_completed_request is to be called when a request has completed and
* returned something to the requesting process, be it an error or data.
*/
* as_completed_request is to be called when a request has completed and
* returned something to the requesting process, be it an error or data.
*/
-static void as_completed_request(
request_queue_t
*q, struct request *rq)
+static void as_completed_request(
struct request_queue
*q, struct request *rq)
{
struct as_data *ad = q->elevator->elevator_data;
{
struct as_data *ad = q->elevator->elevator_data;
@@
-853,7
+875,8
@@
out:
* reference unless it replaces the request at somepart of the elevator
* (ie. the dispatch queue)
*/
* reference unless it replaces the request at somepart of the elevator
* (ie. the dispatch queue)
*/
-static void as_remove_queued_request(request_queue_t *q, struct request *rq)
+static void as_remove_queued_request(struct request_queue *q,
+ struct request *rq)
{
const int data_dir = rq_is_sync(rq);
struct as_data *ad = q->elevator->elevator_data;
{
const int data_dir = rq_is_sync(rq);
struct as_data *ad = q->elevator->elevator_data;
@@
-879,7
+902,7
@@
static void as_remove_queued_request(request_queue_t *q, struct request *rq)
}
/*
}
/*
- * as_fifo_expired returns 0 if there are no expired re
ad
s on the fifo,
+ * as_fifo_expired returns 0 if there are no expired re
quest
s on the fifo,
* 1 otherwise. It is ratelimited so that we only perform the check once per
* `fifo_expire' interval. Otherwise a large number of expired requests
* would create a hopeless seekstorm.
* 1 otherwise. It is ratelimited so that we only perform the check once per
* `fifo_expire' interval. Otherwise a large number of expired requests
* would create a hopeless seekstorm.
@@
-978,7
+1001,7
@@
static void as_move_to_dispatch(struct as_data *ad, struct request *rq)
* read/write expire, batch expire, etc, and moves it to the dispatch
* queue. Returns 1 if a request was found, 0 otherwise.
*/
* read/write expire, batch expire, etc, and moves it to the dispatch
* queue. Returns 1 if a request was found, 0 otherwise.
*/
-static int as_dispatch_request(
request_queue_t
*q, int force)
+static int as_dispatch_request(
struct request_queue
*q, int force)
{
struct as_data *ad = q->elevator->elevator_data;
const int reads = !list_empty(&ad->fifo_list[REQ_SYNC]);
{
struct as_data *ad = q->elevator->elevator_data;
const int reads = !list_empty(&ad->fifo_list[REQ_SYNC]);
@@
-1096,7
+1119,8
@@
dispatch_writes:
ad->batch_data_dir = REQ_ASYNC;
ad->current_write_count = ad->write_batch_count;
ad->write_batch_idled = 0;
ad->batch_data_dir = REQ_ASYNC;
ad->current_write_count = ad->write_batch_count;
ad->write_batch_idled = 0;
- rq = ad->next_rq[ad->batch_data_dir];
+ rq = rq_entry_fifo(ad->fifo_list[REQ_ASYNC].next);
+ ad->last_check_fifo[REQ_ASYNC] = jiffies;
goto dispatch_request;
}
goto dispatch_request;
}
@@
-1139,7
+1163,7
@@
fifo_expired:
/*
* add rq to rbtree and fifo
*/
/*
* add rq to rbtree and fifo
*/
-static void as_add_request(
request_queue_t
*q, struct request *rq)
+static void as_add_request(
struct request_queue
*q, struct request *rq)
{
struct as_data *ad = q->elevator->elevator_data;
int data_dir;
{
struct as_data *ad = q->elevator->elevator_data;
int data_dir;
@@
-1158,7
+1182,7
@@
static void as_add_request(request_queue_t *q, struct request *rq)
as_add_rq_rb(ad, rq);
/*
as_add_rq_rb(ad, rq);
/*
- * set expire time
(only used for reads)
and add to fifo list
+ * set expire time and add to fifo list
*/
rq_set_fifo_time(rq, jiffies + ad->fifo_expire[data_dir]);
list_add_tail(&rq->queuelist, &ad->fifo_list[data_dir]);
*/
rq_set_fifo_time(rq, jiffies + ad->fifo_expire[data_dir]);
list_add_tail(&rq->queuelist, &ad->fifo_list[data_dir]);
@@
-1167,7
+1191,7
@@
static void as_add_request(request_queue_t *q, struct request *rq)
RQ_SET_STATE(rq, AS_RQ_QUEUED);
}
RQ_SET_STATE(rq, AS_RQ_QUEUED);
}
-static void as_activate_request(
request_queue_t
*q, struct request *rq)
+static void as_activate_request(
struct request_queue
*q, struct request *rq)
{
WARN_ON(RQ_STATE(rq) != AS_RQ_DISPATCHED);
RQ_SET_STATE(rq, AS_RQ_REMOVED);
{
WARN_ON(RQ_STATE(rq) != AS_RQ_DISPATCHED);
RQ_SET_STATE(rq, AS_RQ_REMOVED);
@@
-1175,7
+1199,7
@@
static void as_activate_request(request_queue_t *q, struct request *rq)
atomic_dec(&RQ_IOC(rq)->aic->nr_dispatched);
}
atomic_dec(&RQ_IOC(rq)->aic->nr_dispatched);
}
-static void as_deactivate_request(
request_queue_t
*q, struct request *rq)
+static void as_deactivate_request(
struct request_queue
*q, struct request *rq)
{
WARN_ON(RQ_STATE(rq) != AS_RQ_REMOVED);
RQ_SET_STATE(rq, AS_RQ_DISPATCHED);
{
WARN_ON(RQ_STATE(rq) != AS_RQ_REMOVED);
RQ_SET_STATE(rq, AS_RQ_DISPATCHED);
@@
-1189,7
+1213,7
@@
static void as_deactivate_request(request_queue_t *q, struct request *rq)
* is not empty - it is used in the block layer to check for plugging and
* merging opportunities
*/
* is not empty - it is used in the block layer to check for plugging and
* merging opportunities
*/
-static int as_queue_empty(
request_queue_t
*q)
+static int as_queue_empty(
struct request_queue
*q)
{
struct as_data *ad = q->elevator->elevator_data;
{
struct as_data *ad = q->elevator->elevator_data;
@@
-1198,7
+1222,7
@@
static int as_queue_empty(request_queue_t *q)
}
static int
}
static int
-as_merge(
request_queue_t
*q, struct request **req, struct bio *bio)
+as_merge(
struct request_queue
*q, struct request **req, struct bio *bio)
{
struct as_data *ad = q->elevator->elevator_data;
sector_t rb_key = bio->bi_sector + bio_sectors(bio);
{
struct as_data *ad = q->elevator->elevator_data;
sector_t rb_key = bio->bi_sector + bio_sectors(bio);
@@
-1216,7
+1240,8
@@
as_merge(request_queue_t *q, struct request **req, struct bio *bio)
return ELEVATOR_NO_MERGE;
}
return ELEVATOR_NO_MERGE;
}
-static void as_merged_request(request_queue_t *q, struct request *req, int type)
+static void as_merged_request(struct request_queue *q, struct request *req,
+ int type)
{
struct as_data *ad = q->elevator->elevator_data;
{
struct as_data *ad = q->elevator->elevator_data;
@@
-1234,7
+1259,7
@@
static void as_merged_request(request_queue_t *q, struct request *req, int type)
}
}
}
}
-static void as_merged_requests(
request_queue_t
*q, struct request *req,
+static void as_merged_requests(
struct request_queue
*q, struct request *req,
struct request *next)
{
/*
struct request *next)
{
/*
@@
-1243,16
+1268,8
@@
static void as_merged_requests(request_queue_t *q, struct request *req,
*/
if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) {
if (time_before(rq_fifo_time(next), rq_fifo_time(req))) {
*/
if (!list_empty(&req->queuelist) && !list_empty(&next->queuelist)) {
if (time_before(rq_fifo_time(next), rq_fifo_time(req))) {
- struct io_context *rioc = RQ_IOC(req);
- struct io_context *nioc = RQ_IOC(next);
-
list_move(&req->queuelist, &next->queuelist);
rq_set_fifo_time(req, rq_fifo_time(next));
list_move(&req->queuelist, &next->queuelist);
rq_set_fifo_time(req, rq_fifo_time(next));
- /*
- * Don't copy here but swap, because when anext is
- * removed below, it must contain the unused context
- */
- swap_io_context(&rioc, &nioc);
}
}
}
}
@@
-1285,7
+1302,7
@@
static void as_work_handler(struct work_struct *work)
spin_unlock_irqrestore(q->queue_lock, flags);
}
spin_unlock_irqrestore(q->queue_lock, flags);
}
-static int as_may_queue(
request_queue_t
*q, int rw)
+static int as_may_queue(
struct request_queue
*q, int rw)
{
int ret = ELV_MQUEUE_MAY;
struct as_data *ad = q->elevator->elevator_data;
{
int ret = ELV_MQUEUE_MAY;
struct as_data *ad = q->elevator->elevator_data;
@@
-1318,7
+1335,7
@@
static void as_exit_queue(elevator_t *e)
/*
* initialize elevator private data (as_data).
*/
/*
* initialize elevator private data (as_data).
*/
-static void *as_init_queue(
request_queue_t
*q)
+static void *as_init_queue(
struct request_queue
*q)
{
struct as_data *ad;
{
struct as_data *ad;
@@
-1461,7
+1478,9
@@
static struct elevator_type iosched_as = {
static int __init as_init(void)
{
static int __init as_init(void)
{
- return elv_register(&iosched_as);
+ elv_register(&iosched_as);
+
+ return 0;
}
static void __exit as_exit(void)
}
static void __exit as_exit(void)