Lines Matching refs:req

39 static void media_request_clean(struct media_request *req)
44 WARN_ON(req->state != MEDIA_REQUEST_STATE_CLEANING);
45 WARN_ON(req->updating_count);
46 WARN_ON(req->access_count);
48 list_for_each_entry_safe(obj, obj_safe, &req->objects, list) {
53 req->updating_count = 0;
54 req->access_count = 0;
55 WARN_ON(req->num_incomplete_objects);
56 req->num_incomplete_objects = 0;
57 wake_up_interruptible_all(&req->poll_wait);
62 struct media_request *req =
64 struct media_device *mdev = req->mdev;
66 dev_dbg(mdev->dev, "request: release %s\n", req->debug_str);
69 req->state = MEDIA_REQUEST_STATE_CLEANING;
71 media_request_clean(req);
74 mdev->ops->req_free(req);
76 kfree(req);
79 void media_request_put(struct media_request *req)
81 kref_put(&req->kref, media_request_release);
87 struct media_request *req = filp->private_data;
89 media_request_put(req);
96 struct media_request *req = filp->private_data;
103 poll_wait(filp, &req->poll_wait, wait);
104 spin_lock_irqsave(&req->lock, flags);
105 if (req->state == MEDIA_REQUEST_STATE_COMPLETE) {
109 if (req->state != MEDIA_REQUEST_STATE_QUEUED) {
115 spin_unlock_irqrestore(&req->lock, flags);
119 static long media_request_ioctl_queue(struct media_request *req)
121 struct media_device *mdev = req->mdev;
126 dev_dbg(mdev->dev, "request: queue %s\n", req->debug_str);
136 media_request_get(req);
138 spin_lock_irqsave(&req->lock, flags);
139 if (req->state == MEDIA_REQUEST_STATE_IDLE)
140 req->state = MEDIA_REQUEST_STATE_VALIDATING;
141 state = req->state;
142 spin_unlock_irqrestore(&req->lock, flags);
146 req->debug_str, media_request_state_str(state));
147 media_request_put(req);
152 ret = mdev->ops->req_validate(req);
169 spin_lock_irqsave(&req->lock, flags);
170 req->state = ret ? MEDIA_REQUEST_STATE_IDLE
172 spin_unlock_irqrestore(&req->lock, flags);
175 mdev->ops->req_queue(req);
181 req->debug_str, ret);
182 media_request_put(req);
188 static long media_request_ioctl_reinit(struct media_request *req)
190 struct media_device *mdev = req->mdev;
193 spin_lock_irqsave(&req->lock, flags);
194 if (req->state != MEDIA_REQUEST_STATE_IDLE &&
195 req->state != MEDIA_REQUEST_STATE_COMPLETE) {
198 req->debug_str);
199 spin_unlock_irqrestore(&req->lock, flags);
202 if (req->access_count) {
205 req->debug_str);
206 spin_unlock_irqrestore(&req->lock, flags);
209 req->state = MEDIA_REQUEST_STATE_CLEANING;
210 spin_unlock_irqrestore(&req->lock, flags);
212 media_request_clean(req);
214 spin_lock_irqsave(&req->lock, flags);
215 req->state = MEDIA_REQUEST_STATE_IDLE;
216 spin_unlock_irqrestore(&req->lock, flags);
224 struct media_request *req = filp->private_data;
228 return media_request_ioctl_queue(req);
230 return media_request_ioctl_reinit(req);
250 struct media_request *req;
262 req = f.file->private_data;
263 if (req->mdev != mdev)
274 media_request_get(req);
277 return req;
290 struct media_request *req;
300 req = mdev->ops->req_alloc(mdev);
302 req = kzalloc(sizeof(*req), GFP_KERNEL);
303 if (!req)
318 filp->private_data = req;
319 req->mdev = mdev;
320 req->state = MEDIA_REQUEST_STATE_IDLE;
321 req->num_incomplete_objects = 0;
322 kref_init(&req->kref);
323 INIT_LIST_HEAD(&req->objects);
324 spin_lock_init(&req->lock);
325 init_waitqueue_head(&req->poll_wait);
326 req->updating_count = 0;
327 req->access_count = 0;
331 snprintf(req->debug_str, sizeof(req->debug_str), "%u:%d",
333 dev_dbg(mdev->dev, "request: allocated %s\n", req->debug_str);
344 mdev->ops->req_free(req);
346 kfree(req);
355 struct media_request *req = obj->req;
357 if (WARN_ON(req))
363 media_request_object_find(struct media_request *req,
374 spin_lock_irqsave(&req->lock, flags);
375 list_for_each_entry(obj, &req->objects, list) {
382 spin_unlock_irqrestore(&req->lock, flags);
396 obj->req = NULL;
404 int media_request_object_bind(struct media_request *req,
415 spin_lock_irqsave(&req->lock, flags);
417 if (WARN_ON(req->state != MEDIA_REQUEST_STATE_UPDATING &&
418 req->state != MEDIA_REQUEST_STATE_QUEUED))
421 obj->req = req;
426 list_add_tail(&obj->list, &req->objects);
428 list_add(&obj->list, &req->objects);
429 req->num_incomplete_objects++;
433 spin_unlock_irqrestore(&req->lock, flags);
440 struct media_request *req = obj->req;
444 if (WARN_ON(!req))
447 spin_lock_irqsave(&req->lock, flags);
449 obj->req = NULL;
451 if (req->state == MEDIA_REQUEST_STATE_COMPLETE)
454 if (WARN_ON(req->state == MEDIA_REQUEST_STATE_VALIDATING))
457 if (req->state == MEDIA_REQUEST_STATE_CLEANING) {
459 req->num_incomplete_objects--;
463 if (WARN_ON(!req->num_incomplete_objects))
466 req->num_incomplete_objects--;
467 if (req->state == MEDIA_REQUEST_STATE_QUEUED &&
468 !req->num_incomplete_objects) {
469 req->state = MEDIA_REQUEST_STATE_COMPLETE;
471 wake_up_interruptible_all(&req->poll_wait);
475 spin_unlock_irqrestore(&req->lock, flags);
479 media_request_put(req);
485 struct media_request *req = obj->req;
489 spin_lock_irqsave(&req->lock, flags);
493 if (WARN_ON(!req->num_incomplete_objects) ||
494 WARN_ON(req->state != MEDIA_REQUEST_STATE_QUEUED))
497 if (!--req->num_incomplete_objects) {
498 req->state = MEDIA_REQUEST_STATE_COMPLETE;
499 wake_up_interruptible_all(&req->poll_wait);
503 spin_unlock_irqrestore(&req->lock, flags);
505 media_request_put(req);