Lines Matching refs:req

29 	struct xensnd_req *req;
31 req = RING_GET_REQUEST(&evtchnl->u.req.ring,
32 evtchnl->u.req.ring.req_prod_pvt);
33 req->operation = operation;
34 req->id = evtchnl->evt_next_id++;
35 evtchnl->evt_id = req->id;
36 return req;
44 reinit_completion(&evtchnl->u.req.completion);
51 if (wait_for_completion_timeout(&evtchnl->u.req.completion,
55 return evtchnl->u.req.resp_status;
62 struct xensnd_req *req;
65 mutex_lock(&evtchnl->u.req.req_io_lock);
68 req = be_stream_prepare_req(evtchnl, XENSND_OP_HW_PARAM_QUERY);
69 req->op.hw_param = *hw_param_req;
78 *hw_param_resp = evtchnl->u.req.resp.hw_param;
80 mutex_unlock(&evtchnl->u.req.req_io_lock);
90 struct xensnd_req *req;
93 mutex_lock(&evtchnl->u.req.req_io_lock);
96 req = be_stream_prepare_req(evtchnl, XENSND_OP_OPEN);
97 req->op.open.pcm_format = format;
98 req->op.open.pcm_channels = channels;
99 req->op.open.pcm_rate = rate;
100 req->op.open.buffer_sz = buffer_sz;
101 req->op.open.period_sz = period_sz;
102 req->op.open.gref_directory =
111 mutex_unlock(&evtchnl->u.req.req_io_lock);
117 __always_unused struct xensnd_req *req;
120 mutex_lock(&evtchnl->u.req.req_io_lock);
123 req = be_stream_prepare_req(evtchnl, XENSND_OP_CLOSE);
131 mutex_unlock(&evtchnl->u.req.req_io_lock);
138 struct xensnd_req *req;
141 mutex_lock(&evtchnl->u.req.req_io_lock);
144 req = be_stream_prepare_req(evtchnl, XENSND_OP_WRITE);
145 req->op.rw.length = count;
146 req->op.rw.offset = pos;
154 mutex_unlock(&evtchnl->u.req.req_io_lock);
161 struct xensnd_req *req;
164 mutex_lock(&evtchnl->u.req.req_io_lock);
167 req = be_stream_prepare_req(evtchnl, XENSND_OP_READ);
168 req->op.rw.length = count;
169 req->op.rw.offset = pos;
177 mutex_unlock(&evtchnl->u.req.req_io_lock);
184 struct xensnd_req *req;
187 mutex_lock(&evtchnl->u.req.req_io_lock);
190 req = be_stream_prepare_req(evtchnl, XENSND_OP_TRIGGER);
191 req->op.trigger.type = type;
199 mutex_unlock(&evtchnl->u.req.req_io_lock);