Home
last modified time | relevance | path

Searched refs:req (Results 1251 - 1275 of 5442) sorted by relevance

1...<<51525354555657585960>>...218

/third_party/node/test/parallel/
H A Dtest-http-same-map.js11 function onrequest(req, res) {
13 onrequest.requests.push(req);
21 const req = http.get({ host, port });
22 req.once('response', (res) => onresponse(n, req, res));
25 function onresponse(n, req, res) {
34 onresponse.requests.push(req);
H A Dtest-http-upgrade-advertise.js28 const req = http.request({
38 req.on('socket', () => req.socket.on('close', () => {
43 req.on('upgrade', function onUpgrade(res, socket) {
49 req.end();
52 const server = http.createServer(function(req, res) {
58 }).on('upgrade', function(req, socket) {
H A Dtest-http-zero-length-write.js64 const server = http.createServer(function(req, res) {
66 req.setEncoding('utf8');
67 req.on('data', function(c) {
70 req.on('end', function() {
78 const req = http.request({ port: this.address().port, method: 'POST' });
80 req.on('response', function(res) {
89 getSrc().pipe(req);
H A Dtest-http2-backpressure.js13 let req;
20 req._readableState.highWaterMark = 20;
24 assert.strictEqual(await event(req, 'data'), 'A'.repeat(5));
25 assert.strictEqual(await event(req, 'data'), 'A'.repeat(40));
38 req = client.request({ ':path': '/' });
39 req.setEncoding('utf8');
40 req.end();
H A Dtest-http2-priority-event.js41 const req = client.request({ ':path': '/' });
44 req.priority({
51 req.on('priority', common.mustCall(onPriority));
53 req.on('response', common.mustCall());
54 req.resume();
55 req.on('end', common.mustCall(() => {
59 req.end();
H A Dtest-https-agent-session-eviction.js21 https.createServer(options, function(req, res) {
30 const req = https.request({
40 req.end();
46 https.createServer(options, function(req, res) {
55 const req = https.request({
65 req.on('response', common.mustCall(function(res) {
70 req.end();
/kernel/linux/linux-5.10/net/netlink/
H A Ddiag.c45 struct netlink_diag_req *req, in sk_diag_fill()
69 if ((req->ndiag_show & NDIAG_SHOW_GROUPS) && in sk_diag_fill()
73 if ((req->ndiag_show & NDIAG_SHOW_MEMINFO) && in sk_diag_fill()
77 if ((req->ndiag_show & NDIAG_SHOW_FLAGS) && in sk_diag_fill()
95 struct netlink_diag_req *req; in __netlink_diag_dump() local
102 req = nlmsg_data(cb->nlh); in __netlink_diag_dump()
137 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump()
167 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump()
187 struct netlink_diag_req *req; in netlink_diag_dump() local
191 req in netlink_diag_dump()
44 sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct netlink_diag_req *req, u32 portid, u32 seq, u32 flags, int sk_ino) sk_diag_fill() argument
[all...]
/kernel/linux/linux-5.10/fs/nilfs2/
H A Dbmap.h179 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_alloc_ptr()
183 return nilfs_dat_prepare_alloc(dat, &req->bpr_req); in nilfs_bmap_prepare_alloc_ptr()
185 req->bpr_ptr = bmap->b_last_allocated_ptr++; in nilfs_bmap_prepare_alloc_ptr()
190 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_alloc_ptr()
194 nilfs_dat_commit_alloc(dat, &req->bpr_req); in nilfs_bmap_commit_alloc_ptr()
198 union nilfs_bmap_ptr_req *req, in nilfs_bmap_abort_alloc_ptr()
202 nilfs_dat_abort_alloc(dat, &req->bpr_req); in nilfs_bmap_abort_alloc_ptr()
208 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_end_ptr()
211 return dat ? nilfs_dat_prepare_end(dat, &req->bpr_req) : 0; in nilfs_bmap_prepare_end_ptr()
215 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_end_ptr()
178 nilfs_bmap_prepare_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_alloc_ptr() argument
189 nilfs_bmap_commit_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_alloc_ptr() argument
197 nilfs_bmap_abort_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_alloc_ptr() argument
207 nilfs_bmap_prepare_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_end_ptr() argument
214 nilfs_bmap_commit_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_end_ptr() argument
223 nilfs_bmap_abort_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_end_ptr() argument
[all...]
/kernel/linux/linux-6.6/fs/nilfs2/
H A Dbmap.h179 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_alloc_ptr()
183 return nilfs_dat_prepare_alloc(dat, &req->bpr_req); in nilfs_bmap_prepare_alloc_ptr()
185 req->bpr_ptr = bmap->b_last_allocated_ptr++; in nilfs_bmap_prepare_alloc_ptr()
190 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_alloc_ptr()
194 nilfs_dat_commit_alloc(dat, &req->bpr_req); in nilfs_bmap_commit_alloc_ptr()
198 union nilfs_bmap_ptr_req *req, in nilfs_bmap_abort_alloc_ptr()
202 nilfs_dat_abort_alloc(dat, &req->bpr_req); in nilfs_bmap_abort_alloc_ptr()
208 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_end_ptr()
211 return dat ? nilfs_dat_prepare_end(dat, &req->bpr_req) : 0; in nilfs_bmap_prepare_end_ptr()
215 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_end_ptr()
178 nilfs_bmap_prepare_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_alloc_ptr() argument
189 nilfs_bmap_commit_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_alloc_ptr() argument
197 nilfs_bmap_abort_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_alloc_ptr() argument
207 nilfs_bmap_prepare_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_end_ptr() argument
214 nilfs_bmap_commit_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_end_ptr() argument
223 nilfs_bmap_abort_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_end_ptr() argument
[all...]
/kernel/linux/linux-6.6/drivers/virt/acrn/
H A Dioeventfd.c190 struct acrn_io_request *req) in acrn_ioeventfd_handler()
196 if (req->type == ACRN_IOREQ_TYPE_MMIO) { in acrn_ioeventfd_handler()
204 if (req->reqs.mmio_request.direction == ACRN_IOREQ_DIR_READ) { in acrn_ioeventfd_handler()
206 req->reqs.mmio_request.value = 0; in acrn_ioeventfd_handler()
209 addr = req->reqs.mmio_request.address; in acrn_ioeventfd_handler()
210 size = req->reqs.mmio_request.size; in acrn_ioeventfd_handler()
211 val = req->reqs.mmio_request.value; in acrn_ioeventfd_handler()
213 if (req->reqs.pio_request.direction == ACRN_IOREQ_DIR_READ) { in acrn_ioeventfd_handler()
215 req->reqs.pio_request.value = 0; in acrn_ioeventfd_handler()
218 addr = req in acrn_ioeventfd_handler()
189 acrn_ioeventfd_handler(struct acrn_ioreq_client *client, struct acrn_io_request *req) acrn_ioeventfd_handler() argument
[all...]
/kernel/linux/linux-6.6/net/netlink/
H A Ddiag.c45 struct netlink_diag_req *req, in sk_diag_fill()
69 if ((req->ndiag_show & NDIAG_SHOW_GROUPS) && in sk_diag_fill()
73 if ((req->ndiag_show & NDIAG_SHOW_MEMINFO) && in sk_diag_fill()
77 if ((req->ndiag_show & NDIAG_SHOW_FLAGS) && in sk_diag_fill()
95 struct netlink_diag_req *req; in __netlink_diag_dump() local
102 req = nlmsg_data(cb->nlh); in __netlink_diag_dump()
137 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump()
167 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump()
187 struct netlink_diag_req *req; in netlink_diag_dump() local
191 req in netlink_diag_dump()
44 sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct netlink_diag_req *req, u32 portid, u32 seq, u32 flags, int sk_ino) sk_diag_fill() argument
[all...]
/third_party/mesa3d/src/freedreno/drm/virtio/
H A Dvirtio_ringbuffer.c67 /* TODO we can get rid of the extra copy into the req by just in flush_submit_list()
155 * into the req msg in flush_submit_list()
160 struct msm_ccmd_gem_submit_req *req = malloc(req_len); in flush_submit_list() local
162 req->hdr = MSM_CCMD(GEM_SUBMIT, req_len); in flush_submit_list()
163 req->flags = virtio_pipe->pipe; in flush_submit_list()
164 req->queue_id = virtio_pipe->queue_id; in flush_submit_list()
165 req->nr_bos = fd_submit->nr_bos; in flush_submit_list()
166 req->nr_cmds = nr_cmds; in flush_submit_list()
167 req->fence = kfence; in flush_submit_list()
169 memcpy(req in flush_submit_list()
[all...]
/kernel/linux/linux-5.10/crypto/
H A Dalgif_hash.c29 struct ahash_request req; member
39 ds = crypto_ahash_digestsize(crypto_ahash_reqtfm(&ctx->req)); in hash_alloc_result()
57 ds = crypto_ahash_digestsize(crypto_ahash_reqtfm(&ctx->req)); in hash_free_result()
81 err = crypto_wait_req(crypto_ahash_init(&ctx->req), &ctx->wait); in hash_sendmsg()
100 ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, NULL, len); in hash_sendmsg()
102 err = crypto_wait_req(crypto_ahash_update(&ctx->req), in hash_sendmsg()
120 ahash_request_set_crypt(&ctx->req, NULL, ctx->result, 0); in hash_sendmsg()
121 err = crypto_wait_req(crypto_ahash_final(&ctx->req), in hash_sendmsg()
153 ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, ctx->result, size); in hash_sendpage()
157 err = crypto_ahash_finup(&ctx->req); in hash_sendpage()
236 struct ahash_request *req = &ctx->req; hash_accept() local
[all...]
/kernel/linux/linux-6.6/arch/arm64/crypto/
H A Dsm4-neon-glue.c38 static int sm4_ecb_do_crypt(struct skcipher_request *req, const u32 *rkey) in sm4_ecb_do_crypt() argument
44 err = skcipher_walk_virt(&walk, req, false); in sm4_ecb_do_crypt()
66 static int sm4_ecb_encrypt(struct skcipher_request *req) in sm4_ecb_encrypt() argument
68 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_ecb_encrypt()
71 return sm4_ecb_do_crypt(req, ctx->rkey_enc); in sm4_ecb_encrypt()
74 static int sm4_ecb_decrypt(struct skcipher_request *req) in sm4_ecb_decrypt() argument
76 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_ecb_decrypt()
79 return sm4_ecb_do_crypt(req, ctx->rkey_dec); in sm4_ecb_decrypt()
82 static int sm4_cbc_encrypt(struct skcipher_request *req) in sm4_cbc_encrypt() argument
84 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_cbc_encrypt()
114 sm4_cbc_decrypt(struct skcipher_request *req) sm4_cbc_decrypt() argument
145 sm4_cfb_encrypt(struct skcipher_request *req) sm4_cfb_encrypt() argument
185 sm4_cfb_decrypt(struct skcipher_request *req) sm4_cfb_decrypt() argument
229 sm4_ctr_crypt(struct skcipher_request *req) sm4_ctr_crypt() argument
[all...]
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/bpf/
H A Dcmsg.c85 struct cmsg_req_map_alloc_tbl *req; in nfp_bpf_ctrl_alloc_map() local
90 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_alloc_map()
94 req = (void *)skb->data; in nfp_bpf_ctrl_alloc_map()
95 req->key_size = cpu_to_be32(map->key_size); in nfp_bpf_ctrl_alloc_map()
96 req->value_size = cpu_to_be32(map->value_size); in nfp_bpf_ctrl_alloc_map()
97 req->max_entries = cpu_to_be32(map->max_entries); in nfp_bpf_ctrl_alloc_map()
98 req->map_type = cpu_to_be32(map->map_type); in nfp_bpf_ctrl_alloc_map()
99 req->map_flags = 0; in nfp_bpf_ctrl_alloc_map()
123 struct cmsg_req_map_free_tbl *req; in nfp_bpf_ctrl_free_map() local
127 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_free_map()
152 nfp_bpf_ctrl_req_key(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_key() argument
159 nfp_bpf_ctrl_req_val(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_val() argument
300 struct cmsg_req_map_op *req; nfp_bpf_ctrl_entry_op() local
[all...]
/kernel/linux/linux-5.10/drivers/net/wireless/ti/wlcore/
H A Dscan.c53 wl->scan.req = NULL; in wl1271_scan_complete_work()
334 struct cfg80211_scan_request *req) in wlcore_scan()
342 BUG_ON(req->n_channels > WL1271_MAX_CHANNELS); in wlcore_scan()
357 wl->scan.req = req; in wlcore_scan()
365 wl->ops->scan_start(wl, wlvif, req); in wlcore_scan()
373 struct cfg80211_sched_scan_request *req) in wlcore_scan_sched_scan_ssid_list()
376 struct cfg80211_match_set *sets = req->match_sets; in wlcore_scan_sched_scan_ssid_list()
377 struct cfg80211_ssid *ssids = req->ssids; in wlcore_scan_sched_scan_ssid_list()
383 for (i = 0; i < req in wlcore_scan_sched_scan_ssid_list()
332 wlcore_scan(struct wl1271 *wl, struct ieee80211_vif *vif, const u8 *ssid, size_t ssid_len, struct cfg80211_scan_request *req) wlcore_scan() argument
371 wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl, struct wl12xx_vif *wlvif, struct cfg80211_sched_scan_request *req) wlcore_scan_sched_scan_ssid_list() argument
[all...]
/kernel/linux/linux-5.10/drivers/mfd/
H A Dezx-pcap.c269 struct pcap_adc_request *req; in pcap_adc_irq() local
274 req = pcap->adc_queue[pcap->adc_head]; in pcap_adc_irq()
276 if (WARN(!req, "adc irq without pending request\n")) { in pcap_adc_irq()
284 tmp |= (req->ch[0] << PCAP_ADC_ADA1_SHIFT); in pcap_adc_irq()
285 tmp |= (req->ch[1] << PCAP_ADC_ADA2_SHIFT); in pcap_adc_irq()
296 req->callback(req->data, res); in pcap_adc_irq()
297 kfree(req); in pcap_adc_irq()
308 struct pcap_adc_request *req; in pcap_adc_async() local
312 req in pcap_adc_async()
342 struct pcap_adc_sync_request *req = param; pcap_adc_sync_cb() local
[all...]
/kernel/linux/linux-6.6/drivers/mfd/
H A Dezx-pcap.c267 struct pcap_adc_request *req; in pcap_adc_irq() local
272 req = pcap->adc_queue[pcap->adc_head]; in pcap_adc_irq()
274 if (WARN(!req, "adc irq without pending request\n")) { in pcap_adc_irq()
282 tmp |= (req->ch[0] << PCAP_ADC_ADA1_SHIFT); in pcap_adc_irq()
283 tmp |= (req->ch[1] << PCAP_ADC_ADA2_SHIFT); in pcap_adc_irq()
294 req->callback(req->data, res); in pcap_adc_irq()
295 kfree(req); in pcap_adc_irq()
306 struct pcap_adc_request *req; in pcap_adc_async() local
310 req in pcap_adc_async()
340 struct pcap_adc_sync_request *req = param; pcap_adc_sync_cb() local
[all...]
/kernel/linux/linux-6.6/drivers/net/ethernet/netronome/nfp/bpf/
H A Dcmsg.c85 struct cmsg_req_map_alloc_tbl *req; in nfp_bpf_ctrl_alloc_map() local
90 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_alloc_map()
94 req = (void *)skb->data; in nfp_bpf_ctrl_alloc_map()
95 req->key_size = cpu_to_be32(map->key_size); in nfp_bpf_ctrl_alloc_map()
96 req->value_size = cpu_to_be32(map->value_size); in nfp_bpf_ctrl_alloc_map()
97 req->max_entries = cpu_to_be32(map->max_entries); in nfp_bpf_ctrl_alloc_map()
98 req->map_type = cpu_to_be32(map->map_type); in nfp_bpf_ctrl_alloc_map()
99 req->map_flags = 0; in nfp_bpf_ctrl_alloc_map()
123 struct cmsg_req_map_free_tbl *req; in nfp_bpf_ctrl_free_map() local
127 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_free_map()
152 nfp_bpf_ctrl_req_key(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_key() argument
159 nfp_bpf_ctrl_req_val(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_val() argument
300 struct cmsg_req_map_op *req; nfp_bpf_ctrl_entry_op() local
[all...]
/kernel/linux/linux-6.6/drivers/net/wireless/ti/wlcore/
H A Dscan.c53 wl->scan.req = NULL; in wl1271_scan_complete_work()
332 struct cfg80211_scan_request *req) in wlcore_scan()
340 BUG_ON(req->n_channels > WL1271_MAX_CHANNELS); in wlcore_scan()
355 wl->scan.req = req; in wlcore_scan()
363 wl->ops->scan_start(wl, wlvif, req); in wlcore_scan()
371 struct cfg80211_sched_scan_request *req) in wlcore_scan_sched_scan_ssid_list()
374 struct cfg80211_match_set *sets = req->match_sets; in wlcore_scan_sched_scan_ssid_list()
375 struct cfg80211_ssid *ssids = req->ssids; in wlcore_scan_sched_scan_ssid_list()
381 for (i = 0; i < req in wlcore_scan_sched_scan_ssid_list()
330 wlcore_scan(struct wl1271 *wl, struct ieee80211_vif *vif, const u8 *ssid, size_t ssid_len, struct cfg80211_scan_request *req) wlcore_scan() argument
369 wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl, struct wl12xx_vif *wlvif, struct cfg80211_sched_scan_request *req) wlcore_scan_sched_scan_ssid_list() argument
[all...]
/third_party/ffmpeg/libavcodec/arm/
H A Dmdct_vfp.S24 CONTEXT .req a1
25 ORIGOUT .req a2
26 IN .req a3
27 OUT .req v1
28 REVTAB .req v2
29 TCOS .req v3
30 TSIN .req v4
31 OLDFPSCR .req v5
32 J0 .req a2
33 J1 .req a
[all...]
/third_party/libdrm/tests/amdgpu/
H A Dcs_tests.c207 struct amdgpu_bo_alloc_request req = {0}; in amdgpu_cs_uvd_create() local
214 req.alloc_size = 4*1024; in amdgpu_cs_uvd_create()
215 req.preferred_heap = AMDGPU_GEM_DOMAIN_GTT; in amdgpu_cs_uvd_create()
217 r = amdgpu_bo_alloc(device_handle, &req, &buf_handle); in amdgpu_cs_uvd_create()
277 struct amdgpu_bo_alloc_request req = {0}; in amdgpu_cs_uvd_decode() local
285 req.alloc_size = 4*1024; /* msg */ in amdgpu_cs_uvd_decode()
286 req.alloc_size += 4*1024; /* fb */ in amdgpu_cs_uvd_decode()
288 req.alloc_size += 4096; /*it_scaling_table*/ in amdgpu_cs_uvd_decode()
289 req.alloc_size += ALIGN(sizeof(uvd_bitstream), 4*1024); in amdgpu_cs_uvd_decode()
290 req in amdgpu_cs_uvd_decode()
418 struct amdgpu_bo_alloc_request req = {0}; amdgpu_cs_uvd_destroy() local
[all...]
/kernel/linux/linux-5.10/drivers/usb/mtu3/
H A Dmtu3_gadget_ep0.c57 static void ep0_req_giveback(struct mtu3 *mtu, struct usb_request *req) in ep0_req_giveback() argument
59 mtu3_req_complete(mtu->ep0, req, 0); in ep0_req_giveback()
167 static void ep0_dummy_complete(struct usb_ep *ep, struct usb_request *req) in ep0_dummy_complete() argument
170 static void ep0_set_sel_complete(struct usb_ep *ep, struct usb_request *req) in ep0_set_sel_complete() argument
176 memcpy(&sel, req->buf, sizeof(sel)); in ep0_set_sel_complete()
178 mreq = to_mtu3_request(req); in ep0_set_sel_complete()
512 struct usb_request *req; in ep0_rx_state() local
522 req = &mreq->request; in ep0_rx_state()
525 if (req) { in ep0_rx_state()
526 void *buf = req in ep0_rx_state()
566 struct usb_request *req; ep0_tx_state() local
836 mtu3_ep0_queue(struct usb_ep *ep, struct usb_request *req, gfp_t gfp) mtu3_ep0_queue() argument
858 mtu3_ep0_dequeue(struct usb_ep *ep, struct usb_request *req) mtu3_ep0_dequeue() argument
[all...]
/kernel/linux/linux-5.10/net/dccp/
H A Dipv4.c198 struct request_sock *req = inet_reqsk(sk); in dccp_req_err() local
205 if (!between48(seq, dccp_rsk(req)->dreq_iss, dccp_rsk(req)->dreq_gss)) { in dccp_req_err()
214 inet_csk_reqsk_queue_drop(req->rsk_listener, req); in dccp_req_err()
216 reqsk_put(req); in dccp_req_err()
397 struct request_sock *req, in dccp_v4_request_recv_sock()
409 newsk = dccp_create_openreq_child(sk, req, skb); in dccp_v4_request_recv_sock()
414 ireq = inet_rsk(req); in dccp_v4_request_recv_sock()
423 if (dst == NULL && (dst = inet_csk_route_child_sock(sk, newsk, req)) in dccp_v4_request_recv_sock()
395 dccp_v4_request_recv_sock(const struct sock *sk, struct sk_buff *skb, struct request_sock *req, struct dst_entry *dst, struct request_sock *req_unhash, bool *own_req) dccp_v4_request_recv_sock() argument
479 dccp_v4_send_response(const struct sock *sk, struct request_sock *req) dccp_v4_send_response() argument
556 dccp_v4_reqsk_destructor(struct request_sock *req) dccp_v4_reqsk_destructor() argument
562 dccp_syn_ack_timeout(const struct request_sock *req) dccp_syn_ack_timeout() argument
580 struct request_sock *req; dccp_v4_conn_request() local
838 struct request_sock *req = inet_reqsk(sk); dccp_v4_rcv() local
[all...]
/kernel/linux/linux-6.6/drivers/usb/mtu3/
H A Dmtu3_gadget_ep0.c57 static void ep0_req_giveback(struct mtu3 *mtu, struct usb_request *req) in ep0_req_giveback() argument
59 mtu3_req_complete(mtu->ep0, req, 0); in ep0_req_giveback()
167 static void ep0_dummy_complete(struct usb_ep *ep, struct usb_request *req) in ep0_dummy_complete() argument
170 static void ep0_set_sel_complete(struct usb_ep *ep, struct usb_request *req) in ep0_set_sel_complete() argument
176 memcpy(&sel, req->buf, sizeof(sel)); in ep0_set_sel_complete()
178 mreq = to_mtu3_request(req); in ep0_set_sel_complete()
512 struct usb_request *req; in ep0_rx_state() local
522 req = &mreq->request; in ep0_rx_state()
525 if (req) { in ep0_rx_state()
526 void *buf = req in ep0_rx_state()
566 struct usb_request *req; ep0_tx_state() local
836 mtu3_ep0_queue(struct usb_ep *ep, struct usb_request *req, gfp_t gfp) mtu3_ep0_queue() argument
858 mtu3_ep0_dequeue(struct usb_ep *ep, struct usb_request *req) mtu3_ep0_dequeue() argument
[all...]

Completed in 13 milliseconds

1...<<51525354555657585960>>...218