/third_party/node/test/parallel/ |
H A D | test-http-same-map.js | 11 function onrequest(req, res) { 13 onrequest.requests.push(req); 21 const req = http.get({ host, port }); 22 req.once('response', (res) => onresponse(n, req, res)); 25 function onresponse(n, req, res) { 34 onresponse.requests.push(req);
|
H A D | test-http-upgrade-advertise.js | 28 const req = http.request({ 38 req.on('socket', () => req.socket.on('close', () => { 43 req.on('upgrade', function onUpgrade(res, socket) { 49 req.end(); 52 const server = http.createServer(function(req, res) { 58 }).on('upgrade', function(req, socket) {
|
H A D | test-http-zero-length-write.js | 64 const server = http.createServer(function(req, res) { 66 req.setEncoding('utf8'); 67 req.on('data', function(c) { 70 req.on('end', function() { 78 const req = http.request({ port: this.address().port, method: 'POST' }); 80 req.on('response', function(res) { 89 getSrc().pipe(req);
|
H A D | test-http2-backpressure.js | 13 let req; 20 req._readableState.highWaterMark = 20; 24 assert.strictEqual(await event(req, 'data'), 'A'.repeat(5)); 25 assert.strictEqual(await event(req, 'data'), 'A'.repeat(40)); 38 req = client.request({ ':path': '/' }); 39 req.setEncoding('utf8'); 40 req.end();
|
H A D | test-http2-priority-event.js | 41 const req = client.request({ ':path': '/' }); 44 req.priority({ 51 req.on('priority', common.mustCall(onPriority)); 53 req.on('response', common.mustCall()); 54 req.resume(); 55 req.on('end', common.mustCall(() => { 59 req.end();
|
H A D | test-https-agent-session-eviction.js | 21 https.createServer(options, function(req, res) { 30 const req = https.request({ 40 req.end(); 46 https.createServer(options, function(req, res) { 55 const req = https.request({ 65 req.on('response', common.mustCall(function(res) { 70 req.end();
|
/kernel/linux/linux-5.10/net/netlink/ |
H A D | diag.c | 45 struct netlink_diag_req *req, in sk_diag_fill() 69 if ((req->ndiag_show & NDIAG_SHOW_GROUPS) && in sk_diag_fill() 73 if ((req->ndiag_show & NDIAG_SHOW_MEMINFO) && in sk_diag_fill() 77 if ((req->ndiag_show & NDIAG_SHOW_FLAGS) && in sk_diag_fill() 95 struct netlink_diag_req *req; in __netlink_diag_dump() local 102 req = nlmsg_data(cb->nlh); in __netlink_diag_dump() 137 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump() 167 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump() 187 struct netlink_diag_req *req; in netlink_diag_dump() local 191 req in netlink_diag_dump() 44 sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct netlink_diag_req *req, u32 portid, u32 seq, u32 flags, int sk_ino) sk_diag_fill() argument [all...] |
/kernel/linux/linux-5.10/fs/nilfs2/ |
H A D | bmap.h | 179 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_alloc_ptr() 183 return nilfs_dat_prepare_alloc(dat, &req->bpr_req); in nilfs_bmap_prepare_alloc_ptr() 185 req->bpr_ptr = bmap->b_last_allocated_ptr++; in nilfs_bmap_prepare_alloc_ptr() 190 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_alloc_ptr() 194 nilfs_dat_commit_alloc(dat, &req->bpr_req); in nilfs_bmap_commit_alloc_ptr() 198 union nilfs_bmap_ptr_req *req, in nilfs_bmap_abort_alloc_ptr() 202 nilfs_dat_abort_alloc(dat, &req->bpr_req); in nilfs_bmap_abort_alloc_ptr() 208 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_end_ptr() 211 return dat ? nilfs_dat_prepare_end(dat, &req->bpr_req) : 0; in nilfs_bmap_prepare_end_ptr() 215 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_end_ptr() 178 nilfs_bmap_prepare_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_alloc_ptr() argument 189 nilfs_bmap_commit_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_alloc_ptr() argument 197 nilfs_bmap_abort_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_alloc_ptr() argument 207 nilfs_bmap_prepare_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_end_ptr() argument 214 nilfs_bmap_commit_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_end_ptr() argument 223 nilfs_bmap_abort_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_end_ptr() argument [all...] |
/kernel/linux/linux-6.6/fs/nilfs2/ |
H A D | bmap.h | 179 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_alloc_ptr() 183 return nilfs_dat_prepare_alloc(dat, &req->bpr_req); in nilfs_bmap_prepare_alloc_ptr() 185 req->bpr_ptr = bmap->b_last_allocated_ptr++; in nilfs_bmap_prepare_alloc_ptr() 190 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_alloc_ptr() 194 nilfs_dat_commit_alloc(dat, &req->bpr_req); in nilfs_bmap_commit_alloc_ptr() 198 union nilfs_bmap_ptr_req *req, in nilfs_bmap_abort_alloc_ptr() 202 nilfs_dat_abort_alloc(dat, &req->bpr_req); in nilfs_bmap_abort_alloc_ptr() 208 union nilfs_bmap_ptr_req *req, in nilfs_bmap_prepare_end_ptr() 211 return dat ? nilfs_dat_prepare_end(dat, &req->bpr_req) : 0; in nilfs_bmap_prepare_end_ptr() 215 union nilfs_bmap_ptr_req *req, in nilfs_bmap_commit_end_ptr() 178 nilfs_bmap_prepare_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_alloc_ptr() argument 189 nilfs_bmap_commit_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_alloc_ptr() argument 197 nilfs_bmap_abort_alloc_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_alloc_ptr() argument 207 nilfs_bmap_prepare_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_prepare_end_ptr() argument 214 nilfs_bmap_commit_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_commit_end_ptr() argument 223 nilfs_bmap_abort_end_ptr(struct nilfs_bmap *bmap, union nilfs_bmap_ptr_req *req, struct inode *dat) nilfs_bmap_abort_end_ptr() argument [all...] |
/kernel/linux/linux-6.6/drivers/virt/acrn/ |
H A D | ioeventfd.c | 190 struct acrn_io_request *req) in acrn_ioeventfd_handler() 196 if (req->type == ACRN_IOREQ_TYPE_MMIO) { in acrn_ioeventfd_handler() 204 if (req->reqs.mmio_request.direction == ACRN_IOREQ_DIR_READ) { in acrn_ioeventfd_handler() 206 req->reqs.mmio_request.value = 0; in acrn_ioeventfd_handler() 209 addr = req->reqs.mmio_request.address; in acrn_ioeventfd_handler() 210 size = req->reqs.mmio_request.size; in acrn_ioeventfd_handler() 211 val = req->reqs.mmio_request.value; in acrn_ioeventfd_handler() 213 if (req->reqs.pio_request.direction == ACRN_IOREQ_DIR_READ) { in acrn_ioeventfd_handler() 215 req->reqs.pio_request.value = 0; in acrn_ioeventfd_handler() 218 addr = req in acrn_ioeventfd_handler() 189 acrn_ioeventfd_handler(struct acrn_ioreq_client *client, struct acrn_io_request *req) acrn_ioeventfd_handler() argument [all...] |
/kernel/linux/linux-6.6/net/netlink/ |
H A D | diag.c | 45 struct netlink_diag_req *req, in sk_diag_fill() 69 if ((req->ndiag_show & NDIAG_SHOW_GROUPS) && in sk_diag_fill() 73 if ((req->ndiag_show & NDIAG_SHOW_MEMINFO) && in sk_diag_fill() 77 if ((req->ndiag_show & NDIAG_SHOW_FLAGS) && in sk_diag_fill() 95 struct netlink_diag_req *req; in __netlink_diag_dump() local 102 req = nlmsg_data(cb->nlh); in __netlink_diag_dump() 137 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump() 167 if (sk_diag_fill(sk, skb, req, in __netlink_diag_dump() 187 struct netlink_diag_req *req; in netlink_diag_dump() local 191 req in netlink_diag_dump() 44 sk_diag_fill(struct sock *sk, struct sk_buff *skb, struct netlink_diag_req *req, u32 portid, u32 seq, u32 flags, int sk_ino) sk_diag_fill() argument [all...] |
/third_party/mesa3d/src/freedreno/drm/virtio/ |
H A D | virtio_ringbuffer.c | 67 /* TODO we can get rid of the extra copy into the req by just in flush_submit_list() 155 * into the req msg in flush_submit_list() 160 struct msm_ccmd_gem_submit_req *req = malloc(req_len); in flush_submit_list() local 162 req->hdr = MSM_CCMD(GEM_SUBMIT, req_len); in flush_submit_list() 163 req->flags = virtio_pipe->pipe; in flush_submit_list() 164 req->queue_id = virtio_pipe->queue_id; in flush_submit_list() 165 req->nr_bos = fd_submit->nr_bos; in flush_submit_list() 166 req->nr_cmds = nr_cmds; in flush_submit_list() 167 req->fence = kfence; in flush_submit_list() 169 memcpy(req in flush_submit_list() [all...] |
/kernel/linux/linux-5.10/crypto/ |
H A D | algif_hash.c | 29 struct ahash_request req; member 39 ds = crypto_ahash_digestsize(crypto_ahash_reqtfm(&ctx->req)); in hash_alloc_result() 57 ds = crypto_ahash_digestsize(crypto_ahash_reqtfm(&ctx->req)); in hash_free_result() 81 err = crypto_wait_req(crypto_ahash_init(&ctx->req), &ctx->wait); in hash_sendmsg() 100 ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, NULL, len); in hash_sendmsg() 102 err = crypto_wait_req(crypto_ahash_update(&ctx->req), in hash_sendmsg() 120 ahash_request_set_crypt(&ctx->req, NULL, ctx->result, 0); in hash_sendmsg() 121 err = crypto_wait_req(crypto_ahash_final(&ctx->req), in hash_sendmsg() 153 ahash_request_set_crypt(&ctx->req, ctx->sgl.sg, ctx->result, size); in hash_sendpage() 157 err = crypto_ahash_finup(&ctx->req); in hash_sendpage() 236 struct ahash_request *req = &ctx->req; hash_accept() local [all...] |
/kernel/linux/linux-6.6/arch/arm64/crypto/ |
H A D | sm4-neon-glue.c | 38 static int sm4_ecb_do_crypt(struct skcipher_request *req, const u32 *rkey) in sm4_ecb_do_crypt() argument 44 err = skcipher_walk_virt(&walk, req, false); in sm4_ecb_do_crypt() 66 static int sm4_ecb_encrypt(struct skcipher_request *req) in sm4_ecb_encrypt() argument 68 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_ecb_encrypt() 71 return sm4_ecb_do_crypt(req, ctx->rkey_enc); in sm4_ecb_encrypt() 74 static int sm4_ecb_decrypt(struct skcipher_request *req) in sm4_ecb_decrypt() argument 76 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_ecb_decrypt() 79 return sm4_ecb_do_crypt(req, ctx->rkey_dec); in sm4_ecb_decrypt() 82 static int sm4_cbc_encrypt(struct skcipher_request *req) in sm4_cbc_encrypt() argument 84 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req); in sm4_cbc_encrypt() 114 sm4_cbc_decrypt(struct skcipher_request *req) sm4_cbc_decrypt() argument 145 sm4_cfb_encrypt(struct skcipher_request *req) sm4_cfb_encrypt() argument 185 sm4_cfb_decrypt(struct skcipher_request *req) sm4_cfb_decrypt() argument 229 sm4_ctr_crypt(struct skcipher_request *req) sm4_ctr_crypt() argument [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | cmsg.c | 85 struct cmsg_req_map_alloc_tbl *req; in nfp_bpf_ctrl_alloc_map() local 90 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_alloc_map() 94 req = (void *)skb->data; in nfp_bpf_ctrl_alloc_map() 95 req->key_size = cpu_to_be32(map->key_size); in nfp_bpf_ctrl_alloc_map() 96 req->value_size = cpu_to_be32(map->value_size); in nfp_bpf_ctrl_alloc_map() 97 req->max_entries = cpu_to_be32(map->max_entries); in nfp_bpf_ctrl_alloc_map() 98 req->map_type = cpu_to_be32(map->map_type); in nfp_bpf_ctrl_alloc_map() 99 req->map_flags = 0; in nfp_bpf_ctrl_alloc_map() 123 struct cmsg_req_map_free_tbl *req; in nfp_bpf_ctrl_free_map() local 127 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_free_map() 152 nfp_bpf_ctrl_req_key(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_key() argument 159 nfp_bpf_ctrl_req_val(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_val() argument 300 struct cmsg_req_map_op *req; nfp_bpf_ctrl_entry_op() local [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/ti/wlcore/ |
H A D | scan.c | 53 wl->scan.req = NULL; in wl1271_scan_complete_work() 334 struct cfg80211_scan_request *req) in wlcore_scan() 342 BUG_ON(req->n_channels > WL1271_MAX_CHANNELS); in wlcore_scan() 357 wl->scan.req = req; in wlcore_scan() 365 wl->ops->scan_start(wl, wlvif, req); in wlcore_scan() 373 struct cfg80211_sched_scan_request *req) in wlcore_scan_sched_scan_ssid_list() 376 struct cfg80211_match_set *sets = req->match_sets; in wlcore_scan_sched_scan_ssid_list() 377 struct cfg80211_ssid *ssids = req->ssids; in wlcore_scan_sched_scan_ssid_list() 383 for (i = 0; i < req in wlcore_scan_sched_scan_ssid_list() 332 wlcore_scan(struct wl1271 *wl, struct ieee80211_vif *vif, const u8 *ssid, size_t ssid_len, struct cfg80211_scan_request *req) wlcore_scan() argument 371 wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl, struct wl12xx_vif *wlvif, struct cfg80211_sched_scan_request *req) wlcore_scan_sched_scan_ssid_list() argument [all...] |
/kernel/linux/linux-5.10/drivers/mfd/ |
H A D | ezx-pcap.c | 269 struct pcap_adc_request *req; in pcap_adc_irq() local 274 req = pcap->adc_queue[pcap->adc_head]; in pcap_adc_irq() 276 if (WARN(!req, "adc irq without pending request\n")) { in pcap_adc_irq() 284 tmp |= (req->ch[0] << PCAP_ADC_ADA1_SHIFT); in pcap_adc_irq() 285 tmp |= (req->ch[1] << PCAP_ADC_ADA2_SHIFT); in pcap_adc_irq() 296 req->callback(req->data, res); in pcap_adc_irq() 297 kfree(req); in pcap_adc_irq() 308 struct pcap_adc_request *req; in pcap_adc_async() local 312 req in pcap_adc_async() 342 struct pcap_adc_sync_request *req = param; pcap_adc_sync_cb() local [all...] |
/kernel/linux/linux-6.6/drivers/mfd/ |
H A D | ezx-pcap.c | 267 struct pcap_adc_request *req; in pcap_adc_irq() local 272 req = pcap->adc_queue[pcap->adc_head]; in pcap_adc_irq() 274 if (WARN(!req, "adc irq without pending request\n")) { in pcap_adc_irq() 282 tmp |= (req->ch[0] << PCAP_ADC_ADA1_SHIFT); in pcap_adc_irq() 283 tmp |= (req->ch[1] << PCAP_ADC_ADA2_SHIFT); in pcap_adc_irq() 294 req->callback(req->data, res); in pcap_adc_irq() 295 kfree(req); in pcap_adc_irq() 306 struct pcap_adc_request *req; in pcap_adc_async() local 310 req in pcap_adc_async() 340 struct pcap_adc_sync_request *req = param; pcap_adc_sync_cb() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/netronome/nfp/bpf/ |
H A D | cmsg.c | 85 struct cmsg_req_map_alloc_tbl *req; in nfp_bpf_ctrl_alloc_map() local 90 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_alloc_map() 94 req = (void *)skb->data; in nfp_bpf_ctrl_alloc_map() 95 req->key_size = cpu_to_be32(map->key_size); in nfp_bpf_ctrl_alloc_map() 96 req->value_size = cpu_to_be32(map->value_size); in nfp_bpf_ctrl_alloc_map() 97 req->max_entries = cpu_to_be32(map->max_entries); in nfp_bpf_ctrl_alloc_map() 98 req->map_type = cpu_to_be32(map->map_type); in nfp_bpf_ctrl_alloc_map() 99 req->map_flags = 0; in nfp_bpf_ctrl_alloc_map() 123 struct cmsg_req_map_free_tbl *req; in nfp_bpf_ctrl_free_map() local 127 skb = nfp_bpf_cmsg_alloc(bpf, sizeof(*req)); in nfp_bpf_ctrl_free_map() 152 nfp_bpf_ctrl_req_key(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_key() argument 159 nfp_bpf_ctrl_req_val(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req, unsigned int n) nfp_bpf_ctrl_req_val() argument 300 struct cmsg_req_map_op *req; nfp_bpf_ctrl_entry_op() local [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/ti/wlcore/ |
H A D | scan.c | 53 wl->scan.req = NULL; in wl1271_scan_complete_work() 332 struct cfg80211_scan_request *req) in wlcore_scan() 340 BUG_ON(req->n_channels > WL1271_MAX_CHANNELS); in wlcore_scan() 355 wl->scan.req = req; in wlcore_scan() 363 wl->ops->scan_start(wl, wlvif, req); in wlcore_scan() 371 struct cfg80211_sched_scan_request *req) in wlcore_scan_sched_scan_ssid_list() 374 struct cfg80211_match_set *sets = req->match_sets; in wlcore_scan_sched_scan_ssid_list() 375 struct cfg80211_ssid *ssids = req->ssids; in wlcore_scan_sched_scan_ssid_list() 381 for (i = 0; i < req in wlcore_scan_sched_scan_ssid_list() 330 wlcore_scan(struct wl1271 *wl, struct ieee80211_vif *vif, const u8 *ssid, size_t ssid_len, struct cfg80211_scan_request *req) wlcore_scan() argument 369 wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl, struct wl12xx_vif *wlvif, struct cfg80211_sched_scan_request *req) wlcore_scan_sched_scan_ssid_list() argument [all...] |
/third_party/ffmpeg/libavcodec/arm/ |
H A D | mdct_vfp.S | 24 CONTEXT .req a1 25 ORIGOUT .req a2 26 IN .req a3 27 OUT .req v1 28 REVTAB .req v2 29 TCOS .req v3 30 TSIN .req v4 31 OLDFPSCR .req v5 32 J0 .req a2 33 J1 .req a [all...] |
/third_party/libdrm/tests/amdgpu/ |
H A D | cs_tests.c | 207 struct amdgpu_bo_alloc_request req = {0}; in amdgpu_cs_uvd_create() local 214 req.alloc_size = 4*1024; in amdgpu_cs_uvd_create() 215 req.preferred_heap = AMDGPU_GEM_DOMAIN_GTT; in amdgpu_cs_uvd_create() 217 r = amdgpu_bo_alloc(device_handle, &req, &buf_handle); in amdgpu_cs_uvd_create() 277 struct amdgpu_bo_alloc_request req = {0}; in amdgpu_cs_uvd_decode() local 285 req.alloc_size = 4*1024; /* msg */ in amdgpu_cs_uvd_decode() 286 req.alloc_size += 4*1024; /* fb */ in amdgpu_cs_uvd_decode() 288 req.alloc_size += 4096; /*it_scaling_table*/ in amdgpu_cs_uvd_decode() 289 req.alloc_size += ALIGN(sizeof(uvd_bitstream), 4*1024); in amdgpu_cs_uvd_decode() 290 req in amdgpu_cs_uvd_decode() 418 struct amdgpu_bo_alloc_request req = {0}; amdgpu_cs_uvd_destroy() local [all...] |
/kernel/linux/linux-5.10/drivers/usb/mtu3/ |
H A D | mtu3_gadget_ep0.c | 57 static void ep0_req_giveback(struct mtu3 *mtu, struct usb_request *req) in ep0_req_giveback() argument 59 mtu3_req_complete(mtu->ep0, req, 0); in ep0_req_giveback() 167 static void ep0_dummy_complete(struct usb_ep *ep, struct usb_request *req) in ep0_dummy_complete() argument 170 static void ep0_set_sel_complete(struct usb_ep *ep, struct usb_request *req) in ep0_set_sel_complete() argument 176 memcpy(&sel, req->buf, sizeof(sel)); in ep0_set_sel_complete() 178 mreq = to_mtu3_request(req); in ep0_set_sel_complete() 512 struct usb_request *req; in ep0_rx_state() local 522 req = &mreq->request; in ep0_rx_state() 525 if (req) { in ep0_rx_state() 526 void *buf = req in ep0_rx_state() 566 struct usb_request *req; ep0_tx_state() local 836 mtu3_ep0_queue(struct usb_ep *ep, struct usb_request *req, gfp_t gfp) mtu3_ep0_queue() argument 858 mtu3_ep0_dequeue(struct usb_ep *ep, struct usb_request *req) mtu3_ep0_dequeue() argument [all...] |
/kernel/linux/linux-5.10/net/dccp/ |
H A D | ipv4.c | 198 struct request_sock *req = inet_reqsk(sk); in dccp_req_err() local 205 if (!between48(seq, dccp_rsk(req)->dreq_iss, dccp_rsk(req)->dreq_gss)) { in dccp_req_err() 214 inet_csk_reqsk_queue_drop(req->rsk_listener, req); in dccp_req_err() 216 reqsk_put(req); in dccp_req_err() 397 struct request_sock *req, in dccp_v4_request_recv_sock() 409 newsk = dccp_create_openreq_child(sk, req, skb); in dccp_v4_request_recv_sock() 414 ireq = inet_rsk(req); in dccp_v4_request_recv_sock() 423 if (dst == NULL && (dst = inet_csk_route_child_sock(sk, newsk, req)) in dccp_v4_request_recv_sock() 395 dccp_v4_request_recv_sock(const struct sock *sk, struct sk_buff *skb, struct request_sock *req, struct dst_entry *dst, struct request_sock *req_unhash, bool *own_req) dccp_v4_request_recv_sock() argument 479 dccp_v4_send_response(const struct sock *sk, struct request_sock *req) dccp_v4_send_response() argument 556 dccp_v4_reqsk_destructor(struct request_sock *req) dccp_v4_reqsk_destructor() argument 562 dccp_syn_ack_timeout(const struct request_sock *req) dccp_syn_ack_timeout() argument 580 struct request_sock *req; dccp_v4_conn_request() local 838 struct request_sock *req = inet_reqsk(sk); dccp_v4_rcv() local [all...] |
/kernel/linux/linux-6.6/drivers/usb/mtu3/ |
H A D | mtu3_gadget_ep0.c | 57 static void ep0_req_giveback(struct mtu3 *mtu, struct usb_request *req) in ep0_req_giveback() argument 59 mtu3_req_complete(mtu->ep0, req, 0); in ep0_req_giveback() 167 static void ep0_dummy_complete(struct usb_ep *ep, struct usb_request *req) in ep0_dummy_complete() argument 170 static void ep0_set_sel_complete(struct usb_ep *ep, struct usb_request *req) in ep0_set_sel_complete() argument 176 memcpy(&sel, req->buf, sizeof(sel)); in ep0_set_sel_complete() 178 mreq = to_mtu3_request(req); in ep0_set_sel_complete() 512 struct usb_request *req; in ep0_rx_state() local 522 req = &mreq->request; in ep0_rx_state() 525 if (req) { in ep0_rx_state() 526 void *buf = req in ep0_rx_state() 566 struct usb_request *req; ep0_tx_state() local 836 mtu3_ep0_queue(struct usb_ep *ep, struct usb_request *req, gfp_t gfp) mtu3_ep0_queue() argument 858 mtu3_ep0_dequeue(struct usb_ep *ep, struct usb_request *req) mtu3_ep0_dequeue() argument [all...] |