/kernel/linux/linux-6.6/drivers/vhost/ |
H A D | vringh.c | 43 u16 avail_idx, i, head; in __vringh_get_head() local 61 err = getu16(vrh, &head, &vrh->vring.avail->ring[i]); in __vringh_get_head() 63 vringh_bad("Failed to read head: idx %d address %p", in __vringh_get_head() 68 if (head >= vrh->vring.num) { in __vringh_get_head() 70 head, vrh->vring.num); in __vringh_get_head() 75 return head; in __vringh_get_head() 689 * @head: head index we received, for passing to vringh_complete_user(). 695 * *head will be vrh->vring.num. You may be able to ignore an invalid 708 u16 *head) in vringh_getdesc_user() 703 vringh_getdesc_user(struct vringh *vrh, struct vringh_iov *riov, struct vringh_iov *wiov, bool (*getrange)(struct vringh *vrh, u64 addr, struct vringh_range *r), u16 *head) vringh_getdesc_user() argument 808 vringh_complete_user(struct vringh *vrh, u16 head, u32 len) vringh_complete_user() argument 976 vringh_getdesc_kern(struct vringh *vrh, struct vringh_kiov *riov, struct vringh_kiov *wiov, u16 *head, gfp_t gfp) vringh_getdesc_kern() argument 1056 vringh_complete_kern(struct vringh *vrh, u16 head, u32 len) vringh_complete_kern() argument 1480 vringh_getdesc_iotlb(struct vringh *vrh, struct vringh_kiov *riov, struct vringh_kiov *wiov, u16 *head, gfp_t gfp) vringh_getdesc_iotlb() argument 1566 vringh_complete_iotlb(struct vringh *vrh, u16 head, u32 len) vringh_complete_iotlb() argument [all...] |
/kernel/linux/linux-6.6/drivers/nvme/host/ |
H A D | ioctl.c | 315 c.rw.nsid = cpu_to_le32(ns->head->ns_id); in nvme_submit_io() 331 if (ns && nsid != ns->head->ns_id) { in nvme_validate_passthru_nsid() 335 current->comm, nsid, ns->head->ns_id); in nvme_validate_passthru_nsid() 690 return ns->head->ns_id; in nvme_ns_ioctl() 803 void __user *argp, struct nvme_ns_head *head, int srcu_idx, 805 __releases(&head->srcu) 811 srcu_read_unlock(&head->srcu, srcu_idx); 821 struct nvme_ns_head *head = bdev->bd_disk->private_data; in nvme_ns_head_ioctl() local 831 srcu_idx = srcu_read_lock(&head->srcu); in nvme_ns_head_ioctl() 832 ns = nvme_find_path(head); in nvme_ns_head_ioctl() 856 struct nvme_ns_head *head = nvme_ns_head_chr_ioctl() local 881 struct nvme_ns_head *head = container_of(cdev, struct nvme_ns_head, cdev); nvme_ns_head_chr_uring_cmd() local [all...] |
H A D | nvme.h | 475 static inline bool nvme_ns_head_multipath(struct nvme_ns_head *head) in nvme_ns_head_multipath() argument 477 return IS_ENABLED(CONFIG_NVME_MULTIPATH) && head->disk; in nvme_ns_head_multipath() 498 struct nvme_ns_head *head; member 827 struct nvme_ns_head *head) in nvme_is_unique_nsid() 829 return head->shared || in nvme_is_unique_nsid() 855 bool nvme_tryget_ns_head(struct nvme_ns_head *head); 856 void nvme_put_ns_head(struct nvme_ns_head *head); 887 struct nvme_ns *nvme_find_path(struct nvme_ns_head *head); 900 int nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl,struct nvme_ns_head *head); 902 void nvme_mpath_remove_disk(struct nvme_ns_head *head); 826 nvme_is_unique_nsid(struct nvme_ctrl *ctrl, struct nvme_ns_head *head) nvme_is_unique_nsid() argument 940 nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl, struct nvme_ns_head *head) nvme_mpath_alloc_disk() argument 948 nvme_mpath_remove_disk(struct nvme_ns_head *head) nvme_mpath_remove_disk() argument 961 nvme_mpath_shutdown_disk(struct nvme_ns_head *head) nvme_mpath_shutdown_disk() argument [all...] |
/kernel/linux/linux-5.10/drivers/slimbus/ |
H A D | qcom-ctrl.c | 96 int head; member 137 if ((ctrl->rx.tail + 1) % ctrl->rx.n == ctrl->rx.head) { in slim_alloc_rxbuf() 156 idx = ctrl->tx.head; in slim_ack_txn() 157 ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n; in slim_ack_txn() 310 if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) { in slim_alloc_txbuf() 335 u32 *head; in qcom_xfer_msg() local 353 head = (u32 *)pbuf; in qcom_xfer_msg() 356 *head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, in qcom_xfer_msg() 360 *head in qcom_xfer_msg() [all...] |
/kernel/linux/linux-5.10/net/rds/ |
H A D | ib_recv.c | 103 struct rds_ib_cache_head *head; in rds_ib_recv_alloc_cache() local 111 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_recv_alloc_cache() 112 head->first = NULL; in rds_ib_recv_alloc_cache() 113 head->count = 0; in rds_ib_recv_alloc_cache() 138 struct rds_ib_cache_head *head; in rds_ib_cache_splice_all_lists() local 142 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_cache_splice_all_lists() 143 if (head->first) { in rds_ib_cache_splice_all_lists() 144 list_splice_entire_tail(head->first, caller_list); in rds_ib_cache_splice_all_lists() 145 head->first = NULL; in rds_ib_cache_splice_all_lists() 520 struct list_head *head in rds_ib_recv_cache_get() local [all...] |
/kernel/linux/linux-5.10/fs/nilfs2/ |
H A D | recovery.c | 85 * @bhs: buffer head of start block 196 * @bh_sum: buffer head of summary block 238 * @pbh: the current buffer head on summary blocks [in, out] 267 * @pbh: the current buffer head on summary blocks [in, out] 302 * @head: list head to add nilfs_recovery_block struct 306 struct list_head *head) in nilfs_scan_dsync_log() 359 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log() 375 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument 377 while (!list_empty(head)) { in dispose_recovery_list() 304 nilfs_scan_dsync_log(struct the_nilfs *nilfs, sector_t start_blocknr, struct nilfs_segment_summary *sum, struct list_head *head) nilfs_scan_dsync_log() argument 391 nilfs_segment_list_add(struct list_head *head, __u64 segnum) nilfs_segment_list_add() argument 404 nilfs_dispose_segment_list(struct list_head *head) nilfs_dispose_segment_list() argument 419 struct list_head *head = &ri->ri_used_segments; nilfs_prepare_segment_for_recovery() local 492 nilfs_recover_dsync_blocks(struct the_nilfs *nilfs, struct super_block *sb, struct nilfs_root *root, struct list_head *head, unsigned long *nr_salvaged_blocks) nilfs_recover_dsync_blocks() argument [all...] |
/kernel/linux/linux-5.10/fs/hfsplus/ |
H A D | xattr.c | 58 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local 79 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node() 80 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node() 83 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node() 84 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node() 85 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node() 86 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node() 87 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node() 94 if (be32_to_cpu(head in hfsplus_init_header_node() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_fence.c | 120 list_del_init(&fence->head); in vmw_fence_obj_destroy() 292 list_for_each_entry_safe(action, next_action, &list, head) { in vmw_fence_work_func() 293 list_del_init(&action->head); in vmw_fence_work_func() 355 list_add_tail(&fence->head, &fman->fence_list); in vmw_fence_obj_init() 369 list_for_each_entry_safe(action, next_action, list, head) { in vmw_fences_perform_actions() 370 list_del_init(&action->head); in vmw_fences_perform_actions() 380 list_add_tail(&action->head, &fman->cleanup_list); in vmw_fences_perform_actions() 416 list_for_each_entry(fence, &fman->fence_list, head) { in vmw_fence_goal_new_locked() 475 list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) { in __vmw_fences_update() 477 list_del_init(&fence->head); in __vmw_fences_update() [all...] |
/kernel/linux/linux-6.6/fs/hfsplus/ |
H A D | xattr.c | 58 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local 79 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node() 80 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node() 83 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node() 84 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node() 85 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node() 86 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node() 87 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node() 94 if (be32_to_cpu(head in hfsplus_init_header_node() [all...] |
/kernel/linux/linux-6.6/fs/nilfs2/ |
H A D | recovery.c | 85 * @bhs: buffer head of start block 196 * @bh_sum: buffer head of summary block 238 * @pbh: the current buffer head on summary blocks [in, out] 267 * @pbh: the current buffer head on summary blocks [in, out] 302 * @head: list head to add nilfs_recovery_block struct 306 struct list_head *head) in nilfs_scan_dsync_log() 359 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log() 375 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument 377 while (!list_empty(head)) { in dispose_recovery_list() 304 nilfs_scan_dsync_log(struct the_nilfs *nilfs, sector_t start_blocknr, struct nilfs_segment_summary *sum, struct list_head *head) nilfs_scan_dsync_log() argument 391 nilfs_segment_list_add(struct list_head *head, __u64 segnum) nilfs_segment_list_add() argument 404 nilfs_dispose_segment_list(struct list_head *head) nilfs_dispose_segment_list() argument 419 struct list_head *head = &ri->ri_used_segments; nilfs_prepare_segment_for_recovery() local 492 nilfs_recover_dsync_blocks(struct the_nilfs *nilfs, struct super_block *sb, struct nilfs_root *root, struct list_head *head, unsigned long *nr_salvaged_blocks) nilfs_recover_dsync_blocks() argument [all...] |
/kernel/linux/linux-6.6/drivers/slimbus/ |
H A D | qcom-ctrl.c | 96 int head; member 137 if ((ctrl->rx.tail + 1) % ctrl->rx.n == ctrl->rx.head) { in slim_alloc_rxbuf() 156 idx = ctrl->tx.head; in slim_ack_txn() 157 ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n; in slim_ack_txn() 310 if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) { in slim_alloc_txbuf() 335 u32 *head; in qcom_xfer_msg() local 353 head = (u32 *)pbuf; in qcom_xfer_msg() 356 *head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, in qcom_xfer_msg() 360 *head in qcom_xfer_msg() [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 147 tx_ring->head = 0; in hfi1_ipoib_drain_tx_ring() 159 u32 head = tx_ring->head; in hfi1_ipoib_poll_tx_ring() local 162 struct ipoib_txreq *tx = hfi1_txreq_from_idx(tx_ring, head); in hfi1_ipoib_poll_tx_ring() 170 trace_hfi1_tx_produce(tx, head); in hfi1_ipoib_poll_tx_ring() 172 head = CIRC_NEXT(head, max_tx); in hfi1_ipoib_poll_tx_ring() 173 tx = hfi1_txreq_from_idx(tx_ring, head); in hfi1_ipoib_poll_tx_ring() 177 /* Finished freeing tx items so store the head value. */ in hfi1_ipoib_poll_tx_ring() 178 smp_store_release(&tx_ring->head, hea in hfi1_ipoib_poll_tx_ring() 360 u32 head; hfi1_ipoib_send_dma_common() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_fence.c | 131 list_del_init(&fence->head); in vmw_fence_obj_destroy() 299 list_for_each_entry_safe(action, next_action, &list, head) { in vmw_fence_work_func() 300 list_del_init(&action->head); in vmw_fence_work_func() 357 list_add_tail(&fence->head, &fman->fence_list); in vmw_fence_obj_init() 371 list_for_each_entry_safe(action, next_action, list, head) { in vmw_fences_perform_actions() 372 list_del_init(&action->head); in vmw_fences_perform_actions() 382 list_add_tail(&action->head, &fman->cleanup_list); in vmw_fences_perform_actions() 416 list_for_each_entry(fence, &fman->fence_list, head) { in vmw_fence_goal_new_locked() 472 list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) { in __vmw_fences_update() 474 list_del_init(&fence->head); in __vmw_fences_update() [all...] |
/kernel/linux/linux-6.6/drivers/iommu/intel/ |
H A D | svm.c | 486 int head, tail; in intel_drain_pasid_prq() local 511 head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK; in intel_drain_pasid_prq() 512 while (head != tail) { in intel_drain_pasid_prq() 515 req = &iommu->prq[head / sizeof(*req)]; in intel_drain_pasid_prq() 517 head = (head + sizeof(*req)) & PRQ_RING_MASK; in intel_drain_pasid_prq() 656 int head, tail, handled; in prq_event_thread() local 661 * Clear PPR bit before reading head/tail registers, to ensure that in prq_event_thread() 667 head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK; in prq_event_thread() 668 handled = (head ! in prq_event_thread() [all...] |
/kernel/linux/linux-6.6/net/rds/ |
H A D | ib_recv.c | 104 struct rds_ib_cache_head *head; in rds_ib_recv_alloc_cache() local 112 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_recv_alloc_cache() 113 head->first = NULL; in rds_ib_recv_alloc_cache() 114 head->count = 0; in rds_ib_recv_alloc_cache() 139 struct rds_ib_cache_head *head; in rds_ib_cache_splice_all_lists() local 143 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_cache_splice_all_lists() 144 if (head->first) { in rds_ib_cache_splice_all_lists() 145 list_splice_entire_tail(head->first, caller_list); in rds_ib_cache_splice_all_lists() 146 head->first = NULL; in rds_ib_cache_splice_all_lists() 521 struct list_head *head in rds_ib_recv_cache_get() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | handles.h | 14 #define NV50_DISP_HANDLE_CRC_CTX(head, i) (0xfc000000 | head->base.index << 1 | i)
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | handles.h | 14 #define NV50_DISP_HANDLE_CRC_CTX(head, i) (0xfc000000 | head->base.index << 1 | i)
|
/kernel/linux/linux-6.6/sound/usb/ |
H A D | mixer.c | 130 ((cval)->head.mixer->ignore_ctl_error ? 0 : (err)) 310 struct snd_usb_audio *chip = cval->head.mixer->chip; in get_ctl_value_v1() 321 idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8); in get_ctl_value_v1() 346 struct snd_usb_audio *chip = cval->head.mixer->chip; in get_ctl_value_v2() 368 idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8); in get_ctl_value_v2() 411 return (cval->head.mixer->protocol == UAC_VERSION_1) ? in get_ctl_value() 442 if (!cval->head.mixer->ignore_ctl_error) in snd_usb_get_cur_mix_value() 443 usb_audio_dbg(cval->head in snd_usb_get_cur_mix_value() [all...] |
/kernel/linux/linux-5.10/arch/sparc/kernel/ |
H A D | pci_sun4v.h | 54 unsigned long *head); 57 unsigned long head); 60 unsigned long *head);
|
/kernel/linux/linux-5.10/tools/testing/selftests/arm64/signal/testcases/ |
H A D | testcases.h | 68 * head->size accounts both for payload and header _aarch64_ctx size ! 83 struct _aarch64_ctx *get_header(struct _aarch64_ctx *head, uint32_t magic, 86 static inline struct _aarch64_ctx *get_terminator(struct _aarch64_ctx *head, in get_terminator() argument 90 return get_header(head, 0, resv_sz, offset); in get_terminator()
|
/kernel/linux/linux-6.6/arch/sparc/kernel/ |
H A D | pci_sun4v.h | 54 unsigned long *head); 57 unsigned long head); 60 unsigned long *head);
|
/kernel/linux/linux-5.10/drivers/net/wireless/broadcom/brcm80211/brcmsmac/ |
H A D | dma.h | 113 static inline void dma_spin_for_len(uint len, struct sk_buff *head) in dma_spin_for_len() argument 117 while (!(len = *(u16 *) KSEG1ADDR(head->data))) in dma_spin_for_len() 120 *(u16 *) (head->data) = cpu_to_le16((u16) len); in dma_spin_for_len()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/pm/ |
H A D | priv.h | 15 struct list_head head; member 38 struct list_head head; member 76 struct list_head head; member
|
/kernel/linux/linux-5.10/include/drm/ |
H A D | drm_encoder.h | 89 * @head: list management 101 struct list_head head; member 266 list_for_each_entry((encoder), &(dev)->mode_config.encoder_list, head) \ 277 list_for_each_entry(encoder, &(dev)->mode_config.encoder_list, head)
|
/kernel/linux/linux-5.10/include/dfx/ |
H A D | hungtask_base.h | 107 bool hashlist_find(struct hlist_head *head, int count, pid_t tgid); 108 void hashlist_clear(struct hlist_head *head, int count); 109 bool hashlist_insert(struct hlist_head *head, int count, pid_t tgid);
|