Home
last modified time | relevance | path

Searched refs:head (Results 826 - 850 of 4981) sorted by relevance

1...<<31323334353637383940>>...200

/kernel/linux/linux-6.6/drivers/vhost/
H A Dvringh.c43 u16 avail_idx, i, head; in __vringh_get_head() local
61 err = getu16(vrh, &head, &vrh->vring.avail->ring[i]); in __vringh_get_head()
63 vringh_bad("Failed to read head: idx %d address %p", in __vringh_get_head()
68 if (head >= vrh->vring.num) { in __vringh_get_head()
70 head, vrh->vring.num); in __vringh_get_head()
75 return head; in __vringh_get_head()
689 * @head: head index we received, for passing to vringh_complete_user().
695 * *head will be vrh->vring.num. You may be able to ignore an invalid
708 u16 *head) in vringh_getdesc_user()
703 vringh_getdesc_user(struct vringh *vrh, struct vringh_iov *riov, struct vringh_iov *wiov, bool (*getrange)(struct vringh *vrh, u64 addr, struct vringh_range *r), u16 *head) vringh_getdesc_user() argument
808 vringh_complete_user(struct vringh *vrh, u16 head, u32 len) vringh_complete_user() argument
976 vringh_getdesc_kern(struct vringh *vrh, struct vringh_kiov *riov, struct vringh_kiov *wiov, u16 *head, gfp_t gfp) vringh_getdesc_kern() argument
1056 vringh_complete_kern(struct vringh *vrh, u16 head, u32 len) vringh_complete_kern() argument
1480 vringh_getdesc_iotlb(struct vringh *vrh, struct vringh_kiov *riov, struct vringh_kiov *wiov, u16 *head, gfp_t gfp) vringh_getdesc_iotlb() argument
1566 vringh_complete_iotlb(struct vringh *vrh, u16 head, u32 len) vringh_complete_iotlb() argument
[all...]
/kernel/linux/linux-6.6/drivers/nvme/host/
H A Dioctl.c315 c.rw.nsid = cpu_to_le32(ns->head->ns_id); in nvme_submit_io()
331 if (ns && nsid != ns->head->ns_id) { in nvme_validate_passthru_nsid()
335 current->comm, nsid, ns->head->ns_id); in nvme_validate_passthru_nsid()
690 return ns->head->ns_id; in nvme_ns_ioctl()
803 void __user *argp, struct nvme_ns_head *head, int srcu_idx,
805 __releases(&head->srcu)
811 srcu_read_unlock(&head->srcu, srcu_idx);
821 struct nvme_ns_head *head = bdev->bd_disk->private_data; in nvme_ns_head_ioctl() local
831 srcu_idx = srcu_read_lock(&head->srcu); in nvme_ns_head_ioctl()
832 ns = nvme_find_path(head); in nvme_ns_head_ioctl()
856 struct nvme_ns_head *head = nvme_ns_head_chr_ioctl() local
881 struct nvme_ns_head *head = container_of(cdev, struct nvme_ns_head, cdev); nvme_ns_head_chr_uring_cmd() local
[all...]
H A Dnvme.h475 static inline bool nvme_ns_head_multipath(struct nvme_ns_head *head) in nvme_ns_head_multipath() argument
477 return IS_ENABLED(CONFIG_NVME_MULTIPATH) && head->disk; in nvme_ns_head_multipath()
498 struct nvme_ns_head *head; member
827 struct nvme_ns_head *head) in nvme_is_unique_nsid()
829 return head->shared || in nvme_is_unique_nsid()
855 bool nvme_tryget_ns_head(struct nvme_ns_head *head);
856 void nvme_put_ns_head(struct nvme_ns_head *head);
887 struct nvme_ns *nvme_find_path(struct nvme_ns_head *head);
900 int nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl,struct nvme_ns_head *head);
902 void nvme_mpath_remove_disk(struct nvme_ns_head *head);
826 nvme_is_unique_nsid(struct nvme_ctrl *ctrl, struct nvme_ns_head *head) nvme_is_unique_nsid() argument
940 nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl, struct nvme_ns_head *head) nvme_mpath_alloc_disk() argument
948 nvme_mpath_remove_disk(struct nvme_ns_head *head) nvme_mpath_remove_disk() argument
961 nvme_mpath_shutdown_disk(struct nvme_ns_head *head) nvme_mpath_shutdown_disk() argument
[all...]
/kernel/linux/linux-5.10/drivers/slimbus/
H A Dqcom-ctrl.c96 int head; member
137 if ((ctrl->rx.tail + 1) % ctrl->rx.n == ctrl->rx.head) { in slim_alloc_rxbuf()
156 idx = ctrl->tx.head; in slim_ack_txn()
157 ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n; in slim_ack_txn()
310 if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) { in slim_alloc_txbuf()
335 u32 *head; in qcom_xfer_msg() local
353 head = (u32 *)pbuf; in qcom_xfer_msg()
356 *head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, in qcom_xfer_msg()
360 *head in qcom_xfer_msg()
[all...]
/kernel/linux/linux-5.10/net/rds/
H A Dib_recv.c103 struct rds_ib_cache_head *head; in rds_ib_recv_alloc_cache() local
111 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_recv_alloc_cache()
112 head->first = NULL; in rds_ib_recv_alloc_cache()
113 head->count = 0; in rds_ib_recv_alloc_cache()
138 struct rds_ib_cache_head *head; in rds_ib_cache_splice_all_lists() local
142 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_cache_splice_all_lists()
143 if (head->first) { in rds_ib_cache_splice_all_lists()
144 list_splice_entire_tail(head->first, caller_list); in rds_ib_cache_splice_all_lists()
145 head->first = NULL; in rds_ib_cache_splice_all_lists()
520 struct list_head *head in rds_ib_recv_cache_get() local
[all...]
/kernel/linux/linux-5.10/fs/nilfs2/
H A Drecovery.c85 * @bhs: buffer head of start block
196 * @bh_sum: buffer head of summary block
238 * @pbh: the current buffer head on summary blocks [in, out]
267 * @pbh: the current buffer head on summary blocks [in, out]
302 * @head: list head to add nilfs_recovery_block struct
306 struct list_head *head) in nilfs_scan_dsync_log()
359 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log()
375 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument
377 while (!list_empty(head)) { in dispose_recovery_list()
304 nilfs_scan_dsync_log(struct the_nilfs *nilfs, sector_t start_blocknr, struct nilfs_segment_summary *sum, struct list_head *head) nilfs_scan_dsync_log() argument
391 nilfs_segment_list_add(struct list_head *head, __u64 segnum) nilfs_segment_list_add() argument
404 nilfs_dispose_segment_list(struct list_head *head) nilfs_dispose_segment_list() argument
419 struct list_head *head = &ri->ri_used_segments; nilfs_prepare_segment_for_recovery() local
492 nilfs_recover_dsync_blocks(struct the_nilfs *nilfs, struct super_block *sb, struct nilfs_root *root, struct list_head *head, unsigned long *nr_salvaged_blocks) nilfs_recover_dsync_blocks() argument
[all...]
/kernel/linux/linux-5.10/fs/hfsplus/
H A Dxattr.c58 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local
79 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node()
80 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node()
83 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node()
84 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node()
85 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node()
86 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node()
87 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node()
94 if (be32_to_cpu(head in hfsplus_init_header_node()
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_fence.c120 list_del_init(&fence->head); in vmw_fence_obj_destroy()
292 list_for_each_entry_safe(action, next_action, &list, head) { in vmw_fence_work_func()
293 list_del_init(&action->head); in vmw_fence_work_func()
355 list_add_tail(&fence->head, &fman->fence_list); in vmw_fence_obj_init()
369 list_for_each_entry_safe(action, next_action, list, head) { in vmw_fences_perform_actions()
370 list_del_init(&action->head); in vmw_fences_perform_actions()
380 list_add_tail(&action->head, &fman->cleanup_list); in vmw_fences_perform_actions()
416 list_for_each_entry(fence, &fman->fence_list, head) { in vmw_fence_goal_new_locked()
475 list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) { in __vmw_fences_update()
477 list_del_init(&fence->head); in __vmw_fences_update()
[all...]
/kernel/linux/linux-6.6/fs/hfsplus/
H A Dxattr.c58 struct hfs_btree_header_rec *head; in hfsplus_init_header_node() local
79 head = (struct hfs_btree_header_rec *)(buf + offset); in hfsplus_init_header_node()
80 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node()
83 head->node_count = cpu_to_be32(tmp); in hfsplus_init_header_node()
84 head->free_nodes = cpu_to_be32(be32_to_cpu(head->node_count) - 1); in hfsplus_init_header_node()
85 head->clump_size = cpu_to_be32(clump_size); in hfsplus_init_header_node()
86 head->attributes |= cpu_to_be32(HFS_TREE_BIGKEYS | HFS_TREE_VARIDXKEYS); in hfsplus_init_header_node()
87 head->max_key_len = cpu_to_be16(HFSPLUS_ATTR_KEYLEN - sizeof(u16)); in hfsplus_init_header_node()
94 if (be32_to_cpu(head in hfsplus_init_header_node()
[all...]
/kernel/linux/linux-6.6/fs/nilfs2/
H A Drecovery.c85 * @bhs: buffer head of start block
196 * @bh_sum: buffer head of summary block
238 * @pbh: the current buffer head on summary blocks [in, out]
267 * @pbh: the current buffer head on summary blocks [in, out]
302 * @head: list head to add nilfs_recovery_block struct
306 struct list_head *head) in nilfs_scan_dsync_log()
359 list_add_tail(&rb->list, head); in nilfs_scan_dsync_log()
375 static void dispose_recovery_list(struct list_head *head) in dispose_recovery_list() argument
377 while (!list_empty(head)) { in dispose_recovery_list()
304 nilfs_scan_dsync_log(struct the_nilfs *nilfs, sector_t start_blocknr, struct nilfs_segment_summary *sum, struct list_head *head) nilfs_scan_dsync_log() argument
391 nilfs_segment_list_add(struct list_head *head, __u64 segnum) nilfs_segment_list_add() argument
404 nilfs_dispose_segment_list(struct list_head *head) nilfs_dispose_segment_list() argument
419 struct list_head *head = &ri->ri_used_segments; nilfs_prepare_segment_for_recovery() local
492 nilfs_recover_dsync_blocks(struct the_nilfs *nilfs, struct super_block *sb, struct nilfs_root *root, struct list_head *head, unsigned long *nr_salvaged_blocks) nilfs_recover_dsync_blocks() argument
[all...]
/kernel/linux/linux-6.6/drivers/slimbus/
H A Dqcom-ctrl.c96 int head; member
137 if ((ctrl->rx.tail + 1) % ctrl->rx.n == ctrl->rx.head) { in slim_alloc_rxbuf()
156 idx = ctrl->tx.head; in slim_ack_txn()
157 ctrl->tx.head = (ctrl->tx.head + 1) % ctrl->tx.n; in slim_ack_txn()
310 if (((ctrl->tx.head + 1) % ctrl->tx.n) == ctrl->tx.tail) { in slim_alloc_txbuf()
335 u32 *head; in qcom_xfer_msg() local
353 head = (u32 *)pbuf; in qcom_xfer_msg()
356 *head = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, in qcom_xfer_msg()
360 *head in qcom_xfer_msg()
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/hw/hfi1/
H A Dipoib_tx.c147 tx_ring->head = 0; in hfi1_ipoib_drain_tx_ring()
159 u32 head = tx_ring->head; in hfi1_ipoib_poll_tx_ring() local
162 struct ipoib_txreq *tx = hfi1_txreq_from_idx(tx_ring, head); in hfi1_ipoib_poll_tx_ring()
170 trace_hfi1_tx_produce(tx, head); in hfi1_ipoib_poll_tx_ring()
172 head = CIRC_NEXT(head, max_tx); in hfi1_ipoib_poll_tx_ring()
173 tx = hfi1_txreq_from_idx(tx_ring, head); in hfi1_ipoib_poll_tx_ring()
177 /* Finished freeing tx items so store the head value. */ in hfi1_ipoib_poll_tx_ring()
178 smp_store_release(&tx_ring->head, hea in hfi1_ipoib_poll_tx_ring()
360 u32 head; hfi1_ipoib_send_dma_common() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_fence.c131 list_del_init(&fence->head); in vmw_fence_obj_destroy()
299 list_for_each_entry_safe(action, next_action, &list, head) { in vmw_fence_work_func()
300 list_del_init(&action->head); in vmw_fence_work_func()
357 list_add_tail(&fence->head, &fman->fence_list); in vmw_fence_obj_init()
371 list_for_each_entry_safe(action, next_action, list, head) { in vmw_fences_perform_actions()
372 list_del_init(&action->head); in vmw_fences_perform_actions()
382 list_add_tail(&action->head, &fman->cleanup_list); in vmw_fences_perform_actions()
416 list_for_each_entry(fence, &fman->fence_list, head) { in vmw_fence_goal_new_locked()
472 list_for_each_entry_safe(fence, next_fence, &fman->fence_list, head) { in __vmw_fences_update()
474 list_del_init(&fence->head); in __vmw_fences_update()
[all...]
/kernel/linux/linux-6.6/drivers/iommu/intel/
H A Dsvm.c486 int head, tail; in intel_drain_pasid_prq() local
511 head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK; in intel_drain_pasid_prq()
512 while (head != tail) { in intel_drain_pasid_prq()
515 req = &iommu->prq[head / sizeof(*req)]; in intel_drain_pasid_prq()
517 head = (head + sizeof(*req)) & PRQ_RING_MASK; in intel_drain_pasid_prq()
656 int head, tail, handled; in prq_event_thread() local
661 * Clear PPR bit before reading head/tail registers, to ensure that in prq_event_thread()
667 head = dmar_readq(iommu->reg + DMAR_PQH_REG) & PRQ_RING_MASK; in prq_event_thread()
668 handled = (head ! in prq_event_thread()
[all...]
/kernel/linux/linux-6.6/net/rds/
H A Dib_recv.c104 struct rds_ib_cache_head *head; in rds_ib_recv_alloc_cache() local
112 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_recv_alloc_cache()
113 head->first = NULL; in rds_ib_recv_alloc_cache()
114 head->count = 0; in rds_ib_recv_alloc_cache()
139 struct rds_ib_cache_head *head; in rds_ib_cache_splice_all_lists() local
143 head = per_cpu_ptr(cache->percpu, cpu); in rds_ib_cache_splice_all_lists()
144 if (head->first) { in rds_ib_cache_splice_all_lists()
145 list_splice_entire_tail(head->first, caller_list); in rds_ib_cache_splice_all_lists()
146 head->first = NULL; in rds_ib_cache_splice_all_lists()
521 struct list_head *head in rds_ib_recv_cache_get() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/dispnv50/
H A Dhandles.h14 #define NV50_DISP_HANDLE_CRC_CTX(head, i) (0xfc000000 | head->base.index << 1 | i)
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/dispnv50/
H A Dhandles.h14 #define NV50_DISP_HANDLE_CRC_CTX(head, i) (0xfc000000 | head->base.index << 1 | i)
/kernel/linux/linux-6.6/sound/usb/
H A Dmixer.c130 ((cval)->head.mixer->ignore_ctl_error ? 0 : (err))
310 struct snd_usb_audio *chip = cval->head.mixer->chip; in get_ctl_value_v1()
321 idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8); in get_ctl_value_v1()
346 struct snd_usb_audio *chip = cval->head.mixer->chip; in get_ctl_value_v2()
368 idx = mixer_ctrl_intf(cval->head.mixer) | (cval->head.id << 8); in get_ctl_value_v2()
411 return (cval->head.mixer->protocol == UAC_VERSION_1) ? in get_ctl_value()
442 if (!cval->head.mixer->ignore_ctl_error) in snd_usb_get_cur_mix_value()
443 usb_audio_dbg(cval->head in snd_usb_get_cur_mix_value()
[all...]
/kernel/linux/linux-5.10/arch/sparc/kernel/
H A Dpci_sun4v.h54 unsigned long *head);
57 unsigned long head);
60 unsigned long *head);
/kernel/linux/linux-5.10/tools/testing/selftests/arm64/signal/testcases/
H A Dtestcases.h68 * head->size accounts both for payload and header _aarch64_ctx size !
83 struct _aarch64_ctx *get_header(struct _aarch64_ctx *head, uint32_t magic,
86 static inline struct _aarch64_ctx *get_terminator(struct _aarch64_ctx *head, in get_terminator() argument
90 return get_header(head, 0, resv_sz, offset); in get_terminator()
/kernel/linux/linux-6.6/arch/sparc/kernel/
H A Dpci_sun4v.h54 unsigned long *head);
57 unsigned long head);
60 unsigned long *head);
/kernel/linux/linux-5.10/drivers/net/wireless/broadcom/brcm80211/brcmsmac/
H A Ddma.h113 static inline void dma_spin_for_len(uint len, struct sk_buff *head) in dma_spin_for_len() argument
117 while (!(len = *(u16 *) KSEG1ADDR(head->data))) in dma_spin_for_len()
120 *(u16 *) (head->data) = cpu_to_le16((u16) len); in dma_spin_for_len()
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/pm/
H A Dpriv.h15 struct list_head head; member
38 struct list_head head; member
76 struct list_head head; member
/kernel/linux/linux-5.10/include/drm/
H A Ddrm_encoder.h89 * @head: list management
101 struct list_head head; member
266 list_for_each_entry((encoder), &(dev)->mode_config.encoder_list, head) \
277 list_for_each_entry(encoder, &(dev)->mode_config.encoder_list, head)
/kernel/linux/linux-5.10/include/dfx/
H A Dhungtask_base.h107 bool hashlist_find(struct hlist_head *head, int count, pid_t tgid);
108 void hashlist_clear(struct hlist_head *head, int count);
109 bool hashlist_insert(struct hlist_head *head, int count, pid_t tgid);

Completed in 23 milliseconds

1...<<31323334353637383940>>...200