/kernel/linux/linux-6.6/tools/bpf/bpftool/ |
H A D | cfg.c | 30 struct bpf_insn *head; member 94 if (bb->head == insn) in func_append_bb() 96 else if (bb->head > insn) in func_append_bb() 106 new_bb->head = insn; in func_append_bb() 208 bb->tail = bb_next(bb)->head - 1; in func_partition_bb_tail() 249 if (bb->head == insn) in func_search_bb_with_head() 404 start_idx = bb->head - func->start; in draw_bb_node() 405 dump_xlated_for_graph(dd, bb->head, bb->tail, start_idx, in draw_bb_node()
|
/kernel/linux/linux-6.6/security/tomoyo/ |
H A D | audit.c | 436 * @head: Pointer to "struct tomoyo_io_buffer". 440 void tomoyo_read_log(struct tomoyo_io_buffer *head) in tomoyo_read_log() argument 444 if (head->r.w_pos) in tomoyo_read_log() 446 kfree(head->read_buf); in tomoyo_read_log() 447 head->read_buf = NULL; in tomoyo_read_log() 457 head->read_buf = ptr->log; in tomoyo_read_log() 458 head->r.w[head->r.w_pos++] = head->read_buf; in tomoyo_read_log()
|
/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_feedback.c | 18 struct list_head head; member 152 list_add(&feedback_buf->head, &pool->feedback_buffers); in vn_feedback_pool_grow_locked() 180 head) in vn_feedback_pool_fini() 184 &pool->feedback_buffers, head) in vn_feedback_pool_fini() 210 head); in vn_feedback_pool_alloc_locked() 228 list_first_entry(&pool->free_slots, struct vn_feedback_slot, head); in vn_feedback_pool_alloc() 229 list_del(&slot->head); in vn_feedback_pool_alloc() 264 list_add(&slot->head, &pool->free_slots); in vn_feedback_pool_free()
|
/device/soc/hisilicon/hi3861v100/sdk_liteos/boot/flashboot/upg/ |
H A D | boot_upg_check.h | 24 hi_u32 boot_upg_get_common_head(hi_u32 addr, hi_upg_common_head *head); 26 hi_u32 boot_upg_check_common_head(const hi_upg_common_head *head);
|
/device/soc/rockchip/rk3588/kernel/include/trace/hooks/ |
H A D | futex.h | 17 struct plist_head *head, 19 TP_ARGS(node, head, already_on_hb));
|
/kernel/linux/linux-5.10/arch/powerpc/include/asm/ |
H A D | hvcserver.h | 38 extern int hvcs_free_partner_info(struct list_head *head); 40 struct list_head *head, unsigned long *pi_buff);
|
/kernel/linux/linux-6.6/arch/powerpc/include/asm/ |
H A D | hvcserver.h | 38 extern int hvcs_free_partner_info(struct list_head *head); 40 struct list_head *head, unsigned long *pi_buff);
|
/kernel/linux/linux-5.10/tools/include/linux/ |
H A D | ring_buffer.h | 61 u64 head = READ_ONCE(base->data_head); in ring_buffer_read_head() 64 return head; in ring_buffer_read_head()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
H A D | runq.c | 28 list_del(&runq->head); in nvkm_runq_del() 43 list_add_tail(&runq->head, &fifo->runqs); in nvkm_runq_new()
|
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/include/nvif/ |
H A D | head.h | 16 nvif_head_id(struct nvif_head *head) in nvif_head_id() argument 18 return head->object.handle; in nvif_head_id()
|
/kernel/linux/linux-6.6/tools/include/linux/ |
H A D | ring_buffer.h | 61 u64 head = READ_ONCE(base->data_head); in ring_buffer_read_head() 64 return head; in ring_buffer_read_head()
|
/third_party/libcoap/examples/ |
H A D | coap_list.c | 36 coap_insert(coap_list_t **head, coap_list_t *node) { in coap_insert() argument 42 LL_APPEND((*head), node); in coap_insert()
|
/kernel/linux/linux-5.10/kernel/trace/ |
H A D | ring_buffer.c | 492 * head_page == tail_page && head == tail then buffer is empty. 506 struct buffer_page *head_page; /* read from head */ 563 unsigned long head; member 1114 * the head (writes go to the tail). But if a writer is in overwrite 1115 * mode and wraps, it must push the head page forward. 1119 * The reader must be careful to replace only the head page, and 1122 * page after head. It then sets the page after head to point to 1123 * the old reader page. But if the writer moves the head page 1127 * special with the page before head 1248 struct buffer_page *head; rb_head_page_activate() local 1282 rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag, int new_flag) rb_head_page_set() argument 1305 rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_update() argument 1314 rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_head() argument 1323 rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_normal() argument 1343 struct buffer_page *head; rb_set_head_page() local 1482 struct list_head *head = rb_list_head(cpu_buffer->pages); rb_check_pages() local 1665 struct list_head *head = cpu_buffer->pages; rb_free_cpu_buffer() local 3743 struct buffer_page *head = rb_set_head_page(cpu_buffer); rb_per_cpu_empty() local [all...] |
/kernel/linux/linux-6.6/kernel/trace/ |
H A D | ring_buffer.c | 492 * head_page == tail_page && head == tail then buffer is empty. 506 struct buffer_page *head_page; /* read from head */ 564 unsigned long head; member 1197 * the head (writes go to the tail). But if a writer is in overwrite 1198 * mode and wraps, it must push the head page forward. 1202 * The reader must be careful to replace only the head page, and 1205 * page after head. It then sets the page after head to point to 1206 * the old reader page. But if the writer moves the head page 1210 * special with the page before head 1329 struct buffer_page *head; rb_head_page_activate() local 1363 rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag, int new_flag) rb_head_page_set() argument 1386 rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_update() argument 1395 rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_head() argument 1404 rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, struct buffer_page *head, struct buffer_page *prev, int old_flag) rb_head_page_set_normal() argument 1423 struct buffer_page *head; rb_set_head_page() local 1551 struct list_head *head = rb_list_head(cpu_buffer->pages); rb_check_pages() local 1735 struct list_head *head = cpu_buffer->pages; rb_free_cpu_buffer() local 3939 struct buffer_page *head = rb_set_head_page(cpu_buffer); rb_per_cpu_empty() local [all...] |
/kernel/linux/linux-5.10/drivers/vhost/ |
H A D | net.c | 104 int head; member 153 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr() 154 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr() 161 return rxq->tail - rxq->head; in vhost_net_buf_get_size() 166 return rxq->tail == rxq->head; in vhost_net_buf_is_empty() 172 ++rxq->head; in vhost_net_buf_consume() 180 rxq->head = 0; in vhost_net_buf_produce() 191 ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head, in vhost_net_buf_unproduce() 194 rxq->head = rxq->tail = 0; in vhost_net_buf_unproduce() 225 rxq->head in vhost_net_buf_init() 773 int head; handle_tx_copy() local 862 int head; handle_tx_zerocopy() local 987 struct sk_buff *head; peek_head_len() local [all...] |
/kernel/linux/linux-5.10/fs/xfs/ |
H A D | xfs_ioctl.c | 1726 struct fsmap_head head; in xfs_ioc_getfsmap() local 1733 if (copy_from_user(&head, arg, sizeof(struct fsmap_head))) in xfs_ioc_getfsmap() 1735 if (memchr_inv(head.fmh_reserved, 0, sizeof(head.fmh_reserved)) || in xfs_ioc_getfsmap() 1736 memchr_inv(head.fmh_keys[0].fmr_reserved, 0, in xfs_ioc_getfsmap() 1737 sizeof(head.fmh_keys[0].fmr_reserved)) || in xfs_ioc_getfsmap() 1738 memchr_inv(head.fmh_keys[1].fmr_reserved, 0, in xfs_ioc_getfsmap() 1739 sizeof(head.fmh_keys[1].fmr_reserved))) in xfs_ioc_getfsmap() 1747 count = min_t(unsigned int, head.fmh_count, in xfs_ioc_getfsmap() 1751 count = min_t(unsigned int, head in xfs_ioc_getfsmap() [all...] |
/kernel/linux/linux-6.6/fs/xfs/ |
H A D | xfs_ioctl.c | 1518 struct fsmap_head head; in xfs_ioc_getfsmap() local 1525 if (copy_from_user(&head, arg, sizeof(struct fsmap_head))) in xfs_ioc_getfsmap() 1527 if (memchr_inv(head.fmh_reserved, 0, sizeof(head.fmh_reserved)) || in xfs_ioc_getfsmap() 1528 memchr_inv(head.fmh_keys[0].fmr_reserved, 0, in xfs_ioc_getfsmap() 1529 sizeof(head.fmh_keys[0].fmr_reserved)) || in xfs_ioc_getfsmap() 1530 memchr_inv(head.fmh_keys[1].fmr_reserved, 0, in xfs_ioc_getfsmap() 1531 sizeof(head.fmh_keys[1].fmr_reserved))) in xfs_ioc_getfsmap() 1539 count = min_t(unsigned int, head.fmh_count, in xfs_ioc_getfsmap() 1543 count = min_t(unsigned int, head in xfs_ioc_getfsmap() [all...] |
/kernel/linux/linux-6.6/drivers/vhost/ |
H A D | net.c | 105 int head; member 154 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr() 155 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr() 162 return rxq->tail - rxq->head; in vhost_net_buf_get_size() 167 return rxq->tail == rxq->head; in vhost_net_buf_is_empty() 173 ++rxq->head; in vhost_net_buf_consume() 181 rxq->head = 0; in vhost_net_buf_produce() 192 ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head, in vhost_net_buf_unproduce() 195 rxq->head = rxq->tail = 0; in vhost_net_buf_unproduce() 226 rxq->head in vhost_net_buf_init() 774 int head; handle_tx_copy() local 864 int head; handle_tx_zerocopy() local 996 struct sk_buff *head; peek_head_len() local [all...] |
/third_party/skia/src/utils/ |
H A D | SkPolyUtils.cpp | 297 static void remove_node(const OffsetEdge* node, OffsetEdge** head) { in remove_node() argument 301 if (node == *head) { in remove_node() 302 *head = (node->fNext == node) ? nullptr : node->fNext; in remove_node() 374 OffsetEdge* head = &edgeData[0]; in SkInsetConvexPolygon() local 375 OffsetEdge* currEdge = head; in SkInsetConvexPolygon() 380 while (head && prevEdge != currEdge) { in SkInsetConvexPolygon() 394 remove_node(prevEdge, &head); in SkInsetConvexPolygon() 423 remove_node(prevEdge, &head); in SkInsetConvexPolygon() 429 remove_node(currEdge, &head); in SkInsetConvexPolygon() 439 if (!head) { in SkInsetConvexPolygon() 1314 auto head = &edgeData[0]; SkOffsetSimplePolygon() local [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/sfc/ |
H A D | rx_common.c | 33 /* RX maximum head room required. 577 struct list_head *head = &efx->rss_context.list; in efx_alloc_rss_context_entry() local 584 list_for_each_entry(ctx, head, list) { in efx_alloc_rss_context_entry() 610 struct list_head *head = &efx->rss_context.list; in efx_find_rss_context_entry() local 615 list_for_each_entry(ctx, head, list) in efx_find_rss_context_entry() 736 struct hlist_head *head; in efx_rps_hash_find() local 739 head = efx_rps_hash_bucket(efx, spec); in efx_rps_hash_find() 740 if (!head) in efx_rps_hash_find() 742 hlist_for_each(node, head) { in efx_rps_hash_find() 755 struct hlist_head *head; in efx_rps_hash_add() local 780 struct hlist_head *head; efx_rps_hash_del() local [all...] |
/kernel/linux/linux-5.10/drivers/tty/ |
H A D | amiserial.c | 178 if (info->xmit.head != info->xmit.tail in rs_start() 311 if (info->xmit.head == info->xmit.tail in transmit_chars() 325 if (CIRC_CNT(info->xmit.head, in transmit_chars() 333 if (info->xmit.head == info->xmit.tail) { in transmit_chars() 533 info->xmit.head = info->xmit.tail = 0; in startup() 752 if (CIRC_SPACE(info->xmit.head, in rs_put_char() 759 info->xmit.buf[info->xmit.head++] = ch; in rs_put_char() 760 info->xmit.head &= SERIAL_XMIT_SIZE-1; in rs_put_char() 770 if (info->xmit.head == info->xmit.tail in rs_flush_chars() 797 c = CIRC_SPACE_TO_END(info->xmit.head, in rs_write() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/sfc/ |
H A D | rx_common.c | 26 /* RX maximum head room required. 564 struct list_head *head = &efx->rss_context.list; in efx_alloc_rss_context_entry() local 571 list_for_each_entry(ctx, head, list) { in efx_alloc_rss_context_entry() 597 struct list_head *head = &efx->rss_context.list; in efx_find_rss_context_entry() local 602 list_for_each_entry(ctx, head, list) in efx_find_rss_context_entry() 723 struct hlist_head *head; in efx_rps_hash_find() local 726 head = efx_rps_hash_bucket(efx, spec); in efx_rps_hash_find() 727 if (!head) in efx_rps_hash_find() 729 hlist_for_each(node, head) { in efx_rps_hash_find() 742 struct hlist_head *head; in efx_rps_hash_add() local 767 struct hlist_head *head; efx_rps_hash_del() local [all...] |
/kernel/linux/linux-6.6/drivers/xen/ |
H A D | grant-table.c | 183 grant_ref_t head; in get_free_entries() local 193 ref = head = gnttab_free_head; in get_free_entries() 196 bitmap_clear(gnttab_free_bitmap, head, 1); in get_free_entries() 197 if (gnttab_free_tail_ptr == __gnttab_entry(head)) in get_free_entries() 200 head = gnttab_entry(head); in get_free_entries() 202 gnttab_free_head = gnttab_entry(head); in get_free_entries() 203 gnttab_entry(head) = GNTTAB_LIST_END; in get_free_entries() 621 void gnttab_free_grant_references(grant_ref_t head) in gnttab_free_grant_references() argument 627 while (head ! in gnttab_free_grant_references() 637 gnttab_free_grant_reference_seq(grant_ref_t head, unsigned int count) gnttab_free_grant_reference_seq() argument 650 gnttab_alloc_grant_references(u16 count, grant_ref_t *head) gnttab_alloc_grant_references() argument [all...] |
/kernel/linux/linux-6.6/drivers/tty/ |
H A D | amiserial.c | 156 if (info->xmit.head != info->xmit.tail in rs_start() 276 if (info->xmit.head == info->xmit.tail in transmit_chars() 290 if (CIRC_CNT(info->xmit.head, in transmit_chars() 298 if (info->xmit.head == info->xmit.tail) { in transmit_chars() 498 info->xmit.head = info->xmit.tail = 0; in startup() 710 if (CIRC_SPACE(info->xmit.head, in rs_put_char() 717 info->xmit.buf[info->xmit.head++] = ch; in rs_put_char() 718 info->xmit.head &= UART_XMIT_SIZE - 1; in rs_put_char() 728 if (info->xmit.head == info->xmit.tail in rs_flush_chars() 755 c = CIRC_SPACE_TO_END(info->xmit.head, in rs_write() [all...] |
/kernel/linux/linux-6.6/kernel/bpf/ |
H A D | devmap.c | 208 struct hlist_head *head; in dev_map_free() local 211 head = dev_map_index_hash(dtab, i); in dev_map_free() 213 hlist_for_each_entry_safe(dev, next, head, index_hlist) { in dev_map_free() 267 struct hlist_head *head = dev_map_index_hash(dtab, key); in __dev_map_hash_lookup_elem() local 270 hlist_for_each_entry_rcu(dev, head, index_hlist, in __dev_map_hash_lookup_elem() 284 struct hlist_head *head; in dev_map_hash_get_next_key() local 309 head = dev_map_index_hash(dtab, i); in dev_map_hash_get_next_key() 311 next_dev = hlist_entry_safe(rcu_dereference_raw(hlist_first_rcu(head)), in dev_map_hash_get_next_key() 598 struct hlist_head *head; in dev_map_enqueue_multi() local 632 head in dev_map_enqueue_multi() 715 struct hlist_head *head; dev_map_redirect_multi() local 1071 struct hlist_head *head; dev_map_hash_remove_netdev() local [all...] |