/third_party/skia/third_party/externals/zlib/contrib/optimizations/ |
H A D | inflate.c | 135 state->head = Z_NULL; 686 if (state->head != Z_NULL) 687 state->head->done = -1; 731 if (state->head != Z_NULL) 732 state->head->text = (int)((hold >> 8) & 1); 739 if (state->head != Z_NULL) 740 state->head->time = hold; 747 if (state->head != Z_NULL) { 748 state->head->xflags = (int)(hold & 0xff); 749 state->head [all...] |
/third_party/skia/third_party/externals/zlib/ |
H A D | inflate.c | 134 state->head = Z_NULL; 675 if (state->head != Z_NULL) 676 state->head->done = -1; 720 if (state->head != Z_NULL) 721 state->head->text = (int)((hold >> 8) & 1); 728 if (state->head != Z_NULL) 729 state->head->time = hold; 736 if (state->head != Z_NULL) { 737 state->head->xflags = (int)(hold & 0xff); 738 state->head [all...] |
/kernel/linux/linux-5.10/net/sched/ |
H A D | sch_fq_pie.c | 41 * @head: first packet in the flow 50 struct sk_buff *head; member 121 if (!flow->head) in flow_queue_add() 122 flow->head = skb; in flow_queue_add() 228 struct sk_buff *skb = flow->head; in dequeue_head() 230 flow->head = skb->next; in dequeue_head() 240 struct list_head *head; in fq_pie_qdisc_dequeue() local 244 head = &q->new_flows; in fq_pie_qdisc_dequeue() 245 if (list_empty(head)) { in fq_pie_qdisc_dequeue() 246 head in fq_pie_qdisc_dequeue() [all...] |
/kernel/linux/linux-5.10/drivers/nvme/host/ |
H A D | nvme.h | 452 struct nvme_ns_head *head; member 717 struct nvme_ns_head **head, int *srcu_idx); 718 void nvme_put_ns_from_disk(struct nvme_ns_head *head, int idx); 736 int nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl,struct nvme_ns_head *head); 738 void nvme_mpath_remove_disk(struct nvme_ns_head *head); 746 struct nvme_ns *nvme_find_path(struct nvme_ns_head *head); 751 struct nvme_ns_head *head = ns->head; in nvme_mpath_check_last_path() local 753 if (head->disk && list_empty(&head in nvme_mpath_check_last_path() 791 nvme_mpath_alloc_disk(struct nvme_ctrl *ctrl, struct nvme_ns_head *head) nvme_mpath_alloc_disk() argument 800 nvme_mpath_remove_disk(struct nvme_ns_head *head) nvme_mpath_remove_disk() argument [all...] |
/kernel/linux/linux-5.10/drivers/s390/block/ |
H A D | dasd_eer.c | 91 int head; member 106 if (eerb->head < eerb->tail) in dasd_eer_get_free_bytes() 107 return eerb->tail - eerb->head - 1; in dasd_eer_get_free_bytes() 108 return eerb->buffersize - eerb->head + eerb->tail -1; in dasd_eer_get_free_bytes() 118 if (eerb->head >= eerb->tail) in dasd_eer_get_filled_bytes() 119 return eerb->head - eerb->tail; in dasd_eer_get_filled_bytes() 120 return eerb->buffersize - eerb->tail + eerb->head; in dasd_eer_get_filled_bytes() 140 headindex = eerb->head / PAGE_SIZE; in dasd_eer_write_buffer() 141 localhead = eerb->head % PAGE_SIZE; in dasd_eer_write_buffer() 146 eerb->head in dasd_eer_write_buffer() [all...] |
/kernel/linux/linux-6.6/drivers/s390/block/ |
H A D | dasd_eer.c | 91 int head; member 106 if (eerb->head < eerb->tail) in dasd_eer_get_free_bytes() 107 return eerb->tail - eerb->head - 1; in dasd_eer_get_free_bytes() 108 return eerb->buffersize - eerb->head + eerb->tail -1; in dasd_eer_get_free_bytes() 118 if (eerb->head >= eerb->tail) in dasd_eer_get_filled_bytes() 119 return eerb->head - eerb->tail; in dasd_eer_get_filled_bytes() 120 return eerb->buffersize - eerb->tail + eerb->head; in dasd_eer_get_filled_bytes() 140 headindex = eerb->head / PAGE_SIZE; in dasd_eer_write_buffer() 141 localhead = eerb->head % PAGE_SIZE; in dasd_eer_write_buffer() 146 eerb->head in dasd_eer_write_buffer() [all...] |
/kernel/linux/linux-6.6/drivers/crypto/caam/ |
H A D | jr.c | 265 int hw_idx, sw_idx, i, head, tail; in caam_jr_dequeue() local 277 head = READ_ONCE(jrp->head); in caam_jr_dequeue() 282 for (i = 0; CIRC_CNT(head, tail + i, JOBR_DEPTH) >= 1; i++) { in caam_jr_dequeue() 290 BUG_ON(CIRC_CNT(head, tail + i, JOBR_DEPTH) <= 0); in caam_jr_dequeue() 330 } while (CIRC_CNT(head, tail, JOBR_DEPTH) >= 1 && in caam_jr_dequeue() 400 * caam_jr_enqueue() - Enqueue a job descriptor head. Returns -EINPROGRESS 433 int head, tail, desc_size; in caam_jr_enqueue() local 445 head = jrp->head; in caam_jr_enqueue() [all...] |
/kernel/linux/linux-6.6/net/sched/ |
H A D | sch_fq_pie.c | 41 * @head: first packet in the flow 50 struct sk_buff *head; member 121 if (!flow->head) in flow_queue_add() 122 flow->head = skb; in flow_queue_add() 228 struct sk_buff *skb = flow->head; in dequeue_head() 230 flow->head = skb->next; in dequeue_head() 240 struct list_head *head; in fq_pie_qdisc_dequeue() local 244 head = &q->new_flows; in fq_pie_qdisc_dequeue() 245 if (list_empty(head)) { in fq_pie_qdisc_dequeue() 246 head in fq_pie_qdisc_dequeue() [all...] |
/kernel/linux/linux-6.6/net/bridge/ |
H A D | br_cfm.c | 11 hlist_for_each_entry(mep, &br->mep_list, head) in br_mep_find() 23 hlist_for_each_entry_rcu(mep, &br->mep_list, head, in br_mep_find_ifindex() 36 hlist_for_each_entry_rcu(peer_mep, &mep->peer_mep_list, head, in br_peer_mep_find() 564 hlist_add_tail_rcu(&mep->head, &br->mep_list); in br_cfm_mep_create() 578 hlist_for_each_entry_safe(peer_mep, n_store, &mep->peer_mep_list, head) { in mep_delete_implementation() 580 hlist_del_rcu(&peer_mep->head); in mep_delete_implementation() 587 hlist_del_rcu(&mep->head); in mep_delete_implementation() 658 hlist_for_each_entry(peer_mep, &mep->peer_mep_list, head) in br_cfm_cc_config_set() 663 hlist_for_each_entry(peer_mep, &mep->peer_mep_list, head) in br_cfm_cc_config_set() 707 hlist_add_tail_rcu(&peer_mep->head, in br_cfm_cc_peer_mep_add() [all...] |
/third_party/mesa3d/src/gallium/auxiliary/pipebuffer/ |
H A D | pb_buffer_fenced.c | 123 struct list_head head; member 210 fenced_buf = list_entry(curr, struct fenced_buffer, head); in fenced_manager_dump_locked() 225 fenced_buf = list_entry(curr, struct fenced_buffer, head); in fenced_manager_dump_locked() 251 assert(fenced_buf->head.prev); in fenced_buffer_destroy_locked() 252 assert(fenced_buf->head.next); in fenced_buffer_destroy_locked() 253 list_del(&fenced_buf->head); in fenced_buffer_destroy_locked() 279 list_del(&fenced_buf->head); in fenced_buffer_add_locked() 282 list_addtail(&fenced_buf->head, &fenced_mgr->fenced); in fenced_buffer_add_locked() 305 assert(fenced_buf->head.prev); in fenced_buffer_remove_locked() 306 assert(fenced_buf->head in fenced_buffer_remove_locked() [all...] |
/kernel/linux/linux-5.10/fs/hmdfs/ |
H A D | hmdfs_client.c | 395 void hmdfs_client_recv_readpage(struct hmdfs_head_cmd *head, int err, in hmdfs_client_recv_readpage() argument 399 int ret = le32_to_cpu(head->ret_code); in hmdfs_client_recv_readpage() 409 hmdfs_client_resp_statis(async_work->head.peer->sbi, F_READPAGE, in hmdfs_client_recv_readpage() 412 trace_hmdfs_client_recv_readpage(async_work->head.peer, in hmdfs_client_recv_readpage() 1053 struct hmdfs_msg_idr_head *head = NULL; in hmdfs_get_msg_next() local 1056 head = idr_get_next(&peer->msg_idr, id); in hmdfs_get_msg_next() 1057 if (head && head->type < MSG_IDR_MAX && head->type >= 0) in hmdfs_get_msg_next() 1058 kref_get(&head in hmdfs_get_msg_next() 1070 struct hmdfs_msg_idr_head *head = NULL; hmdfs_client_offline_notify() local [all...] |
/kernel/linux/linux-5.10/mm/ |
H A D | memory-failure.c | 857 #define head (1UL << PG_head) macro 880 { head, head, MF_MSG_HUGE, me_huge_page }, 905 #undef head macro 957 struct page *head = compound_head(page); in get_hwpoison_page() local 959 if (!PageHuge(head) && PageTransHuge(head)) { in get_hwpoison_page() 966 if (!PageAnon(head)) { in get_hwpoison_page() 973 if (get_page_unless_zero(head)) { in get_hwpoison_page() 974 if (head in get_hwpoison_page() 1153 struct page *head = compound_head(p); memory_failure_hugetlb() local [all...] |
/kernel/linux/linux-6.6/fs/hmdfs/ |
H A D | hmdfs_client.c | 394 void hmdfs_client_recv_readpage(struct hmdfs_head_cmd *head, int err, in hmdfs_client_recv_readpage() argument 398 int ret = le32_to_cpu(head->ret_code); in hmdfs_client_recv_readpage() 408 hmdfs_client_resp_statis(async_work->head.peer->sbi, F_READPAGE, in hmdfs_client_recv_readpage() 411 trace_hmdfs_client_recv_readpage(async_work->head.peer, in hmdfs_client_recv_readpage() 1052 struct hmdfs_msg_idr_head *head = NULL; in hmdfs_get_msg_next() local 1055 head = idr_get_next(&peer->msg_idr, id); in hmdfs_get_msg_next() 1056 if (head && head->type < MSG_IDR_MAX && head->type >= 0) in hmdfs_get_msg_next() 1057 kref_get(&head in hmdfs_get_msg_next() 1069 struct hmdfs_msg_idr_head *head = NULL; hmdfs_client_offline_notify() local [all...] |
/third_party/skia/third_party/externals/harfbuzz/src/ |
H A D | test-repacker.cc | 794 assert(strncmp (graph.object (3).head, "abc", 3) == 0); in test_sort_kahn_1() 799 assert(strncmp (graph.object (2).head, "def", 3) == 0); in test_sort_kahn_1() 803 assert(strncmp (graph.object (1).head, "jkl", 3) == 0); in test_sort_kahn_1() 806 assert(strncmp (graph.object (0).head, "ghi", 3) == 0); in test_sort_kahn_1() 823 assert(strncmp (graph.object (4).head, "abc", 3) == 0); in test_sort_kahn_2() 829 assert(strncmp (graph.object (3).head, "def", 3) == 0); in test_sort_kahn_2() 833 assert(strncmp (graph.object (2).head, "mn", 2) == 0); in test_sort_kahn_2() 836 assert(strncmp (graph.object (1).head, "ghi", 3) == 0); in test_sort_kahn_2() 840 assert(strncmp (graph.object (0).head, "jkl", 3) == 0); in test_sort_kahn_2() 856 assert(strncmp (graph.object (4).head, "ab in test_sort_shortest() [all...] |
/kernel/linux/linux-5.10/drivers/acpi/ |
H A D | acpi_dbg.c | 28 (CIRC_CNT((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 30 (CIRC_CNT_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 32 (CIRC_SPACE((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 34 (CIRC_SPACE_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 269 p = &crc->buf[crc->head]; in acpi_aml_write_kern() 272 /* sync head after inserting logs */ in acpi_aml_write_kern() 274 crc->head = (crc->head + n) & (ACPI_AML_BUF_SIZE - 1); in acpi_aml_write_kern() 288 /* sync head before removing cmds */ in acpi_aml_readb_kern() 520 acpi_aml_io.out_crc.head in acpi_aml_open() [all...] |
/kernel/linux/linux-5.10/drivers/net/ |
H A D | eql.c | 145 struct list_head *this, *tmp, *head; in eql_timer() local 148 head = &eql->queue.all_slaves; in eql_timer() 149 list_for_each_safe(this, tmp, head) { in eql_timer() 232 struct list_head *head, *tmp, *this; in eql_kill_slave_queue() local 236 head = &queue->all_slaves; in eql_kill_slave_queue() 237 list_for_each_safe(this, tmp, head) { in eql_kill_slave_queue() 299 struct list_head *this, *tmp, *head; in __eql_schedule_slaves() local 305 head = &queue->all_slaves; in __eql_schedule_slaves() 306 list_for_each_safe(this, tmp, head) { in __eql_schedule_slaves() 364 struct list_head *this, *head; in __eql_find_slave_dev() local [all...] |
/kernel/linux/linux-5.10/security/selinux/ss/ |
H A D | mls.c | 38 int i, l, len, head, prev; in mls_compute_context_len() local 52 head = -2; in mls_compute_context_len() 58 if (head != prev) { in mls_compute_context_len() 64 head = i; in mls_compute_context_len() 68 if (prev != head) { in mls_compute_context_len() 94 int i, l, head, prev; in mls_sid_to_context() local 112 head = -2; in mls_sid_to_context() 118 if (prev != head) { in mls_sid_to_context() 119 if (prev - head > 1) in mls_sid_to_context() 134 head in mls_sid_to_context() [all...] |
/kernel/linux/linux-5.10/tools/perf/scripts/python/ |
H A D | compaction-times.py | 121 head = cls.heads[pid] 122 filtered = head.is_filtered() 126 head = cls.heads[pid] = chead(comm, pid, filtered) 129 head.mark_pending(start_secs, start_nsecs) 133 head = cls.heads[pid] 134 if not head.is_filtered(): 135 if head.is_pending(): 136 head.do_increment(migrated, fscan, mscan) 142 head = cls.heads[pid] 143 if not head [all...] |
/kernel/linux/linux-5.10/net/ipv6/ |
H A D | ip6_offload.c | 31 #define indirect_call_gro_receive_l4(f2, f1, cb, head, skb) \ 35 INDIRECT_CALL_L4(cb, f2, f1, head, skb); \ 117 skb->network_header = skb_mac_header(skb) + nhoff - skb->head; in ipv6_gso_segment() 130 skb->head - (unsigned char *)(ipv6h + 1); in ipv6_gso_segment() 134 skb->network_header = (u8 *)ipv6h - skb->head; in ipv6_gso_segment() 184 INDIRECT_CALLABLE_SCOPE struct sk_buff *ipv6_gro_receive(struct list_head *head, in ipv6_gro_receive() argument 235 list_for_each_entry(p, head, list) { in ipv6_gro_receive() 281 ops->callbacks.gro_receive, head, skb); in ipv6_gro_receive() 292 static struct sk_buff *sit_ip6ip6_gro_receive(struct list_head *head, in sit_ip6ip6_gro_receive() argument 304 return ipv6_gro_receive(head, sk in sit_ip6ip6_gro_receive() 307 ip4ip6_gro_receive(struct list_head *head, struct sk_buff *skb) ip4ip6_gro_receive() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/acr/ |
H A D | base.c | 32 list_for_each_entry(hsf, &acr->hsf, head) { in nvkm_acr_hsf_find() 103 list_for_each_entry(lsf, &acr->lsf, head) { in nvkm_acr_load() 128 list_for_each_entry(lsf, &acr->lsf, head) { in nvkm_acr_falcon() 235 list_for_each_entry_safe(lsfw, lsft, &acr->lsfw, head) { in nvkm_acr_oneinit() 258 list_add_tail(&lsf->head, &acr->lsf); in nvkm_acr_oneinit() 266 list_move(&lsf->head, &acr->lsf); in nvkm_acr_oneinit() 274 list_for_each_entry_safe(lsfw, lsft, &acr->lsfw, head) { in nvkm_acr_oneinit() 338 list_for_each_entry(hsfw, &acr->hsfw, head) { in nvkm_acr_oneinit() 357 list_for_each_entry_safe(hsf, hst, &acr->hsf, head) { in nvkm_acr_dtor() 361 list_del(&hsf->head); in nvkm_acr_dtor() [all...] |
/kernel/linux/linux-6.6/drivers/net/ |
H A D | eql.c | 147 struct list_head *this, *tmp, *head; in eql_timer() local 150 head = &eql->queue.all_slaves; in eql_timer() 151 list_for_each_safe(this, tmp, head) { in eql_timer() 234 struct list_head *head, *tmp, *this; in eql_kill_slave_queue() local 238 head = &queue->all_slaves; in eql_kill_slave_queue() 239 list_for_each_safe(this, tmp, head) { in eql_kill_slave_queue() 305 struct list_head *this, *tmp, *head; in __eql_schedule_slaves() local 311 head = &queue->all_slaves; in __eql_schedule_slaves() 312 list_for_each_safe(this, tmp, head) { in __eql_schedule_slaves() 370 struct list_head *this, *head; in __eql_find_slave_dev() local [all...] |
/kernel/linux/linux-6.6/drivers/acpi/ |
H A D | acpi_dbg.c | 28 (CIRC_CNT((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 30 (CIRC_CNT_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 32 (CIRC_SPACE((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 34 (CIRC_SPACE_TO_END((circ)->head, (circ)->tail, ACPI_AML_BUF_SIZE)) 262 p = &crc->buf[crc->head]; in acpi_aml_write_kern() 265 /* sync head after inserting logs */ in acpi_aml_write_kern() 267 crc->head = (crc->head + n) & (ACPI_AML_BUF_SIZE - 1); in acpi_aml_write_kern() 281 /* sync head before removing cmds */ in acpi_aml_readb_kern() 513 acpi_aml_io.out_crc.head in acpi_aml_open() [all...] |
/kernel/linux/linux-6.6/tools/perf/scripts/python/ |
H A D | compaction-times.py | 121 head = cls.heads[pid] 122 filtered = head.is_filtered() 126 head = cls.heads[pid] = chead(comm, pid, filtered) 129 head.mark_pending(start_secs, start_nsecs) 133 head = cls.heads[pid] 134 if not head.is_filtered(): 135 if head.is_pending(): 136 head.do_increment(migrated, fscan, mscan) 142 head = cls.heads[pid] 143 if not head [all...] |
/kernel/linux/linux-6.6/tools/perf/util/ |
H A D | auxtrace.h | 270 * @head: head of buffer list 277 struct list_head head; member 404 u64 *head, u64 *old); 450 * @head: list of address filters 454 struct list_head head; member 469 u64 head; in auxtrace_mmap__read_head() local 475 head = READ_ONCE(pc->aux_head); in auxtrace_mmap__read_head() 477 /* Ensure all reads are done after we read the head */ in auxtrace_mmap__read_head() 479 return head; in auxtrace_mmap__read_head() [all...] |
/kernel/linux/linux-6.6/security/selinux/ss/ |
H A D | mls.c | 38 int i, l, len, head, prev; in mls_compute_context_len() local 52 head = -2; in mls_compute_context_len() 58 if (head != prev) { in mls_compute_context_len() 64 head = i; in mls_compute_context_len() 68 if (prev != head) { in mls_compute_context_len() 94 int i, l, head, prev; in mls_sid_to_context() local 112 head = -2; in mls_sid_to_context() 118 if (prev != head) { in mls_sid_to_context() 119 if (prev - head > 1) in mls_sid_to_context() 134 head in mls_sid_to_context() [all...] |