/kernel/linux/linux-6.6/arch/sparc/boot/ |
H A D | piggyback.c | 184 int image, tail; in main() local 255 if ((tail = open(argv[4], O_RDONLY)) < 0) in main() 257 while ((i = read(tail, buffer, 1024)) > 0) in main() 262 if (close(tail) < 0) in main()
|
/kernel/linux/linux-5.10/security/selinux/ |
H A D | netnode.c | 166 struct sel_netnode *tail; in sel_netnode_insert() local 167 tail = list_entry( in sel_netnode_insert() 171 list_del_rcu(&tail->list); in sel_netnode_insert() 172 kfree_rcu(tail, rcu); in sel_netnode_insert()
|
H A D | netport.c | 114 struct sel_netport *tail; in sel_netport_insert() local 115 tail = list_entry( in sel_netport_insert() 120 list_del_rcu(&tail->list); in sel_netport_insert() 121 kfree_rcu(tail, rcu); in sel_netport_insert()
|
/kernel/linux/linux-6.6/security/selinux/ |
H A D | netnode.c | 165 struct sel_netnode *tail; in sel_netnode_insert() local 166 tail = list_entry( in sel_netnode_insert() 171 list_del_rcu(&tail->list); in sel_netnode_insert() 172 kfree_rcu(tail, rcu); in sel_netnode_insert()
|
H A D | netport.c | 113 struct sel_netport *tail; in sel_netport_insert() local 114 tail = list_entry( in sel_netport_insert() 119 list_del_rcu(&tail->list); in sel_netport_insert() 120 kfree_rcu(tail, rcu); in sel_netport_insert()
|
/kernel/linux/linux-5.10/tools/include/linux/ |
H A D | ring_buffer.h | 32 * READ from the tail WRITE. 69 u64 tail) in ring_buffer_write_tail() 71 smp_store_release(&base->data_tail, tail); in ring_buffer_write_tail() 68 ring_buffer_write_tail(struct perf_event_mmap_page *base, u64 tail) ring_buffer_write_tail() argument
|
/kernel/linux/linux-6.6/tools/include/linux/ |
H A D | ring_buffer.h | 32 * READ from the tail WRITE. 69 u64 tail) in ring_buffer_write_tail() 71 smp_store_release(&base->data_tail, tail); in ring_buffer_write_tail() 68 ring_buffer_write_tail(struct perf_event_mmap_page *base, u64 tail) ring_buffer_write_tail() argument
|
/kernel/linux/linux-5.10/drivers/input/ |
H A D | evdev.c | 42 unsigned int tail; member 44 spinlock_t buffer_lock; /* protects access to buffer, head and tail */ 110 head = client->tail; in __evdev_flush_queue() 111 client->packet_head = client->tail; in __evdev_flush_queue() 116 for (i = client->tail; i != client->head; i = (i + 1) & mask) { in __evdev_flush_queue() 158 if (unlikely(client->head == client->tail)) { in __evdev_queue_syn_dropped() 160 client->tail = (client->head - 1) & (client->bufsize - 1); in __evdev_queue_syn_dropped() 161 client->packet_head = client->tail; in __evdev_queue_syn_dropped() 203 if (client->head != client->tail) { in evdev_set_clk_type() 204 client->packet_head = client->head = client->tail; in evdev_set_clk_type() [all...] |
/kernel/linux/linux-6.6/drivers/input/ |
H A D | evdev.c | 42 unsigned int tail; member 44 spinlock_t buffer_lock; /* protects access to buffer, head and tail */ 110 head = client->tail; in __evdev_flush_queue() 111 client->packet_head = client->tail; in __evdev_flush_queue() 116 for (i = client->tail; i != client->head; i = (i + 1) & mask) { in __evdev_flush_queue() 158 if (unlikely(client->head == client->tail)) { in __evdev_queue_syn_dropped() 160 client->tail = (client->head - 1) & (client->bufsize - 1); in __evdev_queue_syn_dropped() 161 client->packet_head = client->tail; in __evdev_queue_syn_dropped() 203 if (client->head != client->tail) { in evdev_set_clk_type() 204 client->packet_head = client->head = client->tail; in evdev_set_clk_type() [all...] |
/kernel/linux/linux-5.10/drivers/block/zram/zram_group/ |
H A D | zlist.c | 141 struct zlist_node *tail = idx2node(tid, tab); in zlist_add_tail_nolock() local 143 zlist_before_add_check(tab, tail, node, head); in zlist_add_tail_nolock() 152 zlist_node_lock(tail); in zlist_add_tail_nolock() 153 tail->next = idx; in zlist_add_tail_nolock() 155 zlist_node_unlock(tail); in zlist_add_tail_nolock() 156 zlist_after_add_check(tab, tail, node, head); in zlist_add_tail_nolock()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/display/ |
H A D | intel_dsb.c | 210 u32 tail; in intel_dsb_commit() local 226 tail = ALIGN(dsb->free_pos * 4, CACHELINE_BYTES); in intel_dsb_commit() 227 if (tail > dsb->free_pos * 4) in intel_dsb_commit() 229 (tail - dsb->free_pos * 4)); in intel_dsb_commit() 237 "DSB execution started - head 0x%x, tail 0x%x\n", in intel_dsb_commit() 238 i915_ggtt_offset(dsb->vma), tail); in intel_dsb_commit() 240 i915_ggtt_offset(dsb->vma) + tail); in intel_dsb_commit()
|
/kernel/linux/linux-6.6/fs/ntfs3/lib/ |
H A D | lzx_decompress.c | 184 u8 *tail; in lzx_postprocess() local 191 tail = &data[size - 6]; in lzx_postprocess() 192 memcpy(saved_bytes, tail, 6); in lzx_postprocess() 193 memset(tail, 0xE8, 6); in lzx_postprocess() 198 if (p >= tail) in lzx_postprocess() 203 memcpy(tail, saved_bytes, 6); in lzx_postprocess()
|
/kernel/linux/linux-6.6/drivers/block/zram/zram_group/ |
H A D | zlist.c | 141 struct zlist_node *tail = idx2node(tid, tab); in zlist_add_tail_nolock() local 143 zlist_before_add_check(tab, tail, node, head); in zlist_add_tail_nolock() 152 zlist_node_lock(tail); in zlist_add_tail_nolock() 153 tail->next = idx; in zlist_add_tail_nolock() 155 zlist_node_unlock(tail); in zlist_add_tail_nolock() 156 zlist_after_add_check(tab, tail, node, head); in zlist_add_tail_nolock()
|
/kernel/linux/linux-6.6/drivers/dma/ptdma/ |
H A D | ptdma-dev.c | 73 u32 tail; in pt_core_execute_cmd() local 89 /* Write the new tail address back to the queue register */ in pt_core_execute_cmd() 90 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_core_execute_cmd() 91 iowrite32(tail, cmd_q->reg_control + 0x0004); in pt_core_execute_cmd() 129 u32 tail; in pt_do_cmd_complete() local 136 tail = lower_32_bits(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in pt_do_cmd_complete() 138 iowrite32(tail, cmd_q->reg_control + 0x0008); in pt_do_cmd_complete()
|
/kernel/linux/linux-5.10/drivers/firewire/ |
H A D | nosy.c | 63 struct packet *head, *tail; member 121 buffer->tail = (struct packet *) buffer->data; in packet_buffer_init() 194 buffer->tail->length = length; in packet_buffer_put() 196 if (&buffer->tail->data[length] < end) { in packet_buffer_put() 197 memcpy(buffer->tail->data, data, length); in packet_buffer_put() 198 buffer->tail = (struct packet *) &buffer->tail->data[length]; in packet_buffer_put() 200 size_t split = end - buffer->tail->data; in packet_buffer_put() 202 memcpy(buffer->tail->data, data, split); in packet_buffer_put() 204 buffer->tail in packet_buffer_put() [all...] |
/kernel/linux/linux-5.10/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | tdls.c | 398 struct iwl_tdls_channel_switch_cmd_tail *tail = in iwl_mvm_tdls_config_channel_switch() local 416 tail->timing.frame_timestamp = cpu_to_le32(timestamp); in iwl_mvm_tdls_config_channel_switch() 417 tail->timing.switch_time = cpu_to_le32(switch_time); in iwl_mvm_tdls_config_channel_switch() 418 tail->timing.switch_timeout = cpu_to_le32(switch_timeout); in iwl_mvm_tdls_config_channel_switch() 454 tail->timing.max_offchan_duration = in iwl_mvm_tdls_config_channel_switch() 459 tail->frame.switch_time_offset = cpu_to_le32(ch_sw_tm_ie + 2); in iwl_mvm_tdls_config_channel_switch() 469 iwl_mvm_set_tx_cmd_ccmp(info, &tail->frame.tx_cmd); in iwl_mvm_tdls_config_channel_switch() 472 iwl_mvm_set_tx_cmd(mvm, skb, &tail->frame.tx_cmd, info, in iwl_mvm_tdls_config_channel_switch() 475 iwl_mvm_set_tx_cmd_rate(mvm, &tail->frame.tx_cmd, info, sta, in iwl_mvm_tdls_config_channel_switch() 479 memcpy(tail in iwl_mvm_tdls_config_channel_switch() [all...] |
/kernel/linux/linux-6.6/drivers/accel/ivpu/ |
H A D | ivpu_job.c | 58 jobq_header->tail = 0; in ivpu_cmdq_alloc() 147 cmdq->jobq->header.tail = 0; in ivpu_cmdq_reset_locked() 185 u32 tail = READ_ONCE(header->tail); in ivpu_cmdq_push_job() local 186 u32 next_entry = (tail + 1) % cmdq->entry_count; in ivpu_cmdq_push_job() 190 ivpu_dbg(vdev, JOB, "Job queue full: ctx %d engine %d db %d head %d tail %d\n", in ivpu_cmdq_push_job() 191 job->file_priv->ctx.id, job->engine_idx, cmdq->db_id, header->head, tail); in ivpu_cmdq_push_job() 195 entry = &cmdq->jobq->job[tail]; in ivpu_cmdq_push_job() 199 wmb(); /* Ensure that tail is updated after filling entry */ in ivpu_cmdq_push_job() 200 header->tail in ivpu_cmdq_push_job() [all...] |
/kernel/linux/linux-6.6/drivers/firewire/ |
H A D | nosy.c | 63 struct packet *head, *tail; member 121 buffer->tail = (struct packet *) buffer->data; in packet_buffer_init() 194 buffer->tail->length = length; in packet_buffer_put() 196 if (&buffer->tail->data[length] < end) { in packet_buffer_put() 197 memcpy(buffer->tail->data, data, length); in packet_buffer_put() 198 buffer->tail = (struct packet *) &buffer->tail->data[length]; in packet_buffer_put() 200 size_t split = end - buffer->tail->data; in packet_buffer_put() 202 memcpy(buffer->tail->data, data, split); in packet_buffer_put() 204 buffer->tail in packet_buffer_put() [all...] |
/kernel/linux/linux-6.6/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | tdls.c | 342 struct iwl_tdls_channel_switch_cmd_tail *tail = in iwl_mvm_tdls_config_channel_switch() local 360 tail->timing.frame_timestamp = cpu_to_le32(timestamp); in iwl_mvm_tdls_config_channel_switch() 361 tail->timing.switch_time = cpu_to_le32(switch_time); in iwl_mvm_tdls_config_channel_switch() 362 tail->timing.switch_timeout = cpu_to_le32(switch_timeout); in iwl_mvm_tdls_config_channel_switch() 398 tail->timing.max_offchan_duration = in iwl_mvm_tdls_config_channel_switch() 403 tail->frame.switch_time_offset = cpu_to_le32(ch_sw_tm_ie + 2); in iwl_mvm_tdls_config_channel_switch() 413 iwl_mvm_set_tx_cmd_ccmp(info, &tail->frame.tx_cmd); in iwl_mvm_tdls_config_channel_switch() 416 iwl_mvm_set_tx_cmd(mvm, skb, &tail->frame.tx_cmd, info, in iwl_mvm_tdls_config_channel_switch() 419 iwl_mvm_set_tx_cmd_rate(mvm, &tail->frame.tx_cmd, info, sta, in iwl_mvm_tdls_config_channel_switch() 423 memcpy(tail in iwl_mvm_tdls_config_channel_switch() [all...] |
/kernel/linux/linux-5.10/arch/arm/crypto/ |
H A D | aes-ce-glue.c | 401 u8 __aligned(8) tail[AES_BLOCK_SIZE]; in ctr_encrypt() 407 * Tell aes_ctr_encrypt() to process a tail block. in ctr_encrypt() 412 ce_aes_ctr_encrypt(tail, NULL, ctx->key_enc, num_rounds(ctx), in ctr_encrypt() 415 crypto_xor_cpy(tdst, tsrc, tail, nbytes); in ctr_encrypt() 449 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_encrypt() local 460 if (unlikely(tail > 0 && walk.nbytes < walk.total)) { in xts_encrypt() 476 tail = 0; in xts_encrypt() 493 if (err || likely(!tail)) in xts_encrypt() 500 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_encrypt() 521 int tail in xts_decrypt() local [all...] |
/kernel/linux/linux-6.6/arch/arm/crypto/ |
H A D | aes-ce-glue.c | 401 u8 __aligned(8) tail[AES_BLOCK_SIZE]; in ctr_encrypt() 407 * Tell aes_ctr_encrypt() to process a tail block. in ctr_encrypt() 412 ce_aes_ctr_encrypt(tail, NULL, ctx->key_enc, num_rounds(ctx), in ctr_encrypt() 415 crypto_xor_cpy(tdst, tsrc, tail, nbytes); in ctr_encrypt() 449 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_encrypt() local 460 if (unlikely(tail > 0 && walk.nbytes < walk.total)) { in xts_encrypt() 476 tail = 0; in xts_encrypt() 493 if (err || likely(!tail)) in xts_encrypt() 500 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_encrypt() 521 int tail in xts_decrypt() local [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 150 unsigned long tail; in hfi1_ipoib_drain_tx_ring() local 159 tail = tx_ring->tail; in hfi1_ipoib_drain_tx_ring() 162 work_done = min_t(int, CIRC_CNT(head, tail, max_tx), budget); in hfi1_ipoib_drain_tx_ring() 165 hfi1_ipoib_free_tx(tx_ring->items[tail], budget); in hfi1_ipoib_drain_tx_ring() 166 tail = CIRC_NEXT(tail, max_tx); in hfi1_ipoib_drain_tx_ring() 171 /* Finished freeing tx items so store the tail value. */ in hfi1_ipoib_drain_tx_ring() 172 smp_store_release(&tx_ring->tail, tail); in hfi1_ipoib_drain_tx_ring() 198 unsigned long tail; hfi1_ipoib_add_tx() local [all...] |
/kernel/linux/linux-5.10/drivers/thunderbolt/ |
H A D | nhi.c | 177 return ((ring->head + 1) % ring->size) == ring->tail; in ring_full() 182 return ring->head == ring->tail; in ring_empty() 243 if (!(ring->descriptors[ring->tail].flags in ring_work() 250 frame->size = ring->descriptors[ring->tail].length; in ring_work() 251 frame->eof = ring->descriptors[ring->tail].eof; in ring_work() 252 frame->sof = ring->descriptors[ring->tail].sof; in ring_work() 253 frame->flags = ring->descriptors[ring->tail].flags; in ring_work() 255 ring->tail = (ring->tail + 1) % ring->size; in ring_work() 311 if (ring->descriptors[ring->tail] in tb_ring_poll() [all...] |
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/ |
H A D | rpc_rdma.c | 178 return (buf->head[0].iov_len + buf->tail[0].iov_len) < in rpcrdma_nonpayload_inline() 271 /* When encoding a Read chunk, the tail iovec contains an in rpcrdma_convert_iovs() 279 * layer provides space in the tail iovec that may be used in rpcrdma_convert_iovs() 285 if (xdrbuf->tail[0].iov_len) in rpcrdma_convert_iovs() 286 seg = rpcrdma_convert_kvec(&xdrbuf->tail[0], seg, &n); in rpcrdma_convert_iovs() 648 /* The tail iovec may include an XDR pad for the page list, 659 struct page *page = virt_to_page(xdr->tail[0].iov_base); in rpcrdma_prepare_tail_iov() 676 /* Copy the tail to the end of the head buffer. 686 memmove(dst, xdr->tail[0].iov_base, xdr->tail[ in rpcrdma_pullup_tail_iov() 746 struct kvec *tail = &xdr->tail[0]; rpcrdma_prepare_noch_mapped() local [all...] |
/kernel/linux/linux-6.6/drivers/thunderbolt/ |
H A D | nhi.c | 226 return ((ring->head + 1) % ring->size) == ring->tail; in ring_full() 231 return ring->head == ring->tail; in ring_empty() 292 if (!(ring->descriptors[ring->tail].flags in ring_work() 299 frame->size = ring->descriptors[ring->tail].length; in ring_work() 300 frame->eof = ring->descriptors[ring->tail].eof; in ring_work() 301 frame->sof = ring->descriptors[ring->tail].sof; in ring_work() 302 frame->flags = ring->descriptors[ring->tail].flags; in ring_work() 304 ring->tail = (ring->tail + 1) % ring->size; in ring_work() 360 if (ring->descriptors[ring->tail] in tb_ring_poll() [all...] |