/kernel/linux/linux-5.10/drivers/crypto/cavium/zip/ |
H A D | zip_device.c | 59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed() 102 * 1. If there is place for new instructions, push the cmd at sw_head. in zip_load_instr() 104 * at the sw_head. Make sw_head point to the sw_tail to make it in zip_load_instr() 105 * circular. Write sw_head's physical address to the "Next-Chunk in zip_load_instr() 109 zip_dbg("sw_head : %lx", zip_dev->iq[queue].sw_head); in zip_load_instr() 117 memcpy((u8 *)zip_dev->iq[queue].sw_head, (u8 *)instr, in zip_load_instr() 119 zip_dev->iq[queue].sw_head += 16; /* 16 64_bit words = 128B */ in zip_load_instr() 121 /* Now, point the "Next-Chunk Buffer Ptr" to sw_head */ in zip_load_instr() [all...] |
H A D | zip_mem.c | 59 zip->iq[q].sw_head = (u64 *)__get_free_pages((GFP_KERNEL | GFP_DMA), in zip_cmd_qbuf_alloc() 62 if (!zip->iq[q].sw_head) in zip_cmd_qbuf_alloc() 65 memset(zip->iq[q].sw_head, 0, ZIP_CMD_QBUF_SIZE); in zip_cmd_qbuf_alloc() 67 zip_dbg("cmd_qbuf_alloc[%d] Success : %p\n", q, zip->iq[q].sw_head); in zip_cmd_qbuf_alloc()
|
H A D | zip_main.c | 180 zip->iq[q].sw_tail = zip->iq[q].sw_head; in zip_init_hw() 181 zip->iq[q].hw_tail = zip->iq[q].sw_head; in zip_init_hw() 185 que_sbuf_addr.s.ptr = (__pa(zip->iq[q].sw_head) >> in zip_init_hw() 197 zip_dbg("sw_head :0x%lx sw_tail :0x%lx hw_tail :0x%lx", in zip_init_hw() 198 zip->iq[q].sw_head, zip->iq[q].sw_tail, in zip_init_hw() 200 zip_dbg("sw_head phy addr : 0x%lx", que_sbuf_addr.s.ptr); in zip_init_hw()
|
H A D | zip_main.h | 86 u64 *sw_head; member
|
/kernel/linux/linux-6.6/drivers/crypto/cavium/zip/ |
H A D | zip_device.c | 59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed() 102 * 1. If there is place for new instructions, push the cmd at sw_head. in zip_load_instr() 104 * at the sw_head. Make sw_head point to the sw_tail to make it in zip_load_instr() 105 * circular. Write sw_head's physical address to the "Next-Chunk in zip_load_instr() 109 zip_dbg("sw_head : %lx", zip_dev->iq[queue].sw_head); in zip_load_instr() 117 memcpy((u8 *)zip_dev->iq[queue].sw_head, (u8 *)instr, in zip_load_instr() 119 zip_dev->iq[queue].sw_head += 16; /* 16 64_bit words = 128B */ in zip_load_instr() 121 /* Now, point the "Next-Chunk Buffer Ptr" to sw_head */ in zip_load_instr() [all...] |
H A D | zip_mem.c | 59 zip->iq[q].sw_head = (u64 *)__get_free_pages((GFP_KERNEL | GFP_DMA), in zip_cmd_qbuf_alloc() 62 if (!zip->iq[q].sw_head) in zip_cmd_qbuf_alloc() 65 memset(zip->iq[q].sw_head, 0, ZIP_CMD_QBUF_SIZE); in zip_cmd_qbuf_alloc() 67 zip_dbg("cmd_qbuf_alloc[%d] Success : %p\n", q, zip->iq[q].sw_head); in zip_cmd_qbuf_alloc()
|
H A D | zip_main.c | 185 zip->iq[q].sw_tail = zip->iq[q].sw_head; in zip_init_hw() 186 zip->iq[q].hw_tail = zip->iq[q].sw_head; in zip_init_hw() 190 que_sbuf_addr.s.ptr = (__pa(zip->iq[q].sw_head) >> in zip_init_hw() 202 zip_dbg("sw_head :0x%lx sw_tail :0x%lx hw_tail :0x%lx", in zip_init_hw() 203 zip->iq[q].sw_head, zip->iq[q].sw_tail, in zip_init_hw() 205 zip_dbg("sw_head phy addr : 0x%lx", que_sbuf_addr.s.ptr); in zip_init_hw()
|
H A D | zip_main.h | 86 u64 *sw_head; member
|
/kernel/linux/linux-5.10/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_ring.c | 223 self->sw_head = 0; in aq_ring_init() 281 budget && self->sw_head != self->hw_head; budget--) { in aq_ring_tx_clean() 282 struct aq_ring_buff_s *buff = &self->buff_ring[self->sw_head]; in aq_ring_tx_clean() 288 (!aq_ring_dx_in_range(self->sw_head, in aq_ring_tx_clean() 311 self->sw_head = aq_ring_next_dx(self, self->sw_head); in aq_ring_tx_clean() 350 for (; (self->sw_head != self->hw_head) && budget; in aq_ring_rx_clean() 351 self->sw_head = aq_ring_next_dx(self, self->sw_head), in aq_ring_rx_clean() 353 struct aq_ring_buff_s *buff = &self->buff_ring[self->sw_head]; in aq_ring_rx_clean() [all...] |
H A D | aq_ring.h | 130 unsigned int sw_head; member 165 return (((self->sw_tail >= self->sw_head)) ? in aq_ring_avail_dx() 166 (self->size - 1) - self->sw_tail + self->sw_head : in aq_ring_avail_dx() 167 self->sw_head - self->sw_tail - 1); in aq_ring_avail_dx()
|
H A D | aq_vec.c | 59 if (ring[AQ_VEC_TX_ID].sw_head != in aq_vec_poll() 70 if (ring[AQ_VEC_RX_ID].sw_head != in aq_vec_poll()
|
H A D | aq_ptp.c | 670 if (aq_ptp->ptp_tx.sw_head != aq_ptp->ptp_tx.hw_head) { in aq_ptp_poll() 682 if (aq_ptp->hwts_rx.sw_head != aq_ptp->hwts_rx.hw_head) { in aq_ptp_poll() 699 if (aq_ptp->ptp_rx.sw_head != aq_ptp->ptp_rx.hw_head) { in aq_ptp_poll()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/aquantia/atlantic/ |
H A D | aq_ring.c | 233 self->sw_head = 0; in aq_ring_init() 291 budget && self->sw_head != self->hw_head; budget--) { in aq_ring_tx_clean() 292 struct aq_ring_buff_s *buff = &self->buff_ring[self->sw_head]; in aq_ring_tx_clean() 298 (!aq_ring_dx_in_range(self->sw_head, in aq_ring_tx_clean() 333 self->sw_head = aq_ring_next_dx(self, self->sw_head); in aq_ring_tx_clean() 534 for (; (self->sw_head != self->hw_head) && budget; in __aq_ring_rx_clean() 535 self->sw_head = aq_ring_next_dx(self, self->sw_head), in __aq_ring_rx_clean() 537 struct aq_ring_buff_s *buff = &self->buff_ring[self->sw_head]; in __aq_ring_rx_clean() [all...] |
H A D | aq_ring.h | 141 unsigned int sw_head; member 181 return (((self->sw_tail >= self->sw_head)) ? in aq_ring_avail_dx() 182 (self->size - 1) - self->sw_tail + self->sw_head : in aq_ring_avail_dx() 183 self->sw_head - self->sw_tail - 1); in aq_ring_avail_dx()
|
H A D | aq_vec.c | 54 if (ring[AQ_VEC_TX_ID].sw_head != in aq_vec_poll() 65 if (ring[AQ_VEC_RX_ID].sw_head != in aq_vec_poll()
|
H A D | aq_ptp.c | 670 if (aq_ptp->ptp_tx.sw_head != aq_ptp->ptp_tx.hw_head) { in aq_ptp_poll() 682 if (aq_ptp->hwts_rx.sw_head != aq_ptp->hwts_rx.hw_head) { in aq_ptp_poll() 699 if (aq_ptp->ptp_rx.sw_head != aq_ptp->ptp_rx.hw_head) { in aq_ptp_poll()
|