/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/include/nvhw/ |
H A D | drf.h | 140 #define DRF_MR_X(er,ew,ty,p,o,dr,m,v) ({ \ 142 DRF_WR_X(ew, (p), (o), dr, (_t & ~(m)) | (v)); \ 145 #define DRF_MR_N(X,er,ew,ty,p,o,d,r ,m,v) DRF_MR_X(er, ew, ty, (p), (o), d##_##r , (m), (v)) 146 #define DRF_MR_I(X,er,ew,ty,p,o,d,r,i,m,v) DRF_MR_X(er, ew, ty, (p), (o), d##_##r(i), (m), (v)) 174 #define DRF_MV_N(X,er,ew,ty,p,o,d,r, f,v) \ 175 NVVAL_GET_X(DRF_MR_X(er, ew, ty, (p), (o), d##_##r , DRF_SMASK(d##_##r##_##f), \ 177 #define DRF_MV_I(X,er,ew,ty,p,o,d,r,i,f,v) \ 178 NVVAL_GET_X(DRF_MR_X(er, ew, t [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/include/nvhw/ |
H A D | drf.h | 140 #define DRF_MR_X(er,ew,ty,p,o,dr,m,v) ({ \ 142 DRF_WR_X(ew, (p), (o), dr, (_t & ~(m)) | (v)); \ 145 #define DRF_MR_N(X,er,ew,ty,p,o,d,r ,m,v) DRF_MR_X(er, ew, ty, (p), (o), d##_##r , (m), (v)) 146 #define DRF_MR_I(X,er,ew,ty,p,o,d,r,i,m,v) DRF_MR_X(er, ew, ty, (p), (o), d##_##r(i), (m), (v)) 174 #define DRF_MV_N(X,er,ew,ty,p,o,d,r, f,v) \ 175 NVVAL_GET_X(DRF_MR_X(er, ew, ty, (p), (o), d##_##r , DRF_SMASK(d##_##r##_##f), \ 177 #define DRF_MV_I(X,er,ew,ty,p,o,d,r,i,f,v) \ 178 NVVAL_GET_X(DRF_MR_X(er, ew, t [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
H A D | main.c | 3134 struct ib_event_work *ew = in handle_bonded_port_state_event() local 3136 struct mlx4_ib_dev *ibdev = ew->ib_dev; in handle_bonded_port_state_event() 3141 kfree(ew); in handle_bonded_port_state_event() 3184 struct ib_event_work *ew = container_of(work, struct ib_event_work, work); in ib_sl2vl_update_work() local 3185 struct mlx4_ib_dev *mdev = ew->ib_dev; in ib_sl2vl_update_work() 3186 int port = ew->port; in ib_sl2vl_update_work() 3190 kfree(ew); in ib_sl2vl_update_work() 3196 struct ib_event_work *ew; in mlx4_sched_ib_sl2vl_update_work() local 3198 ew = kmalloc(sizeof(*ew), GFP_ATOMI in mlx4_sched_ib_sl2vl_update_work() 3213 struct ib_event_work *ew; mlx4_ib_event() local [all...] |
H A D | mad.c | 1177 struct ib_event_work *ew = container_of(work, struct ib_event_work, work); in handle_port_mgmt_change_event() local 1178 struct mlx4_ib_dev *dev = ew->ib_dev; in handle_port_mgmt_change_event() 1179 struct mlx4_eqe *eqe = &(ew->ib_eqe); in handle_port_mgmt_change_event() 1274 kfree(ew); in handle_port_mgmt_change_event()
|
/kernel/linux/linux-6.6/drivers/infiniband/hw/mlx4/ |
H A D | main.c | 3132 struct ib_event_work *ew = in handle_bonded_port_state_event() local 3134 struct mlx4_ib_dev *ibdev = ew->ib_dev; in handle_bonded_port_state_event() 3139 kfree(ew); in handle_bonded_port_state_event() 3182 struct ib_event_work *ew = container_of(work, struct ib_event_work, work); in ib_sl2vl_update_work() local 3183 struct mlx4_ib_dev *mdev = ew->ib_dev; in ib_sl2vl_update_work() 3184 int port = ew->port; in ib_sl2vl_update_work() 3188 kfree(ew); in ib_sl2vl_update_work() 3194 struct ib_event_work *ew; in mlx4_sched_ib_sl2vl_update_work() local 3196 ew = kmalloc(sizeof(*ew), GFP_ATOMI in mlx4_sched_ib_sl2vl_update_work() 3213 struct ib_event_work *ew; mlx4_ib_event() local [all...] |
H A D | mad.c | 1179 struct ib_event_work *ew = container_of(work, struct ib_event_work, work); in handle_port_mgmt_change_event() local 1180 struct mlx4_ib_dev *dev = ew->ib_dev; in handle_port_mgmt_change_event() 1181 struct mlx4_eqe *eqe = &(ew->ib_eqe); in handle_port_mgmt_change_event() 1276 kfree(ew); in handle_port_mgmt_change_event()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 304 u16 *ew; in ioc3_tcpudp_checksum() local 338 ew = (u16 *)eh; in ioc3_tcpudp_checksum() 339 ehsum = ew[0] + ew[1] + ew[2] + ew[3] + ew[4] + ew[5] + ew[6]; in ioc3_tcpudp_checksum()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/ice/ |
H A D | ice_ddp.c | 396 if (fv->ew[j].prot_id != ICE_PROT_INVALID || in ice_get_prof_index_max() 397 fv->ew[j].off != ICE_FV_OFFSET_INVAL) in ice_get_prof_index_max() 743 if (fv->ew[i].prot_id == (u8)ICE_PROT_UDP_OF && in ice_get_sw_prof_type() 744 fv->ew[i].off == ICE_VNI_OFFSET) in ice_get_sw_prof_type() 748 if (fv->ew[i].prot_id == (u8)ICE_PROT_GRE_OF) in ice_get_sw_prof_type() 843 if (fv->ew[j].prot_id == in ice_get_sw_fv_list() 845 fv->ew[j].off == lkups->fv_words[i].off) in ice_get_sw_fv_list() 912 if (fv->ew[i].prot_id == ICE_PROT_INVALID && in ice_init_prof_result_bm() 913 fv->ew[i].off == ICE_FV_OFFSET_INVAL) in ice_init_prof_result_bm()
|
H A D | ice_ddp.h | 34 struct ice_fv_word ew[ICE_MAX_FV_WORDS]; member
|
H A D | ice_switch.c | 4842 fv_ext = fv->fv_ptr->ew; in ice_fill_fv_word_index()
|
/kernel/linux/linux-6.6/drivers/net/ethernet/sgi/ |
H A D | ioc3-eth.c | 304 u16 *ew; in ioc3_tcpudp_checksum() local 338 ew = (u16 *)eh; in ioc3_tcpudp_checksum() 339 ehsum = ew[0] + ew[1] + ew[2] + ew[3] + ew[4] + ew[5] + ew[6]; in ioc3_tcpudp_checksum()
|
/kernel/linux/linux-5.10/drivers/scsi/ |
H A D | sg.c | 132 struct execute_work ew; member 154 struct execute_work ew; member 1321 struct sg_request *srp = container_of(work, struct sg_request, ew.work); in sg_rq_end_io_usercontext() 1426 INIT_WORK(&srp->ew.work, sg_rq_end_io_usercontext); in sg_rq_end_io() 1427 schedule_work(&srp->ew.work); in sg_rq_end_io() 2229 struct sg_fd *sfp = container_of(work, struct sg_fd, ew.work); in sg_remove_sfp_usercontext() 2272 INIT_WORK(&sfp->ew.work, sg_remove_sfp_usercontext); in sg_remove_sfp() 2273 schedule_work(&sfp->ew.work); in sg_remove_sfp()
|
H A D | scsi_sysfs.c | 455 sdev = container_of(work, struct scsi_device, ew.work); in scsi_device_dev_release_usercontext() 519 &sdp->ew); in scsi_device_dev_release()
|
H A D | scsi_debug.c | 329 struct execute_work ew; member 4846 ew.work); in sdebug_q_cmd_wq_complete() 5068 cancel_work_sync(&sd_dp->ew.work); in stop_qc_helper() 5570 INIT_WORK(&sd_dp->ew.work, sdebug_q_cmd_wq_complete); in schedule_resp() 5578 schedule_work(&sd_dp->ew.work); in schedule_resp()
|
/kernel/linux/linux-6.6/drivers/scsi/ |
H A D | sg.c | 137 struct execute_work ew; member 159 struct execute_work ew; member 1303 struct sg_request *srp = container_of(work, struct sg_request, ew.work); in sg_rq_end_io_usercontext() 1407 INIT_WORK(&srp->ew.work, sg_rq_end_io_usercontext); in sg_rq_end_io() 1408 schedule_work(&srp->ew.work); in sg_rq_end_io() 2209 struct sg_fd *sfp = container_of(work, struct sg_fd, ew.work); in sg_remove_sfp_usercontext() 2252 INIT_WORK(&sfp->ew.work, sg_remove_sfp_usercontext); in sg_remove_sfp() 2253 schedule_work(&sfp->ew.work); in sg_remove_sfp()
|
H A D | scsi_debug.c | 342 struct execute_work ew; member 4950 ew.work); in sdebug_q_cmd_wq_complete() 5201 if (cancel_work_sync(&sd_dp->ew.work)) in stop_qc_helper() 5524 INIT_WORK(&sd_dp->ew.work, sdebug_q_cmd_wq_complete); in sdebug_alloc_queued_cmd() 5683 schedule_work(&sd_dp->ew.work); in schedule_resp()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/ice/ |
H A D | ice_flex_type.h | 18 struct ice_fv_word ew[ICE_MAX_FV_WORDS]; member
|
/kernel/linux/linux-5.10/include/scsi/ |
H A D | scsi_device.h | 228 struct execute_work ew; /* used to get process context on put */ member
|
/kernel/linux/linux-5.10/kernel/ |
H A D | workqueue.c | 3329 * @ew: guaranteed storage for the execute work structure (must 3338 int execute_in_process_context(work_func_t fn, struct execute_work *ew) in execute_in_process_context() argument 3341 fn(&ew->work); in execute_in_process_context() 3345 INIT_WORK(&ew->work, fn); in execute_in_process_context() 3346 schedule_work(&ew->work); in execute_in_process_context()
|
/kernel/linux/linux-6.6/kernel/ |
H A D | workqueue.c | 4069 * @ew: guaranteed storage for the execute work structure (must 4078 int execute_in_process_context(work_func_t fn, struct execute_work *ew) in execute_in_process_context() argument 4081 fn(&ew->work); in execute_in_process_context() 4085 INIT_WORK(&ew->work, fn); in execute_in_process_context() 4086 schedule_work(&ew->work); in execute_in_process_context()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | reg.h | 1968 MLXSW_ITEM32(reg, cwtpm, ew, 36, 1, 1);
|
/kernel/linux/linux-6.6/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | reg.h | 2418 MLXSW_ITEM32(reg, cwtpm, ew, 36, 1, 1);
|