/kernel/linux/linux-5.10/drivers/block/xen-blkback/ |
H A D | common.h | 50 * This is the maximum number of segments that would be allowed in indirect 111 * The maximum number of indirect segments (and pages) that will 114 * feature-max-indirect-segments entry), so the frontend knows how 115 * many indirect segments the backend supports. 126 struct blkif_x86_32_request_indirect indirect; member 160 uint32_t _pad1; /* offsetof(blkif_..,u.indirect.id)==8 */ 167 * The maximum number of indirect segments (and pages) that will 170 * feature-max-indirect-segments entry), so the frontend knows how 171 * many indirect segments the backend supports. 182 struct blkif_x86_64_request_indirect indirect; member [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/amdgpu/ |
H A D | vcn_v2_5.c | 469 static void vcn_v2_5_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v2_5_mc_resume_dpg_mode() argument 476 if (!indirect) { in vcn_v2_5_mc_resume_dpg_mode() 479 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 482 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 484 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 487 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 489 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 491 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 497 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 500 upper_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 682 vcn_v2_5_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v2_5_clock_gating_dpg_mode() argument 792 vcn_v2_6_enable_ras(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v2_6_enable_ras() argument 819 vcn_v2_5_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v2_5_start_dpg_mode() argument [all...] |
H A D | vcn_v4_0_3.c | 58 int inst_idx, bool indirect); 401 * @indirect: indirectly write sram 405 static void vcn_v4_0_3_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v4_0_3_mc_resume_dpg_mode() argument 415 if (!indirect) { in vcn_v4_0_3_mc_resume_dpg_mode() 419 inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 423 inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 425 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 428 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 430 VCN, 0, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 432 VCN, 0, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_3_mc_resume_dpg_mode() 613 vcn_v4_0_3_disable_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v4_0_3_disable_clock_gating_dpg_mode() argument 714 vcn_v4_0_3_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v4_0_3_start_dpg_mode() argument 1747 vcn_v4_0_3_enable_ras(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v4_0_3_enable_ras() argument [all...] |
H A D | vcn_v4_0.c | 432 * @indirect: indirectly write sram 436 static void vcn_v4_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v4_0_mc_resume_dpg_mode() argument 445 if (!indirect) { in vcn_v4_0_mc_resume_dpg_mode() 448 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 451 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 453 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 456 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 458 VCN, inst_idx, regUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 460 VCN, inst_idx, regUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 466 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v4_0_mc_resume_dpg_mode() 775 vcn_v4_0_disable_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v4_0_disable_clock_gating_dpg_mode() argument 883 vcn_v4_0_enable_ras(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v4_0_enable_ras() argument 914 vcn_v4_0_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v4_0_start_dpg_mode() argument [all...] |
H A D | vcn_v2_0.c | 387 static void vcn_v2_0_mc_resume_dpg_mode(struct amdgpu_device *adev, bool indirect) in vcn_v2_0_mc_resume_dpg_mode() argument 394 if (!indirect) { in vcn_v2_0_mc_resume_dpg_mode() 397 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_lo), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 400 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_hi), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 402 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 405 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 407 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 409 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 415 lower_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 418 upper_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 595 vcn_v2_0_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, uint8_t indirect) vcn_v2_0_clock_gating_dpg_mode() argument 795 vcn_v2_0_start_dpg_mode(struct amdgpu_device *adev, bool indirect) vcn_v2_0_start_dpg_mode() argument [all...] |
H A D | vcn_v3_0.c | 498 static void vcn_v3_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v3_0_mc_resume_dpg_mode() argument 505 if (!indirect) { in vcn_v3_0_mc_resume_dpg_mode() 508 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 511 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 513 VCN, inst_idx, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 516 VCN, inst_idx, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 518 VCN, inst_idx, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 520 VCN, inst_idx, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 526 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 529 upper_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 826 vcn_v3_0_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v3_0_clock_gating_dpg_mode() argument 942 vcn_v3_0_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v3_0_start_dpg_mode() argument [all...] |
/kernel/linux/linux-5.10/arch/x86/kernel/ |
H A D | ksysfs.c | 95 struct setup_indirect *indirect; in get_setup_data_size() local 114 indirect = (struct setup_indirect *)data->data; in get_setup_data_size() 116 if (indirect->type != SETUP_INDIRECT) in get_setup_data_size() 117 *size = indirect->len; in get_setup_data_size() 138 struct setup_indirect *indirect; in type_show() local 162 indirect = (struct setup_indirect *)data->data; in type_show() 164 ret = sprintf(buf, "0x%x\n", indirect->type); in type_show() 179 struct setup_indirect *indirect; in setup_data_data_read() local 203 indirect = (struct setup_indirect *)data->data; in setup_data_data_read() 205 if (indirect in setup_data_data_read() [all...] |
H A D | kdebugfs.c | 49 /* Is it direct data or invalid indirect one? */ in setup_data_read() 91 struct setup_indirect *indirect; in create_setup_data_nodes() local 129 indirect = (struct setup_indirect *)data->data; in create_setup_data_nodes() 131 if (indirect->type != SETUP_INDIRECT) { in create_setup_data_nodes() 132 node->paddr = indirect->addr; in create_setup_data_nodes() 133 node->type = indirect->type; in create_setup_data_nodes() 134 node->len = indirect->len; in create_setup_data_nodes()
|
/kernel/linux/linux-6.6/arch/x86/kernel/ |
H A D | ksysfs.c | 95 struct setup_indirect *indirect; in get_setup_data_size() local 114 indirect = (struct setup_indirect *)data->data; in get_setup_data_size() 116 if (indirect->type != SETUP_INDIRECT) in get_setup_data_size() 117 *size = indirect->len; in get_setup_data_size() 138 struct setup_indirect *indirect; in type_show() local 162 indirect = (struct setup_indirect *)data->data; in type_show() 164 ret = sprintf(buf, "0x%x\n", indirect->type); in type_show() 179 struct setup_indirect *indirect; in setup_data_data_read() local 203 indirect = (struct setup_indirect *)data->data; in setup_data_data_read() 205 if (indirect in setup_data_data_read() [all...] |
H A D | kdebugfs.c | 49 /* Is it direct data or invalid indirect one? */ in setup_data_read() 91 struct setup_indirect *indirect; in create_setup_data_nodes() local 129 indirect = (struct setup_indirect *)data->data; in create_setup_data_nodes() 131 if (indirect->type != SETUP_INDIRECT) { in create_setup_data_nodes() 132 node->paddr = indirect->addr; in create_setup_data_nodes() 133 node->type = indirect->type; in create_setup_data_nodes() 134 node->len = indirect->len; in create_setup_data_nodes()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/amdgpu/ |
H A D | vcn_v3_0.c | 459 static void vcn_v3_0_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v3_0_mc_resume_dpg_mode() argument 466 if (!indirect) { in vcn_v3_0_mc_resume_dpg_mode() 469 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 472 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 474 VCN, inst_idx, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 477 VCN, inst_idx, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 479 VCN, inst_idx, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 481 VCN, inst_idx, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 487 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 490 upper_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v3_0_mc_resume_dpg_mode() 780 vcn_v3_0_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v3_0_clock_gating_dpg_mode() argument 896 vcn_v3_0_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v3_0_start_dpg_mode() argument [all...] |
H A D | vcn_v2_5.c | 448 static void vcn_v2_5_mc_resume_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) in vcn_v2_5_mc_resume_dpg_mode() argument 455 if (!indirect) { in vcn_v2_5_mc_resume_dpg_mode() 458 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_lo), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 461 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN + inst_idx].tmr_mc_addr_hi), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 463 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 466 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 468 VCN, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 470 VCN, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 476 lower_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 479 upper_32_bits(adev->vcn.inst[inst_idx].gpu_addr), 0, indirect); in vcn_v2_5_mc_resume_dpg_mode() 661 vcn_v2_5_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, int inst_idx, uint8_t indirect) vcn_v2_5_clock_gating_dpg_mode() argument 771 vcn_v2_5_start_dpg_mode(struct amdgpu_device *adev, int inst_idx, bool indirect) vcn_v2_5_start_dpg_mode() argument [all...] |
H A D | vcn_v2_0.c | 384 static void vcn_v2_0_mc_resume_dpg_mode(struct amdgpu_device *adev, bool indirect) in vcn_v2_0_mc_resume_dpg_mode() argument 391 if (!indirect) { in vcn_v2_0_mc_resume_dpg_mode() 394 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_lo), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 397 (adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].tmr_mc_addr_hi), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 399 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 402 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_LOW), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 404 UVD, 0, mmUVD_LMI_VCPU_CACHE_64BIT_BAR_HIGH), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 406 UVD, 0, mmUVD_VCPU_CACHE_OFFSET0), 0, 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 412 lower_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 415 upper_32_bits(adev->vcn.inst->gpu_addr), 0, indirect); in vcn_v2_0_mc_resume_dpg_mode() 593 vcn_v2_0_clock_gating_dpg_mode(struct amdgpu_device *adev, uint8_t sram_sel, uint8_t indirect) vcn_v2_0_clock_gating_dpg_mode() argument 794 vcn_v2_0_start_dpg_mode(struct amdgpu_device *adev, bool indirect) vcn_v2_0_start_dpg_mode() argument [all...] |
/kernel/linux/linux-6.6/arch/arm64/kvm/hyp/ |
H A D | hyp-entry.S | 216 .macro hyp_ventry indirect, spectrev2 226 .if \indirect != 0 249 .macro generate_vectors indirect, spectrev2 252 hyp_ventry \indirect, \spectrev2 259 generate_vectors indirect = 0, spectrev2 = 1 // HYP_VECTOR_SPECTRE_DIRECT 260 generate_vectors indirect = 1, spectrev2 = 0 // HYP_VECTOR_INDIRECT 261 generate_vectors indirect = 1, spectrev2 = 1 // HYP_VECTOR_SPECTRE_INDIRECT
|
/kernel/linux/linux-5.10/tools/testing/selftests/bpf/progs/ |
H A D | map_ptr_kern.c | 55 static inline int check_bpf_map_ptr(struct bpf_map *indirect, in check_bpf_map_ptr() argument 58 VERIFY(indirect->map_type == direct->map_type); in check_bpf_map_ptr() 59 VERIFY(indirect->key_size == direct->key_size); in check_bpf_map_ptr() 60 VERIFY(indirect->value_size == direct->value_size); in check_bpf_map_ptr() 61 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 62 VERIFY(indirect->id == direct->id); in check_bpf_map_ptr() 63 VERIFY(indirect->memory.pages == direct->memory.pages); in check_bpf_map_ptr() 68 static inline int check(struct bpf_map *indirect, struct bpf_map *direct, in check() argument 71 VERIFY(check_bpf_map_ptr(indirect, direct)); in check() 72 VERIFY(check_bpf_map_fields(indirect, key_siz in check() 77 check_default(struct bpf_map *indirect, struct bpf_map *direct) check_default() argument 86 check_default_noinline(struct bpf_map *indirect, struct bpf_map *direct) check_default_noinline() argument [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
H A D | map_ptr_kern.c | 50 static inline int check_bpf_map_ptr(struct bpf_map *indirect, in check_bpf_map_ptr() argument 53 VERIFY(indirect->map_type == direct->map_type); in check_bpf_map_ptr() 54 VERIFY(indirect->key_size == direct->key_size); in check_bpf_map_ptr() 55 VERIFY(indirect->value_size == direct->value_size); in check_bpf_map_ptr() 56 VERIFY(indirect->max_entries == direct->max_entries); in check_bpf_map_ptr() 57 VERIFY(indirect->id == direct->id); in check_bpf_map_ptr() 62 static inline int check(struct bpf_map *indirect, struct bpf_map *direct, in check() argument 65 VERIFY(check_bpf_map_ptr(indirect, direct)); in check() 66 VERIFY(check_bpf_map_fields(indirect, key_size, value_size, in check() 71 static inline int check_default(struct bpf_map *indirect, in check_default() argument 80 check_default_noinline(struct bpf_map *indirect, struct bpf_map *direct) check_default_noinline() argument [all...] |
/kernel/linux/linux-6.6/drivers/block/xen-blkback/ |
H A D | blkback.c | 74 * the use of indirect descriptors, we set it to a value that provides good 928 pages[i]->gref = req->u.indirect.indirect_grefs[i]; in xen_blkbk_parse_indirect() 938 /* Map indirect segments */ in xen_blkbk_parse_indirect() 1105 dst->u.indirect.indirect_op = src->u.indirect.indirect_op; in blkif_get_x86_32_req() 1106 dst->u.indirect.nr_segments = in blkif_get_x86_32_req() 1107 READ_ONCE(src->u.indirect.nr_segments); in blkif_get_x86_32_req() 1108 dst->u.indirect.handle = src->u.indirect.handle; in blkif_get_x86_32_req() 1109 dst->u.indirect in blkif_get_x86_32_req() [all...] |
/kernel/linux/linux-5.10/drivers/net/can/cc770/ |
H A D | cc770_isa.c | 23 * insmod cc770_isa.ko port=0x310,0x380 indirect=1 irq=7,11 75 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 83 module_param_hw_array(indirect, int, ioport, NULL, 0444); 84 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe()
|
/kernel/linux/linux-5.10/drivers/net/can/sja1000/ |
H A D | sja1000_isa.c | 37 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 38 static spinlock_t indirect_lock[MAXDEV]; /* lock for indirect access mode */ 46 module_param_hw_array(indirect, int, ioport, NULL, 0444); 47 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe()
|
/kernel/linux/linux-6.6/drivers/net/can/sja1000/ |
H A D | sja1000_isa.c | 37 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 38 static spinlock_t indirect_lock[MAXDEV]; /* lock for indirect access mode */ 46 module_param_hw_array(indirect, int, ioport, NULL, 0444); 47 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 139 if (indirect[idx] > 0 || in sja1000_isa_probe() 140 (indirect[idx] == -1 && indirect[0] > 0)) in sja1000_isa_probe()
|
/kernel/linux/linux-6.6/drivers/net/can/cc770/ |
H A D | cc770_isa.c | 23 * insmod cc770_isa.ko port=0x310,0x380 indirect=1 irq=7,11 75 static int indirect[MAXDEV] = {[0 ... (MAXDEV - 1)] = -1}; variable 83 module_param_hw_array(indirect, int, ioport, NULL, 0444); 84 MODULE_PARM_DESC(indirect, "Indirect access via address and data port"); 184 if (indirect[idx] > 0 || in cc770_isa_probe() 185 (indirect[idx] == -1 && indirect[0] > 0)) in cc770_isa_probe()
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | ioremap.c | 642 struct setup_indirect *indirect; in memremap_is_setup_data() local 673 pr_warn("failed to memremap indirect setup_data\n"); in memremap_is_setup_data() 677 indirect = (struct setup_indirect *)data->data; in memremap_is_setup_data() 679 if (indirect->type != SETUP_INDIRECT) { in memremap_is_setup_data() 680 paddr = indirect->addr; in memremap_is_setup_data() 681 len = indirect->len; in memremap_is_setup_data() 703 struct setup_indirect *indirect; in early_memremap_is_setup_data() local 735 pr_warn("failed to early memremap indirect setup_data\n"); in early_memremap_is_setup_data() 739 indirect = (struct setup_indirect *)data->data; in early_memremap_is_setup_data() 741 if (indirect in early_memremap_is_setup_data() [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | ioremap.c | 637 struct setup_indirect *indirect; in memremap_is_setup_data() local 668 pr_warn("failed to memremap indirect setup_data\n"); in memremap_is_setup_data() 672 indirect = (struct setup_indirect *)data->data; in memremap_is_setup_data() 674 if (indirect->type != SETUP_INDIRECT) { in memremap_is_setup_data() 675 paddr = indirect->addr; in memremap_is_setup_data() 676 len = indirect->len; in memremap_is_setup_data() 698 struct setup_indirect *indirect; in early_memremap_is_setup_data() local 730 pr_warn("failed to early memremap indirect setup_data\n"); in early_memremap_is_setup_data() 734 indirect = (struct setup_indirect *)data->data; in early_memremap_is_setup_data() 736 if (indirect in early_memremap_is_setup_data() [all...] |
/kernel/linux/linux-5.10/arch/m68k/math-emu/ |
H A D | fp_decode.h | 53 * a0 - will point to source/dest operand for any indirect mode 121 | .long "addr register indirect" 122 | .long "addr register indirect postincrement" 123 | .long "addr register indirect predecrement" 184 | .long "no memory indirect action/reserved","null outer displacement" 196 | test if %pc is the base register for the indirect addr mode 220 | addressing mode: address register indirect 244 | addressing mode: address register indirect with postincrement 263 | addressing mode: address register indirect with predecrement 289 | addressing mode: address register/programm counter indirect [all...] |
/kernel/linux/linux-6.6/arch/m68k/math-emu/ |
H A D | fp_decode.h | 53 * a0 - will point to source/dest operand for any indirect mode 121 | .long "addr register indirect" 122 | .long "addr register indirect postincrement" 123 | .long "addr register indirect predecrement" 184 | .long "no memory indirect action/reserved","null outer displacement" 196 | test if %pc is the base register for the indirect addr mode 220 | addressing mode: address register indirect 244 | addressing mode: address register indirect with postincrement 263 | addressing mode: address register indirect with predecrement 289 | addressing mode: address register/programm counter indirect [all...] |