/third_party/vk-gl-cts/external/vulkancts/framework/vulkan/ |
H A D | vkResourceInterface.hpp | 50 VkDevice device) = 0; 52 virtual void deinitDevice (VkDevice device) = 0; 76 virtual void registerDeviceFeatures (VkDevice device, 78 virtual void unregisterDeviceFeatures (VkDevice device) const = 0; 79 virtual VkResult createShaderModule (VkDevice device, 84 virtual VkResult createGraphicsPipelines (VkDevice device, 91 virtual VkResult createComputePipelines (VkDevice device, 98 virtual void destroyPipeline (VkDevice device, 101 virtual void createRenderPass (VkDevice device, 105 virtual void createRenderPass2 (VkDevice device, [all...] |
/kernel/linux/linux-5.10/include/linux/ |
H A D | pm_wakeirq.h | 19 extern int dev_pm_set_wake_irq(struct device *dev, int irq); 20 extern int dev_pm_set_dedicated_wake_irq(struct device *dev, int irq); 21 extern int dev_pm_set_dedicated_wake_irq_reverse(struct device *dev, int irq); 22 extern void dev_pm_clear_wake_irq(struct device *dev); 23 extern void dev_pm_enable_wake_irq(struct device *dev); 24 extern void dev_pm_disable_wake_irq(struct device *dev); 28 static inline int dev_pm_set_wake_irq(struct device *dev, int irq) in dev_pm_set_wake_irq() 33 static inline int dev_pm_set_dedicated_wake_irq(struct device *dev, int irq) in dev_pm_set_dedicated_wake_irq() 38 static inline int dev_pm_set_dedicated_wake_irq_reverse(struct device *dev, int irq) in dev_pm_set_dedicated_wake_irq_reverse() 43 static inline void dev_pm_clear_wake_irq(struct device *de [all...] |
H A D | clk.h | 16 struct device; 317 * @dev: device for clock "consumer" 330 struct clk *clk_get(struct device *dev, const char *id); 334 * @dev: device for clock "consumer" 352 int __must_check clk_bulk_get(struct device *dev, int num_clks, 357 * @dev: device for clock "consumer" 372 int __must_check clk_bulk_get_all(struct device *dev, 377 * @dev: device for clock "consumer" 385 int __must_check clk_bulk_get_optional(struct device *dev, int num_clks, 389 * @dev: device fo [all...] |
H A D | cpu.h | 22 struct device; 29 struct device dev; 38 extern struct device *get_cpu_device(unsigned cpu); 50 extern ssize_t cpu_show_meltdown(struct device *dev, 52 extern ssize_t cpu_show_spectre_v1(struct device *dev, 54 extern ssize_t cpu_show_spectre_v2(struct device *dev, 56 extern ssize_t cpu_show_spec_store_bypass(struct device *dev, 58 extern ssize_t cpu_show_l1tf(struct device *dev, 60 extern ssize_t cpu_show_mds(struct device *dev, 62 extern ssize_t cpu_show_tsx_async_abort(struct device *de [all...] |
H A D | of_platform.h | 9 #include <linux/device.h> 16 * struct of_dev_auxdata - lookup table entry for device names & platform_data 23 * the names of devices when creating devices from the device tree. The table 28 * the device name to look up a specific device, but the Linux-specific names 29 * are not encoded into the device tree, so the kernel needs to provide specific 34 * device name will not matter, and drivers should obtain data from the device 54 struct device *parent); 67 struct device *paren [all...] |
/third_party/mesa3d/src/virtio/venus-protocol/ |
H A D | vn_protocol_driver_descriptor_pool.h | 193 static inline size_t vn_sizeof_vkCreateDescriptorPool(VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_sizeof_vkCreateDescriptorPool() argument 199 cmd_size += vn_sizeof_VkDevice(&device); in vn_sizeof_vkCreateDescriptorPool() 213 static inline void vn_encode_vkCreateDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_encode_vkCreateDescriptorPool() argument 220 vn_encode_VkDevice(enc, &device); in vn_encode_vkCreateDescriptorPool() 229 static inline size_t vn_sizeof_vkCreateDescriptorPool_reply(VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_sizeof_vkCreateDescriptorPool_reply() argument 236 /* skip device */ in vn_sizeof_vkCreateDescriptorPool_reply() 246 static inline VkResult vn_decode_vkCreateDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_decode_vkCreateDescriptorPool_reply() argument 254 /* skip device */ in vn_decode_vkCreateDescriptorPool_reply() 266 static inline size_t vn_sizeof_vkDestroyDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) in vn_sizeof_vkDestroyDescriptorPool() argument 272 cmd_size += vn_sizeof_VkDevice(&device); in vn_sizeof_vkDestroyDescriptorPool() 281 vn_encode_vkDestroyDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_encode_vkDestroyDescriptorPool() argument 294 vn_sizeof_vkDestroyDescriptorPool_reply(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_sizeof_vkDestroyDescriptorPool_reply() argument 306 vn_decode_vkDestroyDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_decode_vkDestroyDescriptorPool_reply() argument 317 vn_sizeof_vkResetDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_sizeof_vkResetDescriptorPool() argument 330 vn_encode_vkResetDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_encode_vkResetDescriptorPool() argument 342 vn_sizeof_vkResetDescriptorPool_reply(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_sizeof_vkResetDescriptorPool_reply() argument 356 vn_decode_vkResetDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_decode_vkResetDescriptorPool_reply() argument 371 vn_submit_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool, struct vn_instance_submit_command *submit) vn_submit_vkCreateDescriptorPool() argument 392 vn_submit_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator, struct vn_instance_submit_command *submit) vn_submit_vkDestroyDescriptorPool() argument 413 vn_submit_vkResetDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags, struct vn_instance_submit_command *submit) vn_submit_vkResetDescriptorPool() argument 434 vn_call_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) vn_call_vkCreateDescriptorPool() argument 450 vn_async_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) vn_async_vkCreateDescriptorPool() argument 456 vn_call_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_call_vkDestroyDescriptorPool() argument 469 vn_async_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_async_vkDestroyDescriptorPool() argument 475 vn_call_vkResetDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_call_vkResetDescriptorPool() argument 491 vn_async_vkResetDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_async_vkResetDescriptorPool() argument [all...] |
/third_party/mesa3d/src/vulkan/runtime/ |
H A D | vk_sync.c | 81 vk_sync_init(struct vk_device *device, in vk_sync_init() argument 99 return type->init(device, sync, initial_value); in vk_sync_init() 103 vk_sync_finish(struct vk_device *device, in vk_sync_finish() argument 106 sync->type->finish(device, sync); in vk_sync_finish() 110 vk_sync_create(struct vk_device *device, in vk_sync_create() argument 118 sync = vk_alloc(&device->alloc, type->size, 8, in vk_sync_create() 121 return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY); in vk_sync_create() 123 VkResult result = vk_sync_init(device, sync, type, flags, initial_value); in vk_sync_create() 125 vk_free(&device->alloc, sync); in vk_sync_create() 135 vk_sync_destroy(struct vk_device *device, in vk_sync_destroy() argument 143 vk_sync_signal(struct vk_device *device, struct vk_sync *sync, uint64_t value) vk_sync_signal() argument 158 vk_sync_get_value(struct vk_device *device, struct vk_sync *sync, uint64_t *value) vk_sync_get_value() argument 167 vk_sync_reset(struct vk_device *device, struct vk_sync *sync) vk_sync_reset() argument 175 vk_sync_move(struct vk_device *device, struct vk_sync *dst, struct vk_sync *src) vk_sync_move() argument 214 __vk_sync_wait(struct vk_device *device, struct vk_sync *sync, uint64_t wait_value, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) __vk_sync_wait() argument 240 vk_sync_wait(struct vk_device *device, struct vk_sync *sync, uint64_t wait_value, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) vk_sync_wait() argument 282 __vk_sync_wait_many(struct vk_device *device, uint32_t wait_count, const struct vk_sync_wait *waits, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) __vk_sync_wait_many() argument 328 vk_sync_wait_many(struct vk_device *device, uint32_t wait_count, const struct vk_sync_wait *waits, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) vk_sync_wait_many() argument 349 vk_sync_import_opaque_fd(struct vk_device *device, struct vk_sync *sync, int fd) vk_sync_import_opaque_fd() argument 364 vk_sync_export_opaque_fd(struct vk_device *device, struct vk_sync *sync, int *fd) vk_sync_export_opaque_fd() argument 380 vk_sync_import_sync_file(struct vk_device *device, struct vk_sync *sync, int sync_file) vk_sync_import_sync_file() argument 396 vk_sync_export_sync_file(struct vk_device *device, struct vk_sync *sync, int *sync_file) vk_sync_export_sync_file() argument [all...] |
H A D | vk_semaphore.c | 126 VK_FROM_HANDLE(vk_device, device, _device); in vk_common_CreateSemaphore() 136 assert(device->timeline_mode != VK_DEVICE_TIMELINE_MODE_NONE); in vk_common_CreateSemaphore() 144 get_semaphore_sync_type(device->physical, semaphore_type, handle_types); in vk_common_CreateSemaphore() 147 assert(get_semaphore_sync_type(device->physical, semaphore_type, 0) != NULL); in vk_common_CreateSemaphore() 148 return vk_errorf(device, VK_ERROR_INVALID_EXTERNAL_HANDLE, in vk_common_CreateSemaphore() 159 vk_device_supports_threaded_submit(device)) in vk_common_CreateSemaphore() 167 semaphore = vk_object_zalloc(device, pAllocator, size, in vk_common_CreateSemaphore() 170 return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY); in vk_common_CreateSemaphore() 180 VkResult result = vk_sync_init(device, &semaphore->permanent, in vk_common_CreateSemaphore() 183 vk_object_free(device, pAllocato in vk_common_CreateSemaphore() 193 vk_semaphore_reset_temporary(struct vk_device *device, struct vk_semaphore *semaphore) vk_semaphore_reset_temporary() argument [all...] |
/third_party/backends/frontend/ |
H A D | tstbackend.c | 45 {"device-name", required_argument, NULL, 'd'}, 54 test_options (SANE_Device * device, int can_do_recursive); 212 test_parameters (SANE_Device * device, SANE_Parameters *params) in test_parameters() argument 217 status = sane_get_parameters (device, &p); in test_parameters() 244 test_options_word_list (SANE_Device * device, int option_num, in test_options_word_list() argument 264 status = sane_control_option (device, option_num, in test_options_word_list() 275 test_options(device, can_do_recursive-1); in test_options_word_list() 278 test_parameters(device, NULL); in test_options_word_list() 291 test_options_string_list (SANE_Device * device, int option_num, in test_options_string_list() argument 315 status = sane_control_option (device, option_nu in test_options_string_list() 343 test_options(SANE_Device * device, int can_do_recursive) test_options() argument 860 get_optdesc_by_name(SANE_Handle device, const char *name, int *option_num) get_optdesc_by_name() argument 887 set_min_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_min_value() argument 938 set_max_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_max_value() argument 990 set_random_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_random_value() argument 1048 get_option_value(SANE_Handle device, const char *option_name) get_option_value() argument 1108 display_scan_parameters(SANE_Handle device) display_scan_parameters() argument 1127 test_scan(SANE_Handle device) test_scan() argument 1478 test_scans(SANE_Device * device) test_scans() argument 1666 test_default(SANE_Device * device) test_default() argument 1688 SANE_Handle device; main() local [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | ctxnv40.c | 120 nv40_gr_vs_count(struct nvkm_device *device) in nv40_gr_vs_count() argument 123 switch (device->chipset) { in nv40_gr_vs_count() 161 struct nvkm_device *device = ctx->device; in nv40_gr_construct_general() local 188 if (device->chipset == 0x40) { in nv40_gr_construct_general() 209 if (nv44_gr_class(ctx->device)) { in nv40_gr_construct_general() 218 if (!nv44_gr_class(ctx->device)) { in nv40_gr_construct_general() 224 if (device->chipset == 0x4c || in nv40_gr_construct_general() 225 (device->chipset & 0xf0) == 0x60) in nv40_gr_construct_general() 232 switch (device in nv40_gr_construct_general() 267 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d() local 372 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d_2() local 551 struct nvkm_device *device = ctx->device; nv40_gr_construct_shader() local 661 nv40_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem) nv40_grctx_fill() argument 671 nv40_grctx_init(struct nvkm_device *device, u32 *size) nv40_grctx_init() argument [all...] |
H A D | ctxgv100.c | 115 struct nvkm_device *device = gr->base.engine.subdev.device; in gv100_grctx_generate_rop_mapping() local 120 nvkm_wr32(device, 0x418bb8, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 125 nvkm_wr32(device, 0x418b08 + (i * 4), data); in gv100_grctx_generate_rop_mapping() 126 nvkm_wr32(device, 0x41bf00 + (i * 4), data); in gv100_grctx_generate_rop_mapping() 127 nvkm_wr32(device, 0x40780c + (i * 4), data); in gv100_grctx_generate_rop_mapping() 131 nvkm_wr32(device, 0x41bfd0, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 138 nvkm_wr32(device, 0x41bfb0 + (i * 4), (v22 << 24) | in gv100_grctx_generate_rop_mapping() 145 nvkm_wr32(device, 0x4078bc, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 152 struct nvkm_device *device in gv100_grctx_generate_r400088() local 159 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_sm_id() local 168 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_unkn() local 179 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_unkn88c() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | ctxnv40.c | 120 nv40_gr_vs_count(struct nvkm_device *device) in nv40_gr_vs_count() argument 123 switch (device->chipset) { in nv40_gr_vs_count() 161 struct nvkm_device *device = ctx->device; in nv40_gr_construct_general() local 188 if (device->chipset == 0x40) { in nv40_gr_construct_general() 209 if (nv44_gr_class(ctx->device)) { in nv40_gr_construct_general() 218 if (!nv44_gr_class(ctx->device)) { in nv40_gr_construct_general() 224 if (device->chipset == 0x4c || in nv40_gr_construct_general() 225 (device->chipset & 0xf0) == 0x60) in nv40_gr_construct_general() 232 switch (device in nv40_gr_construct_general() 267 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d() local 372 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d_2() local 551 struct nvkm_device *device = ctx->device; nv40_gr_construct_shader() local 661 nv40_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem) nv40_grctx_fill() argument 671 nv40_grctx_init(struct nvkm_device *device, u32 *size) nv40_grctx_init() argument [all...] |
H A D | ctxgv100.c | 117 struct nvkm_device *device = gr->base.engine.subdev.device; in gv100_grctx_generate_rop_mapping() local 123 nvkm_wr32(device, 0x418bb8, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 128 nvkm_wr32(device, 0x418b08 + (i * 4), data); in gv100_grctx_generate_rop_mapping() 129 nvkm_wr32(device, 0x41bf00 + (i * 4), data); in gv100_grctx_generate_rop_mapping() 130 nvkm_wr32(device, 0x40780c + (i * 4), data); in gv100_grctx_generate_rop_mapping() 134 nvkm_wr32(device, 0x41bfd0, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 141 nvkm_wr32(device, 0x41bfb0 + (i * 4), (v22 << 24) | in gv100_grctx_generate_rop_mapping() 148 nvkm_wr32(device, 0x4078bc, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping() 155 struct nvkm_device *device in gv100_grctx_generate_r400088() local 162 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_sm_id() local 174 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_unkn() local 185 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_unkn88c() local [all...] |
H A D | tu102.c | 30 nvkm_wr32(gr->base.engine.subdev.device, 0x409c24, 0x006e0003); in tu102_gr_init_fecs_exceptions() 36 struct nvkm_device *device = gr->base.engine.subdev.device; in tu102_gr_init_fs() local 45 nvkm_wr32(device, GPC_UNIT(gr->sm[sm].gpc, 0x0c10 + tpc * 4), sm); in tu102_gr_init_fs() 55 struct nvkm_device *device = gr->base.engine.subdev.device; in tu102_gr_init_zcull() local 66 nvkm_wr32(device, GPC_BCAST(0x0980 + ((i / 8) * 4)), data); in tu102_gr_init_zcull() 70 nvkm_wr32(device, GPC_UNIT(gpc, 0x0914), in tu102_gr_init_zcull() 72 nvkm_wr32(device, GPC_UNIT(gpc, 0x0910), 0x00040000 | in tu102_gr_init_zcull() 74 nvkm_wr32(device, GPC_UNI in tu102_gr_init_zcull() 83 struct nvkm_device *device = gr->base.engine.subdev.device; tu102_gr_init_gpc_mmu() local 217 tu102_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) tu102_gr_new() argument [all...] |
/third_party/mesa3d/include/android_stub/hardware/ |
H A D | gralloc1.h | 36 * allocate may be NULL, which instructs the device to report whether the 82 * dequeued buffer before. May be ignored by the gralloc device. */ 277 * gralloc1_capability_t above) supported by this device. This list must not 278 * change after the device has been loaded. 285 * outCapabilities - a list of capabilities supported by this device; may 290 void (*getCapabilities)(struct gralloc1_device* device, uint32_t* outCount, 301 * or NULL if the described function is not supported by this device. 303 gralloc1_function_pointer_t (*getFunction)(struct gralloc1_device* device, 308 gralloc1_device_t** device) { in gralloc1_open() 310 TO_HW_DEVICE_T_OPEN(device)); in gralloc1_open() 307 gralloc1_open(const struct hw_module_t* module, gralloc1_device_t** device) gralloc1_open() argument 313 gralloc1_close(gralloc1_device_t* device) gralloc1_close() argument [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/core/ |
H A D | agent.c | 57 __ib_get_agent_port(const struct ib_device *device, int port_num) in __ib_get_agent_port() argument 62 if (entry->agent[1]->device == device && in __ib_get_agent_port() 70 ib_get_agent_port(const struct ib_device *device, int port_num) in ib_get_agent_port() argument 76 entry = __ib_get_agent_port(device, port_num); in ib_get_agent_port() 82 const struct ib_wc *wc, const struct ib_device *device, in agent_send_response() 91 if (rdma_cap_ib_switch(device)) in agent_send_response() 92 port_priv = ib_get_agent_port(device, 0); in agent_send_response() 94 port_priv = ib_get_agent_port(device, port_num); in agent_send_response() 97 dev_err(&device in agent_send_response() 81 agent_send_response(const struct ib_mad_hdr *mad_hdr, const struct ib_grh *grh, const struct ib_wc *wc, const struct ib_device *device, int port_num, int qpn, size_t resp_mad_len, bool opa) agent_send_response() argument 150 ib_agent_port_open(struct ib_device *device, int port_num) ib_agent_port_open() argument 200 ib_agent_port_close(struct ib_device *device, int port_num) ib_agent_port_close() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
H A D | sorgm200.c | 29 struct nvkm_device *device = sor->disp->engine.subdev.device; in gm200_sor_dp_drive() local 36 data[0] = nvkm_rd32(device, 0x61c118 + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive() 37 data[1] = nvkm_rd32(device, 0x61c120 + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive() 38 data[2] = nvkm_rd32(device, 0x61c130 + loff); in gm200_sor_dp_drive() 41 nvkm_wr32(device, 0x61c118 + loff, data[0] | (dc << shift)); in gm200_sor_dp_drive() 42 nvkm_wr32(device, 0x61c120 + loff, data[1] | (pe << shift)); in gm200_sor_dp_drive() 43 nvkm_wr32(device, 0x61c130 + loff, data[2]); in gm200_sor_dp_drive() 44 data[3] = nvkm_rd32(device, 0x61c13c + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive() 45 nvkm_wr32(device, in gm200_sor_dp_drive() 51 struct nvkm_device *device = outp->disp->engine.subdev.device; gm200_sor_route_set() local 68 struct nvkm_device *device = outp->disp->engine.subdev.device; gm200_sor_route_get() local 151 struct nvkm_device *device = disp->engine.subdev.device; gm200_sor_new() local [all...] |
H A D | coregv100.c | 138 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_idle() local 139 nvkm_msec(device, 2000, in gv100_disp_core_idle() 140 u32 stat = nvkm_rd32(device, 0x610630); in gv100_disp_core_idle() 157 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_intr() local 160 nvkm_mask(device, 0x611dac, mask, data); in gv100_disp_core_intr() 166 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_fini() local 167 nvkm_mask(device, in gv100_disp_core_fini() 177 struct nvkm_device *device = subdev->device; gv100_disp_core_init() local [all...] |
H A D | headnv04.c | 29 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_vblank_put() local 30 nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000000); in nv04_head_vblank_put() 36 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_vblank_get() local 37 nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000001); in nv04_head_vblank_get() 43 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_rgpos() local 44 u32 data = nvkm_rd32(device, 0x600868 + (head->id * 0x2000)); in nv04_head_rgpos() 52 struct nvkm_device *device in nv04_head_state() local [all...] |
/kernel/linux/linux-6.6/include/linux/ |
H A D | cpu.h | 23 struct device; 30 struct device dev; 39 extern struct device *get_cpu_device(unsigned cpu); 51 extern ssize_t cpu_show_meltdown(struct device *dev, 53 extern ssize_t cpu_show_spectre_v1(struct device *dev, 55 extern ssize_t cpu_show_spectre_v2(struct device *dev, 57 extern ssize_t cpu_show_spec_store_bypass(struct device *dev, 59 extern ssize_t cpu_show_l1tf(struct device *dev, 61 extern ssize_t cpu_show_mds(struct device *dev, 63 extern ssize_t cpu_show_tsx_async_abort(struct device *de [all...] |
/kernel/linux/linux-6.6/drivers/infiniband/core/ |
H A D | agent.c | 57 __ib_get_agent_port(const struct ib_device *device, int port_num) in __ib_get_agent_port() argument 62 if (entry->agent[1]->device == device && in __ib_get_agent_port() 70 ib_get_agent_port(const struct ib_device *device, int port_num) in ib_get_agent_port() argument 76 entry = __ib_get_agent_port(device, port_num); in ib_get_agent_port() 82 const struct ib_wc *wc, const struct ib_device *device, in agent_send_response() 91 if (rdma_cap_ib_switch(device)) in agent_send_response() 92 port_priv = ib_get_agent_port(device, 0); in agent_send_response() 94 port_priv = ib_get_agent_port(device, port_num); in agent_send_response() 97 dev_err(&device in agent_send_response() 81 agent_send_response(const struct ib_mad_hdr *mad_hdr, const struct ib_grh *grh, const struct ib_wc *wc, const struct ib_device *device, int port_num, int qpn, size_t resp_mad_len, bool opa) agent_send_response() argument 150 ib_agent_port_open(struct ib_device *device, int port_num) ib_agent_port_open() argument 200 ib_agent_port_close(struct ib_device *device, int port_num) ib_agent_port_close() argument [all...] |
/third_party/skia/third_party/externals/dawn/src/dawn_wire/server/ |
H A D | Server.cpp | 37 for (WGPUDevice device : DeviceObjects().GetAllHandles()) { in ~Server() 38 ClearDeviceCallbacks(device); in ~Server() 49 ObjectData<WGPUDevice>* device = DeviceObjects().Get(deviceId); in InjectTexture() local 50 if (device == nullptr || device->generation != deviceGeneration) { in InjectTexture() 62 data->deviceInfo = device->info.get(); in InjectTexture() 81 ObjectData<WGPUDevice>* device = DeviceObjects().Get(deviceId); in InjectSwapChain() local 82 if (device == nullptr || device->generation != deviceGeneration) { in InjectSwapChain() 94 data->deviceInfo = device in InjectSwapChain() 107 InjectDevice(WGPUDevice device, uint32_t id, uint32_t generation) InjectDevice() argument 164 ClearDeviceCallbacks(WGPUDevice device) ClearDeviceCallbacks() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | nv50.c | 139 struct nvkm_device *device = subdev->device; in nv50_fb_intr() local 140 struct nvkm_fifo *fifo = device->fifo; in nv50_fb_intr() 148 idx = nvkm_rd32(device, 0x100c90); in nv50_fb_intr() 154 nvkm_wr32(device, 0x100c90, idx | i << 24); in nv50_fb_intr() 155 trap[i] = nvkm_rd32(device, 0x100c94); in nv50_fb_intr() 157 nvkm_wr32(device, 0x100c90, idx | 0x80000000); in nv50_fb_intr() 160 if (device->chipset < 0xa3 || in nv50_fb_intr() 161 device->chipset == 0xaa || device in nv50_fb_intr() 199 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_oneinit() local 216 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_init() local 242 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_dtor() local 264 nv50_fb_new_(const struct nv50_fb_func *func, struct nvkm_device *device, int index, struct nvkm_fb **pfb) nv50_fb_new_() argument 286 nv50_fb_new(struct nvkm_device *device, int index, struct nvkm_fb **pfb) nv50_fb_new() argument [all...] |
/third_party/mesa3d/src/amd/vulkan/ |
H A D | radv_spm.c | 33 radv_spm_init_bo(struct radv_device *device) in radv_spm_init_bo() argument 35 struct radeon_winsys *ws = device->ws; in radv_spm_init_bo() 40 device->spm_trace.buffer_size = size; in radv_spm_init_bo() 41 device->spm_trace.sample_interval = sample_interval; in radv_spm_init_bo() 48 device->spm_trace.bo = bo; in radv_spm_init_bo() 52 result = ws->buffer_make_resident(ws, device->spm_trace.bo, true); in radv_spm_init_bo() 56 device->spm_trace.ptr = ws->buffer_map(device->spm_trace.bo); in radv_spm_init_bo() 57 if (!device->spm_trace.ptr) in radv_spm_init_bo() 64 radv_emit_spm_counters(struct radv_device *device, struc argument 104 radv_emit_spm_setup(struct radv_device *device, struct radeon_cmdbuf *cs) radv_emit_spm_setup() argument 187 radv_spm_init(struct radv_device *device) radv_spm_init() argument 220 radv_spm_finish(struct radv_device *device) radv_spm_finish() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/i2c/ |
H A D | busgf119.c | 36 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_drive_scl() local 37 nvkm_mask(device, bus->addr, 0x00000001, state ? 0x00000001 : 0); in gf119_i2c_bus_drive_scl() 44 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_drive_sda() local 45 nvkm_mask(device, bus->addr, 0x00000002, state ? 0x00000002 : 0); in gf119_i2c_bus_drive_sda() 52 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_sense_scl() local 53 return !!(nvkm_rd32(device, bus->addr) & 0x00000010); in gf119_i2c_bus_sense_scl() 60 struct nvkm_device *device in gf119_i2c_bus_sense_sda() local 68 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; gf119_i2c_bus_init() local [all...] |