Home
last modified time | relevance | path

Searched refs:device (Results 476 - 500 of 28745) sorted by relevance

1...<<11121314151617181920>>...1150

/third_party/vk-gl-cts/external/vulkancts/framework/vulkan/
H A DvkResourceInterface.hpp50 VkDevice device) = 0;
52 virtual void deinitDevice (VkDevice device) = 0;
76 virtual void registerDeviceFeatures (VkDevice device,
78 virtual void unregisterDeviceFeatures (VkDevice device) const = 0;
79 virtual VkResult createShaderModule (VkDevice device,
84 virtual VkResult createGraphicsPipelines (VkDevice device,
91 virtual VkResult createComputePipelines (VkDevice device,
98 virtual void destroyPipeline (VkDevice device,
101 virtual void createRenderPass (VkDevice device,
105 virtual void createRenderPass2 (VkDevice device,
[all...]
/kernel/linux/linux-5.10/include/linux/
H A Dpm_wakeirq.h19 extern int dev_pm_set_wake_irq(struct device *dev, int irq);
20 extern int dev_pm_set_dedicated_wake_irq(struct device *dev, int irq);
21 extern int dev_pm_set_dedicated_wake_irq_reverse(struct device *dev, int irq);
22 extern void dev_pm_clear_wake_irq(struct device *dev);
23 extern void dev_pm_enable_wake_irq(struct device *dev);
24 extern void dev_pm_disable_wake_irq(struct device *dev);
28 static inline int dev_pm_set_wake_irq(struct device *dev, int irq) in dev_pm_set_wake_irq()
33 static inline int dev_pm_set_dedicated_wake_irq(struct device *dev, int irq) in dev_pm_set_dedicated_wake_irq()
38 static inline int dev_pm_set_dedicated_wake_irq_reverse(struct device *dev, int irq) in dev_pm_set_dedicated_wake_irq_reverse()
43 static inline void dev_pm_clear_wake_irq(struct device *de
[all...]
H A Dclk.h16 struct device;
317 * @dev: device for clock "consumer"
330 struct clk *clk_get(struct device *dev, const char *id);
334 * @dev: device for clock "consumer"
352 int __must_check clk_bulk_get(struct device *dev, int num_clks,
357 * @dev: device for clock "consumer"
372 int __must_check clk_bulk_get_all(struct device *dev,
377 * @dev: device for clock "consumer"
385 int __must_check clk_bulk_get_optional(struct device *dev, int num_clks,
389 * @dev: device fo
[all...]
H A Dcpu.h22 struct device;
29 struct device dev;
38 extern struct device *get_cpu_device(unsigned cpu);
50 extern ssize_t cpu_show_meltdown(struct device *dev,
52 extern ssize_t cpu_show_spectre_v1(struct device *dev,
54 extern ssize_t cpu_show_spectre_v2(struct device *dev,
56 extern ssize_t cpu_show_spec_store_bypass(struct device *dev,
58 extern ssize_t cpu_show_l1tf(struct device *dev,
60 extern ssize_t cpu_show_mds(struct device *dev,
62 extern ssize_t cpu_show_tsx_async_abort(struct device *de
[all...]
H A Dof_platform.h9 #include <linux/device.h>
16 * struct of_dev_auxdata - lookup table entry for device names & platform_data
23 * the names of devices when creating devices from the device tree. The table
28 * the device name to look up a specific device, but the Linux-specific names
29 * are not encoded into the device tree, so the kernel needs to provide specific
34 * device name will not matter, and drivers should obtain data from the device
54 struct device *parent);
67 struct device *paren
[all...]
/third_party/mesa3d/src/virtio/venus-protocol/
H A Dvn_protocol_driver_descriptor_pool.h193 static inline size_t vn_sizeof_vkCreateDescriptorPool(VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_sizeof_vkCreateDescriptorPool() argument
199 cmd_size += vn_sizeof_VkDevice(&device); in vn_sizeof_vkCreateDescriptorPool()
213 static inline void vn_encode_vkCreateDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_encode_vkCreateDescriptorPool() argument
220 vn_encode_VkDevice(enc, &device); in vn_encode_vkCreateDescriptorPool()
229 static inline size_t vn_sizeof_vkCreateDescriptorPool_reply(VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_sizeof_vkCreateDescriptorPool_reply() argument
236 /* skip device */ in vn_sizeof_vkCreateDescriptorPool_reply()
246 static inline VkResult vn_decode_vkCreateDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) in vn_decode_vkCreateDescriptorPool_reply() argument
254 /* skip device */ in vn_decode_vkCreateDescriptorPool_reply()
266 static inline size_t vn_sizeof_vkDestroyDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) in vn_sizeof_vkDestroyDescriptorPool() argument
272 cmd_size += vn_sizeof_VkDevice(&device); in vn_sizeof_vkDestroyDescriptorPool()
281 vn_encode_vkDestroyDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_encode_vkDestroyDescriptorPool() argument
294 vn_sizeof_vkDestroyDescriptorPool_reply(VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_sizeof_vkDestroyDescriptorPool_reply() argument
306 vn_decode_vkDestroyDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_decode_vkDestroyDescriptorPool_reply() argument
317 vn_sizeof_vkResetDescriptorPool(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_sizeof_vkResetDescriptorPool() argument
330 vn_encode_vkResetDescriptorPool(struct vn_cs_encoder *enc, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_encode_vkResetDescriptorPool() argument
342 vn_sizeof_vkResetDescriptorPool_reply(VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_sizeof_vkResetDescriptorPool_reply() argument
356 vn_decode_vkResetDescriptorPool_reply(struct vn_cs_decoder *dec, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_decode_vkResetDescriptorPool_reply() argument
371 vn_submit_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool, struct vn_instance_submit_command *submit) vn_submit_vkCreateDescriptorPool() argument
392 vn_submit_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator, struct vn_instance_submit_command *submit) vn_submit_vkDestroyDescriptorPool() argument
413 vn_submit_vkResetDescriptorPool(struct vn_instance *vn_instance, VkCommandFlagsEXT cmd_flags, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags, struct vn_instance_submit_command *submit) vn_submit_vkResetDescriptorPool() argument
434 vn_call_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) vn_call_vkCreateDescriptorPool() argument
450 vn_async_vkCreateDescriptorPool(struct vn_instance *vn_instance, VkDevice device, const VkDescriptorPoolCreateInfo* pCreateInfo, const VkAllocationCallbacks* pAllocator, VkDescriptorPool* pDescriptorPool) vn_async_vkCreateDescriptorPool() argument
456 vn_call_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_call_vkDestroyDescriptorPool() argument
469 vn_async_vkDestroyDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, const VkAllocationCallbacks* pAllocator) vn_async_vkDestroyDescriptorPool() argument
475 vn_call_vkResetDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_call_vkResetDescriptorPool() argument
491 vn_async_vkResetDescriptorPool(struct vn_instance *vn_instance, VkDevice device, VkDescriptorPool descriptorPool, VkDescriptorPoolResetFlags flags) vn_async_vkResetDescriptorPool() argument
[all...]
/third_party/mesa3d/src/vulkan/runtime/
H A Dvk_sync.c81 vk_sync_init(struct vk_device *device, in vk_sync_init() argument
99 return type->init(device, sync, initial_value); in vk_sync_init()
103 vk_sync_finish(struct vk_device *device, in vk_sync_finish() argument
106 sync->type->finish(device, sync); in vk_sync_finish()
110 vk_sync_create(struct vk_device *device, in vk_sync_create() argument
118 sync = vk_alloc(&device->alloc, type->size, 8, in vk_sync_create()
121 return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY); in vk_sync_create()
123 VkResult result = vk_sync_init(device, sync, type, flags, initial_value); in vk_sync_create()
125 vk_free(&device->alloc, sync); in vk_sync_create()
135 vk_sync_destroy(struct vk_device *device, in vk_sync_destroy() argument
143 vk_sync_signal(struct vk_device *device, struct vk_sync *sync, uint64_t value) vk_sync_signal() argument
158 vk_sync_get_value(struct vk_device *device, struct vk_sync *sync, uint64_t *value) vk_sync_get_value() argument
167 vk_sync_reset(struct vk_device *device, struct vk_sync *sync) vk_sync_reset() argument
175 vk_sync_move(struct vk_device *device, struct vk_sync *dst, struct vk_sync *src) vk_sync_move() argument
214 __vk_sync_wait(struct vk_device *device, struct vk_sync *sync, uint64_t wait_value, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) __vk_sync_wait() argument
240 vk_sync_wait(struct vk_device *device, struct vk_sync *sync, uint64_t wait_value, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) vk_sync_wait() argument
282 __vk_sync_wait_many(struct vk_device *device, uint32_t wait_count, const struct vk_sync_wait *waits, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) __vk_sync_wait_many() argument
328 vk_sync_wait_many(struct vk_device *device, uint32_t wait_count, const struct vk_sync_wait *waits, enum vk_sync_wait_flags wait_flags, uint64_t abs_timeout_ns) vk_sync_wait_many() argument
349 vk_sync_import_opaque_fd(struct vk_device *device, struct vk_sync *sync, int fd) vk_sync_import_opaque_fd() argument
364 vk_sync_export_opaque_fd(struct vk_device *device, struct vk_sync *sync, int *fd) vk_sync_export_opaque_fd() argument
380 vk_sync_import_sync_file(struct vk_device *device, struct vk_sync *sync, int sync_file) vk_sync_import_sync_file() argument
396 vk_sync_export_sync_file(struct vk_device *device, struct vk_sync *sync, int *sync_file) vk_sync_export_sync_file() argument
[all...]
H A Dvk_semaphore.c126 VK_FROM_HANDLE(vk_device, device, _device); in vk_common_CreateSemaphore()
136 assert(device->timeline_mode != VK_DEVICE_TIMELINE_MODE_NONE); in vk_common_CreateSemaphore()
144 get_semaphore_sync_type(device->physical, semaphore_type, handle_types); in vk_common_CreateSemaphore()
147 assert(get_semaphore_sync_type(device->physical, semaphore_type, 0) != NULL); in vk_common_CreateSemaphore()
148 return vk_errorf(device, VK_ERROR_INVALID_EXTERNAL_HANDLE, in vk_common_CreateSemaphore()
159 vk_device_supports_threaded_submit(device)) in vk_common_CreateSemaphore()
167 semaphore = vk_object_zalloc(device, pAllocator, size, in vk_common_CreateSemaphore()
170 return vk_error(device, VK_ERROR_OUT_OF_HOST_MEMORY); in vk_common_CreateSemaphore()
180 VkResult result = vk_sync_init(device, &semaphore->permanent, in vk_common_CreateSemaphore()
183 vk_object_free(device, pAllocato in vk_common_CreateSemaphore()
193 vk_semaphore_reset_temporary(struct vk_device *device, struct vk_semaphore *semaphore) vk_semaphore_reset_temporary() argument
[all...]
/third_party/backends/frontend/
H A Dtstbackend.c45 {"device-name", required_argument, NULL, 'd'},
54 test_options (SANE_Device * device, int can_do_recursive);
212 test_parameters (SANE_Device * device, SANE_Parameters *params) in test_parameters() argument
217 status = sane_get_parameters (device, &p); in test_parameters()
244 test_options_word_list (SANE_Device * device, int option_num, in test_options_word_list() argument
264 status = sane_control_option (device, option_num, in test_options_word_list()
275 test_options(device, can_do_recursive-1); in test_options_word_list()
278 test_parameters(device, NULL); in test_options_word_list()
291 test_options_string_list (SANE_Device * device, int option_num, in test_options_string_list() argument
315 status = sane_control_option (device, option_nu in test_options_string_list()
343 test_options(SANE_Device * device, int can_do_recursive) test_options() argument
860 get_optdesc_by_name(SANE_Handle device, const char *name, int *option_num) get_optdesc_by_name() argument
887 set_min_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_min_value() argument
938 set_max_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_max_value() argument
990 set_random_value(SANE_Handle device, int option_num, const SANE_Option_Descriptor *opt) set_random_value() argument
1048 get_option_value(SANE_Handle device, const char *option_name) get_option_value() argument
1108 display_scan_parameters(SANE_Handle device) display_scan_parameters() argument
1127 test_scan(SANE_Handle device) test_scan() argument
1478 test_scans(SANE_Device * device) test_scans() argument
1666 test_default(SANE_Device * device) test_default() argument
1688 SANE_Handle device; main() local
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dctxnv40.c120 nv40_gr_vs_count(struct nvkm_device *device) in nv40_gr_vs_count() argument
123 switch (device->chipset) { in nv40_gr_vs_count()
161 struct nvkm_device *device = ctx->device; in nv40_gr_construct_general() local
188 if (device->chipset == 0x40) { in nv40_gr_construct_general()
209 if (nv44_gr_class(ctx->device)) { in nv40_gr_construct_general()
218 if (!nv44_gr_class(ctx->device)) { in nv40_gr_construct_general()
224 if (device->chipset == 0x4c || in nv40_gr_construct_general()
225 (device->chipset & 0xf0) == 0x60) in nv40_gr_construct_general()
232 switch (device in nv40_gr_construct_general()
267 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d() local
372 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d_2() local
551 struct nvkm_device *device = ctx->device; nv40_gr_construct_shader() local
661 nv40_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem) nv40_grctx_fill() argument
671 nv40_grctx_init(struct nvkm_device *device, u32 *size) nv40_grctx_init() argument
[all...]
H A Dctxgv100.c115 struct nvkm_device *device = gr->base.engine.subdev.device; in gv100_grctx_generate_rop_mapping() local
120 nvkm_wr32(device, 0x418bb8, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
125 nvkm_wr32(device, 0x418b08 + (i * 4), data); in gv100_grctx_generate_rop_mapping()
126 nvkm_wr32(device, 0x41bf00 + (i * 4), data); in gv100_grctx_generate_rop_mapping()
127 nvkm_wr32(device, 0x40780c + (i * 4), data); in gv100_grctx_generate_rop_mapping()
131 nvkm_wr32(device, 0x41bfd0, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
138 nvkm_wr32(device, 0x41bfb0 + (i * 4), (v22 << 24) | in gv100_grctx_generate_rop_mapping()
145 nvkm_wr32(device, 0x4078bc, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
152 struct nvkm_device *device in gv100_grctx_generate_r400088() local
159 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_sm_id() local
168 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_unkn() local
179 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_unkn88c() local
[all...]
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/engine/gr/
H A Dctxnv40.c120 nv40_gr_vs_count(struct nvkm_device *device) in nv40_gr_vs_count() argument
123 switch (device->chipset) { in nv40_gr_vs_count()
161 struct nvkm_device *device = ctx->device; in nv40_gr_construct_general() local
188 if (device->chipset == 0x40) { in nv40_gr_construct_general()
209 if (nv44_gr_class(ctx->device)) { in nv40_gr_construct_general()
218 if (!nv44_gr_class(ctx->device)) { in nv40_gr_construct_general()
224 if (device->chipset == 0x4c || in nv40_gr_construct_general()
225 (device->chipset & 0xf0) == 0x60) in nv40_gr_construct_general()
232 switch (device in nv40_gr_construct_general()
267 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d() local
372 struct nvkm_device *device = ctx->device; nv40_gr_construct_state3d_2() local
551 struct nvkm_device *device = ctx->device; nv40_gr_construct_shader() local
661 nv40_grctx_fill(struct nvkm_device *device, struct nvkm_gpuobj *mem) nv40_grctx_fill() argument
671 nv40_grctx_init(struct nvkm_device *device, u32 *size) nv40_grctx_init() argument
[all...]
H A Dctxgv100.c117 struct nvkm_device *device = gr->base.engine.subdev.device; in gv100_grctx_generate_rop_mapping() local
123 nvkm_wr32(device, 0x418bb8, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
128 nvkm_wr32(device, 0x418b08 + (i * 4), data); in gv100_grctx_generate_rop_mapping()
129 nvkm_wr32(device, 0x41bf00 + (i * 4), data); in gv100_grctx_generate_rop_mapping()
130 nvkm_wr32(device, 0x40780c + (i * 4), data); in gv100_grctx_generate_rop_mapping()
134 nvkm_wr32(device, 0x41bfd0, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
141 nvkm_wr32(device, 0x41bfb0 + (i * 4), (v22 << 24) | in gv100_grctx_generate_rop_mapping()
148 nvkm_wr32(device, 0x4078bc, (gr->tpc_total << 8) | in gv100_grctx_generate_rop_mapping()
155 struct nvkm_device *device in gv100_grctx_generate_r400088() local
162 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_sm_id() local
174 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_generate_unkn() local
185 struct nvkm_device *device = gr->base.engine.subdev.device; gv100_grctx_unkn88c() local
[all...]
H A Dtu102.c30 nvkm_wr32(gr->base.engine.subdev.device, 0x409c24, 0x006e0003); in tu102_gr_init_fecs_exceptions()
36 struct nvkm_device *device = gr->base.engine.subdev.device; in tu102_gr_init_fs() local
45 nvkm_wr32(device, GPC_UNIT(gr->sm[sm].gpc, 0x0c10 + tpc * 4), sm); in tu102_gr_init_fs()
55 struct nvkm_device *device = gr->base.engine.subdev.device; in tu102_gr_init_zcull() local
66 nvkm_wr32(device, GPC_BCAST(0x0980 + ((i / 8) * 4)), data); in tu102_gr_init_zcull()
70 nvkm_wr32(device, GPC_UNIT(gpc, 0x0914), in tu102_gr_init_zcull()
72 nvkm_wr32(device, GPC_UNIT(gpc, 0x0910), 0x00040000 | in tu102_gr_init_zcull()
74 nvkm_wr32(device, GPC_UNI in tu102_gr_init_zcull()
83 struct nvkm_device *device = gr->base.engine.subdev.device; tu102_gr_init_gpc_mmu() local
217 tu102_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) tu102_gr_new() argument
[all...]
/third_party/mesa3d/include/android_stub/hardware/
H A Dgralloc1.h36 * allocate may be NULL, which instructs the device to report whether the
82 * dequeued buffer before. May be ignored by the gralloc device. */
277 * gralloc1_capability_t above) supported by this device. This list must not
278 * change after the device has been loaded.
285 * outCapabilities - a list of capabilities supported by this device; may
290 void (*getCapabilities)(struct gralloc1_device* device, uint32_t* outCount,
301 * or NULL if the described function is not supported by this device.
303 gralloc1_function_pointer_t (*getFunction)(struct gralloc1_device* device,
308 gralloc1_device_t** device) { in gralloc1_open()
310 TO_HW_DEVICE_T_OPEN(device)); in gralloc1_open()
307 gralloc1_open(const struct hw_module_t* module, gralloc1_device_t** device) gralloc1_open() argument
313 gralloc1_close(gralloc1_device_t* device) gralloc1_close() argument
[all...]
/kernel/linux/linux-5.10/drivers/infiniband/core/
H A Dagent.c57 __ib_get_agent_port(const struct ib_device *device, int port_num) in __ib_get_agent_port() argument
62 if (entry->agent[1]->device == device && in __ib_get_agent_port()
70 ib_get_agent_port(const struct ib_device *device, int port_num) in ib_get_agent_port() argument
76 entry = __ib_get_agent_port(device, port_num); in ib_get_agent_port()
82 const struct ib_wc *wc, const struct ib_device *device, in agent_send_response()
91 if (rdma_cap_ib_switch(device)) in agent_send_response()
92 port_priv = ib_get_agent_port(device, 0); in agent_send_response()
94 port_priv = ib_get_agent_port(device, port_num); in agent_send_response()
97 dev_err(&device in agent_send_response()
81 agent_send_response(const struct ib_mad_hdr *mad_hdr, const struct ib_grh *grh, const struct ib_wc *wc, const struct ib_device *device, int port_num, int qpn, size_t resp_mad_len, bool opa) agent_send_response() argument
150 ib_agent_port_open(struct ib_device *device, int port_num) ib_agent_port_open() argument
200 ib_agent_port_close(struct ib_device *device, int port_num) ib_agent_port_close() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/disp/
H A Dsorgm200.c29 struct nvkm_device *device = sor->disp->engine.subdev.device; in gm200_sor_dp_drive() local
36 data[0] = nvkm_rd32(device, 0x61c118 + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive()
37 data[1] = nvkm_rd32(device, 0x61c120 + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive()
38 data[2] = nvkm_rd32(device, 0x61c130 + loff); in gm200_sor_dp_drive()
41 nvkm_wr32(device, 0x61c118 + loff, data[0] | (dc << shift)); in gm200_sor_dp_drive()
42 nvkm_wr32(device, 0x61c120 + loff, data[1] | (pe << shift)); in gm200_sor_dp_drive()
43 nvkm_wr32(device, 0x61c130 + loff, data[2]); in gm200_sor_dp_drive()
44 data[3] = nvkm_rd32(device, 0x61c13c + loff) & ~(0x000000ff << shift); in gm200_sor_dp_drive()
45 nvkm_wr32(device, in gm200_sor_dp_drive()
51 struct nvkm_device *device = outp->disp->engine.subdev.device; gm200_sor_route_set() local
68 struct nvkm_device *device = outp->disp->engine.subdev.device; gm200_sor_route_get() local
151 struct nvkm_device *device = disp->engine.subdev.device; gm200_sor_new() local
[all...]
H A Dcoregv100.c138 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_idle() local
139 nvkm_msec(device, 2000, in gv100_disp_core_idle()
140 u32 stat = nvkm_rd32(device, 0x610630); in gv100_disp_core_idle()
157 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_intr() local
160 nvkm_mask(device, 0x611dac, mask, data); in gv100_disp_core_intr()
166 struct nvkm_device *device = chan->disp->base.engine.subdev.device; in gv100_disp_core_fini() local
167 nvkm_mask(device, in gv100_disp_core_fini()
177 struct nvkm_device *device = subdev->device; gv100_disp_core_init() local
[all...]
H A Dheadnv04.c29 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_vblank_put() local
30 nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000000); in nv04_head_vblank_put()
36 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_vblank_get() local
37 nvkm_wr32(device, 0x600140 + (head->id * 0x2000) , 0x00000001); in nv04_head_vblank_get()
43 struct nvkm_device *device = head->disp->engine.subdev.device; in nv04_head_rgpos() local
44 u32 data = nvkm_rd32(device, 0x600868 + (head->id * 0x2000)); in nv04_head_rgpos()
52 struct nvkm_device *device in nv04_head_state() local
[all...]
/kernel/linux/linux-6.6/include/linux/
H A Dcpu.h23 struct device;
30 struct device dev;
39 extern struct device *get_cpu_device(unsigned cpu);
51 extern ssize_t cpu_show_meltdown(struct device *dev,
53 extern ssize_t cpu_show_spectre_v1(struct device *dev,
55 extern ssize_t cpu_show_spectre_v2(struct device *dev,
57 extern ssize_t cpu_show_spec_store_bypass(struct device *dev,
59 extern ssize_t cpu_show_l1tf(struct device *dev,
61 extern ssize_t cpu_show_mds(struct device *dev,
63 extern ssize_t cpu_show_tsx_async_abort(struct device *de
[all...]
/kernel/linux/linux-6.6/drivers/infiniband/core/
H A Dagent.c57 __ib_get_agent_port(const struct ib_device *device, int port_num) in __ib_get_agent_port() argument
62 if (entry->agent[1]->device == device && in __ib_get_agent_port()
70 ib_get_agent_port(const struct ib_device *device, int port_num) in ib_get_agent_port() argument
76 entry = __ib_get_agent_port(device, port_num); in ib_get_agent_port()
82 const struct ib_wc *wc, const struct ib_device *device, in agent_send_response()
91 if (rdma_cap_ib_switch(device)) in agent_send_response()
92 port_priv = ib_get_agent_port(device, 0); in agent_send_response()
94 port_priv = ib_get_agent_port(device, port_num); in agent_send_response()
97 dev_err(&device in agent_send_response()
81 agent_send_response(const struct ib_mad_hdr *mad_hdr, const struct ib_grh *grh, const struct ib_wc *wc, const struct ib_device *device, int port_num, int qpn, size_t resp_mad_len, bool opa) agent_send_response() argument
150 ib_agent_port_open(struct ib_device *device, int port_num) ib_agent_port_open() argument
200 ib_agent_port_close(struct ib_device *device, int port_num) ib_agent_port_close() argument
[all...]
/third_party/skia/third_party/externals/dawn/src/dawn_wire/server/
H A DServer.cpp37 for (WGPUDevice device : DeviceObjects().GetAllHandles()) { in ~Server()
38 ClearDeviceCallbacks(device); in ~Server()
49 ObjectData<WGPUDevice>* device = DeviceObjects().Get(deviceId); in InjectTexture() local
50 if (device == nullptr || device->generation != deviceGeneration) { in InjectTexture()
62 data->deviceInfo = device->info.get(); in InjectTexture()
81 ObjectData<WGPUDevice>* device = DeviceObjects().Get(deviceId); in InjectSwapChain() local
82 if (device == nullptr || device->generation != deviceGeneration) { in InjectSwapChain()
94 data->deviceInfo = device in InjectSwapChain()
107 InjectDevice(WGPUDevice device, uint32_t id, uint32_t generation) InjectDevice() argument
164 ClearDeviceCallbacks(WGPUDevice device) ClearDeviceCallbacks() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dnv50.c139 struct nvkm_device *device = subdev->device; in nv50_fb_intr() local
140 struct nvkm_fifo *fifo = device->fifo; in nv50_fb_intr()
148 idx = nvkm_rd32(device, 0x100c90); in nv50_fb_intr()
154 nvkm_wr32(device, 0x100c90, idx | i << 24); in nv50_fb_intr()
155 trap[i] = nvkm_rd32(device, 0x100c94); in nv50_fb_intr()
157 nvkm_wr32(device, 0x100c90, idx | 0x80000000); in nv50_fb_intr()
160 if (device->chipset < 0xa3 || in nv50_fb_intr()
161 device->chipset == 0xaa || device in nv50_fb_intr()
199 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_oneinit() local
216 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_init() local
242 struct nvkm_device *device = fb->base.subdev.device; nv50_fb_dtor() local
264 nv50_fb_new_(const struct nv50_fb_func *func, struct nvkm_device *device, int index, struct nvkm_fb **pfb) nv50_fb_new_() argument
286 nv50_fb_new(struct nvkm_device *device, int index, struct nvkm_fb **pfb) nv50_fb_new() argument
[all...]
/third_party/mesa3d/src/amd/vulkan/
H A Dradv_spm.c33 radv_spm_init_bo(struct radv_device *device) in radv_spm_init_bo() argument
35 struct radeon_winsys *ws = device->ws; in radv_spm_init_bo()
40 device->spm_trace.buffer_size = size; in radv_spm_init_bo()
41 device->spm_trace.sample_interval = sample_interval; in radv_spm_init_bo()
48 device->spm_trace.bo = bo; in radv_spm_init_bo()
52 result = ws->buffer_make_resident(ws, device->spm_trace.bo, true); in radv_spm_init_bo()
56 device->spm_trace.ptr = ws->buffer_map(device->spm_trace.bo); in radv_spm_init_bo()
57 if (!device->spm_trace.ptr) in radv_spm_init_bo()
64 radv_emit_spm_counters(struct radv_device *device, struc argument
104 radv_emit_spm_setup(struct radv_device *device, struct radeon_cmdbuf *cs) radv_emit_spm_setup() argument
187 radv_spm_init(struct radv_device *device) radv_spm_init() argument
220 radv_spm_finish(struct radv_device *device) radv_spm_finish() argument
[all...]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/i2c/
H A Dbusgf119.c36 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_drive_scl() local
37 nvkm_mask(device, bus->addr, 0x00000001, state ? 0x00000001 : 0); in gf119_i2c_bus_drive_scl()
44 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_drive_sda() local
45 nvkm_mask(device, bus->addr, 0x00000002, state ? 0x00000002 : 0); in gf119_i2c_bus_drive_sda()
52 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; in gf119_i2c_bus_sense_scl() local
53 return !!(nvkm_rd32(device, bus->addr) & 0x00000010); in gf119_i2c_bus_sense_scl()
60 struct nvkm_device *device in gf119_i2c_bus_sense_sda() local
68 struct nvkm_device *device = bus->base.pad->i2c->subdev.device; gf119_i2c_bus_init() local
[all...]

Completed in 14 milliseconds

1...<<11121314151617181920>>...1150