/third_party/mesa3d/src/virtio/vulkan/ |
H A D | vn_renderer_internal.c | 42 list_for_each_entry_safe(struct vn_renderer_shmem, shmem, in vn_renderer_shmem_cache_fini() 44 cache->destroy_func(cache->renderer, shmem); in vn_renderer_shmem_cache_fini() 81 list_for_each_entry_safe(struct vn_renderer_shmem, shmem, in vn_renderer_shmem_cache_remove_expired_locked() 83 if (shmem == last_shmem || in vn_renderer_shmem_cache_remove_expired_locked() 84 now - shmem->cache_timestamp < VN_RENDERER_SHMEM_CACHE_EXPIRACY) in vn_renderer_shmem_cache_remove_expired_locked() 87 list_del(&shmem->cache_head); in vn_renderer_shmem_cache_remove_expired_locked() 88 cache->destroy_func(cache->renderer, shmem); in vn_renderer_shmem_cache_remove_expired_locked() 95 struct vn_renderer_shmem *shmem) in vn_renderer_shmem_cache_add() 97 assert(!vn_refcount_is_valid(&shmem->refcount)); in vn_renderer_shmem_cache_add() 101 choose_bucket(cache, shmem in vn_renderer_shmem_cache_add() 94 vn_renderer_shmem_cache_add(struct vn_renderer_shmem_cache *cache, struct vn_renderer_shmem *shmem) vn_renderer_shmem_cache_add() argument 134 struct vn_renderer_shmem *shmem = NULL; vn_renderer_shmem_cache_get() local [all...] |
H A D | vn_renderer_util.c | 53 /* power-of-two to hit shmem cache */ in vn_renderer_shmem_pool_init() 62 if (pool->shmem) in vn_renderer_shmem_pool_fini() 63 vn_renderer_shmem_unref(renderer, pool->shmem); in vn_renderer_shmem_pool_fini() 72 /* power-of-two to hit shmem cache */ in vn_renderer_shmem_pool_grow() 80 struct vn_renderer_shmem *shmem = in vn_renderer_shmem_pool_grow() local 82 if (!shmem) in vn_renderer_shmem_pool_grow() 85 if (pool->shmem) in vn_renderer_shmem_pool_grow() 86 vn_renderer_shmem_unref(renderer, pool->shmem); in vn_renderer_shmem_pool_grow() 88 pool->shmem = shmem; in vn_renderer_shmem_pool_grow() 108 struct vn_renderer_shmem *shmem = vn_renderer_shmem_pool_alloc() local [all...] |
H A D | vn_cs.c | 76 struct vn_renderer_shmem *shmem, in vn_cs_encoder_add_buffer() 84 /* shmem ownership transferred */ in vn_cs_encoder_add_buffer() 85 cur_buf->shmem = shmem; in vn_cs_encoder_add_buffer() 113 /* when the shmem pool is used, no need to cache the shmem in cs */ in vn_cs_encoder_gc_buffers() 117 enc->buffers[i].shmem); in vn_cs_encoder_gc_buffers() 135 vn_renderer_shmem_unref(enc->instance->renderer, enc->buffers[i].shmem); in vn_cs_encoder_gc_buffers() 140 vn_cs_encoder_add_buffer(enc, cur_buf->shmem, used, in vn_cs_encoder_gc_buffers() 169 vn_renderer_shmem_unref(enc->instance->renderer, enc->buffers[i].shmem); in vn_cs_encoder_fini() 75 vn_cs_encoder_add_buffer(struct vn_cs_encoder *enc, struct vn_renderer_shmem *shmem, size_t offset, void *base, size_t size) vn_cs_encoder_add_buffer() argument 262 struct vn_renderer_shmem *shmem; vn_cs_encoder_reserve_internal() local [all...] |
H A D | vn_renderer.h | 154 struct vn_renderer_shmem *shmem); 285 struct vn_renderer_shmem *shmem = in vn_renderer_shmem_create() local 287 if (shmem) { in vn_renderer_shmem_create() 288 assert(vn_refcount_is_valid(&shmem->refcount)); in vn_renderer_shmem_create() 289 assert(shmem->res_id); in vn_renderer_shmem_create() 290 assert(shmem->mmap_size >= size); in vn_renderer_shmem_create() 291 assert(shmem->mmap_ptr); in vn_renderer_shmem_create() 294 return shmem; in vn_renderer_shmem_create() 299 struct vn_renderer_shmem *shmem) in vn_renderer_shmem_ref() 301 vn_refcount_inc(&shmem in vn_renderer_shmem_ref() 298 vn_renderer_shmem_ref(struct vn_renderer *renderer, struct vn_renderer_shmem *shmem) vn_renderer_shmem_ref() argument 306 vn_renderer_shmem_unref(struct vn_renderer *renderer, struct vn_renderer_shmem *shmem) vn_renderer_shmem_unref() argument [all...] |
H A D | vn_instance.c | 128 instance->ring.shmem = in vn_instance_init_ring() 130 if (!instance->ring.shmem) { in vn_instance_init_ring() 132 vn_log(instance, "failed to allocate/map ring shmem"); in vn_instance_init_ring() 140 instance->ring.shmem->mmap_ptr); in vn_instance_init_ring() 146 .resourceId = instance->ring.shmem->res_id, in vn_instance_init_ring() 191 /* prepare the reply shmem */ in vn_instance_init_experimental_features() 380 .resourceId = buf->shmem->res_id, in vn_instance_submission_get_cs() 428 vn_renderer_shmem_ref(ring->renderer, cs->buffers[i].shmem); in vn_instance_submission_get_ring_submit() 563 const struct vn_renderer_shmem *saved_pool_shmem = pool->shmem; in vn_instance_get_reply_shmem_locked() 566 struct vn_renderer_shmem *shmem in vn_instance_get_reply_shmem_locked() local [all...] |
H A D | vn_instance.h | 49 struct vn_renderer_shmem *shmem; member 172 struct vn_renderer_shmem *shmem; in vn_instance_cs_shmem_alloc() local 175 shmem = vn_renderer_shmem_pool_alloc( in vn_instance_cs_shmem_alloc() 179 return shmem; in vn_instance_cs_shmem_alloc()
|
H A D | vn_renderer_internal.h | 12 struct vn_renderer *renderer, struct vn_renderer_shmem *shmem); 48 struct vn_renderer_shmem *shmem);
|
H A D | vn_renderer_vtest.c | 786 struct vtest_shmem *shmem = (struct vtest_shmem *)_shmem; in vtest_shmem_destroy_now() local 788 munmap(shmem->base.mmap_ptr, shmem->base.mmap_size); in vtest_shmem_destroy_now() 791 vtest_vcmd_resource_unref(vtest, shmem->base.res_id); in vtest_shmem_destroy_now() 797 struct vn_renderer_shmem *shmem) in vtest_shmem_destroy() 801 if (vn_renderer_shmem_cache_add(&vtest->shmem_cache, shmem)) in vtest_shmem_destroy() 804 vtest_shmem_destroy_now(&vtest->base, shmem); in vtest_shmem_destroy() 837 struct vtest_shmem *shmem = in vtest_shmem_create() local 839 *shmem = (struct vtest_shmem){ in vtest_shmem_create() 848 return &shmem in vtest_shmem_create() 796 vtest_shmem_destroy(struct vn_renderer *renderer, struct vn_renderer_shmem *shmem) vtest_shmem_destroy() argument [all...] |
H A D | vn_renderer_util.h | 15 struct vn_renderer_shmem *shmem; member
|
H A D | vn_renderer_virtgpu.c | 1304 struct virtgpu_shmem *shmem = (struct virtgpu_shmem *)_shmem; in virtgpu_shmem_destroy_now() local 1306 munmap(shmem->base.mmap_ptr, shmem->base.mmap_size); in virtgpu_shmem_destroy_now() 1307 virtgpu_ioctl_gem_close(gpu, shmem->gem_handle); in virtgpu_shmem_destroy_now() 1312 struct vn_renderer_shmem *shmem) in virtgpu_shmem_destroy() 1316 if (vn_renderer_shmem_cache_add(&gpu->shmem_cache, shmem)) in virtgpu_shmem_destroy() 1319 virtgpu_shmem_destroy_now(&gpu->base, shmem); in virtgpu_shmem_destroy() 1347 struct virtgpu_shmem *shmem = in virtgpu_shmem_create() local 1349 *shmem = (struct virtgpu_shmem){ in virtgpu_shmem_create() 1359 return &shmem in virtgpu_shmem_create() 1311 virtgpu_shmem_destroy(struct vn_renderer *renderer, struct vn_renderer_shmem *shmem) virtgpu_shmem_destroy() argument [all...] |
H A D | vn_cs.h | 50 struct vn_renderer_shmem *shmem; member 57 struct vn_instance *instance; /* TODO shmem cache */ 68 /* the current buffer is buffers[buffer_count - 1].shmem */
|
/third_party/node/deps/v8/src/d8/ |
H A D | cov.cc | 27 struct shmem_data* shmem; variable 45 shmem = (struct shmem_data*)v8::base::Malloc(SHM_SIZE); in __sanitizer_cov_trace_pc_guard_init() 53 shmem = (struct shmem_data*)mmap(0, SHM_SIZE, PROT_READ | PROT_WRITE, in __sanitizer_cov_trace_pc_guard_init() 55 if (shmem == MAP_FAILED) { in __sanitizer_cov_trace_pc_guard_init() 65 shmem->num_edges = static_cast<uint32_t>(stop - start); in __sanitizer_cov_trace_pc_guard_init() 66 builtins_start = 1 + shmem->num_edges; in __sanitizer_cov_trace_pc_guard_init() 68 shm_key, shmem->num_edges); in __sanitizer_cov_trace_pc_guard_init() 75 if (shmem->edges[i / 8] & (1 << (i % 8))) { in sanitizer_cov_count_discovered_edges() 89 shmem->edges[index / 8] |= 1 << (index % 8); in __sanitizer_cov_trace_pc_guard() 94 if (num_edges + shmem in cov_init_builtins_edges() [all...] |
/third_party/ltp/testcases/kernel/containers/sysvipc/ |
H A D | shm_comm.c | 41 char *shmem; in chld1_shm() local 45 shmem = SAFE_SHMAT(id, NULL, 0); in chld1_shm() 46 *shmem = 'A'; in chld1_shm() 50 if (*shmem != 'A') in chld1_shm() 57 SAFE_SHMDT(shmem); in chld1_shm() 64 char *shmem; in chld2_shm() local 68 shmem = SAFE_SHMAT(id, NULL, 0); in chld2_shm() 72 *shmem = 'B'; in chld2_shm() 76 SAFE_SHMDT(shmem); in chld2_shm()
|
/third_party/mesa3d/src/freedreno/drm/virtio/ |
H A D | virtio_pipe.c | 211 if (unlikely(!virtio_dev->shmem)) { in init_shmem() 213 _FD_BO_VIRTIO_SHM, "shmem"); in init_shmem() 214 virtio_dev->shmem = fd_bo_map(virtio_dev->shmem_bo); in init_shmem() 217 uint32_t offset = virtio_dev->shmem->rsp_mem_offset; in init_shmem() 219 virtio_dev->rsp_mem = &((uint8_t *)virtio_dev->shmem)[offset]; in init_shmem()
|
H A D | msm_proto.h | 14 * 2) Response (rsp) messages are returned via shmem->rsp_mem, at an offset 23 * Defines the layout of shmem buffer used for host->guest communication. 32 * Offset to the start of rsp memory region in the shmem buffer. This 33 * is set by the host when the shmem buffer is allocated, to allow for 34 * extending the shmem buffer with new fields. The size of the rsp 35 * memory region is the size of the shmem buffer (controlled by the 47 #define msm_shmem_has_field(shmem, field) ({ \ 48 struct msm_shmem *_shmem = (shmem); \ 89 /* Offset into shmem ctrl buffer to write response. The host ensures
|
H A D | virtio_priv.h | 50 struct msm_shmem *shmem; member
|
H A D | virtio_device.c | 378 while (fd_fence_before(virtio_dev->shmem->seqno, req->seqno)) in virtio_host_sync()
|