1/*
2 * Copyright © 2021 Google, Inc.
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
20 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
21 * SOFTWARE.
22 */
23
24#ifdef X
25#undef X
26#endif
27
28#if PTRSZ == 32
29#define X(n) n##_32
30#else
31#define X(n) n##_64
32#endif
33
34static void X(emit_reloc_common)(struct fd_ringbuffer *ring,
35                                 const struct fd_reloc *reloc)
36{
37   (*ring->cur++) = (uint32_t)reloc->iova;
38#if PTRSZ == 64
39   (*ring->cur++) = (uint32_t)(reloc->iova >> 32);
40#endif
41}
42
43static void X(fd_ringbuffer_sp_emit_reloc_nonobj)(struct fd_ringbuffer *ring,
44                                                  const struct fd_reloc *reloc)
45{
46   X(emit_reloc_common)(ring, reloc);
47
48   assert(!(ring->flags & _FD_RINGBUFFER_OBJECT));
49
50   struct fd_ringbuffer_sp *fd_ring = to_fd_ringbuffer_sp(ring);
51
52   struct fd_submit_sp *fd_submit = to_fd_submit_sp(fd_ring->u.submit);
53
54   fd_submit_append_bo(fd_submit, reloc->bo);
55}
56
57static void X(fd_ringbuffer_sp_emit_reloc_obj)(struct fd_ringbuffer *ring,
58                                               const struct fd_reloc *reloc)
59{
60   X(emit_reloc_common)(ring, reloc);
61
62   assert(ring->flags & _FD_RINGBUFFER_OBJECT);
63
64   struct fd_ringbuffer_sp *fd_ring = to_fd_ringbuffer_sp(ring);
65
66   /* Avoid emitting duplicate BO references into the list.  Ringbuffer
67    * objects are long-lived, so this saves ongoing work at draw time in
68    * exchange for a bit at context setup/first draw.  And the number of
69    * relocs per ringbuffer object is fairly small, so the O(n^2) doesn't
70    * hurt much.
71    */
72   if (!fd_ringbuffer_references_bo(ring, reloc->bo)) {
73      APPEND(&fd_ring->u, reloc_bos, fd_bo_ref(reloc->bo));
74   }
75}
76
77static uint32_t X(fd_ringbuffer_sp_emit_reloc_ring)(
78   struct fd_ringbuffer *ring, struct fd_ringbuffer *target, uint32_t cmd_idx)
79{
80   struct fd_ringbuffer_sp *fd_target = to_fd_ringbuffer_sp(target);
81   struct fd_bo *bo;
82   uint32_t size;
83
84   if ((target->flags & FD_RINGBUFFER_GROWABLE) &&
85       (cmd_idx < fd_target->u.nr_cmds)) {
86      bo = fd_target->u.cmds[cmd_idx].ring_bo;
87      size = fd_target->u.cmds[cmd_idx].size;
88   } else {
89      bo = fd_target->ring_bo;
90      size = offset_bytes(target->cur, target->start);
91   }
92
93   if (ring->flags & _FD_RINGBUFFER_OBJECT) {
94      X(fd_ringbuffer_sp_emit_reloc_obj)(ring, &(struct fd_reloc){
95                .bo = bo,
96                .iova = bo->iova + fd_target->offset,
97                .offset = fd_target->offset,
98             });
99   } else {
100      X(fd_ringbuffer_sp_emit_reloc_nonobj)(ring, &(struct fd_reloc){
101                .bo = bo,
102                .iova = bo->iova + fd_target->offset,
103                .offset = fd_target->offset,
104             });
105   }
106
107   if (!(target->flags & _FD_RINGBUFFER_OBJECT))
108      return size;
109
110   struct fd_ringbuffer_sp *fd_ring = to_fd_ringbuffer_sp(ring);
111
112   if (ring->flags & _FD_RINGBUFFER_OBJECT) {
113      for (unsigned i = 0; i < fd_target->u.nr_reloc_bos; i++) {
114         struct fd_bo *target_bo = fd_target->u.reloc_bos[i];
115         if (!fd_ringbuffer_references_bo(ring, target_bo))
116            APPEND(&fd_ring->u, reloc_bos, fd_bo_ref(target_bo));
117      }
118   } else {
119      // TODO it would be nice to know whether we have already
120      // seen this target before.  But hopefully we hit the
121      // append_bo() fast path enough for this to not matter:
122      struct fd_submit_sp *fd_submit = to_fd_submit_sp(fd_ring->u.submit);
123
124      for (unsigned i = 0; i < fd_target->u.nr_reloc_bos; i++) {
125         fd_submit_append_bo(fd_submit, fd_target->u.reloc_bos[i]);
126      }
127   }
128
129   return size;
130}
131