1/* SPDX-License-Identifier: GPL-2.0 */
2/*
3 * Copyright (C) 2015-2018 Etnaviv Project
4 */
5
6#ifndef __ETNAVIV_GPU_H__
7#define __ETNAVIV_GPU_H__
8
9#include "etnaviv_cmdbuf.h"
10#include "etnaviv_gem.h"
11#include "etnaviv_mmu.h"
12#include "etnaviv_drv.h"
13#include "common.xml.h"
14
15struct etnaviv_gem_submit;
16struct etnaviv_vram_mapping;
17
18struct etnaviv_chip_identity {
19	u32 model;
20	u32 revision;
21	u32 product_id;
22	u32 customer_id;
23	u32 eco_id;
24
25	/* Supported feature fields. */
26	u32 features;
27
28	/* Supported minor feature fields. */
29	u32 minor_features0;
30	u32 minor_features1;
31	u32 minor_features2;
32	u32 minor_features3;
33	u32 minor_features4;
34	u32 minor_features5;
35	u32 minor_features6;
36	u32 minor_features7;
37	u32 minor_features8;
38	u32 minor_features9;
39	u32 minor_features10;
40	u32 minor_features11;
41
42	/* Number of streams supported. */
43	u32 stream_count;
44
45	/* Total number of temporary registers per thread. */
46	u32 register_max;
47
48	/* Maximum number of threads. */
49	u32 thread_count;
50
51	/* Number of shader cores. */
52	u32 shader_core_count;
53
54	/* Number of Neural Network cores. */
55	u32 nn_core_count;
56
57	/* Size of the vertex cache. */
58	u32 vertex_cache_size;
59
60	/* Number of entries in the vertex output buffer. */
61	u32 vertex_output_buffer_size;
62
63	/* Number of pixel pipes. */
64	u32 pixel_pipes;
65
66	/* Number of instructions. */
67	u32 instruction_count;
68
69	/* Number of constants. */
70	u32 num_constants;
71
72	/* Buffer size */
73	u32 buffer_size;
74
75	/* Number of varyings */
76	u8 varyings_count;
77};
78
79enum etnaviv_sec_mode {
80	ETNA_SEC_NONE = 0,
81	ETNA_SEC_KERNEL,
82	ETNA_SEC_TZ
83};
84
85struct etnaviv_event {
86	struct dma_fence *fence;
87	struct etnaviv_gem_submit *submit;
88
89	void (*sync_point)(struct etnaviv_gpu *gpu, struct etnaviv_event *event);
90};
91
92struct etnaviv_cmdbuf_suballoc;
93struct regulator;
94struct clk;
95
96#define ETNA_NR_EVENTS 30
97
98enum etnaviv_gpu_state {
99	ETNA_GPU_STATE_UNKNOWN = 0,
100	ETNA_GPU_STATE_IDENTIFIED,
101	ETNA_GPU_STATE_RESET,
102	ETNA_GPU_STATE_INITIALIZED,
103	ETNA_GPU_STATE_RUNNING,
104	ETNA_GPU_STATE_FAULT,
105};
106
107struct etnaviv_gpu {
108	struct drm_device *drm;
109	struct thermal_cooling_device *cooling;
110	struct device *dev;
111	struct mutex lock;
112	struct etnaviv_chip_identity identity;
113	enum etnaviv_sec_mode sec_mode;
114	struct workqueue_struct *wq;
115	struct mutex sched_lock;
116	struct drm_gpu_scheduler sched;
117	enum etnaviv_gpu_state state;
118
119	/* 'ring'-buffer: */
120	struct etnaviv_cmdbuf buffer;
121	int exec_state;
122
123	/* event management: */
124	DECLARE_BITMAP(event_bitmap, ETNA_NR_EVENTS);
125	struct etnaviv_event event[ETNA_NR_EVENTS];
126	struct completion event_free;
127	spinlock_t event_spinlock;
128
129	u32 idle_mask;
130
131	/* Fencing support */
132	struct xarray user_fences;
133	u32 next_user_fence;
134	u32 next_fence;
135	u32 completed_fence;
136	wait_queue_head_t fence_event;
137	u64 fence_context;
138	spinlock_t fence_spinlock;
139
140	/* worker for handling 'sync' points: */
141	struct work_struct sync_point_work;
142	int sync_point_event;
143
144	/* hang detection */
145	u32 hangcheck_dma_addr;
146	u32 hangcheck_fence;
147
148	void __iomem *mmio;
149	int irq;
150
151	struct etnaviv_iommu_context *mmu_context;
152	unsigned int flush_seq;
153
154	/* Power Control: */
155	struct clk *clk_bus;
156	struct clk *clk_reg;
157	struct clk *clk_core;
158	struct clk *clk_shader;
159
160	unsigned int freq_scale;
161	unsigned int fe_waitcycles;
162	unsigned long base_rate_core;
163	unsigned long base_rate_shader;
164};
165
166static inline void gpu_write(struct etnaviv_gpu *gpu, u32 reg, u32 data)
167{
168	writel(data, gpu->mmio + reg);
169}
170
171static inline u32 gpu_read(struct etnaviv_gpu *gpu, u32 reg)
172{
173	return readl(gpu->mmio + reg);
174}
175
176static inline u32 gpu_fix_power_address(struct etnaviv_gpu *gpu, u32 reg)
177{
178	/* Power registers in GC300 < 2.0 are offset by 0x100 */
179	if (gpu->identity.model == chipModel_GC300 &&
180	    gpu->identity.revision < 0x2000)
181		reg += 0x100;
182
183	return reg;
184}
185
186static inline void gpu_write_power(struct etnaviv_gpu *gpu, u32 reg, u32 data)
187{
188	writel(data, gpu->mmio + gpu_fix_power_address(gpu, reg));
189}
190
191static inline u32 gpu_read_power(struct etnaviv_gpu *gpu, u32 reg)
192{
193	return readl(gpu->mmio + gpu_fix_power_address(gpu, reg));
194}
195
196int etnaviv_gpu_get_param(struct etnaviv_gpu *gpu, u32 param, u64 *value);
197
198int etnaviv_gpu_init(struct etnaviv_gpu *gpu);
199bool etnaviv_fill_identity_from_hwdb(struct etnaviv_gpu *gpu);
200
201#ifdef CONFIG_DEBUG_FS
202int etnaviv_gpu_debugfs(struct etnaviv_gpu *gpu, struct seq_file *m);
203#endif
204
205void etnaviv_gpu_recover_hang(struct etnaviv_gem_submit *submit);
206void etnaviv_gpu_retire(struct etnaviv_gpu *gpu);
207int etnaviv_gpu_wait_fence_interruptible(struct etnaviv_gpu *gpu,
208	u32 fence, struct drm_etnaviv_timespec *timeout);
209int etnaviv_gpu_wait_obj_inactive(struct etnaviv_gpu *gpu,
210	struct etnaviv_gem_object *etnaviv_obj,
211	struct drm_etnaviv_timespec *timeout);
212struct dma_fence *etnaviv_gpu_submit(struct etnaviv_gem_submit *submit);
213int etnaviv_gpu_pm_get_sync(struct etnaviv_gpu *gpu);
214void etnaviv_gpu_pm_put(struct etnaviv_gpu *gpu);
215int etnaviv_gpu_wait_idle(struct etnaviv_gpu *gpu, unsigned int timeout_ms);
216void etnaviv_gpu_start_fe(struct etnaviv_gpu *gpu, u32 address, u16 prefetch);
217
218extern struct platform_driver etnaviv_gpu_driver;
219
220#endif /* __ETNAVIV_GPU_H__ */
221