blob: e87bb915a6de4ca96fac655d2a5f5c88f6c89d67 [file] [log] [blame]
Jerome Glisse771fe6b2009-06-05 14:42:42 +02001/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28#ifndef __RADEON_ASIC_H__
29#define __RADEON_ASIC_H__
30
31/*
32 * common functions
33 */
34void radeon_legacy_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
35void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable);
36
37void radeon_atom_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
38void radeon_atom_set_memory_clock(struct radeon_device *rdev, uint32_t mem_clock);
39void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable);
40
41/*
42 * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280
43 */
Jerome Glisse068a1172009-06-17 13:28:30 +020044int r100_init(struct radeon_device *rdev);
Dave Airlie551ebd82009-09-01 15:25:57 +100045int r200_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020046uint32_t r100_mm_rreg(struct radeon_device *rdev, uint32_t reg);
47void r100_mm_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
48void r100_errata(struct radeon_device *rdev);
49void r100_vram_info(struct radeon_device *rdev);
50int r100_gpu_reset(struct radeon_device *rdev);
51int r100_mc_init(struct radeon_device *rdev);
52void r100_mc_fini(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +020053u32 r100_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020054int r100_wb_init(struct radeon_device *rdev);
55void r100_wb_fini(struct radeon_device *rdev);
56int r100_gart_enable(struct radeon_device *rdev);
57void r100_pci_gart_disable(struct radeon_device *rdev);
58void r100_pci_gart_tlb_flush(struct radeon_device *rdev);
59int r100_pci_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
60int r100_cp_init(struct radeon_device *rdev, unsigned ring_size);
61void r100_cp_fini(struct radeon_device *rdev);
62void r100_cp_disable(struct radeon_device *rdev);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100063void r100_cp_commit(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020064void r100_ring_start(struct radeon_device *rdev);
65int r100_irq_set(struct radeon_device *rdev);
66int r100_irq_process(struct radeon_device *rdev);
67void r100_fence_ring_emit(struct radeon_device *rdev,
68 struct radeon_fence *fence);
69int r100_cs_parse(struct radeon_cs_parser *p);
70void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
71uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg);
72int r100_copy_blit(struct radeon_device *rdev,
73 uint64_t src_offset,
74 uint64_t dst_offset,
75 unsigned num_pages,
76 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +100077int r100_set_surface_reg(struct radeon_device *rdev, int reg,
78 uint32_t tiling_flags, uint32_t pitch,
79 uint32_t offset, uint32_t obj_size);
80int r100_clear_surface_reg(struct radeon_device *rdev, int reg);
Jerome Glissec93bb852009-07-13 21:04:08 +020081void r100_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100082void r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
83int r100_ib_test(struct radeon_device *rdev);
84int r100_ring_test(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020085
86static struct radeon_asic r100_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +020087 .init = &r100_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020088 .errata = &r100_errata,
89 .vram_info = &r100_vram_info,
90 .gpu_reset = &r100_gpu_reset,
91 .mc_init = &r100_mc_init,
92 .mc_fini = &r100_mc_fini,
93 .wb_init = &r100_wb_init,
94 .wb_fini = &r100_wb_fini,
95 .gart_enable = &r100_gart_enable,
96 .gart_disable = &r100_pci_gart_disable,
97 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
98 .gart_set_page = &r100_pci_gart_set_page,
99 .cp_init = &r100_cp_init,
100 .cp_fini = &r100_cp_fini,
101 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000102 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200103 .ring_start = &r100_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000104 .ring_test = &r100_ring_test,
105 .ring_ib_execute = &r100_ring_ib_execute,
106 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200107 .irq_set = &r100_irq_set,
108 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200109 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200110 .fence_ring_emit = &r100_fence_ring_emit,
111 .cs_parse = &r100_cs_parse,
112 .copy_blit = &r100_copy_blit,
113 .copy_dma = NULL,
114 .copy = &r100_copy_blit,
115 .set_engine_clock = &radeon_legacy_set_engine_clock,
116 .set_memory_clock = NULL,
117 .set_pcie_lanes = NULL,
118 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000119 .set_surface_reg = r100_set_surface_reg,
120 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200121 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200122};
123
124
125/*
126 * r300,r350,rv350,rv380
127 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200128int r300_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200129void r300_errata(struct radeon_device *rdev);
130void r300_vram_info(struct radeon_device *rdev);
131int r300_gpu_reset(struct radeon_device *rdev);
132int r300_mc_init(struct radeon_device *rdev);
133void r300_mc_fini(struct radeon_device *rdev);
134void r300_ring_start(struct radeon_device *rdev);
135void r300_fence_ring_emit(struct radeon_device *rdev,
136 struct radeon_fence *fence);
137int r300_cs_parse(struct radeon_cs_parser *p);
138int r300_gart_enable(struct radeon_device *rdev);
139void rv370_pcie_gart_disable(struct radeon_device *rdev);
140void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev);
141int rv370_pcie_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
142uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
143void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
144void rv370_set_pcie_lanes(struct radeon_device *rdev, int lanes);
145int r300_copy_dma(struct radeon_device *rdev,
146 uint64_t src_offset,
147 uint64_t dst_offset,
148 unsigned num_pages,
149 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +1000150
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200151static struct radeon_asic r300_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200152 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200153 .errata = &r300_errata,
154 .vram_info = &r300_vram_info,
155 .gpu_reset = &r300_gpu_reset,
156 .mc_init = &r300_mc_init,
157 .mc_fini = &r300_mc_fini,
158 .wb_init = &r100_wb_init,
159 .wb_fini = &r100_wb_fini,
160 .gart_enable = &r300_gart_enable,
161 .gart_disable = &r100_pci_gart_disable,
162 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
163 .gart_set_page = &r100_pci_gart_set_page,
164 .cp_init = &r100_cp_init,
165 .cp_fini = &r100_cp_fini,
166 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000167 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200168 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000169 .ring_test = &r100_ring_test,
170 .ring_ib_execute = &r100_ring_ib_execute,
171 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200172 .irq_set = &r100_irq_set,
173 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200174 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200175 .fence_ring_emit = &r300_fence_ring_emit,
176 .cs_parse = &r300_cs_parse,
177 .copy_blit = &r100_copy_blit,
178 .copy_dma = &r300_copy_dma,
179 .copy = &r100_copy_blit,
180 .set_engine_clock = &radeon_legacy_set_engine_clock,
181 .set_memory_clock = NULL,
182 .set_pcie_lanes = &rv370_set_pcie_lanes,
183 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000184 .set_surface_reg = r100_set_surface_reg,
185 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200186 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200187};
188
189/*
190 * r420,r423,rv410
191 */
192void r420_errata(struct radeon_device *rdev);
193void r420_vram_info(struct radeon_device *rdev);
194int r420_mc_init(struct radeon_device *rdev);
195void r420_mc_fini(struct radeon_device *rdev);
196static struct radeon_asic r420_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200197 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200198 .errata = &r420_errata,
199 .vram_info = &r420_vram_info,
200 .gpu_reset = &r300_gpu_reset,
201 .mc_init = &r420_mc_init,
202 .mc_fini = &r420_mc_fini,
203 .wb_init = &r100_wb_init,
204 .wb_fini = &r100_wb_fini,
205 .gart_enable = &r300_gart_enable,
206 .gart_disable = &rv370_pcie_gart_disable,
207 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
208 .gart_set_page = &rv370_pcie_gart_set_page,
209 .cp_init = &r100_cp_init,
210 .cp_fini = &r100_cp_fini,
211 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000212 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200213 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000214 .ring_test = &r100_ring_test,
215 .ring_ib_execute = &r100_ring_ib_execute,
216 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200217 .irq_set = &r100_irq_set,
218 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200219 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200220 .fence_ring_emit = &r300_fence_ring_emit,
221 .cs_parse = &r300_cs_parse,
222 .copy_blit = &r100_copy_blit,
223 .copy_dma = &r300_copy_dma,
224 .copy = &r100_copy_blit,
225 .set_engine_clock = &radeon_atom_set_engine_clock,
226 .set_memory_clock = &radeon_atom_set_memory_clock,
227 .set_pcie_lanes = &rv370_set_pcie_lanes,
228 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000229 .set_surface_reg = r100_set_surface_reg,
230 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200231 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200232};
233
234
235/*
236 * rs400,rs480
237 */
238void rs400_errata(struct radeon_device *rdev);
239void rs400_vram_info(struct radeon_device *rdev);
240int rs400_mc_init(struct radeon_device *rdev);
241void rs400_mc_fini(struct radeon_device *rdev);
242int rs400_gart_enable(struct radeon_device *rdev);
243void rs400_gart_disable(struct radeon_device *rdev);
244void rs400_gart_tlb_flush(struct radeon_device *rdev);
245int rs400_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
246uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg);
247void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
248static struct radeon_asic rs400_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200249 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200250 .errata = &rs400_errata,
251 .vram_info = &rs400_vram_info,
252 .gpu_reset = &r300_gpu_reset,
253 .mc_init = &rs400_mc_init,
254 .mc_fini = &rs400_mc_fini,
255 .wb_init = &r100_wb_init,
256 .wb_fini = &r100_wb_fini,
257 .gart_enable = &rs400_gart_enable,
258 .gart_disable = &rs400_gart_disable,
259 .gart_tlb_flush = &rs400_gart_tlb_flush,
260 .gart_set_page = &rs400_gart_set_page,
261 .cp_init = &r100_cp_init,
262 .cp_fini = &r100_cp_fini,
263 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000264 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200265 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000266 .ring_test = &r100_ring_test,
267 .ring_ib_execute = &r100_ring_ib_execute,
268 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200269 .irq_set = &r100_irq_set,
270 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200271 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200272 .fence_ring_emit = &r300_fence_ring_emit,
273 .cs_parse = &r300_cs_parse,
274 .copy_blit = &r100_copy_blit,
275 .copy_dma = &r300_copy_dma,
276 .copy = &r100_copy_blit,
277 .set_engine_clock = &radeon_legacy_set_engine_clock,
278 .set_memory_clock = NULL,
279 .set_pcie_lanes = NULL,
280 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000281 .set_surface_reg = r100_set_surface_reg,
282 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200283 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200284};
285
286
287/*
288 * rs600.
289 */
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000290int rs600_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200291void rs600_errata(struct radeon_device *rdev);
292void rs600_vram_info(struct radeon_device *rdev);
293int rs600_mc_init(struct radeon_device *rdev);
294void rs600_mc_fini(struct radeon_device *rdev);
295int rs600_irq_set(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200296int rs600_irq_process(struct radeon_device *rdev);
297u32 rs600_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200298int rs600_gart_enable(struct radeon_device *rdev);
299void rs600_gart_disable(struct radeon_device *rdev);
300void rs600_gart_tlb_flush(struct radeon_device *rdev);
301int rs600_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
302uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg);
303void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200304void rs600_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200305static struct radeon_asic rs600_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000306 .init = &rs600_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200307 .errata = &rs600_errata,
308 .vram_info = &rs600_vram_info,
309 .gpu_reset = &r300_gpu_reset,
310 .mc_init = &rs600_mc_init,
311 .mc_fini = &rs600_mc_fini,
312 .wb_init = &r100_wb_init,
313 .wb_fini = &r100_wb_fini,
314 .gart_enable = &rs600_gart_enable,
315 .gart_disable = &rs600_gart_disable,
316 .gart_tlb_flush = &rs600_gart_tlb_flush,
317 .gart_set_page = &rs600_gart_set_page,
318 .cp_init = &r100_cp_init,
319 .cp_fini = &r100_cp_fini,
320 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000321 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200322 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000323 .ring_test = &r100_ring_test,
324 .ring_ib_execute = &r100_ring_ib_execute,
325 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200326 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200327 .irq_process = &rs600_irq_process,
328 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200329 .fence_ring_emit = &r300_fence_ring_emit,
330 .cs_parse = &r300_cs_parse,
331 .copy_blit = &r100_copy_blit,
332 .copy_dma = &r300_copy_dma,
333 .copy = &r100_copy_blit,
334 .set_engine_clock = &radeon_atom_set_engine_clock,
335 .set_memory_clock = &radeon_atom_set_memory_clock,
336 .set_pcie_lanes = NULL,
337 .set_clock_gating = &radeon_atom_set_clock_gating,
Jerome Glissec93bb852009-07-13 21:04:08 +0200338 .bandwidth_update = &rs600_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200339};
340
341
342/*
343 * rs690,rs740
344 */
345void rs690_errata(struct radeon_device *rdev);
346void rs690_vram_info(struct radeon_device *rdev);
347int rs690_mc_init(struct radeon_device *rdev);
348void rs690_mc_fini(struct radeon_device *rdev);
349uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg);
350void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200351void rs690_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200352static struct radeon_asic rs690_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000353 .init = &rs600_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200354 .errata = &rs690_errata,
355 .vram_info = &rs690_vram_info,
356 .gpu_reset = &r300_gpu_reset,
357 .mc_init = &rs690_mc_init,
358 .mc_fini = &rs690_mc_fini,
359 .wb_init = &r100_wb_init,
360 .wb_fini = &r100_wb_fini,
361 .gart_enable = &rs400_gart_enable,
362 .gart_disable = &rs400_gart_disable,
363 .gart_tlb_flush = &rs400_gart_tlb_flush,
364 .gart_set_page = &rs400_gart_set_page,
365 .cp_init = &r100_cp_init,
366 .cp_fini = &r100_cp_fini,
367 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000368 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200369 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000370 .ring_test = &r100_ring_test,
371 .ring_ib_execute = &r100_ring_ib_execute,
372 .ib_test = &r100_ib_test,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200373 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200374 .irq_process = &rs600_irq_process,
375 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200376 .fence_ring_emit = &r300_fence_ring_emit,
377 .cs_parse = &r300_cs_parse,
378 .copy_blit = &r100_copy_blit,
379 .copy_dma = &r300_copy_dma,
380 .copy = &r300_copy_dma,
381 .set_engine_clock = &radeon_atom_set_engine_clock,
382 .set_memory_clock = &radeon_atom_set_memory_clock,
383 .set_pcie_lanes = NULL,
384 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000385 .set_surface_reg = r100_set_surface_reg,
386 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200387 .bandwidth_update = &rs690_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200388};
389
390
391/*
392 * rv515
393 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200394int rv515_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200395void rv515_errata(struct radeon_device *rdev);
396void rv515_vram_info(struct radeon_device *rdev);
397int rv515_gpu_reset(struct radeon_device *rdev);
398int rv515_mc_init(struct radeon_device *rdev);
399void rv515_mc_fini(struct radeon_device *rdev);
400uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg);
401void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
402void rv515_ring_start(struct radeon_device *rdev);
403uint32_t rv515_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
404void rv515_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200405void rv515_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200406static struct radeon_asic rv515_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200407 .init = &rv515_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200408 .errata = &rv515_errata,
409 .vram_info = &rv515_vram_info,
410 .gpu_reset = &rv515_gpu_reset,
411 .mc_init = &rv515_mc_init,
412 .mc_fini = &rv515_mc_fini,
413 .wb_init = &r100_wb_init,
414 .wb_fini = &r100_wb_fini,
415 .gart_enable = &r300_gart_enable,
416 .gart_disable = &rv370_pcie_gart_disable,
417 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
418 .gart_set_page = &rv370_pcie_gart_set_page,
419 .cp_init = &r100_cp_init,
420 .cp_fini = &r100_cp_fini,
421 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000422 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200423 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000424 .ring_test = &r100_ring_test,
425 .ring_ib_execute = &r100_ring_ib_execute,
426 .ib_test = &r100_ib_test,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200427 .irq_set = &rs600_irq_set,
428 .irq_process = &rs600_irq_process,
429 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200430 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200431 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200432 .copy_blit = &r100_copy_blit,
433 .copy_dma = &r300_copy_dma,
434 .copy = &r100_copy_blit,
435 .set_engine_clock = &radeon_atom_set_engine_clock,
436 .set_memory_clock = &radeon_atom_set_memory_clock,
437 .set_pcie_lanes = &rv370_set_pcie_lanes,
438 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000439 .set_surface_reg = r100_set_surface_reg,
440 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200441 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200442};
443
444
445/*
446 * r520,rv530,rv560,rv570,r580
447 */
448void r520_errata(struct radeon_device *rdev);
449void r520_vram_info(struct radeon_device *rdev);
450int r520_mc_init(struct radeon_device *rdev);
451void r520_mc_fini(struct radeon_device *rdev);
Jerome Glissec93bb852009-07-13 21:04:08 +0200452void r520_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200453static struct radeon_asic r520_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200454 .init = &rv515_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200455 .errata = &r520_errata,
456 .vram_info = &r520_vram_info,
457 .gpu_reset = &rv515_gpu_reset,
458 .mc_init = &r520_mc_init,
459 .mc_fini = &r520_mc_fini,
460 .wb_init = &r100_wb_init,
461 .wb_fini = &r100_wb_fini,
462 .gart_enable = &r300_gart_enable,
463 .gart_disable = &rv370_pcie_gart_disable,
464 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
465 .gart_set_page = &rv370_pcie_gart_set_page,
466 .cp_init = &r100_cp_init,
467 .cp_fini = &r100_cp_fini,
468 .cp_disable = &r100_cp_disable,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000469 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200470 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000471 .ring_test = &r100_ring_test,
472 .ring_ib_execute = &r100_ring_ib_execute,
473 .ib_test = &r100_ib_test,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200474 .irq_set = &rs600_irq_set,
475 .irq_process = &rs600_irq_process,
476 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200477 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200478 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200479 .copy_blit = &r100_copy_blit,
480 .copy_dma = &r300_copy_dma,
481 .copy = &r100_copy_blit,
482 .set_engine_clock = &radeon_atom_set_engine_clock,
483 .set_memory_clock = &radeon_atom_set_memory_clock,
484 .set_pcie_lanes = &rv370_set_pcie_lanes,
485 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000486 .set_surface_reg = r100_set_surface_reg,
487 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200488 .bandwidth_update = &r520_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200489};
490
491/*
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000492 * r600,rv610,rv630,rv620,rv635,rv670,rs780,rs880
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200493 */
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000494int r600_init(struct radeon_device *rdev);
495void r600_fini(struct radeon_device *rdev);
496int r600_suspend(struct radeon_device *rdev);
497int r600_resume(struct radeon_device *rdev);
498int r600_wb_init(struct radeon_device *rdev);
499void r600_wb_fini(struct radeon_device *rdev);
500void r600_cp_commit(struct radeon_device *rdev);
501void r600_pcie_gart_tlb_flush(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200502uint32_t r600_pciep_rreg(struct radeon_device *rdev, uint32_t reg);
503void r600_pciep_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000504int r600_cs_parse(struct radeon_cs_parser *p);
505void r600_fence_ring_emit(struct radeon_device *rdev,
506 struct radeon_fence *fence);
507int r600_copy_dma(struct radeon_device *rdev,
508 uint64_t src_offset,
509 uint64_t dst_offset,
510 unsigned num_pages,
511 struct radeon_fence *fence);
512int r600_irq_process(struct radeon_device *rdev);
513int r600_irq_set(struct radeon_device *rdev);
514int r600_gpu_reset(struct radeon_device *rdev);
515int r600_set_surface_reg(struct radeon_device *rdev, int reg,
516 uint32_t tiling_flags, uint32_t pitch,
517 uint32_t offset, uint32_t obj_size);
518int r600_clear_surface_reg(struct radeon_device *rdev, int reg);
519void r600_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
520int r600_ib_test(struct radeon_device *rdev);
521int r600_ring_test(struct radeon_device *rdev);
522int r600_copy_blit(struct radeon_device *rdev,
523 uint64_t src_offset, uint64_t dst_offset,
524 unsigned num_pages, struct radeon_fence *fence);
525
526static struct radeon_asic r600_asic = {
527 .errata = NULL,
528 .init = &r600_init,
529 .fini = &r600_fini,
530 .suspend = &r600_suspend,
531 .resume = &r600_resume,
532 .cp_commit = &r600_cp_commit,
533 .vram_info = NULL,
534 .gpu_reset = &r600_gpu_reset,
535 .mc_init = NULL,
536 .mc_fini = NULL,
537 .wb_init = &r600_wb_init,
538 .wb_fini = &r600_wb_fini,
539 .gart_enable = NULL,
540 .gart_disable = NULL,
541 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
542 .gart_set_page = &rs600_gart_set_page,
543 .cp_init = NULL,
544 .cp_fini = NULL,
545 .cp_disable = NULL,
546 .ring_start = NULL,
547 .ring_test = &r600_ring_test,
548 .ring_ib_execute = &r600_ring_ib_execute,
549 .ib_test = &r600_ib_test,
550 .irq_set = &r600_irq_set,
551 .irq_process = &r600_irq_process,
552 .fence_ring_emit = &r600_fence_ring_emit,
553 .cs_parse = &r600_cs_parse,
554 .copy_blit = &r600_copy_blit,
555 .copy_dma = &r600_copy_blit,
556 .copy = NULL,
557 .set_engine_clock = &radeon_atom_set_engine_clock,
558 .set_memory_clock = &radeon_atom_set_memory_clock,
559 .set_pcie_lanes = NULL,
560 .set_clock_gating = &radeon_atom_set_clock_gating,
561 .set_surface_reg = r600_set_surface_reg,
562 .clear_surface_reg = r600_clear_surface_reg,
563 .bandwidth_update = &r520_bandwidth_update,
564};
565
566/*
567 * rv770,rv730,rv710,rv740
568 */
569int rv770_init(struct radeon_device *rdev);
570void rv770_fini(struct radeon_device *rdev);
571int rv770_suspend(struct radeon_device *rdev);
572int rv770_resume(struct radeon_device *rdev);
573int rv770_gpu_reset(struct radeon_device *rdev);
574
575static struct radeon_asic rv770_asic = {
576 .errata = NULL,
577 .init = &rv770_init,
578 .fini = &rv770_fini,
579 .suspend = &rv770_suspend,
580 .resume = &rv770_resume,
581 .cp_commit = &r600_cp_commit,
582 .vram_info = NULL,
583 .gpu_reset = &rv770_gpu_reset,
584 .mc_init = NULL,
585 .mc_fini = NULL,
586 .wb_init = &r600_wb_init,
587 .wb_fini = &r600_wb_fini,
588 .gart_enable = NULL,
589 .gart_disable = NULL,
590 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
591 .gart_set_page = &rs600_gart_set_page,
592 .cp_init = NULL,
593 .cp_fini = NULL,
594 .cp_disable = NULL,
595 .ring_start = NULL,
596 .ring_test = &r600_ring_test,
597 .ring_ib_execute = &r600_ring_ib_execute,
598 .ib_test = &r600_ib_test,
599 .irq_set = &r600_irq_set,
600 .irq_process = &r600_irq_process,
601 .fence_ring_emit = &r600_fence_ring_emit,
602 .cs_parse = &r600_cs_parse,
603 .copy_blit = &r600_copy_blit,
604 .copy_dma = &r600_copy_blit,
605 .copy = NULL,
606 .set_engine_clock = &radeon_atom_set_engine_clock,
607 .set_memory_clock = &radeon_atom_set_memory_clock,
608 .set_pcie_lanes = NULL,
609 .set_clock_gating = &radeon_atom_set_clock_gating,
610 .set_surface_reg = r600_set_surface_reg,
611 .clear_surface_reg = r600_clear_surface_reg,
612 .bandwidth_update = &r520_bandwidth_update,
613};
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200614
615#endif