blob: c9cbd8ae1f9533ff6be7a44c64bac0734092cc26 [file] [log] [blame]
Jerome Glisse771fe6b2009-06-05 14:42:42 +02001/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28#ifndef __RADEON_ASIC_H__
29#define __RADEON_ASIC_H__
30
31/*
32 * common functions
33 */
34void radeon_legacy_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
35void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable);
36
37void radeon_atom_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
38void radeon_atom_set_memory_clock(struct radeon_device *rdev, uint32_t mem_clock);
39void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable);
40
41/*
42 * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280
43 */
Jerome Glisse068a1172009-06-17 13:28:30 +020044int r100_init(struct radeon_device *rdev);
Dave Airlie551ebd82009-09-01 15:25:57 +100045int r200_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020046uint32_t r100_mm_rreg(struct radeon_device *rdev, uint32_t reg);
47void r100_mm_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
48void r100_errata(struct radeon_device *rdev);
49void r100_vram_info(struct radeon_device *rdev);
50int r100_gpu_reset(struct radeon_device *rdev);
51int r100_mc_init(struct radeon_device *rdev);
52void r100_mc_fini(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +020053u32 r100_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020054int r100_wb_init(struct radeon_device *rdev);
55void r100_wb_fini(struct radeon_device *rdev);
56int r100_gart_enable(struct radeon_device *rdev);
57void r100_pci_gart_disable(struct radeon_device *rdev);
58void r100_pci_gart_tlb_flush(struct radeon_device *rdev);
59int r100_pci_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
60int r100_cp_init(struct radeon_device *rdev, unsigned ring_size);
61void r100_cp_fini(struct radeon_device *rdev);
62void r100_cp_disable(struct radeon_device *rdev);
63void r100_ring_start(struct radeon_device *rdev);
64int r100_irq_set(struct radeon_device *rdev);
65int r100_irq_process(struct radeon_device *rdev);
66void r100_fence_ring_emit(struct radeon_device *rdev,
67 struct radeon_fence *fence);
68int r100_cs_parse(struct radeon_cs_parser *p);
69void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
70uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg);
71int r100_copy_blit(struct radeon_device *rdev,
72 uint64_t src_offset,
73 uint64_t dst_offset,
74 unsigned num_pages,
75 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +100076int r100_set_surface_reg(struct radeon_device *rdev, int reg,
77 uint32_t tiling_flags, uint32_t pitch,
78 uint32_t offset, uint32_t obj_size);
79int r100_clear_surface_reg(struct radeon_device *rdev, int reg);
Jerome Glissec93bb852009-07-13 21:04:08 +020080void r100_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020081
82static struct radeon_asic r100_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +020083 .init = &r100_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020084 .errata = &r100_errata,
85 .vram_info = &r100_vram_info,
86 .gpu_reset = &r100_gpu_reset,
87 .mc_init = &r100_mc_init,
88 .mc_fini = &r100_mc_fini,
89 .wb_init = &r100_wb_init,
90 .wb_fini = &r100_wb_fini,
91 .gart_enable = &r100_gart_enable,
92 .gart_disable = &r100_pci_gart_disable,
93 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
94 .gart_set_page = &r100_pci_gart_set_page,
95 .cp_init = &r100_cp_init,
96 .cp_fini = &r100_cp_fini,
97 .cp_disable = &r100_cp_disable,
98 .ring_start = &r100_ring_start,
99 .irq_set = &r100_irq_set,
100 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200101 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200102 .fence_ring_emit = &r100_fence_ring_emit,
103 .cs_parse = &r100_cs_parse,
104 .copy_blit = &r100_copy_blit,
105 .copy_dma = NULL,
106 .copy = &r100_copy_blit,
107 .set_engine_clock = &radeon_legacy_set_engine_clock,
108 .set_memory_clock = NULL,
109 .set_pcie_lanes = NULL,
110 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000111 .set_surface_reg = r100_set_surface_reg,
112 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200113 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200114};
115
116
117/*
118 * r300,r350,rv350,rv380
119 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200120int r300_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200121void r300_errata(struct radeon_device *rdev);
122void r300_vram_info(struct radeon_device *rdev);
123int r300_gpu_reset(struct radeon_device *rdev);
124int r300_mc_init(struct radeon_device *rdev);
125void r300_mc_fini(struct radeon_device *rdev);
126void r300_ring_start(struct radeon_device *rdev);
127void r300_fence_ring_emit(struct radeon_device *rdev,
128 struct radeon_fence *fence);
129int r300_cs_parse(struct radeon_cs_parser *p);
130int r300_gart_enable(struct radeon_device *rdev);
131void rv370_pcie_gart_disable(struct radeon_device *rdev);
132void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev);
133int rv370_pcie_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
134uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
135void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
136void rv370_set_pcie_lanes(struct radeon_device *rdev, int lanes);
137int r300_copy_dma(struct radeon_device *rdev,
138 uint64_t src_offset,
139 uint64_t dst_offset,
140 unsigned num_pages,
141 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +1000142
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200143static struct radeon_asic r300_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200144 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200145 .errata = &r300_errata,
146 .vram_info = &r300_vram_info,
147 .gpu_reset = &r300_gpu_reset,
148 .mc_init = &r300_mc_init,
149 .mc_fini = &r300_mc_fini,
150 .wb_init = &r100_wb_init,
151 .wb_fini = &r100_wb_fini,
152 .gart_enable = &r300_gart_enable,
153 .gart_disable = &r100_pci_gart_disable,
154 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
155 .gart_set_page = &r100_pci_gart_set_page,
156 .cp_init = &r100_cp_init,
157 .cp_fini = &r100_cp_fini,
158 .cp_disable = &r100_cp_disable,
159 .ring_start = &r300_ring_start,
160 .irq_set = &r100_irq_set,
161 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200162 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200163 .fence_ring_emit = &r300_fence_ring_emit,
164 .cs_parse = &r300_cs_parse,
165 .copy_blit = &r100_copy_blit,
166 .copy_dma = &r300_copy_dma,
167 .copy = &r100_copy_blit,
168 .set_engine_clock = &radeon_legacy_set_engine_clock,
169 .set_memory_clock = NULL,
170 .set_pcie_lanes = &rv370_set_pcie_lanes,
171 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000172 .set_surface_reg = r100_set_surface_reg,
173 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200174 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200175};
176
177/*
178 * r420,r423,rv410
179 */
180void r420_errata(struct radeon_device *rdev);
181void r420_vram_info(struct radeon_device *rdev);
182int r420_mc_init(struct radeon_device *rdev);
183void r420_mc_fini(struct radeon_device *rdev);
184static struct radeon_asic r420_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200185 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200186 .errata = &r420_errata,
187 .vram_info = &r420_vram_info,
188 .gpu_reset = &r300_gpu_reset,
189 .mc_init = &r420_mc_init,
190 .mc_fini = &r420_mc_fini,
191 .wb_init = &r100_wb_init,
192 .wb_fini = &r100_wb_fini,
193 .gart_enable = &r300_gart_enable,
194 .gart_disable = &rv370_pcie_gart_disable,
195 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
196 .gart_set_page = &rv370_pcie_gart_set_page,
197 .cp_init = &r100_cp_init,
198 .cp_fini = &r100_cp_fini,
199 .cp_disable = &r100_cp_disable,
200 .ring_start = &r300_ring_start,
201 .irq_set = &r100_irq_set,
202 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200203 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200204 .fence_ring_emit = &r300_fence_ring_emit,
205 .cs_parse = &r300_cs_parse,
206 .copy_blit = &r100_copy_blit,
207 .copy_dma = &r300_copy_dma,
208 .copy = &r100_copy_blit,
209 .set_engine_clock = &radeon_atom_set_engine_clock,
210 .set_memory_clock = &radeon_atom_set_memory_clock,
211 .set_pcie_lanes = &rv370_set_pcie_lanes,
212 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000213 .set_surface_reg = r100_set_surface_reg,
214 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200215 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200216};
217
218
219/*
220 * rs400,rs480
221 */
222void rs400_errata(struct radeon_device *rdev);
223void rs400_vram_info(struct radeon_device *rdev);
224int rs400_mc_init(struct radeon_device *rdev);
225void rs400_mc_fini(struct radeon_device *rdev);
226int rs400_gart_enable(struct radeon_device *rdev);
227void rs400_gart_disable(struct radeon_device *rdev);
228void rs400_gart_tlb_flush(struct radeon_device *rdev);
229int rs400_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
230uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg);
231void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
232static struct radeon_asic rs400_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200233 .init = &r300_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200234 .errata = &rs400_errata,
235 .vram_info = &rs400_vram_info,
236 .gpu_reset = &r300_gpu_reset,
237 .mc_init = &rs400_mc_init,
238 .mc_fini = &rs400_mc_fini,
239 .wb_init = &r100_wb_init,
240 .wb_fini = &r100_wb_fini,
241 .gart_enable = &rs400_gart_enable,
242 .gart_disable = &rs400_gart_disable,
243 .gart_tlb_flush = &rs400_gart_tlb_flush,
244 .gart_set_page = &rs400_gart_set_page,
245 .cp_init = &r100_cp_init,
246 .cp_fini = &r100_cp_fini,
247 .cp_disable = &r100_cp_disable,
248 .ring_start = &r300_ring_start,
249 .irq_set = &r100_irq_set,
250 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200251 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200252 .fence_ring_emit = &r300_fence_ring_emit,
253 .cs_parse = &r300_cs_parse,
254 .copy_blit = &r100_copy_blit,
255 .copy_dma = &r300_copy_dma,
256 .copy = &r100_copy_blit,
257 .set_engine_clock = &radeon_legacy_set_engine_clock,
258 .set_memory_clock = NULL,
259 .set_pcie_lanes = NULL,
260 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000261 .set_surface_reg = r100_set_surface_reg,
262 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200263 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200264};
265
266
267/*
268 * rs600.
269 */
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000270int rs600_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200271void rs600_errata(struct radeon_device *rdev);
272void rs600_vram_info(struct radeon_device *rdev);
273int rs600_mc_init(struct radeon_device *rdev);
274void rs600_mc_fini(struct radeon_device *rdev);
275int rs600_irq_set(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200276int rs600_irq_process(struct radeon_device *rdev);
277u32 rs600_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200278int rs600_gart_enable(struct radeon_device *rdev);
279void rs600_gart_disable(struct radeon_device *rdev);
280void rs600_gart_tlb_flush(struct radeon_device *rdev);
281int rs600_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
282uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg);
283void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200284void rs600_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200285static struct radeon_asic rs600_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000286 .init = &rs600_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200287 .errata = &rs600_errata,
288 .vram_info = &rs600_vram_info,
289 .gpu_reset = &r300_gpu_reset,
290 .mc_init = &rs600_mc_init,
291 .mc_fini = &rs600_mc_fini,
292 .wb_init = &r100_wb_init,
293 .wb_fini = &r100_wb_fini,
294 .gart_enable = &rs600_gart_enable,
295 .gart_disable = &rs600_gart_disable,
296 .gart_tlb_flush = &rs600_gart_tlb_flush,
297 .gart_set_page = &rs600_gart_set_page,
298 .cp_init = &r100_cp_init,
299 .cp_fini = &r100_cp_fini,
300 .cp_disable = &r100_cp_disable,
301 .ring_start = &r300_ring_start,
302 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200303 .irq_process = &rs600_irq_process,
304 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200305 .fence_ring_emit = &r300_fence_ring_emit,
306 .cs_parse = &r300_cs_parse,
307 .copy_blit = &r100_copy_blit,
308 .copy_dma = &r300_copy_dma,
309 .copy = &r100_copy_blit,
310 .set_engine_clock = &radeon_atom_set_engine_clock,
311 .set_memory_clock = &radeon_atom_set_memory_clock,
312 .set_pcie_lanes = NULL,
313 .set_clock_gating = &radeon_atom_set_clock_gating,
Jerome Glissec93bb852009-07-13 21:04:08 +0200314 .bandwidth_update = &rs600_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200315};
316
317
318/*
319 * rs690,rs740
320 */
321void rs690_errata(struct radeon_device *rdev);
322void rs690_vram_info(struct radeon_device *rdev);
323int rs690_mc_init(struct radeon_device *rdev);
324void rs690_mc_fini(struct radeon_device *rdev);
325uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg);
326void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200327void rs690_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200328static struct radeon_asic rs690_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000329 .init = &rs600_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200330 .errata = &rs690_errata,
331 .vram_info = &rs690_vram_info,
332 .gpu_reset = &r300_gpu_reset,
333 .mc_init = &rs690_mc_init,
334 .mc_fini = &rs690_mc_fini,
335 .wb_init = &r100_wb_init,
336 .wb_fini = &r100_wb_fini,
337 .gart_enable = &rs400_gart_enable,
338 .gart_disable = &rs400_gart_disable,
339 .gart_tlb_flush = &rs400_gart_tlb_flush,
340 .gart_set_page = &rs400_gart_set_page,
341 .cp_init = &r100_cp_init,
342 .cp_fini = &r100_cp_fini,
343 .cp_disable = &r100_cp_disable,
344 .ring_start = &r300_ring_start,
345 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200346 .irq_process = &rs600_irq_process,
347 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200348 .fence_ring_emit = &r300_fence_ring_emit,
349 .cs_parse = &r300_cs_parse,
350 .copy_blit = &r100_copy_blit,
351 .copy_dma = &r300_copy_dma,
352 .copy = &r300_copy_dma,
353 .set_engine_clock = &radeon_atom_set_engine_clock,
354 .set_memory_clock = &radeon_atom_set_memory_clock,
355 .set_pcie_lanes = NULL,
356 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000357 .set_surface_reg = r100_set_surface_reg,
358 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200359 .bandwidth_update = &rs690_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200360};
361
362
363/*
364 * rv515
365 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200366int rv515_init(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200367void rv515_errata(struct radeon_device *rdev);
368void rv515_vram_info(struct radeon_device *rdev);
369int rv515_gpu_reset(struct radeon_device *rdev);
370int rv515_mc_init(struct radeon_device *rdev);
371void rv515_mc_fini(struct radeon_device *rdev);
372uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg);
373void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
374void rv515_ring_start(struct radeon_device *rdev);
375uint32_t rv515_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
376void rv515_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200377void rv515_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200378static struct radeon_asic rv515_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200379 .init = &rv515_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200380 .errata = &rv515_errata,
381 .vram_info = &rv515_vram_info,
382 .gpu_reset = &rv515_gpu_reset,
383 .mc_init = &rv515_mc_init,
384 .mc_fini = &rv515_mc_fini,
385 .wb_init = &r100_wb_init,
386 .wb_fini = &r100_wb_fini,
387 .gart_enable = &r300_gart_enable,
388 .gart_disable = &rv370_pcie_gart_disable,
389 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
390 .gart_set_page = &rv370_pcie_gart_set_page,
391 .cp_init = &r100_cp_init,
392 .cp_fini = &r100_cp_fini,
393 .cp_disable = &r100_cp_disable,
394 .ring_start = &rv515_ring_start,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200395 .irq_set = &rs600_irq_set,
396 .irq_process = &rs600_irq_process,
397 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200398 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200399 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200400 .copy_blit = &r100_copy_blit,
401 .copy_dma = &r300_copy_dma,
402 .copy = &r100_copy_blit,
403 .set_engine_clock = &radeon_atom_set_engine_clock,
404 .set_memory_clock = &radeon_atom_set_memory_clock,
405 .set_pcie_lanes = &rv370_set_pcie_lanes,
406 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000407 .set_surface_reg = r100_set_surface_reg,
408 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200409 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200410};
411
412
413/*
414 * r520,rv530,rv560,rv570,r580
415 */
416void r520_errata(struct radeon_device *rdev);
417void r520_vram_info(struct radeon_device *rdev);
418int r520_mc_init(struct radeon_device *rdev);
419void r520_mc_fini(struct radeon_device *rdev);
Jerome Glissec93bb852009-07-13 21:04:08 +0200420void r520_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200421static struct radeon_asic r520_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200422 .init = &rv515_init,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200423 .errata = &r520_errata,
424 .vram_info = &r520_vram_info,
425 .gpu_reset = &rv515_gpu_reset,
426 .mc_init = &r520_mc_init,
427 .mc_fini = &r520_mc_fini,
428 .wb_init = &r100_wb_init,
429 .wb_fini = &r100_wb_fini,
430 .gart_enable = &r300_gart_enable,
431 .gart_disable = &rv370_pcie_gart_disable,
432 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
433 .gart_set_page = &rv370_pcie_gart_set_page,
434 .cp_init = &r100_cp_init,
435 .cp_fini = &r100_cp_fini,
436 .cp_disable = &r100_cp_disable,
437 .ring_start = &rv515_ring_start,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200438 .irq_set = &rs600_irq_set,
439 .irq_process = &rs600_irq_process,
440 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200441 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200442 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200443 .copy_blit = &r100_copy_blit,
444 .copy_dma = &r300_copy_dma,
445 .copy = &r100_copy_blit,
446 .set_engine_clock = &radeon_atom_set_engine_clock,
447 .set_memory_clock = &radeon_atom_set_memory_clock,
448 .set_pcie_lanes = &rv370_set_pcie_lanes,
449 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000450 .set_surface_reg = r100_set_surface_reg,
451 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200452 .bandwidth_update = &r520_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200453};
454
455/*
456 * r600,rv610,rv630,rv620,rv635,rv670,rs780,rv770,rv730,rv710
457 */
458uint32_t r600_pciep_rreg(struct radeon_device *rdev, uint32_t reg);
459void r600_pciep_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
460
461#endif