blob: d38f996328270e88ced365c50aedc895c538ce17 [file] [log] [blame]
Jerome Glisse771fe6b2009-06-05 14:42:42 +02001/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28#ifndef __RADEON_ASIC_H__
29#define __RADEON_ASIC_H__
30
31/*
32 * common functions
33 */
34void radeon_legacy_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
35void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable);
36
37void radeon_atom_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
38void radeon_atom_set_memory_clock(struct radeon_device *rdev, uint32_t mem_clock);
39void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable);
40
41/*
42 * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280
43 */
Jerome Glissed4550902009-10-01 10:12:06 +020044extern int r100_init(struct radeon_device *rdev);
45extern void r100_fini(struct radeon_device *rdev);
46extern int r100_suspend(struct radeon_device *rdev);
47extern int r100_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020048uint32_t r100_mm_rreg(struct radeon_device *rdev, uint32_t reg);
49void r100_mm_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020050int r100_gpu_reset(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +020051u32 r100_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020052void r100_pci_gart_tlb_flush(struct radeon_device *rdev);
53int r100_pci_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100054void r100_cp_commit(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020055void r100_ring_start(struct radeon_device *rdev);
56int r100_irq_set(struct radeon_device *rdev);
57int r100_irq_process(struct radeon_device *rdev);
58void r100_fence_ring_emit(struct radeon_device *rdev,
59 struct radeon_fence *fence);
60int r100_cs_parse(struct radeon_cs_parser *p);
61void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
62uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg);
63int r100_copy_blit(struct radeon_device *rdev,
64 uint64_t src_offset,
65 uint64_t dst_offset,
66 unsigned num_pages,
67 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +100068int r100_set_surface_reg(struct radeon_device *rdev, int reg,
69 uint32_t tiling_flags, uint32_t pitch,
70 uint32_t offset, uint32_t obj_size);
71int r100_clear_surface_reg(struct radeon_device *rdev, int reg);
Jerome Glissec93bb852009-07-13 21:04:08 +020072void r100_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100073void r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100074int r100_ring_test(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020075
76static struct radeon_asic r100_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +020077 .init = &r100_init,
Jerome Glissed4550902009-10-01 10:12:06 +020078 .fini = &r100_fini,
79 .suspend = &r100_suspend,
80 .resume = &r100_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020081 .gpu_reset = &r100_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020082 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
83 .gart_set_page = &r100_pci_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +100084 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020085 .ring_start = &r100_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +100086 .ring_test = &r100_ring_test,
87 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020088 .irq_set = &r100_irq_set,
89 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +020090 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020091 .fence_ring_emit = &r100_fence_ring_emit,
92 .cs_parse = &r100_cs_parse,
93 .copy_blit = &r100_copy_blit,
94 .copy_dma = NULL,
95 .copy = &r100_copy_blit,
96 .set_engine_clock = &radeon_legacy_set_engine_clock,
97 .set_memory_clock = NULL,
98 .set_pcie_lanes = NULL,
99 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000100 .set_surface_reg = r100_set_surface_reg,
101 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200102 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200103};
104
105
106/*
107 * r300,r350,rv350,rv380
108 */
Jerome Glisse207bf9e2009-09-30 15:35:32 +0200109extern int r300_init(struct radeon_device *rdev);
110extern void r300_fini(struct radeon_device *rdev);
111extern int r300_suspend(struct radeon_device *rdev);
112extern int r300_resume(struct radeon_device *rdev);
113extern int r300_gpu_reset(struct radeon_device *rdev);
114extern void r300_ring_start(struct radeon_device *rdev);
115extern void r300_fence_ring_emit(struct radeon_device *rdev,
116 struct radeon_fence *fence);
117extern int r300_cs_parse(struct radeon_cs_parser *p);
118extern void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev);
119extern int rv370_pcie_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
120extern uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
121extern void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
122extern void rv370_set_pcie_lanes(struct radeon_device *rdev, int lanes);
123extern int r300_copy_dma(struct radeon_device *rdev,
124 uint64_t src_offset,
125 uint64_t dst_offset,
126 unsigned num_pages,
127 struct radeon_fence *fence);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200128static struct radeon_asic r300_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200129 .init = &r300_init,
Jerome Glisse207bf9e2009-09-30 15:35:32 +0200130 .fini = &r300_fini,
131 .suspend = &r300_suspend,
132 .resume = &r300_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200133 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200134 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
135 .gart_set_page = &r100_pci_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000136 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200137 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000138 .ring_test = &r100_ring_test,
139 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200140 .irq_set = &r100_irq_set,
141 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200142 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200143 .fence_ring_emit = &r300_fence_ring_emit,
144 .cs_parse = &r300_cs_parse,
145 .copy_blit = &r100_copy_blit,
146 .copy_dma = &r300_copy_dma,
147 .copy = &r100_copy_blit,
148 .set_engine_clock = &radeon_legacy_set_engine_clock,
149 .set_memory_clock = NULL,
150 .set_pcie_lanes = &rv370_set_pcie_lanes,
151 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000152 .set_surface_reg = r100_set_surface_reg,
153 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200154 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200155};
156
157/*
158 * r420,r423,rv410
159 */
Jerome Glisse9f022dd2009-09-11 15:35:22 +0200160extern int r420_init(struct radeon_device *rdev);
161extern void r420_fini(struct radeon_device *rdev);
162extern int r420_suspend(struct radeon_device *rdev);
163extern int r420_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200164static struct radeon_asic r420_asic = {
Jerome Glisse9f022dd2009-09-11 15:35:22 +0200165 .init = &r420_init,
166 .fini = &r420_fini,
167 .suspend = &r420_suspend,
168 .resume = &r420_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200169 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200170 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
171 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000172 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200173 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000174 .ring_test = &r100_ring_test,
175 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200176 .irq_set = &r100_irq_set,
177 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200178 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200179 .fence_ring_emit = &r300_fence_ring_emit,
180 .cs_parse = &r300_cs_parse,
181 .copy_blit = &r100_copy_blit,
182 .copy_dma = &r300_copy_dma,
183 .copy = &r100_copy_blit,
184 .set_engine_clock = &radeon_atom_set_engine_clock,
185 .set_memory_clock = &radeon_atom_set_memory_clock,
186 .set_pcie_lanes = &rv370_set_pcie_lanes,
187 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000188 .set_surface_reg = r100_set_surface_reg,
189 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200190 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200191};
192
193
194/*
195 * rs400,rs480
196 */
Jerome Glisseca6ffc62009-10-01 10:20:52 +0200197extern int rs400_init(struct radeon_device *rdev);
198extern void rs400_fini(struct radeon_device *rdev);
199extern int rs400_suspend(struct radeon_device *rdev);
200extern int rs400_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200201void rs400_gart_tlb_flush(struct radeon_device *rdev);
202int rs400_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
203uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg);
204void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
205static struct radeon_asic rs400_asic = {
Jerome Glisseca6ffc62009-10-01 10:20:52 +0200206 .init = &rs400_init,
207 .fini = &rs400_fini,
208 .suspend = &rs400_suspend,
209 .resume = &rs400_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200210 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200211 .gart_tlb_flush = &rs400_gart_tlb_flush,
212 .gart_set_page = &rs400_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000213 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200214 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000215 .ring_test = &r100_ring_test,
216 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200217 .irq_set = &r100_irq_set,
218 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200219 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200220 .fence_ring_emit = &r300_fence_ring_emit,
221 .cs_parse = &r300_cs_parse,
222 .copy_blit = &r100_copy_blit,
223 .copy_dma = &r300_copy_dma,
224 .copy = &r100_copy_blit,
225 .set_engine_clock = &radeon_legacy_set_engine_clock,
226 .set_memory_clock = NULL,
227 .set_pcie_lanes = NULL,
228 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000229 .set_surface_reg = r100_set_surface_reg,
230 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200231 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200232};
233
234
235/*
236 * rs600.
237 */
Jerome Glissec010f802009-09-30 22:09:06 +0200238extern int rs600_init(struct radeon_device *rdev);
239extern void rs600_fini(struct radeon_device *rdev);
240extern int rs600_suspend(struct radeon_device *rdev);
241extern int rs600_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200242int rs600_irq_set(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200243int rs600_irq_process(struct radeon_device *rdev);
244u32 rs600_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200245void rs600_gart_tlb_flush(struct radeon_device *rdev);
246int rs600_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
247uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg);
248void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200249void rs600_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200250static struct radeon_asic rs600_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000251 .init = &rs600_init,
Jerome Glissec010f802009-09-30 22:09:06 +0200252 .fini = &rs600_fini,
253 .suspend = &rs600_suspend,
254 .resume = &rs600_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200255 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200256 .gart_tlb_flush = &rs600_gart_tlb_flush,
257 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000258 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200259 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000260 .ring_test = &r100_ring_test,
261 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200262 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200263 .irq_process = &rs600_irq_process,
264 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200265 .fence_ring_emit = &r300_fence_ring_emit,
266 .cs_parse = &r300_cs_parse,
267 .copy_blit = &r100_copy_blit,
268 .copy_dma = &r300_copy_dma,
269 .copy = &r100_copy_blit,
270 .set_engine_clock = &radeon_atom_set_engine_clock,
271 .set_memory_clock = &radeon_atom_set_memory_clock,
272 .set_pcie_lanes = NULL,
273 .set_clock_gating = &radeon_atom_set_clock_gating,
Jerome Glissec93bb852009-07-13 21:04:08 +0200274 .bandwidth_update = &rs600_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200275};
276
277
278/*
279 * rs690,rs740
280 */
Jerome Glisse3bc68532009-10-01 09:39:24 +0200281int rs690_init(struct radeon_device *rdev);
282void rs690_fini(struct radeon_device *rdev);
283int rs690_resume(struct radeon_device *rdev);
284int rs690_suspend(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200285uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg);
286void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200287void rs690_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200288static struct radeon_asic rs690_asic = {
Jerome Glisse3bc68532009-10-01 09:39:24 +0200289 .init = &rs690_init,
290 .fini = &rs690_fini,
291 .suspend = &rs690_suspend,
292 .resume = &rs690_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200293 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200294 .gart_tlb_flush = &rs400_gart_tlb_flush,
295 .gart_set_page = &rs400_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000296 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200297 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000298 .ring_test = &r100_ring_test,
299 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200300 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200301 .irq_process = &rs600_irq_process,
302 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200303 .fence_ring_emit = &r300_fence_ring_emit,
304 .cs_parse = &r300_cs_parse,
305 .copy_blit = &r100_copy_blit,
306 .copy_dma = &r300_copy_dma,
307 .copy = &r300_copy_dma,
308 .set_engine_clock = &radeon_atom_set_engine_clock,
309 .set_memory_clock = &radeon_atom_set_memory_clock,
310 .set_pcie_lanes = NULL,
311 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000312 .set_surface_reg = r100_set_surface_reg,
313 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200314 .bandwidth_update = &rs690_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200315};
316
317
318/*
319 * rv515
320 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200321int rv515_init(struct radeon_device *rdev);
Jerome Glissed39c3b82009-09-28 18:34:43 +0200322void rv515_fini(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200323int rv515_gpu_reset(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200324uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg);
325void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
326void rv515_ring_start(struct radeon_device *rdev);
327uint32_t rv515_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
328void rv515_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200329void rv515_bandwidth_update(struct radeon_device *rdev);
Jerome Glissed39c3b82009-09-28 18:34:43 +0200330int rv515_resume(struct radeon_device *rdev);
331int rv515_suspend(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200332static struct radeon_asic rv515_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200333 .init = &rv515_init,
Jerome Glissed39c3b82009-09-28 18:34:43 +0200334 .fini = &rv515_fini,
335 .suspend = &rv515_suspend,
336 .resume = &rv515_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200337 .gpu_reset = &rv515_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200338 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
339 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000340 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200341 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000342 .ring_test = &r100_ring_test,
343 .ring_ib_execute = &r100_ring_ib_execute,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200344 .irq_set = &rs600_irq_set,
345 .irq_process = &rs600_irq_process,
346 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200347 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200348 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200349 .copy_blit = &r100_copy_blit,
350 .copy_dma = &r300_copy_dma,
351 .copy = &r100_copy_blit,
352 .set_engine_clock = &radeon_atom_set_engine_clock,
353 .set_memory_clock = &radeon_atom_set_memory_clock,
354 .set_pcie_lanes = &rv370_set_pcie_lanes,
355 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000356 .set_surface_reg = r100_set_surface_reg,
357 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200358 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200359};
360
361
362/*
363 * r520,rv530,rv560,rv570,r580
364 */
Jerome Glissed39c3b82009-09-28 18:34:43 +0200365int r520_init(struct radeon_device *rdev);
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200366int r520_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200367static struct radeon_asic r520_asic = {
Jerome Glissed39c3b82009-09-28 18:34:43 +0200368 .init = &r520_init,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200369 .fini = &rv515_fini,
370 .suspend = &rv515_suspend,
371 .resume = &r520_resume,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200372 .gpu_reset = &rv515_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200373 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
374 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000375 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200376 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000377 .ring_test = &r100_ring_test,
378 .ring_ib_execute = &r100_ring_ib_execute,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200379 .irq_set = &rs600_irq_set,
380 .irq_process = &rs600_irq_process,
381 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200382 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200383 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200384 .copy_blit = &r100_copy_blit,
385 .copy_dma = &r300_copy_dma,
386 .copy = &r100_copy_blit,
387 .set_engine_clock = &radeon_atom_set_engine_clock,
388 .set_memory_clock = &radeon_atom_set_memory_clock,
389 .set_pcie_lanes = &rv370_set_pcie_lanes,
390 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000391 .set_surface_reg = r100_set_surface_reg,
392 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200393 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200394};
395
396/*
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000397 * r600,rv610,rv630,rv620,rv635,rv670,rs780,rs880
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200398 */
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000399int r600_init(struct radeon_device *rdev);
400void r600_fini(struct radeon_device *rdev);
401int r600_suspend(struct radeon_device *rdev);
402int r600_resume(struct radeon_device *rdev);
403int r600_wb_init(struct radeon_device *rdev);
404void r600_wb_fini(struct radeon_device *rdev);
405void r600_cp_commit(struct radeon_device *rdev);
406void r600_pcie_gart_tlb_flush(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200407uint32_t r600_pciep_rreg(struct radeon_device *rdev, uint32_t reg);
408void r600_pciep_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000409int r600_cs_parse(struct radeon_cs_parser *p);
410void r600_fence_ring_emit(struct radeon_device *rdev,
411 struct radeon_fence *fence);
412int r600_copy_dma(struct radeon_device *rdev,
413 uint64_t src_offset,
414 uint64_t dst_offset,
415 unsigned num_pages,
416 struct radeon_fence *fence);
417int r600_irq_process(struct radeon_device *rdev);
418int r600_irq_set(struct radeon_device *rdev);
419int r600_gpu_reset(struct radeon_device *rdev);
420int r600_set_surface_reg(struct radeon_device *rdev, int reg,
421 uint32_t tiling_flags, uint32_t pitch,
422 uint32_t offset, uint32_t obj_size);
423int r600_clear_surface_reg(struct radeon_device *rdev, int reg);
424void r600_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000425int r600_ring_test(struct radeon_device *rdev);
426int r600_copy_blit(struct radeon_device *rdev,
427 uint64_t src_offset, uint64_t dst_offset,
428 unsigned num_pages, struct radeon_fence *fence);
429
430static struct radeon_asic r600_asic = {
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000431 .init = &r600_init,
432 .fini = &r600_fini,
433 .suspend = &r600_suspend,
434 .resume = &r600_resume,
435 .cp_commit = &r600_cp_commit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000436 .gpu_reset = &r600_gpu_reset,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000437 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
438 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000439 .ring_test = &r600_ring_test,
440 .ring_ib_execute = &r600_ring_ib_execute,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000441 .irq_set = &r600_irq_set,
442 .irq_process = &r600_irq_process,
443 .fence_ring_emit = &r600_fence_ring_emit,
444 .cs_parse = &r600_cs_parse,
445 .copy_blit = &r600_copy_blit,
446 .copy_dma = &r600_copy_blit,
Alex Deuchera3812872009-09-10 15:54:35 -0400447 .copy = &r600_copy_blit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000448 .set_engine_clock = &radeon_atom_set_engine_clock,
449 .set_memory_clock = &radeon_atom_set_memory_clock,
450 .set_pcie_lanes = NULL,
451 .set_clock_gating = &radeon_atom_set_clock_gating,
452 .set_surface_reg = r600_set_surface_reg,
453 .clear_surface_reg = r600_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200454 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000455};
456
457/*
458 * rv770,rv730,rv710,rv740
459 */
460int rv770_init(struct radeon_device *rdev);
461void rv770_fini(struct radeon_device *rdev);
462int rv770_suspend(struct radeon_device *rdev);
463int rv770_resume(struct radeon_device *rdev);
464int rv770_gpu_reset(struct radeon_device *rdev);
465
466static struct radeon_asic rv770_asic = {
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000467 .init = &rv770_init,
468 .fini = &rv770_fini,
469 .suspend = &rv770_suspend,
470 .resume = &rv770_resume,
471 .cp_commit = &r600_cp_commit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000472 .gpu_reset = &rv770_gpu_reset,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000473 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
474 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000475 .ring_test = &r600_ring_test,
476 .ring_ib_execute = &r600_ring_ib_execute,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000477 .irq_set = &r600_irq_set,
478 .irq_process = &r600_irq_process,
479 .fence_ring_emit = &r600_fence_ring_emit,
480 .cs_parse = &r600_cs_parse,
481 .copy_blit = &r600_copy_blit,
482 .copy_dma = &r600_copy_blit,
Alex Deuchera3812872009-09-10 15:54:35 -0400483 .copy = &r600_copy_blit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000484 .set_engine_clock = &radeon_atom_set_engine_clock,
485 .set_memory_clock = &radeon_atom_set_memory_clock,
486 .set_pcie_lanes = NULL,
487 .set_clock_gating = &radeon_atom_set_clock_gating,
488 .set_surface_reg = r600_set_surface_reg,
489 .clear_surface_reg = r600_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200490 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000491};
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200492
493#endif