blob: c3532c7a6f3f04f74d26d6ef97ece8fea7da053e [file] [log] [blame]
Jerome Glisse771fe6b2009-06-05 14:42:42 +02001/*
2 * Copyright 2008 Advanced Micro Devices, Inc.
3 * Copyright 2008 Red Hat Inc.
4 * Copyright 2009 Jerome Glisse.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
20 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
21 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
22 * OTHER DEALINGS IN THE SOFTWARE.
23 *
24 * Authors: Dave Airlie
25 * Alex Deucher
26 * Jerome Glisse
27 */
28#ifndef __RADEON_ASIC_H__
29#define __RADEON_ASIC_H__
30
31/*
32 * common functions
33 */
34void radeon_legacy_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
35void radeon_legacy_set_clock_gating(struct radeon_device *rdev, int enable);
36
37void radeon_atom_set_engine_clock(struct radeon_device *rdev, uint32_t eng_clock);
38void radeon_atom_set_memory_clock(struct radeon_device *rdev, uint32_t mem_clock);
39void radeon_atom_set_clock_gating(struct radeon_device *rdev, int enable);
40
41/*
42 * r100,rv100,rs100,rv200,rs200,r200,rv250,rs300,rv280
43 */
Jerome Glissed4550902009-10-01 10:12:06 +020044extern int r100_init(struct radeon_device *rdev);
45extern void r100_fini(struct radeon_device *rdev);
46extern int r100_suspend(struct radeon_device *rdev);
47extern int r100_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020048uint32_t r100_mm_rreg(struct radeon_device *rdev, uint32_t reg);
49void r100_mm_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Dave Airlie28d52042009-09-21 14:33:58 +100050void r100_vga_set_state(struct radeon_device *rdev, bool state);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020051int r100_gpu_reset(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +020052u32 r100_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020053void r100_pci_gart_tlb_flush(struct radeon_device *rdev);
54int r100_pci_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100055void r100_cp_commit(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020056void r100_ring_start(struct radeon_device *rdev);
57int r100_irq_set(struct radeon_device *rdev);
58int r100_irq_process(struct radeon_device *rdev);
59void r100_fence_ring_emit(struct radeon_device *rdev,
60 struct radeon_fence *fence);
61int r100_cs_parse(struct radeon_cs_parser *p);
62void r100_pll_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
63uint32_t r100_pll_rreg(struct radeon_device *rdev, uint32_t reg);
64int r100_copy_blit(struct radeon_device *rdev,
65 uint64_t src_offset,
66 uint64_t dst_offset,
67 unsigned num_pages,
68 struct radeon_fence *fence);
Dave Airliee024e112009-06-24 09:48:08 +100069int r100_set_surface_reg(struct radeon_device *rdev, int reg,
70 uint32_t tiling_flags, uint32_t pitch,
71 uint32_t offset, uint32_t obj_size);
72int r100_clear_surface_reg(struct radeon_device *rdev, int reg);
Jerome Glissec93bb852009-07-13 21:04:08 +020073void r100_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100074void r100_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
Jerome Glisse3ce0a232009-09-08 10:10:24 +100075int r100_ring_test(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +020076
77static struct radeon_asic r100_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +020078 .init = &r100_init,
Jerome Glissed4550902009-10-01 10:12:06 +020079 .fini = &r100_fini,
80 .suspend = &r100_suspend,
81 .resume = &r100_resume,
Dave Airlie28d52042009-09-21 14:33:58 +100082 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020083 .gpu_reset = &r100_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020084 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
85 .gart_set_page = &r100_pci_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +100086 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020087 .ring_start = &r100_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +100088 .ring_test = &r100_ring_test,
89 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020090 .irq_set = &r100_irq_set,
91 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +020092 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +020093 .fence_ring_emit = &r100_fence_ring_emit,
94 .cs_parse = &r100_cs_parse,
95 .copy_blit = &r100_copy_blit,
96 .copy_dma = NULL,
97 .copy = &r100_copy_blit,
98 .set_engine_clock = &radeon_legacy_set_engine_clock,
99 .set_memory_clock = NULL,
100 .set_pcie_lanes = NULL,
101 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000102 .set_surface_reg = r100_set_surface_reg,
103 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200104 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200105};
106
107
108/*
109 * r300,r350,rv350,rv380
110 */
Jerome Glisse207bf9e2009-09-30 15:35:32 +0200111extern int r300_init(struct radeon_device *rdev);
112extern void r300_fini(struct radeon_device *rdev);
113extern int r300_suspend(struct radeon_device *rdev);
114extern int r300_resume(struct radeon_device *rdev);
115extern int r300_gpu_reset(struct radeon_device *rdev);
116extern void r300_ring_start(struct radeon_device *rdev);
117extern void r300_fence_ring_emit(struct radeon_device *rdev,
118 struct radeon_fence *fence);
119extern int r300_cs_parse(struct radeon_cs_parser *p);
120extern void rv370_pcie_gart_tlb_flush(struct radeon_device *rdev);
121extern int rv370_pcie_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
122extern uint32_t rv370_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
123extern void rv370_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
124extern void rv370_set_pcie_lanes(struct radeon_device *rdev, int lanes);
125extern int r300_copy_dma(struct radeon_device *rdev,
126 uint64_t src_offset,
127 uint64_t dst_offset,
128 unsigned num_pages,
129 struct radeon_fence *fence);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200130static struct radeon_asic r300_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200131 .init = &r300_init,
Jerome Glisse207bf9e2009-09-30 15:35:32 +0200132 .fini = &r300_fini,
133 .suspend = &r300_suspend,
134 .resume = &r300_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000135 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200136 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200137 .gart_tlb_flush = &r100_pci_gart_tlb_flush,
138 .gart_set_page = &r100_pci_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000139 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200140 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000141 .ring_test = &r100_ring_test,
142 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200143 .irq_set = &r100_irq_set,
144 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200145 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200146 .fence_ring_emit = &r300_fence_ring_emit,
147 .cs_parse = &r300_cs_parse,
148 .copy_blit = &r100_copy_blit,
149 .copy_dma = &r300_copy_dma,
150 .copy = &r100_copy_blit,
151 .set_engine_clock = &radeon_legacy_set_engine_clock,
152 .set_memory_clock = NULL,
153 .set_pcie_lanes = &rv370_set_pcie_lanes,
154 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000155 .set_surface_reg = r100_set_surface_reg,
156 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200157 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200158};
159
160/*
161 * r420,r423,rv410
162 */
Jerome Glisse9f022dd2009-09-11 15:35:22 +0200163extern int r420_init(struct radeon_device *rdev);
164extern void r420_fini(struct radeon_device *rdev);
165extern int r420_suspend(struct radeon_device *rdev);
166extern int r420_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200167static struct radeon_asic r420_asic = {
Jerome Glisse9f022dd2009-09-11 15:35:22 +0200168 .init = &r420_init,
169 .fini = &r420_fini,
170 .suspend = &r420_suspend,
171 .resume = &r420_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000172 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200173 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200174 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
175 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000176 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200177 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000178 .ring_test = &r100_ring_test,
179 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200180 .irq_set = &r100_irq_set,
181 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200182 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200183 .fence_ring_emit = &r300_fence_ring_emit,
184 .cs_parse = &r300_cs_parse,
185 .copy_blit = &r100_copy_blit,
186 .copy_dma = &r300_copy_dma,
187 .copy = &r100_copy_blit,
188 .set_engine_clock = &radeon_atom_set_engine_clock,
189 .set_memory_clock = &radeon_atom_set_memory_clock,
190 .set_pcie_lanes = &rv370_set_pcie_lanes,
191 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000192 .set_surface_reg = r100_set_surface_reg,
193 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200194 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200195};
196
197
198/*
199 * rs400,rs480
200 */
Jerome Glisseca6ffc62009-10-01 10:20:52 +0200201extern int rs400_init(struct radeon_device *rdev);
202extern void rs400_fini(struct radeon_device *rdev);
203extern int rs400_suspend(struct radeon_device *rdev);
204extern int rs400_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200205void rs400_gart_tlb_flush(struct radeon_device *rdev);
206int rs400_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
207uint32_t rs400_mc_rreg(struct radeon_device *rdev, uint32_t reg);
208void rs400_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
209static struct radeon_asic rs400_asic = {
Jerome Glisseca6ffc62009-10-01 10:20:52 +0200210 .init = &rs400_init,
211 .fini = &rs400_fini,
212 .suspend = &rs400_suspend,
213 .resume = &rs400_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000214 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200215 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200216 .gart_tlb_flush = &rs400_gart_tlb_flush,
217 .gart_set_page = &rs400_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000218 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200219 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000220 .ring_test = &r100_ring_test,
221 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200222 .irq_set = &r100_irq_set,
223 .irq_process = &r100_irq_process,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200224 .get_vblank_counter = &r100_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200225 .fence_ring_emit = &r300_fence_ring_emit,
226 .cs_parse = &r300_cs_parse,
227 .copy_blit = &r100_copy_blit,
228 .copy_dma = &r300_copy_dma,
229 .copy = &r100_copy_blit,
230 .set_engine_clock = &radeon_legacy_set_engine_clock,
231 .set_memory_clock = NULL,
232 .set_pcie_lanes = NULL,
233 .set_clock_gating = &radeon_legacy_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000234 .set_surface_reg = r100_set_surface_reg,
235 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200236 .bandwidth_update = &r100_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200237};
238
239
240/*
241 * rs600.
242 */
Jerome Glissec010f802009-09-30 22:09:06 +0200243extern int rs600_init(struct radeon_device *rdev);
244extern void rs600_fini(struct radeon_device *rdev);
245extern int rs600_suspend(struct radeon_device *rdev);
246extern int rs600_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200247int rs600_irq_set(struct radeon_device *rdev);
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200248int rs600_irq_process(struct radeon_device *rdev);
249u32 rs600_get_vblank_counter(struct radeon_device *rdev, int crtc);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200250void rs600_gart_tlb_flush(struct radeon_device *rdev);
251int rs600_gart_set_page(struct radeon_device *rdev, int i, uint64_t addr);
252uint32_t rs600_mc_rreg(struct radeon_device *rdev, uint32_t reg);
253void rs600_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200254void rs600_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200255static struct radeon_asic rs600_asic = {
Dave Airlie3f7dc91a2009-08-27 11:10:15 +1000256 .init = &rs600_init,
Jerome Glissec010f802009-09-30 22:09:06 +0200257 .fini = &rs600_fini,
258 .suspend = &rs600_suspend,
259 .resume = &rs600_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000260 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200261 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200262 .gart_tlb_flush = &rs600_gart_tlb_flush,
263 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000264 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200265 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000266 .ring_test = &r100_ring_test,
267 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200268 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200269 .irq_process = &rs600_irq_process,
270 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200271 .fence_ring_emit = &r300_fence_ring_emit,
272 .cs_parse = &r300_cs_parse,
273 .copy_blit = &r100_copy_blit,
274 .copy_dma = &r300_copy_dma,
275 .copy = &r100_copy_blit,
276 .set_engine_clock = &radeon_atom_set_engine_clock,
277 .set_memory_clock = &radeon_atom_set_memory_clock,
278 .set_pcie_lanes = NULL,
279 .set_clock_gating = &radeon_atom_set_clock_gating,
Jerome Glissec93bb852009-07-13 21:04:08 +0200280 .bandwidth_update = &rs600_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200281};
282
283
284/*
285 * rs690,rs740
286 */
Jerome Glisse3bc68532009-10-01 09:39:24 +0200287int rs690_init(struct radeon_device *rdev);
288void rs690_fini(struct radeon_device *rdev);
289int rs690_resume(struct radeon_device *rdev);
290int rs690_suspend(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200291uint32_t rs690_mc_rreg(struct radeon_device *rdev, uint32_t reg);
292void rs690_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200293void rs690_bandwidth_update(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200294static struct radeon_asic rs690_asic = {
Jerome Glisse3bc68532009-10-01 09:39:24 +0200295 .init = &rs690_init,
296 .fini = &rs690_fini,
297 .suspend = &rs690_suspend,
298 .resume = &rs690_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000299 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200300 .gpu_reset = &r300_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200301 .gart_tlb_flush = &rs400_gart_tlb_flush,
302 .gart_set_page = &rs400_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000303 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200304 .ring_start = &r300_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000305 .ring_test = &r100_ring_test,
306 .ring_ib_execute = &r100_ring_ib_execute,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200307 .irq_set = &rs600_irq_set,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200308 .irq_process = &rs600_irq_process,
309 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200310 .fence_ring_emit = &r300_fence_ring_emit,
311 .cs_parse = &r300_cs_parse,
312 .copy_blit = &r100_copy_blit,
313 .copy_dma = &r300_copy_dma,
314 .copy = &r300_copy_dma,
315 .set_engine_clock = &radeon_atom_set_engine_clock,
316 .set_memory_clock = &radeon_atom_set_memory_clock,
317 .set_pcie_lanes = NULL,
318 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000319 .set_surface_reg = r100_set_surface_reg,
320 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200321 .bandwidth_update = &rs690_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200322};
323
324
325/*
326 * rv515
327 */
Jerome Glisse068a1172009-06-17 13:28:30 +0200328int rv515_init(struct radeon_device *rdev);
Jerome Glissed39c3b82009-09-28 18:34:43 +0200329void rv515_fini(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200330int rv515_gpu_reset(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200331uint32_t rv515_mc_rreg(struct radeon_device *rdev, uint32_t reg);
332void rv515_mc_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
333void rv515_ring_start(struct radeon_device *rdev);
334uint32_t rv515_pcie_rreg(struct radeon_device *rdev, uint32_t reg);
335void rv515_pcie_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glissec93bb852009-07-13 21:04:08 +0200336void rv515_bandwidth_update(struct radeon_device *rdev);
Jerome Glissed39c3b82009-09-28 18:34:43 +0200337int rv515_resume(struct radeon_device *rdev);
338int rv515_suspend(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200339static struct radeon_asic rv515_asic = {
Jerome Glisse068a1172009-06-17 13:28:30 +0200340 .init = &rv515_init,
Jerome Glissed39c3b82009-09-28 18:34:43 +0200341 .fini = &rv515_fini,
342 .suspend = &rv515_suspend,
343 .resume = &rv515_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000344 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200345 .gpu_reset = &rv515_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200346 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
347 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000348 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200349 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000350 .ring_test = &r100_ring_test,
351 .ring_ib_execute = &r100_ring_ib_execute,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200352 .irq_set = &rs600_irq_set,
353 .irq_process = &rs600_irq_process,
354 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200355 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200356 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200357 .copy_blit = &r100_copy_blit,
358 .copy_dma = &r300_copy_dma,
359 .copy = &r100_copy_blit,
360 .set_engine_clock = &radeon_atom_set_engine_clock,
361 .set_memory_clock = &radeon_atom_set_memory_clock,
362 .set_pcie_lanes = &rv370_set_pcie_lanes,
363 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000364 .set_surface_reg = r100_set_surface_reg,
365 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissec93bb852009-07-13 21:04:08 +0200366 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200367};
368
369
370/*
371 * r520,rv530,rv560,rv570,r580
372 */
Jerome Glissed39c3b82009-09-28 18:34:43 +0200373int r520_init(struct radeon_device *rdev);
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200374int r520_resume(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200375static struct radeon_asic r520_asic = {
Jerome Glissed39c3b82009-09-28 18:34:43 +0200376 .init = &r520_init,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200377 .fini = &rv515_fini,
378 .suspend = &rv515_suspend,
379 .resume = &r520_resume,
Dave Airlie28d52042009-09-21 14:33:58 +1000380 .vga_set_state = &r100_vga_set_state,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200381 .gpu_reset = &rv515_gpu_reset,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200382 .gart_tlb_flush = &rv370_pcie_gart_tlb_flush,
383 .gart_set_page = &rv370_pcie_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000384 .cp_commit = &r100_cp_commit,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200385 .ring_start = &rv515_ring_start,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000386 .ring_test = &r100_ring_test,
387 .ring_ib_execute = &r100_ring_ib_execute,
Michel Dänzer7ed220d2009-08-13 11:10:51 +0200388 .irq_set = &rs600_irq_set,
389 .irq_process = &rs600_irq_process,
390 .get_vblank_counter = &rs600_get_vblank_counter,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200391 .fence_ring_emit = &r300_fence_ring_emit,
Jerome Glisse068a1172009-06-17 13:28:30 +0200392 .cs_parse = &r300_cs_parse,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200393 .copy_blit = &r100_copy_blit,
394 .copy_dma = &r300_copy_dma,
395 .copy = &r100_copy_blit,
396 .set_engine_clock = &radeon_atom_set_engine_clock,
397 .set_memory_clock = &radeon_atom_set_memory_clock,
398 .set_pcie_lanes = &rv370_set_pcie_lanes,
399 .set_clock_gating = &radeon_atom_set_clock_gating,
Dave Airliee024e112009-06-24 09:48:08 +1000400 .set_surface_reg = r100_set_surface_reg,
401 .clear_surface_reg = r100_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200402 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200403};
404
405/*
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000406 * r600,rv610,rv630,rv620,rv635,rv670,rs780,rs880
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200407 */
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000408int r600_init(struct radeon_device *rdev);
409void r600_fini(struct radeon_device *rdev);
410int r600_suspend(struct radeon_device *rdev);
411int r600_resume(struct radeon_device *rdev);
Dave Airlie28d52042009-09-21 14:33:58 +1000412void r600_vga_set_state(struct radeon_device *rdev, bool state);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000413int r600_wb_init(struct radeon_device *rdev);
414void r600_wb_fini(struct radeon_device *rdev);
415void r600_cp_commit(struct radeon_device *rdev);
416void r600_pcie_gart_tlb_flush(struct radeon_device *rdev);
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200417uint32_t r600_pciep_rreg(struct radeon_device *rdev, uint32_t reg);
418void r600_pciep_wreg(struct radeon_device *rdev, uint32_t reg, uint32_t v);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000419int r600_cs_parse(struct radeon_cs_parser *p);
420void r600_fence_ring_emit(struct radeon_device *rdev,
421 struct radeon_fence *fence);
422int r600_copy_dma(struct radeon_device *rdev,
423 uint64_t src_offset,
424 uint64_t dst_offset,
425 unsigned num_pages,
426 struct radeon_fence *fence);
427int r600_irq_process(struct radeon_device *rdev);
428int r600_irq_set(struct radeon_device *rdev);
429int r600_gpu_reset(struct radeon_device *rdev);
430int r600_set_surface_reg(struct radeon_device *rdev, int reg,
431 uint32_t tiling_flags, uint32_t pitch,
432 uint32_t offset, uint32_t obj_size);
433int r600_clear_surface_reg(struct radeon_device *rdev, int reg);
434void r600_ring_ib_execute(struct radeon_device *rdev, struct radeon_ib *ib);
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000435int r600_ring_test(struct radeon_device *rdev);
436int r600_copy_blit(struct radeon_device *rdev,
437 uint64_t src_offset, uint64_t dst_offset,
438 unsigned num_pages, struct radeon_fence *fence);
439
440static struct radeon_asic r600_asic = {
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000441 .init = &r600_init,
442 .fini = &r600_fini,
443 .suspend = &r600_suspend,
444 .resume = &r600_resume,
445 .cp_commit = &r600_cp_commit,
Dave Airlie28d52042009-09-21 14:33:58 +1000446 .vga_set_state = &r600_vga_set_state,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000447 .gpu_reset = &r600_gpu_reset,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000448 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
449 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000450 .ring_test = &r600_ring_test,
451 .ring_ib_execute = &r600_ring_ib_execute,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000452 .irq_set = &r600_irq_set,
453 .irq_process = &r600_irq_process,
454 .fence_ring_emit = &r600_fence_ring_emit,
455 .cs_parse = &r600_cs_parse,
456 .copy_blit = &r600_copy_blit,
457 .copy_dma = &r600_copy_blit,
Alex Deuchera3812872009-09-10 15:54:35 -0400458 .copy = &r600_copy_blit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000459 .set_engine_clock = &radeon_atom_set_engine_clock,
460 .set_memory_clock = &radeon_atom_set_memory_clock,
461 .set_pcie_lanes = NULL,
462 .set_clock_gating = &radeon_atom_set_clock_gating,
463 .set_surface_reg = r600_set_surface_reg,
464 .clear_surface_reg = r600_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200465 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000466};
467
468/*
469 * rv770,rv730,rv710,rv740
470 */
471int rv770_init(struct radeon_device *rdev);
472void rv770_fini(struct radeon_device *rdev);
473int rv770_suspend(struct radeon_device *rdev);
474int rv770_resume(struct radeon_device *rdev);
475int rv770_gpu_reset(struct radeon_device *rdev);
476
477static struct radeon_asic rv770_asic = {
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000478 .init = &rv770_init,
479 .fini = &rv770_fini,
480 .suspend = &rv770_suspend,
481 .resume = &rv770_resume,
482 .cp_commit = &r600_cp_commit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000483 .gpu_reset = &rv770_gpu_reset,
Dave Airlie28d52042009-09-21 14:33:58 +1000484 .vga_set_state = &r600_vga_set_state,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000485 .gart_tlb_flush = &r600_pcie_gart_tlb_flush,
486 .gart_set_page = &rs600_gart_set_page,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000487 .ring_test = &r600_ring_test,
488 .ring_ib_execute = &r600_ring_ib_execute,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000489 .irq_set = &r600_irq_set,
490 .irq_process = &r600_irq_process,
491 .fence_ring_emit = &r600_fence_ring_emit,
492 .cs_parse = &r600_cs_parse,
493 .copy_blit = &r600_copy_blit,
494 .copy_dma = &r600_copy_blit,
Alex Deuchera3812872009-09-10 15:54:35 -0400495 .copy = &r600_copy_blit,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000496 .set_engine_clock = &radeon_atom_set_engine_clock,
497 .set_memory_clock = &radeon_atom_set_memory_clock,
498 .set_pcie_lanes = NULL,
499 .set_clock_gating = &radeon_atom_set_clock_gating,
500 .set_surface_reg = r600_set_surface_reg,
501 .clear_surface_reg = r600_clear_surface_reg,
Jerome Glissef0ed1f62009-09-28 20:39:19 +0200502 .bandwidth_update = &rv515_bandwidth_update,
Jerome Glisse3ce0a232009-09-08 10:10:24 +1000503};
Jerome Glisse771fe6b2009-06-05 14:42:42 +0200504
505#endif