blob: a027b11bbdb38b5040d679982762959c03df97d2 [file] [log] [blame]
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +02001/******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2008 - 2014 Intel Corporation. All rights reserved.
9 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
Golan Ben-Ami5b086412016-02-09 12:57:16 +020010 * Copyright(c) 2015 - 2016 Intel Deutschland GmbH
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +020011 *
12 * This program is free software; you can redistribute it and/or modify
13 * it under the terms of version 2 of the GNU General Public License as
14 * published by the Free Software Foundation.
15 *
16 * This program is distributed in the hope that it will be useful, but
17 * WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
19 * General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program;
23 *
24 * The full GNU General Public License is included in this distribution
25 * in the file called COPYING.
26 *
27 * Contact Information:
Emmanuel Grumbachd01c5362015-11-17 15:39:56 +020028 * Intel Linux Wireless <linuxwifi@intel.com>
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +020029 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
30 *
31 * BSD LICENSE
32 *
33 * Copyright(c) 2005 - 2014 Intel Corporation. All rights reserved.
34 * Copyright(c) 2013 - 2015 Intel Mobile Communications GmbH
Golan Ben-Ami5b086412016-02-09 12:57:16 +020035 * Copyright(c) 2015 - 2016 Intel Deutschland GmbH
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +020036 * All rights reserved.
37 *
38 * Redistribution and use in source and binary forms, with or without
39 * modification, are permitted provided that the following conditions
40 * are met:
41 *
42 * * Redistributions of source code must retain the above copyright
43 * notice, this list of conditions and the following disclaimer.
44 * * Redistributions in binary form must reproduce the above copyright
45 * notice, this list of conditions and the following disclaimer in
46 * the documentation and/or other materials provided with the
47 * distribution.
48 * * Neither the name Intel Corporation nor the names of its
49 * contributors may be used to endorse or promote products derived
50 * from this software without specific prior written permission.
51 *
52 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
53 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
54 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
55 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
56 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
57 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
58 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
59 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
60 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
61 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
62 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
63 *
64 *****************************************************************************/
65#include <linux/devcoredump.h>
66
67#include "fw-dbg.h"
68#include "iwl-io.h"
69#include "mvm.h"
70#include "iwl-prph.h"
71#include "iwl-csr.h"
72
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +020073#define RADIO_REG_MAX_READ 0x2ad
74static void iwl_mvm_read_radio_reg(struct iwl_mvm *mvm,
75 struct iwl_fw_error_dump_data **dump_data)
76{
77 u8 *pos = (void *)(*dump_data)->data;
78 unsigned long flags;
79 int i;
80
81 if (!iwl_trans_grab_nic_access(mvm->trans, &flags))
82 return;
83
84 (*dump_data)->type = cpu_to_le32(IWL_FW_ERROR_DUMP_RADIO_REG);
85 (*dump_data)->len = cpu_to_le32(RADIO_REG_MAX_READ);
86
87 for (i = 0; i < RADIO_REG_MAX_READ; i++) {
88 u32 rd_cmd = RADIO_RSP_RD_CMD;
89
90 rd_cmd |= i << RADIO_RSP_ADDR_POS;
91 iwl_write_prph_no_grab(mvm->trans, RSP_RADIO_CMD, rd_cmd);
92 *pos = (u8)iwl_read_prph_no_grab(mvm->trans, RSP_RADIO_RDDAT);
93
94 pos++;
95 }
96
97 *dump_data = iwl_fw_error_next_data(*dump_data);
98
99 iwl_trans_release_nic_access(mvm->trans, &flags);
100}
101
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200102static void iwl_mvm_dump_fifos(struct iwl_mvm *mvm,
103 struct iwl_fw_error_dump_data **dump_data)
104{
105 struct iwl_fw_error_dump_fifo *fifo_hdr;
106 u32 *fifo_data;
107 u32 fifo_len;
108 unsigned long flags;
109 int i, j;
110
Emmanuel Grumbach23ba9342015-12-17 11:55:13 +0200111 if (!iwl_trans_grab_nic_access(mvm->trans, &flags))
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200112 return;
113
114 /* Pull RXF data from all RXFs */
115 for (i = 0; i < ARRAY_SIZE(mvm->shared_mem_cfg.rxfifo_size); i++) {
116 /*
117 * Keep aside the additional offset that might be needed for
118 * next RXF
119 */
120 u32 offset_diff = RXF_DIFF_FROM_PREV * i;
121
122 fifo_hdr = (void *)(*dump_data)->data;
123 fifo_data = (void *)fifo_hdr->data;
124 fifo_len = mvm->shared_mem_cfg.rxfifo_size[i];
125
126 /* No need to try to read the data if the length is 0 */
127 if (fifo_len == 0)
128 continue;
129
130 /* Add a TLV for the RXF */
131 (*dump_data)->type = cpu_to_le32(IWL_FW_ERROR_DUMP_RXF);
132 (*dump_data)->len = cpu_to_le32(fifo_len + sizeof(*fifo_hdr));
133
134 fifo_hdr->fifo_num = cpu_to_le32(i);
135 fifo_hdr->available_bytes =
136 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
137 RXF_RD_D_SPACE +
138 offset_diff));
139 fifo_hdr->wr_ptr =
140 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
141 RXF_RD_WR_PTR +
142 offset_diff));
143 fifo_hdr->rd_ptr =
144 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
145 RXF_RD_RD_PTR +
146 offset_diff));
147 fifo_hdr->fence_ptr =
148 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
149 RXF_RD_FENCE_PTR +
150 offset_diff));
151 fifo_hdr->fence_mode =
152 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
153 RXF_SET_FENCE_MODE +
154 offset_diff));
155
156 /* Lock fence */
157 iwl_trans_write_prph(mvm->trans,
158 RXF_SET_FENCE_MODE + offset_diff, 0x1);
159 /* Set fence pointer to the same place like WR pointer */
160 iwl_trans_write_prph(mvm->trans,
161 RXF_LD_WR2FENCE + offset_diff, 0x1);
162 /* Set fence offset */
163 iwl_trans_write_prph(mvm->trans,
164 RXF_LD_FENCE_OFFSET_ADDR + offset_diff,
165 0x0);
166
167 /* Read FIFO */
168 fifo_len /= sizeof(u32); /* Size in DWORDS */
169 for (j = 0; j < fifo_len; j++)
170 fifo_data[j] = iwl_trans_read_prph(mvm->trans,
171 RXF_FIFO_RD_FENCE_INC +
172 offset_diff);
173 *dump_data = iwl_fw_error_next_data(*dump_data);
174 }
175
176 /* Pull TXF data from all TXFs */
177 for (i = 0; i < ARRAY_SIZE(mvm->shared_mem_cfg.txfifo_size); i++) {
178 /* Mark the number of TXF we're pulling now */
179 iwl_trans_write_prph(mvm->trans, TXF_LARC_NUM, i);
180
181 fifo_hdr = (void *)(*dump_data)->data;
182 fifo_data = (void *)fifo_hdr->data;
183 fifo_len = mvm->shared_mem_cfg.txfifo_size[i];
184
185 /* No need to try to read the data if the length is 0 */
186 if (fifo_len == 0)
187 continue;
188
189 /* Add a TLV for the FIFO */
190 (*dump_data)->type = cpu_to_le32(IWL_FW_ERROR_DUMP_TXF);
191 (*dump_data)->len = cpu_to_le32(fifo_len + sizeof(*fifo_hdr));
192
193 fifo_hdr->fifo_num = cpu_to_le32(i);
194 fifo_hdr->available_bytes =
195 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
196 TXF_FIFO_ITEM_CNT));
197 fifo_hdr->wr_ptr =
198 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
199 TXF_WR_PTR));
200 fifo_hdr->rd_ptr =
201 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
202 TXF_RD_PTR));
203 fifo_hdr->fence_ptr =
204 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
205 TXF_FENCE_PTR));
206 fifo_hdr->fence_mode =
207 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
208 TXF_LOCK_FENCE));
209
210 /* Set the TXF_READ_MODIFY_ADDR to TXF_WR_PTR */
211 iwl_trans_write_prph(mvm->trans, TXF_READ_MODIFY_ADDR,
212 TXF_WR_PTR);
213
214 /* Dummy-read to advance the read pointer to the head */
215 iwl_trans_read_prph(mvm->trans, TXF_READ_MODIFY_DATA);
216
217 /* Read FIFO */
218 fifo_len /= sizeof(u32); /* Size in DWORDS */
219 for (j = 0; j < fifo_len; j++)
220 fifo_data[j] = iwl_trans_read_prph(mvm->trans,
221 TXF_READ_MODIFY_DATA);
222 *dump_data = iwl_fw_error_next_data(*dump_data);
223 }
224
Golan Ben-Ami5b086412016-02-09 12:57:16 +0200225 if (fw_has_capa(&mvm->fw->ucode_capa,
226 IWL_UCODE_TLV_CAPA_EXTEND_SHARED_MEM_CFG)) {
227 /* Pull UMAC internal TXF data from all TXFs */
228 for (i = 0;
229 i < ARRAY_SIZE(mvm->shared_mem_cfg.internal_txfifo_size);
230 i++) {
Golan Ben-Ami5b086412016-02-09 12:57:16 +0200231 fifo_hdr = (void *)(*dump_data)->data;
232 fifo_data = (void *)fifo_hdr->data;
233 fifo_len = mvm->shared_mem_cfg.internal_txfifo_size[i];
234
235 /* No need to try to read the data if the length is 0 */
236 if (fifo_len == 0)
237 continue;
238
239 /* Add a TLV for the internal FIFOs */
240 (*dump_data)->type =
241 cpu_to_le32(IWL_FW_ERROR_DUMP_INTERNAL_TXF);
242 (*dump_data)->len =
243 cpu_to_le32(fifo_len + sizeof(*fifo_hdr));
244
245 fifo_hdr->fifo_num = cpu_to_le32(i);
Emmanuel Grumbach39654cb32016-04-12 13:07:52 +0300246
247 /* Mark the number of TXF we're pulling now */
Golan Ben-Amie7c9bd12016-06-15 09:16:24 +0300248 iwl_trans_write_prph(mvm->trans, TXF_CPU2_NUM, i +
249 ARRAY_SIZE(mvm->shared_mem_cfg.txfifo_size));
Emmanuel Grumbach39654cb32016-04-12 13:07:52 +0300250
Golan Ben-Ami5b086412016-02-09 12:57:16 +0200251 fifo_hdr->available_bytes =
252 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
253 TXF_CPU2_FIFO_ITEM_CNT));
254 fifo_hdr->wr_ptr =
255 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
256 TXF_CPU2_WR_PTR));
257 fifo_hdr->rd_ptr =
258 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
259 TXF_CPU2_RD_PTR));
260 fifo_hdr->fence_ptr =
261 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
262 TXF_CPU2_FENCE_PTR));
263 fifo_hdr->fence_mode =
264 cpu_to_le32(iwl_trans_read_prph(mvm->trans,
265 TXF_CPU2_LOCK_FENCE));
266
267 /* Set TXF_CPU2_READ_MODIFY_ADDR to TXF_CPU2_WR_PTR */
268 iwl_trans_write_prph(mvm->trans,
269 TXF_CPU2_READ_MODIFY_ADDR,
270 TXF_CPU2_WR_PTR);
271
272 /* Dummy-read to advance the read pointer to head */
273 iwl_trans_read_prph(mvm->trans,
274 TXF_CPU2_READ_MODIFY_DATA);
275
276 /* Read FIFO */
277 fifo_len /= sizeof(u32); /* Size in DWORDS */
278 for (j = 0; j < fifo_len; j++)
279 fifo_data[j] =
280 iwl_trans_read_prph(mvm->trans,
281 TXF_CPU2_READ_MODIFY_DATA);
282 *dump_data = iwl_fw_error_next_data(*dump_data);
283 }
284 }
285
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200286 iwl_trans_release_nic_access(mvm->trans, &flags);
287}
288
289void iwl_mvm_free_fw_dump_desc(struct iwl_mvm *mvm)
290{
Emmanuel Grumbach9fb78072016-01-05 09:35:21 +0200291 if (mvm->fw_dump_desc == &iwl_mvm_dump_desc_assert)
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200292 return;
293
294 kfree(mvm->fw_dump_desc);
295 mvm->fw_dump_desc = NULL;
296}
297
298#define IWL8260_ICCM_OFFSET 0x44000 /* Only for B-step */
299#define IWL8260_ICCM_LEN 0xC000 /* Only for B-step */
300
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300301struct iwl_prph_range {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200302 u32 start, end;
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300303};
304
305static const struct iwl_prph_range iwl_prph_dump_addr_comm[] = {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200306 { .start = 0x00a00000, .end = 0x00a00000 },
307 { .start = 0x00a0000c, .end = 0x00a00024 },
308 { .start = 0x00a0002c, .end = 0x00a0003c },
309 { .start = 0x00a00410, .end = 0x00a00418 },
310 { .start = 0x00a00420, .end = 0x00a00420 },
311 { .start = 0x00a00428, .end = 0x00a00428 },
312 { .start = 0x00a00430, .end = 0x00a0043c },
313 { .start = 0x00a00444, .end = 0x00a00444 },
314 { .start = 0x00a004c0, .end = 0x00a004cc },
315 { .start = 0x00a004d8, .end = 0x00a004d8 },
316 { .start = 0x00a004e0, .end = 0x00a004f0 },
317 { .start = 0x00a00840, .end = 0x00a00840 },
318 { .start = 0x00a00850, .end = 0x00a00858 },
319 { .start = 0x00a01004, .end = 0x00a01008 },
320 { .start = 0x00a01010, .end = 0x00a01010 },
321 { .start = 0x00a01018, .end = 0x00a01018 },
322 { .start = 0x00a01024, .end = 0x00a01024 },
323 { .start = 0x00a0102c, .end = 0x00a01034 },
324 { .start = 0x00a0103c, .end = 0x00a01040 },
325 { .start = 0x00a01048, .end = 0x00a01094 },
326 { .start = 0x00a01c00, .end = 0x00a01c20 },
327 { .start = 0x00a01c58, .end = 0x00a01c58 },
328 { .start = 0x00a01c7c, .end = 0x00a01c7c },
329 { .start = 0x00a01c28, .end = 0x00a01c54 },
330 { .start = 0x00a01c5c, .end = 0x00a01c5c },
331 { .start = 0x00a01c60, .end = 0x00a01cdc },
332 { .start = 0x00a01ce0, .end = 0x00a01d0c },
333 { .start = 0x00a01d18, .end = 0x00a01d20 },
334 { .start = 0x00a01d2c, .end = 0x00a01d30 },
335 { .start = 0x00a01d40, .end = 0x00a01d5c },
336 { .start = 0x00a01d80, .end = 0x00a01d80 },
337 { .start = 0x00a01d98, .end = 0x00a01d9c },
338 { .start = 0x00a01da8, .end = 0x00a01da8 },
339 { .start = 0x00a01db8, .end = 0x00a01df4 },
340 { .start = 0x00a01dc0, .end = 0x00a01dfc },
341 { .start = 0x00a01e00, .end = 0x00a01e2c },
342 { .start = 0x00a01e40, .end = 0x00a01e60 },
343 { .start = 0x00a01e68, .end = 0x00a01e6c },
344 { .start = 0x00a01e74, .end = 0x00a01e74 },
345 { .start = 0x00a01e84, .end = 0x00a01e90 },
346 { .start = 0x00a01e9c, .end = 0x00a01ec4 },
347 { .start = 0x00a01ed0, .end = 0x00a01ee0 },
348 { .start = 0x00a01f00, .end = 0x00a01f1c },
349 { .start = 0x00a01f44, .end = 0x00a01ffc },
350 { .start = 0x00a02000, .end = 0x00a02048 },
351 { .start = 0x00a02068, .end = 0x00a020f0 },
352 { .start = 0x00a02100, .end = 0x00a02118 },
353 { .start = 0x00a02140, .end = 0x00a0214c },
354 { .start = 0x00a02168, .end = 0x00a0218c },
355 { .start = 0x00a021c0, .end = 0x00a021c0 },
356 { .start = 0x00a02400, .end = 0x00a02410 },
357 { .start = 0x00a02418, .end = 0x00a02420 },
358 { .start = 0x00a02428, .end = 0x00a0242c },
359 { .start = 0x00a02434, .end = 0x00a02434 },
360 { .start = 0x00a02440, .end = 0x00a02460 },
361 { .start = 0x00a02468, .end = 0x00a024b0 },
362 { .start = 0x00a024c8, .end = 0x00a024cc },
363 { .start = 0x00a02500, .end = 0x00a02504 },
364 { .start = 0x00a0250c, .end = 0x00a02510 },
365 { .start = 0x00a02540, .end = 0x00a02554 },
366 { .start = 0x00a02580, .end = 0x00a025f4 },
367 { .start = 0x00a02600, .end = 0x00a0260c },
368 { .start = 0x00a02648, .end = 0x00a02650 },
369 { .start = 0x00a02680, .end = 0x00a02680 },
370 { .start = 0x00a026c0, .end = 0x00a026d0 },
371 { .start = 0x00a02700, .end = 0x00a0270c },
372 { .start = 0x00a02804, .end = 0x00a02804 },
373 { .start = 0x00a02818, .end = 0x00a0281c },
374 { .start = 0x00a02c00, .end = 0x00a02db4 },
375 { .start = 0x00a02df4, .end = 0x00a02fb0 },
376 { .start = 0x00a03000, .end = 0x00a03014 },
377 { .start = 0x00a0301c, .end = 0x00a0302c },
378 { .start = 0x00a03034, .end = 0x00a03038 },
379 { .start = 0x00a03040, .end = 0x00a03048 },
380 { .start = 0x00a03060, .end = 0x00a03068 },
381 { .start = 0x00a03070, .end = 0x00a03074 },
382 { .start = 0x00a0307c, .end = 0x00a0307c },
383 { .start = 0x00a03080, .end = 0x00a03084 },
384 { .start = 0x00a0308c, .end = 0x00a03090 },
385 { .start = 0x00a03098, .end = 0x00a03098 },
386 { .start = 0x00a030a0, .end = 0x00a030a0 },
387 { .start = 0x00a030a8, .end = 0x00a030b4 },
388 { .start = 0x00a030bc, .end = 0x00a030bc },
389 { .start = 0x00a030c0, .end = 0x00a0312c },
390 { .start = 0x00a03c00, .end = 0x00a03c5c },
391 { .start = 0x00a04400, .end = 0x00a04454 },
392 { .start = 0x00a04460, .end = 0x00a04474 },
393 { .start = 0x00a044c0, .end = 0x00a044ec },
394 { .start = 0x00a04500, .end = 0x00a04504 },
395 { .start = 0x00a04510, .end = 0x00a04538 },
396 { .start = 0x00a04540, .end = 0x00a04548 },
397 { .start = 0x00a04560, .end = 0x00a0457c },
398 { .start = 0x00a04590, .end = 0x00a04598 },
399 { .start = 0x00a045c0, .end = 0x00a045f4 },
400};
401
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300402static const struct iwl_prph_range iwl_prph_dump_addr_9000[] = {
403 { .start = 0x00a05c00, .end = 0x00a05c18 },
404 { .start = 0x00a05400, .end = 0x00a056e8 },
405 { .start = 0x00a08000, .end = 0x00a098bc },
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300406 { .start = 0x00a02400, .end = 0x00a02758 },
407};
408
Johannes Berg5bdaa0e2016-10-20 10:01:43 +0200409static void _iwl_read_prph_block(struct iwl_trans *trans, u32 start,
410 u32 len_bytes, __le32 *data)
411{
412 u32 i;
413
414 for (i = 0; i < len_bytes; i += 4)
415 *data++ = cpu_to_le32(iwl_read_prph_no_grab(trans, start + i));
416}
417
418static bool iwl_read_prph_block(struct iwl_trans *trans, u32 start,
419 u32 len_bytes, __le32 *data)
420{
421 unsigned long flags;
422 bool success = false;
423
424 if (iwl_trans_grab_nic_access(trans, &flags)) {
425 success = true;
426 _iwl_read_prph_block(trans, start, len_bytes, data);
427 iwl_trans_release_nic_access(trans, &flags);
428 }
429
430 return success;
431}
432
Johannes Berg1110f8e2016-10-20 09:44:05 +0200433static void iwl_dump_prph(struct iwl_trans *trans,
434 struct iwl_fw_error_dump_data **data,
435 const struct iwl_prph_range *iwl_prph_dump_addr,
436 u32 range_len)
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200437{
438 struct iwl_fw_error_dump_prph *prph;
439 unsigned long flags;
Johannes Berg1110f8e2016-10-20 09:44:05 +0200440 u32 i;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200441
Emmanuel Grumbach23ba9342015-12-17 11:55:13 +0200442 if (!iwl_trans_grab_nic_access(trans, &flags))
Johannes Berg1110f8e2016-10-20 09:44:05 +0200443 return;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200444
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300445 for (i = 0; i < range_len; i++) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200446 /* The range includes both boundaries */
447 int num_bytes_in_chunk = iwl_prph_dump_addr[i].end -
448 iwl_prph_dump_addr[i].start + 4;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200449
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200450 (*data)->type = cpu_to_le32(IWL_FW_ERROR_DUMP_PRPH);
451 (*data)->len = cpu_to_le32(sizeof(*prph) +
452 num_bytes_in_chunk);
453 prph = (void *)(*data)->data;
454 prph->prph_start = cpu_to_le32(iwl_prph_dump_addr[i].start);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200455
Johannes Berg5bdaa0e2016-10-20 10:01:43 +0200456 _iwl_read_prph_block(trans, iwl_prph_dump_addr[i].start,
457 /* our range is inclusive, hence + 4 */
458 iwl_prph_dump_addr[i].end -
459 iwl_prph_dump_addr[i].start + 4,
460 (void *)prph->data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200461
Dan Carpenter95a451c2015-12-09 13:26:08 +0300462 *data = iwl_fw_error_next_data(*data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200463 }
464
465 iwl_trans_release_nic_access(trans, &flags);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200466}
467
Aviya Erenfeld7e62a692016-09-20 18:07:44 +0300468/*
469 * alloc_sgtable - allocates scallerlist table in the given size,
470 * fills it with pages and returns it
471 * @size: the size (in bytes) of the table
472*/
473static struct scatterlist *alloc_sgtable(int size)
474{
475 int alloc_size, nents, i;
476 struct page *new_page;
477 struct scatterlist *iter;
478 struct scatterlist *table;
479
480 nents = DIV_ROUND_UP(size, PAGE_SIZE);
481 table = kcalloc(nents, sizeof(*table), GFP_KERNEL);
482 if (!table)
483 return NULL;
484 sg_init_table(table, nents);
485 iter = table;
486 for_each_sg(table, iter, sg_nents(table), i) {
487 new_page = alloc_page(GFP_KERNEL);
488 if (!new_page) {
489 /* release all previous allocated pages in the table */
490 iter = table;
491 for_each_sg(table, iter, sg_nents(table), i) {
492 new_page = sg_page(iter);
493 if (new_page)
494 __free_page(new_page);
495 }
496 return NULL;
497 }
498 alloc_size = min_t(int, size, PAGE_SIZE);
499 size -= PAGE_SIZE;
500 sg_set_page(iter, new_page, alloc_size, 0);
501 }
502 return table;
503}
504
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200505void iwl_mvm_fw_error_dump(struct iwl_mvm *mvm)
506{
507 struct iwl_fw_error_dump_file *dump_file;
508 struct iwl_fw_error_dump_data *dump_data;
509 struct iwl_fw_error_dump_info *dump_info;
510 struct iwl_fw_error_dump_mem *dump_mem;
511 struct iwl_fw_error_dump_trigger_desc *dump_trig;
512 struct iwl_mvm_dump_ptrs *fw_error_dump;
Aviya Erenfeld7e62a692016-09-20 18:07:44 +0300513 struct scatterlist *sg_dump_data;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200514 u32 sram_len, sram_ofs;
Johannes Berg2ed1e012016-10-20 09:41:14 +0200515 const struct iwl_fw_dbg_mem_seg_tlv *fw_dbg_mem = mvm->fw->dbg_mem_tlv;
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200516 u32 file_len, fifo_data_len = 0, prph_len = 0, radio_len = 0;
Johannes Berg2ed1e012016-10-20 09:41:14 +0200517 u32 smem_len = mvm->fw->n_dbg_mem_tlv ? 0 : mvm->cfg->smem_len;
518 u32 sram2_len = mvm->fw->n_dbg_mem_tlv ? 0 : mvm->cfg->dccm2_len;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200519 bool monitor_dump_only = false;
520 int i;
521
Golan Ben-Ami33efe942015-12-23 17:53:27 +0200522 if (!IWL_MVM_COLLECT_FW_ERR_DUMP &&
523 !mvm->trans->dbg_dest_tlv)
524 return;
525
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200526 lockdep_assert_held(&mvm->mutex);
527
528 /* there's no point in fw dump if the bus is dead */
529 if (test_bit(STATUS_TRANS_DEAD, &mvm->trans->status)) {
530 IWL_ERR(mvm, "Skip fw error dump since bus is dead\n");
Emmanuel Grumbach9fb78072016-01-05 09:35:21 +0200531 goto out;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200532 }
533
534 if (mvm->fw_dump_trig &&
535 mvm->fw_dump_trig->mode & IWL_FW_DBG_TRIGGER_MONITOR_ONLY)
536 monitor_dump_only = true;
537
538 fw_error_dump = kzalloc(sizeof(*fw_error_dump), GFP_KERNEL);
539 if (!fw_error_dump)
Emmanuel Grumbach9fb78072016-01-05 09:35:21 +0200540 goto out;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200541
542 /* SRAM - include stack CCM if driver knows the values for it */
543 if (!mvm->cfg->dccm_offset || !mvm->cfg->dccm_len) {
544 const struct fw_img *img;
545
546 img = &mvm->fw->img[mvm->cur_ucode];
547 sram_ofs = img->sec[IWL_UCODE_SECTION_DATA].offset;
548 sram_len = img->sec[IWL_UCODE_SECTION_DATA].len;
549 } else {
550 sram_ofs = mvm->cfg->dccm_offset;
551 sram_len = mvm->cfg->dccm_len;
552 }
553
554 /* reading RXF/TXF sizes */
555 if (test_bit(STATUS_FW_ERROR, &mvm->trans->status)) {
556 struct iwl_mvm_shared_mem_cfg *mem_cfg = &mvm->shared_mem_cfg;
557
558 fifo_data_len = 0;
559
560 /* Count RXF size */
561 for (i = 0; i < ARRAY_SIZE(mem_cfg->rxfifo_size); i++) {
562 if (!mem_cfg->rxfifo_size[i])
563 continue;
564
565 /* Add header info */
566 fifo_data_len += mem_cfg->rxfifo_size[i] +
567 sizeof(*dump_data) +
568 sizeof(struct iwl_fw_error_dump_fifo);
569 }
570
Sara Sharondb06f042016-07-05 17:37:58 +0300571 for (i = 0; i < mem_cfg->num_txfifo_entries; i++) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200572 if (!mem_cfg->txfifo_size[i])
573 continue;
574
575 /* Add header info */
576 fifo_data_len += mem_cfg->txfifo_size[i] +
577 sizeof(*dump_data) +
578 sizeof(struct iwl_fw_error_dump_fifo);
579 }
Emmanuel Grumbache8f0c4d2015-12-16 13:42:17 +0200580
Golan Ben-Ami5b086412016-02-09 12:57:16 +0200581 if (fw_has_capa(&mvm->fw->ucode_capa,
582 IWL_UCODE_TLV_CAPA_EXTEND_SHARED_MEM_CFG)) {
583 for (i = 0;
584 i < ARRAY_SIZE(mem_cfg->internal_txfifo_size);
585 i++) {
586 if (!mem_cfg->internal_txfifo_size[i])
587 continue;
588
589 /* Add header info */
590 fifo_data_len +=
591 mem_cfg->internal_txfifo_size[i] +
592 sizeof(*dump_data) +
593 sizeof(struct iwl_fw_error_dump_fifo);
594 }
595 }
596
Emmanuel Grumbache8f0c4d2015-12-16 13:42:17 +0200597 /* Make room for PRPH registers */
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300598 for (i = 0; i < ARRAY_SIZE(iwl_prph_dump_addr_comm); i++) {
Emmanuel Grumbache8f0c4d2015-12-16 13:42:17 +0200599 /* The range includes both boundaries */
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300600 int num_bytes_in_chunk =
601 iwl_prph_dump_addr_comm[i].end -
602 iwl_prph_dump_addr_comm[i].start + 4;
Emmanuel Grumbache8f0c4d2015-12-16 13:42:17 +0200603
604 prph_len += sizeof(*dump_data) +
605 sizeof(struct iwl_fw_error_dump_prph) +
606 num_bytes_in_chunk;
607 }
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200608
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300609 if (mvm->cfg->mq_rx_supported) {
610 for (i = 0; i <
611 ARRAY_SIZE(iwl_prph_dump_addr_9000); i++) {
612 /* The range includes both boundaries */
613 int num_bytes_in_chunk =
614 iwl_prph_dump_addr_9000[i].end -
615 iwl_prph_dump_addr_9000[i].start + 4;
616
617 prph_len += sizeof(*dump_data) +
618 sizeof(struct iwl_fw_error_dump_prph) +
619 num_bytes_in_chunk;
620 }
621 }
622
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200623 if (mvm->cfg->device_family == IWL_DEVICE_FAMILY_7000)
624 radio_len = sizeof(*dump_data) + RADIO_REG_MAX_READ;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200625 }
626
627 file_len = sizeof(*dump_file) +
628 sizeof(*dump_data) * 2 +
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200629 fifo_data_len +
Emmanuel Grumbache8f0c4d2015-12-16 13:42:17 +0200630 prph_len +
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200631 radio_len +
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200632 sizeof(*dump_info);
633
634 /* Make room for the SMEM, if it exists */
635 if (smem_len)
636 file_len += sizeof(*dump_data) + sizeof(*dump_mem) + smem_len;
637
638 /* Make room for the secondary SRAM, if it exists */
639 if (sram2_len)
640 file_len += sizeof(*dump_data) + sizeof(*dump_mem) + sram2_len;
641
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200642 /* Make room for MEM segments */
Johannes Berg2ed1e012016-10-20 09:41:14 +0200643 for (i = 0; i < mvm->fw->n_dbg_mem_tlv; i++) {
644 file_len += sizeof(*dump_data) + sizeof(*dump_mem) +
645 le32_to_cpu(fw_dbg_mem[i].len);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200646 }
647
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200648 /* Make room for fw's virtual image pages, if it exists */
Matti Gottliebf742aaf2016-04-10 10:53:57 +0300649 if (mvm->fw->img[mvm->cur_ucode].paging_mem_size &&
650 mvm->fw_paging_db[0].fw_paging_block)
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200651 file_len += mvm->num_of_paging_blk *
652 (sizeof(*dump_data) +
653 sizeof(struct iwl_fw_error_dump_paging) +
654 PAGING_BLOCK_SIZE);
655
656 /* If we only want a monitor dump, reset the file length */
657 if (monitor_dump_only) {
658 file_len = sizeof(*dump_file) + sizeof(*dump_data) +
659 sizeof(*dump_info);
660 }
661
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200662 /*
663 * In 8000 HW family B-step include the ICCM (which resides separately)
664 */
665 if (mvm->cfg->device_family == IWL_DEVICE_FAMILY_8000 &&
666 CSR_HW_REV_STEP(mvm->trans->hw_rev) == SILICON_B_STEP)
667 file_len += sizeof(*dump_data) + sizeof(*dump_mem) +
668 IWL8260_ICCM_LEN;
669
670 if (mvm->fw_dump_desc)
671 file_len += sizeof(*dump_data) + sizeof(*dump_trig) +
672 mvm->fw_dump_desc->len;
673
Johannes Berg2ed1e012016-10-20 09:41:14 +0200674 if (!mvm->fw->n_dbg_mem_tlv)
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200675 file_len += sram_len + sizeof(*dump_mem);
676
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200677 dump_file = vzalloc(file_len);
678 if (!dump_file) {
679 kfree(fw_error_dump);
Emmanuel Grumbach9fb78072016-01-05 09:35:21 +0200680 goto out;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200681 }
682
683 fw_error_dump->op_mode_ptr = dump_file;
684
685 dump_file->barker = cpu_to_le32(IWL_FW_ERROR_DUMP_BARKER);
686 dump_data = (void *)dump_file->data;
687
688 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_DEV_FW_INFO);
689 dump_data->len = cpu_to_le32(sizeof(*dump_info));
690 dump_info = (void *)dump_data->data;
691 dump_info->device_family =
692 mvm->cfg->device_family == IWL_DEVICE_FAMILY_7000 ?
693 cpu_to_le32(IWL_FW_ERROR_DUMP_FAMILY_7) :
694 cpu_to_le32(IWL_FW_ERROR_DUMP_FAMILY_8);
695 dump_info->hw_step = cpu_to_le32(CSR_HW_REV_STEP(mvm->trans->hw_rev));
696 memcpy(dump_info->fw_human_readable, mvm->fw->human_readable,
697 sizeof(dump_info->fw_human_readable));
698 strncpy(dump_info->dev_human_readable, mvm->cfg->name,
699 sizeof(dump_info->dev_human_readable));
700 strncpy(dump_info->bus_human_readable, mvm->dev->bus->name,
701 sizeof(dump_info->bus_human_readable));
702
703 dump_data = iwl_fw_error_next_data(dump_data);
704 /* We only dump the FIFOs if the FW is in error state */
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200705 if (test_bit(STATUS_FW_ERROR, &mvm->trans->status)) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200706 iwl_mvm_dump_fifos(mvm, &dump_data);
Emmanuel Grumbach976f15a2015-12-28 15:22:28 +0200707 if (radio_len)
708 iwl_mvm_read_radio_reg(mvm, &dump_data);
709 }
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200710
711 if (mvm->fw_dump_desc) {
712 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_ERROR_INFO);
713 dump_data->len = cpu_to_le32(sizeof(*dump_trig) +
714 mvm->fw_dump_desc->len);
715 dump_trig = (void *)dump_data->data;
716 memcpy(dump_trig, &mvm->fw_dump_desc->trig_desc,
717 sizeof(*dump_trig) + mvm->fw_dump_desc->len);
718
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200719 dump_data = iwl_fw_error_next_data(dump_data);
720 }
721
722 /* In case we only want monitor dump, skip to dump trasport data */
723 if (monitor_dump_only)
724 goto dump_trans_data;
725
Johannes Berg2ed1e012016-10-20 09:41:14 +0200726 if (!mvm->fw->n_dbg_mem_tlv) {
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200727 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM);
728 dump_data->len = cpu_to_le32(sram_len + sizeof(*dump_mem));
729 dump_mem = (void *)dump_data->data;
730 dump_mem->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM);
731 dump_mem->offset = cpu_to_le32(sram_ofs);
732 iwl_trans_read_mem_bytes(mvm->trans, sram_ofs, dump_mem->data,
733 sram_len);
734 dump_data = iwl_fw_error_next_data(dump_data);
735 }
736
Johannes Berg2ed1e012016-10-20 09:41:14 +0200737 for (i = 0; i < mvm->fw->n_dbg_mem_tlv; i++) {
738 u32 len = le32_to_cpu(fw_dbg_mem[i].len);
739 u32 ofs = le32_to_cpu(fw_dbg_mem[i].ofs);
Johannes Berg5bdaa0e2016-10-20 10:01:43 +0200740 bool success;
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200741
Johannes Berg2ed1e012016-10-20 09:41:14 +0200742 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM);
743 dump_data->len = cpu_to_le32(len + sizeof(*dump_mem));
744 dump_mem = (void *)dump_data->data;
745 dump_mem->type = fw_dbg_mem[i].data_type;
746 dump_mem->offset = cpu_to_le32(ofs);
Johannes Berg5bdaa0e2016-10-20 10:01:43 +0200747
748 switch (dump_mem->type & cpu_to_le32(FW_DBG_MEM_TYPE_MASK)) {
749 case cpu_to_le32(FW_DBG_MEM_TYPE_REGULAR):
750 iwl_trans_read_mem_bytes(mvm->trans, ofs,
751 dump_mem->data,
752 len);
753 success = true;
754 break;
755 case cpu_to_le32(FW_DBG_MEM_TYPE_PRPH):
756 success = iwl_read_prph_block(mvm->trans, ofs, len,
757 (void *)dump_mem->data);
758 break;
759 default:
760 /*
761 * shouldn't get here, we ignored this kind
762 * of TLV earlier during the TLV parsing?!
763 */
764 WARN_ON(1);
765 success = false;
766 }
767
768 if (success)
769 dump_data = iwl_fw_error_next_data(dump_data);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200770 }
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200771
772 if (smem_len) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200773 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM);
774 dump_data->len = cpu_to_le32(smem_len + sizeof(*dump_mem));
775 dump_mem = (void *)dump_data->data;
776 dump_mem->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SMEM);
777 dump_mem->offset = cpu_to_le32(mvm->cfg->smem_offset);
778 iwl_trans_read_mem_bytes(mvm->trans, mvm->cfg->smem_offset,
779 dump_mem->data, smem_len);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200780 dump_data = iwl_fw_error_next_data(dump_data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200781 }
782
783 if (sram2_len) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200784 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM);
785 dump_data->len = cpu_to_le32(sram2_len + sizeof(*dump_mem));
786 dump_mem = (void *)dump_data->data;
787 dump_mem->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM);
788 dump_mem->offset = cpu_to_le32(mvm->cfg->dccm2_offset);
789 iwl_trans_read_mem_bytes(mvm->trans, mvm->cfg->dccm2_offset,
790 dump_mem->data, sram2_len);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200791 dump_data = iwl_fw_error_next_data(dump_data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200792 }
793
794 if (mvm->cfg->device_family == IWL_DEVICE_FAMILY_8000 &&
795 CSR_HW_REV_STEP(mvm->trans->hw_rev) == SILICON_B_STEP) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200796 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM);
797 dump_data->len = cpu_to_le32(IWL8260_ICCM_LEN +
798 sizeof(*dump_mem));
799 dump_mem = (void *)dump_data->data;
800 dump_mem->type = cpu_to_le32(IWL_FW_ERROR_DUMP_MEM_SRAM);
801 dump_mem->offset = cpu_to_le32(IWL8260_ICCM_OFFSET);
802 iwl_trans_read_mem_bytes(mvm->trans, IWL8260_ICCM_OFFSET,
803 dump_mem->data, IWL8260_ICCM_LEN);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200804 dump_data = iwl_fw_error_next_data(dump_data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200805 }
806
807 /* Dump fw's virtual image */
Matti Gottliebf742aaf2016-04-10 10:53:57 +0300808 if (mvm->fw->img[mvm->cur_ucode].paging_mem_size &&
809 mvm->fw_paging_db[0].fw_paging_block) {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200810 for (i = 1; i < mvm->num_of_paging_blk + 1; i++) {
811 struct iwl_fw_error_dump_paging *paging;
812 struct page *pages =
813 mvm->fw_paging_db[i].fw_paging_block;
Sara Sharon4b70f072016-11-30 16:49:11 +0200814 dma_addr_t addr = mvm->fw_paging_db[i].fw_paging_phys;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200815
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200816 dump_data->type = cpu_to_le32(IWL_FW_ERROR_DUMP_PAGING);
817 dump_data->len = cpu_to_le32(sizeof(*paging) +
818 PAGING_BLOCK_SIZE);
819 paging = (void *)dump_data->data;
820 paging->index = cpu_to_le32(i);
Sara Sharon4b70f072016-11-30 16:49:11 +0200821 dma_sync_single_for_cpu(mvm->trans->dev, addr,
822 PAGING_BLOCK_SIZE,
823 DMA_BIDIRECTIONAL);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200824 memcpy(paging->data, page_address(pages),
825 PAGING_BLOCK_SIZE);
Golan Ben-Amia6017b92016-03-14 12:24:20 +0200826 dump_data = iwl_fw_error_next_data(dump_data);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200827 }
828 }
829
Golan Ben-Amie87e26392016-04-06 11:59:50 +0300830 if (prph_len) {
831 iwl_dump_prph(mvm->trans, &dump_data,
832 iwl_prph_dump_addr_comm,
833 ARRAY_SIZE(iwl_prph_dump_addr_comm));
834
835 if (mvm->cfg->mq_rx_supported)
836 iwl_dump_prph(mvm->trans, &dump_data,
837 iwl_prph_dump_addr_9000,
838 ARRAY_SIZE(iwl_prph_dump_addr_9000));
839 }
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200840
841dump_trans_data:
842 fw_error_dump->trans_ptr = iwl_trans_dump_data(mvm->trans,
843 mvm->fw_dump_trig);
844 fw_error_dump->op_mode_len = file_len;
845 if (fw_error_dump->trans_ptr)
846 file_len += fw_error_dump->trans_ptr->len;
847 dump_file->file_len = cpu_to_le32(file_len);
848
Aviya Erenfeld7e62a692016-09-20 18:07:44 +0300849 sg_dump_data = alloc_sgtable(file_len);
850 if (sg_dump_data) {
851 sg_pcopy_from_buffer(sg_dump_data,
852 sg_nents(sg_dump_data),
853 fw_error_dump->op_mode_ptr,
854 fw_error_dump->op_mode_len, 0);
Johannes Bergc2e27e12016-10-20 15:25:00 +0200855 if (fw_error_dump->trans_ptr)
856 sg_pcopy_from_buffer(sg_dump_data,
857 sg_nents(sg_dump_data),
858 fw_error_dump->trans_ptr->data,
859 fw_error_dump->trans_ptr->len,
860 fw_error_dump->op_mode_len);
Aviya Erenfeld7e62a692016-09-20 18:07:44 +0300861 dev_coredumpsg(mvm->trans->dev, sg_dump_data, file_len,
862 GFP_KERNEL);
863 }
864 vfree(fw_error_dump->op_mode_ptr);
865 vfree(fw_error_dump->trans_ptr);
866 kfree(fw_error_dump);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200867
Emmanuel Grumbach9fb78072016-01-05 09:35:21 +0200868out:
869 iwl_mvm_free_fw_dump_desc(mvm);
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200870 mvm->fw_dump_trig = NULL;
871 clear_bit(IWL_MVM_STATUS_DUMPING_FW_LOG, &mvm->status);
872}
873
Emmanuel Grumbacha80c7a62016-01-05 09:14:08 +0200874const struct iwl_mvm_dump_desc iwl_mvm_dump_desc_assert = {
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200875 .trig_desc = {
876 .type = cpu_to_le32(FW_DBG_TRIGGER_FW_ASSERT),
877 },
878};
879
880int iwl_mvm_fw_dbg_collect_desc(struct iwl_mvm *mvm,
Emmanuel Grumbacha80c7a62016-01-05 09:14:08 +0200881 const struct iwl_mvm_dump_desc *desc,
882 const struct iwl_fw_dbg_trigger_tlv *trigger)
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200883{
884 unsigned int delay = 0;
885
886 if (trigger)
887 delay = msecs_to_jiffies(le32_to_cpu(trigger->stop_delay));
888
889 if (test_and_set_bit(IWL_MVM_STATUS_DUMPING_FW_LOG, &mvm->status))
890 return -EBUSY;
891
892 if (WARN_ON(mvm->fw_dump_desc))
893 iwl_mvm_free_fw_dump_desc(mvm);
894
895 IWL_WARN(mvm, "Collecting data: trigger %d fired.\n",
896 le32_to_cpu(desc->trig_desc.type));
897
898 mvm->fw_dump_desc = desc;
899 mvm->fw_dump_trig = trigger;
900
901 queue_delayed_work(system_wq, &mvm->fw_dump_wk, delay);
902
903 return 0;
904}
905
906int iwl_mvm_fw_dbg_collect(struct iwl_mvm *mvm, enum iwl_fw_dbg_trigger trig,
907 const char *str, size_t len,
Emmanuel Grumbacha80c7a62016-01-05 09:14:08 +0200908 const struct iwl_fw_dbg_trigger_tlv *trigger)
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +0200909{
910 struct iwl_mvm_dump_desc *desc;
911
912 desc = kzalloc(sizeof(*desc) + len, GFP_ATOMIC);
913 if (!desc)
914 return -ENOMEM;
915
916 desc->len = len;
917 desc->trig_desc.type = cpu_to_le32(trig);
918 memcpy(desc->trig_desc.data, str, len);
919
920 return iwl_mvm_fw_dbg_collect_desc(mvm, desc, trigger);
921}
922
923int iwl_mvm_fw_dbg_collect_trig(struct iwl_mvm *mvm,
924 struct iwl_fw_dbg_trigger_tlv *trigger,
925 const char *fmt, ...)
926{
927 u16 occurrences = le16_to_cpu(trigger->occurrences);
928 int ret, len = 0;
929 char buf[64];
930
931 if (!occurrences)
932 return 0;
933
934 if (fmt) {
935 va_list ap;
936
937 buf[sizeof(buf) - 1] = '\0';
938
939 va_start(ap, fmt);
940 vsnprintf(buf, sizeof(buf), fmt, ap);
941 va_end(ap);
942
943 /* check for truncation */
944 if (WARN_ON_ONCE(buf[sizeof(buf) - 1]))
945 buf[sizeof(buf) - 1] = '\0';
946
947 len = strlen(buf) + 1;
948 }
949
950 ret = iwl_mvm_fw_dbg_collect(mvm, le32_to_cpu(trigger->id), buf, len,
951 trigger);
952
953 if (ret)
954 return ret;
955
956 trigger->occurrences = cpu_to_le16(occurrences - 1);
957 return 0;
958}
959
960static inline void iwl_mvm_restart_early_start(struct iwl_mvm *mvm)
961{
962 if (mvm->cfg->device_family == IWL_DEVICE_FAMILY_7000)
963 iwl_clear_bits_prph(mvm->trans, MON_BUFF_SAMPLE_CTL, 0x100);
964 else
965 iwl_write_prph(mvm->trans, DBGC_IN_SAMPLE, 1);
966}
967
968int iwl_mvm_start_fw_dbg_conf(struct iwl_mvm *mvm, u8 conf_id)
969{
970 u8 *ptr;
971 int ret;
972 int i;
973
974 if (WARN_ONCE(conf_id >= ARRAY_SIZE(mvm->fw->dbg_conf_tlv),
975 "Invalid configuration %d\n", conf_id))
976 return -EINVAL;
977
978 /* EARLY START - firmware's configuration is hard coded */
979 if ((!mvm->fw->dbg_conf_tlv[conf_id] ||
980 !mvm->fw->dbg_conf_tlv[conf_id]->num_of_hcmds) &&
981 conf_id == FW_DBG_START_FROM_ALIVE) {
982 iwl_mvm_restart_early_start(mvm);
983 return 0;
984 }
985
986 if (!mvm->fw->dbg_conf_tlv[conf_id])
987 return -EINVAL;
988
989 if (mvm->fw_dbg_conf != FW_DBG_INVALID)
990 IWL_WARN(mvm, "FW already configured (%d) - re-configuring\n",
991 mvm->fw_dbg_conf);
992
993 /* Send all HCMDs for configuring the FW debug */
994 ptr = (void *)&mvm->fw->dbg_conf_tlv[conf_id]->hcmd;
995 for (i = 0; i < mvm->fw->dbg_conf_tlv[conf_id]->num_of_hcmds; i++) {
996 struct iwl_fw_dbg_conf_hcmd *cmd = (void *)ptr;
997
998 ret = iwl_mvm_send_cmd_pdu(mvm, cmd->id, 0,
999 le16_to_cpu(cmd->len), cmd->data);
1000 if (ret)
1001 return ret;
1002
1003 ptr += sizeof(*cmd);
1004 ptr += le16_to_cpu(cmd->len);
1005 }
1006
1007 mvm->fw_dbg_conf = conf_id;
Emmanuel Grumbachff6e58e2016-08-03 22:06:43 +03001008
1009 return 0;
Golan Ben-Ami2f89a5d2015-10-27 19:17:14 +02001010}