blob: 0299e5bbb9022a457fa398ca7a401bc73bc6c1b8 [file] [log] [blame]
Jeff Kirsherae06c702018-03-22 10:08:48 -07001// SPDX-License-Identifier: GPL-2.0
Jeff Kirsher51dce242018-04-26 08:08:09 -07002/* Copyright(c) 2013 - 2018 Intel Corporation. */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +00003
4#include "i40e_prototype.h"
5
6/**
Shannon Nelson3e261862014-02-06 05:51:06 +00007 * i40e_init_nvm_ops - Initialize NVM function pointers
8 * @hw: pointer to the HW structure
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +00009 *
Shannon Nelson3e261862014-02-06 05:51:06 +000010 * Setup the function pointers and the NVM info structure. Should be called
11 * once per NVM initialization, e.g. inside the i40e_init_shared_code().
12 * Please notice that the NVM term is used here (& in all methods covered
13 * in this file) as an equivalent of the FLASH part mapped into the SR.
14 * We are accessing FLASH always thru the Shadow RAM.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000015 **/
16i40e_status i40e_init_nvm(struct i40e_hw *hw)
17{
18 struct i40e_nvm_info *nvm = &hw->nvm;
19 i40e_status ret_code = 0;
20 u32 fla, gens;
21 u8 sr_size;
22
23 /* The SR size is stored regardless of the nvm programming mode
24 * as the blank mode may be used in the factory line.
25 */
26 gens = rd32(hw, I40E_GLNVM_GENS);
27 sr_size = ((gens & I40E_GLNVM_GENS_SR_SIZE_MASK) >>
28 I40E_GLNVM_GENS_SR_SIZE_SHIFT);
Shannon Nelson3e261862014-02-06 05:51:06 +000029 /* Switching to words (sr_size contains power of 2KB) */
Jesse Brandeburg41a1d042015-06-04 16:24:02 -040030 nvm->sr_size = BIT(sr_size) * I40E_SR_WORDS_IN_1KB;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000031
Shannon Nelson3e261862014-02-06 05:51:06 +000032 /* Check if we are in the normal or blank NVM programming mode */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000033 fla = rd32(hw, I40E_GLNVM_FLA);
Shannon Nelson3e261862014-02-06 05:51:06 +000034 if (fla & I40E_GLNVM_FLA_LOCKED_MASK) { /* Normal programming mode */
35 /* Max NVM timeout */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000036 nvm->timeout = I40E_MAX_NVM_TIMEOUT;
37 nvm->blank_nvm_mode = false;
Shannon Nelson3e261862014-02-06 05:51:06 +000038 } else { /* Blank programming mode */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000039 nvm->blank_nvm_mode = true;
40 ret_code = I40E_ERR_NVM_BLANK_MODE;
Shannon Nelson74d0d0e2014-11-13 08:23:15 +000041 i40e_debug(hw, I40E_DEBUG_NVM, "NVM init error: unsupported blank mode.\n");
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000042 }
43
44 return ret_code;
45}
46
47/**
Shannon Nelson3e261862014-02-06 05:51:06 +000048 * i40e_acquire_nvm - Generic request for acquiring the NVM ownership
49 * @hw: pointer to the HW structure
50 * @access: NVM access type (read or write)
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000051 *
Shannon Nelson3e261862014-02-06 05:51:06 +000052 * This function will request NVM ownership for reading
53 * via the proper Admin Command.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000054 **/
55i40e_status i40e_acquire_nvm(struct i40e_hw *hw,
56 enum i40e_aq_resource_access_type access)
57{
58 i40e_status ret_code = 0;
59 u64 gtime, timeout;
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000060 u64 time_left = 0;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000061
62 if (hw->nvm.blank_nvm_mode)
63 goto i40e_i40e_acquire_nvm_exit;
64
65 ret_code = i40e_aq_request_resource(hw, I40E_NVM_RESOURCE_ID, access,
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000066 0, &time_left, NULL);
Shannon Nelson3e261862014-02-06 05:51:06 +000067 /* Reading the Global Device Timer */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000068 gtime = rd32(hw, I40E_GLVFGEN_TIMER);
69
Shannon Nelson3e261862014-02-06 05:51:06 +000070 /* Store the timeout */
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000071 hw->nvm.hw_semaphore_timeout = I40E_MS_TO_GTIME(time_left) + gtime;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000072
Shannon Nelsona3f0b382014-11-13 08:23:21 +000073 if (ret_code)
74 i40e_debug(hw, I40E_DEBUG_NVM,
75 "NVM acquire type %d failed time_left=%llu ret=%d aq_err=%d\n",
76 access, time_left, ret_code, hw->aq.asq_last_status);
77
78 if (ret_code && time_left) {
Shannon Nelson3e261862014-02-06 05:51:06 +000079 /* Poll until the current NVM owner timeouts */
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000080 timeout = I40E_MS_TO_GTIME(I40E_MAX_NVM_TIMEOUT) + gtime;
Shannon Nelsona3f0b382014-11-13 08:23:21 +000081 while ((gtime < timeout) && time_left) {
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000082 usleep_range(10000, 20000);
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000083 gtime = rd32(hw, I40E_GLVFGEN_TIMER);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000084 ret_code = i40e_aq_request_resource(hw,
85 I40E_NVM_RESOURCE_ID,
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000086 access, 0, &time_left,
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000087 NULL);
88 if (!ret_code) {
89 hw->nvm.hw_semaphore_timeout =
Shannon Nelsonc509c1d2014-11-13 08:23:19 +000090 I40E_MS_TO_GTIME(time_left) + gtime;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000091 break;
92 }
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000093 }
94 if (ret_code) {
95 hw->nvm.hw_semaphore_timeout = 0;
Shannon Nelson74d0d0e2014-11-13 08:23:15 +000096 i40e_debug(hw, I40E_DEBUG_NVM,
Shannon Nelsona3f0b382014-11-13 08:23:21 +000097 "NVM acquire timed out, wait %llu ms before trying again. status=%d aq_err=%d\n",
98 time_left, ret_code, hw->aq.asq_last_status);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +000099 }
100 }
101
102i40e_i40e_acquire_nvm_exit:
103 return ret_code;
104}
105
106/**
Shannon Nelson3e261862014-02-06 05:51:06 +0000107 * i40e_release_nvm - Generic request for releasing the NVM ownership
108 * @hw: pointer to the HW structure
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000109 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000110 * This function will release NVM resource via the proper Admin Command.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000111 **/
112void i40e_release_nvm(struct i40e_hw *hw)
113{
Paul M Stillwell Jr981e25c2017-06-20 15:16:55 -0700114 i40e_status ret_code = I40E_SUCCESS;
115 u32 total_delay = 0;
116
117 if (hw->nvm.blank_nvm_mode)
118 return;
119
120 ret_code = i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
121
122 /* there are some rare cases when trying to release the resource
123 * results in an admin Q timeout, so handle them correctly
124 */
125 while ((ret_code == I40E_ERR_ADMIN_QUEUE_TIMEOUT) &&
126 (total_delay < hw->aq.asq_cmd_timeout)) {
127 usleep_range(1000, 2000);
128 ret_code = i40e_aq_release_resource(hw,
129 I40E_NVM_RESOURCE_ID,
130 0, NULL);
131 total_delay++;
132 }
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000133}
134
135/**
Shannon Nelson3e261862014-02-06 05:51:06 +0000136 * i40e_poll_sr_srctl_done_bit - Polls the GLNVM_SRCTL done bit
137 * @hw: pointer to the HW structure
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000138 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000139 * Polls the SRCTL Shadow RAM register done bit.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000140 **/
141static i40e_status i40e_poll_sr_srctl_done_bit(struct i40e_hw *hw)
142{
143 i40e_status ret_code = I40E_ERR_TIMEOUT;
144 u32 srctl, wait_cnt;
145
Shannon Nelson3e261862014-02-06 05:51:06 +0000146 /* Poll the I40E_GLNVM_SRCTL until the done bit is set */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000147 for (wait_cnt = 0; wait_cnt < I40E_SRRD_SRCTL_ATTEMPTS; wait_cnt++) {
148 srctl = rd32(hw, I40E_GLNVM_SRCTL);
149 if (srctl & I40E_GLNVM_SRCTL_DONE_MASK) {
150 ret_code = 0;
151 break;
152 }
153 udelay(5);
154 }
155 if (ret_code == I40E_ERR_TIMEOUT)
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000156 i40e_debug(hw, I40E_DEBUG_NVM, "Done bit in GLNVM_SRCTL not set");
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000157 return ret_code;
158}
159
160/**
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000161 * i40e_read_nvm_word_srctl - Reads Shadow RAM via SRCTL register
Shannon Nelson3e261862014-02-06 05:51:06 +0000162 * @hw: pointer to the HW structure
163 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF)
164 * @data: word read from the Shadow RAM
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000165 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000166 * Reads one 16 bit word from the Shadow RAM using the GLNVM_SRCTL register.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000167 **/
Shannon Nelson37a29732015-02-27 09:15:19 +0000168static i40e_status i40e_read_nvm_word_srctl(struct i40e_hw *hw, u16 offset,
169 u16 *data)
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000170{
171 i40e_status ret_code = I40E_ERR_TIMEOUT;
172 u32 sr_reg;
173
174 if (offset >= hw->nvm.sr_size) {
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000175 i40e_debug(hw, I40E_DEBUG_NVM,
176 "NVM read error: offset %d beyond Shadow RAM limit %d\n",
177 offset, hw->nvm.sr_size);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000178 ret_code = I40E_ERR_PARAM;
179 goto read_nvm_exit;
180 }
181
Shannon Nelson3e261862014-02-06 05:51:06 +0000182 /* Poll the done bit first */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000183 ret_code = i40e_poll_sr_srctl_done_bit(hw);
184 if (!ret_code) {
Shannon Nelson3e261862014-02-06 05:51:06 +0000185 /* Write the address and start reading */
Jesse Brandeburg41a1d042015-06-04 16:24:02 -0400186 sr_reg = ((u32)offset << I40E_GLNVM_SRCTL_ADDR_SHIFT) |
187 BIT(I40E_GLNVM_SRCTL_START_SHIFT);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000188 wr32(hw, I40E_GLNVM_SRCTL, sr_reg);
189
Shannon Nelson3e261862014-02-06 05:51:06 +0000190 /* Poll I40E_GLNVM_SRCTL until the done bit is set */
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000191 ret_code = i40e_poll_sr_srctl_done_bit(hw);
192 if (!ret_code) {
193 sr_reg = rd32(hw, I40E_GLNVM_SRDATA);
194 *data = (u16)((sr_reg &
195 I40E_GLNVM_SRDATA_RDDATA_MASK)
196 >> I40E_GLNVM_SRDATA_RDDATA_SHIFT);
197 }
198 }
199 if (ret_code)
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000200 i40e_debug(hw, I40E_DEBUG_NVM,
201 "NVM read error: Couldn't access Shadow RAM address: 0x%x\n",
202 offset);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000203
204read_nvm_exit:
205 return ret_code;
206}
207
208/**
Shannon Nelson7073f462015-06-05 12:20:34 -0400209 * i40e_read_nvm_aq - Read Shadow RAM.
210 * @hw: pointer to the HW structure.
211 * @module_pointer: module pointer location in words from the NVM beginning
212 * @offset: offset in words from module start
213 * @words: number of words to write
214 * @data: buffer with words to write to the Shadow RAM
215 * @last_command: tells the AdminQ that this is the last command
216 *
217 * Writes a 16 bit words buffer to the Shadow RAM using the admin command.
218 **/
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500219static i40e_status i40e_read_nvm_aq(struct i40e_hw *hw,
220 u8 module_pointer, u32 offset,
221 u16 words, void *data,
Shannon Nelson7073f462015-06-05 12:20:34 -0400222 bool last_command)
223{
224 i40e_status ret_code = I40E_ERR_NVM;
225 struct i40e_asq_cmd_details cmd_details;
226
227 memset(&cmd_details, 0, sizeof(cmd_details));
Jacob Keller3c8f3e92017-09-01 13:43:08 -0700228 cmd_details.wb_desc = &hw->nvm_wb_desc;
Shannon Nelson7073f462015-06-05 12:20:34 -0400229
230 /* Here we are checking the SR limit only for the flat memory model.
231 * We cannot do it for the module-based model, as we did not acquire
232 * the NVM resource yet (we cannot get the module pointer value).
233 * Firmware will check the module-based model.
234 */
235 if ((offset + words) > hw->nvm.sr_size)
236 i40e_debug(hw, I40E_DEBUG_NVM,
237 "NVM write error: offset %d beyond Shadow RAM limit %d\n",
238 (offset + words), hw->nvm.sr_size);
239 else if (words > I40E_SR_SECTOR_SIZE_IN_WORDS)
240 /* We can write only up to 4KB (one sector), in one AQ write */
241 i40e_debug(hw, I40E_DEBUG_NVM,
242 "NVM write fail error: tried to write %d words, limit is %d.\n",
243 words, I40E_SR_SECTOR_SIZE_IN_WORDS);
244 else if (((offset + (words - 1)) / I40E_SR_SECTOR_SIZE_IN_WORDS)
245 != (offset / I40E_SR_SECTOR_SIZE_IN_WORDS))
246 /* A single write cannot spread over two sectors */
247 i40e_debug(hw, I40E_DEBUG_NVM,
248 "NVM write error: cannot spread over two sectors in a single write offset=%d words=%d\n",
249 offset, words);
250 else
251 ret_code = i40e_aq_read_nvm(hw, module_pointer,
252 2 * offset, /*bytes*/
253 2 * words, /*bytes*/
254 data, last_command, &cmd_details);
255
256 return ret_code;
257}
258
259/**
260 * i40e_read_nvm_word_aq - Reads Shadow RAM via AQ
261 * @hw: pointer to the HW structure
262 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF)
263 * @data: word read from the Shadow RAM
264 *
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700265 * Reads one 16 bit word from the Shadow RAM using the AdminQ
Shannon Nelson7073f462015-06-05 12:20:34 -0400266 **/
267static i40e_status i40e_read_nvm_word_aq(struct i40e_hw *hw, u16 offset,
268 u16 *data)
269{
270 i40e_status ret_code = I40E_ERR_TIMEOUT;
271
272 ret_code = i40e_read_nvm_aq(hw, 0x0, offset, 1, data, true);
273 *data = le16_to_cpu(*(__le16 *)data);
274
275 return ret_code;
276}
277
278/**
Stefano Brivioe836e322017-09-06 10:11:38 +0200279 * __i40e_read_nvm_word - Reads nvm word, assumes caller does the locking
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000280 * @hw: pointer to the HW structure
281 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF)
282 * @data: word read from the Shadow RAM
283 *
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700284 * Reads one 16 bit word from the Shadow RAM.
285 *
286 * Do not use this function except in cases where the nvm lock is already
287 * taken via i40e_acquire_nvm().
288 **/
289static i40e_status __i40e_read_nvm_word(struct i40e_hw *hw,
290 u16 offset, u16 *data)
291{
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700292 if (hw->flags & I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE)
Stefano Brivio2c4d36b2017-09-06 10:11:39 +0200293 return i40e_read_nvm_word_aq(hw, offset, data);
294
295 return i40e_read_nvm_word_srctl(hw, offset, data);
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700296}
297
298/**
299 * i40e_read_nvm_word - Reads nvm word and acquire lock if necessary
300 * @hw: pointer to the HW structure
301 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF)
302 * @data: word read from the Shadow RAM
303 *
304 * Reads one 16 bit word from the Shadow RAM.
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000305 **/
306i40e_status i40e_read_nvm_word(struct i40e_hw *hw, u16 offset,
307 u16 *data)
308{
Jacob Keller3d72aeb2017-10-27 11:06:55 -0400309 i40e_status ret_code = 0;
Anjali Singhai07f89be2015-09-24 15:26:32 -0700310
Jacob Keller3d72aeb2017-10-27 11:06:55 -0400311 if (hw->flags & I40E_HW_FLAG_NVM_READ_REQUIRES_LOCK)
312 ret_code = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700313 if (ret_code)
314 return ret_code;
315
316 ret_code = __i40e_read_nvm_word(hw, offset, data);
317
Jacob Keller3d72aeb2017-10-27 11:06:55 -0400318 if (hw->flags & I40E_HW_FLAG_NVM_READ_REQUIRES_LOCK)
319 i40e_release_nvm(hw);
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700320
Anjali Singhai07f89be2015-09-24 15:26:32 -0700321 return ret_code;
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000322}
323
324/**
325 * i40e_read_nvm_buffer_srctl - Reads Shadow RAM buffer via SRCTL register
326 * @hw: pointer to the HW structure
327 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF).
328 * @words: (in) number of words to read; (out) number of words actually read
329 * @data: words read from the Shadow RAM
330 *
331 * Reads 16 bit words (data buffer) from the SR using the i40e_read_nvm_srrd()
332 * method. The buffer read is preceded by the NVM ownership take
333 * and followed by the release.
334 **/
Shannon Nelson37a29732015-02-27 09:15:19 +0000335static i40e_status i40e_read_nvm_buffer_srctl(struct i40e_hw *hw, u16 offset,
336 u16 *words, u16 *data)
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000337{
338 i40e_status ret_code = 0;
339 u16 index, word;
340
341 /* Loop thru the selected region */
342 for (word = 0; word < *words; word++) {
343 index = offset + word;
344 ret_code = i40e_read_nvm_word_srctl(hw, index, &data[word]);
345 if (ret_code)
346 break;
347 }
348
349 /* Update the number of words read from the Shadow RAM */
350 *words = word;
351
352 return ret_code;
353}
354
355/**
Shannon Nelson7073f462015-06-05 12:20:34 -0400356 * i40e_read_nvm_buffer_aq - Reads Shadow RAM buffer via AQ
357 * @hw: pointer to the HW structure
358 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF).
359 * @words: (in) number of words to read; (out) number of words actually read
360 * @data: words read from the Shadow RAM
361 *
362 * Reads 16 bit words (data buffer) from the SR using the i40e_read_nvm_aq()
363 * method. The buffer read is preceded by the NVM ownership take
364 * and followed by the release.
365 **/
366static i40e_status i40e_read_nvm_buffer_aq(struct i40e_hw *hw, u16 offset,
367 u16 *words, u16 *data)
368{
369 i40e_status ret_code;
Colin Ian King793c6f82017-11-05 13:04:29 +0000370 u16 read_size;
Shannon Nelson7073f462015-06-05 12:20:34 -0400371 bool last_cmd = false;
372 u16 words_read = 0;
373 u16 i = 0;
374
375 do {
376 /* Calculate number of bytes we should read in this step.
377 * FVL AQ do not allow to read more than one page at a time or
378 * to cross page boundaries.
379 */
380 if (offset % I40E_SR_SECTOR_SIZE_IN_WORDS)
381 read_size = min(*words,
382 (u16)(I40E_SR_SECTOR_SIZE_IN_WORDS -
383 (offset % I40E_SR_SECTOR_SIZE_IN_WORDS)));
384 else
385 read_size = min((*words - words_read),
386 I40E_SR_SECTOR_SIZE_IN_WORDS);
387
388 /* Check if this is last command, if so set proper flag */
389 if ((words_read + read_size) >= *words)
390 last_cmd = true;
391
392 ret_code = i40e_read_nvm_aq(hw, 0x0, offset, read_size,
393 data + words_read, last_cmd);
394 if (ret_code)
395 goto read_nvm_buffer_aq_exit;
396
397 /* Increment counter for words already read and move offset to
398 * new read location
399 */
400 words_read += read_size;
401 offset += read_size;
402 } while (words_read < *words);
403
404 for (i = 0; i < *words; i++)
405 data[i] = le16_to_cpu(((__le16 *)data)[i]);
406
407read_nvm_buffer_aq_exit:
408 *words = words_read;
409 return ret_code;
410}
411
412/**
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700413 * __i40e_read_nvm_buffer - Reads nvm buffer, caller must acquire lock
Shannon Nelson3e261862014-02-06 05:51:06 +0000414 * @hw: pointer to the HW structure
415 * @offset: offset of the Shadow RAM word to read (0x000000 - 0x001FFF).
416 * @words: (in) number of words to read; (out) number of words actually read
417 * @data: words read from the Shadow RAM
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000418 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000419 * Reads 16 bit words (data buffer) from the SR using the i40e_read_nvm_srrd()
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700420 * method.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000421 **/
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700422static i40e_status __i40e_read_nvm_buffer(struct i40e_hw *hw,
423 u16 offset, u16 *words,
424 u16 *data)
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000425{
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700426 if (hw->flags & I40E_HW_FLAG_AQ_SRCTL_ACCESS_ENABLE)
Stefano Brivio2c4d36b2017-09-06 10:11:39 +0200427 return i40e_read_nvm_buffer_aq(hw, offset, words, data);
428
429 return i40e_read_nvm_buffer_srctl(hw, offset, words, data);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000430}
431
432/**
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000433 * i40e_write_nvm_aq - Writes Shadow RAM.
434 * @hw: pointer to the HW structure.
435 * @module_pointer: module pointer location in words from the NVM beginning
436 * @offset: offset in words from module start
437 * @words: number of words to write
438 * @data: buffer with words to write to the Shadow RAM
439 * @last_command: tells the AdminQ that this is the last command
440 *
441 * Writes a 16 bit words buffer to the Shadow RAM using the admin command.
442 **/
Wei Yongjun952d9632014-07-30 09:02:53 +0000443static i40e_status i40e_write_nvm_aq(struct i40e_hw *hw, u8 module_pointer,
444 u32 offset, u16 words, void *data,
445 bool last_command)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000446{
447 i40e_status ret_code = I40E_ERR_NVM;
Shannon Nelson6b5c1b82015-08-28 17:55:47 -0400448 struct i40e_asq_cmd_details cmd_details;
449
450 memset(&cmd_details, 0, sizeof(cmd_details));
451 cmd_details.wb_desc = &hw->nvm_wb_desc;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000452
453 /* Here we are checking the SR limit only for the flat memory model.
454 * We cannot do it for the module-based model, as we did not acquire
455 * the NVM resource yet (we cannot get the module pointer value).
456 * Firmware will check the module-based model.
457 */
458 if ((offset + words) > hw->nvm.sr_size)
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000459 i40e_debug(hw, I40E_DEBUG_NVM,
460 "NVM write error: offset %d beyond Shadow RAM limit %d\n",
461 (offset + words), hw->nvm.sr_size);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000462 else if (words > I40E_SR_SECTOR_SIZE_IN_WORDS)
463 /* We can write only up to 4KB (one sector), in one AQ write */
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000464 i40e_debug(hw, I40E_DEBUG_NVM,
465 "NVM write fail error: tried to write %d words, limit is %d.\n",
466 words, I40E_SR_SECTOR_SIZE_IN_WORDS);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000467 else if (((offset + (words - 1)) / I40E_SR_SECTOR_SIZE_IN_WORDS)
468 != (offset / I40E_SR_SECTOR_SIZE_IN_WORDS))
469 /* A single write cannot spread over two sectors */
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000470 i40e_debug(hw, I40E_DEBUG_NVM,
471 "NVM write error: cannot spread over two sectors in a single write offset=%d words=%d\n",
472 offset, words);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000473 else
474 ret_code = i40e_aq_update_nvm(hw, module_pointer,
475 2 * offset, /*bytes*/
476 2 * words, /*bytes*/
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500477 data, last_command, 0,
478 &cmd_details);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000479
480 return ret_code;
481}
482
483/**
Shannon Nelson3e261862014-02-06 05:51:06 +0000484 * i40e_calc_nvm_checksum - Calculates and returns the checksum
485 * @hw: pointer to hardware structure
486 * @checksum: pointer to the checksum
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000487 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000488 * This function calculates SW Checksum that covers the whole 64kB shadow RAM
489 * except the VPD and PCIe ALT Auto-load modules. The structure and size of VPD
490 * is customer specific and unknown. Therefore, this function skips all maximum
491 * possible size of VPD (1kB).
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000492 **/
493static i40e_status i40e_calc_nvm_checksum(struct i40e_hw *hw,
494 u16 *checksum)
495{
Jean Sacren0e5229c2015-10-13 01:06:31 -0600496 i40e_status ret_code;
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000497 struct i40e_virt_mem vmem;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000498 u16 pcie_alt_module = 0;
499 u16 checksum_local = 0;
500 u16 vpd_module = 0;
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000501 u16 *data;
502 u16 i = 0;
503
504 ret_code = i40e_allocate_virt_mem(hw, &vmem,
505 I40E_SR_SECTOR_SIZE_IN_WORDS * sizeof(u16));
506 if (ret_code)
507 goto i40e_calc_nvm_checksum_exit;
508 data = (u16 *)vmem.va;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000509
510 /* read pointer to VPD area */
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700511 ret_code = __i40e_read_nvm_word(hw, I40E_SR_VPD_PTR, &vpd_module);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000512 if (ret_code) {
513 ret_code = I40E_ERR_NVM_CHECKSUM;
514 goto i40e_calc_nvm_checksum_exit;
515 }
516
517 /* read pointer to PCIe Alt Auto-load module */
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700518 ret_code = __i40e_read_nvm_word(hw, I40E_SR_PCIE_ALT_AUTO_LOAD_PTR,
519 &pcie_alt_module);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000520 if (ret_code) {
521 ret_code = I40E_ERR_NVM_CHECKSUM;
522 goto i40e_calc_nvm_checksum_exit;
523 }
524
525 /* Calculate SW checksum that covers the whole 64kB shadow RAM
526 * except the VPD and PCIe ALT Auto-load modules
527 */
528 for (i = 0; i < hw->nvm.sr_size; i++) {
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000529 /* Read SR page */
530 if ((i % I40E_SR_SECTOR_SIZE_IN_WORDS) == 0) {
531 u16 words = I40E_SR_SECTOR_SIZE_IN_WORDS;
532
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700533 ret_code = __i40e_read_nvm_buffer(hw, i, &words, data);
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000534 if (ret_code) {
535 ret_code = I40E_ERR_NVM_CHECKSUM;
536 goto i40e_calc_nvm_checksum_exit;
537 }
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000538 }
539
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000540 /* Skip Checksum word */
541 if (i == I40E_SR_SW_CHECKSUM_WORD)
542 continue;
543 /* Skip VPD module (convert byte size to word count) */
544 if ((i >= (u32)vpd_module) &&
545 (i < ((u32)vpd_module +
546 (I40E_SR_VPD_MODULE_MAX_SIZE / 2)))) {
547 continue;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000548 }
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000549 /* Skip PCIe ALT module (convert byte size to word count) */
550 if ((i >= (u32)pcie_alt_module) &&
551 (i < ((u32)pcie_alt_module +
552 (I40E_SR_PCIE_ALT_MODULE_MAX_SIZE / 2)))) {
553 continue;
554 }
555
556 checksum_local += data[i % I40E_SR_SECTOR_SIZE_IN_WORDS];
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000557 }
558
559 *checksum = (u16)I40E_SR_SW_CHECKSUM_BASE - checksum_local;
560
561i40e_calc_nvm_checksum_exit:
Kamil Krawczykd1bbe0e2015-01-24 09:58:33 +0000562 i40e_free_virt_mem(hw, &vmem);
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000563 return ret_code;
564}
565
566/**
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000567 * i40e_update_nvm_checksum - Updates the NVM checksum
568 * @hw: pointer to hardware structure
569 *
570 * NVM ownership must be acquired before calling this function and released
571 * on ARQ completion event reception by caller.
572 * This function will commit SR to NVM.
573 **/
574i40e_status i40e_update_nvm_checksum(struct i40e_hw *hw)
575{
Jean Sacren0e5229c2015-10-13 01:06:31 -0600576 i40e_status ret_code;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000577 u16 checksum;
Jesse Brandeburgdd38c582015-08-26 15:14:18 -0400578 __le16 le_sum;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000579
580 ret_code = i40e_calc_nvm_checksum(hw, &checksum);
Jean Sacren2fc4cd52015-10-13 01:06:32 -0600581 if (!ret_code) {
582 le_sum = cpu_to_le16(checksum);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000583 ret_code = i40e_write_nvm_aq(hw, 0x00, I40E_SR_SW_CHECKSUM_WORD,
Jesse Brandeburgdd38c582015-08-26 15:14:18 -0400584 1, &le_sum, true);
Jean Sacren2fc4cd52015-10-13 01:06:32 -0600585 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000586
587 return ret_code;
588}
589
590/**
Shannon Nelson3e261862014-02-06 05:51:06 +0000591 * i40e_validate_nvm_checksum - Validate EEPROM checksum
592 * @hw: pointer to hardware structure
593 * @checksum: calculated checksum
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000594 *
Shannon Nelson3e261862014-02-06 05:51:06 +0000595 * Performs checksum calculation and validates the NVM SW checksum. If the
596 * caller does not need checksum, the value can be NULL.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000597 **/
598i40e_status i40e_validate_nvm_checksum(struct i40e_hw *hw,
599 u16 *checksum)
600{
601 i40e_status ret_code = 0;
602 u16 checksum_sr = 0;
Jesse Brandeburge15c9fa2014-01-17 15:36:31 -0800603 u16 checksum_local = 0;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000604
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700605 /* We must acquire the NVM lock in order to correctly synchronize the
606 * NVM accesses across multiple PFs. Without doing so it is possible
607 * for one of the PFs to read invalid data potentially indicating that
608 * the checksum is invalid.
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000609 */
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -0700610 ret_code = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
611 if (ret_code)
612 return ret_code;
613 ret_code = i40e_calc_nvm_checksum(hw, &checksum_local);
614 __i40e_read_nvm_word(hw, I40E_SR_SW_CHECKSUM_WORD, &checksum_sr);
615 i40e_release_nvm(hw);
616 if (ret_code)
617 return ret_code;
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000618
619 /* Verify read checksum from EEPROM is the same as
620 * calculated checksum
621 */
622 if (checksum_local != checksum_sr)
623 ret_code = I40E_ERR_NVM_CHECKSUM;
624
625 /* If the user cares, return the calculated checksum */
626 if (checksum)
627 *checksum = checksum_local;
628
Jesse Brandeburg56a62fc2013-09-11 08:40:12 +0000629 return ret_code;
630}
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000631
632static i40e_status i40e_nvmupd_state_init(struct i40e_hw *hw,
633 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400634 u8 *bytes, int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000635static i40e_status i40e_nvmupd_state_reading(struct i40e_hw *hw,
636 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400637 u8 *bytes, int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000638static i40e_status i40e_nvmupd_state_writing(struct i40e_hw *hw,
639 struct i40e_nvm_access *cmd,
640 u8 *bytes, int *errno);
641static enum i40e_nvmupd_cmd i40e_nvmupd_validate_command(struct i40e_hw *hw,
642 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400643 int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000644static i40e_status i40e_nvmupd_nvm_erase(struct i40e_hw *hw,
645 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400646 int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000647static i40e_status i40e_nvmupd_nvm_write(struct i40e_hw *hw,
648 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400649 u8 *bytes, int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000650static i40e_status i40e_nvmupd_nvm_read(struct i40e_hw *hw,
651 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400652 u8 *bytes, int *perrno);
Shannon Nelsone4c83c22015-08-28 17:55:50 -0400653static i40e_status i40e_nvmupd_exec_aq(struct i40e_hw *hw,
654 struct i40e_nvm_access *cmd,
655 u8 *bytes, int *perrno);
Shannon Nelsonb72dc7b2015-08-28 17:55:51 -0400656static i40e_status i40e_nvmupd_get_aq_result(struct i40e_hw *hw,
657 struct i40e_nvm_access *cmd,
658 u8 *bytes, int *perrno);
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500659static i40e_status i40e_nvmupd_get_aq_event(struct i40e_hw *hw,
660 struct i40e_nvm_access *cmd,
661 u8 *bytes, int *perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000662static inline u8 i40e_nvmupd_get_module(u32 val)
663{
664 return (u8)(val & I40E_NVM_MOD_PNT_MASK);
665}
666static inline u8 i40e_nvmupd_get_transaction(u32 val)
667{
668 return (u8)((val & I40E_NVM_TRANS_MASK) >> I40E_NVM_TRANS_SHIFT);
669}
670
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500671static inline u8 i40e_nvmupd_get_preservation_flags(u32 val)
672{
673 return (u8)((val & I40E_NVM_PRESERVATION_FLAGS_MASK) >>
674 I40E_NVM_PRESERVATION_FLAGS_SHIFT);
675}
676
Jingjing Wu4e68adfe2015-09-28 14:12:31 -0400677static const char * const i40e_nvm_update_state_str[] = {
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000678 "I40E_NVMUPD_INVALID",
679 "I40E_NVMUPD_READ_CON",
680 "I40E_NVMUPD_READ_SNT",
681 "I40E_NVMUPD_READ_LCB",
682 "I40E_NVMUPD_READ_SA",
683 "I40E_NVMUPD_WRITE_ERA",
684 "I40E_NVMUPD_WRITE_CON",
685 "I40E_NVMUPD_WRITE_SNT",
686 "I40E_NVMUPD_WRITE_LCB",
687 "I40E_NVMUPD_WRITE_SA",
688 "I40E_NVMUPD_CSUM_CON",
689 "I40E_NVMUPD_CSUM_SA",
690 "I40E_NVMUPD_CSUM_LCB",
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400691 "I40E_NVMUPD_STATUS",
Shannon Nelsone4c83c22015-08-28 17:55:50 -0400692 "I40E_NVMUPD_EXEC_AQ",
Shannon Nelsonb72dc7b2015-08-28 17:55:51 -0400693 "I40E_NVMUPD_GET_AQ_RESULT",
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500694 "I40E_NVMUPD_GET_AQ_EVENT",
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000695};
696
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000697/**
698 * i40e_nvmupd_command - Process an NVM update command
699 * @hw: pointer to hardware structure
700 * @cmd: pointer to nvm update command
701 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -0400702 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000703 *
704 * Dispatches command depending on what update state is current
705 **/
706i40e_status i40e_nvmupd_command(struct i40e_hw *hw,
707 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400708 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000709{
710 i40e_status status;
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400711 enum i40e_nvmupd_cmd upd_cmd;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000712
713 /* assume success */
Shannon Nelson79afe832015-07-23 16:54:33 -0400714 *perrno = 0;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000715
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400716 /* early check for status command and debug msgs */
717 upd_cmd = i40e_nvmupd_validate_command(hw, cmd, perrno);
718
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700719 i40e_debug(hw, I40E_DEBUG_NVM, "%s state %d nvm_release_on_hold %d opc 0x%04x cmd 0x%08x config 0x%08x offset 0x%08x data_size 0x%08x\n",
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400720 i40e_nvm_update_state_str[upd_cmd],
721 hw->nvmupd_state,
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700722 hw->nvm_release_on_done, hw->nvm_wait_opcode,
Shannon Nelson1d73b2d2015-12-23 12:05:51 -0800723 cmd->command, cmd->config, cmd->offset, cmd->data_size);
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400724
725 if (upd_cmd == I40E_NVMUPD_INVALID) {
726 *perrno = -EFAULT;
727 i40e_debug(hw, I40E_DEBUG_NVM,
728 "i40e_nvmupd_validate_command returns %d errno %d\n",
729 upd_cmd, *perrno);
730 }
731
732 /* a status request returns immediately rather than
733 * going into the state machine
734 */
735 if (upd_cmd == I40E_NVMUPD_STATUS) {
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700736 if (!cmd->data_size) {
737 *perrno = -EFAULT;
738 return I40E_ERR_BUF_TOO_SHORT;
739 }
740
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400741 bytes[0] = hw->nvmupd_state;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700742
743 if (cmd->data_size >= 4) {
744 bytes[1] = 0;
745 *((u16 *)&bytes[2]) = hw->nvm_wait_opcode;
746 }
747
Maciej Sosin81fa7c92016-10-11 15:26:57 -0700748 /* Clear error status on read */
749 if (hw->nvmupd_state == I40E_NVMUPD_STATE_ERROR)
750 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
751
Shannon Nelson0af8e9d2015-08-28 17:55:48 -0400752 return 0;
753 }
754
Maciej Sosin81fa7c92016-10-11 15:26:57 -0700755 /* Clear status even it is not read and log */
756 if (hw->nvmupd_state == I40E_NVMUPD_STATE_ERROR) {
757 i40e_debug(hw, I40E_DEBUG_NVM,
758 "Clearing I40E_NVMUPD_STATE_ERROR state without reading\n");
759 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
760 }
761
Sudheer Mogilappagari2bf01932017-07-12 05:46:07 -0400762 /* Acquire lock to prevent race condition where adminq_task
763 * can execute after i40e_nvmupd_nvm_read/write but before state
Sudheer Mogilappagari167d52e2017-08-27 15:07:47 -0700764 * variables (nvm_wait_opcode, nvm_release_on_done) are updated.
765 *
766 * During NVMUpdate, it is observed that lock could be held for
767 * ~5ms for most commands. However lock is held for ~60ms for
768 * NVMUPD_CSUM_LCB command.
Sudheer Mogilappagari2bf01932017-07-12 05:46:07 -0400769 */
770 mutex_lock(&hw->aq.arq_mutex);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000771 switch (hw->nvmupd_state) {
772 case I40E_NVMUPD_STATE_INIT:
Shannon Nelson79afe832015-07-23 16:54:33 -0400773 status = i40e_nvmupd_state_init(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000774 break;
775
776 case I40E_NVMUPD_STATE_READING:
Shannon Nelson79afe832015-07-23 16:54:33 -0400777 status = i40e_nvmupd_state_reading(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000778 break;
779
780 case I40E_NVMUPD_STATE_WRITING:
Shannon Nelson79afe832015-07-23 16:54:33 -0400781 status = i40e_nvmupd_state_writing(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000782 break;
783
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400784 case I40E_NVMUPD_STATE_INIT_WAIT:
785 case I40E_NVMUPD_STATE_WRITE_WAIT:
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700786 /* if we need to stop waiting for an event, clear
787 * the wait info and return before doing anything else
788 */
789 if (cmd->offset == 0xffff) {
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500790 i40e_nvmupd_clear_wait_state(hw);
Sudheer Mogilappagari167d52e2017-08-27 15:07:47 -0700791 status = 0;
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500792 break;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700793 }
794
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400795 status = I40E_ERR_NOT_READY;
796 *perrno = -EBUSY;
797 break;
798
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000799 default:
800 /* invalid state, should never happen */
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000801 i40e_debug(hw, I40E_DEBUG_NVM,
802 "NVMUPD: no such state %d\n", hw->nvmupd_state);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000803 status = I40E_NOT_SUPPORTED;
Shannon Nelson79afe832015-07-23 16:54:33 -0400804 *perrno = -ESRCH;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000805 break;
806 }
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500807
Sudheer Mogilappagari2bf01932017-07-12 05:46:07 -0400808 mutex_unlock(&hw->aq.arq_mutex);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000809 return status;
810}
811
812/**
813 * i40e_nvmupd_state_init - Handle NVM update state Init
814 * @hw: pointer to hardware structure
815 * @cmd: pointer to nvm update command buffer
816 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -0400817 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000818 *
819 * Process legitimate commands of the Init state and conditionally set next
820 * state. Reject all other commands.
821 **/
822static i40e_status i40e_nvmupd_state_init(struct i40e_hw *hw,
823 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400824 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000825{
826 i40e_status status = 0;
827 enum i40e_nvmupd_cmd upd_cmd;
828
Shannon Nelson79afe832015-07-23 16:54:33 -0400829 upd_cmd = i40e_nvmupd_validate_command(hw, cmd, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000830
831 switch (upd_cmd) {
832 case I40E_NVMUPD_READ_SA:
833 status = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
834 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400835 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000836 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000837 } else {
Shannon Nelson79afe832015-07-23 16:54:33 -0400838 status = i40e_nvmupd_nvm_read(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000839 i40e_release_nvm(hw);
840 }
841 break;
842
843 case I40E_NVMUPD_READ_SNT:
844 status = i40e_acquire_nvm(hw, I40E_RESOURCE_READ);
845 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400846 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000847 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000848 } else {
Shannon Nelson79afe832015-07-23 16:54:33 -0400849 status = i40e_nvmupd_nvm_read(hw, cmd, bytes, perrno);
Shannon Nelson0fdd0522014-11-13 08:23:20 +0000850 if (status)
851 i40e_release_nvm(hw);
852 else
853 hw->nvmupd_state = I40E_NVMUPD_STATE_READING;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000854 }
855 break;
856
857 case I40E_NVMUPD_WRITE_ERA:
858 status = i40e_acquire_nvm(hw, I40E_RESOURCE_WRITE);
859 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400860 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000861 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000862 } else {
Shannon Nelson79afe832015-07-23 16:54:33 -0400863 status = i40e_nvmupd_nvm_erase(hw, cmd, perrno);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400864 if (status) {
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000865 i40e_release_nvm(hw);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400866 } else {
Shannon Nelson437f82a2016-04-01 03:56:09 -0700867 hw->nvm_release_on_done = true;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700868 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_erase;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400869 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
870 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000871 }
872 break;
873
874 case I40E_NVMUPD_WRITE_SA:
875 status = i40e_acquire_nvm(hw, I40E_RESOURCE_WRITE);
876 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400877 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000878 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000879 } else {
Shannon Nelson79afe832015-07-23 16:54:33 -0400880 status = i40e_nvmupd_nvm_write(hw, cmd, bytes, perrno);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400881 if (status) {
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000882 i40e_release_nvm(hw);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400883 } else {
Shannon Nelson437f82a2016-04-01 03:56:09 -0700884 hw->nvm_release_on_done = true;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700885 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400886 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
887 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000888 }
889 break;
890
891 case I40E_NVMUPD_WRITE_SNT:
892 status = i40e_acquire_nvm(hw, I40E_RESOURCE_WRITE);
893 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400894 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000895 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000896 } else {
Shannon Nelson79afe832015-07-23 16:54:33 -0400897 status = i40e_nvmupd_nvm_write(hw, cmd, bytes, perrno);
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700898 if (status) {
Shannon Nelson0fdd0522014-11-13 08:23:20 +0000899 i40e_release_nvm(hw);
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700900 } else {
901 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400902 hw->nvmupd_state = I40E_NVMUPD_STATE_WRITE_WAIT;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700903 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000904 }
905 break;
906
907 case I40E_NVMUPD_CSUM_SA:
908 status = i40e_acquire_nvm(hw, I40E_RESOURCE_WRITE);
909 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400910 *perrno = i40e_aq_rc_to_posix(status,
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000911 hw->aq.asq_last_status);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000912 } else {
913 status = i40e_update_nvm_checksum(hw);
914 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -0400915 *perrno = hw->aq.asq_last_status ?
Shannon Nelsonbf848f32014-11-13 08:23:22 +0000916 i40e_aq_rc_to_posix(status,
917 hw->aq.asq_last_status) :
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000918 -EIO;
919 i40e_release_nvm(hw);
920 } else {
Shannon Nelson437f82a2016-04-01 03:56:09 -0700921 hw->nvm_release_on_done = true;
Shannon Nelsonfed2db92016-04-12 08:30:43 -0700922 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400923 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000924 }
925 }
926 break;
927
Shannon Nelsone4c83c22015-08-28 17:55:50 -0400928 case I40E_NVMUPD_EXEC_AQ:
929 status = i40e_nvmupd_exec_aq(hw, cmd, bytes, perrno);
930 break;
931
Shannon Nelsonb72dc7b2015-08-28 17:55:51 -0400932 case I40E_NVMUPD_GET_AQ_RESULT:
933 status = i40e_nvmupd_get_aq_result(hw, cmd, bytes, perrno);
934 break;
935
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -0500936 case I40E_NVMUPD_GET_AQ_EVENT:
937 status = i40e_nvmupd_get_aq_event(hw, cmd, bytes, perrno);
938 break;
939
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000940 default:
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000941 i40e_debug(hw, I40E_DEBUG_NVM,
942 "NVMUPD: bad cmd %s in init state\n",
943 i40e_nvm_update_state_str[upd_cmd]);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000944 status = I40E_ERR_NVM;
Shannon Nelson79afe832015-07-23 16:54:33 -0400945 *perrno = -ESRCH;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000946 break;
947 }
948 return status;
949}
950
951/**
952 * i40e_nvmupd_state_reading - Handle NVM update state Reading
953 * @hw: pointer to hardware structure
954 * @cmd: pointer to nvm update command buffer
955 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -0400956 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000957 *
958 * NVM ownership is already held. Process legitimate commands and set any
959 * change in state; reject all other commands.
960 **/
961static i40e_status i40e_nvmupd_state_reading(struct i40e_hw *hw,
962 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -0400963 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000964{
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -0400965 i40e_status status = 0;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000966 enum i40e_nvmupd_cmd upd_cmd;
967
Shannon Nelson79afe832015-07-23 16:54:33 -0400968 upd_cmd = i40e_nvmupd_validate_command(hw, cmd, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000969
970 switch (upd_cmd) {
971 case I40E_NVMUPD_READ_SA:
972 case I40E_NVMUPD_READ_CON:
Shannon Nelson79afe832015-07-23 16:54:33 -0400973 status = i40e_nvmupd_nvm_read(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000974 break;
975
976 case I40E_NVMUPD_READ_LCB:
Shannon Nelson79afe832015-07-23 16:54:33 -0400977 status = i40e_nvmupd_nvm_read(hw, cmd, bytes, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000978 i40e_release_nvm(hw);
979 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
980 break;
981
982 default:
Shannon Nelson74d0d0e2014-11-13 08:23:15 +0000983 i40e_debug(hw, I40E_DEBUG_NVM,
984 "NVMUPD: bad cmd %s in reading state.\n",
985 i40e_nvm_update_state_str[upd_cmd]);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000986 status = I40E_NOT_SUPPORTED;
Shannon Nelson79afe832015-07-23 16:54:33 -0400987 *perrno = -ESRCH;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000988 break;
989 }
990 return status;
991}
992
993/**
994 * i40e_nvmupd_state_writing - Handle NVM update state Writing
995 * @hw: pointer to hardware structure
996 * @cmd: pointer to nvm update command buffer
997 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -0400998 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +0000999 *
1000 * NVM ownership is already held. Process legitimate commands and set any
1001 * change in state; reject all other commands
1002 **/
1003static i40e_status i40e_nvmupd_state_writing(struct i40e_hw *hw,
1004 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -04001005 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001006{
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001007 i40e_status status = 0;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001008 enum i40e_nvmupd_cmd upd_cmd;
Shannon Nelson2c47e352015-02-21 06:45:10 +00001009 bool retry_attempt = false;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001010
Shannon Nelson79afe832015-07-23 16:54:33 -04001011 upd_cmd = i40e_nvmupd_validate_command(hw, cmd, perrno);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001012
Shannon Nelson2c47e352015-02-21 06:45:10 +00001013retry:
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001014 switch (upd_cmd) {
1015 case I40E_NVMUPD_WRITE_CON:
Shannon Nelson79afe832015-07-23 16:54:33 -04001016 status = i40e_nvmupd_nvm_write(hw, cmd, bytes, perrno);
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001017 if (!status) {
1018 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001019 hw->nvmupd_state = I40E_NVMUPD_STATE_WRITE_WAIT;
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001020 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001021 break;
1022
1023 case I40E_NVMUPD_WRITE_LCB:
Shannon Nelson79afe832015-07-23 16:54:33 -04001024 status = i40e_nvmupd_nvm_write(hw, cmd, bytes, perrno);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001025 if (status) {
1026 *perrno = hw->aq.asq_last_status ?
1027 i40e_aq_rc_to_posix(status,
1028 hw->aq.asq_last_status) :
1029 -EIO;
1030 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1031 } else {
Shannon Nelson437f82a2016-04-01 03:56:09 -07001032 hw->nvm_release_on_done = true;
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001033 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001034 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
1035 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001036 break;
1037
1038 case I40E_NVMUPD_CSUM_CON:
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -07001039 /* Assumes the caller has acquired the nvm */
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001040 status = i40e_update_nvm_checksum(hw);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001041 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -04001042 *perrno = hw->aq.asq_last_status ?
Shannon Nelsonbf848f32014-11-13 08:23:22 +00001043 i40e_aq_rc_to_posix(status,
1044 hw->aq.asq_last_status) :
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001045 -EIO;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001046 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001047 } else {
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001048 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001049 hw->nvmupd_state = I40E_NVMUPD_STATE_WRITE_WAIT;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001050 }
1051 break;
1052
Shannon Nelson0fdd0522014-11-13 08:23:20 +00001053 case I40E_NVMUPD_CSUM_LCB:
Anjali Singhai Jain09f79fd2017-09-01 13:42:49 -07001054 /* Assumes the caller has acquired the nvm */
Shannon Nelson0fdd0522014-11-13 08:23:20 +00001055 status = i40e_update_nvm_checksum(hw);
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001056 if (status) {
Shannon Nelson79afe832015-07-23 16:54:33 -04001057 *perrno = hw->aq.asq_last_status ?
Shannon Nelsonbf848f32014-11-13 08:23:22 +00001058 i40e_aq_rc_to_posix(status,
1059 hw->aq.asq_last_status) :
Shannon Nelson0fdd0522014-11-13 08:23:20 +00001060 -EIO;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001061 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1062 } else {
Shannon Nelson437f82a2016-04-01 03:56:09 -07001063 hw->nvm_release_on_done = true;
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001064 hw->nvm_wait_opcode = i40e_aqc_opc_nvm_update;
Shannon Nelson2f1b5bc2015-08-28 17:55:49 -04001065 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
1066 }
Shannon Nelson0fdd0522014-11-13 08:23:20 +00001067 break;
1068
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001069 default:
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001070 i40e_debug(hw, I40E_DEBUG_NVM,
1071 "NVMUPD: bad cmd %s in writing state.\n",
1072 i40e_nvm_update_state_str[upd_cmd]);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001073 status = I40E_NOT_SUPPORTED;
Shannon Nelson79afe832015-07-23 16:54:33 -04001074 *perrno = -ESRCH;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001075 break;
1076 }
Shannon Nelson2c47e352015-02-21 06:45:10 +00001077
1078 /* In some circumstances, a multi-write transaction takes longer
1079 * than the default 3 minute timeout on the write semaphore. If
1080 * the write failed with an EBUSY status, this is likely the problem,
1081 * so here we try to reacquire the semaphore then retry the write.
1082 * We only do one retry, then give up.
1083 */
1084 if (status && (hw->aq.asq_last_status == I40E_AQ_RC_EBUSY) &&
1085 !retry_attempt) {
1086 i40e_status old_status = status;
1087 u32 old_asq_status = hw->aq.asq_last_status;
1088 u32 gtime;
1089
1090 gtime = rd32(hw, I40E_GLVFGEN_TIMER);
1091 if (gtime >= hw->nvm.hw_semaphore_timeout) {
1092 i40e_debug(hw, I40E_DEBUG_ALL,
1093 "NVMUPD: write semaphore expired (%d >= %lld), retrying\n",
1094 gtime, hw->nvm.hw_semaphore_timeout);
1095 i40e_release_nvm(hw);
1096 status = i40e_acquire_nvm(hw, I40E_RESOURCE_WRITE);
1097 if (status) {
1098 i40e_debug(hw, I40E_DEBUG_ALL,
1099 "NVMUPD: write semaphore reacquire failed aq_err = %d\n",
1100 hw->aq.asq_last_status);
1101 status = old_status;
1102 hw->aq.asq_last_status = old_asq_status;
1103 } else {
1104 retry_attempt = true;
1105 goto retry;
1106 }
1107 }
1108 }
1109
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001110 return status;
1111}
1112
1113/**
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001114 * i40e_nvmupd_clear_wait_state - clear wait state on hw
1115 * @hw: pointer to the hardware structure
1116 **/
1117void i40e_nvmupd_clear_wait_state(struct i40e_hw *hw)
1118{
1119 i40e_debug(hw, I40E_DEBUG_NVM,
1120 "NVMUPD: clearing wait on opcode 0x%04x\n",
1121 hw->nvm_wait_opcode);
1122
1123 if (hw->nvm_release_on_done) {
1124 i40e_release_nvm(hw);
1125 hw->nvm_release_on_done = false;
1126 }
1127 hw->nvm_wait_opcode = 0;
1128
1129 if (hw->aq.arq_last_status) {
1130 hw->nvmupd_state = I40E_NVMUPD_STATE_ERROR;
1131 return;
1132 }
1133
1134 switch (hw->nvmupd_state) {
1135 case I40E_NVMUPD_STATE_INIT_WAIT:
1136 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1137 break;
1138
1139 case I40E_NVMUPD_STATE_WRITE_WAIT:
1140 hw->nvmupd_state = I40E_NVMUPD_STATE_WRITING;
1141 break;
1142
1143 default:
1144 break;
1145 }
1146}
1147
1148/**
Shannon Nelsonbab2fb62016-04-01 03:56:11 -07001149 * i40e_nvmupd_check_wait_event - handle NVM update operation events
1150 * @hw: pointer to the hardware structure
1151 * @opcode: the event that just happened
Jacob Kellerf5254422018-04-20 01:41:33 -07001152 * @desc: AdminQ descriptor
Shannon Nelsonbab2fb62016-04-01 03:56:11 -07001153 **/
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001154void i40e_nvmupd_check_wait_event(struct i40e_hw *hw, u16 opcode,
1155 struct i40e_aq_desc *desc)
Shannon Nelsonbab2fb62016-04-01 03:56:11 -07001156{
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001157 u32 aq_desc_len = sizeof(struct i40e_aq_desc);
1158
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001159 if (opcode == hw->nvm_wait_opcode) {
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001160 memcpy(&hw->nvm_aq_event_desc, desc, aq_desc_len);
1161 i40e_nvmupd_clear_wait_state(hw);
Shannon Nelsonbab2fb62016-04-01 03:56:11 -07001162 }
1163}
1164
1165/**
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001166 * i40e_nvmupd_validate_command - Validate given command
1167 * @hw: pointer to hardware structure
1168 * @cmd: pointer to nvm update command buffer
Shannon Nelson79afe832015-07-23 16:54:33 -04001169 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001170 *
1171 * Return one of the valid command types or I40E_NVMUPD_INVALID
1172 **/
1173static enum i40e_nvmupd_cmd i40e_nvmupd_validate_command(struct i40e_hw *hw,
1174 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -04001175 int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001176{
1177 enum i40e_nvmupd_cmd upd_cmd;
Shannon Nelson0af8e9d2015-08-28 17:55:48 -04001178 u8 module, transaction;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001179
1180 /* anything that doesn't match a recognized case is an error */
1181 upd_cmd = I40E_NVMUPD_INVALID;
1182
1183 transaction = i40e_nvmupd_get_transaction(cmd->config);
Shannon Nelson0af8e9d2015-08-28 17:55:48 -04001184 module = i40e_nvmupd_get_module(cmd->config);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001185
1186 /* limits on data size */
1187 if ((cmd->data_size < 1) ||
1188 (cmd->data_size > I40E_NVMUPD_MAX_DATA)) {
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001189 i40e_debug(hw, I40E_DEBUG_NVM,
1190 "i40e_nvmupd_validate_command data_size %d\n",
1191 cmd->data_size);
Shannon Nelson79afe832015-07-23 16:54:33 -04001192 *perrno = -EFAULT;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001193 return I40E_NVMUPD_INVALID;
1194 }
1195
1196 switch (cmd->command) {
1197 case I40E_NVM_READ:
1198 switch (transaction) {
1199 case I40E_NVM_CON:
1200 upd_cmd = I40E_NVMUPD_READ_CON;
1201 break;
1202 case I40E_NVM_SNT:
1203 upd_cmd = I40E_NVMUPD_READ_SNT;
1204 break;
1205 case I40E_NVM_LCB:
1206 upd_cmd = I40E_NVMUPD_READ_LCB;
1207 break;
1208 case I40E_NVM_SA:
1209 upd_cmd = I40E_NVMUPD_READ_SA;
1210 break;
Shannon Nelson0af8e9d2015-08-28 17:55:48 -04001211 case I40E_NVM_EXEC:
1212 if (module == 0xf)
1213 upd_cmd = I40E_NVMUPD_STATUS;
Shannon Nelsonb72dc7b2015-08-28 17:55:51 -04001214 else if (module == 0)
1215 upd_cmd = I40E_NVMUPD_GET_AQ_RESULT;
Shannon Nelson0af8e9d2015-08-28 17:55:48 -04001216 break;
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001217 case I40E_NVM_AQE:
1218 upd_cmd = I40E_NVMUPD_GET_AQ_EVENT;
1219 break;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001220 }
1221 break;
1222
1223 case I40E_NVM_WRITE:
1224 switch (transaction) {
1225 case I40E_NVM_CON:
1226 upd_cmd = I40E_NVMUPD_WRITE_CON;
1227 break;
1228 case I40E_NVM_SNT:
1229 upd_cmd = I40E_NVMUPD_WRITE_SNT;
1230 break;
1231 case I40E_NVM_LCB:
1232 upd_cmd = I40E_NVMUPD_WRITE_LCB;
1233 break;
1234 case I40E_NVM_SA:
1235 upd_cmd = I40E_NVMUPD_WRITE_SA;
1236 break;
1237 case I40E_NVM_ERA:
1238 upd_cmd = I40E_NVMUPD_WRITE_ERA;
1239 break;
1240 case I40E_NVM_CSUM:
1241 upd_cmd = I40E_NVMUPD_CSUM_CON;
1242 break;
1243 case (I40E_NVM_CSUM|I40E_NVM_SA):
1244 upd_cmd = I40E_NVMUPD_CSUM_SA;
1245 break;
1246 case (I40E_NVM_CSUM|I40E_NVM_LCB):
1247 upd_cmd = I40E_NVMUPD_CSUM_LCB;
1248 break;
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001249 case I40E_NVM_EXEC:
1250 if (module == 0)
1251 upd_cmd = I40E_NVMUPD_EXEC_AQ;
1252 break;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001253 }
1254 break;
1255 }
1256
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001257 return upd_cmd;
1258}
1259
1260/**
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001261 * i40e_nvmupd_exec_aq - Run an AQ command
1262 * @hw: pointer to hardware structure
1263 * @cmd: pointer to nvm update command buffer
1264 * @bytes: pointer to the data buffer
1265 * @perrno: pointer to return error code
1266 *
1267 * cmd structure contains identifiers and data buffer
1268 **/
1269static i40e_status i40e_nvmupd_exec_aq(struct i40e_hw *hw,
1270 struct i40e_nvm_access *cmd,
1271 u8 *bytes, int *perrno)
1272{
1273 struct i40e_asq_cmd_details cmd_details;
1274 i40e_status status;
1275 struct i40e_aq_desc *aq_desc;
1276 u32 buff_size = 0;
1277 u8 *buff = NULL;
1278 u32 aq_desc_len;
1279 u32 aq_data_len;
1280
1281 i40e_debug(hw, I40E_DEBUG_NVM, "NVMUPD: %s\n", __func__);
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001282 if (cmd->offset == 0xffff)
1283 return 0;
1284
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001285 memset(&cmd_details, 0, sizeof(cmd_details));
1286 cmd_details.wb_desc = &hw->nvm_wb_desc;
1287
1288 aq_desc_len = sizeof(struct i40e_aq_desc);
1289 memset(&hw->nvm_wb_desc, 0, aq_desc_len);
1290
1291 /* get the aq descriptor */
1292 if (cmd->data_size < aq_desc_len) {
1293 i40e_debug(hw, I40E_DEBUG_NVM,
1294 "NVMUPD: not enough aq desc bytes for exec, size %d < %d\n",
1295 cmd->data_size, aq_desc_len);
1296 *perrno = -EINVAL;
1297 return I40E_ERR_PARAM;
1298 }
1299 aq_desc = (struct i40e_aq_desc *)bytes;
1300
1301 /* if data buffer needed, make sure it's ready */
1302 aq_data_len = cmd->data_size - aq_desc_len;
1303 buff_size = max_t(u32, aq_data_len, le16_to_cpu(aq_desc->datalen));
1304 if (buff_size) {
1305 if (!hw->nvm_buff.va) {
1306 status = i40e_allocate_virt_mem(hw, &hw->nvm_buff,
1307 hw->aq.asq_buf_size);
1308 if (status)
1309 i40e_debug(hw, I40E_DEBUG_NVM,
1310 "NVMUPD: i40e_allocate_virt_mem for exec buff failed, %d\n",
1311 status);
1312 }
1313
1314 if (hw->nvm_buff.va) {
1315 buff = hw->nvm_buff.va;
1316 memcpy(buff, &bytes[aq_desc_len], aq_data_len);
1317 }
1318 }
1319
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001320 if (cmd->offset)
1321 memset(&hw->nvm_aq_event_desc, 0, aq_desc_len);
1322
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001323 /* and away we go! */
1324 status = i40e_asq_send_command(hw, aq_desc, buff,
1325 buff_size, &cmd_details);
1326 if (status) {
1327 i40e_debug(hw, I40E_DEBUG_NVM,
1328 "i40e_nvmupd_exec_aq err %s aq_err %s\n",
1329 i40e_stat_str(hw, status),
1330 i40e_aq_str(hw, hw->aq.asq_last_status));
1331 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001332 return status;
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001333 }
1334
Shannon Nelsonfed2db92016-04-12 08:30:43 -07001335 /* should we wait for a followup event? */
1336 if (cmd->offset) {
1337 hw->nvm_wait_opcode = cmd->offset;
1338 hw->nvmupd_state = I40E_NVMUPD_STATE_INIT_WAIT;
1339 }
1340
Shannon Nelsone4c83c22015-08-28 17:55:50 -04001341 return status;
1342}
1343
1344/**
Shannon Nelsonb72dc7b2015-08-28 17:55:51 -04001345 * i40e_nvmupd_get_aq_result - Get the results from the previous exec_aq
1346 * @hw: pointer to hardware structure
1347 * @cmd: pointer to nvm update command buffer
1348 * @bytes: pointer to the data buffer
1349 * @perrno: pointer to return error code
1350 *
1351 * cmd structure contains identifiers and data buffer
1352 **/
1353static i40e_status i40e_nvmupd_get_aq_result(struct i40e_hw *hw,
1354 struct i40e_nvm_access *cmd,
1355 u8 *bytes, int *perrno)
1356{
1357 u32 aq_total_len;
1358 u32 aq_desc_len;
1359 int remainder;
1360 u8 *buff;
1361
1362 i40e_debug(hw, I40E_DEBUG_NVM, "NVMUPD: %s\n", __func__);
1363
1364 aq_desc_len = sizeof(struct i40e_aq_desc);
1365 aq_total_len = aq_desc_len + le16_to_cpu(hw->nvm_wb_desc.datalen);
1366
1367 /* check offset range */
1368 if (cmd->offset > aq_total_len) {
1369 i40e_debug(hw, I40E_DEBUG_NVM, "%s: offset too big %d > %d\n",
1370 __func__, cmd->offset, aq_total_len);
1371 *perrno = -EINVAL;
1372 return I40E_ERR_PARAM;
1373 }
1374
1375 /* check copylength range */
1376 if (cmd->data_size > (aq_total_len - cmd->offset)) {
1377 int new_len = aq_total_len - cmd->offset;
1378
1379 i40e_debug(hw, I40E_DEBUG_NVM, "%s: copy length %d too big, trimming to %d\n",
1380 __func__, cmd->data_size, new_len);
1381 cmd->data_size = new_len;
1382 }
1383
1384 remainder = cmd->data_size;
1385 if (cmd->offset < aq_desc_len) {
1386 u32 len = aq_desc_len - cmd->offset;
1387
1388 len = min(len, cmd->data_size);
1389 i40e_debug(hw, I40E_DEBUG_NVM, "%s: aq_desc bytes %d to %d\n",
1390 __func__, cmd->offset, cmd->offset + len);
1391
1392 buff = ((u8 *)&hw->nvm_wb_desc) + cmd->offset;
1393 memcpy(bytes, buff, len);
1394
1395 bytes += len;
1396 remainder -= len;
1397 buff = hw->nvm_buff.va;
1398 } else {
1399 buff = hw->nvm_buff.va + (cmd->offset - aq_desc_len);
1400 }
1401
1402 if (remainder > 0) {
1403 int start_byte = buff - (u8 *)hw->nvm_buff.va;
1404
1405 i40e_debug(hw, I40E_DEBUG_NVM, "%s: databuf bytes %d to %d\n",
1406 __func__, start_byte, start_byte + remainder);
1407 memcpy(bytes, buff, remainder);
1408 }
1409
1410 return 0;
1411}
1412
1413/**
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001414 * i40e_nvmupd_get_aq_event - Get the Admin Queue event from previous exec_aq
1415 * @hw: pointer to hardware structure
1416 * @cmd: pointer to nvm update command buffer
1417 * @bytes: pointer to the data buffer
1418 * @perrno: pointer to return error code
1419 *
1420 * cmd structure contains identifiers and data buffer
1421 **/
1422static i40e_status i40e_nvmupd_get_aq_event(struct i40e_hw *hw,
1423 struct i40e_nvm_access *cmd,
1424 u8 *bytes, int *perrno)
1425{
1426 u32 aq_total_len;
1427 u32 aq_desc_len;
1428
1429 i40e_debug(hw, I40E_DEBUG_NVM, "NVMUPD: %s\n", __func__);
1430
1431 aq_desc_len = sizeof(struct i40e_aq_desc);
1432 aq_total_len = aq_desc_len + le16_to_cpu(hw->nvm_aq_event_desc.datalen);
1433
1434 /* check copylength range */
1435 if (cmd->data_size > aq_total_len) {
1436 i40e_debug(hw, I40E_DEBUG_NVM,
1437 "%s: copy length %d too big, trimming to %d\n",
1438 __func__, cmd->data_size, aq_total_len);
1439 cmd->data_size = aq_total_len;
1440 }
1441
1442 memcpy(bytes, &hw->nvm_aq_event_desc, cmd->data_size);
1443
1444 return 0;
1445}
1446
1447/**
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001448 * i40e_nvmupd_nvm_read - Read NVM
1449 * @hw: pointer to hardware structure
1450 * @cmd: pointer to nvm update command buffer
1451 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -04001452 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001453 *
1454 * cmd structure contains identifiers and data buffer
1455 **/
1456static i40e_status i40e_nvmupd_nvm_read(struct i40e_hw *hw,
1457 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -04001458 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001459{
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001460 struct i40e_asq_cmd_details cmd_details;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001461 i40e_status status;
1462 u8 module, transaction;
1463 bool last;
1464
1465 transaction = i40e_nvmupd_get_transaction(cmd->config);
1466 module = i40e_nvmupd_get_module(cmd->config);
1467 last = (transaction == I40E_NVM_LCB) || (transaction == I40E_NVM_SA);
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001468
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001469 memset(&cmd_details, 0, sizeof(cmd_details));
1470 cmd_details.wb_desc = &hw->nvm_wb_desc;
1471
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001472 status = i40e_aq_read_nvm(hw, module, cmd->offset, (u16)cmd->data_size,
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001473 bytes, last, &cmd_details);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001474 if (status) {
1475 i40e_debug(hw, I40E_DEBUG_NVM,
1476 "i40e_nvmupd_nvm_read mod 0x%x off 0x%x len 0x%x\n",
1477 module, cmd->offset, cmd->data_size);
1478 i40e_debug(hw, I40E_DEBUG_NVM,
1479 "i40e_nvmupd_nvm_read status %d aq %d\n",
1480 status, hw->aq.asq_last_status);
Shannon Nelson79afe832015-07-23 16:54:33 -04001481 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001482 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001483
1484 return status;
1485}
1486
1487/**
1488 * i40e_nvmupd_nvm_erase - Erase an NVM module
1489 * @hw: pointer to hardware structure
1490 * @cmd: pointer to nvm update command buffer
Shannon Nelson79afe832015-07-23 16:54:33 -04001491 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001492 *
1493 * module, offset, data_size and data are in cmd structure
1494 **/
1495static i40e_status i40e_nvmupd_nvm_erase(struct i40e_hw *hw,
1496 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -04001497 int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001498{
1499 i40e_status status = 0;
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001500 struct i40e_asq_cmd_details cmd_details;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001501 u8 module, transaction;
1502 bool last;
1503
1504 transaction = i40e_nvmupd_get_transaction(cmd->config);
1505 module = i40e_nvmupd_get_module(cmd->config);
1506 last = (transaction & I40E_NVM_LCB);
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001507
1508 memset(&cmd_details, 0, sizeof(cmd_details));
1509 cmd_details.wb_desc = &hw->nvm_wb_desc;
1510
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001511 status = i40e_aq_erase_nvm(hw, module, cmd->offset, (u16)cmd->data_size,
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001512 last, &cmd_details);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001513 if (status) {
1514 i40e_debug(hw, I40E_DEBUG_NVM,
1515 "i40e_nvmupd_nvm_erase mod 0x%x off 0x%x len 0x%x\n",
1516 module, cmd->offset, cmd->data_size);
1517 i40e_debug(hw, I40E_DEBUG_NVM,
1518 "i40e_nvmupd_nvm_erase status %d aq %d\n",
1519 status, hw->aq.asq_last_status);
Shannon Nelson79afe832015-07-23 16:54:33 -04001520 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001521 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001522
1523 return status;
1524}
1525
1526/**
1527 * i40e_nvmupd_nvm_write - Write NVM
1528 * @hw: pointer to hardware structure
1529 * @cmd: pointer to nvm update command buffer
1530 * @bytes: pointer to the data buffer
Shannon Nelson79afe832015-07-23 16:54:33 -04001531 * @perrno: pointer to return error code
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001532 *
1533 * module, offset, data_size and data are in cmd structure
1534 **/
1535static i40e_status i40e_nvmupd_nvm_write(struct i40e_hw *hw,
1536 struct i40e_nvm_access *cmd,
Shannon Nelson79afe832015-07-23 16:54:33 -04001537 u8 *bytes, int *perrno)
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001538{
1539 i40e_status status = 0;
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001540 struct i40e_asq_cmd_details cmd_details;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001541 u8 module, transaction;
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001542 u8 preservation_flags;
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001543 bool last;
1544
1545 transaction = i40e_nvmupd_get_transaction(cmd->config);
1546 module = i40e_nvmupd_get_module(cmd->config);
1547 last = (transaction & I40E_NVM_LCB);
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001548 preservation_flags = i40e_nvmupd_get_preservation_flags(cmd->config);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001549
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001550 memset(&cmd_details, 0, sizeof(cmd_details));
1551 cmd_details.wb_desc = &hw->nvm_wb_desc;
1552
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001553 status = i40e_aq_update_nvm(hw, module, cmd->offset,
Shannon Nelson6b5c1b82015-08-28 17:55:47 -04001554 (u16)cmd->data_size, bytes, last,
Pawel Jablonskie3a5d6e2017-12-18 05:14:44 -05001555 preservation_flags, &cmd_details);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001556 if (status) {
1557 i40e_debug(hw, I40E_DEBUG_NVM,
1558 "i40e_nvmupd_nvm_write mod 0x%x off 0x%x len 0x%x\n",
1559 module, cmd->offset, cmd->data_size);
1560 i40e_debug(hw, I40E_DEBUG_NVM,
1561 "i40e_nvmupd_nvm_write status %d aq %d\n",
1562 status, hw->aq.asq_last_status);
Shannon Nelson79afe832015-07-23 16:54:33 -04001563 *perrno = i40e_aq_rc_to_posix(status, hw->aq.asq_last_status);
Shannon Nelson74d0d0e2014-11-13 08:23:15 +00001564 }
Shannon Nelsoncd552cb2014-07-09 07:46:09 +00001565
1566 return status;
1567}