blob: bf512ac75b9e640f058cf1d1b8ddbd2d9490134a [file] [log] [blame]
Stefan Weinhuber20c64462006-03-24 03:15:25 -08001/*
2 * Character device driver for extended error reporting.
3 *
4 * Copyright (C) 2005 IBM Corporation
5 * extended error reporting for DASD ECKD devices
6 * Author(s): Stefan Weinhuber <wein@de.ibm.com>
7 */
8
9#include <linux/init.h>
10#include <linux/fs.h>
11#include <linux/kernel.h>
12#include <linux/miscdevice.h>
13#include <linux/module.h>
14#include <linux/moduleparam.h>
15#include <linux/device.h>
16#include <linux/poll.h>
Christoph Hellwig3006d7c2007-05-31 17:38:02 +020017#include <linux/mutex.h>
Arnd Bergmann986837b2008-05-20 19:15:38 +020018#include <linux/smp_lock.h>
Julien Brunel0983e562008-08-21 19:46:30 +020019#include <linux/err.h>
Stefan Weinhuber20c64462006-03-24 03:15:25 -080020
21#include <asm/uaccess.h>
Stefan Weinhuber20c64462006-03-24 03:15:25 -080022#include <asm/atomic.h>
23#include <asm/ebcdic.h>
24
25#include "dasd_int.h"
26#include "dasd_eckd.h"
27
28#ifdef PRINTK_HEADER
29#undef PRINTK_HEADER
30#endif /* PRINTK_HEADER */
31#define PRINTK_HEADER "dasd(eer):"
32
33/*
34 * SECTION: the internal buffer
35 */
36
37/*
38 * The internal buffer is meant to store obaque blobs of data, so it does
39 * not know of higher level concepts like triggers.
40 * It consists of a number of pages that are used as a ringbuffer. Each data
41 * blob is stored in a simple record that consists of an integer, which
42 * contains the size of the following data, and the data bytes themselfes.
43 *
44 * To allow for multiple independent readers we create one internal buffer
45 * each time the device is opened and destroy the buffer when the file is
46 * closed again. The number of pages used for this buffer is determined by
47 * the module parmeter eer_pages.
48 *
49 * One record can be written to a buffer by using the functions
50 * - dasd_eer_start_record (one time per record to write the size to the
51 * buffer and reserve the space for the data)
52 * - dasd_eer_write_buffer (one or more times per record to write the data)
53 * The data can be written in several steps but you will have to compute
54 * the total size up front for the invocation of dasd_eer_start_record.
55 * If the ringbuffer is full, dasd_eer_start_record will remove the required
56 * number of old records.
57 *
58 * A record is typically read in two steps, first read the integer that
59 * specifies the size of the following data, then read the data.
60 * Both can be done by
61 * - dasd_eer_read_buffer
62 *
63 * For all mentioned functions you need to get the bufferlock first and keep
64 * it until a complete record is written or read.
65 *
66 * All information necessary to keep track of an internal buffer is kept in
67 * a struct eerbuffer. The buffer specific to a file pointer is strored in
68 * the private_data field of that file. To be able to write data to all
69 * existing buffers, each buffer is also added to the bufferlist.
70 * If the user does not want to read a complete record in one go, we have to
71 * keep track of the rest of the record. residual stores the number of bytes
72 * that are still to deliver. If the rest of the record is invalidated between
73 * two reads then residual will be set to -1 so that the next read will fail.
74 * All entries in the eerbuffer structure are protected with the bufferlock.
75 * To avoid races between writing to a buffer on the one side and creating
76 * and destroying buffers on the other side, the bufferlock must also be used
77 * to protect the bufferlist.
78 */
79
80static int eer_pages = 5;
81module_param(eer_pages, int, S_IRUGO|S_IWUSR);
82
83struct eerbuffer {
84 struct list_head list;
85 char **buffer;
86 int buffersize;
87 int buffer_page_count;
88 int head;
89 int tail;
90 int residual;
91};
92
93static LIST_HEAD(bufferlist);
Ingo Molnar34af9462006-06-27 02:53:55 -070094static DEFINE_SPINLOCK(bufferlock);
Stefan Weinhuber20c64462006-03-24 03:15:25 -080095static DECLARE_WAIT_QUEUE_HEAD(dasd_eer_read_wait_queue);
96
97/*
98 * How many free bytes are available on the buffer.
99 * Needs to be called with bufferlock held.
100 */
101static int dasd_eer_get_free_bytes(struct eerbuffer *eerb)
102{
103 if (eerb->head < eerb->tail)
104 return eerb->tail - eerb->head - 1;
105 return eerb->buffersize - eerb->head + eerb->tail -1;
106}
107
108/*
109 * How many bytes of buffer space are used.
110 * Needs to be called with bufferlock held.
111 */
112static int dasd_eer_get_filled_bytes(struct eerbuffer *eerb)
113{
114
115 if (eerb->head >= eerb->tail)
116 return eerb->head - eerb->tail;
117 return eerb->buffersize - eerb->tail + eerb->head;
118}
119
120/*
121 * The dasd_eer_write_buffer function just copies count bytes of data
122 * to the buffer. Make sure to call dasd_eer_start_record first, to
123 * make sure that enough free space is available.
124 * Needs to be called with bufferlock held.
125 */
126static void dasd_eer_write_buffer(struct eerbuffer *eerb,
127 char *data, int count)
128{
129
130 unsigned long headindex,localhead;
131 unsigned long rest, len;
132 char *nextdata;
133
134 nextdata = data;
135 rest = count;
136 while (rest > 0) {
137 headindex = eerb->head / PAGE_SIZE;
138 localhead = eerb->head % PAGE_SIZE;
139 len = min(rest, PAGE_SIZE - localhead);
140 memcpy(eerb->buffer[headindex]+localhead, nextdata, len);
141 nextdata += len;
142 rest -= len;
143 eerb->head += len;
144 if (eerb->head == eerb->buffersize)
145 eerb->head = 0; /* wrap around */
146 BUG_ON(eerb->head > eerb->buffersize);
147 }
148}
149
150/*
151 * Needs to be called with bufferlock held.
152 */
153static int dasd_eer_read_buffer(struct eerbuffer *eerb, char *data, int count)
154{
155
156 unsigned long tailindex,localtail;
157 unsigned long rest, len, finalcount;
158 char *nextdata;
159
160 finalcount = min(count, dasd_eer_get_filled_bytes(eerb));
161 nextdata = data;
162 rest = finalcount;
163 while (rest > 0) {
164 tailindex = eerb->tail / PAGE_SIZE;
165 localtail = eerb->tail % PAGE_SIZE;
166 len = min(rest, PAGE_SIZE - localtail);
167 memcpy(nextdata, eerb->buffer[tailindex] + localtail, len);
168 nextdata += len;
169 rest -= len;
170 eerb->tail += len;
171 if (eerb->tail == eerb->buffersize)
172 eerb->tail = 0; /* wrap around */
173 BUG_ON(eerb->tail > eerb->buffersize);
174 }
175 return finalcount;
176}
177
178/*
179 * Whenever you want to write a blob of data to the internal buffer you
180 * have to start by using this function first. It will write the number
181 * of bytes that will be written to the buffer. If necessary it will remove
182 * old records to make room for the new one.
183 * Needs to be called with bufferlock held.
184 */
185static int dasd_eer_start_record(struct eerbuffer *eerb, int count)
186{
187 int tailcount;
188
189 if (count + sizeof(count) > eerb->buffersize)
190 return -ENOMEM;
191 while (dasd_eer_get_free_bytes(eerb) < count + sizeof(count)) {
192 if (eerb->residual > 0) {
193 eerb->tail += eerb->residual;
194 if (eerb->tail >= eerb->buffersize)
195 eerb->tail -= eerb->buffersize;
196 eerb->residual = -1;
197 }
198 dasd_eer_read_buffer(eerb, (char *) &tailcount,
199 sizeof(tailcount));
200 eerb->tail += tailcount;
201 if (eerb->tail >= eerb->buffersize)
202 eerb->tail -= eerb->buffersize;
203 }
204 dasd_eer_write_buffer(eerb, (char*) &count, sizeof(count));
205
206 return 0;
207};
208
209/*
210 * Release pages that are not used anymore.
211 */
212static void dasd_eer_free_buffer_pages(char **buf, int no_pages)
213{
214 int i;
215
216 for (i = 0; i < no_pages; i++)
217 free_page((unsigned long) buf[i]);
218}
219
220/*
221 * Allocate a new set of memory pages.
222 */
223static int dasd_eer_allocate_buffer_pages(char **buf, int no_pages)
224{
225 int i;
226
227 for (i = 0; i < no_pages; i++) {
228 buf[i] = (char *) get_zeroed_page(GFP_KERNEL);
229 if (!buf[i]) {
230 dasd_eer_free_buffer_pages(buf, i);
231 return -ENOMEM;
232 }
233 }
234 return 0;
235}
236
237/*
238 * SECTION: The extended error reporting functionality
239 */
240
241/*
242 * When a DASD device driver wants to report an error, it calls the
243 * function dasd_eer_write and gives the respective trigger ID as
244 * parameter. Currently there are four kinds of triggers:
245 *
246 * DASD_EER_FATALERROR: all kinds of unrecoverable I/O problems
247 * DASD_EER_PPRCSUSPEND: PPRC was suspended
248 * DASD_EER_NOPATH: There is no path to the device left.
249 * DASD_EER_STATECHANGE: The state of the device has changed.
250 *
251 * For the first three triggers all required information can be supplied by
252 * the caller. For these triggers a record is written by the function
253 * dasd_eer_write_standard_trigger.
254 *
255 * The DASD_EER_STATECHANGE trigger is special since a sense subsystem
256 * status ccw need to be executed to gather the necessary sense data first.
257 * The dasd_eer_snss function will queue the SNSS request and the request
258 * callback will then call dasd_eer_write with the DASD_EER_STATCHANGE
259 * trigger.
260 *
261 * To avoid memory allocations at runtime, the necessary memory is allocated
262 * when the extended error reporting is enabled for a device (by
263 * dasd_eer_probe). There is one sense subsystem status request for each
264 * eer enabled DASD device. The presence of the cqr in device->eer_cqr
265 * indicates that eer is enable for the device. The use of the snss request
266 * is protected by the DASD_FLAG_EER_IN_USE bit. When this flag indicates
267 * that the cqr is currently in use, dasd_eer_snss cannot start a second
268 * request but sets the DASD_FLAG_EER_SNSS flag instead. The callback of
269 * the SNSS request will check the bit and call dasd_eer_snss again.
270 */
271
272#define SNSS_DATA_SIZE 44
273
274#define DASD_EER_BUSID_SIZE 10
275struct dasd_eer_header {
276 __u32 total_size;
277 __u32 trigger;
278 __u64 tv_sec;
279 __u64 tv_usec;
280 char busid[DASD_EER_BUSID_SIZE];
Stefan Weinhuber774fc4e2006-06-29 15:02:59 +0200281} __attribute__ ((packed));
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800282
283/*
284 * The following function can be used for those triggers that have
285 * all necessary data available when the function is called.
286 * If the parameter cqr is not NULL, the chain of requests will be searched
287 * for valid sense data, and all valid sense data sets will be added to
288 * the triggers data.
289 */
290static void dasd_eer_write_standard_trigger(struct dasd_device *device,
291 struct dasd_ccw_req *cqr,
292 int trigger)
293{
294 struct dasd_ccw_req *temp_cqr;
295 int data_size;
296 struct timeval tv;
297 struct dasd_eer_header header;
298 unsigned long flags;
299 struct eerbuffer *eerb;
300
301 /* go through cqr chain and count the valid sense data sets */
302 data_size = 0;
303 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers)
304 if (temp_cqr->irb.esw.esw0.erw.cons)
305 data_size += 32;
306
307 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */
308 header.trigger = trigger;
309 do_gettimeofday(&tv);
310 header.tv_sec = tv.tv_sec;
311 header.tv_usec = tv.tv_usec;
312 strncpy(header.busid, device->cdev->dev.bus_id, DASD_EER_BUSID_SIZE);
313
314 spin_lock_irqsave(&bufferlock, flags);
315 list_for_each_entry(eerb, &bufferlist, list) {
316 dasd_eer_start_record(eerb, header.total_size);
317 dasd_eer_write_buffer(eerb, (char *) &header, sizeof(header));
318 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers)
319 if (temp_cqr->irb.esw.esw0.erw.cons)
320 dasd_eer_write_buffer(eerb, cqr->irb.ecw, 32);
321 dasd_eer_write_buffer(eerb, "EOR", 4);
322 }
323 spin_unlock_irqrestore(&bufferlock, flags);
324 wake_up_interruptible(&dasd_eer_read_wait_queue);
325}
326
327/*
328 * This function writes a DASD_EER_STATECHANGE trigger.
329 */
330static void dasd_eer_write_snss_trigger(struct dasd_device *device,
331 struct dasd_ccw_req *cqr,
332 int trigger)
333{
334 int data_size;
335 int snss_rc;
336 struct timeval tv;
337 struct dasd_eer_header header;
338 unsigned long flags;
339 struct eerbuffer *eerb;
340
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100341 snss_rc = (cqr->status == DASD_CQR_DONE) ? 0 : -EIO;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800342 if (snss_rc)
343 data_size = 0;
344 else
345 data_size = SNSS_DATA_SIZE;
346
347 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */
348 header.trigger = DASD_EER_STATECHANGE;
349 do_gettimeofday(&tv);
350 header.tv_sec = tv.tv_sec;
351 header.tv_usec = tv.tv_usec;
352 strncpy(header.busid, device->cdev->dev.bus_id, DASD_EER_BUSID_SIZE);
353
354 spin_lock_irqsave(&bufferlock, flags);
355 list_for_each_entry(eerb, &bufferlist, list) {
356 dasd_eer_start_record(eerb, header.total_size);
357 dasd_eer_write_buffer(eerb, (char *) &header , sizeof(header));
358 if (!snss_rc)
359 dasd_eer_write_buffer(eerb, cqr->data, SNSS_DATA_SIZE);
360 dasd_eer_write_buffer(eerb, "EOR", 4);
361 }
362 spin_unlock_irqrestore(&bufferlock, flags);
363 wake_up_interruptible(&dasd_eer_read_wait_queue);
364}
365
366/*
367 * This function is called for all triggers. It calls the appropriate
368 * function that writes the actual trigger records.
369 */
370void dasd_eer_write(struct dasd_device *device, struct dasd_ccw_req *cqr,
371 unsigned int id)
372{
373 if (!device->eer_cqr)
374 return;
375 switch (id) {
376 case DASD_EER_FATALERROR:
377 case DASD_EER_PPRCSUSPEND:
378 dasd_eer_write_standard_trigger(device, cqr, id);
379 break;
380 case DASD_EER_NOPATH:
381 dasd_eer_write_standard_trigger(device, NULL, id);
382 break;
383 case DASD_EER_STATECHANGE:
384 dasd_eer_write_snss_trigger(device, cqr, id);
385 break;
386 default: /* unknown trigger, so we write it without any sense data */
387 dasd_eer_write_standard_trigger(device, NULL, id);
388 break;
389 }
390}
391EXPORT_SYMBOL(dasd_eer_write);
392
393/*
394 * Start a sense subsystem status request.
395 * Needs to be called with the device held.
396 */
397void dasd_eer_snss(struct dasd_device *device)
398{
399 struct dasd_ccw_req *cqr;
400
401 cqr = device->eer_cqr;
402 if (!cqr) /* Device not eer enabled. */
403 return;
404 if (test_and_set_bit(DASD_FLAG_EER_IN_USE, &device->flags)) {
405 /* Sense subsystem status request in use. */
406 set_bit(DASD_FLAG_EER_SNSS, &device->flags);
407 return;
408 }
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100409 /* cdev is already locked, can't use dasd_add_request_head */
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800410 clear_bit(DASD_FLAG_EER_SNSS, &device->flags);
411 cqr->status = DASD_CQR_QUEUED;
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100412 list_add(&cqr->devlist, &device->ccw_queue);
413 dasd_schedule_device_bh(device);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800414}
415
416/*
417 * Callback function for use with sense subsystem status request.
418 */
419static void dasd_eer_snss_cb(struct dasd_ccw_req *cqr, void *data)
420{
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100421 struct dasd_device *device = cqr->startdev;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800422 unsigned long flags;
423
424 dasd_eer_write(device, cqr, DASD_EER_STATECHANGE);
425 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
426 if (device->eer_cqr == cqr) {
427 clear_bit(DASD_FLAG_EER_IN_USE, &device->flags);
428 if (test_bit(DASD_FLAG_EER_SNSS, &device->flags))
429 /* Another SNSS has been requested in the meantime. */
430 dasd_eer_snss(device);
431 cqr = NULL;
432 }
433 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
434 if (cqr)
435 /*
436 * Extended error recovery has been switched off while
437 * the SNSS request was running. It could even have
438 * been switched off and on again in which case there
439 * is a new ccw in device->eer_cqr. Free the "old"
440 * snss request now.
441 */
442 dasd_kfree_request(cqr, device);
443}
444
445/*
446 * Enable error reporting on a given device.
447 */
448int dasd_eer_enable(struct dasd_device *device)
449{
450 struct dasd_ccw_req *cqr;
451 unsigned long flags;
452
453 if (device->eer_cqr)
454 return 0;
455
456 if (!device->discipline || strcmp(device->discipline->name, "ECKD"))
457 return -EPERM; /* FIXME: -EMEDIUMTYPE ? */
458
459 cqr = dasd_kmalloc_request("ECKD", 1 /* SNSS */,
460 SNSS_DATA_SIZE, device);
Julien Brunel0983e562008-08-21 19:46:30 +0200461 if (IS_ERR(cqr))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800462 return -ENOMEM;
463
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100464 cqr->startdev = device;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800465 cqr->retries = 255;
466 cqr->expires = 10 * HZ;
Stefan Weinhuber046f3e82007-03-05 23:35:52 +0100467 clear_bit(DASD_CQR_FLAGS_USE_ERP, &cqr->flags);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800468
469 cqr->cpaddr->cmd_code = DASD_ECKD_CCW_SNSS;
470 cqr->cpaddr->count = SNSS_DATA_SIZE;
471 cqr->cpaddr->flags = 0;
472 cqr->cpaddr->cda = (__u32)(addr_t) cqr->data;
473
474 cqr->buildclk = get_clock();
475 cqr->status = DASD_CQR_FILLED;
476 cqr->callback = dasd_eer_snss_cb;
477
478 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
479 if (!device->eer_cqr) {
480 device->eer_cqr = cqr;
481 cqr = NULL;
482 }
483 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
484 if (cqr)
485 dasd_kfree_request(cqr, device);
486 return 0;
487}
488
489/*
490 * Disable error reporting on a given device.
491 */
492void dasd_eer_disable(struct dasd_device *device)
493{
494 struct dasd_ccw_req *cqr;
495 unsigned long flags;
496 int in_use;
497
498 if (!device->eer_cqr)
499 return;
500 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
501 cqr = device->eer_cqr;
502 device->eer_cqr = NULL;
503 clear_bit(DASD_FLAG_EER_SNSS, &device->flags);
504 in_use = test_and_clear_bit(DASD_FLAG_EER_IN_USE, &device->flags);
505 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
506 if (cqr && !in_use)
507 dasd_kfree_request(cqr, device);
508}
509
510/*
511 * SECTION: the device operations
512 */
513
514/*
515 * On the one side we need a lock to access our internal buffer, on the
516 * other side a copy_to_user can sleep. So we need to copy the data we have
517 * to transfer in a readbuffer, which is protected by the readbuffer_mutex.
518 */
519static char readbuffer[PAGE_SIZE];
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200520static DEFINE_MUTEX(readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800521
522static int dasd_eer_open(struct inode *inp, struct file *filp)
523{
524 struct eerbuffer *eerb;
525 unsigned long flags;
526
527 eerb = kzalloc(sizeof(struct eerbuffer), GFP_KERNEL);
Stefan Weinhuberf45a43d2006-06-29 14:57:46 +0200528 if (!eerb)
529 return -ENOMEM;
Arnd Bergmann986837b2008-05-20 19:15:38 +0200530 lock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800531 eerb->buffer_page_count = eer_pages;
532 if (eerb->buffer_page_count < 1 ||
533 eerb->buffer_page_count > INT_MAX / PAGE_SIZE) {
534 kfree(eerb);
535 MESSAGE(KERN_WARNING, "can't open device since module "
536 "parameter eer_pages is smaller then 1 or"
537 " bigger then %d", (int)(INT_MAX / PAGE_SIZE));
Arnd Bergmann986837b2008-05-20 19:15:38 +0200538 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800539 return -EINVAL;
540 }
541 eerb->buffersize = eerb->buffer_page_count * PAGE_SIZE;
542 eerb->buffer = kmalloc(eerb->buffer_page_count * sizeof(char *),
543 GFP_KERNEL);
544 if (!eerb->buffer) {
545 kfree(eerb);
Arnd Bergmann986837b2008-05-20 19:15:38 +0200546 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800547 return -ENOMEM;
548 }
549 if (dasd_eer_allocate_buffer_pages(eerb->buffer,
550 eerb->buffer_page_count)) {
551 kfree(eerb->buffer);
552 kfree(eerb);
Arnd Bergmann986837b2008-05-20 19:15:38 +0200553 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800554 return -ENOMEM;
555 }
556 filp->private_data = eerb;
557 spin_lock_irqsave(&bufferlock, flags);
558 list_add(&eerb->list, &bufferlist);
559 spin_unlock_irqrestore(&bufferlock, flags);
560
Arnd Bergmann986837b2008-05-20 19:15:38 +0200561 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800562 return nonseekable_open(inp,filp);
563}
564
565static int dasd_eer_close(struct inode *inp, struct file *filp)
566{
567 struct eerbuffer *eerb;
568 unsigned long flags;
569
570 eerb = (struct eerbuffer *) filp->private_data;
571 spin_lock_irqsave(&bufferlock, flags);
572 list_del(&eerb->list);
573 spin_unlock_irqrestore(&bufferlock, flags);
574 dasd_eer_free_buffer_pages(eerb->buffer, eerb->buffer_page_count);
575 kfree(eerb->buffer);
576 kfree(eerb);
577
578 return 0;
579}
580
581static ssize_t dasd_eer_read(struct file *filp, char __user *buf,
582 size_t count, loff_t *ppos)
583{
584 int tc,rc;
585 int tailcount,effective_count;
586 unsigned long flags;
587 struct eerbuffer *eerb;
588
589 eerb = (struct eerbuffer *) filp->private_data;
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200590 if (mutex_lock_interruptible(&readbuffer_mutex))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800591 return -ERESTARTSYS;
592
593 spin_lock_irqsave(&bufferlock, flags);
594
595 if (eerb->residual < 0) { /* the remainder of this record */
596 /* has been deleted */
597 eerb->residual = 0;
598 spin_unlock_irqrestore(&bufferlock, flags);
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200599 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800600 return -EIO;
601 } else if (eerb->residual > 0) {
602 /* OK we still have a second half of a record to deliver */
603 effective_count = min(eerb->residual, (int) count);
604 eerb->residual -= effective_count;
605 } else {
606 tc = 0;
607 while (!tc) {
608 tc = dasd_eer_read_buffer(eerb, (char *) &tailcount,
609 sizeof(tailcount));
610 if (!tc) {
611 /* no data available */
612 spin_unlock_irqrestore(&bufferlock, flags);
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200613 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800614 if (filp->f_flags & O_NONBLOCK)
615 return -EAGAIN;
616 rc = wait_event_interruptible(
617 dasd_eer_read_wait_queue,
618 eerb->head != eerb->tail);
619 if (rc)
620 return rc;
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200621 if (mutex_lock_interruptible(&readbuffer_mutex))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800622 return -ERESTARTSYS;
623 spin_lock_irqsave(&bufferlock, flags);
624 }
625 }
626 WARN_ON(tc != sizeof(tailcount));
627 effective_count = min(tailcount,(int)count);
628 eerb->residual = tailcount - effective_count;
629 }
630
631 tc = dasd_eer_read_buffer(eerb, readbuffer, effective_count);
632 WARN_ON(tc != effective_count);
633
634 spin_unlock_irqrestore(&bufferlock, flags);
635
636 if (copy_to_user(buf, readbuffer, effective_count)) {
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200637 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800638 return -EFAULT;
639 }
640
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200641 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800642 return effective_count;
643}
644
645static unsigned int dasd_eer_poll(struct file *filp, poll_table *ptable)
646{
647 unsigned int mask;
648 unsigned long flags;
649 struct eerbuffer *eerb;
650
651 eerb = (struct eerbuffer *) filp->private_data;
652 poll_wait(filp, &dasd_eer_read_wait_queue, ptable);
653 spin_lock_irqsave(&bufferlock, flags);
654 if (eerb->head != eerb->tail)
655 mask = POLLIN | POLLRDNORM ;
656 else
657 mask = 0;
658 spin_unlock_irqrestore(&bufferlock, flags);
659 return mask;
660}
661
Arjan van de Vend54b1fd2007-02-12 00:55:34 -0800662static const struct file_operations dasd_eer_fops = {
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800663 .open = &dasd_eer_open,
664 .release = &dasd_eer_close,
665 .read = &dasd_eer_read,
666 .poll = &dasd_eer_poll,
667 .owner = THIS_MODULE,
668};
669
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100670static struct miscdevice *dasd_eer_dev = NULL;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800671
672int __init dasd_eer_init(void)
673{
674 int rc;
675
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100676 dasd_eer_dev = kzalloc(sizeof(*dasd_eer_dev), GFP_KERNEL);
677 if (!dasd_eer_dev)
678 return -ENOMEM;
679
680 dasd_eer_dev->minor = MISC_DYNAMIC_MINOR;
681 dasd_eer_dev->name = "dasd_eer";
682 dasd_eer_dev->fops = &dasd_eer_fops;
683
684 rc = misc_register(dasd_eer_dev);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800685 if (rc) {
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100686 kfree(dasd_eer_dev);
687 dasd_eer_dev = NULL;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800688 MESSAGE(KERN_ERR, "%s", "dasd_eer_init could not "
689 "register misc device");
690 return rc;
691 }
692
693 return 0;
694}
695
Heiko Carstens1375fc12006-09-20 15:59:12 +0200696void dasd_eer_exit(void)
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800697{
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100698 if (dasd_eer_dev) {
699 WARN_ON(misc_deregister(dasd_eer_dev) != 0);
700 kfree(dasd_eer_dev);
701 dasd_eer_dev = NULL;
702 }
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800703}