blob: 9ce4209552ae3b2c6b16095656d40d2b5aebc2bb [file] [log] [blame]
Stefan Weinhuber20c64462006-03-24 03:15:25 -08001/*
2 * Character device driver for extended error reporting.
3 *
4 * Copyright (C) 2005 IBM Corporation
5 * extended error reporting for DASD ECKD devices
6 * Author(s): Stefan Weinhuber <wein@de.ibm.com>
7 */
8
9#include <linux/init.h>
10#include <linux/fs.h>
11#include <linux/kernel.h>
12#include <linux/miscdevice.h>
13#include <linux/module.h>
14#include <linux/moduleparam.h>
15#include <linux/device.h>
16#include <linux/poll.h>
Christoph Hellwig3006d7c2007-05-31 17:38:02 +020017#include <linux/mutex.h>
Arnd Bergmann986837b2008-05-20 19:15:38 +020018#include <linux/smp_lock.h>
Julien Brunel0983e562008-08-21 19:46:30 +020019#include <linux/err.h>
Stefan Weinhuber20c64462006-03-24 03:15:25 -080020
21#include <asm/uaccess.h>
Stefan Weinhuber20c64462006-03-24 03:15:25 -080022#include <asm/atomic.h>
23#include <asm/ebcdic.h>
24
25#include "dasd_int.h"
26#include "dasd_eckd.h"
27
28#ifdef PRINTK_HEADER
29#undef PRINTK_HEADER
30#endif /* PRINTK_HEADER */
31#define PRINTK_HEADER "dasd(eer):"
32
33/*
34 * SECTION: the internal buffer
35 */
36
37/*
38 * The internal buffer is meant to store obaque blobs of data, so it does
39 * not know of higher level concepts like triggers.
40 * It consists of a number of pages that are used as a ringbuffer. Each data
41 * blob is stored in a simple record that consists of an integer, which
42 * contains the size of the following data, and the data bytes themselfes.
43 *
44 * To allow for multiple independent readers we create one internal buffer
45 * each time the device is opened and destroy the buffer when the file is
46 * closed again. The number of pages used for this buffer is determined by
47 * the module parmeter eer_pages.
48 *
49 * One record can be written to a buffer by using the functions
50 * - dasd_eer_start_record (one time per record to write the size to the
51 * buffer and reserve the space for the data)
52 * - dasd_eer_write_buffer (one or more times per record to write the data)
53 * The data can be written in several steps but you will have to compute
54 * the total size up front for the invocation of dasd_eer_start_record.
55 * If the ringbuffer is full, dasd_eer_start_record will remove the required
56 * number of old records.
57 *
58 * A record is typically read in two steps, first read the integer that
59 * specifies the size of the following data, then read the data.
60 * Both can be done by
61 * - dasd_eer_read_buffer
62 *
63 * For all mentioned functions you need to get the bufferlock first and keep
64 * it until a complete record is written or read.
65 *
66 * All information necessary to keep track of an internal buffer is kept in
67 * a struct eerbuffer. The buffer specific to a file pointer is strored in
68 * the private_data field of that file. To be able to write data to all
69 * existing buffers, each buffer is also added to the bufferlist.
70 * If the user does not want to read a complete record in one go, we have to
71 * keep track of the rest of the record. residual stores the number of bytes
72 * that are still to deliver. If the rest of the record is invalidated between
73 * two reads then residual will be set to -1 so that the next read will fail.
74 * All entries in the eerbuffer structure are protected with the bufferlock.
75 * To avoid races between writing to a buffer on the one side and creating
76 * and destroying buffers on the other side, the bufferlock must also be used
77 * to protect the bufferlist.
78 */
79
80static int eer_pages = 5;
81module_param(eer_pages, int, S_IRUGO|S_IWUSR);
82
83struct eerbuffer {
84 struct list_head list;
85 char **buffer;
86 int buffersize;
87 int buffer_page_count;
88 int head;
89 int tail;
90 int residual;
91};
92
93static LIST_HEAD(bufferlist);
Ingo Molnar34af9462006-06-27 02:53:55 -070094static DEFINE_SPINLOCK(bufferlock);
Stefan Weinhuber20c64462006-03-24 03:15:25 -080095static DECLARE_WAIT_QUEUE_HEAD(dasd_eer_read_wait_queue);
96
97/*
98 * How many free bytes are available on the buffer.
99 * Needs to be called with bufferlock held.
100 */
101static int dasd_eer_get_free_bytes(struct eerbuffer *eerb)
102{
103 if (eerb->head < eerb->tail)
104 return eerb->tail - eerb->head - 1;
105 return eerb->buffersize - eerb->head + eerb->tail -1;
106}
107
108/*
109 * How many bytes of buffer space are used.
110 * Needs to be called with bufferlock held.
111 */
112static int dasd_eer_get_filled_bytes(struct eerbuffer *eerb)
113{
114
115 if (eerb->head >= eerb->tail)
116 return eerb->head - eerb->tail;
117 return eerb->buffersize - eerb->tail + eerb->head;
118}
119
120/*
121 * The dasd_eer_write_buffer function just copies count bytes of data
122 * to the buffer. Make sure to call dasd_eer_start_record first, to
123 * make sure that enough free space is available.
124 * Needs to be called with bufferlock held.
125 */
126static void dasd_eer_write_buffer(struct eerbuffer *eerb,
127 char *data, int count)
128{
129
130 unsigned long headindex,localhead;
131 unsigned long rest, len;
132 char *nextdata;
133
134 nextdata = data;
135 rest = count;
136 while (rest > 0) {
137 headindex = eerb->head / PAGE_SIZE;
138 localhead = eerb->head % PAGE_SIZE;
139 len = min(rest, PAGE_SIZE - localhead);
140 memcpy(eerb->buffer[headindex]+localhead, nextdata, len);
141 nextdata += len;
142 rest -= len;
143 eerb->head += len;
144 if (eerb->head == eerb->buffersize)
145 eerb->head = 0; /* wrap around */
146 BUG_ON(eerb->head > eerb->buffersize);
147 }
148}
149
150/*
151 * Needs to be called with bufferlock held.
152 */
153static int dasd_eer_read_buffer(struct eerbuffer *eerb, char *data, int count)
154{
155
156 unsigned long tailindex,localtail;
157 unsigned long rest, len, finalcount;
158 char *nextdata;
159
160 finalcount = min(count, dasd_eer_get_filled_bytes(eerb));
161 nextdata = data;
162 rest = finalcount;
163 while (rest > 0) {
164 tailindex = eerb->tail / PAGE_SIZE;
165 localtail = eerb->tail % PAGE_SIZE;
166 len = min(rest, PAGE_SIZE - localtail);
167 memcpy(nextdata, eerb->buffer[tailindex] + localtail, len);
168 nextdata += len;
169 rest -= len;
170 eerb->tail += len;
171 if (eerb->tail == eerb->buffersize)
172 eerb->tail = 0; /* wrap around */
173 BUG_ON(eerb->tail > eerb->buffersize);
174 }
175 return finalcount;
176}
177
178/*
179 * Whenever you want to write a blob of data to the internal buffer you
180 * have to start by using this function first. It will write the number
181 * of bytes that will be written to the buffer. If necessary it will remove
182 * old records to make room for the new one.
183 * Needs to be called with bufferlock held.
184 */
185static int dasd_eer_start_record(struct eerbuffer *eerb, int count)
186{
187 int tailcount;
188
189 if (count + sizeof(count) > eerb->buffersize)
190 return -ENOMEM;
191 while (dasd_eer_get_free_bytes(eerb) < count + sizeof(count)) {
192 if (eerb->residual > 0) {
193 eerb->tail += eerb->residual;
194 if (eerb->tail >= eerb->buffersize)
195 eerb->tail -= eerb->buffersize;
196 eerb->residual = -1;
197 }
198 dasd_eer_read_buffer(eerb, (char *) &tailcount,
199 sizeof(tailcount));
200 eerb->tail += tailcount;
201 if (eerb->tail >= eerb->buffersize)
202 eerb->tail -= eerb->buffersize;
203 }
204 dasd_eer_write_buffer(eerb, (char*) &count, sizeof(count));
205
206 return 0;
207};
208
209/*
210 * Release pages that are not used anymore.
211 */
212static void dasd_eer_free_buffer_pages(char **buf, int no_pages)
213{
214 int i;
215
216 for (i = 0; i < no_pages; i++)
217 free_page((unsigned long) buf[i]);
218}
219
220/*
221 * Allocate a new set of memory pages.
222 */
223static int dasd_eer_allocate_buffer_pages(char **buf, int no_pages)
224{
225 int i;
226
227 for (i = 0; i < no_pages; i++) {
228 buf[i] = (char *) get_zeroed_page(GFP_KERNEL);
229 if (!buf[i]) {
230 dasd_eer_free_buffer_pages(buf, i);
231 return -ENOMEM;
232 }
233 }
234 return 0;
235}
236
237/*
238 * SECTION: The extended error reporting functionality
239 */
240
241/*
242 * When a DASD device driver wants to report an error, it calls the
243 * function dasd_eer_write and gives the respective trigger ID as
244 * parameter. Currently there are four kinds of triggers:
245 *
246 * DASD_EER_FATALERROR: all kinds of unrecoverable I/O problems
247 * DASD_EER_PPRCSUSPEND: PPRC was suspended
248 * DASD_EER_NOPATH: There is no path to the device left.
249 * DASD_EER_STATECHANGE: The state of the device has changed.
250 *
251 * For the first three triggers all required information can be supplied by
252 * the caller. For these triggers a record is written by the function
253 * dasd_eer_write_standard_trigger.
254 *
255 * The DASD_EER_STATECHANGE trigger is special since a sense subsystem
256 * status ccw need to be executed to gather the necessary sense data first.
257 * The dasd_eer_snss function will queue the SNSS request and the request
258 * callback will then call dasd_eer_write with the DASD_EER_STATCHANGE
259 * trigger.
260 *
261 * To avoid memory allocations at runtime, the necessary memory is allocated
262 * when the extended error reporting is enabled for a device (by
263 * dasd_eer_probe). There is one sense subsystem status request for each
264 * eer enabled DASD device. The presence of the cqr in device->eer_cqr
265 * indicates that eer is enable for the device. The use of the snss request
266 * is protected by the DASD_FLAG_EER_IN_USE bit. When this flag indicates
267 * that the cqr is currently in use, dasd_eer_snss cannot start a second
268 * request but sets the DASD_FLAG_EER_SNSS flag instead. The callback of
269 * the SNSS request will check the bit and call dasd_eer_snss again.
270 */
271
272#define SNSS_DATA_SIZE 44
273
274#define DASD_EER_BUSID_SIZE 10
275struct dasd_eer_header {
276 __u32 total_size;
277 __u32 trigger;
278 __u64 tv_sec;
279 __u64 tv_usec;
280 char busid[DASD_EER_BUSID_SIZE];
Stefan Weinhuber774fc4e2006-06-29 15:02:59 +0200281} __attribute__ ((packed));
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800282
283/*
284 * The following function can be used for those triggers that have
285 * all necessary data available when the function is called.
286 * If the parameter cqr is not NULL, the chain of requests will be searched
287 * for valid sense data, and all valid sense data sets will be added to
288 * the triggers data.
289 */
290static void dasd_eer_write_standard_trigger(struct dasd_device *device,
291 struct dasd_ccw_req *cqr,
292 int trigger)
293{
294 struct dasd_ccw_req *temp_cqr;
295 int data_size;
296 struct timeval tv;
297 struct dasd_eer_header header;
298 unsigned long flags;
299 struct eerbuffer *eerb;
Stefan Weinhuberf3eb5382009-03-26 15:23:48 +0100300 char *sense;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800301
302 /* go through cqr chain and count the valid sense data sets */
303 data_size = 0;
304 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers)
Stefan Weinhuberf3eb5382009-03-26 15:23:48 +0100305 if (dasd_get_sense(&temp_cqr->irb))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800306 data_size += 32;
307
308 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */
309 header.trigger = trigger;
310 do_gettimeofday(&tv);
311 header.tv_sec = tv.tv_sec;
312 header.tv_usec = tv.tv_usec;
Kay Sievers2a0217d2008-10-10 21:33:09 +0200313 strncpy(header.busid, dev_name(&device->cdev->dev),
314 DASD_EER_BUSID_SIZE);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800315
316 spin_lock_irqsave(&bufferlock, flags);
317 list_for_each_entry(eerb, &bufferlist, list) {
318 dasd_eer_start_record(eerb, header.total_size);
319 dasd_eer_write_buffer(eerb, (char *) &header, sizeof(header));
Stefan Weinhuberf3eb5382009-03-26 15:23:48 +0100320 for (temp_cqr = cqr; temp_cqr; temp_cqr = temp_cqr->refers) {
321 sense = dasd_get_sense(&temp_cqr->irb);
322 if (sense)
323 dasd_eer_write_buffer(eerb, sense, 32);
324 }
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800325 dasd_eer_write_buffer(eerb, "EOR", 4);
326 }
327 spin_unlock_irqrestore(&bufferlock, flags);
328 wake_up_interruptible(&dasd_eer_read_wait_queue);
329}
330
331/*
332 * This function writes a DASD_EER_STATECHANGE trigger.
333 */
334static void dasd_eer_write_snss_trigger(struct dasd_device *device,
335 struct dasd_ccw_req *cqr,
336 int trigger)
337{
338 int data_size;
339 int snss_rc;
340 struct timeval tv;
341 struct dasd_eer_header header;
342 unsigned long flags;
343 struct eerbuffer *eerb;
344
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100345 snss_rc = (cqr->status == DASD_CQR_DONE) ? 0 : -EIO;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800346 if (snss_rc)
347 data_size = 0;
348 else
349 data_size = SNSS_DATA_SIZE;
350
351 header.total_size = sizeof(header) + data_size + 4; /* "EOR" */
352 header.trigger = DASD_EER_STATECHANGE;
353 do_gettimeofday(&tv);
354 header.tv_sec = tv.tv_sec;
355 header.tv_usec = tv.tv_usec;
Kay Sievers2a0217d2008-10-10 21:33:09 +0200356 strncpy(header.busid, dev_name(&device->cdev->dev),
357 DASD_EER_BUSID_SIZE);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800358
359 spin_lock_irqsave(&bufferlock, flags);
360 list_for_each_entry(eerb, &bufferlist, list) {
361 dasd_eer_start_record(eerb, header.total_size);
362 dasd_eer_write_buffer(eerb, (char *) &header , sizeof(header));
363 if (!snss_rc)
364 dasd_eer_write_buffer(eerb, cqr->data, SNSS_DATA_SIZE);
365 dasd_eer_write_buffer(eerb, "EOR", 4);
366 }
367 spin_unlock_irqrestore(&bufferlock, flags);
368 wake_up_interruptible(&dasd_eer_read_wait_queue);
369}
370
371/*
372 * This function is called for all triggers. It calls the appropriate
373 * function that writes the actual trigger records.
374 */
375void dasd_eer_write(struct dasd_device *device, struct dasd_ccw_req *cqr,
376 unsigned int id)
377{
378 if (!device->eer_cqr)
379 return;
380 switch (id) {
381 case DASD_EER_FATALERROR:
382 case DASD_EER_PPRCSUSPEND:
383 dasd_eer_write_standard_trigger(device, cqr, id);
384 break;
385 case DASD_EER_NOPATH:
386 dasd_eer_write_standard_trigger(device, NULL, id);
387 break;
388 case DASD_EER_STATECHANGE:
389 dasd_eer_write_snss_trigger(device, cqr, id);
390 break;
391 default: /* unknown trigger, so we write it without any sense data */
392 dasd_eer_write_standard_trigger(device, NULL, id);
393 break;
394 }
395}
396EXPORT_SYMBOL(dasd_eer_write);
397
398/*
399 * Start a sense subsystem status request.
400 * Needs to be called with the device held.
401 */
402void dasd_eer_snss(struct dasd_device *device)
403{
404 struct dasd_ccw_req *cqr;
405
406 cqr = device->eer_cqr;
407 if (!cqr) /* Device not eer enabled. */
408 return;
409 if (test_and_set_bit(DASD_FLAG_EER_IN_USE, &device->flags)) {
410 /* Sense subsystem status request in use. */
411 set_bit(DASD_FLAG_EER_SNSS, &device->flags);
412 return;
413 }
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100414 /* cdev is already locked, can't use dasd_add_request_head */
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800415 clear_bit(DASD_FLAG_EER_SNSS, &device->flags);
416 cqr->status = DASD_CQR_QUEUED;
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100417 list_add(&cqr->devlist, &device->ccw_queue);
418 dasd_schedule_device_bh(device);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800419}
420
421/*
422 * Callback function for use with sense subsystem status request.
423 */
424static void dasd_eer_snss_cb(struct dasd_ccw_req *cqr, void *data)
425{
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100426 struct dasd_device *device = cqr->startdev;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800427 unsigned long flags;
428
429 dasd_eer_write(device, cqr, DASD_EER_STATECHANGE);
430 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
431 if (device->eer_cqr == cqr) {
432 clear_bit(DASD_FLAG_EER_IN_USE, &device->flags);
433 if (test_bit(DASD_FLAG_EER_SNSS, &device->flags))
434 /* Another SNSS has been requested in the meantime. */
435 dasd_eer_snss(device);
436 cqr = NULL;
437 }
438 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
439 if (cqr)
440 /*
441 * Extended error recovery has been switched off while
442 * the SNSS request was running. It could even have
443 * been switched off and on again in which case there
444 * is a new ccw in device->eer_cqr. Free the "old"
445 * snss request now.
446 */
447 dasd_kfree_request(cqr, device);
448}
449
450/*
451 * Enable error reporting on a given device.
452 */
453int dasd_eer_enable(struct dasd_device *device)
454{
455 struct dasd_ccw_req *cqr;
456 unsigned long flags;
Stefan Weinhuberf3eb5382009-03-26 15:23:48 +0100457 struct ccw1 *ccw;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800458
459 if (device->eer_cqr)
460 return 0;
461
462 if (!device->discipline || strcmp(device->discipline->name, "ECKD"))
463 return -EPERM; /* FIXME: -EMEDIUMTYPE ? */
464
465 cqr = dasd_kmalloc_request("ECKD", 1 /* SNSS */,
466 SNSS_DATA_SIZE, device);
Julien Brunel0983e562008-08-21 19:46:30 +0200467 if (IS_ERR(cqr))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800468 return -ENOMEM;
469
Stefan Weinhuber8e09f212008-01-26 14:11:23 +0100470 cqr->startdev = device;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800471 cqr->retries = 255;
472 cqr->expires = 10 * HZ;
Stefan Weinhuber046f3e82007-03-05 23:35:52 +0100473 clear_bit(DASD_CQR_FLAGS_USE_ERP, &cqr->flags);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800474
Stefan Weinhuberf3eb5382009-03-26 15:23:48 +0100475 ccw = cqr->cpaddr;
476 ccw->cmd_code = DASD_ECKD_CCW_SNSS;
477 ccw->count = SNSS_DATA_SIZE;
478 ccw->flags = 0;
479 ccw->cda = (__u32)(addr_t) cqr->data;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800480
481 cqr->buildclk = get_clock();
482 cqr->status = DASD_CQR_FILLED;
483 cqr->callback = dasd_eer_snss_cb;
484
485 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
486 if (!device->eer_cqr) {
487 device->eer_cqr = cqr;
488 cqr = NULL;
489 }
490 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
491 if (cqr)
492 dasd_kfree_request(cqr, device);
493 return 0;
494}
495
496/*
497 * Disable error reporting on a given device.
498 */
499void dasd_eer_disable(struct dasd_device *device)
500{
501 struct dasd_ccw_req *cqr;
502 unsigned long flags;
503 int in_use;
504
505 if (!device->eer_cqr)
506 return;
507 spin_lock_irqsave(get_ccwdev_lock(device->cdev), flags);
508 cqr = device->eer_cqr;
509 device->eer_cqr = NULL;
510 clear_bit(DASD_FLAG_EER_SNSS, &device->flags);
511 in_use = test_and_clear_bit(DASD_FLAG_EER_IN_USE, &device->flags);
512 spin_unlock_irqrestore(get_ccwdev_lock(device->cdev), flags);
513 if (cqr && !in_use)
514 dasd_kfree_request(cqr, device);
515}
516
517/*
518 * SECTION: the device operations
519 */
520
521/*
522 * On the one side we need a lock to access our internal buffer, on the
523 * other side a copy_to_user can sleep. So we need to copy the data we have
524 * to transfer in a readbuffer, which is protected by the readbuffer_mutex.
525 */
526static char readbuffer[PAGE_SIZE];
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200527static DEFINE_MUTEX(readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800528
529static int dasd_eer_open(struct inode *inp, struct file *filp)
530{
531 struct eerbuffer *eerb;
532 unsigned long flags;
533
534 eerb = kzalloc(sizeof(struct eerbuffer), GFP_KERNEL);
Stefan Weinhuberf45a43d2006-06-29 14:57:46 +0200535 if (!eerb)
536 return -ENOMEM;
Arnd Bergmann986837b2008-05-20 19:15:38 +0200537 lock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800538 eerb->buffer_page_count = eer_pages;
539 if (eerb->buffer_page_count < 1 ||
540 eerb->buffer_page_count > INT_MAX / PAGE_SIZE) {
541 kfree(eerb);
542 MESSAGE(KERN_WARNING, "can't open device since module "
Frederik Schwarzer025dfda2008-10-16 19:02:37 +0200543 "parameter eer_pages is smaller than 1 or"
544 " bigger than %d", (int)(INT_MAX / PAGE_SIZE));
Arnd Bergmann986837b2008-05-20 19:15:38 +0200545 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800546 return -EINVAL;
547 }
548 eerb->buffersize = eerb->buffer_page_count * PAGE_SIZE;
549 eerb->buffer = kmalloc(eerb->buffer_page_count * sizeof(char *),
550 GFP_KERNEL);
551 if (!eerb->buffer) {
552 kfree(eerb);
Arnd Bergmann986837b2008-05-20 19:15:38 +0200553 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800554 return -ENOMEM;
555 }
556 if (dasd_eer_allocate_buffer_pages(eerb->buffer,
557 eerb->buffer_page_count)) {
558 kfree(eerb->buffer);
559 kfree(eerb);
Arnd Bergmann986837b2008-05-20 19:15:38 +0200560 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800561 return -ENOMEM;
562 }
563 filp->private_data = eerb;
564 spin_lock_irqsave(&bufferlock, flags);
565 list_add(&eerb->list, &bufferlist);
566 spin_unlock_irqrestore(&bufferlock, flags);
567
Arnd Bergmann986837b2008-05-20 19:15:38 +0200568 unlock_kernel();
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800569 return nonseekable_open(inp,filp);
570}
571
572static int dasd_eer_close(struct inode *inp, struct file *filp)
573{
574 struct eerbuffer *eerb;
575 unsigned long flags;
576
577 eerb = (struct eerbuffer *) filp->private_data;
578 spin_lock_irqsave(&bufferlock, flags);
579 list_del(&eerb->list);
580 spin_unlock_irqrestore(&bufferlock, flags);
581 dasd_eer_free_buffer_pages(eerb->buffer, eerb->buffer_page_count);
582 kfree(eerb->buffer);
583 kfree(eerb);
584
585 return 0;
586}
587
588static ssize_t dasd_eer_read(struct file *filp, char __user *buf,
589 size_t count, loff_t *ppos)
590{
591 int tc,rc;
592 int tailcount,effective_count;
593 unsigned long flags;
594 struct eerbuffer *eerb;
595
596 eerb = (struct eerbuffer *) filp->private_data;
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200597 if (mutex_lock_interruptible(&readbuffer_mutex))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800598 return -ERESTARTSYS;
599
600 spin_lock_irqsave(&bufferlock, flags);
601
602 if (eerb->residual < 0) { /* the remainder of this record */
603 /* has been deleted */
604 eerb->residual = 0;
605 spin_unlock_irqrestore(&bufferlock, flags);
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200606 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800607 return -EIO;
608 } else if (eerb->residual > 0) {
609 /* OK we still have a second half of a record to deliver */
610 effective_count = min(eerb->residual, (int) count);
611 eerb->residual -= effective_count;
612 } else {
613 tc = 0;
614 while (!tc) {
615 tc = dasd_eer_read_buffer(eerb, (char *) &tailcount,
616 sizeof(tailcount));
617 if (!tc) {
618 /* no data available */
619 spin_unlock_irqrestore(&bufferlock, flags);
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200620 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800621 if (filp->f_flags & O_NONBLOCK)
622 return -EAGAIN;
623 rc = wait_event_interruptible(
624 dasd_eer_read_wait_queue,
625 eerb->head != eerb->tail);
626 if (rc)
627 return rc;
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200628 if (mutex_lock_interruptible(&readbuffer_mutex))
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800629 return -ERESTARTSYS;
630 spin_lock_irqsave(&bufferlock, flags);
631 }
632 }
633 WARN_ON(tc != sizeof(tailcount));
634 effective_count = min(tailcount,(int)count);
635 eerb->residual = tailcount - effective_count;
636 }
637
638 tc = dasd_eer_read_buffer(eerb, readbuffer, effective_count);
639 WARN_ON(tc != effective_count);
640
641 spin_unlock_irqrestore(&bufferlock, flags);
642
643 if (copy_to_user(buf, readbuffer, effective_count)) {
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200644 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800645 return -EFAULT;
646 }
647
Christoph Hellwig3006d7c2007-05-31 17:38:02 +0200648 mutex_unlock(&readbuffer_mutex);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800649 return effective_count;
650}
651
652static unsigned int dasd_eer_poll(struct file *filp, poll_table *ptable)
653{
654 unsigned int mask;
655 unsigned long flags;
656 struct eerbuffer *eerb;
657
658 eerb = (struct eerbuffer *) filp->private_data;
659 poll_wait(filp, &dasd_eer_read_wait_queue, ptable);
660 spin_lock_irqsave(&bufferlock, flags);
661 if (eerb->head != eerb->tail)
662 mask = POLLIN | POLLRDNORM ;
663 else
664 mask = 0;
665 spin_unlock_irqrestore(&bufferlock, flags);
666 return mask;
667}
668
Arjan van de Vend54b1fd2007-02-12 00:55:34 -0800669static const struct file_operations dasd_eer_fops = {
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800670 .open = &dasd_eer_open,
671 .release = &dasd_eer_close,
672 .read = &dasd_eer_read,
673 .poll = &dasd_eer_poll,
674 .owner = THIS_MODULE,
675};
676
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100677static struct miscdevice *dasd_eer_dev = NULL;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800678
679int __init dasd_eer_init(void)
680{
681 int rc;
682
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100683 dasd_eer_dev = kzalloc(sizeof(*dasd_eer_dev), GFP_KERNEL);
684 if (!dasd_eer_dev)
685 return -ENOMEM;
686
687 dasd_eer_dev->minor = MISC_DYNAMIC_MINOR;
688 dasd_eer_dev->name = "dasd_eer";
689 dasd_eer_dev->fops = &dasd_eer_fops;
690
691 rc = misc_register(dasd_eer_dev);
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800692 if (rc) {
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100693 kfree(dasd_eer_dev);
694 dasd_eer_dev = NULL;
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800695 MESSAGE(KERN_ERR, "%s", "dasd_eer_init could not "
696 "register misc device");
697 return rc;
698 }
699
700 return 0;
701}
702
Heiko Carstens1375fc12006-09-20 15:59:12 +0200703void dasd_eer_exit(void)
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800704{
Stefan Weinhubere3c699b2007-02-05 21:17:04 +0100705 if (dasd_eer_dev) {
706 WARN_ON(misc_deregister(dasd_eer_dev) != 0);
707 kfree(dasd_eer_dev);
708 dasd_eer_dev = NULL;
709 }
Stefan Weinhuber20c64462006-03-24 03:15:25 -0800710}