2 * character device driver for reading z/VM system service records
5 * Copyright IBM Corp. 2004, 2009
6 * character device driver for reading z/VM system service records,
8 * Author(s): Xenia Tkatschow <xenia@us.ibm.com>
9 * Stefan Weinhuber <wein@de.ibm.com>
13 #define KMSG_COMPONENT "vmlogrdr"
14 #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
16 #include <linux/module.h>
17 #include <linux/init.h>
18 #include <linux/slab.h>
19 #include <linux/errno.h>
20 #include <linux/types.h>
21 #include <linux/interrupt.h>
22 #include <linux/spinlock.h>
23 #include <linux/atomic.h>
24 #include <asm/uaccess.h>
25 #include <asm/cpcmd.h>
26 #include <asm/debug.h>
27 #include <asm/ebcdic.h>
28 #include <net/iucv/iucv.h>
29 #include <linux/kmod.h>
30 #include <linux/cdev.h>
31 #include <linux/device.h>
32 #include <linux/string.h>
35 ("(C) 2004 IBM Corporation by Xenia Tkatschow (xenia@us.ibm.com)\n"
36 " Stefan Weinhuber (wein@de.ibm.com)");
37 MODULE_DESCRIPTION ("Character device driver for reading z/VM "
38 "system service records.");
39 MODULE_LICENSE("GPL");
43 * The size of the buffer for iucv data transfer is one page,
44 * but in addition to the data we read from iucv we also
45 * place an integer and some characters into that buffer,
46 * so the maximum size for record data is a little less then
49 #define NET_BUFFER_SIZE (PAGE_SIZE - sizeof(int) - sizeof(FENCE))
52 * The elements that are concurrently accessed by bottom halves are
53 * connection_established, iucv_path_severed, local_interrupt_buffer
54 * and receive_ready. The first three can be protected by
55 * priv_lock. receive_ready is atomic, so it can be incremented and
56 * decremented without holding a lock.
57 * The variable dev_in_use needs to be protected by the lock, since
58 * it's a flag used by open to make sure that the device is opened only
59 * by one user at the same time.
61 struct vmlogrdr_priv_t
{
62 char system_service
[8];
63 char internal_name
[8];
64 char recording_name
[8];
65 struct iucv_path
*path
;
66 int connection_established
;
67 int iucv_path_severed
;
68 struct iucv_message local_interrupt_buffer
;
69 atomic_t receive_ready
;
72 char * current_position
;
74 ulong residual_length
;
76 int dev_in_use
; /* 1: already opened, 0: not opened*/
78 struct device
*device
;
79 struct device
*class_device
;
86 * File operation structure for vmlogrdr devices
88 static int vmlogrdr_open(struct inode
*, struct file
*);
89 static int vmlogrdr_release(struct inode
*, struct file
*);
90 static ssize_t
vmlogrdr_read (struct file
*filp
, char __user
*data
,
91 size_t count
, loff_t
* ppos
);
93 static const struct file_operations vmlogrdr_fops
= {
95 .open
= vmlogrdr_open
,
96 .release
= vmlogrdr_release
,
97 .read
= vmlogrdr_read
,
102 static void vmlogrdr_iucv_path_complete(struct iucv_path
*, u8 ipuser
[16]);
103 static void vmlogrdr_iucv_path_severed(struct iucv_path
*, u8 ipuser
[16]);
104 static void vmlogrdr_iucv_message_pending(struct iucv_path
*,
105 struct iucv_message
*);
108 static struct iucv_handler vmlogrdr_iucv_handler
= {
109 .path_complete
= vmlogrdr_iucv_path_complete
,
110 .path_severed
= vmlogrdr_iucv_path_severed
,
111 .message_pending
= vmlogrdr_iucv_message_pending
,
115 static DECLARE_WAIT_QUEUE_HEAD(conn_wait_queue
);
116 static DECLARE_WAIT_QUEUE_HEAD(read_wait_queue
);
119 * pointer to system service private structure
120 * minor number 0 --> logrec
121 * minor number 1 --> account
122 * minor number 2 --> symptom
125 static struct vmlogrdr_priv_t sys_ser
[] = {
126 { .system_service
= "*LOGREC ",
127 .internal_name
= "logrec",
128 .recording_name
= "EREP",
131 .priv_lock
= __SPIN_LOCK_UNLOCKED(sys_ser
[0].priv_lock
),
135 { .system_service
= "*ACCOUNT",
136 .internal_name
= "account",
137 .recording_name
= "ACCOUNT",
140 .priv_lock
= __SPIN_LOCK_UNLOCKED(sys_ser
[1].priv_lock
),
144 { .system_service
= "*SYMPTOM",
145 .internal_name
= "symptom",
146 .recording_name
= "SYMPTOM",
149 .priv_lock
= __SPIN_LOCK_UNLOCKED(sys_ser
[2].priv_lock
),
155 #define MAXMINOR (sizeof(sys_ser)/sizeof(struct vmlogrdr_priv_t))
157 static char FENCE
[] = {"EOR"};
158 static int vmlogrdr_major
= 0;
159 static struct cdev
*vmlogrdr_cdev
= NULL
;
160 static int recording_class_AB
;
163 static void vmlogrdr_iucv_path_complete(struct iucv_path
*path
, u8 ipuser
[16])
165 struct vmlogrdr_priv_t
* logptr
= path
->private;
167 spin_lock(&logptr
->priv_lock
);
168 logptr
->connection_established
= 1;
169 spin_unlock(&logptr
->priv_lock
);
170 wake_up(&conn_wait_queue
);
174 static void vmlogrdr_iucv_path_severed(struct iucv_path
*path
, u8 ipuser
[16])
176 struct vmlogrdr_priv_t
* logptr
= path
->private;
177 u8 reason
= (u8
) ipuser
[8];
179 pr_err("vmlogrdr: connection severed with reason %i\n", reason
);
181 iucv_path_sever(path
, NULL
);
185 spin_lock(&logptr
->priv_lock
);
186 logptr
->connection_established
= 0;
187 logptr
->iucv_path_severed
= 1;
188 spin_unlock(&logptr
->priv_lock
);
190 wake_up(&conn_wait_queue
);
191 /* just in case we're sleeping waiting for a record */
192 wake_up_interruptible(&read_wait_queue
);
196 static void vmlogrdr_iucv_message_pending(struct iucv_path
*path
,
197 struct iucv_message
*msg
)
199 struct vmlogrdr_priv_t
* logptr
= path
->private;
202 * This function is the bottom half so it should be quick.
203 * Copy the external interrupt data into our local eib and increment
206 spin_lock(&logptr
->priv_lock
);
207 memcpy(&logptr
->local_interrupt_buffer
, msg
, sizeof(*msg
));
208 atomic_inc(&logptr
->receive_ready
);
209 spin_unlock(&logptr
->priv_lock
);
210 wake_up_interruptible(&read_wait_queue
);
214 static int vmlogrdr_get_recording_class_AB(void)
216 static const char cp_command
[] = "QUERY COMMAND RECORDING ";
217 char cp_response
[80];
221 cpcmd(cp_command
, cp_response
, sizeof(cp_response
), NULL
);
222 len
= strnlen(cp_response
,sizeof(cp_response
));
224 tail
=strnchr(cp_response
,len
,'=');
228 if (!strncmp("ANY",tail
,3))
230 if (!strncmp("NONE",tail
,4))
233 * expect comma separated list of classes here, if one of them
234 * is A or B return 1 otherwise 0
236 for (i
=tail
-cp_response
; i
<len
; i
++)
237 if ( cp_response
[i
]=='A' || cp_response
[i
]=='B' )
243 static int vmlogrdr_recording(struct vmlogrdr_priv_t
* logptr
,
244 int action
, int purge
)
248 char cp_response
[160];
249 char *onoff
, *qid_string
;
252 onoff
= ((action
== 1) ? "ON" : "OFF");
253 qid_string
= ((recording_class_AB
== 1) ? " QID * " : "");
256 * The recording commands needs to be called with option QID
257 * for guests that have previlege classes A or B.
258 * Purging has to be done as separate step, because recording
259 * can't be switched on as long as records are on the queue.
260 * Doing both at the same time doesn't work.
262 if (purge
&& (action
== 1)) {
263 memset(cp_command
, 0x00, sizeof(cp_command
));
264 memset(cp_response
, 0x00, sizeof(cp_response
));
265 snprintf(cp_command
, sizeof(cp_command
),
266 "RECORDING %s PURGE %s",
267 logptr
->recording_name
,
269 cpcmd(cp_command
, cp_response
, sizeof(cp_response
), NULL
);
272 memset(cp_command
, 0x00, sizeof(cp_command
));
273 memset(cp_response
, 0x00, sizeof(cp_response
));
274 snprintf(cp_command
, sizeof(cp_command
), "RECORDING %s %s %s",
275 logptr
->recording_name
,
278 cpcmd(cp_command
, cp_response
, sizeof(cp_response
), NULL
);
279 /* The recording command will usually answer with 'Command complete'
280 * on success, but when the specific service was never connected
281 * before then there might be an additional informational message
282 * 'HCPCRC8072I Recording entry not found' before the
283 * 'Command complete'. So I use strstr rather then the strncmp.
285 if (strstr(cp_response
,"Command complete"))
290 * If we turn recording off, we have to purge any remaining records
291 * afterwards, as a large number of queued records may impact z/VM
294 if (purge
&& (action
== 0)) {
295 memset(cp_command
, 0x00, sizeof(cp_command
));
296 memset(cp_response
, 0x00, sizeof(cp_response
));
297 snprintf(cp_command
, sizeof(cp_command
),
298 "RECORDING %s PURGE %s",
299 logptr
->recording_name
,
301 cpcmd(cp_command
, cp_response
, sizeof(cp_response
), NULL
);
308 static int vmlogrdr_open (struct inode
*inode
, struct file
*filp
)
311 struct vmlogrdr_priv_t
* logptr
= NULL
;
315 dev_num
= iminor(inode
);
316 if (dev_num
>= MAXMINOR
)
318 logptr
= &sys_ser
[dev_num
];
321 * only allow for blocking reads to be open
323 if (filp
->f_flags
& O_NONBLOCK
)
326 /* Besure this device hasn't already been opened */
327 spin_lock_bh(&logptr
->priv_lock
);
328 if (logptr
->dev_in_use
) {
329 spin_unlock_bh(&logptr
->priv_lock
);
332 logptr
->dev_in_use
= 1;
333 logptr
->connection_established
= 0;
334 logptr
->iucv_path_severed
= 0;
335 atomic_set(&logptr
->receive_ready
, 0);
336 logptr
->buffer_free
= 1;
337 spin_unlock_bh(&logptr
->priv_lock
);
339 /* set the file options */
340 filp
->private_data
= logptr
;
341 filp
->f_op
= &vmlogrdr_fops
;
343 /* start recording for this service*/
344 if (logptr
->autorecording
) {
345 ret
= vmlogrdr_recording(logptr
,1,logptr
->autopurge
);
347 pr_warning("vmlogrdr: failed to start "
348 "recording automatically\n");
351 /* create connection to the system service */
352 logptr
->path
= iucv_path_alloc(10, 0, GFP_KERNEL
);
355 connect_rc
= iucv_path_connect(logptr
->path
, &vmlogrdr_iucv_handler
,
356 logptr
->system_service
, NULL
, NULL
,
359 pr_err("vmlogrdr: iucv connection to %s "
360 "failed with rc %i \n",
361 logptr
->system_service
, connect_rc
);
365 /* We've issued the connect and now we must wait for a
366 * ConnectionComplete or ConnectinSevered Interrupt
367 * before we can continue to process.
369 wait_event(conn_wait_queue
, (logptr
->connection_established
)
370 || (logptr
->iucv_path_severed
));
371 if (logptr
->iucv_path_severed
)
373 nonseekable_open(inode
, filp
);
377 if (logptr
->autorecording
)
378 vmlogrdr_recording(logptr
,0,logptr
->autopurge
);
380 kfree(logptr
->path
); /* kfree(NULL) is ok. */
383 logptr
->dev_in_use
= 0;
388 static int vmlogrdr_release (struct inode
*inode
, struct file
*filp
)
392 struct vmlogrdr_priv_t
* logptr
= filp
->private_data
;
394 iucv_path_sever(logptr
->path
, NULL
);
397 if (logptr
->autorecording
) {
398 ret
= vmlogrdr_recording(logptr
,0,logptr
->autopurge
);
400 pr_warning("vmlogrdr: failed to stop "
401 "recording automatically\n");
403 logptr
->dev_in_use
= 0;
409 static int vmlogrdr_receive_data(struct vmlogrdr_priv_t
*priv
)
412 /* we need to keep track of two data sizes here:
413 * The number of bytes we need to receive from iucv and
414 * the total number of bytes we actually write into the buffer.
416 int user_data_count
, iucv_data_count
;
419 if (atomic_read(&priv
->receive_ready
)) {
420 spin_lock_bh(&priv
->priv_lock
);
421 if (priv
->residual_length
){
422 /* receive second half of a record */
423 iucv_data_count
= priv
->residual_length
;
425 buffer
= priv
->buffer
;
427 /* receive a new record:
428 * We need to return the total length of the record
429 * + size of FENCE in the first 4 bytes of the buffer.
431 iucv_data_count
= priv
->local_interrupt_buffer
.length
;
432 user_data_count
= sizeof(int);
433 temp
= (int*)priv
->buffer
;
434 *temp
= iucv_data_count
+ sizeof(FENCE
);
435 buffer
= priv
->buffer
+ sizeof(int);
438 * If the record is bigger than our buffer, we receive only
439 * a part of it. We can get the rest later.
441 if (iucv_data_count
> NET_BUFFER_SIZE
)
442 iucv_data_count
= NET_BUFFER_SIZE
;
443 rc
= iucv_message_receive(priv
->path
,
444 &priv
->local_interrupt_buffer
,
445 0, buffer
, iucv_data_count
,
446 &priv
->residual_length
);
447 spin_unlock_bh(&priv
->priv_lock
);
448 /* An rc of 5 indicates that the record was bigger than
449 * the buffer, which is OK for us. A 9 indicates that the
450 * record was purged befor we could receive it.
455 atomic_set(&priv
->receive_ready
, 0);
460 priv
->buffer_free
= 0;
461 user_data_count
+= iucv_data_count
;
462 priv
->current_position
= priv
->buffer
;
463 if (priv
->residual_length
== 0){
464 /* the whole record has been captured,
465 * now add the fence */
466 atomic_dec(&priv
->receive_ready
);
467 buffer
= priv
->buffer
+ user_data_count
;
468 memcpy(buffer
, FENCE
, sizeof(FENCE
));
469 user_data_count
+= sizeof(FENCE
);
471 priv
->remaining
= user_data_count
;
478 static ssize_t
vmlogrdr_read(struct file
*filp
, char __user
*data
,
479 size_t count
, loff_t
* ppos
)
482 struct vmlogrdr_priv_t
* priv
= filp
->private_data
;
484 while (priv
->buffer_free
) {
485 rc
= vmlogrdr_receive_data(priv
);
487 rc
= wait_event_interruptible(read_wait_queue
,
488 atomic_read(&priv
->receive_ready
));
493 /* copy only up to end of record */
494 if (count
> priv
->remaining
)
495 count
= priv
->remaining
;
497 if (copy_to_user(data
, priv
->current_position
, count
))
501 priv
->current_position
+= count
;
502 priv
->remaining
-= count
;
504 /* if all data has been transferred, set buffer free */
505 if (priv
->remaining
== 0)
506 priv
->buffer_free
= 1;
511 static ssize_t
vmlogrdr_autopurge_store(struct device
* dev
,
512 struct device_attribute
*attr
,
513 const char * buf
, size_t count
)
515 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
532 static ssize_t
vmlogrdr_autopurge_show(struct device
*dev
,
533 struct device_attribute
*attr
,
536 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
537 return sprintf(buf
, "%u\n", priv
->autopurge
);
541 static DEVICE_ATTR(autopurge
, 0644, vmlogrdr_autopurge_show
,
542 vmlogrdr_autopurge_store
);
545 static ssize_t
vmlogrdr_purge_store(struct device
* dev
,
546 struct device_attribute
*attr
,
547 const char * buf
, size_t count
)
551 char cp_response
[80];
552 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
557 memset(cp_command
, 0x00, sizeof(cp_command
));
558 memset(cp_response
, 0x00, sizeof(cp_response
));
561 * The recording command needs to be called with option QID
562 * for guests that have previlege classes A or B.
563 * Other guests will not recognize the command and we have to
564 * issue the same command without the QID parameter.
567 if (recording_class_AB
)
568 snprintf(cp_command
, sizeof(cp_command
),
569 "RECORDING %s PURGE QID * ",
570 priv
->recording_name
);
572 snprintf(cp_command
, sizeof(cp_command
),
573 "RECORDING %s PURGE ",
574 priv
->recording_name
);
576 cpcmd(cp_command
, cp_response
, sizeof(cp_response
), NULL
);
582 static DEVICE_ATTR(purge
, 0200, NULL
, vmlogrdr_purge_store
);
585 static ssize_t
vmlogrdr_autorecording_store(struct device
*dev
,
586 struct device_attribute
*attr
,
587 const char *buf
, size_t count
)
589 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
594 priv
->autorecording
=0;
597 priv
->autorecording
=1;
606 static ssize_t
vmlogrdr_autorecording_show(struct device
*dev
,
607 struct device_attribute
*attr
,
610 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
611 return sprintf(buf
, "%u\n", priv
->autorecording
);
615 static DEVICE_ATTR(autorecording
, 0644, vmlogrdr_autorecording_show
,
616 vmlogrdr_autorecording_store
);
619 static ssize_t
vmlogrdr_recording_store(struct device
* dev
,
620 struct device_attribute
*attr
,
621 const char * buf
, size_t count
)
623 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
628 ret
= vmlogrdr_recording(priv
,0,0);
631 ret
= vmlogrdr_recording(priv
,1,0);
644 static DEVICE_ATTR(recording
, 0200, NULL
, vmlogrdr_recording_store
);
647 static ssize_t
vmlogrdr_recording_status_show(struct device_driver
*driver
,
651 static const char cp_command
[] = "QUERY RECORDING ";
654 cpcmd(cp_command
, buf
, 4096, NULL
);
658 static DRIVER_ATTR(recording_status
, 0444, vmlogrdr_recording_status_show
,
660 static struct attribute
*vmlogrdr_drv_attrs
[] = {
661 &driver_attr_recording_status
.attr
,
664 static struct attribute_group vmlogrdr_drv_attr_group
= {
665 .attrs
= vmlogrdr_drv_attrs
,
667 static const struct attribute_group
*vmlogrdr_drv_attr_groups
[] = {
668 &vmlogrdr_drv_attr_group
,
672 static struct attribute
*vmlogrdr_attrs
[] = {
673 &dev_attr_autopurge
.attr
,
674 &dev_attr_purge
.attr
,
675 &dev_attr_autorecording
.attr
,
676 &dev_attr_recording
.attr
,
679 static struct attribute_group vmlogrdr_attr_group
= {
680 .attrs
= vmlogrdr_attrs
,
682 static const struct attribute_group
*vmlogrdr_attr_groups
[] = {
683 &vmlogrdr_attr_group
,
687 static int vmlogrdr_pm_prepare(struct device
*dev
)
690 struct vmlogrdr_priv_t
*priv
= dev_get_drvdata(dev
);
694 spin_lock_bh(&priv
->priv_lock
);
695 if (priv
->dev_in_use
)
697 spin_unlock_bh(&priv
->priv_lock
);
700 pr_err("vmlogrdr: device %s is busy. Refuse to suspend.\n",
706 static const struct dev_pm_ops vmlogrdr_pm_ops
= {
707 .prepare
= vmlogrdr_pm_prepare
,
710 static struct class *vmlogrdr_class
;
711 static struct device_driver vmlogrdr_driver
= {
714 .pm
= &vmlogrdr_pm_ops
,
715 .groups
= vmlogrdr_drv_attr_groups
,
718 static int vmlogrdr_register_driver(void)
722 /* Register with iucv driver */
723 ret
= iucv_register(&vmlogrdr_iucv_handler
, 1);
727 ret
= driver_register(&vmlogrdr_driver
);
731 vmlogrdr_class
= class_create(THIS_MODULE
, "vmlogrdr");
732 if (IS_ERR(vmlogrdr_class
)) {
733 ret
= PTR_ERR(vmlogrdr_class
);
734 vmlogrdr_class
= NULL
;
740 driver_unregister(&vmlogrdr_driver
);
742 iucv_unregister(&vmlogrdr_iucv_handler
, 1);
748 static void vmlogrdr_unregister_driver(void)
750 class_destroy(vmlogrdr_class
);
751 vmlogrdr_class
= NULL
;
752 driver_unregister(&vmlogrdr_driver
);
753 iucv_unregister(&vmlogrdr_iucv_handler
, 1);
757 static int vmlogrdr_register_device(struct vmlogrdr_priv_t
*priv
)
762 dev
= kzalloc(sizeof(struct device
), GFP_KERNEL
);
764 dev_set_name(dev
, priv
->internal_name
);
765 dev
->bus
= &iucv_bus
;
766 dev
->parent
= iucv_root
;
767 dev
->driver
= &vmlogrdr_driver
;
768 dev
->groups
= vmlogrdr_attr_groups
;
769 dev_set_drvdata(dev
, priv
);
771 * The release function could be called after the
772 * module has been unloaded. It's _only_ task is to
773 * free the struct. Therefore, we specify kfree()
774 * directly here. (Probably a little bit obfuscating
777 dev
->release
= (void (*)(struct device
*))kfree
;
780 ret
= device_register(dev
);
786 priv
->class_device
= device_create(vmlogrdr_class
, dev
,
787 MKDEV(vmlogrdr_major
,
789 priv
, "%s", dev_name(dev
));
790 if (IS_ERR(priv
->class_device
)) {
791 ret
= PTR_ERR(priv
->class_device
);
792 priv
->class_device
=NULL
;
793 device_unregister(dev
);
801 static int vmlogrdr_unregister_device(struct vmlogrdr_priv_t
*priv
)
803 device_destroy(vmlogrdr_class
, MKDEV(vmlogrdr_major
, priv
->minor_num
));
804 if (priv
->device
!= NULL
) {
805 device_unregister(priv
->device
);
812 static int vmlogrdr_register_cdev(dev_t dev
)
815 vmlogrdr_cdev
= cdev_alloc();
816 if (!vmlogrdr_cdev
) {
819 vmlogrdr_cdev
->owner
= THIS_MODULE
;
820 vmlogrdr_cdev
->ops
= &vmlogrdr_fops
;
821 vmlogrdr_cdev
->dev
= dev
;
822 rc
= cdev_add(vmlogrdr_cdev
, vmlogrdr_cdev
->dev
, MAXMINOR
);
826 // cleanup: cdev is not fully registered, no cdev_del here!
827 kobject_put(&vmlogrdr_cdev
->kobj
);
833 static void vmlogrdr_cleanup(void)
838 cdev_del(vmlogrdr_cdev
);
841 for (i
=0; i
< MAXMINOR
; ++i
) {
842 vmlogrdr_unregister_device(&sys_ser
[i
]);
843 free_page((unsigned long)sys_ser
[i
].buffer
);
845 vmlogrdr_unregister_driver();
846 if (vmlogrdr_major
) {
847 unregister_chrdev_region(MKDEV(vmlogrdr_major
, 0), MAXMINOR
);
853 static int __init
vmlogrdr_init(void)
859 if (! MACHINE_IS_VM
) {
860 pr_err("not running under VM, driver not loaded.\n");
864 recording_class_AB
= vmlogrdr_get_recording_class_AB();
866 rc
= alloc_chrdev_region(&dev
, 0, MAXMINOR
, "vmlogrdr");
869 vmlogrdr_major
= MAJOR(dev
);
871 rc
=vmlogrdr_register_driver();
875 for (i
=0; i
< MAXMINOR
; ++i
) {
876 sys_ser
[i
].buffer
= (char *) get_zeroed_page(GFP_KERNEL
);
877 if (!sys_ser
[i
].buffer
) {
881 sys_ser
[i
].current_position
= sys_ser
[i
].buffer
;
882 rc
=vmlogrdr_register_device(&sys_ser
[i
]);
889 rc
= vmlogrdr_register_cdev(dev
);
900 static void __exit
vmlogrdr_exit(void)
907 module_init(vmlogrdr_init
);
908 module_exit(vmlogrdr_exit
);