| /* |
| * arch/s390/appldata/appldata_base.c |
| * |
| * Base infrastructure for Linux-z/VM Monitor Stream, Stage 1. |
| * Exports appldata_register_ops() and appldata_unregister_ops() for the |
| * data gathering modules. |
| * |
| * Copyright (C) 2003,2006 IBM Corporation, IBM Deutschland Entwicklung GmbH. |
| * |
| * Author: Gerald Schaefer <gerald.schaefer@de.ibm.com> |
| */ |
| |
| #include <linux/config.h> |
| #include <linux/module.h> |
| #include <linux/init.h> |
| #include <linux/slab.h> |
| #include <linux/errno.h> |
| #include <asm/uaccess.h> |
| #include <asm/io.h> |
| #include <asm/smp.h> |
| #include <linux/interrupt.h> |
| #include <linux/proc_fs.h> |
| #include <linux/page-flags.h> |
| #include <linux/swap.h> |
| #include <linux/pagemap.h> |
| #include <linux/sysctl.h> |
| #include <asm/timer.h> |
| //#include <linux/kernel_stat.h> |
| #include <linux/notifier.h> |
| #include <linux/cpu.h> |
| #include <linux/workqueue.h> |
| |
| #include "appldata.h" |
| |
| |
| #define MY_PRINT_NAME "appldata" /* for debug messages, etc. */ |
| #define APPLDATA_CPU_INTERVAL 10000 /* default (CPU) time for |
| sampling interval in |
| milliseconds */ |
| |
| #define TOD_MICRO 0x01000 /* nr. of TOD clock units |
| for 1 microsecond */ |
| |
| /* |
| * Parameter list for DIAGNOSE X'DC' |
| */ |
| #ifndef CONFIG_64BIT |
| struct appldata_parameter_list { |
| u16 diag; /* The DIAGNOSE code X'00DC' */ |
| u8 function; /* The function code for the DIAGNOSE */ |
| u8 parlist_length; /* Length of the parameter list */ |
| u32 product_id_addr; /* Address of the 16-byte product ID */ |
| u16 reserved; |
| u16 buffer_length; /* Length of the application data buffer */ |
| u32 buffer_addr; /* Address of the application data buffer */ |
| }; |
| #else |
| struct appldata_parameter_list { |
| u16 diag; |
| u8 function; |
| u8 parlist_length; |
| u32 unused01; |
| u16 reserved; |
| u16 buffer_length; |
| u32 unused02; |
| u64 product_id_addr; |
| u64 buffer_addr; |
| }; |
| #endif /* CONFIG_64BIT */ |
| |
| /* |
| * /proc entries (sysctl) |
| */ |
| static const char appldata_proc_name[APPLDATA_PROC_NAME_LENGTH] = "appldata"; |
| static int appldata_timer_handler(ctl_table *ctl, int write, struct file *filp, |
| void __user *buffer, size_t *lenp, loff_t *ppos); |
| static int appldata_interval_handler(ctl_table *ctl, int write, |
| struct file *filp, |
| void __user *buffer, |
| size_t *lenp, loff_t *ppos); |
| |
| static struct ctl_table_header *appldata_sysctl_header; |
| static struct ctl_table appldata_table[] = { |
| { |
| .ctl_name = CTL_APPLDATA_TIMER, |
| .procname = "timer", |
| .mode = S_IRUGO | S_IWUSR, |
| .proc_handler = &appldata_timer_handler, |
| }, |
| { |
| .ctl_name = CTL_APPLDATA_INTERVAL, |
| .procname = "interval", |
| .mode = S_IRUGO | S_IWUSR, |
| .proc_handler = &appldata_interval_handler, |
| }, |
| { .ctl_name = 0 } |
| }; |
| |
| static struct ctl_table appldata_dir_table[] = { |
| { |
| .ctl_name = CTL_APPLDATA, |
| .procname = appldata_proc_name, |
| .maxlen = 0, |
| .mode = S_IRUGO | S_IXUGO, |
| .child = appldata_table, |
| }, |
| { .ctl_name = 0 } |
| }; |
| |
| /* |
| * Timer |
| */ |
| DEFINE_PER_CPU(struct vtimer_list, appldata_timer); |
| static atomic_t appldata_expire_count = ATOMIC_INIT(0); |
| |
| static DEFINE_SPINLOCK(appldata_timer_lock); |
| static int appldata_interval = APPLDATA_CPU_INTERVAL; |
| static int appldata_timer_active; |
| |
| /* |
| * Work queue |
| */ |
| static struct workqueue_struct *appldata_wq; |
| static void appldata_work_fn(void *data); |
| static DECLARE_WORK(appldata_work, appldata_work_fn, NULL); |
| |
| |
| /* |
| * Ops list |
| */ |
| static DEFINE_SPINLOCK(appldata_ops_lock); |
| static LIST_HEAD(appldata_ops_list); |
| |
| |
| /*************************** timer, work, DIAG *******************************/ |
| /* |
| * appldata_timer_function() |
| * |
| * schedule work and reschedule timer |
| */ |
| static void appldata_timer_function(unsigned long data, struct pt_regs *regs) |
| { |
| P_DEBUG(" -= Timer =-\n"); |
| P_DEBUG("CPU: %i, expire_count: %i\n", smp_processor_id(), |
| atomic_read(&appldata_expire_count)); |
| if (atomic_dec_and_test(&appldata_expire_count)) { |
| atomic_set(&appldata_expire_count, num_online_cpus()); |
| queue_work(appldata_wq, (struct work_struct *) data); |
| } |
| } |
| |
| /* |
| * appldata_work_fn() |
| * |
| * call data gathering function for each (active) module |
| */ |
| static void appldata_work_fn(void *data) |
| { |
| struct list_head *lh; |
| struct appldata_ops *ops; |
| int i; |
| |
| P_DEBUG(" -= Work Queue =-\n"); |
| i = 0; |
| spin_lock(&appldata_ops_lock); |
| list_for_each(lh, &appldata_ops_list) { |
| ops = list_entry(lh, struct appldata_ops, list); |
| P_DEBUG("list_for_each loop: %i) active = %u, name = %s\n", |
| ++i, ops->active, ops->name); |
| if (ops->active == 1) { |
| ops->callback(ops->data); |
| } |
| } |
| spin_unlock(&appldata_ops_lock); |
| } |
| |
| /* |
| * appldata_diag() |
| * |
| * prepare parameter list, issue DIAG 0xDC |
| */ |
| int appldata_diag(char record_nr, u16 function, unsigned long buffer, |
| u16 length, char *mod_lvl) |
| { |
| unsigned long ry; |
| struct appldata_product_id { |
| char prod_nr[7]; /* product nr. */ |
| char prod_fn[2]; /* product function */ |
| char record_nr; /* record nr. */ |
| char version_nr[2]; /* version */ |
| char release_nr[2]; /* release */ |
| char mod_lvl[2]; /* modification lvl. */ |
| } appldata_product_id = { |
| /* all strings are EBCDIC, record_nr is byte */ |
| .prod_nr = {0xD3, 0xC9, 0xD5, 0xE4, |
| 0xE7, 0xD2, 0xD9}, /* "LINUXKR" */ |
| .prod_fn = {0xD5, 0xD3}, /* "NL" */ |
| .record_nr = record_nr, |
| .version_nr = {0xF2, 0xF6}, /* "26" */ |
| .release_nr = {0xF0, 0xF1}, /* "01" */ |
| .mod_lvl = {mod_lvl[0], mod_lvl[1]}, |
| }; |
| struct appldata_parameter_list appldata_parameter_list = { |
| .diag = 0xDC, |
| .function = function, |
| .parlist_length = |
| sizeof(appldata_parameter_list), |
| .buffer_length = length, |
| .product_id_addr = |
| (unsigned long) &appldata_product_id, |
| .buffer_addr = virt_to_phys((void *) buffer) |
| }; |
| |
| if (!MACHINE_IS_VM) |
| return -ENOSYS; |
| ry = -1; |
| asm volatile( |
| "diag %1,%0,0xDC\n\t" |
| : "=d" (ry) |
| : "d" (&appldata_parameter_list), |
| "m" (appldata_parameter_list), |
| "m" (appldata_product_id) |
| : "cc"); |
| return (int) ry; |
| } |
| /************************ timer, work, DIAG <END> ****************************/ |
| |
| |
| /****************************** /proc stuff **********************************/ |
| |
| /* |
| * appldata_mod_vtimer_wrap() |
| * |
| * wrapper function for mod_virt_timer(), because smp_call_function_on() |
| * accepts only one parameter. |
| */ |
| static void __appldata_mod_vtimer_wrap(void *p) { |
| struct { |
| struct vtimer_list *timer; |
| u64 expires; |
| } *args = p; |
| mod_virt_timer(args->timer, args->expires); |
| } |
| |
| #define APPLDATA_ADD_TIMER 0 |
| #define APPLDATA_DEL_TIMER 1 |
| #define APPLDATA_MOD_TIMER 2 |
| |
| /* |
| * __appldata_vtimer_setup() |
| * |
| * Add, delete or modify virtual timers on all online cpus. |
| * The caller needs to get the appldata_timer_lock spinlock. |
| */ |
| static void |
| __appldata_vtimer_setup(int cmd) |
| { |
| u64 per_cpu_interval; |
| int i; |
| |
| switch (cmd) { |
| case APPLDATA_ADD_TIMER: |
| if (appldata_timer_active) |
| break; |
| per_cpu_interval = (u64) (appldata_interval*1000 / |
| num_online_cpus()) * TOD_MICRO; |
| for_each_online_cpu(i) { |
| per_cpu(appldata_timer, i).expires = per_cpu_interval; |
| smp_call_function_on(add_virt_timer_periodic, |
| &per_cpu(appldata_timer, i), |
| 0, 1, i); |
| } |
| appldata_timer_active = 1; |
| P_INFO("Monitoring timer started.\n"); |
| break; |
| case APPLDATA_DEL_TIMER: |
| for_each_online_cpu(i) |
| del_virt_timer(&per_cpu(appldata_timer, i)); |
| if (!appldata_timer_active) |
| break; |
| appldata_timer_active = 0; |
| atomic_set(&appldata_expire_count, num_online_cpus()); |
| P_INFO("Monitoring timer stopped.\n"); |
| break; |
| case APPLDATA_MOD_TIMER: |
| per_cpu_interval = (u64) (appldata_interval*1000 / |
| num_online_cpus()) * TOD_MICRO; |
| if (!appldata_timer_active) |
| break; |
| for_each_online_cpu(i) { |
| struct { |
| struct vtimer_list *timer; |
| u64 expires; |
| } args; |
| args.timer = &per_cpu(appldata_timer, i); |
| args.expires = per_cpu_interval; |
| smp_call_function_on(__appldata_mod_vtimer_wrap, |
| &args, 0, 1, i); |
| } |
| } |
| } |
| |
| /* |
| * appldata_timer_handler() |
| * |
| * Start/Stop timer, show status of timer (0 = not active, 1 = active) |
| */ |
| static int |
| appldata_timer_handler(ctl_table *ctl, int write, struct file *filp, |
| void __user *buffer, size_t *lenp, loff_t *ppos) |
| { |
| int len; |
| char buf[2]; |
| |
| if (!*lenp || *ppos) { |
| *lenp = 0; |
| return 0; |
| } |
| if (!write) { |
| len = sprintf(buf, appldata_timer_active ? "1\n" : "0\n"); |
| if (len > *lenp) |
| len = *lenp; |
| if (copy_to_user(buffer, buf, len)) |
| return -EFAULT; |
| goto out; |
| } |
| len = *lenp; |
| if (copy_from_user(buf, buffer, len > sizeof(buf) ? sizeof(buf) : len)) |
| return -EFAULT; |
| spin_lock(&appldata_timer_lock); |
| if (buf[0] == '1') |
| __appldata_vtimer_setup(APPLDATA_ADD_TIMER); |
| else if (buf[0] == '0') |
| __appldata_vtimer_setup(APPLDATA_DEL_TIMER); |
| spin_unlock(&appldata_timer_lock); |
| out: |
| *lenp = len; |
| *ppos += len; |
| return 0; |
| } |
| |
| /* |
| * appldata_interval_handler() |
| * |
| * Set (CPU) timer interval for collection of data (in milliseconds), show |
| * current timer interval. |
| */ |
| static int |
| appldata_interval_handler(ctl_table *ctl, int write, struct file *filp, |
| void __user *buffer, size_t *lenp, loff_t *ppos) |
| { |
| int len, interval; |
| char buf[16]; |
| |
| if (!*lenp || *ppos) { |
| *lenp = 0; |
| return 0; |
| } |
| if (!write) { |
| len = sprintf(buf, "%i\n", appldata_interval); |
| if (len > *lenp) |
| len = *lenp; |
| if (copy_to_user(buffer, buf, len)) |
| return -EFAULT; |
| goto out; |
| } |
| len = *lenp; |
| if (copy_from_user(buf, buffer, len > sizeof(buf) ? sizeof(buf) : len)) { |
| return -EFAULT; |
| } |
| sscanf(buf, "%i", &interval); |
| if (interval <= 0) { |
| P_ERROR("Timer CPU interval has to be > 0!\n"); |
| return -EINVAL; |
| } |
| |
| spin_lock(&appldata_timer_lock); |
| appldata_interval = interval; |
| __appldata_vtimer_setup(APPLDATA_MOD_TIMER); |
| spin_unlock(&appldata_timer_lock); |
| |
| P_INFO("Monitoring CPU interval set to %u milliseconds.\n", |
| interval); |
| out: |
| *lenp = len; |
| *ppos += len; |
| return 0; |
| } |
| |
| /* |
| * appldata_generic_handler() |
| * |
| * Generic start/stop monitoring and DIAG, show status of |
| * monitoring (0 = not in process, 1 = in process) |
| */ |
| static int |
| appldata_generic_handler(ctl_table *ctl, int write, struct file *filp, |
| void __user *buffer, size_t *lenp, loff_t *ppos) |
| { |
| struct appldata_ops *ops = NULL, *tmp_ops; |
| int rc, len, found; |
| char buf[2]; |
| struct list_head *lh; |
| |
| found = 0; |
| spin_lock(&appldata_ops_lock); |
| list_for_each(lh, &appldata_ops_list) { |
| tmp_ops = list_entry(lh, struct appldata_ops, list); |
| if (&tmp_ops->ctl_table[2] == ctl) { |
| found = 1; |
| } |
| } |
| if (!found) { |
| spin_unlock(&appldata_ops_lock); |
| return -ENODEV; |
| } |
| ops = ctl->data; |
| if (!try_module_get(ops->owner)) { // protect this function |
| spin_unlock(&appldata_ops_lock); |
| return -ENODEV; |
| } |
| spin_unlock(&appldata_ops_lock); |
| |
| if (!*lenp || *ppos) { |
| *lenp = 0; |
| module_put(ops->owner); |
| return 0; |
| } |
| if (!write) { |
| len = sprintf(buf, ops->active ? "1\n" : "0\n"); |
| if (len > *lenp) |
| len = *lenp; |
| if (copy_to_user(buffer, buf, len)) { |
| module_put(ops->owner); |
| return -EFAULT; |
| } |
| goto out; |
| } |
| len = *lenp; |
| if (copy_from_user(buf, buffer, |
| len > sizeof(buf) ? sizeof(buf) : len)) { |
| module_put(ops->owner); |
| return -EFAULT; |
| } |
| |
| spin_lock(&appldata_ops_lock); |
| if ((buf[0] == '1') && (ops->active == 0)) { |
| // protect work queue callback |
| if (!try_module_get(ops->owner)) { |
| spin_unlock(&appldata_ops_lock); |
| module_put(ops->owner); |
| return -ENODEV; |
| } |
| ops->callback(ops->data); // init record |
| rc = appldata_diag(ops->record_nr, |
| APPLDATA_START_INTERVAL_REC, |
| (unsigned long) ops->data, ops->size, |
| ops->mod_lvl); |
| if (rc != 0) { |
| P_ERROR("START DIAG 0xDC for %s failed, " |
| "return code: %d\n", ops->name, rc); |
| module_put(ops->owner); |
| } else { |
| P_INFO("Monitoring %s data enabled, " |
| "DIAG 0xDC started.\n", ops->name); |
| ops->active = 1; |
| } |
| } else if ((buf[0] == '0') && (ops->active == 1)) { |
| ops->active = 0; |
| rc = appldata_diag(ops->record_nr, APPLDATA_STOP_REC, |
| (unsigned long) ops->data, ops->size, |
| ops->mod_lvl); |
| if (rc != 0) { |
| P_ERROR("STOP DIAG 0xDC for %s failed, " |
| "return code: %d\n", ops->name, rc); |
| } else { |
| P_INFO("Monitoring %s data disabled, " |
| "DIAG 0xDC stopped.\n", ops->name); |
| } |
| module_put(ops->owner); |
| } |
| spin_unlock(&appldata_ops_lock); |
| out: |
| *lenp = len; |
| *ppos += len; |
| module_put(ops->owner); |
| return 0; |
| } |
| |
| /*************************** /proc stuff <END> *******************************/ |
| |
| |
| /************************* module-ops management *****************************/ |
| /* |
| * appldata_register_ops() |
| * |
| * update ops list, register /proc/sys entries |
| */ |
| int appldata_register_ops(struct appldata_ops *ops) |
| { |
| struct list_head *lh; |
| struct appldata_ops *tmp_ops; |
| int i; |
| |
| i = 0; |
| |
| if ((ops->size > APPLDATA_MAX_REC_SIZE) || |
| (ops->size < 0)){ |
| P_ERROR("Invalid size of %s record = %i, maximum = %i!\n", |
| ops->name, ops->size, APPLDATA_MAX_REC_SIZE); |
| return -ENOMEM; |
| } |
| if ((ops->ctl_nr == CTL_APPLDATA) || |
| (ops->ctl_nr == CTL_APPLDATA_TIMER) || |
| (ops->ctl_nr == CTL_APPLDATA_INTERVAL)) { |
| P_ERROR("ctl_nr %i already in use!\n", ops->ctl_nr); |
| return -EBUSY; |
| } |
| ops->ctl_table = kzalloc(4*sizeof(struct ctl_table), GFP_KERNEL); |
| if (ops->ctl_table == NULL) { |
| P_ERROR("Not enough memory for %s ctl_table!\n", ops->name); |
| return -ENOMEM; |
| } |
| |
| spin_lock(&appldata_ops_lock); |
| list_for_each(lh, &appldata_ops_list) { |
| tmp_ops = list_entry(lh, struct appldata_ops, list); |
| P_DEBUG("register_ops loop: %i) name = %s, ctl = %i\n", |
| ++i, tmp_ops->name, tmp_ops->ctl_nr); |
| P_DEBUG("Comparing %s (ctl %i) with %s (ctl %i)\n", |
| tmp_ops->name, tmp_ops->ctl_nr, ops->name, |
| ops->ctl_nr); |
| if (strncmp(tmp_ops->name, ops->name, |
| APPLDATA_PROC_NAME_LENGTH) == 0) { |
| P_ERROR("Name \"%s\" already registered!\n", ops->name); |
| kfree(ops->ctl_table); |
| spin_unlock(&appldata_ops_lock); |
| return -EBUSY; |
| } |
| if (tmp_ops->ctl_nr == ops->ctl_nr) { |
| P_ERROR("ctl_nr %i already registered!\n", ops->ctl_nr); |
| kfree(ops->ctl_table); |
| spin_unlock(&appldata_ops_lock); |
| return -EBUSY; |
| } |
| } |
| list_add(&ops->list, &appldata_ops_list); |
| spin_unlock(&appldata_ops_lock); |
| |
| ops->ctl_table[0].ctl_name = CTL_APPLDATA; |
| ops->ctl_table[0].procname = appldata_proc_name; |
| ops->ctl_table[0].maxlen = 0; |
| ops->ctl_table[0].mode = S_IRUGO | S_IXUGO; |
| ops->ctl_table[0].child = &ops->ctl_table[2]; |
| |
| ops->ctl_table[1].ctl_name = 0; |
| |
| ops->ctl_table[2].ctl_name = ops->ctl_nr; |
| ops->ctl_table[2].procname = ops->name; |
| ops->ctl_table[2].mode = S_IRUGO | S_IWUSR; |
| ops->ctl_table[2].proc_handler = appldata_generic_handler; |
| ops->ctl_table[2].data = ops; |
| |
| ops->ctl_table[3].ctl_name = 0; |
| |
| ops->sysctl_header = register_sysctl_table(ops->ctl_table,1); |
| |
| P_INFO("%s-ops registered!\n", ops->name); |
| return 0; |
| } |
| |
| /* |
| * appldata_unregister_ops() |
| * |
| * update ops list, unregister /proc entries, stop DIAG if necessary |
| */ |
| void appldata_unregister_ops(struct appldata_ops *ops) |
| { |
| void *table; |
| spin_lock(&appldata_ops_lock); |
| list_del(&ops->list); |
| /* at that point any incoming access will fail */ |
| table = ops->ctl_table; |
| ops->ctl_table = NULL; |
| spin_unlock(&appldata_ops_lock); |
| unregister_sysctl_table(ops->sysctl_header); |
| kfree(table); |
| P_INFO("%s-ops unregistered!\n", ops->name); |
| } |
| /********************** module-ops management <END> **************************/ |
| |
| |
| /******************************* init / exit *********************************/ |
| |
| static void |
| appldata_online_cpu(int cpu) |
| { |
| init_virt_timer(&per_cpu(appldata_timer, cpu)); |
| per_cpu(appldata_timer, cpu).function = appldata_timer_function; |
| per_cpu(appldata_timer, cpu).data = (unsigned long) |
| &appldata_work; |
| atomic_inc(&appldata_expire_count); |
| spin_lock(&appldata_timer_lock); |
| __appldata_vtimer_setup(APPLDATA_MOD_TIMER); |
| spin_unlock(&appldata_timer_lock); |
| } |
| |
| static void |
| appldata_offline_cpu(int cpu) |
| { |
| del_virt_timer(&per_cpu(appldata_timer, cpu)); |
| if (atomic_dec_and_test(&appldata_expire_count)) { |
| atomic_set(&appldata_expire_count, num_online_cpus()); |
| queue_work(appldata_wq, &appldata_work); |
| } |
| spin_lock(&appldata_timer_lock); |
| __appldata_vtimer_setup(APPLDATA_MOD_TIMER); |
| spin_unlock(&appldata_timer_lock); |
| } |
| |
| static int __cpuinit |
| appldata_cpu_notify(struct notifier_block *self, |
| unsigned long action, void *hcpu) |
| { |
| switch (action) { |
| case CPU_ONLINE: |
| appldata_online_cpu((long) hcpu); |
| break; |
| #ifdef CONFIG_HOTPLUG_CPU |
| case CPU_DEAD: |
| appldata_offline_cpu((long) hcpu); |
| break; |
| #endif |
| default: |
| break; |
| } |
| return NOTIFY_OK; |
| } |
| |
| static struct notifier_block __devinitdata appldata_nb = { |
| .notifier_call = appldata_cpu_notify, |
| }; |
| |
| /* |
| * appldata_init() |
| * |
| * init timer, register /proc entries |
| */ |
| static int __init appldata_init(void) |
| { |
| int i; |
| |
| P_DEBUG("sizeof(parameter_list) = %lu\n", |
| sizeof(struct appldata_parameter_list)); |
| |
| appldata_wq = create_singlethread_workqueue("appldata"); |
| if (!appldata_wq) { |
| P_ERROR("Could not create work queue\n"); |
| return -ENOMEM; |
| } |
| |
| for_each_online_cpu(i) |
| appldata_online_cpu(i); |
| |
| /* Register cpu hotplug notifier */ |
| register_cpu_notifier(&appldata_nb); |
| |
| appldata_sysctl_header = register_sysctl_table(appldata_dir_table, 1); |
| #ifdef MODULE |
| appldata_dir_table[0].de->owner = THIS_MODULE; |
| appldata_table[0].de->owner = THIS_MODULE; |
| appldata_table[1].de->owner = THIS_MODULE; |
| #endif |
| |
| P_DEBUG("Base interface initialized.\n"); |
| return 0; |
| } |
| |
| /* |
| * appldata_exit() |
| * |
| * stop timer, unregister /proc entries |
| */ |
| static void __exit appldata_exit(void) |
| { |
| struct list_head *lh; |
| struct appldata_ops *ops; |
| int rc, i; |
| |
| P_DEBUG("Unloading module ...\n"); |
| /* |
| * ops list should be empty, but just in case something went wrong... |
| */ |
| spin_lock(&appldata_ops_lock); |
| list_for_each(lh, &appldata_ops_list) { |
| ops = list_entry(lh, struct appldata_ops, list); |
| rc = appldata_diag(ops->record_nr, APPLDATA_STOP_REC, |
| (unsigned long) ops->data, ops->size, |
| ops->mod_lvl); |
| if (rc != 0) { |
| P_ERROR("STOP DIAG 0xDC for %s failed, " |
| "return code: %d\n", ops->name, rc); |
| } |
| } |
| spin_unlock(&appldata_ops_lock); |
| |
| for_each_online_cpu(i) |
| appldata_offline_cpu(i); |
| |
| appldata_timer_active = 0; |
| |
| unregister_sysctl_table(appldata_sysctl_header); |
| |
| destroy_workqueue(appldata_wq); |
| P_DEBUG("... module unloaded!\n"); |
| } |
| /**************************** init / exit <END> ******************************/ |
| |
| |
| module_init(appldata_init); |
| module_exit(appldata_exit); |
| MODULE_LICENSE("GPL"); |
| MODULE_AUTHOR("Gerald Schaefer"); |
| MODULE_DESCRIPTION("Linux-VM Monitor Stream, base infrastructure"); |
| |
| EXPORT_SYMBOL_GPL(appldata_register_ops); |
| EXPORT_SYMBOL_GPL(appldata_unregister_ops); |
| EXPORT_SYMBOL_GPL(appldata_diag); |
| |
| #ifdef MODULE |
| /* |
| * Kernel symbols needed by appldata_mem and appldata_os modules. |
| * However, if this file is compiled as a module (for testing only), these |
| * symbols are not exported. In this case, we define them locally and export |
| * those. |
| */ |
| void si_swapinfo(struct sysinfo *val) |
| { |
| val->freeswap = -1ul; |
| val->totalswap = -1ul; |
| } |
| |
| unsigned long avenrun[3] = {-1 - FIXED_1/200, -1 - FIXED_1/200, |
| -1 - FIXED_1/200}; |
| int nr_threads = -1; |
| |
| void get_full_page_state(struct page_state *ps) |
| { |
| memset(ps, -1, sizeof(struct page_state)); |
| } |
| |
| unsigned long nr_running(void) |
| { |
| return -1; |
| } |
| |
| unsigned long nr_iowait(void) |
| { |
| return -1; |
| } |
| |
| /*unsigned long nr_context_switches(void) |
| { |
| return -1; |
| }*/ |
| #endif /* MODULE */ |
| EXPORT_SYMBOL_GPL(si_swapinfo); |
| EXPORT_SYMBOL_GPL(nr_threads); |
| EXPORT_SYMBOL_GPL(nr_running); |
| EXPORT_SYMBOL_GPL(nr_iowait); |
| //EXPORT_SYMBOL_GPL(nr_context_switches); |