| /* |
| * linux/kernel/power/swsusp.c |
| * |
| * This file provides code to write suspend image to swap and read it back. |
| * |
| * Copyright (C) 1998-2001 Gabor Kuti <seasons@fornax.hu> |
| * Copyright (C) 1998,2001-2005 Pavel Machek <pavel@suse.cz> |
| * |
| * This file is released under the GPLv2. |
| * |
| * I'd like to thank the following people for their work: |
| * |
| * Pavel Machek <pavel@ucw.cz>: |
| * Modifications, defectiveness pointing, being with me at the very beginning, |
| * suspend to swap space, stop all tasks. Port to 2.4.18-ac and 2.5.17. |
| * |
| * Steve Doddi <dirk@loth.demon.co.uk>: |
| * Support the possibility of hardware state restoring. |
| * |
| * Raph <grey.havens@earthling.net>: |
| * Support for preserving states of network devices and virtual console |
| * (including X and svgatextmode) |
| * |
| * Kurt Garloff <garloff@suse.de>: |
| * Straightened the critical function in order to prevent compilers from |
| * playing tricks with local variables. |
| * |
| * Andreas Mohr <a.mohr@mailto.de> |
| * |
| * Alex Badea <vampire@go.ro>: |
| * Fixed runaway init |
| * |
| * Rafael J. Wysocki <rjw@sisk.pl> |
| * Added the swap map data structure and reworked the handling of swap |
| * |
| * More state savers are welcome. Especially for the scsi layer... |
| * |
| * For TODOs,FIXMEs also look in Documentation/power/swsusp.txt |
| */ |
| |
| #include <linux/module.h> |
| #include <linux/mm.h> |
| #include <linux/suspend.h> |
| #include <linux/smp_lock.h> |
| #include <linux/file.h> |
| #include <linux/utsname.h> |
| #include <linux/version.h> |
| #include <linux/delay.h> |
| #include <linux/bitops.h> |
| #include <linux/spinlock.h> |
| #include <linux/genhd.h> |
| #include <linux/kernel.h> |
| #include <linux/major.h> |
| #include <linux/swap.h> |
| #include <linux/pm.h> |
| #include <linux/device.h> |
| #include <linux/buffer_head.h> |
| #include <linux/swapops.h> |
| #include <linux/bootmem.h> |
| #include <linux/syscalls.h> |
| #include <linux/highmem.h> |
| #include <linux/bio.h> |
| |
| #include <asm/uaccess.h> |
| #include <asm/mmu_context.h> |
| #include <asm/pgtable.h> |
| #include <asm/tlbflush.h> |
| #include <asm/io.h> |
| |
| #include "power.h" |
| |
| /* |
| * Preferred image size in bytes (tunable via /sys/power/image_size). |
| * When it is set to N, swsusp will do its best to ensure the image |
| * size will not exceed N bytes, but if that is impossible, it will |
| * try to create the smallest image possible. |
| */ |
| unsigned long image_size = 500 * 1024 * 1024; |
| |
| #ifdef CONFIG_HIGHMEM |
| unsigned int count_highmem_pages(void); |
| int save_highmem(void); |
| int restore_highmem(void); |
| #else |
| static int save_highmem(void) { return 0; } |
| static int restore_highmem(void) { return 0; } |
| static unsigned int count_highmem_pages(void) { return 0; } |
| #endif |
| |
| extern char resume_file[]; |
| |
| #define SWSUSP_SIG "S1SUSPEND" |
| |
| static struct swsusp_header { |
| char reserved[PAGE_SIZE - 20 - sizeof(swp_entry_t)]; |
| swp_entry_t image; |
| char orig_sig[10]; |
| char sig[10]; |
| } __attribute__((packed, aligned(PAGE_SIZE))) swsusp_header; |
| |
| static struct swsusp_info swsusp_info; |
| |
| /* |
| * Saving part... |
| */ |
| |
| static unsigned short root_swap = 0xffff; |
| |
| static int mark_swapfiles(swp_entry_t start) |
| { |
| int error; |
| |
| rw_swap_page_sync(READ, |
| swp_entry(root_swap, 0), |
| virt_to_page((unsigned long)&swsusp_header)); |
| if (!memcmp("SWAP-SPACE",swsusp_header.sig, 10) || |
| !memcmp("SWAPSPACE2",swsusp_header.sig, 10)) { |
| memcpy(swsusp_header.orig_sig,swsusp_header.sig, 10); |
| memcpy(swsusp_header.sig,SWSUSP_SIG, 10); |
| swsusp_header.image = start; |
| error = rw_swap_page_sync(WRITE, |
| swp_entry(root_swap, 0), |
| virt_to_page((unsigned long) |
| &swsusp_header)); |
| } else { |
| pr_debug("swsusp: Partition is not swap space.\n"); |
| error = -ENODEV; |
| } |
| return error; |
| } |
| |
| /* |
| * Check whether the swap device is the specified resume |
| * device, irrespective of whether they are specified by |
| * identical names. |
| * |
| * (Thus, device inode aliasing is allowed. You can say /dev/hda4 |
| * instead of /dev/ide/host0/bus0/target0/lun0/part4 [if using devfs] |
| * and they'll be considered the same device. This is *necessary* for |
| * devfs, since the resume code can only recognize the form /dev/hda4, |
| * but the suspend code would see the long name.) |
| */ |
| static inline int is_resume_device(const struct swap_info_struct *swap_info) |
| { |
| struct file *file = swap_info->swap_file; |
| struct inode *inode = file->f_dentry->d_inode; |
| |
| return S_ISBLK(inode->i_mode) && |
| swsusp_resume_device == MKDEV(imajor(inode), iminor(inode)); |
| } |
| |
| static int swsusp_swap_check(void) /* This is called before saving image */ |
| { |
| int i; |
| |
| spin_lock(&swap_lock); |
| for (i = 0; i < MAX_SWAPFILES; i++) { |
| if (!(swap_info[i].flags & SWP_WRITEOK)) |
| continue; |
| if (!swsusp_resume_device || is_resume_device(swap_info + i)) { |
| spin_unlock(&swap_lock); |
| root_swap = i; |
| return 0; |
| } |
| } |
| spin_unlock(&swap_lock); |
| return -ENODEV; |
| } |
| |
| /** |
| * write_page - Write one page to a fresh swap location. |
| * @addr: Address we're writing. |
| * @loc: Place to store the entry we used. |
| * |
| * Allocate a new swap entry and 'sync' it. Note we discard -EIO |
| * errors. That is an artifact left over from swsusp. It did not |
| * check the return of rw_swap_page_sync() at all, since most pages |
| * written back to swap would return -EIO. |
| * This is a partial improvement, since we will at least return other |
| * errors, though we need to eventually fix the damn code. |
| */ |
| static int write_page(unsigned long addr, swp_entry_t *loc) |
| { |
| swp_entry_t entry; |
| int error = -ENOSPC; |
| |
| entry = get_swap_page_of_type(root_swap); |
| if (swp_offset(entry)) { |
| error = rw_swap_page_sync(WRITE, entry, virt_to_page(addr)); |
| if (!error || error == -EIO) |
| *loc = entry; |
| } |
| return error; |
| } |
| |
| /** |
| * Swap map-handling functions |
| * |
| * The swap map is a data structure used for keeping track of each page |
| * written to the swap. It consists of many swap_map_page structures |
| * that contain each an array of MAP_PAGE_SIZE swap entries. |
| * These structures are linked together with the help of either the |
| * .next (in memory) or the .next_swap (in swap) member. |
| * |
| * The swap map is created during suspend. At that time we need to keep |
| * it in memory, because we have to free all of the allocated swap |
| * entries if an error occurs. The memory needed is preallocated |
| * so that we know in advance if there's enough of it. |
| * |
| * The first swap_map_page structure is filled with the swap entries that |
| * correspond to the first MAP_PAGE_SIZE data pages written to swap and |
| * so on. After the all of the data pages have been written, the order |
| * of the swap_map_page structures in the map is reversed so that they |
| * can be read from swap in the original order. This causes the data |
| * pages to be loaded in exactly the same order in which they have been |
| * saved. |
| * |
| * During resume we only need to use one swap_map_page structure |
| * at a time, which means that we only need to use two memory pages for |
| * reading the image - one for reading the swap_map_page structures |
| * and the second for reading the data pages from swap. |
| */ |
| |
| #define MAP_PAGE_SIZE ((PAGE_SIZE - sizeof(swp_entry_t) - sizeof(void *)) \ |
| / sizeof(swp_entry_t)) |
| |
| struct swap_map_page { |
| swp_entry_t entries[MAP_PAGE_SIZE]; |
| swp_entry_t next_swap; |
| struct swap_map_page *next; |
| }; |
| |
| static inline void free_swap_map(struct swap_map_page *swap_map) |
| { |
| struct swap_map_page *swp; |
| |
| while (swap_map) { |
| swp = swap_map->next; |
| free_page((unsigned long)swap_map); |
| swap_map = swp; |
| } |
| } |
| |
| static struct swap_map_page *alloc_swap_map(unsigned int nr_pages) |
| { |
| struct swap_map_page *swap_map, *swp; |
| unsigned n = 0; |
| |
| if (!nr_pages) |
| return NULL; |
| |
| pr_debug("alloc_swap_map(): nr_pages = %d\n", nr_pages); |
| swap_map = (struct swap_map_page *)get_zeroed_page(GFP_ATOMIC); |
| swp = swap_map; |
| for (n = MAP_PAGE_SIZE; n < nr_pages; n += MAP_PAGE_SIZE) { |
| swp->next = (struct swap_map_page *)get_zeroed_page(GFP_ATOMIC); |
| swp = swp->next; |
| if (!swp) { |
| free_swap_map(swap_map); |
| return NULL; |
| } |
| } |
| return swap_map; |
| } |
| |
| /** |
| * reverse_swap_map - reverse the order of pages in the swap map |
| * @swap_map |
| */ |
| |
| static inline struct swap_map_page *reverse_swap_map(struct swap_map_page *swap_map) |
| { |
| struct swap_map_page *prev, *next; |
| |
| prev = NULL; |
| while (swap_map) { |
| next = swap_map->next; |
| swap_map->next = prev; |
| prev = swap_map; |
| swap_map = next; |
| } |
| return prev; |
| } |
| |
| /** |
| * free_swap_map_entries - free the swap entries allocated to store |
| * the swap map @swap_map (this is only called in case of an error) |
| */ |
| static inline void free_swap_map_entries(struct swap_map_page *swap_map) |
| { |
| while (swap_map) { |
| if (swap_map->next_swap.val) |
| swap_free(swap_map->next_swap); |
| swap_map = swap_map->next; |
| } |
| } |
| |
| /** |
| * save_swap_map - save the swap map used for tracing the data pages |
| * stored in the swap |
| */ |
| |
| static int save_swap_map(struct swap_map_page *swap_map, swp_entry_t *start) |
| { |
| swp_entry_t entry = (swp_entry_t){0}; |
| int error; |
| |
| while (swap_map) { |
| swap_map->next_swap = entry; |
| if ((error = write_page((unsigned long)swap_map, &entry))) |
| return error; |
| swap_map = swap_map->next; |
| } |
| *start = entry; |
| return 0; |
| } |
| |
| /** |
| * free_image_entries - free the swap entries allocated to store |
| * the image data pages (this is only called in case of an error) |
| */ |
| |
| static inline void free_image_entries(struct swap_map_page *swp) |
| { |
| unsigned k; |
| |
| while (swp) { |
| for (k = 0; k < MAP_PAGE_SIZE; k++) |
| if (swp->entries[k].val) |
| swap_free(swp->entries[k]); |
| swp = swp->next; |
| } |
| } |
| |
| /** |
| * The swap_map_handle structure is used for handling the swap map in |
| * a file-alike way |
| */ |
| |
| struct swap_map_handle { |
| struct swap_map_page *cur; |
| unsigned int k; |
| }; |
| |
| static inline void init_swap_map_handle(struct swap_map_handle *handle, |
| struct swap_map_page *map) |
| { |
| handle->cur = map; |
| handle->k = 0; |
| } |
| |
| static inline int swap_map_write_page(struct swap_map_handle *handle, |
| unsigned long addr) |
| { |
| int error; |
| |
| error = write_page(addr, handle->cur->entries + handle->k); |
| if (error) |
| return error; |
| if (++handle->k >= MAP_PAGE_SIZE) { |
| handle->cur = handle->cur->next; |
| handle->k = 0; |
| } |
| return 0; |
| } |
| |
| /** |
| * save_image_data - save the data pages pointed to by the PBEs |
| * from the list @pblist using the swap map handle @handle |
| * (assume there are @nr_pages data pages to save) |
| */ |
| |
| static int save_image_data(struct pbe *pblist, |
| struct swap_map_handle *handle, |
| unsigned int nr_pages) |
| { |
| unsigned int m; |
| struct pbe *p; |
| int error = 0; |
| |
| printk("Saving image data pages (%u pages) ... ", nr_pages); |
| m = nr_pages / 100; |
| if (!m) |
| m = 1; |
| nr_pages = 0; |
| for_each_pbe (p, pblist) { |
| error = swap_map_write_page(handle, p->address); |
| if (error) |
| break; |
| if (!(nr_pages % m)) |
| printk("\b\b\b\b%3d%%", nr_pages / m); |
| nr_pages++; |
| } |
| if (!error) |
| printk("\b\b\b\bdone\n"); |
| return error; |
| } |
| |
| static void dump_info(void) |
| { |
| pr_debug(" swsusp: Version: %u\n",swsusp_info.version_code); |
| pr_debug(" swsusp: Num Pages: %ld\n",swsusp_info.num_physpages); |
| pr_debug(" swsusp: UTS Sys: %s\n",swsusp_info.uts.sysname); |
| pr_debug(" swsusp: UTS Node: %s\n",swsusp_info.uts.nodename); |
| pr_debug(" swsusp: UTS Release: %s\n",swsusp_info.uts.release); |
| pr_debug(" swsusp: UTS Version: %s\n",swsusp_info.uts.version); |
| pr_debug(" swsusp: UTS Machine: %s\n",swsusp_info.uts.machine); |
| pr_debug(" swsusp: UTS Domain: %s\n",swsusp_info.uts.domainname); |
| pr_debug(" swsusp: CPUs: %d\n",swsusp_info.cpus); |
| pr_debug(" swsusp: Image: %ld Pages\n",swsusp_info.image_pages); |
| pr_debug(" swsusp: Total: %ld Pages\n", swsusp_info.pages); |
| } |
| |
| static void init_header(unsigned int nr_pages) |
| { |
| memset(&swsusp_info, 0, sizeof(swsusp_info)); |
| swsusp_info.version_code = LINUX_VERSION_CODE; |
| swsusp_info.num_physpages = num_physpages; |
| memcpy(&swsusp_info.uts, &system_utsname, sizeof(system_utsname)); |
| |
| swsusp_info.cpus = num_online_cpus(); |
| swsusp_info.image_pages = nr_pages; |
| swsusp_info.pages = nr_pages + |
| ((nr_pages * sizeof(long) + PAGE_SIZE - 1) >> PAGE_SHIFT) + 1; |
| } |
| |
| /** |
| * pack_orig_addresses - the .orig_address fields of the PBEs from the |
| * list starting at @pbe are stored in the array @buf[] (1 page) |
| */ |
| |
| static inline struct pbe *pack_orig_addresses(unsigned long *buf, |
| struct pbe *pbe) |
| { |
| int j; |
| |
| for (j = 0; j < PAGE_SIZE / sizeof(long) && pbe; j++) { |
| buf[j] = pbe->orig_address; |
| pbe = pbe->next; |
| } |
| if (!pbe) |
| for (; j < PAGE_SIZE / sizeof(long); j++) |
| buf[j] = 0; |
| return pbe; |
| } |
| |
| /** |
| * save_image_metadata - save the .orig_address fields of the PBEs |
| * from the list @pblist using the swap map handle @handle |
| */ |
| |
| static int save_image_metadata(struct pbe *pblist, |
| struct swap_map_handle *handle) |
| { |
| unsigned long *buf; |
| unsigned int n = 0; |
| struct pbe *p; |
| int error = 0; |
| |
| printk("Saving image metadata ... "); |
| buf = (unsigned long *)get_zeroed_page(GFP_ATOMIC); |
| if (!buf) |
| return -ENOMEM; |
| p = pblist; |
| while (p) { |
| p = pack_orig_addresses(buf, p); |
| error = swap_map_write_page(handle, (unsigned long)buf); |
| if (error) |
| break; |
| n++; |
| } |
| free_page((unsigned long)buf); |
| if (!error) |
| printk("done (%u pages saved)\n", n); |
| return error; |
| } |
| |
| /** |
| * enough_swap - Make sure we have enough swap to save the image. |
| * |
| * Returns TRUE or FALSE after checking the total amount of swap |
| * space avaiable from the resume partition. |
| */ |
| |
| static int enough_swap(unsigned int nr_pages) |
| { |
| unsigned int free_swap = swap_info[root_swap].pages - |
| swap_info[root_swap].inuse_pages; |
| |
| pr_debug("swsusp: free swap pages: %u\n", free_swap); |
| return free_swap > (nr_pages + PAGES_FOR_IO + |
| (nr_pages + PBES_PER_PAGE - 1) / PBES_PER_PAGE); |
| } |
| |
| /** |
| * swsusp_write - Write entire image and metadata. |
| * |
| * It is important _NOT_ to umount filesystems at this point. We want |
| * them synced (in case something goes wrong) but we DO not want to mark |
| * filesystem clean: it is not. (And it does not matter, if we resume |
| * correctly, we'll mark system clean, anyway.) |
| */ |
| |
| int swsusp_write(struct pbe *pblist, unsigned int nr_pages) |
| { |
| struct swap_map_page *swap_map; |
| struct swap_map_handle handle; |
| swp_entry_t start; |
| int error; |
| |
| if ((error = swsusp_swap_check())) { |
| printk(KERN_ERR "swsusp: Cannot find swap device, try swapon -a.\n"); |
| return error; |
| } |
| if (!enough_swap(nr_pages)) { |
| printk(KERN_ERR "swsusp: Not enough free swap\n"); |
| return -ENOSPC; |
| } |
| |
| init_header(nr_pages); |
| swap_map = alloc_swap_map(swsusp_info.pages); |
| if (!swap_map) |
| return -ENOMEM; |
| init_swap_map_handle(&handle, swap_map); |
| |
| error = swap_map_write_page(&handle, (unsigned long)&swsusp_info); |
| if (!error) |
| error = save_image_metadata(pblist, &handle); |
| if (!error) |
| error = save_image_data(pblist, &handle, nr_pages); |
| if (error) |
| goto Free_image_entries; |
| |
| swap_map = reverse_swap_map(swap_map); |
| error = save_swap_map(swap_map, &start); |
| if (error) |
| goto Free_map_entries; |
| |
| dump_info(); |
| printk( "S" ); |
| error = mark_swapfiles(start); |
| printk( "|\n" ); |
| if (error) |
| goto Free_map_entries; |
| |
| Free_swap_map: |
| free_swap_map(swap_map); |
| return error; |
| |
| Free_map_entries: |
| free_swap_map_entries(swap_map); |
| Free_image_entries: |
| free_image_entries(swap_map); |
| goto Free_swap_map; |
| } |
| |
| /** |
| * swsusp_shrink_memory - Try to free as much memory as needed |
| * |
| * ... but do not OOM-kill anyone |
| * |
| * Notice: all userland should be stopped before it is called, or |
| * livelock is possible. |
| */ |
| |
| #define SHRINK_BITE 10000 |
| |
| int swsusp_shrink_memory(void) |
| { |
| long size, tmp; |
| struct zone *zone; |
| unsigned long pages = 0; |
| unsigned int i = 0; |
| char *p = "-\\|/"; |
| |
| printk("Shrinking memory... "); |
| do { |
| size = 2 * count_highmem_pages(); |
| size += size / 50 + count_data_pages(); |
| size += (size + PBES_PER_PAGE - 1) / PBES_PER_PAGE + |
| PAGES_FOR_IO; |
| tmp = size; |
| for_each_zone (zone) |
| if (!is_highmem(zone)) |
| tmp -= zone->free_pages; |
| if (tmp > 0) { |
| tmp = shrink_all_memory(SHRINK_BITE); |
| if (!tmp) |
| return -ENOMEM; |
| pages += tmp; |
| } else if (size > image_size / PAGE_SIZE) { |
| tmp = shrink_all_memory(SHRINK_BITE); |
| pages += tmp; |
| } |
| printk("\b%c", p[i++%4]); |
| } while (tmp > 0); |
| printk("\bdone (%lu pages freed)\n", pages); |
| |
| return 0; |
| } |
| |
| int swsusp_suspend(void) |
| { |
| int error; |
| |
| if ((error = arch_prepare_suspend())) |
| return error; |
| local_irq_disable(); |
| /* At this point, device_suspend() has been called, but *not* |
| * device_power_down(). We *must* device_power_down() now. |
| * Otherwise, drivers for some devices (e.g. interrupt controllers) |
| * become desynchronized with the actual state of the hardware |
| * at resume time, and evil weirdness ensues. |
| */ |
| if ((error = device_power_down(PMSG_FREEZE))) { |
| printk(KERN_ERR "Some devices failed to power down, aborting suspend\n"); |
| goto Enable_irqs; |
| } |
| |
| if ((error = save_highmem())) { |
| printk(KERN_ERR "swsusp: Not enough free pages for highmem\n"); |
| goto Restore_highmem; |
| } |
| |
| save_processor_state(); |
| if ((error = swsusp_arch_suspend())) |
| printk(KERN_ERR "Error %d suspending\n", error); |
| /* Restore control flow magically appears here */ |
| restore_processor_state(); |
| Restore_highmem: |
| restore_highmem(); |
| device_power_up(); |
| Enable_irqs: |
| local_irq_enable(); |
| return error; |
| } |
| |
| int swsusp_resume(void) |
| { |
| int error; |
| local_irq_disable(); |
| if (device_power_down(PMSG_FREEZE)) |
| printk(KERN_ERR "Some devices failed to power down, very bad\n"); |
| /* We'll ignore saved state, but this gets preempt count (etc) right */ |
| save_processor_state(); |
| error = swsusp_arch_resume(); |
| /* Code below is only ever reached in case of failure. Otherwise |
| * execution continues at place where swsusp_arch_suspend was called |
| */ |
| BUG_ON(!error); |
| /* The only reason why swsusp_arch_resume() can fail is memory being |
| * very tight, so we have to free it as soon as we can to avoid |
| * subsequent failures |
| */ |
| swsusp_free(); |
| restore_processor_state(); |
| restore_highmem(); |
| touch_softlockup_watchdog(); |
| device_power_up(); |
| local_irq_enable(); |
| return error; |
| } |
| |
| /** |
| * mark_unsafe_pages - mark the pages that cannot be used for storing |
| * the image during resume, because they conflict with the pages that |
| * had been used before suspend |
| */ |
| |
| static void mark_unsafe_pages(struct pbe *pblist) |
| { |
| struct zone *zone; |
| unsigned long zone_pfn; |
| struct pbe *p; |
| |
| if (!pblist) /* a sanity check */ |
| return; |
| |
| /* Clear page flags */ |
| for_each_zone (zone) { |
| for (zone_pfn = 0; zone_pfn < zone->spanned_pages; ++zone_pfn) |
| if (pfn_valid(zone_pfn + zone->zone_start_pfn)) |
| ClearPageNosaveFree(pfn_to_page(zone_pfn + |
| zone->zone_start_pfn)); |
| } |
| |
| /* Mark orig addresses */ |
| for_each_pbe (p, pblist) |
| SetPageNosaveFree(virt_to_page(p->orig_address)); |
| |
| } |
| |
| static void copy_page_backup_list(struct pbe *dst, struct pbe *src) |
| { |
| /* We assume both lists contain the same number of elements */ |
| while (src) { |
| dst->orig_address = src->orig_address; |
| dst = dst->next; |
| src = src->next; |
| } |
| } |
| |
| /* |
| * Using bio to read from swap. |
| * This code requires a bit more work than just using buffer heads |
| * but, it is the recommended way for 2.5/2.6. |
| * The following are to signal the beginning and end of I/O. Bios |
| * finish asynchronously, while we want them to happen synchronously. |
| * A simple atomic_t, and a wait loop take care of this problem. |
| */ |
| |
| static atomic_t io_done = ATOMIC_INIT(0); |
| |
| static int end_io(struct bio *bio, unsigned int num, int err) |
| { |
| if (!test_bit(BIO_UPTODATE, &bio->bi_flags)) |
| panic("I/O error reading memory image"); |
| atomic_set(&io_done, 0); |
| return 0; |
| } |
| |
| static struct block_device *resume_bdev; |
| |
| /** |
| * submit - submit BIO request. |
| * @rw: READ or WRITE. |
| * @off physical offset of page. |
| * @page: page we're reading or writing. |
| * |
| * Straight from the textbook - allocate and initialize the bio. |
| * If we're writing, make sure the page is marked as dirty. |
| * Then submit it and wait. |
| */ |
| |
| static int submit(int rw, pgoff_t page_off, void *page) |
| { |
| int error = 0; |
| struct bio *bio; |
| |
| bio = bio_alloc(GFP_ATOMIC, 1); |
| if (!bio) |
| return -ENOMEM; |
| bio->bi_sector = page_off * (PAGE_SIZE >> 9); |
| bio->bi_bdev = resume_bdev; |
| bio->bi_end_io = end_io; |
| |
| if (bio_add_page(bio, virt_to_page(page), PAGE_SIZE, 0) < PAGE_SIZE) { |
| printk("swsusp: ERROR: adding page to bio at %ld\n",page_off); |
| error = -EFAULT; |
| goto Done; |
| } |
| |
| |
| atomic_set(&io_done, 1); |
| submit_bio(rw | (1 << BIO_RW_SYNC), bio); |
| while (atomic_read(&io_done)) |
| yield(); |
| if (rw == READ) |
| bio_set_pages_dirty(bio); |
| Done: |
| bio_put(bio); |
| return error; |
| } |
| |
| static int bio_read_page(pgoff_t page_off, void *page) |
| { |
| return submit(READ, page_off, page); |
| } |
| |
| static int bio_write_page(pgoff_t page_off, void *page) |
| { |
| return submit(WRITE, page_off, page); |
| } |
| |
| /** |
| * The following functions allow us to read data using a swap map |
| * in a file-alike way |
| */ |
| |
| static inline void release_swap_map_reader(struct swap_map_handle *handle) |
| { |
| if (handle->cur) |
| free_page((unsigned long)handle->cur); |
| handle->cur = NULL; |
| } |
| |
| static inline int get_swap_map_reader(struct swap_map_handle *handle, |
| swp_entry_t start) |
| { |
| int error; |
| |
| if (!swp_offset(start)) |
| return -EINVAL; |
| handle->cur = (struct swap_map_page *)get_zeroed_page(GFP_ATOMIC); |
| if (!handle->cur) |
| return -ENOMEM; |
| error = bio_read_page(swp_offset(start), handle->cur); |
| if (error) { |
| release_swap_map_reader(handle); |
| return error; |
| } |
| handle->k = 0; |
| return 0; |
| } |
| |
| static inline int swap_map_read_page(struct swap_map_handle *handle, void *buf) |
| { |
| unsigned long offset; |
| int error; |
| |
| if (!handle->cur) |
| return -EINVAL; |
| offset = swp_offset(handle->cur->entries[handle->k]); |
| if (!offset) |
| return -EINVAL; |
| error = bio_read_page(offset, buf); |
| if (error) |
| return error; |
| if (++handle->k >= MAP_PAGE_SIZE) { |
| handle->k = 0; |
| offset = swp_offset(handle->cur->next_swap); |
| if (!offset) |
| release_swap_map_reader(handle); |
| else |
| error = bio_read_page(offset, handle->cur); |
| } |
| return error; |
| } |
| |
| static int check_header(void) |
| { |
| char *reason = NULL; |
| |
| dump_info(); |
| if (swsusp_info.version_code != LINUX_VERSION_CODE) |
| reason = "kernel version"; |
| if (swsusp_info.num_physpages != num_physpages) |
| reason = "memory size"; |
| if (strcmp(swsusp_info.uts.sysname,system_utsname.sysname)) |
| reason = "system type"; |
| if (strcmp(swsusp_info.uts.release,system_utsname.release)) |
| reason = "kernel release"; |
| if (strcmp(swsusp_info.uts.version,system_utsname.version)) |
| reason = "version"; |
| if (strcmp(swsusp_info.uts.machine,system_utsname.machine)) |
| reason = "machine"; |
| if (reason) { |
| printk(KERN_ERR "swsusp: Resume mismatch: %s\n", reason); |
| return -EPERM; |
| } |
| return 0; |
| } |
| |
| /** |
| * load_image_data - load the image data using the swap map handle |
| * @handle and store them using the page backup list @pblist |
| * (assume there are @nr_pages pages to load) |
| */ |
| |
| static int load_image_data(struct pbe *pblist, |
| struct swap_map_handle *handle, |
| unsigned int nr_pages) |
| { |
| int error; |
| unsigned int m; |
| struct pbe *p; |
| |
| if (!pblist) |
| return -EINVAL; |
| printk("Loading image data pages (%u pages) ... ", nr_pages); |
| m = nr_pages / 100; |
| if (!m) |
| m = 1; |
| nr_pages = 0; |
| p = pblist; |
| while (p) { |
| error = swap_map_read_page(handle, (void *)p->address); |
| if (error) |
| break; |
| p = p->next; |
| if (!(nr_pages % m)) |
| printk("\b\b\b\b%3d%%", nr_pages / m); |
| nr_pages++; |
| } |
| if (!error) |
| printk("\b\b\b\bdone\n"); |
| return error; |
| } |
| |
| /** |
| * unpack_orig_addresses - copy the elements of @buf[] (1 page) to |
| * the PBEs in the list starting at @pbe |
| */ |
| |
| static inline struct pbe *unpack_orig_addresses(unsigned long *buf, |
| struct pbe *pbe) |
| { |
| int j; |
| |
| for (j = 0; j < PAGE_SIZE / sizeof(long) && pbe; j++) { |
| pbe->orig_address = buf[j]; |
| pbe = pbe->next; |
| } |
| return pbe; |
| } |
| |
| /** |
| * load_image_metadata - load the image metadata using the swap map |
| * handle @handle and put them into the PBEs in the list @pblist |
| */ |
| |
| static int load_image_metadata(struct pbe *pblist, struct swap_map_handle *handle) |
| { |
| struct pbe *p; |
| unsigned long *buf; |
| unsigned int n = 0; |
| int error = 0; |
| |
| printk("Loading image metadata ... "); |
| buf = (unsigned long *)get_zeroed_page(GFP_ATOMIC); |
| if (!buf) |
| return -ENOMEM; |
| p = pblist; |
| while (p) { |
| error = swap_map_read_page(handle, buf); |
| if (error) |
| break; |
| p = unpack_orig_addresses(buf, p); |
| n++; |
| } |
| free_page((unsigned long)buf); |
| if (!error) |
| printk("done (%u pages loaded)\n", n); |
| return error; |
| } |
| |
| int swsusp_read(struct pbe **pblist_ptr) |
| { |
| int error; |
| struct pbe *p, *pblist; |
| struct swap_map_handle handle; |
| unsigned int nr_pages; |
| |
| if (IS_ERR(resume_bdev)) { |
| pr_debug("swsusp: block device not initialised\n"); |
| return PTR_ERR(resume_bdev); |
| } |
| |
| error = get_swap_map_reader(&handle, swsusp_header.image); |
| if (!error) |
| error = swap_map_read_page(&handle, &swsusp_info); |
| if (!error) |
| error = check_header(); |
| if (error) |
| return error; |
| nr_pages = swsusp_info.image_pages; |
| p = alloc_pagedir(nr_pages, GFP_ATOMIC, 0); |
| if (!p) |
| return -ENOMEM; |
| error = load_image_metadata(p, &handle); |
| if (!error) { |
| mark_unsafe_pages(p); |
| pblist = alloc_pagedir(nr_pages, GFP_ATOMIC, 1); |
| if (pblist) |
| copy_page_backup_list(pblist, p); |
| free_pagedir(p); |
| if (!pblist) |
| error = -ENOMEM; |
| |
| /* Allocate memory for the image and read the data from swap */ |
| if (!error) |
| error = alloc_data_pages(pblist, GFP_ATOMIC, 1); |
| if (!error) { |
| release_eaten_pages(); |
| error = load_image_data(pblist, &handle, nr_pages); |
| } |
| if (!error) |
| *pblist_ptr = pblist; |
| } |
| release_swap_map_reader(&handle); |
| |
| blkdev_put(resume_bdev); |
| |
| if (!error) |
| pr_debug("swsusp: Reading resume file was successful\n"); |
| else |
| pr_debug("swsusp: Error %d resuming\n", error); |
| return error; |
| } |
| |
| /** |
| * swsusp_check - Check for swsusp signature in the resume device |
| */ |
| |
| int swsusp_check(void) |
| { |
| int error; |
| |
| resume_bdev = open_by_devnum(swsusp_resume_device, FMODE_READ); |
| if (!IS_ERR(resume_bdev)) { |
| set_blocksize(resume_bdev, PAGE_SIZE); |
| memset(&swsusp_header, 0, sizeof(swsusp_header)); |
| if ((error = bio_read_page(0, &swsusp_header))) |
| return error; |
| if (!memcmp(SWSUSP_SIG, swsusp_header.sig, 10)) { |
| memcpy(swsusp_header.sig, swsusp_header.orig_sig, 10); |
| /* Reset swap signature now */ |
| error = bio_write_page(0, &swsusp_header); |
| } else { |
| return -EINVAL; |
| } |
| if (error) |
| blkdev_put(resume_bdev); |
| else |
| pr_debug("swsusp: Signature found, resuming\n"); |
| } else { |
| error = PTR_ERR(resume_bdev); |
| } |
| |
| if (error) |
| pr_debug("swsusp: Error %d check for resume file\n", error); |
| |
| return error; |
| } |
| |
| /** |
| * swsusp_close - close swap device. |
| */ |
| |
| void swsusp_close(void) |
| { |
| if (IS_ERR(resume_bdev)) { |
| pr_debug("swsusp: block device not initialised\n"); |
| return; |
| } |
| |
| blkdev_put(resume_bdev); |
| } |