/* * zswap.c - zswap driver file * * zswap is a backend for frontswap that takes pages that are in the process * of being swapped out and attempts to compress and store them in a * RAM-based memory pool. This can result in a significant I/O reduction on * the swap device and, in the case where decompressing from RAM is faster * than reading from the swap device, can also improve workload performance. * * Copyright (C) 2012 Seth Jennings * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License * as published by the Free Software Foundation; either version 2 * of the License, or (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING #include #endif /********************************* * statistics **********************************/ /* Total bytes used by the compressed storage */ static u64 zswap_pool_total_size; /* Number of memory pages used by the compressed pool */ u64 zswap_pool_pages; /* The number of compressed pages currently stored in zswap */ atomic_t zswap_stored_pages = ATOMIC_INIT(0); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING /* The number of swapped out pages which are identified as duplicate to the existing zswap pages. Compression and storing of these pages is avoided */ static atomic_t zswap_duplicate_pages = ATOMIC_INIT(0); #endif /* * The statistics below are not protected from concurrent access for * performance reasons so they may not be a 100% accurate. However, * they do provide useful information on roughly how many times a * certain event is occurring. */ /* Pool limit was hit (see zswap_max_pool_percent) */ static u64 zswap_pool_limit_hit; /* Pages written back when pool limit was reached */ static u64 zswap_written_back_pages; /* Store failed due to a reclaim failure after pool limit was reached */ static u64 zswap_reject_reclaim_fail; /* Compressed page was too big for the allocator to (optimally) store */ static u64 zswap_reject_compress_poor; /* Store failed because underlying allocator could not get memory */ static u64 zswap_reject_alloc_fail; /* Store failed because the entry metadata could not be allocated (rare) */ static u64 zswap_reject_kmemcache_fail; /* Duplicate store was encountered (rare) */ static u64 zswap_duplicate_entry; /* The number of zero pages currently stored in zswap */ static atomic_t zswap_zero_pages = ATOMIC_INIT(0); /********************************* * tunables **********************************/ /* Enable/disable zswap (disabled by default) */ static bool zswap_enabled = 1; static int zswap_enabled_param_set(const char *, const struct kernel_param *); static struct kernel_param_ops zswap_enabled_param_ops = { .set = zswap_enabled_param_set, .get = param_get_bool, }; module_param_cb(enabled, &zswap_enabled_param_ops, &zswap_enabled, 0644); /* Crypto compressor to use */ #define ZSWAP_COMPRESSOR_DEFAULT "lzo" #if IS_ENABLED(CONFIG_CRYPTO_ZSTD) #define ZSWAP_COMPRESSOR "zstd" #else #define ZSWAP_COMPRESSOR "lz4" #endif static char *zswap_compressor = ZSWAP_COMPRESSOR; static int zswap_compressor_param_set(const char *, const struct kernel_param *); static struct kernel_param_ops zswap_compressor_param_ops = { .set = zswap_compressor_param_set, .get = param_get_charp, .free = param_free_charp, }; module_param_cb(compressor, &zswap_compressor_param_ops, &zswap_compressor, 0644); /* Compressed storage zpool to use */ #define ZSWAP_ZPOOL_DEFAULT "zsmalloc" static char *zswap_zpool_type = ZSWAP_ZPOOL_DEFAULT; static int zswap_zpool_param_set(const char *, const struct kernel_param *); static struct kernel_param_ops zswap_zpool_param_ops = { .set = zswap_zpool_param_set, .get = param_get_charp, .free = param_free_charp, }; module_param_cb(zpool, &zswap_zpool_param_ops, &zswap_zpool_type, 0644); /* The maximum percentage of memory that the compressed pool can occupy */ static unsigned int zswap_max_pool_percent = 50; module_param_named(max_pool_percent, zswap_max_pool_percent, uint, 0644); /********************************* * data structures **********************************/ #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING /* * struct zswap_handle * This structure contains the metadata for tracking single zpool * allocation. * * rbnode - links the handle into red-black tree * checksum - 32-bit checksum value of the page swapped to zswap * ref_count - number of pages sharing this handle * length - the length in bytes of the compressed page data. * Needed during decompression. * handle - zpool allocation handle that stores the compressed page data */ struct zswap_handle { struct rb_node rbnode; u32 checksum; u32 ref_count; unsigned int length; unsigned long handle; }; #endif struct zswap_pool { struct zpool *zpool; struct crypto_comp * __percpu *tfm; struct kref kref; struct list_head list; struct work_struct work; struct notifier_block notifier; char tfm_name[CRYPTO_MAX_ALG_NAME]; }; /* * struct zswap_entry * * This structure contains the metadata for tracking a single compressed * page within zswap. * * #ifndef CONFIG_ZSWAP_SAME_PAGE_SHARING * rbnode - links the entry into red-black tree for the appropriate swap type * offset - the swap offset for the entry. Index into the red-black tree. * refcount - the number of outstanding reference to the entry. This is needed * to protect against premature freeing of the entry by code * concurrent calls to load, invalidate, and writeback. The lock * for the zswap_tree structure that contains the entry must * be held while changing the refcount. Since the lock must * be held, there is no reason to also make refcount atomic. * length - the length in bytes of the compressed page data. Needed during * decompression * pool - the zswap_pool the entry's data is in * handle - zpool allocation handle that stores the compressed page data * zero_flag - the flag indicating the page for the zswap_entry is a zero page. * zswap does not store the page during compression. * It memsets the page with 0 during decompression. * #else * zhandle - pointer to struct zswap_handle where length and handle are moved into. * #endif */ #ifndef CONFIG_ZSWAP_SAME_PAGE_SHARING struct zswap_entry { struct rb_node rbnode; pgoff_t offset; int refcount; unsigned int length; struct zswap_pool *pool; unsigned long handle; unsigned char zero_flag; }; #else struct zswap_entry { struct rb_node rbnode; pgoff_t offset; int refcount; struct zswap_pool *pool; struct zswap_handle *zhandle; unsigned char zero_flag; }; #endif #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK struct zswap_header { swp_entry_t swpentry; }; #endif /* * The tree lock in the zswap_tree struct protects a few things: * - the rbtree * - the refcount field of each entry in the tree */ struct zswap_tree { struct rb_root rbroot; #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING struct rb_root zhandleroot; void *buffer; #endif spinlock_t lock; }; static struct zswap_tree *zswap_trees[MAX_SWAPFILES]; /* RCU-protected iteration */ static LIST_HEAD(zswap_pools); /* protects zswap_pools list modification */ static DEFINE_SPINLOCK(zswap_pools_lock); /* pool counter to provide unique names to zpool */ static atomic_t zswap_pools_count = ATOMIC_INIT(0); /* used by param callback function */ static bool zswap_init_started; /* fatal error during init */ static bool zswap_init_failed; /********************************* * helpers and fwd declarations **********************************/ #define zswap_pool_debug(msg, p) \ pr_debug("%s pool %s/%s\n", msg, (p)->tfm_name, \ zpool_get_type((p)->zpool)) static int zswap_writeback_entry(struct zpool *pool, unsigned long handle); static int zswap_pool_get(struct zswap_pool *pool); static void zswap_pool_put(struct zswap_pool *pool); static const struct zpool_ops zswap_zpool_ops = { .evict = zswap_writeback_entry }; static bool zswap_is_full(void) { return totalram_pages * zswap_max_pool_percent / 100 < DIV_ROUND_UP(zswap_pool_total_size, PAGE_SIZE); } static void zswap_update_total_size(void) { struct zswap_pool *pool; u64 total = 0; rcu_read_lock(); list_for_each_entry_rcu(pool, &zswap_pools, list) total += zpool_get_total_size(pool->zpool); rcu_read_unlock(); zswap_pool_total_size = total; zswap_pool_pages = zswap_pool_total_size >> PAGE_SHIFT; } /********************************* * zswap entry functions **********************************/ static struct kmem_cache *zswap_entry_cache; static int __init zswap_entry_cache_create(void) { zswap_entry_cache = KMEM_CACHE(zswap_entry, 0); return zswap_entry_cache == NULL; } static void __init zswap_entry_cache_destroy(void) { kmem_cache_destroy(zswap_entry_cache); } static struct zswap_entry *zswap_entry_cache_alloc(gfp_t gfp) { struct zswap_entry *entry; entry = kmem_cache_alloc(zswap_entry_cache, gfp); if (!entry) return NULL; entry->refcount = 1; entry->zero_flag = 0; #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING entry->zhandle = NULL; #endif RB_CLEAR_NODE(&entry->rbnode); return entry; } static void zswap_entry_cache_free(struct zswap_entry *entry) { kmem_cache_free(zswap_entry_cache, entry); } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING /********************************* * zswap handle functions **********************************/ static struct kmem_cache *zswap_handle_cache; static int zswap_handle_cache_create(void) { zswap_handle_cache = KMEM_CACHE(zswap_handle, 0); return zswap_handle_cache == NULL; } static void __init zswap_handle_cache_destroy(void) { kmem_cache_destroy(zswap_handle_cache); } static struct zswap_handle *zswap_handle_cache_alloc(gfp_t gfp) { struct zswap_handle *zhandle; zhandle = kmem_cache_alloc(zswap_handle_cache, gfp); if (!zhandle) return NULL; zhandle->ref_count = 1; RB_CLEAR_NODE(&zhandle->rbnode); return zhandle; } static void zswap_handle_cache_free(struct zswap_handle *zhandle) { kmem_cache_free(zswap_handle_cache, zhandle); } #endif /********************************* * rbtree functions **********************************/ static struct zswap_entry *zswap_rb_search(struct rb_root *root, pgoff_t offset) { struct rb_node *node = root->rb_node; struct zswap_entry *entry; while (node) { entry = rb_entry(node, struct zswap_entry, rbnode); if (entry->offset > offset) node = node->rb_left; else if (entry->offset < offset) node = node->rb_right; else return entry; } return NULL; } /* * In the case that a entry with the same offset is found, a pointer to * the existing entry is stored in dupentry and the function returns -EEXIST */ static int zswap_rb_insert(struct rb_root *root, struct zswap_entry *entry, struct zswap_entry **dupentry) { struct rb_node **link = &root->rb_node, *parent = NULL; struct zswap_entry *myentry; while (*link) { parent = *link; myentry = rb_entry(parent, struct zswap_entry, rbnode); if (myentry->offset > entry->offset) link = &(*link)->rb_left; else if (myentry->offset < entry->offset) link = &(*link)->rb_right; else { *dupentry = myentry; return -EEXIST; } } rb_link_node(&entry->rbnode, parent, link); rb_insert_color(&entry->rbnode, root); return 0; } static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry) { if (!RB_EMPTY_NODE(&entry->rbnode)) { rb_erase(&entry->rbnode, root); RB_CLEAR_NODE(&entry->rbnode); } } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING static struct zswap_handle *zswap_handle_rb_search(struct rb_root *root, u32 checksum) { struct rb_node *node = root->rb_node; struct zswap_handle *zhandle; while (node) { zhandle = rb_entry(node, struct zswap_handle, rbnode); if (zhandle->checksum > checksum) node = node->rb_left; else if (zhandle->checksum < checksum) node = node->rb_right; else return zhandle; } return NULL; } /* * In the case that zhandle with the same checksum is found, a pointer to * the existing zhandle is stored in duphandle and the function returns -EEXIST */ static int zswap_handle_rb_insert(struct rb_root *root, struct zswap_handle *zhandle, struct zswap_handle **duphandle) { struct rb_node **link = &root->rb_node, *parent = NULL; struct zswap_handle *myhandle; while (*link) { parent = *link; myhandle = rb_entry(parent, struct zswap_handle, rbnode); if (myhandle->checksum > zhandle->checksum) link = &parent->rb_left; else if (myhandle->checksum < zhandle->checksum) link = &parent->rb_right; else { *duphandle = myhandle; return -EEXIST; } } rb_link_node(&zhandle->rbnode, parent, link); rb_insert_color(&zhandle->rbnode, root); return 0; } static void zswap_handle_erase(struct rb_root *root, struct zswap_handle *zhandle) { if (!RB_EMPTY_NODE(&zhandle->rbnode)) { rb_erase(&zhandle->rbnode, root); RB_CLEAR_NODE(&zhandle->rbnode); } } static void zswap_free_handle(struct zswap_pool *pool, struct zswap_handle *zhandle) { zpool_free(pool->zpool, zhandle->handle); zswap_handle_cache_free(zhandle); } /* This function searches for the same page in the zhandle RB-Tree based on the * checksum value of the new page. If the same page is found the zhandle of that * page is returned. */ static struct zswap_handle *zswap_same_page_search(struct zswap_pool *pool, struct zswap_tree *tree, u8 *uncmem, u32 checksum) { int ret = 0; unsigned int dlen = PAGE_SIZE; u8 *src = NULL, *dst = NULL; struct zswap_handle *myhandle = NULL; struct crypto_comp *tfm; myhandle = zswap_handle_rb_search(&tree->zhandleroot, checksum); if (myhandle) { /* Compare memory contents */ dst = (u8 *)tree->buffer; src = (u8 *)zpool_map_handle(pool->zpool, myhandle->handle, ZPOOL_MM_RO_NOWAIT); if (!src) return NULL; if (myhandle->length == PAGE_SIZE) copy_page(dst, src); else { tfm = *get_cpu_ptr(pool->tfm); ret = crypto_comp_decompress(tfm, src, myhandle->length, dst, &dlen); put_cpu_ptr(pool->tfm); } zpool_unmap_handle(pool->zpool, myhandle->handle); BUG_ON(ret); ret = memcmp(dst, uncmem, PAGE_SIZE); if (ret) myhandle = NULL; } return myhandle; } #endif /* * Carries out the common pattern of freeing and entry's zpool allocation, * freeing the entry itself, and decrementing the number of stored pages. */ static void zswap_free_entry(struct zswap_entry *entry) { if (entry->zero_flag == 1) { atomic_dec(&zswap_zero_pages); goto zeropage_out; } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING entry->zhandle->ref_count--; if (!entry->zhandle->ref_count) zswap_free_handle(entry->pool, entry->zhandle); else atomic_dec(&zswap_duplicate_pages); #else zpool_free(entry->pool->zpool, entry->handle); #endif zeropage_out: zswap_pool_put(entry->pool); zswap_entry_cache_free(entry); atomic_dec(&zswap_stored_pages); zswap_update_total_size(); } /* caller must hold the tree lock */ static void zswap_entry_get(struct zswap_entry *entry) { entry->refcount++; } /* caller must hold the tree lock * remove from the tree and free it, if nobody reference the entry */ static void zswap_entry_put(struct zswap_tree *tree, struct zswap_entry *entry) { int refcount = --entry->refcount; BUG_ON(refcount < 0); if (refcount == 0) { #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING if (entry->zhandle && entry->zhandle->ref_count == 1) zswap_handle_erase(&tree->zhandleroot, entry->zhandle); #endif zswap_rb_erase(&tree->rbroot, entry); zswap_free_entry(entry); } } /* caller must hold the tree lock */ static struct zswap_entry *zswap_entry_find_get(struct rb_root *root, pgoff_t offset) { struct zswap_entry *entry; entry = zswap_rb_search(root, offset); if (entry) zswap_entry_get(entry); return entry; } /********************************* * per-cpu code **********************************/ static DEFINE_PER_CPU(u8 *, zswap_dstmem); static int __zswap_cpu_dstmem_notifier(unsigned long action, unsigned long cpu) { u8 *dst; switch (action) { case CPU_UP_PREPARE: dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); if (!dst) { pr_err("can't allocate compressor buffer\n"); return NOTIFY_BAD; } per_cpu(zswap_dstmem, cpu) = dst; break; case CPU_DEAD: case CPU_UP_CANCELED: dst = per_cpu(zswap_dstmem, cpu); kfree(dst); per_cpu(zswap_dstmem, cpu) = NULL; break; default: break; } return NOTIFY_OK; } static int zswap_cpu_dstmem_notifier(struct notifier_block *nb, unsigned long action, void *pcpu) { return __zswap_cpu_dstmem_notifier(action, (unsigned long)pcpu); } static struct notifier_block zswap_dstmem_notifier = { .notifier_call = zswap_cpu_dstmem_notifier, }; static int __init zswap_cpu_dstmem_init(void) { unsigned long cpu; cpu_notifier_register_begin(); for_each_online_cpu(cpu) if (__zswap_cpu_dstmem_notifier(CPU_UP_PREPARE, cpu) == NOTIFY_BAD) goto cleanup; __register_cpu_notifier(&zswap_dstmem_notifier); cpu_notifier_register_done(); return 0; cleanup: for_each_online_cpu(cpu) __zswap_cpu_dstmem_notifier(CPU_UP_CANCELED, cpu); cpu_notifier_register_done(); return -ENOMEM; } static void zswap_cpu_dstmem_destroy(void) { unsigned long cpu; cpu_notifier_register_begin(); for_each_online_cpu(cpu) __zswap_cpu_dstmem_notifier(CPU_UP_CANCELED, cpu); __unregister_cpu_notifier(&zswap_dstmem_notifier); cpu_notifier_register_done(); } static int __zswap_cpu_comp_notifier(struct zswap_pool *pool, unsigned long action, unsigned long cpu) { struct crypto_comp *tfm; switch (action) { case CPU_UP_PREPARE: if (WARN_ON(*per_cpu_ptr(pool->tfm, cpu))) break; tfm = crypto_alloc_comp(pool->tfm_name, 0, 0); if (IS_ERR_OR_NULL(tfm)) { pr_err("could not alloc crypto comp %s : %ld\n", pool->tfm_name, PTR_ERR(tfm)); return NOTIFY_BAD; } *per_cpu_ptr(pool->tfm, cpu) = tfm; break; case CPU_DEAD: case CPU_UP_CANCELED: tfm = *per_cpu_ptr(pool->tfm, cpu); if (!IS_ERR_OR_NULL(tfm)) crypto_free_comp(tfm); *per_cpu_ptr(pool->tfm, cpu) = NULL; break; default: break; } return NOTIFY_OK; } static int zswap_cpu_comp_notifier(struct notifier_block *nb, unsigned long action, void *pcpu) { unsigned long cpu = (unsigned long)pcpu; struct zswap_pool *pool = container_of(nb, typeof(*pool), notifier); return __zswap_cpu_comp_notifier(pool, action, cpu); } static int zswap_cpu_comp_init(struct zswap_pool *pool) { unsigned long cpu; memset(&pool->notifier, 0, sizeof(pool->notifier)); pool->notifier.notifier_call = zswap_cpu_comp_notifier; cpu_notifier_register_begin(); for_each_online_cpu(cpu) if (__zswap_cpu_comp_notifier(pool, CPU_UP_PREPARE, cpu) == NOTIFY_BAD) goto cleanup; __register_cpu_notifier(&pool->notifier); cpu_notifier_register_done(); return 0; cleanup: for_each_online_cpu(cpu) __zswap_cpu_comp_notifier(pool, CPU_UP_CANCELED, cpu); cpu_notifier_register_done(); return -ENOMEM; } static void zswap_cpu_comp_destroy(struct zswap_pool *pool) { unsigned long cpu; cpu_notifier_register_begin(); for_each_online_cpu(cpu) __zswap_cpu_comp_notifier(pool, CPU_UP_CANCELED, cpu); __unregister_cpu_notifier(&pool->notifier); cpu_notifier_register_done(); } /********************************* * pool functions **********************************/ static struct zswap_pool *__zswap_pool_current(void) { struct zswap_pool *pool; pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list); WARN_ON(!pool); return pool; } static struct zswap_pool *zswap_pool_current(void) { assert_spin_locked(&zswap_pools_lock); return __zswap_pool_current(); } static struct zswap_pool *zswap_pool_current_get(void) { struct zswap_pool *pool; rcu_read_lock(); pool = __zswap_pool_current(); if (!pool || !zswap_pool_get(pool)) pool = NULL; rcu_read_unlock(); return pool; } #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK static struct zswap_pool *zswap_pool_last_get(void) { struct zswap_pool *pool, *last = NULL; rcu_read_lock(); list_for_each_entry_rcu(pool, &zswap_pools, list) last = pool; if (!WARN_ON(!last) && !zswap_pool_get(last)) last = NULL; rcu_read_unlock(); return last; } #endif /* type and compressor must be null-terminated */ static struct zswap_pool *zswap_pool_find_get(char *type, char *compressor) { struct zswap_pool *pool; assert_spin_locked(&zswap_pools_lock); list_for_each_entry_rcu(pool, &zswap_pools, list) { if (strcmp(pool->tfm_name, compressor)) continue; if (strcmp(zpool_get_type(pool->zpool), type)) continue; /* if we can't get it, it's about to be destroyed */ if (!zswap_pool_get(pool)) continue; return pool; } return NULL; } static struct zswap_pool *zswap_pool_create(char *type, char *compressor) { struct zswap_pool *pool; char name[38]; /* 'zswap' + 32 char (max) num + \0 */ #ifdef CONFIG_ZSWAP_MIGRATION_SUPPORT gfp_t gfp = __GFP_NORETRY | __GFP_NOWARN | __GFP_KSWAPD_RECLAIM | __GFP_HIGHMEM | __GFP_MOVABLE; #else gfp_t gfp = __GFP_NORETRY | __GFP_NOWARN | __GFP_KSWAPD_RECLAIM | __GFP_HIGHMEM; #endif pool = kzalloc(sizeof(*pool), GFP_KERNEL); if (!pool) { pr_err("pool alloc failed\n"); return NULL; } /* unique name for each pool specifically required by zsmalloc */ snprintf(name, 38, "zswap%x", atomic_inc_return(&zswap_pools_count)); pool->zpool = zpool_create_pool(type, name, gfp, &zswap_zpool_ops); if (!pool->zpool) { pr_err("%s zpool not available\n", type); goto error; } pr_debug("using %s zpool\n", zpool_get_type(pool->zpool)); strlcpy(pool->tfm_name, compressor, sizeof(pool->tfm_name)); pool->tfm = alloc_percpu(struct crypto_comp *); if (!pool->tfm) { pr_err("percpu alloc failed\n"); goto error; } if (zswap_cpu_comp_init(pool)) goto error; pr_debug("using %s compressor\n", pool->tfm_name); /* being the current pool takes 1 ref; this func expects the * caller to always add the new pool as the current pool */ kref_init(&pool->kref); INIT_LIST_HEAD(&pool->list); zswap_pool_debug("created", pool); return pool; error: free_percpu(pool->tfm); if (pool->zpool) zpool_destroy_pool(pool->zpool); kfree(pool); return NULL; } static __init struct zswap_pool *__zswap_pool_create_fallback(void) { if (!crypto_has_comp(zswap_compressor, 0, 0)) { if (!strcmp(zswap_compressor, ZSWAP_COMPRESSOR_DEFAULT)) { pr_err("default compressor %s not available\n", zswap_compressor); return NULL; } pr_err("compressor %s not available, using default %s\n", zswap_compressor, ZSWAP_COMPRESSOR_DEFAULT); param_free_charp(&zswap_compressor); zswap_compressor = ZSWAP_COMPRESSOR_DEFAULT; } if (!zpool_has_pool(zswap_zpool_type)) { if (!strcmp(zswap_zpool_type, ZSWAP_ZPOOL_DEFAULT)) { pr_err("default zpool %s not available\n", zswap_zpool_type); return NULL; } pr_err("zpool %s not available, using default %s\n", zswap_zpool_type, ZSWAP_ZPOOL_DEFAULT); param_free_charp(&zswap_zpool_type); zswap_zpool_type = ZSWAP_ZPOOL_DEFAULT; } return zswap_pool_create(zswap_zpool_type, zswap_compressor); } static void zswap_pool_destroy(struct zswap_pool *pool) { zswap_pool_debug("destroying", pool); zswap_cpu_comp_destroy(pool); free_percpu(pool->tfm); zpool_destroy_pool(pool->zpool); kfree(pool); } static int __must_check zswap_pool_get(struct zswap_pool *pool) { return kref_get_unless_zero(&pool->kref); } static void __zswap_pool_release(struct work_struct *work) { struct zswap_pool *pool = container_of(work, typeof(*pool), work); synchronize_rcu(); /* nobody should have been able to get a kref... */ WARN_ON(kref_get_unless_zero(&pool->kref)); /* pool is now off zswap_pools list and has no references. */ zswap_pool_destroy(pool); } static void __zswap_pool_empty(struct kref *kref) { struct zswap_pool *pool; pool = container_of(kref, typeof(*pool), kref); spin_lock(&zswap_pools_lock); WARN_ON(pool == zswap_pool_current()); list_del_rcu(&pool->list); INIT_WORK(&pool->work, __zswap_pool_release); schedule_work(&pool->work); spin_unlock(&zswap_pools_lock); } static void zswap_pool_put(struct zswap_pool *pool) { kref_put(&pool->kref, __zswap_pool_empty); } /********************************* * param callbacks **********************************/ /* val must be a null-terminated string */ static int __zswap_param_set(const char *val, const struct kernel_param *kp, char *type, char *compressor) { struct zswap_pool *pool, *put_pool = NULL; char *s = strstrip((char *)val); int ret; if (zswap_init_failed) { pr_err("can't set param, initialization failed\n"); return -ENODEV; } /* no change required */ if (!strcmp(s, *(char **)kp->arg)) return 0; /* if this is load-time (pre-init) param setting, * don't create a pool; that's done during init. */ if (!zswap_init_started) return param_set_charp(s, kp); if (!type) { if (!zpool_has_pool(s)) { pr_err("zpool %s not available\n", s); return -ENOENT; } type = s; } else if (!compressor) { if (!crypto_has_comp(s, 0, 0)) { pr_err("compressor %s not available\n", s); return -ENOENT; } compressor = s; } else { WARN_ON(1); return -EINVAL; } spin_lock(&zswap_pools_lock); pool = zswap_pool_find_get(type, compressor); if (pool) { zswap_pool_debug("using existing", pool); WARN_ON(pool == zswap_pool_current()); list_del_rcu(&pool->list); } spin_unlock(&zswap_pools_lock); if (!pool) pool = zswap_pool_create(type, compressor); if (pool) ret = param_set_charp(s, kp); else ret = -EINVAL; spin_lock(&zswap_pools_lock); if (!ret) { put_pool = zswap_pool_current(); list_add_rcu(&pool->list, &zswap_pools); } else if (pool) { /* add the possibly pre-existing pool to the end of the pools * list; if it's new (and empty) then it'll be removed and * destroyed by the put after we drop the lock */ list_add_tail_rcu(&pool->list, &zswap_pools); put_pool = pool; } spin_unlock(&zswap_pools_lock); /* drop the ref from either the old current pool, * or the new pool we failed to add */ if (put_pool) zswap_pool_put(put_pool); return ret; } static int zswap_compressor_param_set(const char *val, const struct kernel_param *kp) { return __zswap_param_set(val, kp, zswap_zpool_type, NULL); } static int zswap_zpool_param_set(const char *val, const struct kernel_param *kp) { return __zswap_param_set(val, kp, NULL, zswap_compressor); } static int zswap_enabled_param_set(const char *val, const struct kernel_param *kp) { if (zswap_init_failed) { pr_err("can't enable, initialization failed\n"); return -ENODEV; } return param_set_bool(val, kp); } /********************************* * writeback code **********************************/ /* return enum for zswap_get_swap_cache_page */ enum zswap_get_swap_ret { ZSWAP_SWAPCACHE_NEW, ZSWAP_SWAPCACHE_EXIST, ZSWAP_SWAPCACHE_FAIL, }; #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK /* * zswap_get_swap_cache_page * * This is an adaption of read_swap_cache_async() * * This function tries to find a page with the given swap entry * in the swapper_space address space (the swap cache). If the page * is found, it is returned in retpage. Otherwise, a page is allocated, * added to the swap cache, and returned in retpage. * * If success, the swap cache page is returned in retpage * Returns ZSWAP_SWAPCACHE_EXIST if page was already in the swap cache * Returns ZSWAP_SWAPCACHE_NEW if the new page needs to be populated, * the new page is added to swapcache and locked * Returns ZSWAP_SWAPCACHE_FAIL on error */ static int zswap_get_swap_cache_page(swp_entry_t entry, struct page **retpage) { bool page_was_allocated; *retpage = __read_swap_cache_async(entry, GFP_KERNEL, NULL, 0, &page_was_allocated); if (page_was_allocated) return ZSWAP_SWAPCACHE_NEW; if (!*retpage) return ZSWAP_SWAPCACHE_FAIL; return ZSWAP_SWAPCACHE_EXIST; } /* * Attempts to free an entry by adding a page to the swap cache, * decompressing the entry data into the page, and issuing a * bio write to write the page back to the swap device. * * This can be thought of as a "resumed writeback" of the page * to the swap device. We are basically resuming the same swap * writeback path that was intercepted with the frontswap_store() * in the first place. After the page has been decompressed into * the swap cache, the compressed version stored by zswap can be * freed. */ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle) { struct zswap_header *zhdr; swp_entry_t swpentry; struct zswap_tree *tree; pgoff_t offset; struct zswap_entry *entry; struct page *page; struct crypto_comp *tfm; u8 *src, *dst; unsigned int dlen; int ret; struct writeback_control wbc = { .sync_mode = WB_SYNC_NONE, }; /* extract swpentry from data */ zhdr = zpool_map_handle(pool, handle, ZPOOL_MM_RO); swpentry = zhdr->swpentry; /* here */ zpool_unmap_handle(pool, handle); tree = zswap_trees[swp_type(swpentry)]; offset = swp_offset(swpentry); /* find and ref zswap entry */ spin_lock(&tree->lock); entry = zswap_entry_find_get(&tree->rbroot, offset); if (!entry) { /* entry was invalidated */ spin_unlock(&tree->lock); return 0; } spin_unlock(&tree->lock); BUG_ON(offset != entry->offset); /* try to allocate swap cache page */ switch (zswap_get_swap_cache_page(swpentry, &page)) { case ZSWAP_SWAPCACHE_FAIL: /* no memory or invalidate happened */ ret = -ENOMEM; goto fail; case ZSWAP_SWAPCACHE_EXIST: /* page is already in the swap cache, ignore for now */ page_cache_release(page); ret = -EEXIST; goto fail; case ZSWAP_SWAPCACHE_NEW: /* page is locked */ /* decompress */ dlen = PAGE_SIZE; src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->handle, ZPOOL_MM_RO) + sizeof(struct zswap_header); dst = kmap_atomic(page); tfm = *get_cpu_ptr(entry->pool->tfm); ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen); put_cpu_ptr(entry->pool->tfm); kunmap_atomic(dst); zpool_unmap_handle(entry->pool->zpool, entry->handle); BUG_ON(ret); BUG_ON(dlen != PAGE_SIZE); /* page is up to date */ SetPageUptodate(page); } /* move it to the tail of the inactive list after end_writeback */ SetPageReclaim(page); /* start writeback */ __swap_writepage(page, &wbc, end_swap_bio_write); page_cache_release(page); zswap_written_back_pages++; spin_lock(&tree->lock); /* drop local reference */ zswap_entry_put(tree, entry); /* * There are two possible situations for entry here: * (1) refcount is 1(normal case), entry is valid and on the tree * (2) refcount is 0, entry is freed and not on the tree * because invalidate happened during writeback * search the tree and free the entry if find entry */ if (entry == zswap_rb_search(&tree->rbroot, offset)) zswap_entry_put(tree, entry); spin_unlock(&tree->lock); goto end; /* * if we get here due to ZSWAP_SWAPCACHE_EXIST * a load may happening concurrently * it is safe and okay to not free the entry * if we free the entry in the following put * it it either okay to return !0 */ fail: spin_lock(&tree->lock); zswap_entry_put(tree, entry); spin_unlock(&tree->lock); end: return ret; } static int zswap_shrink(void) { struct zswap_pool *pool; int ret; pool = zswap_pool_last_get(); if (!pool) return -ENOENT; ret = zpool_shrink(pool->zpool, 1, NULL); zswap_pool_put(pool); return ret; } #else static int zswap_writeback_entry(struct zpool *pool, unsigned long handle) { return -EINVAL; } static int zswap_shrink(void) { return -EINVAL; } #endif /* CONFIG_ZSWAP_ENABLE_WRITEBACK */ static int page_zero_filled(void *ptr) { unsigned int pos; unsigned long *page; page = (unsigned long *)ptr; for (pos = 0; pos != PAGE_SIZE / sizeof(*page); pos++) { if (page[pos]) return 0; } return 1; } /********************************* * frontswap hooks **********************************/ /* attempts to compress and store an single page */ static int zswap_frontswap_store(unsigned type, pgoff_t offset, struct page *page) { struct zswap_tree *tree = zswap_trees[type]; struct zswap_entry *entry, *dupentry; struct crypto_comp *tfm; int ret; unsigned int dlen = PAGE_SIZE, len; unsigned long handle; char *buf; u8 *src, *dst; #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK struct zswap_header *zhdr; #endif #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING struct zswap_handle *zhandle = NULL, *duphandle = NULL; u32 checksum = 0; #endif if (!zswap_enabled || !tree) { ret = -ENODEV; goto reject; } /* if this page got EIO on pageout before, give up immediately */ if (PageError(page)) { ret = -ENOMEM; goto reject; } /* reclaim space if needed */ if (zswap_is_full()) { zswap_pool_limit_hit++; if (zswap_shrink()) { zswap_reject_reclaim_fail++; ret = -ENOMEM; goto reject; } /* A second zswap_is_full() check after * zswap_shrink() to make sure it's now * under the max_pool_percent */ if (zswap_is_full()) { ret = -ENOMEM; goto reject; } } /* allocate entry */ entry = zswap_entry_cache_alloc(GFP_KERNEL); if (!entry) { zswap_reject_kmemcache_fail++; ret = -ENOMEM; goto reject; } /* if entry is successfully added, it keeps the reference */ entry->pool = zswap_pool_current_get(); if (!entry->pool) { ret = -EINVAL; goto freepage; } /* compress */ src = kmap_atomic(page); if (page_zero_filled(src)) { atomic_inc(&zswap_zero_pages); entry->zero_flag = 1; kunmap_atomic(src); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING entry->offset = offset; goto insert_entry; #else handle = 0; dlen = PAGE_SIZE; goto zeropage_out; #endif } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING checksum = jhash2((const u32 *)src, PAGE_SIZE / 4, 17); spin_lock(&tree->lock); zhandle = zswap_same_page_search(entry->pool, tree, src, checksum); if (zhandle) { entry->offset = offset; entry->zhandle = zhandle; entry->zhandle->ref_count++; spin_unlock(&tree->lock); kunmap_atomic(src); atomic_inc(&zswap_duplicate_pages); goto insert_entry; } spin_unlock(&tree->lock); #endif /* compress */ dst = get_cpu_var(zswap_dstmem); tfm = *get_cpu_ptr(entry->pool->tfm); ret = crypto_comp_compress(tfm, src, PAGE_SIZE, dst, &dlen); kunmap_atomic(src); put_cpu_ptr(entry->pool->tfm); if (ret) { ret = -EINVAL; goto put_dstmem; } /* store */ if (dlen > PAGE_SIZE) dlen = PAGE_SIZE; len = dlen; #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK len += sizeof(struct zswap_header); #endif ret = zpool_malloc(entry->pool->zpool, len, __GFP_NORETRY | __GFP_NOWARN | __GFP_KSWAPD_RECLAIM, &handle); if (ret == -ENOSPC) { zswap_reject_compress_poor++; goto put_dstmem; } if (ret) { zswap_reject_alloc_fail++; goto put_dstmem; } #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK zhdr = zpool_map_handle(entry->pool->zpool, handle, ZPOOL_MM_RW); zhdr->swpentry = swp_entry(type, offset); buf = (u8 *)(zhdr + 1); memcpy(buf, dst, dlen); #else buf = (u8 *)zpool_map_handle(entry->pool->zpool, handle, ZPOOL_MM_RW); if (dlen == PAGE_SIZE) { src = kmap_atomic(page); copy_page(buf, src); kunmap_atomic(src); } else memcpy(buf, dst, dlen); #endif zpool_unmap_handle(entry->pool->zpool, handle); put_cpu_var(zswap_dstmem); #ifndef CONFIG_ZSWAP_SAME_PAGE_SHARING zeropage_out: #endif /* populate entry */ entry->offset = offset; #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING zhandle = zswap_handle_cache_alloc(GFP_KERNEL); if (!zhandle) { ret = -ENOMEM; goto freeentry; } entry->zhandle = zhandle; entry->zhandle->handle = handle; entry->zhandle->length = dlen; entry->zhandle->checksum = checksum; #else entry->handle = handle; entry->length = dlen; #endif #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING spin_lock(&tree->lock); ret = zswap_handle_rb_insert(&tree->zhandleroot, entry->zhandle, &duphandle); spin_unlock(&tree->lock); insert_entry: #endif /* map */ spin_lock(&tree->lock); do { ret = zswap_rb_insert(&tree->rbroot, entry, &dupentry); if (ret == -EEXIST) { zswap_duplicate_entry++; /* remove from rbtree */ zswap_rb_erase(&tree->rbroot, dupentry); zswap_entry_put(tree, dupentry); } } while (ret == -EEXIST); spin_unlock(&tree->lock); /* update stats */ atomic_inc(&zswap_stored_pages); zswap_update_total_size(); return 0; put_dstmem: put_cpu_var(zswap_dstmem); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING freeentry: #endif zswap_pool_put(entry->pool); freepage: zswap_entry_cache_free(entry); reject: return ret; } static void hexdump(char *title, u8 *data, int len) { int i; printk("%s: length = %d @ %p\n", title, len, data); for (i = 0; i < len; i++) { printk("%02x ", data[i]); if ((i & 0xf) == 0xf) printk("\n"); } printk("\n"); } /* * returns 0 if the page was successfully decompressed * return -1 on entry not found or error */ static int zswap_frontswap_load(unsigned type, pgoff_t offset, struct page *page) { struct zswap_tree *tree = zswap_trees[type]; struct zswap_entry *entry; struct crypto_comp *tfm; u8 *src, *dst; unsigned int dlen; int ret = 0; /* find */ spin_lock(&tree->lock); entry = zswap_entry_find_get(&tree->rbroot, offset); if (!entry) { /* entry was written back */ spin_unlock(&tree->lock); return -1; } spin_unlock(&tree->lock); if (entry->zero_flag == 1) { dst = kmap_atomic(page); memset(dst, 0, PAGE_SIZE); kunmap_atomic(dst); goto zeropage_out; } /* decompress */ dlen = PAGE_SIZE; #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->zhandle->handle, ZPOOL_MM_RO); #else src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->handle, ZPOOL_MM_RO); #endif dst = kmap_atomic(page); #ifdef CONFIG_ZSWAP_ENABLE_WRITEBACK src += sizeof(struct zswap_header); tfm = *get_cpu_ptr(entry->pool->tfm); ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen); put_cpu_ptr(entry->pool->tfm); #else #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING if (entry->zhandle->length == PAGE_SIZE) copy_page(dst, src); else { tfm = *get_cpu_ptr(entry->pool->tfm); ret = crypto_comp_decompress(tfm, src, entry->zhandle->length, dst, &dlen); put_cpu_ptr(entry->pool->tfm); } #else if (entry->length == PAGE_SIZE) copy_page(dst, src); else { tfm = *get_cpu_ptr(entry->pool->tfm); ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen); put_cpu_ptr(entry->pool->tfm); } #endif #endif if (ret) { #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING hexdump("src buffer", src, entry->zhandle->length); #else hexdump("src buffer", src, entry->length); #endif if (dlen) hexdump("dest buffer", dst, dlen); printk("zswap_comp_op returned %d\n", ret); } kunmap_atomic(dst); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING zpool_unmap_handle(entry->pool->zpool, entry->zhandle->handle); #else zpool_unmap_handle(entry->pool->zpool, entry->handle); #endif BUG_ON(ret); zeropage_out: spin_lock(&tree->lock); zswap_entry_put(tree, entry); spin_unlock(&tree->lock); return 0; } /* frees an entry in zswap */ static void zswap_frontswap_invalidate_page(unsigned type, pgoff_t offset) { struct zswap_tree *tree = zswap_trees[type]; struct zswap_entry *entry; /* find */ spin_lock(&tree->lock); entry = zswap_rb_search(&tree->rbroot, offset); if (!entry) { /* entry was written back */ spin_unlock(&tree->lock); return; } /* remove from rbtree */ zswap_rb_erase(&tree->rbroot, entry); /* drop the initial reference from entry creation */ zswap_entry_put(tree, entry); spin_unlock(&tree->lock); } /* frees all zswap entries for the given swap type */ static void zswap_frontswap_invalidate_area(unsigned type) { struct zswap_tree *tree = zswap_trees[type]; struct zswap_entry *entry, *n; if (!tree) return; /* walk the tree and free everything */ spin_lock(&tree->lock); rbtree_postorder_for_each_entry_safe(entry, n, &tree->rbroot, rbnode) zswap_free_entry(entry); tree->rbroot = RB_ROOT; #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING tree->zhandleroot = RB_ROOT; #endif spin_unlock(&tree->lock); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING free_page((unsigned long)tree->buffer); #endif kfree(tree); zswap_trees[type] = NULL; } static void zswap_frontswap_init(unsigned type) { struct zswap_tree *tree; tree = kzalloc(sizeof(struct zswap_tree), GFP_KERNEL); if (!tree) { pr_err("alloc failed, zswap disabled for swap type %d\n", type); return; } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING tree->buffer = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO); if (!tree->buffer) { pr_err("zswap: Error allocating compressor buffer\n"); kfree(tree); return; } tree->zhandleroot = RB_ROOT; #endif tree->rbroot = RB_ROOT; spin_lock_init(&tree->lock); zswap_trees[type] = tree; } static struct frontswap_ops zswap_frontswap_ops = { .store = zswap_frontswap_store, .load = zswap_frontswap_load, .invalidate_page = zswap_frontswap_invalidate_page, .invalidate_area = zswap_frontswap_invalidate_area, .init = zswap_frontswap_init }; /********************************* * debugfs functions **********************************/ #ifdef CONFIG_DEBUG_FS #include static struct dentry *zswap_debugfs_root; static int __init zswap_debugfs_init(void) { if (!debugfs_initialized()) return -ENODEV; zswap_debugfs_root = debugfs_create_dir("zswap", NULL); if (!zswap_debugfs_root) return -ENOMEM; debugfs_create_u64("pool_limit_hit", S_IRUGO, zswap_debugfs_root, &zswap_pool_limit_hit); debugfs_create_u64("reject_reclaim_fail", S_IRUGO, zswap_debugfs_root, &zswap_reject_reclaim_fail); debugfs_create_u64("reject_alloc_fail", S_IRUGO, zswap_debugfs_root, &zswap_reject_alloc_fail); debugfs_create_u64("reject_kmemcache_fail", S_IRUGO, zswap_debugfs_root, &zswap_reject_kmemcache_fail); debugfs_create_u64("reject_compress_poor", S_IRUGO, zswap_debugfs_root, &zswap_reject_compress_poor); debugfs_create_u64("written_back_pages", S_IRUGO, zswap_debugfs_root, &zswap_written_back_pages); debugfs_create_u64("duplicate_entry", S_IRUGO, zswap_debugfs_root, &zswap_duplicate_entry); debugfs_create_u64("pool_total_size", S_IRUGO, zswap_debugfs_root, &zswap_pool_total_size); debugfs_create_u64("pool_pages", S_IRUGO, zswap_debugfs_root, &zswap_pool_pages); debugfs_create_atomic_t("stored_pages", S_IRUGO, zswap_debugfs_root, &zswap_stored_pages); #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING debugfs_create_atomic_t("duplicate_pages", S_IRUGO, zswap_debugfs_root, &zswap_duplicate_pages); #endif debugfs_create_atomic_t("zero_pages", S_IRUGO, zswap_debugfs_root, &zswap_zero_pages); return 0; } static void __exit zswap_debugfs_exit(void) { debugfs_remove_recursive(zswap_debugfs_root); } #else static int __init zswap_debugfs_init(void) { return 0; } static void __exit zswap_debugfs_exit(void) { } #endif static int zswap_size_notifier(struct notifier_block *nb, unsigned long action, void *data) { struct seq_file *s; s = (struct seq_file *)data; if (s) seq_printf(s, "ZSwapDevice: %8lu kB\n", (unsigned long)zswap_pool_pages << (PAGE_SHIFT - 10)); else pr_cont("ZSwapDevice:%lukB ", (unsigned long)zswap_pool_pages << (PAGE_SHIFT - 10)); return 0; } static struct notifier_block zswap_size_nb = { .notifier_call = zswap_size_notifier, }; /********************************* * module init and exit **********************************/ static int __init init_zswap(void) { struct zswap_pool *pool; zswap_init_started = true; if (zswap_entry_cache_create()) { pr_err("entry cache creation failed\n"); goto cache_fail; } #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING if (zswap_handle_cache_create()) { pr_err("handle cache creation failed\n"); goto handlecachefail; } #endif if (zswap_cpu_dstmem_init()) { pr_err("dstmem alloc failed\n"); goto dstmem_fail; } pool = __zswap_pool_create_fallback(); if (!pool) { pr_err("pool creation failed\n"); goto pool_fail; } pr_info("loaded using pool %s/%s\n", pool->tfm_name, zpool_get_type(pool->zpool)); list_add(&pool->list, &zswap_pools); frontswap_register_ops(&zswap_frontswap_ops); if (zswap_debugfs_init()) pr_warn("debugfs initialization failed\n"); show_mem_extra_notifier_register(&zswap_size_nb); return 0; pool_fail: zswap_cpu_dstmem_destroy(); dstmem_fail: #ifdef CONFIG_ZSWAP_SAME_PAGE_SHARING zswap_handle_cache_destroy(); handlecachefail: #endif zswap_entry_cache_destroy(); cache_fail: /* if built-in, we aren't unloaded on failure; don't allow use */ zswap_init_failed = true; zswap_enabled = false; return -ENOMEM; } /* must be late so crypto has time to come up */ late_initcall(init_zswap); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Seth Jennings "); MODULE_DESCRIPTION("Compressed cache for swap pages");