zswap: dynamic pool creation

Add dynamic creation of pools.  Move the static crypto compression per-cpu
transforms into each pool.  Add a pointer to zswap_entry to the pool it's
in.

This is required by the following patch which enables changing the zswap
zpool and compressor params at runtime.

[akpm@linux-foundation.org: fix merge snafus]
Signed-off-by: Dan Streetman <ddstreet@ieee.org>
Acked-by: Seth Jennings <sjennings@variantweb.net>
Cc: Sergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Dan Streetman 2015-09-09 15:35:19 -07:00 committed by Linus Torvalds
parent 3f0e131221
commit f1c54846ee
1 changed files with 413 additions and 151 deletions

View File

@ -98,67 +98,20 @@ module_param_named(zpool, zswap_zpool_type, charp, 0444);
/* zpool is shared by all of zswap backend */
static struct zpool *zswap_pool;
/*********************************
* compression functions
**********************************/
/* per-cpu compression transforms */
static struct crypto_comp * __percpu *zswap_comp_pcpu_tfms;
enum comp_op {
ZSWAP_COMPOP_COMPRESS,
ZSWAP_COMPOP_DECOMPRESS
};
static int zswap_comp_op(enum comp_op op, const u8 *src, unsigned int slen,
u8 *dst, unsigned int *dlen)
{
struct crypto_comp *tfm;
int ret;
tfm = *per_cpu_ptr(zswap_comp_pcpu_tfms, get_cpu());
switch (op) {
case ZSWAP_COMPOP_COMPRESS:
ret = crypto_comp_compress(tfm, src, slen, dst, dlen);
break;
case ZSWAP_COMPOP_DECOMPRESS:
ret = crypto_comp_decompress(tfm, src, slen, dst, dlen);
break;
default:
ret = -EINVAL;
}
put_cpu();
return ret;
}
static int __init zswap_comp_init(void)
{
if (!crypto_has_comp(zswap_compressor, 0, 0)) {
pr_info("%s compressor not available\n", zswap_compressor);
/* fall back to default compressor */
zswap_compressor = ZSWAP_COMPRESSOR_DEFAULT;
if (!crypto_has_comp(zswap_compressor, 0, 0))
/* can't even load the default compressor */
return -ENODEV;
}
pr_info("using %s compressor\n", zswap_compressor);
/* alloc percpu transforms */
zswap_comp_pcpu_tfms = alloc_percpu(struct crypto_comp *);
if (!zswap_comp_pcpu_tfms)
return -ENOMEM;
return 0;
}
static void __init zswap_comp_exit(void)
{
/* free percpu transforms */
free_percpu(zswap_comp_pcpu_tfms);
}
/*********************************
* data structures
**********************************/
struct zswap_pool {
struct zpool *zpool;
struct crypto_comp * __percpu *tfm;
struct kref kref;
struct list_head list;
struct rcu_head rcu_head;
struct notifier_block notifier;
char tfm_name[CRYPTO_MAX_ALG_NAME];
};
/*
* struct zswap_entry
*
@ -166,22 +119,24 @@ static void __init zswap_comp_exit(void)
* page within zswap.
*
* rbnode - links the entry into red-black tree for the appropriate swap type
* offset - the swap offset for the entry. Index into the red-black tree.
* refcount - the number of outstanding reference to the entry. This is needed
* to protect against premature freeing of the entry by code
* concurrent calls to load, invalidate, and writeback. The lock
* for the zswap_tree structure that contains the entry must
* be held while changing the refcount. Since the lock must
* be held, there is no reason to also make refcount atomic.
* offset - the swap offset for the entry. Index into the red-black tree.
* handle - zpool allocation handle that stores the compressed page data
* length - the length in bytes of the compressed page data. Needed during
* decompression
* pool - the zswap_pool the entry's data is in
* handle - zpool allocation handle that stores the compressed page data
*/
struct zswap_entry {
struct rb_node rbnode;
pgoff_t offset;
int refcount;
unsigned int length;
struct zswap_pool *pool;
unsigned long handle;
};
@ -201,6 +156,48 @@ struct zswap_tree {
static struct zswap_tree *zswap_trees[MAX_SWAPFILES];
/* RCU-protected iteration */
static LIST_HEAD(zswap_pools);
/* protects zswap_pools list modification */
static DEFINE_SPINLOCK(zswap_pools_lock);
/*********************************
* helpers and fwd declarations
**********************************/
#define zswap_pool_debug(msg, p) \
pr_debug("%s pool %s/%s\n", msg, (p)->tfm_name, \
zpool_get_type((p)->zpool))
static int zswap_writeback_entry(struct zpool *pool, unsigned long handle);
static int zswap_pool_get(struct zswap_pool *pool);
static void zswap_pool_put(struct zswap_pool *pool);
static const struct zpool_ops zswap_zpool_ops = {
.evict = zswap_writeback_entry
};
static bool zswap_is_full(void)
{
return totalram_pages * zswap_max_pool_percent / 100 <
DIV_ROUND_UP(zswap_pool_total_size, PAGE_SIZE);
}
static void zswap_update_total_size(void)
{
struct zswap_pool *pool;
u64 total = 0;
rcu_read_lock();
list_for_each_entry_rcu(pool, &zswap_pools, list)
total += zpool_get_total_size(pool->zpool);
rcu_read_unlock();
zswap_pool_total_size = total;
}
/*********************************
* zswap entry functions
**********************************/
@ -294,10 +291,11 @@ static void zswap_rb_erase(struct rb_root *root, struct zswap_entry *entry)
*/
static void zswap_free_entry(struct zswap_entry *entry)
{
zpool_free(zswap_pool, entry->handle);
zpool_free(entry->pool->zpool, entry->handle);
zswap_pool_put(entry->pool);
zswap_entry_cache_free(entry);
atomic_dec(&zswap_stored_pages);
zswap_pool_total_size = zpool_get_total_size(zswap_pool);
zswap_update_total_size();
}
/* caller must hold the tree lock */
@ -339,35 +337,21 @@ static struct zswap_entry *zswap_entry_find_get(struct rb_root *root,
**********************************/
static DEFINE_PER_CPU(u8 *, zswap_dstmem);
static int __zswap_cpu_notifier(unsigned long action, unsigned long cpu)
static int __zswap_cpu_dstmem_notifier(unsigned long action, unsigned long cpu)
{
struct crypto_comp *tfm;
u8 *dst;
switch (action) {
case CPU_UP_PREPARE:
tfm = crypto_alloc_comp(zswap_compressor, 0, 0);
if (IS_ERR(tfm)) {
pr_err("can't allocate compressor transform\n");
return NOTIFY_BAD;
}
*per_cpu_ptr(zswap_comp_pcpu_tfms, cpu) = tfm;
dst = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu));
if (!dst) {
pr_err("can't allocate compressor buffer\n");
crypto_free_comp(tfm);
*per_cpu_ptr(zswap_comp_pcpu_tfms, cpu) = NULL;
return NOTIFY_BAD;
}
per_cpu(zswap_dstmem, cpu) = dst;
break;
case CPU_DEAD:
case CPU_UP_CANCELED:
tfm = *per_cpu_ptr(zswap_comp_pcpu_tfms, cpu);
if (tfm) {
crypto_free_comp(tfm);
*per_cpu_ptr(zswap_comp_pcpu_tfms, cpu) = NULL;
}
dst = per_cpu(zswap_dstmem, cpu);
kfree(dst);
per_cpu(zswap_dstmem, cpu) = NULL;
@ -378,43 +362,303 @@ static int __zswap_cpu_notifier(unsigned long action, unsigned long cpu)
return NOTIFY_OK;
}
static int zswap_cpu_notifier(struct notifier_block *nb,
unsigned long action, void *pcpu)
static int zswap_cpu_dstmem_notifier(struct notifier_block *nb,
unsigned long action, void *pcpu)
{
unsigned long cpu = (unsigned long)pcpu;
return __zswap_cpu_notifier(action, cpu);
return __zswap_cpu_dstmem_notifier(action, (unsigned long)pcpu);
}
static struct notifier_block zswap_cpu_notifier_block = {
.notifier_call = zswap_cpu_notifier
static struct notifier_block zswap_dstmem_notifier = {
.notifier_call = zswap_cpu_dstmem_notifier,
};
static int __init zswap_cpu_init(void)
static int __init zswap_cpu_dstmem_init(void)
{
unsigned long cpu;
cpu_notifier_register_begin();
for_each_online_cpu(cpu)
if (__zswap_cpu_notifier(CPU_UP_PREPARE, cpu) != NOTIFY_OK)
if (__zswap_cpu_dstmem_notifier(CPU_UP_PREPARE, cpu) ==
NOTIFY_BAD)
goto cleanup;
__register_cpu_notifier(&zswap_cpu_notifier_block);
__register_cpu_notifier(&zswap_dstmem_notifier);
cpu_notifier_register_done();
return 0;
cleanup:
for_each_online_cpu(cpu)
__zswap_cpu_notifier(CPU_UP_CANCELED, cpu);
__zswap_cpu_dstmem_notifier(CPU_UP_CANCELED, cpu);
cpu_notifier_register_done();
return -ENOMEM;
}
/*********************************
* helpers
**********************************/
static bool zswap_is_full(void)
static void zswap_cpu_dstmem_destroy(void)
{
return totalram_pages * zswap_max_pool_percent / 100 <
DIV_ROUND_UP(zswap_pool_total_size, PAGE_SIZE);
unsigned long cpu;
cpu_notifier_register_begin();
for_each_online_cpu(cpu)
__zswap_cpu_dstmem_notifier(CPU_UP_CANCELED, cpu);
__unregister_cpu_notifier(&zswap_dstmem_notifier);
cpu_notifier_register_done();
}
static int __zswap_cpu_comp_notifier(struct zswap_pool *pool,
unsigned long action, unsigned long cpu)
{
struct crypto_comp *tfm;
switch (action) {
case CPU_UP_PREPARE:
if (WARN_ON(*per_cpu_ptr(pool->tfm, cpu)))
break;
tfm = crypto_alloc_comp(pool->tfm_name, 0, 0);
if (IS_ERR_OR_NULL(tfm)) {
pr_err("could not alloc crypto comp %s : %ld\n",
pool->tfm_name, PTR_ERR(tfm));
return NOTIFY_BAD;
}
*per_cpu_ptr(pool->tfm, cpu) = tfm;
break;
case CPU_DEAD:
case CPU_UP_CANCELED:
tfm = *per_cpu_ptr(pool->tfm, cpu);
if (!IS_ERR_OR_NULL(tfm))
crypto_free_comp(tfm);
*per_cpu_ptr(pool->tfm, cpu) = NULL;
break;
default:
break;
}
return NOTIFY_OK;
}
static int zswap_cpu_comp_notifier(struct notifier_block *nb,
unsigned long action, void *pcpu)
{
unsigned long cpu = (unsigned long)pcpu;
struct zswap_pool *pool = container_of(nb, typeof(*pool), notifier);
return __zswap_cpu_comp_notifier(pool, action, cpu);
}
static int zswap_cpu_comp_init(struct zswap_pool *pool)
{
unsigned long cpu;
memset(&pool->notifier, 0, sizeof(pool->notifier));
pool->notifier.notifier_call = zswap_cpu_comp_notifier;
cpu_notifier_register_begin();
for_each_online_cpu(cpu)
if (__zswap_cpu_comp_notifier(pool, CPU_UP_PREPARE, cpu) ==
NOTIFY_BAD)
goto cleanup;
__register_cpu_notifier(&pool->notifier);
cpu_notifier_register_done();
return 0;
cleanup:
for_each_online_cpu(cpu)
__zswap_cpu_comp_notifier(pool, CPU_UP_CANCELED, cpu);
cpu_notifier_register_done();
return -ENOMEM;
}
static void zswap_cpu_comp_destroy(struct zswap_pool *pool)
{
unsigned long cpu;
cpu_notifier_register_begin();
for_each_online_cpu(cpu)
__zswap_cpu_comp_notifier(pool, CPU_UP_CANCELED, cpu);
__unregister_cpu_notifier(&pool->notifier);
cpu_notifier_register_done();
}
/*********************************
* pool functions
**********************************/
static struct zswap_pool *__zswap_pool_current(void)
{
struct zswap_pool *pool;
pool = list_first_or_null_rcu(&zswap_pools, typeof(*pool), list);
WARN_ON(!pool);
return pool;
}
static struct zswap_pool *zswap_pool_current(void)
{
assert_spin_locked(&zswap_pools_lock);
return __zswap_pool_current();
}
static struct zswap_pool *zswap_pool_current_get(void)
{
struct zswap_pool *pool;
rcu_read_lock();
pool = __zswap_pool_current();
if (!pool || !zswap_pool_get(pool))
pool = NULL;
rcu_read_unlock();
return pool;
}
static struct zswap_pool *zswap_pool_last_get(void)
{
struct zswap_pool *pool, *last = NULL;
rcu_read_lock();
list_for_each_entry_rcu(pool, &zswap_pools, list)
last = pool;
if (!WARN_ON(!last) && !zswap_pool_get(last))
last = NULL;
rcu_read_unlock();
return last;
}
static struct zswap_pool *zswap_pool_find_get(char *type, char *compressor)
{
struct zswap_pool *pool;
assert_spin_locked(&zswap_pools_lock);
list_for_each_entry_rcu(pool, &zswap_pools, list) {
if (strncmp(pool->tfm_name, compressor, sizeof(pool->tfm_name)))
continue;
if (strncmp(zpool_get_type(pool->zpool), type,
sizeof(zswap_zpool_type)))
continue;
/* if we can't get it, it's about to be destroyed */
if (!zswap_pool_get(pool))
continue;
return pool;
}
return NULL;
}
static struct zswap_pool *zswap_pool_create(char *type, char *compressor)
{
struct zswap_pool *pool;
gfp_t gfp = __GFP_NORETRY | __GFP_NOWARN;
pool = kzalloc(sizeof(*pool), GFP_KERNEL);
if (!pool) {
pr_err("pool alloc failed\n");
return NULL;
}
pool->zpool = zpool_create_pool(type, "zswap", gfp, &zswap_zpool_ops);
if (!pool->zpool) {
pr_err("%s zpool not available\n", type);
goto error;
}
pr_debug("using %s zpool\n", zpool_get_type(pool->zpool));
strlcpy(pool->tfm_name, compressor, sizeof(pool->tfm_name));
pool->tfm = alloc_percpu(struct crypto_comp *);
if (!pool->tfm) {
pr_err("percpu alloc failed\n");
goto error;
}
if (zswap_cpu_comp_init(pool))
goto error;
pr_debug("using %s compressor\n", pool->tfm_name);
/* being the current pool takes 1 ref; this func expects the
* caller to always add the new pool as the current pool
*/
kref_init(&pool->kref);
INIT_LIST_HEAD(&pool->list);
zswap_pool_debug("created", pool);
return pool;
error:
free_percpu(pool->tfm);
if (pool->zpool)
zpool_destroy_pool(pool->zpool);
kfree(pool);
return NULL;
}
static struct zswap_pool *__zswap_pool_create_fallback(void)
{
if (!crypto_has_comp(zswap_compressor, 0, 0)) {
pr_err("compressor %s not available, using default %s\n",
zswap_compressor, ZSWAP_COMPRESSOR_DEFAULT);
strncpy(zswap_compressor, ZSWAP_COMPRESSOR_DEFAULT,
sizeof(zswap_compressor));
}
if (!zpool_has_pool(zswap_zpool_type)) {
pr_err("zpool %s not available, using default %s\n",
zswap_zpool_type, ZSWAP_ZPOOL_DEFAULT);
strncpy(zswap_zpool_type, ZSWAP_ZPOOL_DEFAULT,
sizeof(zswap_zpool_type));
}
return zswap_pool_create(zswap_zpool_type, zswap_compressor);
}
static void zswap_pool_destroy(struct zswap_pool *pool)
{
zswap_pool_debug("destroying", pool);
zswap_cpu_comp_destroy(pool);
free_percpu(pool->tfm);
zpool_destroy_pool(pool->zpool);
kfree(pool);
}
static int __must_check zswap_pool_get(struct zswap_pool *pool)
{
return kref_get_unless_zero(&pool->kref);
}
static void __zswap_pool_release(struct rcu_head *head)
{
struct zswap_pool *pool = container_of(head, typeof(*pool), rcu_head);
/* nobody should have been able to get a kref... */
WARN_ON(kref_get_unless_zero(&pool->kref));
/* pool is now off zswap_pools list and has no references. */
zswap_pool_destroy(pool);
}
static void __zswap_pool_empty(struct kref *kref)
{
struct zswap_pool *pool;
pool = container_of(kref, typeof(*pool), kref);
spin_lock(&zswap_pools_lock);
WARN_ON(pool == zswap_pool_current());
list_del_rcu(&pool->list);
call_rcu(&pool->rcu_head, __zswap_pool_release);
spin_unlock(&zswap_pools_lock);
}
static void zswap_pool_put(struct zswap_pool *pool)
{
kref_put(&pool->kref, __zswap_pool_empty);
}
/*********************************
@ -477,6 +721,7 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
pgoff_t offset;
struct zswap_entry *entry;
struct page *page;
struct crypto_comp *tfm;
u8 *src, *dst;
unsigned int dlen;
int ret;
@ -517,13 +762,15 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
case ZSWAP_SWAPCACHE_NEW: /* page is locked */
/* decompress */
dlen = PAGE_SIZE;
src = (u8 *)zpool_map_handle(zswap_pool, entry->handle,
src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->handle,
ZPOOL_MM_RO) + sizeof(struct zswap_header);
dst = kmap_atomic(page);
ret = zswap_comp_op(ZSWAP_COMPOP_DECOMPRESS, src,
entry->length, dst, &dlen);
tfm = *get_cpu_ptr(entry->pool->tfm);
ret = crypto_comp_decompress(tfm, src, entry->length,
dst, &dlen);
put_cpu_ptr(entry->pool->tfm);
kunmap_atomic(dst);
zpool_unmap_handle(zswap_pool, entry->handle);
zpool_unmap_handle(entry->pool->zpool, entry->handle);
BUG_ON(ret);
BUG_ON(dlen != PAGE_SIZE);
@ -572,6 +819,22 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
return ret;
}
static int zswap_shrink(void)
{
struct zswap_pool *pool;
int ret;
pool = zswap_pool_last_get();
if (!pool)
return -ENOENT;
ret = zpool_shrink(pool->zpool, 1, NULL);
zswap_pool_put(pool);
return ret;
}
/*********************************
* frontswap hooks
**********************************/
@ -581,6 +844,7 @@ static int zswap_frontswap_store(unsigned type, pgoff_t offset,
{
struct zswap_tree *tree = zswap_trees[type];
struct zswap_entry *entry, *dupentry;
struct crypto_comp *tfm;
int ret;
unsigned int dlen = PAGE_SIZE, len;
unsigned long handle;
@ -596,7 +860,7 @@ static int zswap_frontswap_store(unsigned type, pgoff_t offset,
/* reclaim space if needed */
if (zswap_is_full()) {
zswap_pool_limit_hit++;
if (zpool_shrink(zswap_pool, 1, NULL)) {
if (zswap_shrink()) {
zswap_reject_reclaim_fail++;
ret = -ENOMEM;
goto reject;
@ -611,33 +875,42 @@ static int zswap_frontswap_store(unsigned type, pgoff_t offset,
goto reject;
}
/* compress */
dst = get_cpu_var(zswap_dstmem);
src = kmap_atomic(page);
ret = zswap_comp_op(ZSWAP_COMPOP_COMPRESS, src, PAGE_SIZE, dst, &dlen);
kunmap_atomic(src);
if (ret) {
/* if entry is successfully added, it keeps the reference */
entry->pool = zswap_pool_current_get();
if (!entry->pool) {
ret = -EINVAL;
goto freepage;
}
/* compress */
dst = get_cpu_var(zswap_dstmem);
tfm = *get_cpu_ptr(entry->pool->tfm);
src = kmap_atomic(page);
ret = crypto_comp_compress(tfm, src, PAGE_SIZE, dst, &dlen);
kunmap_atomic(src);
put_cpu_ptr(entry->pool->tfm);
if (ret) {
ret = -EINVAL;
goto put_dstmem;
}
/* store */
len = dlen + sizeof(struct zswap_header);
ret = zpool_malloc(zswap_pool, len, __GFP_NORETRY | __GFP_NOWARN,
&handle);
ret = zpool_malloc(entry->pool->zpool, len,
__GFP_NORETRY | __GFP_NOWARN, &handle);
if (ret == -ENOSPC) {
zswap_reject_compress_poor++;
goto freepage;
goto put_dstmem;
}
if (ret) {
zswap_reject_alloc_fail++;
goto freepage;
goto put_dstmem;
}
zhdr = zpool_map_handle(zswap_pool, handle, ZPOOL_MM_RW);
zhdr = zpool_map_handle(entry->pool->zpool, handle, ZPOOL_MM_RW);
zhdr->swpentry = swp_entry(type, offset);
buf = (u8 *)(zhdr + 1);
memcpy(buf, dst, dlen);
zpool_unmap_handle(zswap_pool, handle);
zpool_unmap_handle(entry->pool->zpool, handle);
put_cpu_var(zswap_dstmem);
/* populate entry */
@ -660,12 +933,14 @@ static int zswap_frontswap_store(unsigned type, pgoff_t offset,
/* update stats */
atomic_inc(&zswap_stored_pages);
zswap_pool_total_size = zpool_get_total_size(zswap_pool);
zswap_update_total_size();
return 0;
freepage:
put_dstmem:
put_cpu_var(zswap_dstmem);
zswap_pool_put(entry->pool);
freepage:
zswap_entry_cache_free(entry);
reject:
return ret;
@ -680,6 +955,7 @@ static int zswap_frontswap_load(unsigned type, pgoff_t offset,
{
struct zswap_tree *tree = zswap_trees[type];
struct zswap_entry *entry;
struct crypto_comp *tfm;
u8 *src, *dst;
unsigned int dlen;
int ret;
@ -696,13 +972,14 @@ static int zswap_frontswap_load(unsigned type, pgoff_t offset,
/* decompress */
dlen = PAGE_SIZE;
src = (u8 *)zpool_map_handle(zswap_pool, entry->handle,
src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->handle,
ZPOOL_MM_RO) + sizeof(struct zswap_header);
dst = kmap_atomic(page);
ret = zswap_comp_op(ZSWAP_COMPOP_DECOMPRESS, src, entry->length,
dst, &dlen);
tfm = *get_cpu_ptr(entry->pool->tfm);
ret = crypto_comp_decompress(tfm, src, entry->length, dst, &dlen);
put_cpu_ptr(entry->pool->tfm);
kunmap_atomic(dst);
zpool_unmap_handle(zswap_pool, entry->handle);
zpool_unmap_handle(entry->pool->zpool, entry->handle);
BUG_ON(ret);
spin_lock(&tree->lock);
@ -755,10 +1032,6 @@ static void zswap_frontswap_invalidate_area(unsigned type)
zswap_trees[type] = NULL;
}
static const struct zpool_ops zswap_zpool_ops = {
.evict = zswap_writeback_entry
};
static void zswap_frontswap_init(unsigned type)
{
struct zswap_tree *tree;
@ -839,49 +1112,38 @@ static void __exit zswap_debugfs_exit(void) { }
**********************************/
static int __init init_zswap(void)
{
gfp_t gfp = __GFP_NORETRY | __GFP_NOWARN;
pr_info("loading zswap\n");
zswap_pool = zpool_create_pool(zswap_zpool_type, "zswap", gfp,
&zswap_zpool_ops);
if (!zswap_pool && strcmp(zswap_zpool_type, ZSWAP_ZPOOL_DEFAULT)) {
pr_info("%s zpool not available\n", zswap_zpool_type);
zswap_zpool_type = ZSWAP_ZPOOL_DEFAULT;
zswap_pool = zpool_create_pool(zswap_zpool_type, "zswap", gfp,
&zswap_zpool_ops);
}
if (!zswap_pool) {
pr_err("%s zpool not available\n", zswap_zpool_type);
pr_err("zpool creation failed\n");
goto error;
}
pr_info("using %s pool\n", zswap_zpool_type);
struct zswap_pool *pool;
if (zswap_entry_cache_create()) {
pr_err("entry cache creation failed\n");
goto cachefail;
goto cache_fail;
}
if (zswap_comp_init()) {
pr_err("compressor initialization failed\n");
goto compfail;
if (zswap_cpu_dstmem_init()) {
pr_err("dstmem alloc failed\n");
goto dstmem_fail;
}
if (zswap_cpu_init()) {
pr_err("per-cpu initialization failed\n");
goto pcpufail;
pool = __zswap_pool_create_fallback();
if (!pool) {
pr_err("pool creation failed\n");
goto pool_fail;
}
pr_info("loaded using pool %s/%s\n", pool->tfm_name,
zpool_get_type(pool->zpool));
list_add(&pool->list, &zswap_pools);
frontswap_register_ops(&zswap_frontswap_ops);
if (zswap_debugfs_init())
pr_warn("debugfs initialization failed\n");
return 0;
pcpufail:
zswap_comp_exit();
compfail:
pool_fail:
zswap_cpu_dstmem_destroy();
dstmem_fail:
zswap_entry_cache_destroy();
cachefail:
zpool_destroy_pool(zswap_pool);
error:
cache_fail:
return -ENOMEM;
}
/* must be late so crypto has time to come up */