zswap: do not map same object twice

zswap_writeback_entry() maps a handle to read swpentry first, and
then in the most common case it would map the same handle again.
This is ok when zbud is the backend since its mapping callback is
plain and simple, but it slows things down for z3fold.

Since there's hardly a point in unmapping a handle _that_ fast as
zswap_writeback_entry() does when it reads swpentry, the
suggestion is to keep the handle mapped till the end.

Link: http://lkml.kernel.org/r/20190916004640.b453167d3556c4093af4cf7d@gmail.com
Signed-off-by: Vitaly Wool <vitalywool@gmail.com>
Reviewed-by: Dan Streetman <ddstreet@ieee.org>
Cc: Shakeel Butt <shakeelb@google.com>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Sergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
Cc: Seth Jennings <sjenning@redhat.com>
Cc: Vitaly Wool <vitalywool@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Vitaly Wool 2019-09-23 15:39:43 -07:00 committed by Linus Torvalds
parent d2fcd82bb8
commit 068619e32f
1 changed files with 3 additions and 4 deletions

View File

@ -856,7 +856,6 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
/* extract swpentry from data */ /* extract swpentry from data */
zhdr = zpool_map_handle(pool, handle, ZPOOL_MM_RO); zhdr = zpool_map_handle(pool, handle, ZPOOL_MM_RO);
swpentry = zhdr->swpentry; /* here */ swpentry = zhdr->swpentry; /* here */
zpool_unmap_handle(pool, handle);
tree = zswap_trees[swp_type(swpentry)]; tree = zswap_trees[swp_type(swpentry)];
offset = swp_offset(swpentry); offset = swp_offset(swpentry);
@ -866,6 +865,7 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
if (!entry) { if (!entry) {
/* entry was invalidated */ /* entry was invalidated */
spin_unlock(&tree->lock); spin_unlock(&tree->lock);
zpool_unmap_handle(pool, handle);
return 0; return 0;
} }
spin_unlock(&tree->lock); spin_unlock(&tree->lock);
@ -886,15 +886,13 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
case ZSWAP_SWAPCACHE_NEW: /* page is locked */ case ZSWAP_SWAPCACHE_NEW: /* page is locked */
/* decompress */ /* decompress */
dlen = PAGE_SIZE; dlen = PAGE_SIZE;
src = (u8 *)zpool_map_handle(entry->pool->zpool, entry->handle, src = (u8 *)zhdr + sizeof(struct zswap_header);
ZPOOL_MM_RO) + sizeof(struct zswap_header);
dst = kmap_atomic(page); dst = kmap_atomic(page);
tfm = *get_cpu_ptr(entry->pool->tfm); tfm = *get_cpu_ptr(entry->pool->tfm);
ret = crypto_comp_decompress(tfm, src, entry->length, ret = crypto_comp_decompress(tfm, src, entry->length,
dst, &dlen); dst, &dlen);
put_cpu_ptr(entry->pool->tfm); put_cpu_ptr(entry->pool->tfm);
kunmap_atomic(dst); kunmap_atomic(dst);
zpool_unmap_handle(entry->pool->zpool, entry->handle);
BUG_ON(ret); BUG_ON(ret);
BUG_ON(dlen != PAGE_SIZE); BUG_ON(dlen != PAGE_SIZE);
@ -940,6 +938,7 @@ static int zswap_writeback_entry(struct zpool *pool, unsigned long handle)
spin_unlock(&tree->lock); spin_unlock(&tree->lock);
end: end:
zpool_unmap_handle(pool, handle);
return ret; return ret;
} }