mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-24 09:25:45 +07:00
3d5fe03a3e
We can end up allocating a new compression stream with GFP_KERNEL from within the IO path, which may result is nested (recursive) IO operations. That can introduce problems if the IO path in question is a reclaimer, holding some locks that will deadlock nested IOs. Allocate streams and working memory using GFP_NOIO flag, forbidding recursive IO and FS operations. An example: inconsistent {IN-RECLAIM_FS-W} -> {RECLAIM_FS-ON-W} usage. git/20158 [HC0[0]:SC0[0]:HE1:SE1] takes: (jbd2_handle){+.+.?.}, at: start_this_handle+0x4ca/0x555 {IN-RECLAIM_FS-W} state was registered at: __lock_acquire+0x8da/0x117b lock_acquire+0x10c/0x1a7 start_this_handle+0x52d/0x555 jbd2__journal_start+0xb4/0x237 __ext4_journal_start_sb+0x108/0x17e ext4_dirty_inode+0x32/0x61 __mark_inode_dirty+0x16b/0x60c iput+0x11e/0x274 __dentry_kill+0x148/0x1b8 shrink_dentry_list+0x274/0x44a prune_dcache_sb+0x4a/0x55 super_cache_scan+0xfc/0x176 shrink_slab.part.14.constprop.25+0x2a2/0x4d3 shrink_zone+0x74/0x140 kswapd+0x6b7/0x930 kthread+0x107/0x10f ret_from_fork+0x3f/0x70 irq event stamp: 138297 hardirqs last enabled at (138297): debug_check_no_locks_freed+0x113/0x12f hardirqs last disabled at (138296): debug_check_no_locks_freed+0x33/0x12f softirqs last enabled at (137818): __do_softirq+0x2d3/0x3e9 softirqs last disabled at (137813): irq_exit+0x41/0x95 other info that might help us debug this: Possible unsafe locking scenario: CPU0 ---- lock(jbd2_handle); <Interrupt> lock(jbd2_handle); *** DEADLOCK *** 5 locks held by git/20158: #0: (sb_writers#7){.+.+.+}, at: [<ffffffff81155411>] mnt_want_write+0x24/0x4b #1: (&type->i_mutex_dir_key#2/1){+.+.+.}, at: [<ffffffff81145087>] lock_rename+0xd9/0xe3 #2: (&sb->s_type->i_mutex_key#11){+.+.+.}, at: [<ffffffff8114f8e2>] lock_two_nondirectories+0x3f/0x6b #3: (&sb->s_type->i_mutex_key#11/4){+.+.+.}, at: [<ffffffff8114f909>] lock_two_nondirectories+0x66/0x6b #4: (jbd2_handle){+.+.?.}, at: [<ffffffff811e31db>] start_this_handle+0x4ca/0x555 stack backtrace: CPU: 2 PID: 20158 Comm: git Not tainted 4.1.0-rc7-next-20150615-dbg-00016-g8bdf555-dirty #211 Call Trace: dump_stack+0x4c/0x6e mark_lock+0x384/0x56d mark_held_locks+0x5f/0x76 lockdep_trace_alloc+0xb2/0xb5 kmem_cache_alloc_trace+0x32/0x1e2 zcomp_strm_alloc+0x25/0x73 [zram] zcomp_strm_multi_find+0xe7/0x173 [zram] zcomp_strm_find+0xc/0xe [zram] zram_bvec_rw+0x2ca/0x7e0 [zram] zram_make_request+0x1fa/0x301 [zram] generic_make_request+0x9c/0xdb submit_bio+0xf7/0x120 ext4_io_submit+0x2e/0x43 ext4_bio_write_page+0x1b7/0x300 mpage_submit_page+0x60/0x77 mpage_map_and_submit_buffers+0x10f/0x21d ext4_writepages+0xc8c/0xe1b do_writepages+0x23/0x2c __filemap_fdatawrite_range+0x84/0x8b filemap_flush+0x1c/0x1e ext4_alloc_da_blocks+0xb8/0x117 ext4_rename+0x132/0x6dc ? mark_held_locks+0x5f/0x76 ext4_rename2+0x29/0x2b vfs_rename+0x540/0x636 SyS_renameat2+0x359/0x44d SyS_rename+0x1e/0x20 entry_SYSCALL_64_fastpath+0x12/0x6f [minchan@kernel.org: add stable mark] Signed-off-by: Sergey Senozhatsky <sergey.senozhatsky@gmail.com> Acked-by: Minchan Kim <minchan@kernel.org> Cc: Kyeongdon Kim <kyeongdon.kim@lge.com> Cc: <stable@vger.kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
361 lines
8.4 KiB
C
361 lines
8.4 KiB
C
/*
|
|
* Copyright (C) 2014 Sergey Senozhatsky.
|
|
*
|
|
* This program is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU General Public License
|
|
* as published by the Free Software Foundation; either version
|
|
* 2 of the License, or (at your option) any later version.
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/string.h>
|
|
#include <linux/err.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/wait.h>
|
|
#include <linux/sched.h>
|
|
|
|
#include "zcomp.h"
|
|
#include "zcomp_lzo.h"
|
|
#ifdef CONFIG_ZRAM_LZ4_COMPRESS
|
|
#include "zcomp_lz4.h"
|
|
#endif
|
|
|
|
/*
|
|
* single zcomp_strm backend
|
|
*/
|
|
struct zcomp_strm_single {
|
|
struct mutex strm_lock;
|
|
struct zcomp_strm *zstrm;
|
|
};
|
|
|
|
/*
|
|
* multi zcomp_strm backend
|
|
*/
|
|
struct zcomp_strm_multi {
|
|
/* protect strm list */
|
|
spinlock_t strm_lock;
|
|
/* max possible number of zstrm streams */
|
|
int max_strm;
|
|
/* number of available zstrm streams */
|
|
int avail_strm;
|
|
/* list of available strms */
|
|
struct list_head idle_strm;
|
|
wait_queue_head_t strm_wait;
|
|
};
|
|
|
|
static struct zcomp_backend *backends[] = {
|
|
&zcomp_lzo,
|
|
#ifdef CONFIG_ZRAM_LZ4_COMPRESS
|
|
&zcomp_lz4,
|
|
#endif
|
|
NULL
|
|
};
|
|
|
|
static struct zcomp_backend *find_backend(const char *compress)
|
|
{
|
|
int i = 0;
|
|
while (backends[i]) {
|
|
if (sysfs_streq(compress, backends[i]->name))
|
|
break;
|
|
i++;
|
|
}
|
|
return backends[i];
|
|
}
|
|
|
|
static void zcomp_strm_free(struct zcomp *comp, struct zcomp_strm *zstrm)
|
|
{
|
|
if (zstrm->private)
|
|
comp->backend->destroy(zstrm->private);
|
|
free_pages((unsigned long)zstrm->buffer, 1);
|
|
kfree(zstrm);
|
|
}
|
|
|
|
/*
|
|
* allocate new zcomp_strm structure with ->private initialized by
|
|
* backend, return NULL on error
|
|
*/
|
|
static struct zcomp_strm *zcomp_strm_alloc(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm *zstrm = kmalloc(sizeof(*zstrm), GFP_NOIO);
|
|
if (!zstrm)
|
|
return NULL;
|
|
|
|
zstrm->private = comp->backend->create();
|
|
/*
|
|
* allocate 2 pages. 1 for compressed data, plus 1 extra for the
|
|
* case when compressed size is larger than the original one
|
|
*/
|
|
zstrm->buffer = (void *)__get_free_pages(GFP_NOIO | __GFP_ZERO, 1);
|
|
if (!zstrm->private || !zstrm->buffer) {
|
|
zcomp_strm_free(comp, zstrm);
|
|
zstrm = NULL;
|
|
}
|
|
return zstrm;
|
|
}
|
|
|
|
/*
|
|
* get idle zcomp_strm or wait until other process release
|
|
* (zcomp_strm_release()) one for us
|
|
*/
|
|
static struct zcomp_strm *zcomp_strm_multi_find(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm_multi *zs = comp->stream;
|
|
struct zcomp_strm *zstrm;
|
|
|
|
while (1) {
|
|
spin_lock(&zs->strm_lock);
|
|
if (!list_empty(&zs->idle_strm)) {
|
|
zstrm = list_entry(zs->idle_strm.next,
|
|
struct zcomp_strm, list);
|
|
list_del(&zstrm->list);
|
|
spin_unlock(&zs->strm_lock);
|
|
return zstrm;
|
|
}
|
|
/* zstrm streams limit reached, wait for idle stream */
|
|
if (zs->avail_strm >= zs->max_strm) {
|
|
spin_unlock(&zs->strm_lock);
|
|
wait_event(zs->strm_wait, !list_empty(&zs->idle_strm));
|
|
continue;
|
|
}
|
|
/* allocate new zstrm stream */
|
|
zs->avail_strm++;
|
|
spin_unlock(&zs->strm_lock);
|
|
|
|
zstrm = zcomp_strm_alloc(comp);
|
|
if (!zstrm) {
|
|
spin_lock(&zs->strm_lock);
|
|
zs->avail_strm--;
|
|
spin_unlock(&zs->strm_lock);
|
|
wait_event(zs->strm_wait, !list_empty(&zs->idle_strm));
|
|
continue;
|
|
}
|
|
break;
|
|
}
|
|
return zstrm;
|
|
}
|
|
|
|
/* add stream back to idle list and wake up waiter or free the stream */
|
|
static void zcomp_strm_multi_release(struct zcomp *comp, struct zcomp_strm *zstrm)
|
|
{
|
|
struct zcomp_strm_multi *zs = comp->stream;
|
|
|
|
spin_lock(&zs->strm_lock);
|
|
if (zs->avail_strm <= zs->max_strm) {
|
|
list_add(&zstrm->list, &zs->idle_strm);
|
|
spin_unlock(&zs->strm_lock);
|
|
wake_up(&zs->strm_wait);
|
|
return;
|
|
}
|
|
|
|
zs->avail_strm--;
|
|
spin_unlock(&zs->strm_lock);
|
|
zcomp_strm_free(comp, zstrm);
|
|
}
|
|
|
|
/* change max_strm limit */
|
|
static bool zcomp_strm_multi_set_max_streams(struct zcomp *comp, int num_strm)
|
|
{
|
|
struct zcomp_strm_multi *zs = comp->stream;
|
|
struct zcomp_strm *zstrm;
|
|
|
|
spin_lock(&zs->strm_lock);
|
|
zs->max_strm = num_strm;
|
|
/*
|
|
* if user has lowered the limit and there are idle streams,
|
|
* immediately free as much streams (and memory) as we can.
|
|
*/
|
|
while (zs->avail_strm > num_strm && !list_empty(&zs->idle_strm)) {
|
|
zstrm = list_entry(zs->idle_strm.next,
|
|
struct zcomp_strm, list);
|
|
list_del(&zstrm->list);
|
|
zcomp_strm_free(comp, zstrm);
|
|
zs->avail_strm--;
|
|
}
|
|
spin_unlock(&zs->strm_lock);
|
|
return true;
|
|
}
|
|
|
|
static void zcomp_strm_multi_destroy(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm_multi *zs = comp->stream;
|
|
struct zcomp_strm *zstrm;
|
|
|
|
while (!list_empty(&zs->idle_strm)) {
|
|
zstrm = list_entry(zs->idle_strm.next,
|
|
struct zcomp_strm, list);
|
|
list_del(&zstrm->list);
|
|
zcomp_strm_free(comp, zstrm);
|
|
}
|
|
kfree(zs);
|
|
}
|
|
|
|
static int zcomp_strm_multi_create(struct zcomp *comp, int max_strm)
|
|
{
|
|
struct zcomp_strm *zstrm;
|
|
struct zcomp_strm_multi *zs;
|
|
|
|
comp->destroy = zcomp_strm_multi_destroy;
|
|
comp->strm_find = zcomp_strm_multi_find;
|
|
comp->strm_release = zcomp_strm_multi_release;
|
|
comp->set_max_streams = zcomp_strm_multi_set_max_streams;
|
|
zs = kmalloc(sizeof(struct zcomp_strm_multi), GFP_KERNEL);
|
|
if (!zs)
|
|
return -ENOMEM;
|
|
|
|
comp->stream = zs;
|
|
spin_lock_init(&zs->strm_lock);
|
|
INIT_LIST_HEAD(&zs->idle_strm);
|
|
init_waitqueue_head(&zs->strm_wait);
|
|
zs->max_strm = max_strm;
|
|
zs->avail_strm = 1;
|
|
|
|
zstrm = zcomp_strm_alloc(comp);
|
|
if (!zstrm) {
|
|
kfree(zs);
|
|
return -ENOMEM;
|
|
}
|
|
list_add(&zstrm->list, &zs->idle_strm);
|
|
return 0;
|
|
}
|
|
|
|
static struct zcomp_strm *zcomp_strm_single_find(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm_single *zs = comp->stream;
|
|
mutex_lock(&zs->strm_lock);
|
|
return zs->zstrm;
|
|
}
|
|
|
|
static void zcomp_strm_single_release(struct zcomp *comp,
|
|
struct zcomp_strm *zstrm)
|
|
{
|
|
struct zcomp_strm_single *zs = comp->stream;
|
|
mutex_unlock(&zs->strm_lock);
|
|
}
|
|
|
|
static bool zcomp_strm_single_set_max_streams(struct zcomp *comp, int num_strm)
|
|
{
|
|
/* zcomp_strm_single support only max_comp_streams == 1 */
|
|
return false;
|
|
}
|
|
|
|
static void zcomp_strm_single_destroy(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm_single *zs = comp->stream;
|
|
zcomp_strm_free(comp, zs->zstrm);
|
|
kfree(zs);
|
|
}
|
|
|
|
static int zcomp_strm_single_create(struct zcomp *comp)
|
|
{
|
|
struct zcomp_strm_single *zs;
|
|
|
|
comp->destroy = zcomp_strm_single_destroy;
|
|
comp->strm_find = zcomp_strm_single_find;
|
|
comp->strm_release = zcomp_strm_single_release;
|
|
comp->set_max_streams = zcomp_strm_single_set_max_streams;
|
|
zs = kmalloc(sizeof(struct zcomp_strm_single), GFP_KERNEL);
|
|
if (!zs)
|
|
return -ENOMEM;
|
|
|
|
comp->stream = zs;
|
|
mutex_init(&zs->strm_lock);
|
|
zs->zstrm = zcomp_strm_alloc(comp);
|
|
if (!zs->zstrm) {
|
|
kfree(zs);
|
|
return -ENOMEM;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/* show available compressors */
|
|
ssize_t zcomp_available_show(const char *comp, char *buf)
|
|
{
|
|
ssize_t sz = 0;
|
|
int i = 0;
|
|
|
|
while (backends[i]) {
|
|
if (!strcmp(comp, backends[i]->name))
|
|
sz += scnprintf(buf + sz, PAGE_SIZE - sz - 2,
|
|
"[%s] ", backends[i]->name);
|
|
else
|
|
sz += scnprintf(buf + sz, PAGE_SIZE - sz - 2,
|
|
"%s ", backends[i]->name);
|
|
i++;
|
|
}
|
|
sz += scnprintf(buf + sz, PAGE_SIZE - sz, "\n");
|
|
return sz;
|
|
}
|
|
|
|
bool zcomp_available_algorithm(const char *comp)
|
|
{
|
|
return find_backend(comp) != NULL;
|
|
}
|
|
|
|
bool zcomp_set_max_streams(struct zcomp *comp, int num_strm)
|
|
{
|
|
return comp->set_max_streams(comp, num_strm);
|
|
}
|
|
|
|
struct zcomp_strm *zcomp_strm_find(struct zcomp *comp)
|
|
{
|
|
return comp->strm_find(comp);
|
|
}
|
|
|
|
void zcomp_strm_release(struct zcomp *comp, struct zcomp_strm *zstrm)
|
|
{
|
|
comp->strm_release(comp, zstrm);
|
|
}
|
|
|
|
int zcomp_compress(struct zcomp *comp, struct zcomp_strm *zstrm,
|
|
const unsigned char *src, size_t *dst_len)
|
|
{
|
|
return comp->backend->compress(src, zstrm->buffer, dst_len,
|
|
zstrm->private);
|
|
}
|
|
|
|
int zcomp_decompress(struct zcomp *comp, const unsigned char *src,
|
|
size_t src_len, unsigned char *dst)
|
|
{
|
|
return comp->backend->decompress(src, src_len, dst);
|
|
}
|
|
|
|
void zcomp_destroy(struct zcomp *comp)
|
|
{
|
|
comp->destroy(comp);
|
|
kfree(comp);
|
|
}
|
|
|
|
/*
|
|
* search available compressors for requested algorithm.
|
|
* allocate new zcomp and initialize it. return compressing
|
|
* backend pointer or ERR_PTR if things went bad. ERR_PTR(-EINVAL)
|
|
* if requested algorithm is not supported, ERR_PTR(-ENOMEM) in
|
|
* case of allocation error, or any other error potentially
|
|
* returned by functions zcomp_strm_{multi,single}_create.
|
|
*/
|
|
struct zcomp *zcomp_create(const char *compress, int max_strm)
|
|
{
|
|
struct zcomp *comp;
|
|
struct zcomp_backend *backend;
|
|
int error;
|
|
|
|
backend = find_backend(compress);
|
|
if (!backend)
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
comp = kzalloc(sizeof(struct zcomp), GFP_KERNEL);
|
|
if (!comp)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
comp->backend = backend;
|
|
if (max_strm > 1)
|
|
error = zcomp_strm_multi_create(comp, max_strm);
|
|
else
|
|
error = zcomp_strm_single_create(comp);
|
|
if (error) {
|
|
kfree(comp);
|
|
return ERR_PTR(error);
|
|
}
|
|
return comp;
|
|
}
|