mirror of
https://github.com/torvalds/linux.git
synced 2026-03-08 01:04:41 +01:00
This was done entirely with mindless brute force, using
git grep -l '\<k[vmz]*alloc_objs*(.*, GFP_KERNEL)' |
xargs sed -i 's/\(alloc_objs*(.*\), GFP_KERNEL)/\1)/'
to convert the new alloc_obj() users that had a simple GFP_KERNEL
argument to just drop that argument.
Note that due to the extreme simplicity of the scripting, any slightly
more complex cases spread over multiple lines would not be triggered:
they definitely exist, but this covers the vast bulk of the cases, and
the resulting diff is also then easier to check automatically.
For the same reason the 'flex' versions will be done as a separate
conversion.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
220 lines
5.8 KiB
C
220 lines
5.8 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
#include <linux/zstd.h>
|
|
#include "compress.h"
|
|
|
|
struct z_erofs_zstd {
|
|
struct z_erofs_zstd *next;
|
|
u8 bounce[PAGE_SIZE];
|
|
void *wksp;
|
|
unsigned int wkspsz;
|
|
};
|
|
|
|
static DEFINE_SPINLOCK(z_erofs_zstd_lock);
|
|
static unsigned int z_erofs_zstd_max_dictsize;
|
|
static unsigned int z_erofs_zstd_nstrms, z_erofs_zstd_avail_strms;
|
|
static struct z_erofs_zstd *z_erofs_zstd_head;
|
|
static DECLARE_WAIT_QUEUE_HEAD(z_erofs_zstd_wq);
|
|
|
|
module_param_named(zstd_streams, z_erofs_zstd_nstrms, uint, 0444);
|
|
|
|
static struct z_erofs_zstd *z_erofs_isolate_strms(bool all)
|
|
{
|
|
struct z_erofs_zstd *strm;
|
|
|
|
again:
|
|
spin_lock(&z_erofs_zstd_lock);
|
|
strm = z_erofs_zstd_head;
|
|
if (!strm) {
|
|
spin_unlock(&z_erofs_zstd_lock);
|
|
wait_event(z_erofs_zstd_wq, READ_ONCE(z_erofs_zstd_head));
|
|
goto again;
|
|
}
|
|
z_erofs_zstd_head = all ? NULL : strm->next;
|
|
spin_unlock(&z_erofs_zstd_lock);
|
|
return strm;
|
|
}
|
|
|
|
static void z_erofs_zstd_exit(void)
|
|
{
|
|
while (z_erofs_zstd_avail_strms) {
|
|
struct z_erofs_zstd *strm, *n;
|
|
|
|
for (strm = z_erofs_isolate_strms(true); strm; strm = n) {
|
|
n = strm->next;
|
|
|
|
kvfree(strm->wksp);
|
|
kfree(strm);
|
|
--z_erofs_zstd_avail_strms;
|
|
}
|
|
}
|
|
}
|
|
|
|
static int __init z_erofs_zstd_init(void)
|
|
{
|
|
/* by default, use # of possible CPUs instead */
|
|
if (!z_erofs_zstd_nstrms)
|
|
z_erofs_zstd_nstrms = num_possible_cpus();
|
|
|
|
for (; z_erofs_zstd_avail_strms < z_erofs_zstd_nstrms;
|
|
++z_erofs_zstd_avail_strms) {
|
|
struct z_erofs_zstd *strm;
|
|
|
|
strm = kzalloc_obj(*strm);
|
|
if (!strm) {
|
|
z_erofs_zstd_exit();
|
|
return -ENOMEM;
|
|
}
|
|
spin_lock(&z_erofs_zstd_lock);
|
|
strm->next = z_erofs_zstd_head;
|
|
z_erofs_zstd_head = strm;
|
|
spin_unlock(&z_erofs_zstd_lock);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int z_erofs_load_zstd_config(struct super_block *sb,
|
|
struct erofs_super_block *dsb, void *data, int size)
|
|
{
|
|
static DEFINE_MUTEX(zstd_resize_mutex);
|
|
struct z_erofs_zstd_cfgs *zstd = data;
|
|
unsigned int dict_size, wkspsz;
|
|
struct z_erofs_zstd *strm, *head = NULL;
|
|
void *wksp;
|
|
|
|
if (!zstd || size < sizeof(struct z_erofs_zstd_cfgs) || zstd->format) {
|
|
erofs_err(sb, "unsupported zstd format, size=%u", size);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (zstd->windowlog > ilog2(Z_EROFS_ZSTD_MAX_DICT_SIZE) - 10) {
|
|
erofs_err(sb, "unsupported zstd window log %u", zstd->windowlog);
|
|
return -EINVAL;
|
|
}
|
|
dict_size = 1U << (zstd->windowlog + 10);
|
|
|
|
/* in case 2 z_erofs_load_zstd_config() race to avoid deadlock */
|
|
mutex_lock(&zstd_resize_mutex);
|
|
if (z_erofs_zstd_max_dictsize >= dict_size) {
|
|
mutex_unlock(&zstd_resize_mutex);
|
|
return 0;
|
|
}
|
|
|
|
/* 1. collect/isolate all streams for the following check */
|
|
while (z_erofs_zstd_avail_strms) {
|
|
struct z_erofs_zstd *n;
|
|
|
|
for (strm = z_erofs_isolate_strms(true); strm; strm = n) {
|
|
n = strm->next;
|
|
strm->next = head;
|
|
head = strm;
|
|
--z_erofs_zstd_avail_strms;
|
|
}
|
|
}
|
|
|
|
/* 2. walk each isolated stream and grow max dict_size if needed */
|
|
wkspsz = zstd_dstream_workspace_bound(dict_size);
|
|
for (strm = head; strm; strm = strm->next) {
|
|
wksp = kvmalloc(wkspsz, GFP_KERNEL);
|
|
if (!wksp)
|
|
break;
|
|
kvfree(strm->wksp);
|
|
strm->wksp = wksp;
|
|
strm->wkspsz = wkspsz;
|
|
}
|
|
|
|
/* 3. push back all to the global list and update max dict_size */
|
|
spin_lock(&z_erofs_zstd_lock);
|
|
DBG_BUGON(z_erofs_zstd_head);
|
|
z_erofs_zstd_head = head;
|
|
spin_unlock(&z_erofs_zstd_lock);
|
|
z_erofs_zstd_avail_strms = z_erofs_zstd_nstrms;
|
|
wake_up_all(&z_erofs_zstd_wq);
|
|
if (!strm)
|
|
z_erofs_zstd_max_dictsize = dict_size;
|
|
mutex_unlock(&zstd_resize_mutex);
|
|
return strm ? -ENOMEM : 0;
|
|
}
|
|
|
|
static const char *z_erofs_zstd_decompress(struct z_erofs_decompress_req *rq,
|
|
struct page **pgpl)
|
|
{
|
|
struct super_block *sb = rq->sb;
|
|
struct z_erofs_stream_dctx dctx = { .rq = rq, .no = -1, .ni = 0 };
|
|
zstd_in_buffer in_buf = { NULL, 0, 0 };
|
|
zstd_out_buffer out_buf = { NULL, 0, 0 };
|
|
struct z_erofs_zstd *strm;
|
|
zstd_dstream *stream;
|
|
const char *reason;
|
|
int zerr;
|
|
|
|
/* 1. get the exact compressed size */
|
|
dctx.kin = kmap_local_page(*rq->in);
|
|
reason = z_erofs_fixup_insize(rq, dctx.kin + rq->pageofs_in,
|
|
min(rq->inputsize, sb->s_blocksize - rq->pageofs_in));
|
|
if (reason) {
|
|
kunmap_local(dctx.kin);
|
|
return reason;
|
|
}
|
|
|
|
/* 2. get an available ZSTD context */
|
|
strm = z_erofs_isolate_strms(false);
|
|
|
|
/* 3. multi-call decompress */
|
|
stream = zstd_init_dstream(z_erofs_zstd_max_dictsize, strm->wksp, strm->wkspsz);
|
|
if (!stream) {
|
|
reason = ERR_PTR(-ENOMEM);
|
|
goto failed_zinit;
|
|
}
|
|
|
|
rq->fillgaps = true; /* ZSTD doesn't support NULL output buffer */
|
|
in_buf.size = min_t(u32, rq->inputsize, PAGE_SIZE - rq->pageofs_in);
|
|
rq->inputsize -= in_buf.size;
|
|
in_buf.src = dctx.kin + rq->pageofs_in;
|
|
dctx.bounce = strm->bounce;
|
|
|
|
do {
|
|
dctx.inbuf_sz = in_buf.size;
|
|
dctx.inbuf_pos = in_buf.pos;
|
|
reason = z_erofs_stream_switch_bufs(&dctx, &out_buf.dst,
|
|
(void **)&in_buf.src, pgpl);
|
|
if (reason)
|
|
break;
|
|
|
|
if (out_buf.size == out_buf.pos) {
|
|
out_buf.size = dctx.avail_out;
|
|
out_buf.pos = 0;
|
|
}
|
|
in_buf.size = dctx.inbuf_sz;
|
|
in_buf.pos = dctx.inbuf_pos;
|
|
|
|
zerr = zstd_decompress_stream(stream, &out_buf, &in_buf);
|
|
dctx.avail_out = out_buf.size - out_buf.pos;
|
|
if (zstd_is_error(zerr) ||
|
|
((rq->outputsize + dctx.avail_out) && (!zerr || (zerr > 0 &&
|
|
!(rq->inputsize + in_buf.size - in_buf.pos))))) {
|
|
reason = zstd_is_error(zerr) ? zstd_get_error_name(zerr) :
|
|
"unexpected end of stream";
|
|
break;
|
|
}
|
|
} while (rq->outputsize + dctx.avail_out);
|
|
|
|
if (dctx.kout)
|
|
kunmap_local(dctx.kout);
|
|
failed_zinit:
|
|
kunmap_local(dctx.kin);
|
|
/* 4. push back ZSTD stream context to the global list */
|
|
spin_lock(&z_erofs_zstd_lock);
|
|
strm->next = z_erofs_zstd_head;
|
|
z_erofs_zstd_head = strm;
|
|
spin_unlock(&z_erofs_zstd_lock);
|
|
wake_up(&z_erofs_zstd_wq);
|
|
return reason;
|
|
}
|
|
|
|
const struct z_erofs_decompressor z_erofs_zstd_decomp = {
|
|
.config = z_erofs_load_zstd_config,
|
|
.decompress = z_erofs_zstd_decompress,
|
|
.init = z_erofs_zstd_init,
|
|
.exit = z_erofs_zstd_exit,
|
|
.name = "zstd",
|
|
};
|