mirror of
https://github.com/torvalds/linux.git
synced 2026-03-08 01:24:47 +01:00
This was done entirely with mindless brute force, using
git grep -l '\<k[vmz]*alloc_objs*(.*, GFP_KERNEL)' |
xargs sed -i 's/\(alloc_objs*(.*\), GFP_KERNEL)/\1)/'
to convert the new alloc_obj() users that had a simple GFP_KERNEL
argument to just drop that argument.
Note that due to the extreme simplicity of the scripting, any slightly
more complex cases spread over multiple lines would not be triggered:
they definitely exist, but this covers the vast bulk of the cases, and
the resulting diff is also then easier to check automatically.
For the same reason the 'flex' versions will be done as a separate
conversion.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
770 lines
20 KiB
C
770 lines
20 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2015 Google, Inc.
|
|
*
|
|
* Author: Sami Tolvanen <samitolvanen@google.com>
|
|
*/
|
|
|
|
#include "dm-verity-fec.h"
|
|
#include <linux/math64.h>
|
|
|
|
#define DM_MSG_PREFIX "verity-fec"
|
|
|
|
/*
|
|
* When correcting a data block, the FEC code performs optimally when it can
|
|
* collect all the associated RS blocks at the same time. As each byte is part
|
|
* of a different RS block, there are '1 << data_dev_block_bits' RS blocks.
|
|
* There are '1 << DM_VERITY_FEC_BUF_RS_BITS' RS blocks per buffer, so that
|
|
* gives '1 << (data_dev_block_bits - DM_VERITY_FEC_BUF_RS_BITS)' buffers.
|
|
*/
|
|
static inline unsigned int fec_max_nbufs(struct dm_verity *v)
|
|
{
|
|
return 1 << (v->data_dev_block_bits - DM_VERITY_FEC_BUF_RS_BITS);
|
|
}
|
|
|
|
/*
|
|
* Return an interleaved offset for a byte in RS block.
|
|
*/
|
|
static inline u64 fec_interleave(struct dm_verity *v, u64 offset)
|
|
{
|
|
u32 mod;
|
|
|
|
mod = do_div(offset, v->fec->rsn);
|
|
return offset + mod * (v->fec->rounds << v->data_dev_block_bits);
|
|
}
|
|
|
|
/*
|
|
* Read error-correcting codes for the requested RS block. Returns a pointer
|
|
* to the data block. Caller is responsible for releasing buf.
|
|
*/
|
|
static u8 *fec_read_parity(struct dm_verity *v, u64 rsb, int index,
|
|
unsigned int *offset, unsigned int par_buf_offset,
|
|
struct dm_buffer **buf, unsigned short ioprio)
|
|
{
|
|
u64 position, block, rem;
|
|
u8 *res;
|
|
|
|
/* We have already part of parity bytes read, skip to the next block */
|
|
if (par_buf_offset)
|
|
index++;
|
|
|
|
position = (index + rsb) * v->fec->roots;
|
|
block = div64_u64_rem(position, v->fec->io_size, &rem);
|
|
*offset = par_buf_offset ? 0 : (unsigned int)rem;
|
|
|
|
res = dm_bufio_read_with_ioprio(v->fec->bufio, block, buf, ioprio);
|
|
if (IS_ERR(res)) {
|
|
DMERR("%s: FEC %llu: parity read failed (block %llu): %ld",
|
|
v->data_dev->name, (unsigned long long)rsb,
|
|
(unsigned long long)block, PTR_ERR(res));
|
|
*buf = NULL;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
/* Loop over each allocated buffer. */
|
|
#define fec_for_each_buffer(io, __i) \
|
|
for (__i = 0; __i < (io)->nbufs; __i++)
|
|
|
|
/* Loop over each RS block in each allocated buffer. */
|
|
#define fec_for_each_buffer_rs_block(io, __i, __j) \
|
|
fec_for_each_buffer(io, __i) \
|
|
for (__j = 0; __j < 1 << DM_VERITY_FEC_BUF_RS_BITS; __j++)
|
|
|
|
/*
|
|
* Return a pointer to the current RS block when called inside
|
|
* fec_for_each_buffer_rs_block.
|
|
*/
|
|
static inline u8 *fec_buffer_rs_block(struct dm_verity *v,
|
|
struct dm_verity_fec_io *fio,
|
|
unsigned int i, unsigned int j)
|
|
{
|
|
return &fio->bufs[i][j * v->fec->rsn];
|
|
}
|
|
|
|
/*
|
|
* Return an index to the current RS block when called inside
|
|
* fec_for_each_buffer_rs_block.
|
|
*/
|
|
static inline unsigned int fec_buffer_rs_index(unsigned int i, unsigned int j)
|
|
{
|
|
return (i << DM_VERITY_FEC_BUF_RS_BITS) + j;
|
|
}
|
|
|
|
/*
|
|
* Decode all RS blocks from buffers and copy corrected bytes into fio->output
|
|
* starting from block_offset.
|
|
*/
|
|
static int fec_decode_bufs(struct dm_verity *v, struct dm_verity_io *io,
|
|
struct dm_verity_fec_io *fio, u64 rsb, int byte_index,
|
|
unsigned int block_offset, int neras)
|
|
{
|
|
int r, corrected = 0, res;
|
|
struct dm_buffer *buf;
|
|
unsigned int n, i, j, offset, par_buf_offset = 0;
|
|
uint16_t par_buf[DM_VERITY_FEC_RSM - DM_VERITY_FEC_MIN_RSN];
|
|
u8 *par, *block;
|
|
struct bio *bio = dm_bio_from_per_bio_data(io, v->ti->per_io_data_size);
|
|
|
|
par = fec_read_parity(v, rsb, block_offset, &offset,
|
|
par_buf_offset, &buf, bio->bi_ioprio);
|
|
if (IS_ERR(par))
|
|
return PTR_ERR(par);
|
|
|
|
/*
|
|
* Decode the RS blocks we have in bufs. Each RS block results in
|
|
* one corrected target byte and consumes fec->roots parity bytes.
|
|
*/
|
|
fec_for_each_buffer_rs_block(fio, n, i) {
|
|
block = fec_buffer_rs_block(v, fio, n, i);
|
|
for (j = 0; j < v->fec->roots - par_buf_offset; j++)
|
|
par_buf[par_buf_offset + j] = par[offset + j];
|
|
/* Decode an RS block using Reed-Solomon */
|
|
res = decode_rs8(fio->rs, block, par_buf, v->fec->rsn,
|
|
NULL, neras, fio->erasures, 0, NULL);
|
|
if (res < 0) {
|
|
r = res;
|
|
goto error;
|
|
}
|
|
|
|
corrected += res;
|
|
fio->output[block_offset] = block[byte_index];
|
|
|
|
block_offset++;
|
|
if (block_offset >= 1 << v->data_dev_block_bits)
|
|
goto done;
|
|
|
|
/* Read the next block when we run out of parity bytes */
|
|
offset += (v->fec->roots - par_buf_offset);
|
|
/* Check if parity bytes are split between blocks */
|
|
if (offset < v->fec->io_size && (offset + v->fec->roots) > v->fec->io_size) {
|
|
par_buf_offset = v->fec->io_size - offset;
|
|
for (j = 0; j < par_buf_offset; j++)
|
|
par_buf[j] = par[offset + j];
|
|
offset += par_buf_offset;
|
|
} else
|
|
par_buf_offset = 0;
|
|
|
|
if (offset >= v->fec->io_size) {
|
|
dm_bufio_release(buf);
|
|
|
|
par = fec_read_parity(v, rsb, block_offset, &offset,
|
|
par_buf_offset, &buf, bio->bi_ioprio);
|
|
if (IS_ERR(par))
|
|
return PTR_ERR(par);
|
|
}
|
|
}
|
|
done:
|
|
r = corrected;
|
|
error:
|
|
dm_bufio_release(buf);
|
|
|
|
if (r < 0 && neras)
|
|
DMERR_LIMIT("%s: FEC %llu: failed to correct: %d",
|
|
v->data_dev->name, (unsigned long long)rsb, r);
|
|
else if (r > 0) {
|
|
DMWARN_LIMIT("%s: FEC %llu: corrected %d errors",
|
|
v->data_dev->name, (unsigned long long)rsb, r);
|
|
atomic64_inc(&v->fec->corrected);
|
|
}
|
|
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Locate data block erasures using verity hashes.
|
|
*/
|
|
static int fec_is_erasure(struct dm_verity *v, struct dm_verity_io *io,
|
|
const u8 *want_digest, const u8 *data)
|
|
{
|
|
if (unlikely(verity_hash(v, io, data, 1 << v->data_dev_block_bits,
|
|
io->tmp_digest)))
|
|
return 0;
|
|
|
|
return memcmp(io->tmp_digest, want_digest, v->digest_size) != 0;
|
|
}
|
|
|
|
/*
|
|
* Read data blocks that are part of the RS block and deinterleave as much as
|
|
* fits into buffers. Check for erasure locations if @neras is non-NULL.
|
|
*/
|
|
static int fec_read_bufs(struct dm_verity *v, struct dm_verity_io *io,
|
|
u64 rsb, u64 target, unsigned int block_offset,
|
|
int *neras)
|
|
{
|
|
bool is_zero;
|
|
int i, j, target_index = -1;
|
|
struct dm_buffer *buf;
|
|
struct dm_bufio_client *bufio;
|
|
struct dm_verity_fec_io *fio = io->fec_io;
|
|
u64 block, ileaved;
|
|
u8 *bbuf, *rs_block;
|
|
u8 want_digest[HASH_MAX_DIGESTSIZE];
|
|
unsigned int n, k;
|
|
struct bio *bio = dm_bio_from_per_bio_data(io, v->ti->per_io_data_size);
|
|
|
|
if (neras)
|
|
*neras = 0;
|
|
|
|
if (WARN_ON(v->digest_size > sizeof(want_digest)))
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* read each of the rsn data blocks that are part of the RS block, and
|
|
* interleave contents to available bufs
|
|
*/
|
|
for (i = 0; i < v->fec->rsn; i++) {
|
|
ileaved = fec_interleave(v, rsb * v->fec->rsn + i);
|
|
|
|
/*
|
|
* target is the data block we want to correct, target_index is
|
|
* the index of this block within the rsn RS blocks
|
|
*/
|
|
if (ileaved == target)
|
|
target_index = i;
|
|
|
|
block = ileaved >> v->data_dev_block_bits;
|
|
bufio = v->fec->data_bufio;
|
|
|
|
if (block >= v->data_blocks) {
|
|
block -= v->data_blocks;
|
|
|
|
/*
|
|
* blocks outside the area were assumed to contain
|
|
* zeros when encoding data was generated
|
|
*/
|
|
if (unlikely(block >= v->fec->hash_blocks))
|
|
continue;
|
|
|
|
block += v->hash_start;
|
|
bufio = v->bufio;
|
|
}
|
|
|
|
bbuf = dm_bufio_read_with_ioprio(bufio, block, &buf, bio->bi_ioprio);
|
|
if (IS_ERR(bbuf)) {
|
|
DMWARN_LIMIT("%s: FEC %llu: read failed (%llu): %ld",
|
|
v->data_dev->name,
|
|
(unsigned long long)rsb,
|
|
(unsigned long long)block, PTR_ERR(bbuf));
|
|
|
|
/* assume the block is corrupted */
|
|
if (neras && *neras <= v->fec->roots)
|
|
fio->erasures[(*neras)++] = i;
|
|
|
|
continue;
|
|
}
|
|
|
|
/* locate erasures if the block is on the data device */
|
|
if (bufio == v->fec->data_bufio &&
|
|
verity_hash_for_block(v, io, block, want_digest,
|
|
&is_zero) == 0) {
|
|
/* skip known zero blocks entirely */
|
|
if (is_zero)
|
|
goto done;
|
|
|
|
/*
|
|
* skip if we have already found the theoretical
|
|
* maximum number (i.e. fec->roots) of erasures
|
|
*/
|
|
if (neras && *neras <= v->fec->roots &&
|
|
fec_is_erasure(v, io, want_digest, bbuf))
|
|
fio->erasures[(*neras)++] = i;
|
|
}
|
|
|
|
/*
|
|
* deinterleave and copy the bytes that fit into bufs,
|
|
* starting from block_offset
|
|
*/
|
|
fec_for_each_buffer_rs_block(fio, n, j) {
|
|
k = fec_buffer_rs_index(n, j) + block_offset;
|
|
|
|
if (k >= 1 << v->data_dev_block_bits)
|
|
goto done;
|
|
|
|
rs_block = fec_buffer_rs_block(v, fio, n, j);
|
|
rs_block[i] = bbuf[k];
|
|
}
|
|
done:
|
|
dm_bufio_release(buf);
|
|
}
|
|
|
|
return target_index;
|
|
}
|
|
|
|
/*
|
|
* Allocate and initialize a struct dm_verity_fec_io to use for FEC for a bio.
|
|
* This runs the first time a block needs to be corrected for a bio. In the
|
|
* common case where no block needs to be corrected, this code never runs.
|
|
*
|
|
* This always succeeds, as all required allocations are done from mempools.
|
|
* Additional buffers are also allocated opportunistically to improve error
|
|
* correction performance, but these aren't required to succeed.
|
|
*/
|
|
static struct dm_verity_fec_io *fec_alloc_and_init_io(struct dm_verity *v)
|
|
{
|
|
const unsigned int max_nbufs = fec_max_nbufs(v);
|
|
struct dm_verity_fec *f = v->fec;
|
|
struct dm_verity_fec_io *fio;
|
|
unsigned int n;
|
|
|
|
fio = mempool_alloc(&f->fio_pool, GFP_NOIO);
|
|
fio->rs = mempool_alloc(&f->rs_pool, GFP_NOIO);
|
|
|
|
fio->bufs[0] = mempool_alloc(&f->prealloc_pool, GFP_NOIO);
|
|
|
|
/* try to allocate the maximum number of buffers */
|
|
for (n = 1; n < max_nbufs; n++) {
|
|
fio->bufs[n] = kmem_cache_alloc(f->cache, GFP_NOWAIT);
|
|
/* we can manage with even one buffer if necessary */
|
|
if (unlikely(!fio->bufs[n]))
|
|
break;
|
|
}
|
|
fio->nbufs = n;
|
|
|
|
fio->output = mempool_alloc(&f->output_pool, GFP_NOIO);
|
|
fio->level = 0;
|
|
return fio;
|
|
}
|
|
|
|
/*
|
|
* Initialize buffers and clear erasures. fec_read_bufs() assumes buffers are
|
|
* zeroed before deinterleaving.
|
|
*/
|
|
static void fec_init_bufs(struct dm_verity *v, struct dm_verity_fec_io *fio)
|
|
{
|
|
unsigned int n;
|
|
|
|
fec_for_each_buffer(fio, n)
|
|
memset(fio->bufs[n], 0, v->fec->rsn << DM_VERITY_FEC_BUF_RS_BITS);
|
|
|
|
memset(fio->erasures, 0, sizeof(fio->erasures));
|
|
}
|
|
|
|
/*
|
|
* Decode all RS blocks in a single data block and return the target block
|
|
* (indicated by @offset) in fio->output. If @use_erasures is non-zero, uses
|
|
* hashes to locate erasures.
|
|
*/
|
|
static int fec_decode_rsb(struct dm_verity *v, struct dm_verity_io *io,
|
|
struct dm_verity_fec_io *fio, u64 rsb, u64 offset,
|
|
const u8 *want_digest, bool use_erasures)
|
|
{
|
|
int r, neras = 0;
|
|
unsigned int pos;
|
|
|
|
for (pos = 0; pos < 1 << v->data_dev_block_bits; ) {
|
|
fec_init_bufs(v, fio);
|
|
|
|
r = fec_read_bufs(v, io, rsb, offset, pos,
|
|
use_erasures ? &neras : NULL);
|
|
if (unlikely(r < 0))
|
|
return r;
|
|
|
|
r = fec_decode_bufs(v, io, fio, rsb, r, pos, neras);
|
|
if (r < 0)
|
|
return r;
|
|
|
|
pos += fio->nbufs << DM_VERITY_FEC_BUF_RS_BITS;
|
|
}
|
|
|
|
/* Always re-validate the corrected block against the expected hash */
|
|
r = verity_hash(v, io, fio->output, 1 << v->data_dev_block_bits,
|
|
io->tmp_digest);
|
|
if (unlikely(r < 0))
|
|
return r;
|
|
|
|
if (memcmp(io->tmp_digest, want_digest, v->digest_size)) {
|
|
DMERR_LIMIT("%s: FEC %llu: failed to correct (%d erasures)",
|
|
v->data_dev->name, (unsigned long long)rsb, neras);
|
|
return -EILSEQ;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Correct errors in a block. Copies corrected block to dest. */
|
|
int verity_fec_decode(struct dm_verity *v, struct dm_verity_io *io,
|
|
enum verity_block_type type, const u8 *want_digest,
|
|
sector_t block, u8 *dest)
|
|
{
|
|
int r;
|
|
struct dm_verity_fec_io *fio;
|
|
u64 offset, res, rsb;
|
|
|
|
if (!verity_fec_is_enabled(v))
|
|
return -EOPNOTSUPP;
|
|
|
|
fio = io->fec_io;
|
|
if (!fio)
|
|
fio = io->fec_io = fec_alloc_and_init_io(v);
|
|
|
|
if (fio->level)
|
|
return -EIO;
|
|
|
|
fio->level++;
|
|
|
|
if (type == DM_VERITY_BLOCK_TYPE_METADATA)
|
|
block = block - v->hash_start + v->data_blocks;
|
|
|
|
/*
|
|
* For RS(M, N), the continuous FEC data is divided into blocks of N
|
|
* bytes. Since block size may not be divisible by N, the last block
|
|
* is zero padded when decoding.
|
|
*
|
|
* Each byte of the block is covered by a different RS(M, N) code,
|
|
* and each code is interleaved over N blocks to make it less likely
|
|
* that bursty corruption will leave us in unrecoverable state.
|
|
*/
|
|
|
|
offset = block << v->data_dev_block_bits;
|
|
res = div64_u64(offset, v->fec->rounds << v->data_dev_block_bits);
|
|
|
|
/*
|
|
* The base RS block we can feed to the interleaver to find out all
|
|
* blocks required for decoding.
|
|
*/
|
|
rsb = offset - res * (v->fec->rounds << v->data_dev_block_bits);
|
|
|
|
/*
|
|
* Locating erasures is slow, so attempt to recover the block without
|
|
* them first. Do a second attempt with erasures if the corruption is
|
|
* bad enough.
|
|
*/
|
|
r = fec_decode_rsb(v, io, fio, rsb, offset, want_digest, false);
|
|
if (r < 0) {
|
|
r = fec_decode_rsb(v, io, fio, rsb, offset, want_digest, true);
|
|
if (r < 0)
|
|
goto done;
|
|
}
|
|
|
|
memcpy(dest, fio->output, 1 << v->data_dev_block_bits);
|
|
|
|
done:
|
|
fio->level--;
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* Clean up per-bio data.
|
|
*/
|
|
void __verity_fec_finish_io(struct dm_verity_io *io)
|
|
{
|
|
unsigned int n;
|
|
struct dm_verity_fec *f = io->v->fec;
|
|
struct dm_verity_fec_io *fio = io->fec_io;
|
|
|
|
mempool_free(fio->rs, &f->rs_pool);
|
|
|
|
mempool_free(fio->bufs[0], &f->prealloc_pool);
|
|
|
|
for (n = 1; n < fio->nbufs; n++)
|
|
kmem_cache_free(f->cache, fio->bufs[n]);
|
|
|
|
mempool_free(fio->output, &f->output_pool);
|
|
|
|
mempool_free(fio, &f->fio_pool);
|
|
io->fec_io = NULL;
|
|
}
|
|
|
|
/*
|
|
* Append feature arguments and values to the status table.
|
|
*/
|
|
unsigned int verity_fec_status_table(struct dm_verity *v, unsigned int sz,
|
|
char *result, unsigned int maxlen)
|
|
{
|
|
if (!verity_fec_is_enabled(v))
|
|
return sz;
|
|
|
|
DMEMIT(" " DM_VERITY_OPT_FEC_DEV " %s "
|
|
DM_VERITY_OPT_FEC_BLOCKS " %llu "
|
|
DM_VERITY_OPT_FEC_START " %llu "
|
|
DM_VERITY_OPT_FEC_ROOTS " %d",
|
|
v->fec->dev->name,
|
|
(unsigned long long)v->fec->blocks,
|
|
(unsigned long long)v->fec->start,
|
|
v->fec->roots);
|
|
|
|
return sz;
|
|
}
|
|
|
|
void verity_fec_dtr(struct dm_verity *v)
|
|
{
|
|
struct dm_verity_fec *f = v->fec;
|
|
|
|
if (!verity_fec_is_enabled(v))
|
|
goto out;
|
|
|
|
mempool_exit(&f->fio_pool);
|
|
mempool_exit(&f->rs_pool);
|
|
mempool_exit(&f->prealloc_pool);
|
|
mempool_exit(&f->output_pool);
|
|
kmem_cache_destroy(f->cache);
|
|
|
|
if (!IS_ERR_OR_NULL(f->data_bufio))
|
|
dm_bufio_client_destroy(f->data_bufio);
|
|
if (!IS_ERR_OR_NULL(f->bufio))
|
|
dm_bufio_client_destroy(f->bufio);
|
|
|
|
if (f->dev)
|
|
dm_put_device(v->ti, f->dev);
|
|
out:
|
|
kfree(f);
|
|
v->fec = NULL;
|
|
}
|
|
|
|
static void *fec_rs_alloc(gfp_t gfp_mask, void *pool_data)
|
|
{
|
|
struct dm_verity *v = pool_data;
|
|
|
|
return init_rs_gfp(8, 0x11d, 0, 1, v->fec->roots, gfp_mask);
|
|
}
|
|
|
|
static void fec_rs_free(void *element, void *pool_data)
|
|
{
|
|
struct rs_control *rs = element;
|
|
|
|
if (rs)
|
|
free_rs(rs);
|
|
}
|
|
|
|
bool verity_is_fec_opt_arg(const char *arg_name)
|
|
{
|
|
return (!strcasecmp(arg_name, DM_VERITY_OPT_FEC_DEV) ||
|
|
!strcasecmp(arg_name, DM_VERITY_OPT_FEC_BLOCKS) ||
|
|
!strcasecmp(arg_name, DM_VERITY_OPT_FEC_START) ||
|
|
!strcasecmp(arg_name, DM_VERITY_OPT_FEC_ROOTS));
|
|
}
|
|
|
|
int verity_fec_parse_opt_args(struct dm_arg_set *as, struct dm_verity *v,
|
|
unsigned int *argc, const char *arg_name)
|
|
{
|
|
int r;
|
|
struct dm_target *ti = v->ti;
|
|
const char *arg_value;
|
|
unsigned long long num_ll;
|
|
unsigned char num_c;
|
|
char dummy;
|
|
|
|
if (!*argc) {
|
|
ti->error = "FEC feature arguments require a value";
|
|
return -EINVAL;
|
|
}
|
|
|
|
arg_value = dm_shift_arg(as);
|
|
(*argc)--;
|
|
|
|
if (!strcasecmp(arg_name, DM_VERITY_OPT_FEC_DEV)) {
|
|
if (v->fec->dev) {
|
|
ti->error = "FEC device already specified";
|
|
return -EINVAL;
|
|
}
|
|
r = dm_get_device(ti, arg_value, BLK_OPEN_READ, &v->fec->dev);
|
|
if (r) {
|
|
ti->error = "FEC device lookup failed";
|
|
return r;
|
|
}
|
|
|
|
} else if (!strcasecmp(arg_name, DM_VERITY_OPT_FEC_BLOCKS)) {
|
|
if (sscanf(arg_value, "%llu%c", &num_ll, &dummy) != 1 ||
|
|
((sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT))
|
|
>> (v->data_dev_block_bits - SECTOR_SHIFT) != num_ll)) {
|
|
ti->error = "Invalid " DM_VERITY_OPT_FEC_BLOCKS;
|
|
return -EINVAL;
|
|
}
|
|
v->fec->blocks = num_ll;
|
|
|
|
} else if (!strcasecmp(arg_name, DM_VERITY_OPT_FEC_START)) {
|
|
if (sscanf(arg_value, "%llu%c", &num_ll, &dummy) != 1 ||
|
|
((sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT)) >>
|
|
(v->data_dev_block_bits - SECTOR_SHIFT) != num_ll)) {
|
|
ti->error = "Invalid " DM_VERITY_OPT_FEC_START;
|
|
return -EINVAL;
|
|
}
|
|
v->fec->start = num_ll;
|
|
|
|
} else if (!strcasecmp(arg_name, DM_VERITY_OPT_FEC_ROOTS)) {
|
|
if (sscanf(arg_value, "%hhu%c", &num_c, &dummy) != 1 || !num_c ||
|
|
num_c < (DM_VERITY_FEC_RSM - DM_VERITY_FEC_MAX_RSN) ||
|
|
num_c > (DM_VERITY_FEC_RSM - DM_VERITY_FEC_MIN_RSN)) {
|
|
ti->error = "Invalid " DM_VERITY_OPT_FEC_ROOTS;
|
|
return -EINVAL;
|
|
}
|
|
v->fec->roots = num_c;
|
|
|
|
} else {
|
|
ti->error = "Unrecognized verity FEC feature request";
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Allocate dm_verity_fec for v->fec. Must be called before verity_fec_ctr.
|
|
*/
|
|
int verity_fec_ctr_alloc(struct dm_verity *v)
|
|
{
|
|
struct dm_verity_fec *f;
|
|
|
|
f = kzalloc_obj(struct dm_verity_fec);
|
|
if (!f) {
|
|
v->ti->error = "Cannot allocate FEC structure";
|
|
return -ENOMEM;
|
|
}
|
|
v->fec = f;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Validate arguments and preallocate memory. Must be called after arguments
|
|
* have been parsed using verity_fec_parse_opt_args.
|
|
*/
|
|
int verity_fec_ctr(struct dm_verity *v)
|
|
{
|
|
struct dm_verity_fec *f = v->fec;
|
|
struct dm_target *ti = v->ti;
|
|
u64 hash_blocks, fec_blocks;
|
|
int ret;
|
|
|
|
if (!verity_fec_is_enabled(v)) {
|
|
verity_fec_dtr(v);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* FEC is computed over data blocks, possible metadata, and
|
|
* hash blocks. In other words, FEC covers total of fec_blocks
|
|
* blocks consisting of the following:
|
|
*
|
|
* data blocks | hash blocks | metadata (optional)
|
|
*
|
|
* We allow metadata after hash blocks to support a use case
|
|
* where all data is stored on the same device and FEC covers
|
|
* the entire area.
|
|
*
|
|
* If metadata is included, we require it to be available on the
|
|
* hash device after the hash blocks.
|
|
*/
|
|
|
|
hash_blocks = v->hash_blocks - v->hash_start;
|
|
|
|
/*
|
|
* Require matching block sizes for data and hash devices for
|
|
* simplicity.
|
|
*/
|
|
if (v->data_dev_block_bits != v->hash_dev_block_bits) {
|
|
ti->error = "Block sizes must match to use FEC";
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!f->roots) {
|
|
ti->error = "Missing " DM_VERITY_OPT_FEC_ROOTS;
|
|
return -EINVAL;
|
|
}
|
|
f->rsn = DM_VERITY_FEC_RSM - f->roots;
|
|
|
|
if (!f->blocks) {
|
|
ti->error = "Missing " DM_VERITY_OPT_FEC_BLOCKS;
|
|
return -EINVAL;
|
|
}
|
|
|
|
f->rounds = f->blocks;
|
|
if (sector_div(f->rounds, f->rsn))
|
|
f->rounds++;
|
|
|
|
/*
|
|
* Due to optional metadata, f->blocks can be larger than
|
|
* data_blocks and hash_blocks combined.
|
|
*/
|
|
if (f->blocks < v->data_blocks + hash_blocks || !f->rounds) {
|
|
ti->error = "Invalid " DM_VERITY_OPT_FEC_BLOCKS;
|
|
return -EINVAL;
|
|
}
|
|
|
|
/*
|
|
* Metadata is accessed through the hash device, so we require
|
|
* it to be large enough.
|
|
*/
|
|
f->hash_blocks = f->blocks - v->data_blocks;
|
|
if (dm_bufio_get_device_size(v->bufio) < f->hash_blocks) {
|
|
ti->error = "Hash device is too small for "
|
|
DM_VERITY_OPT_FEC_BLOCKS;
|
|
return -E2BIG;
|
|
}
|
|
|
|
f->io_size = 1 << v->data_dev_block_bits;
|
|
|
|
f->bufio = dm_bufio_client_create(f->dev->bdev,
|
|
f->io_size,
|
|
1, 0, NULL, NULL, 0);
|
|
if (IS_ERR(f->bufio)) {
|
|
ti->error = "Cannot initialize FEC bufio client";
|
|
return PTR_ERR(f->bufio);
|
|
}
|
|
|
|
dm_bufio_set_sector_offset(f->bufio, f->start << (v->data_dev_block_bits - SECTOR_SHIFT));
|
|
|
|
fec_blocks = div64_u64(f->rounds * f->roots, v->fec->roots << SECTOR_SHIFT);
|
|
if (dm_bufio_get_device_size(f->bufio) < fec_blocks) {
|
|
ti->error = "FEC device is too small";
|
|
return -E2BIG;
|
|
}
|
|
|
|
f->data_bufio = dm_bufio_client_create(v->data_dev->bdev,
|
|
1 << v->data_dev_block_bits,
|
|
1, 0, NULL, NULL, 0);
|
|
if (IS_ERR(f->data_bufio)) {
|
|
ti->error = "Cannot initialize FEC data bufio client";
|
|
return PTR_ERR(f->data_bufio);
|
|
}
|
|
|
|
if (dm_bufio_get_device_size(f->data_bufio) < v->data_blocks) {
|
|
ti->error = "Data device is too small";
|
|
return -E2BIG;
|
|
}
|
|
|
|
/* Preallocate some dm_verity_fec_io structures */
|
|
ret = mempool_init_kmalloc_pool(&f->fio_pool, num_online_cpus(),
|
|
struct_size((struct dm_verity_fec_io *)0,
|
|
bufs, fec_max_nbufs(v)));
|
|
if (ret) {
|
|
ti->error = "Cannot allocate FEC IO pool";
|
|
return ret;
|
|
}
|
|
|
|
/* Preallocate an rs_control structure for each worker thread */
|
|
ret = mempool_init(&f->rs_pool, num_online_cpus(), fec_rs_alloc,
|
|
fec_rs_free, (void *) v);
|
|
if (ret) {
|
|
ti->error = "Cannot allocate RS pool";
|
|
return ret;
|
|
}
|
|
|
|
f->cache = kmem_cache_create("dm_verity_fec_buffers",
|
|
f->rsn << DM_VERITY_FEC_BUF_RS_BITS,
|
|
0, 0, NULL);
|
|
if (!f->cache) {
|
|
ti->error = "Cannot create FEC buffer cache";
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Preallocate one buffer for each thread */
|
|
ret = mempool_init_slab_pool(&f->prealloc_pool, num_online_cpus(),
|
|
f->cache);
|
|
if (ret) {
|
|
ti->error = "Cannot allocate FEC buffer prealloc pool";
|
|
return ret;
|
|
}
|
|
|
|
/* Preallocate an output buffer for each thread */
|
|
ret = mempool_init_kmalloc_pool(&f->output_pool, num_online_cpus(),
|
|
1 << v->data_dev_block_bits);
|
|
if (ret) {
|
|
ti->error = "Cannot allocate FEC output pool";
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|