mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-10-29 23:53:32 +00:00
cffb5152ee
Remove tabs before spaces from comment as recommended by kernel coding style. Checkpatch also warn about these. Signed-off-by: Kari Argillander <kari.argillander@gmail.com> Signed-off-by: Konstantin Komarov <almaz.alexandrovich@paragon-software.com>
453 lines
9.4 KiB
C
453 lines
9.4 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
*
|
|
* Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
|
|
*
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/stddef.h>
|
|
#include <linux/string.h>
|
|
#include <linux/types.h>
|
|
|
|
#include "debug.h"
|
|
#include "ntfs_fs.h"
|
|
|
|
// clang-format off
|
|
/* Src buffer is zero. */
|
|
#define LZNT_ERROR_ALL_ZEROS 1
|
|
#define LZNT_CHUNK_SIZE 0x1000
|
|
// clang-format on
|
|
|
|
struct lznt_hash {
|
|
const u8 *p1;
|
|
const u8 *p2;
|
|
};
|
|
|
|
struct lznt {
|
|
const u8 *unc;
|
|
const u8 *unc_end;
|
|
const u8 *best_match;
|
|
size_t max_len;
|
|
bool std;
|
|
|
|
struct lznt_hash hash[LZNT_CHUNK_SIZE];
|
|
};
|
|
|
|
static inline size_t get_match_len(const u8 *ptr, const u8 *end, const u8 *prev,
|
|
size_t max_len)
|
|
{
|
|
size_t len = 0;
|
|
|
|
while (ptr + len < end && ptr[len] == prev[len] && ++len < max_len)
|
|
;
|
|
return len;
|
|
}
|
|
|
|
static size_t longest_match_std(const u8 *src, struct lznt *ctx)
|
|
{
|
|
size_t hash_index;
|
|
size_t len1 = 0, len2 = 0;
|
|
const u8 **hash;
|
|
|
|
hash_index =
|
|
((40543U * ((((src[0] << 4) ^ src[1]) << 4) ^ src[2])) >> 4) &
|
|
(LZNT_CHUNK_SIZE - 1);
|
|
|
|
hash = &(ctx->hash[hash_index].p1);
|
|
|
|
if (hash[0] >= ctx->unc && hash[0] < src && hash[0][0] == src[0] &&
|
|
hash[0][1] == src[1] && hash[0][2] == src[2]) {
|
|
len1 = 3;
|
|
if (ctx->max_len > 3)
|
|
len1 += get_match_len(src + 3, ctx->unc_end,
|
|
hash[0] + 3, ctx->max_len - 3);
|
|
}
|
|
|
|
if (hash[1] >= ctx->unc && hash[1] < src && hash[1][0] == src[0] &&
|
|
hash[1][1] == src[1] && hash[1][2] == src[2]) {
|
|
len2 = 3;
|
|
if (ctx->max_len > 3)
|
|
len2 += get_match_len(src + 3, ctx->unc_end,
|
|
hash[1] + 3, ctx->max_len - 3);
|
|
}
|
|
|
|
/* Compare two matches and select the best one. */
|
|
if (len1 < len2) {
|
|
ctx->best_match = hash[1];
|
|
len1 = len2;
|
|
} else {
|
|
ctx->best_match = hash[0];
|
|
}
|
|
|
|
hash[1] = hash[0];
|
|
hash[0] = src;
|
|
return len1;
|
|
}
|
|
|
|
static size_t longest_match_best(const u8 *src, struct lznt *ctx)
|
|
{
|
|
size_t max_len;
|
|
const u8 *ptr;
|
|
|
|
if (ctx->unc >= src || !ctx->max_len)
|
|
return 0;
|
|
|
|
max_len = 0;
|
|
for (ptr = ctx->unc; ptr < src; ++ptr) {
|
|
size_t len =
|
|
get_match_len(src, ctx->unc_end, ptr, ctx->max_len);
|
|
if (len >= max_len) {
|
|
max_len = len;
|
|
ctx->best_match = ptr;
|
|
}
|
|
}
|
|
|
|
return max_len >= 3 ? max_len : 0;
|
|
}
|
|
|
|
static const size_t s_max_len[] = {
|
|
0x1002, 0x802, 0x402, 0x202, 0x102, 0x82, 0x42, 0x22, 0x12,
|
|
};
|
|
|
|
static const size_t s_max_off[] = {
|
|
0x10, 0x20, 0x40, 0x80, 0x100, 0x200, 0x400, 0x800, 0x1000,
|
|
};
|
|
|
|
static inline u16 make_pair(size_t offset, size_t len, size_t index)
|
|
{
|
|
return ((offset - 1) << (12 - index)) |
|
|
((len - 3) & (((1 << (12 - index)) - 1)));
|
|
}
|
|
|
|
static inline size_t parse_pair(u16 pair, size_t *offset, size_t index)
|
|
{
|
|
*offset = 1 + (pair >> (12 - index));
|
|
return 3 + (pair & ((1 << (12 - index)) - 1));
|
|
}
|
|
|
|
/*
|
|
* compress_chunk
|
|
*
|
|
* Return:
|
|
* * 0 - Ok, @cmpr contains @cmpr_chunk_size bytes of compressed data.
|
|
* * 1 - Input buffer is full zero.
|
|
* * -2 - The compressed buffer is too small to hold the compressed data.
|
|
*/
|
|
static inline int compress_chunk(size_t (*match)(const u8 *, struct lznt *),
|
|
const u8 *unc, const u8 *unc_end, u8 *cmpr,
|
|
u8 *cmpr_end, size_t *cmpr_chunk_size,
|
|
struct lznt *ctx)
|
|
{
|
|
size_t cnt = 0;
|
|
size_t idx = 0;
|
|
const u8 *up = unc;
|
|
u8 *cp = cmpr + 3;
|
|
u8 *cp2 = cmpr + 2;
|
|
u8 not_zero = 0;
|
|
/* Control byte of 8-bit values: ( 0 - means byte as is, 1 - short pair ). */
|
|
u8 ohdr = 0;
|
|
u8 *last;
|
|
u16 t16;
|
|
|
|
if (unc + LZNT_CHUNK_SIZE < unc_end)
|
|
unc_end = unc + LZNT_CHUNK_SIZE;
|
|
|
|
last = min(cmpr + LZNT_CHUNK_SIZE + sizeof(short), cmpr_end);
|
|
|
|
ctx->unc = unc;
|
|
ctx->unc_end = unc_end;
|
|
ctx->max_len = s_max_len[0];
|
|
|
|
while (up < unc_end) {
|
|
size_t max_len;
|
|
|
|
while (unc + s_max_off[idx] < up)
|
|
ctx->max_len = s_max_len[++idx];
|
|
|
|
/* Find match. */
|
|
max_len = up + 3 <= unc_end ? (*match)(up, ctx) : 0;
|
|
|
|
if (!max_len) {
|
|
if (cp >= last)
|
|
goto NotCompressed;
|
|
not_zero |= *cp++ = *up++;
|
|
} else if (cp + 1 >= last) {
|
|
goto NotCompressed;
|
|
} else {
|
|
t16 = make_pair(up - ctx->best_match, max_len, idx);
|
|
*cp++ = t16;
|
|
*cp++ = t16 >> 8;
|
|
|
|
ohdr |= 1 << cnt;
|
|
up += max_len;
|
|
}
|
|
|
|
cnt = (cnt + 1) & 7;
|
|
if (!cnt) {
|
|
*cp2 = ohdr;
|
|
ohdr = 0;
|
|
cp2 = cp;
|
|
cp += 1;
|
|
}
|
|
}
|
|
|
|
if (cp2 < last)
|
|
*cp2 = ohdr;
|
|
else
|
|
cp -= 1;
|
|
|
|
*cmpr_chunk_size = cp - cmpr;
|
|
|
|
t16 = (*cmpr_chunk_size - 3) | 0xB000;
|
|
cmpr[0] = t16;
|
|
cmpr[1] = t16 >> 8;
|
|
|
|
return not_zero ? 0 : LZNT_ERROR_ALL_ZEROS;
|
|
|
|
NotCompressed:
|
|
|
|
if ((cmpr + LZNT_CHUNK_SIZE + sizeof(short)) > last)
|
|
return -2;
|
|
|
|
/*
|
|
* Copy non cmpr data.
|
|
* 0x3FFF == ((LZNT_CHUNK_SIZE + 2 - 3) | 0x3000)
|
|
*/
|
|
cmpr[0] = 0xff;
|
|
cmpr[1] = 0x3f;
|
|
|
|
memcpy(cmpr + sizeof(short), unc, LZNT_CHUNK_SIZE);
|
|
*cmpr_chunk_size = LZNT_CHUNK_SIZE + sizeof(short);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline ssize_t decompress_chunk(u8 *unc, u8 *unc_end, const u8 *cmpr,
|
|
const u8 *cmpr_end)
|
|
{
|
|
u8 *up = unc;
|
|
u8 ch = *cmpr++;
|
|
size_t bit = 0;
|
|
size_t index = 0;
|
|
u16 pair;
|
|
size_t offset, length;
|
|
|
|
/* Do decompression until pointers are inside range. */
|
|
while (up < unc_end && cmpr < cmpr_end) {
|
|
/* Correct index */
|
|
while (unc + s_max_off[index] < up)
|
|
index += 1;
|
|
|
|
/* Check the current flag for zero. */
|
|
if (!(ch & (1 << bit))) {
|
|
/* Just copy byte. */
|
|
*up++ = *cmpr++;
|
|
goto next;
|
|
}
|
|
|
|
/* Check for boundary. */
|
|
if (cmpr + 1 >= cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/* Read a short from little endian stream. */
|
|
pair = cmpr[1];
|
|
pair <<= 8;
|
|
pair |= cmpr[0];
|
|
|
|
cmpr += 2;
|
|
|
|
/* Translate packed information into offset and length. */
|
|
length = parse_pair(pair, &offset, index);
|
|
|
|
/* Check offset for boundary. */
|
|
if (unc + offset > up)
|
|
return -EINVAL;
|
|
|
|
/* Truncate the length if necessary. */
|
|
if (up + length >= unc_end)
|
|
length = unc_end - up;
|
|
|
|
/* Now we copy bytes. This is the heart of LZ algorithm. */
|
|
for (; length > 0; length--, up++)
|
|
*up = *(up - offset);
|
|
|
|
next:
|
|
/* Advance flag bit value. */
|
|
bit = (bit + 1) & 7;
|
|
|
|
if (!bit) {
|
|
if (cmpr >= cmpr_end)
|
|
break;
|
|
|
|
ch = *cmpr++;
|
|
}
|
|
}
|
|
|
|
/* Return the size of uncompressed data. */
|
|
return up - unc;
|
|
}
|
|
|
|
/*
|
|
* get_lznt_ctx
|
|
* @level: 0 - Standard compression.
|
|
* !0 - Best compression, requires a lot of cpu.
|
|
*/
|
|
struct lznt *get_lznt_ctx(int level)
|
|
{
|
|
struct lznt *r = kzalloc(level ? offsetof(struct lznt, hash)
|
|
: sizeof(struct lznt),
|
|
GFP_NOFS);
|
|
|
|
if (r)
|
|
r->std = !level;
|
|
return r;
|
|
}
|
|
|
|
/*
|
|
* compress_lznt - Compresses @unc into @cmpr
|
|
*
|
|
* Return:
|
|
* * +x - Ok, @cmpr contains 'final_compressed_size' bytes of compressed data.
|
|
* * 0 - Input buffer is full zero.
|
|
*/
|
|
size_t compress_lznt(const void *unc, size_t unc_size, void *cmpr,
|
|
size_t cmpr_size, struct lznt *ctx)
|
|
{
|
|
int err;
|
|
size_t (*match)(const u8 *src, struct lznt *ctx);
|
|
u8 *p = cmpr;
|
|
u8 *end = p + cmpr_size;
|
|
const u8 *unc_chunk = unc;
|
|
const u8 *unc_end = unc_chunk + unc_size;
|
|
bool is_zero = true;
|
|
|
|
if (ctx->std) {
|
|
match = &longest_match_std;
|
|
memset(ctx->hash, 0, sizeof(ctx->hash));
|
|
} else {
|
|
match = &longest_match_best;
|
|
}
|
|
|
|
/* Compression cycle. */
|
|
for (; unc_chunk < unc_end; unc_chunk += LZNT_CHUNK_SIZE) {
|
|
cmpr_size = 0;
|
|
err = compress_chunk(match, unc_chunk, unc_end, p, end,
|
|
&cmpr_size, ctx);
|
|
if (err < 0)
|
|
return unc_size;
|
|
|
|
if (is_zero && err != LZNT_ERROR_ALL_ZEROS)
|
|
is_zero = false;
|
|
|
|
p += cmpr_size;
|
|
}
|
|
|
|
if (p <= end - 2)
|
|
p[0] = p[1] = 0;
|
|
|
|
return is_zero ? 0 : PtrOffset(cmpr, p);
|
|
}
|
|
|
|
/*
|
|
* decompress_lznt - Decompress @cmpr into @unc.
|
|
*/
|
|
ssize_t decompress_lznt(const void *cmpr, size_t cmpr_size, void *unc,
|
|
size_t unc_size)
|
|
{
|
|
const u8 *cmpr_chunk = cmpr;
|
|
const u8 *cmpr_end = cmpr_chunk + cmpr_size;
|
|
u8 *unc_chunk = unc;
|
|
u8 *unc_end = unc_chunk + unc_size;
|
|
u16 chunk_hdr;
|
|
|
|
if (cmpr_size < sizeof(short))
|
|
return -EINVAL;
|
|
|
|
/* Read chunk header. */
|
|
chunk_hdr = cmpr_chunk[1];
|
|
chunk_hdr <<= 8;
|
|
chunk_hdr |= cmpr_chunk[0];
|
|
|
|
/* Loop through decompressing chunks. */
|
|
for (;;) {
|
|
size_t chunk_size_saved;
|
|
size_t unc_use;
|
|
size_t cmpr_use = 3 + (chunk_hdr & (LZNT_CHUNK_SIZE - 1));
|
|
|
|
/* Check that the chunk actually fits the supplied buffer. */
|
|
if (cmpr_chunk + cmpr_use > cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/* First make sure the chunk contains compressed data. */
|
|
if (chunk_hdr & 0x8000) {
|
|
/* Decompress a chunk and return if we get an error. */
|
|
ssize_t err =
|
|
decompress_chunk(unc_chunk, unc_end,
|
|
cmpr_chunk + sizeof(chunk_hdr),
|
|
cmpr_chunk + cmpr_use);
|
|
if (err < 0)
|
|
return err;
|
|
unc_use = err;
|
|
} else {
|
|
/* This chunk does not contain compressed data. */
|
|
unc_use = unc_chunk + LZNT_CHUNK_SIZE > unc_end
|
|
? unc_end - unc_chunk
|
|
: LZNT_CHUNK_SIZE;
|
|
|
|
if (cmpr_chunk + sizeof(chunk_hdr) + unc_use >
|
|
cmpr_end) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
memcpy(unc_chunk, cmpr_chunk + sizeof(chunk_hdr),
|
|
unc_use);
|
|
}
|
|
|
|
/* Advance pointers. */
|
|
cmpr_chunk += cmpr_use;
|
|
unc_chunk += unc_use;
|
|
|
|
/* Check for the end of unc buffer. */
|
|
if (unc_chunk >= unc_end)
|
|
break;
|
|
|
|
/* Proceed the next chunk. */
|
|
if (cmpr_chunk > cmpr_end - 2)
|
|
break;
|
|
|
|
chunk_size_saved = LZNT_CHUNK_SIZE;
|
|
|
|
/* Read chunk header. */
|
|
chunk_hdr = cmpr_chunk[1];
|
|
chunk_hdr <<= 8;
|
|
chunk_hdr |= cmpr_chunk[0];
|
|
|
|
if (!chunk_hdr)
|
|
break;
|
|
|
|
/* Check the size of unc buffer. */
|
|
if (unc_use < chunk_size_saved) {
|
|
size_t t1 = chunk_size_saved - unc_use;
|
|
u8 *t2 = unc_chunk + t1;
|
|
|
|
/* 'Zero' memory. */
|
|
if (t2 >= unc_end)
|
|
break;
|
|
|
|
memset(unc_chunk, 0, t1);
|
|
unc_chunk = t2;
|
|
}
|
|
}
|
|
|
|
/* Check compression boundary. */
|
|
if (cmpr_chunk > cmpr_end)
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* The unc size is just a difference between current
|
|
* pointer and original one.
|
|
*/
|
|
return PtrOffset(unc, unc_chunk);
|
|
}
|