mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2024-11-15 16:24:13 +08:00
e0c1b49f5b
Upgrade to the latest upstream zstd version 1.4.10. This patch is 100% generated from upstream zstd commit 20821a46f412 [0]. This patch is very large because it is transitioning from the custom kernel zstd to using upstream directly. The new zstd follows upstreams file structure which is different. Future update patches will be much smaller because they will only contain the changes from one upstream zstd release. As an aid for review I've created a commit [1] that shows the diff between upstream zstd as-is (which doesn't compile), and the zstd code imported in this patch. The verion of zstd in this patch is generated from upstream with changes applied by automation to replace upstreams libc dependencies, remove unnecessary portability macros, replace `/**` comments with `/*` comments, and use the kernel's xxhash instead of bundling it. The benefits of this patch are as follows: 1. Using upstream directly with automated script to generate kernel code. This allows us to update the kernel every upstream release, so the kernel gets the latest bug fixes and performance improvements, and doesn't get 3 years out of date again. The automation and the translated code are tested every upstream commit to ensure it continues to work. 2. Upgrades from a custom zstd based on 1.3.1 to 1.4.10, getting 3 years of performance improvements and bug fixes. On x86_64 I've measured 15% faster BtrFS and SquashFS decompression+read speeds, 35% faster kernel decompression, and 30% faster ZRAM decompression+read speeds. 3. Zstd-1.4.10 supports negative compression levels, which allow zstd to match or subsume lzo's performance. 4. Maintains the same kernel-specific wrapper API, so no callers have to be modified with zstd version updates. One concern that was brought up was stack usage. Upstream zstd had already removed most of its heavy stack usage functions, but I just removed the last functions that allocate arrays on the stack. I've measured the high water mark for both compression and decompression before and after this patch. Decompression is approximately neutral, using about 1.2KB of stack space. Compression levels up to 3 regressed from 1.4KB -> 1.6KB, and higher compression levels regressed from 1.5KB -> 2KB. We've added unit tests upstream to prevent further regression. I believe that this is a reasonable increase, and if it does end up causing problems, this commit can be cleanly reverted, because it only touches zstd. I chose the bulk update instead of replaying upstream commits because there have been ~3500 upstream commits since the 1.3.1 release, zstd wasn't ready to be used in the kernel as-is before a month ago, and not all upstream zstd commits build. The bulk update preserves bisectablity because bugs can be bisected to the zstd version update. At that point the update can be reverted, and we can work with upstream to find and fix the bug. Note that upstream zstd release 1.4.10 doesn't exist yet. I have cut a staging branch at 20821a46f412 [0] and will apply any changes requested to the staging branch. Once we're ready to merge this update I will cut a zstd release at the commit we merge, so we have a known zstd release in the kernel. The implementation of the kernel API is contained in zstd_compress_module.c and zstd_decompress_module.c. [0]20821a46f4
[1]e0fa481d0e
Signed-off-by: Nick Terrell <terrelln@fb.com> Tested By: Paul Jones <paul@pauljones.id.au> Tested-by: Oleksandr Natalenko <oleksandr@natalenko.name> Tested-by: Sedat Dilek <sedat.dilek@gmail.com> # LLVM/Clang v13.0.0 on x86-64 Tested-by: Jean-Denis Girard <jd.girard@sysnux.pf>
166 lines
6.6 KiB
C
166 lines
6.6 KiB
C
/* ******************************************************************
|
|
* hist : Histogram functions
|
|
* part of Finite State Entropy project
|
|
* Copyright (c) Yann Collet, Facebook, Inc.
|
|
*
|
|
* You can contact the author at :
|
|
* - FSE source repository : https://github.com/Cyan4973/FiniteStateEntropy
|
|
* - Public forum : https://groups.google.com/forum/#!forum/lz4c
|
|
*
|
|
* This source code is licensed under both the BSD-style license (found in the
|
|
* LICENSE file in the root directory of this source tree) and the GPLv2 (found
|
|
* in the COPYING file in the root directory of this source tree).
|
|
* You may select, at your option, one of the above-listed licenses.
|
|
****************************************************************** */
|
|
|
|
/* --- dependencies --- */
|
|
#include "../common/mem.h" /* U32, BYTE, etc. */
|
|
#include "../common/debug.h" /* assert, DEBUGLOG */
|
|
#include "../common/error_private.h" /* ERROR */
|
|
#include "hist.h"
|
|
|
|
|
|
/* --- Error management --- */
|
|
unsigned HIST_isError(size_t code) { return ERR_isError(code); }
|
|
|
|
/*-**************************************************************
|
|
* Histogram functions
|
|
****************************************************************/
|
|
unsigned HIST_count_simple(unsigned* count, unsigned* maxSymbolValuePtr,
|
|
const void* src, size_t srcSize)
|
|
{
|
|
const BYTE* ip = (const BYTE*)src;
|
|
const BYTE* const end = ip + srcSize;
|
|
unsigned maxSymbolValue = *maxSymbolValuePtr;
|
|
unsigned largestCount=0;
|
|
|
|
ZSTD_memset(count, 0, (maxSymbolValue+1) * sizeof(*count));
|
|
if (srcSize==0) { *maxSymbolValuePtr = 0; return 0; }
|
|
|
|
while (ip<end) {
|
|
assert(*ip <= maxSymbolValue);
|
|
count[*ip++]++;
|
|
}
|
|
|
|
while (!count[maxSymbolValue]) maxSymbolValue--;
|
|
*maxSymbolValuePtr = maxSymbolValue;
|
|
|
|
{ U32 s;
|
|
for (s=0; s<=maxSymbolValue; s++)
|
|
if (count[s] > largestCount) largestCount = count[s];
|
|
}
|
|
|
|
return largestCount;
|
|
}
|
|
|
|
typedef enum { trustInput, checkMaxSymbolValue } HIST_checkInput_e;
|
|
|
|
/* HIST_count_parallel_wksp() :
|
|
* store histogram into 4 intermediate tables, recombined at the end.
|
|
* this design makes better use of OoO cpus,
|
|
* and is noticeably faster when some values are heavily repeated.
|
|
* But it needs some additional workspace for intermediate tables.
|
|
* `workSpace` must be a U32 table of size >= HIST_WKSP_SIZE_U32.
|
|
* @return : largest histogram frequency,
|
|
* or an error code (notably when histogram's alphabet is larger than *maxSymbolValuePtr) */
|
|
static size_t HIST_count_parallel_wksp(
|
|
unsigned* count, unsigned* maxSymbolValuePtr,
|
|
const void* source, size_t sourceSize,
|
|
HIST_checkInput_e check,
|
|
U32* const workSpace)
|
|
{
|
|
const BYTE* ip = (const BYTE*)source;
|
|
const BYTE* const iend = ip+sourceSize;
|
|
size_t const countSize = (*maxSymbolValuePtr + 1) * sizeof(*count);
|
|
unsigned max=0;
|
|
U32* const Counting1 = workSpace;
|
|
U32* const Counting2 = Counting1 + 256;
|
|
U32* const Counting3 = Counting2 + 256;
|
|
U32* const Counting4 = Counting3 + 256;
|
|
|
|
/* safety checks */
|
|
assert(*maxSymbolValuePtr <= 255);
|
|
if (!sourceSize) {
|
|
ZSTD_memset(count, 0, countSize);
|
|
*maxSymbolValuePtr = 0;
|
|
return 0;
|
|
}
|
|
ZSTD_memset(workSpace, 0, 4*256*sizeof(unsigned));
|
|
|
|
/* by stripes of 16 bytes */
|
|
{ U32 cached = MEM_read32(ip); ip += 4;
|
|
while (ip < iend-15) {
|
|
U32 c = cached; cached = MEM_read32(ip); ip += 4;
|
|
Counting1[(BYTE) c ]++;
|
|
Counting2[(BYTE)(c>>8) ]++;
|
|
Counting3[(BYTE)(c>>16)]++;
|
|
Counting4[ c>>24 ]++;
|
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
|
Counting1[(BYTE) c ]++;
|
|
Counting2[(BYTE)(c>>8) ]++;
|
|
Counting3[(BYTE)(c>>16)]++;
|
|
Counting4[ c>>24 ]++;
|
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
|
Counting1[(BYTE) c ]++;
|
|
Counting2[(BYTE)(c>>8) ]++;
|
|
Counting3[(BYTE)(c>>16)]++;
|
|
Counting4[ c>>24 ]++;
|
|
c = cached; cached = MEM_read32(ip); ip += 4;
|
|
Counting1[(BYTE) c ]++;
|
|
Counting2[(BYTE)(c>>8) ]++;
|
|
Counting3[(BYTE)(c>>16)]++;
|
|
Counting4[ c>>24 ]++;
|
|
}
|
|
ip-=4;
|
|
}
|
|
|
|
/* finish last symbols */
|
|
while (ip<iend) Counting1[*ip++]++;
|
|
|
|
{ U32 s;
|
|
for (s=0; s<256; s++) {
|
|
Counting1[s] += Counting2[s] + Counting3[s] + Counting4[s];
|
|
if (Counting1[s] > max) max = Counting1[s];
|
|
} }
|
|
|
|
{ unsigned maxSymbolValue = 255;
|
|
while (!Counting1[maxSymbolValue]) maxSymbolValue--;
|
|
if (check && maxSymbolValue > *maxSymbolValuePtr) return ERROR(maxSymbolValue_tooSmall);
|
|
*maxSymbolValuePtr = maxSymbolValue;
|
|
ZSTD_memmove(count, Counting1, countSize); /* in case count & Counting1 are overlapping */
|
|
}
|
|
return (size_t)max;
|
|
}
|
|
|
|
/* HIST_countFast_wksp() :
|
|
* Same as HIST_countFast(), but using an externally provided scratch buffer.
|
|
* `workSpace` is a writable buffer which must be 4-bytes aligned,
|
|
* `workSpaceSize` must be >= HIST_WKSP_SIZE
|
|
*/
|
|
size_t HIST_countFast_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
|
const void* source, size_t sourceSize,
|
|
void* workSpace, size_t workSpaceSize)
|
|
{
|
|
if (sourceSize < 1500) /* heuristic threshold */
|
|
return HIST_count_simple(count, maxSymbolValuePtr, source, sourceSize);
|
|
if ((size_t)workSpace & 3) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
|
if (workSpaceSize < HIST_WKSP_SIZE) return ERROR(workSpace_tooSmall);
|
|
return HIST_count_parallel_wksp(count, maxSymbolValuePtr, source, sourceSize, trustInput, (U32*)workSpace);
|
|
}
|
|
|
|
/* HIST_count_wksp() :
|
|
* Same as HIST_count(), but using an externally provided scratch buffer.
|
|
* `workSpace` size must be table of >= HIST_WKSP_SIZE_U32 unsigned */
|
|
size_t HIST_count_wksp(unsigned* count, unsigned* maxSymbolValuePtr,
|
|
const void* source, size_t sourceSize,
|
|
void* workSpace, size_t workSpaceSize)
|
|
{
|
|
if ((size_t)workSpace & 3) return ERROR(GENERIC); /* must be aligned on 4-bytes boundaries */
|
|
if (workSpaceSize < HIST_WKSP_SIZE) return ERROR(workSpace_tooSmall);
|
|
if (*maxSymbolValuePtr < 255)
|
|
return HIST_count_parallel_wksp(count, maxSymbolValuePtr, source, sourceSize, checkMaxSymbolValue, (U32*)workSpace);
|
|
*maxSymbolValuePtr = 255;
|
|
return HIST_countFast_wksp(count, maxSymbolValuePtr, source, sourceSize, workSpace, workSpaceSize);
|
|
}
|
|
|