Scott Baker | eee8dd8 | 2019-09-24 12:52:34 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (c) 2016-present, Yann Collet, Facebook, Inc. |
| 3 | * All rights reserved. |
| 4 | * |
| 5 | * This source code is licensed under both the BSD-style license (found in the |
| 6 | * LICENSE file in the root directory of this source tree) and the GPLv2 (found |
| 7 | * in the COPYING file in the root directory of this source tree). |
| 8 | * You may select, at your option, one of the above-listed licenses. |
| 9 | */ |
| 10 | |
| 11 | |
| 12 | /****************************************** |
| 13 | * Includes |
| 14 | ******************************************/ |
| 15 | #include <stddef.h> /* size_t, ptrdiff_t */ |
| 16 | #include "zstd_v01.h" |
| 17 | #include "error_private.h" |
| 18 | |
| 19 | |
| 20 | /****************************************** |
| 21 | * Static allocation |
| 22 | ******************************************/ |
| 23 | /* You can statically allocate FSE CTable/DTable as a table of unsigned using below macro */ |
| 24 | #define FSE_DTABLE_SIZE_U32(maxTableLog) (1 + (1<<maxTableLog)) |
| 25 | |
| 26 | /* You can statically allocate Huff0 DTable as a table of unsigned short using below macro */ |
| 27 | #define HUF_DTABLE_SIZE_U16(maxTableLog) (1 + (1<<maxTableLog)) |
| 28 | #define HUF_CREATE_STATIC_DTABLE(DTable, maxTableLog) \ |
| 29 | unsigned short DTable[HUF_DTABLE_SIZE_U16(maxTableLog)] = { maxTableLog } |
| 30 | |
| 31 | |
| 32 | /****************************************** |
| 33 | * Error Management |
| 34 | ******************************************/ |
| 35 | #define FSE_LIST_ERRORS(ITEM) \ |
| 36 | ITEM(FSE_OK_NoError) ITEM(FSE_ERROR_GENERIC) \ |
| 37 | ITEM(FSE_ERROR_tableLog_tooLarge) ITEM(FSE_ERROR_maxSymbolValue_tooLarge) ITEM(FSE_ERROR_maxSymbolValue_tooSmall) \ |
| 38 | ITEM(FSE_ERROR_dstSize_tooSmall) ITEM(FSE_ERROR_srcSize_wrong)\ |
| 39 | ITEM(FSE_ERROR_corruptionDetected) \ |
| 40 | ITEM(FSE_ERROR_maxCode) |
| 41 | |
| 42 | #define FSE_GENERATE_ENUM(ENUM) ENUM, |
| 43 | typedef enum { FSE_LIST_ERRORS(FSE_GENERATE_ENUM) } FSE_errorCodes; /* enum is exposed, to detect & handle specific errors; compare function result to -enum value */ |
| 44 | |
| 45 | |
| 46 | /****************************************** |
| 47 | * FSE symbol compression API |
| 48 | ******************************************/ |
| 49 | /* |
| 50 | This API consists of small unitary functions, which highly benefit from being inlined. |
| 51 | You will want to enable link-time-optimization to ensure these functions are properly inlined in your binary. |
| 52 | Visual seems to do it automatically. |
| 53 | For gcc or clang, you'll need to add -flto flag at compilation and linking stages. |
| 54 | If none of these solutions is applicable, include "fse.c" directly. |
| 55 | */ |
| 56 | |
| 57 | typedef unsigned FSE_CTable; /* don't allocate that. It's just a way to be more restrictive than void* */ |
| 58 | typedef unsigned FSE_DTable; /* don't allocate that. It's just a way to be more restrictive than void* */ |
| 59 | |
| 60 | typedef struct |
| 61 | { |
| 62 | size_t bitContainer; |
| 63 | int bitPos; |
| 64 | char* startPtr; |
| 65 | char* ptr; |
| 66 | char* endPtr; |
| 67 | } FSE_CStream_t; |
| 68 | |
| 69 | typedef struct |
| 70 | { |
| 71 | ptrdiff_t value; |
| 72 | const void* stateTable; |
| 73 | const void* symbolTT; |
| 74 | unsigned stateLog; |
| 75 | } FSE_CState_t; |
| 76 | |
| 77 | typedef struct |
| 78 | { |
| 79 | size_t bitContainer; |
| 80 | unsigned bitsConsumed; |
| 81 | const char* ptr; |
| 82 | const char* start; |
| 83 | } FSE_DStream_t; |
| 84 | |
| 85 | typedef struct |
| 86 | { |
| 87 | size_t state; |
| 88 | const void* table; /* precise table may vary, depending on U16 */ |
| 89 | } FSE_DState_t; |
| 90 | |
| 91 | typedef enum { FSE_DStream_unfinished = 0, |
| 92 | FSE_DStream_endOfBuffer = 1, |
| 93 | FSE_DStream_completed = 2, |
| 94 | FSE_DStream_tooFar = 3 } FSE_DStream_status; /* result of FSE_reloadDStream() */ |
| 95 | /* 1,2,4,8 would be better for bitmap combinations, but slows down performance a bit ... ?! */ |
| 96 | |
| 97 | |
| 98 | /**************************************************************** |
| 99 | * Tuning parameters |
| 100 | ****************************************************************/ |
| 101 | /* MEMORY_USAGE : |
| 102 | * Memory usage formula : N->2^N Bytes (examples : 10 -> 1KB; 12 -> 4KB ; 16 -> 64KB; 20 -> 1MB; etc.) |
| 103 | * Increasing memory usage improves compression ratio |
| 104 | * Reduced memory usage can improve speed, due to cache effect |
| 105 | * Recommended max value is 14, for 16KB, which nicely fits into Intel x86 L1 cache */ |
| 106 | #define FSE_MAX_MEMORY_USAGE 14 |
| 107 | #define FSE_DEFAULT_MEMORY_USAGE 13 |
| 108 | |
| 109 | /* FSE_MAX_SYMBOL_VALUE : |
| 110 | * Maximum symbol value authorized. |
| 111 | * Required for proper stack allocation */ |
| 112 | #define FSE_MAX_SYMBOL_VALUE 255 |
| 113 | |
| 114 | |
| 115 | /**************************************************************** |
| 116 | * template functions type & suffix |
| 117 | ****************************************************************/ |
| 118 | #define FSE_FUNCTION_TYPE BYTE |
| 119 | #define FSE_FUNCTION_EXTENSION |
| 120 | |
| 121 | |
| 122 | /**************************************************************** |
| 123 | * Byte symbol type |
| 124 | ****************************************************************/ |
| 125 | typedef struct |
| 126 | { |
| 127 | unsigned short newState; |
| 128 | unsigned char symbol; |
| 129 | unsigned char nbBits; |
| 130 | } FSE_decode_t; /* size == U32 */ |
| 131 | |
| 132 | |
| 133 | |
| 134 | /**************************************************************** |
| 135 | * Compiler specifics |
| 136 | ****************************************************************/ |
| 137 | #ifdef _MSC_VER /* Visual Studio */ |
| 138 | # define FORCE_INLINE static __forceinline |
| 139 | # include <intrin.h> /* For Visual 2005 */ |
| 140 | # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */ |
| 141 | # pragma warning(disable : 4214) /* disable: C4214: non-int bitfields */ |
| 142 | #else |
| 143 | # define GCC_VERSION (__GNUC__ * 100 + __GNUC_MINOR__) |
| 144 | # if defined (__cplusplus) || defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */ |
| 145 | # ifdef __GNUC__ |
| 146 | # define FORCE_INLINE static inline __attribute__((always_inline)) |
| 147 | # else |
| 148 | # define FORCE_INLINE static inline |
| 149 | # endif |
| 150 | # else |
| 151 | # define FORCE_INLINE static |
| 152 | # endif /* __STDC_VERSION__ */ |
| 153 | #endif |
| 154 | |
| 155 | |
| 156 | /**************************************************************** |
| 157 | * Includes |
| 158 | ****************************************************************/ |
| 159 | #include <stdlib.h> /* malloc, free, qsort */ |
| 160 | #include <string.h> /* memcpy, memset */ |
| 161 | #include <stdio.h> /* printf (debug) */ |
| 162 | |
| 163 | |
| 164 | #ifndef MEM_ACCESS_MODULE |
| 165 | #define MEM_ACCESS_MODULE |
| 166 | /**************************************************************** |
| 167 | * Basic Types |
| 168 | *****************************************************************/ |
| 169 | #if defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */ |
| 170 | # include <stdint.h> |
| 171 | typedef uint8_t BYTE; |
| 172 | typedef uint16_t U16; |
| 173 | typedef int16_t S16; |
| 174 | typedef uint32_t U32; |
| 175 | typedef int32_t S32; |
| 176 | typedef uint64_t U64; |
| 177 | typedef int64_t S64; |
| 178 | #else |
| 179 | typedef unsigned char BYTE; |
| 180 | typedef unsigned short U16; |
| 181 | typedef signed short S16; |
| 182 | typedef unsigned int U32; |
| 183 | typedef signed int S32; |
| 184 | typedef unsigned long long U64; |
| 185 | typedef signed long long S64; |
| 186 | #endif |
| 187 | |
| 188 | #endif /* MEM_ACCESS_MODULE */ |
| 189 | |
| 190 | /**************************************************************** |
| 191 | * Memory I/O |
| 192 | *****************************************************************/ |
| 193 | /* FSE_FORCE_MEMORY_ACCESS |
| 194 | * By default, access to unaligned memory is controlled by `memcpy()`, which is safe and portable. |
| 195 | * Unfortunately, on some target/compiler combinations, the generated assembly is sub-optimal. |
| 196 | * The below switch allow to select different access method for improved performance. |
| 197 | * Method 0 (default) : use `memcpy()`. Safe and portable. |
| 198 | * Method 1 : `__packed` statement. It depends on compiler extension (ie, not portable). |
| 199 | * This method is safe if your compiler supports it, and *generally* as fast or faster than `memcpy`. |
| 200 | * Method 2 : direct access. This method is portable but violate C standard. |
| 201 | * It can generate buggy code on targets generating assembly depending on alignment. |
| 202 | * But in some circumstances, it's the only known way to get the most performance (ie GCC + ARMv6) |
| 203 | * See http://fastcompression.blogspot.fr/2015/08/accessing-unaligned-memory.html for details. |
| 204 | * Prefer these methods in priority order (0 > 1 > 2) |
| 205 | */ |
| 206 | #ifndef FSE_FORCE_MEMORY_ACCESS /* can be defined externally, on command line for example */ |
| 207 | # if defined(__GNUC__) && ( defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__) || defined(__ARM_ARCH_6T2__) ) |
| 208 | # define FSE_FORCE_MEMORY_ACCESS 2 |
| 209 | # elif (defined(__INTEL_COMPILER) && !defined(WIN32)) || \ |
| 210 | (defined(__GNUC__) && ( defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_7A__) || defined(__ARM_ARCH_7R__) || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7S__) )) |
| 211 | # define FSE_FORCE_MEMORY_ACCESS 1 |
| 212 | # endif |
| 213 | #endif |
| 214 | |
| 215 | |
| 216 | static unsigned FSE_32bits(void) |
| 217 | { |
| 218 | return sizeof(void*)==4; |
| 219 | } |
| 220 | |
| 221 | static unsigned FSE_isLittleEndian(void) |
| 222 | { |
| 223 | const union { U32 i; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */ |
| 224 | return one.c[0]; |
| 225 | } |
| 226 | |
| 227 | #if defined(FSE_FORCE_MEMORY_ACCESS) && (FSE_FORCE_MEMORY_ACCESS==2) |
| 228 | |
| 229 | static U16 FSE_read16(const void* memPtr) { return *(const U16*) memPtr; } |
| 230 | static U32 FSE_read32(const void* memPtr) { return *(const U32*) memPtr; } |
| 231 | static U64 FSE_read64(const void* memPtr) { return *(const U64*) memPtr; } |
| 232 | |
| 233 | #elif defined(FSE_FORCE_MEMORY_ACCESS) && (FSE_FORCE_MEMORY_ACCESS==1) |
| 234 | |
| 235 | /* __pack instructions are safer, but compiler specific, hence potentially problematic for some compilers */ |
| 236 | /* currently only defined for gcc and icc */ |
| 237 | typedef union { U16 u16; U32 u32; U64 u64; } __attribute__((packed)) unalign; |
| 238 | |
| 239 | static U16 FSE_read16(const void* ptr) { return ((const unalign*)ptr)->u16; } |
| 240 | static U32 FSE_read32(const void* ptr) { return ((const unalign*)ptr)->u32; } |
| 241 | static U64 FSE_read64(const void* ptr) { return ((const unalign*)ptr)->u64; } |
| 242 | |
| 243 | #else |
| 244 | |
| 245 | static U16 FSE_read16(const void* memPtr) |
| 246 | { |
| 247 | U16 val; memcpy(&val, memPtr, sizeof(val)); return val; |
| 248 | } |
| 249 | |
| 250 | static U32 FSE_read32(const void* memPtr) |
| 251 | { |
| 252 | U32 val; memcpy(&val, memPtr, sizeof(val)); return val; |
| 253 | } |
| 254 | |
| 255 | static U64 FSE_read64(const void* memPtr) |
| 256 | { |
| 257 | U64 val; memcpy(&val, memPtr, sizeof(val)); return val; |
| 258 | } |
| 259 | |
| 260 | #endif // FSE_FORCE_MEMORY_ACCESS |
| 261 | |
| 262 | static U16 FSE_readLE16(const void* memPtr) |
| 263 | { |
| 264 | if (FSE_isLittleEndian()) |
| 265 | return FSE_read16(memPtr); |
| 266 | else |
| 267 | { |
| 268 | const BYTE* p = (const BYTE*)memPtr; |
| 269 | return (U16)(p[0] + (p[1]<<8)); |
| 270 | } |
| 271 | } |
| 272 | |
| 273 | static U32 FSE_readLE32(const void* memPtr) |
| 274 | { |
| 275 | if (FSE_isLittleEndian()) |
| 276 | return FSE_read32(memPtr); |
| 277 | else |
| 278 | { |
| 279 | const BYTE* p = (const BYTE*)memPtr; |
| 280 | return (U32)((U32)p[0] + ((U32)p[1]<<8) + ((U32)p[2]<<16) + ((U32)p[3]<<24)); |
| 281 | } |
| 282 | } |
| 283 | |
| 284 | |
| 285 | static U64 FSE_readLE64(const void* memPtr) |
| 286 | { |
| 287 | if (FSE_isLittleEndian()) |
| 288 | return FSE_read64(memPtr); |
| 289 | else |
| 290 | { |
| 291 | const BYTE* p = (const BYTE*)memPtr; |
| 292 | return (U64)((U64)p[0] + ((U64)p[1]<<8) + ((U64)p[2]<<16) + ((U64)p[3]<<24) |
| 293 | + ((U64)p[4]<<32) + ((U64)p[5]<<40) + ((U64)p[6]<<48) + ((U64)p[7]<<56)); |
| 294 | } |
| 295 | } |
| 296 | |
| 297 | static size_t FSE_readLEST(const void* memPtr) |
| 298 | { |
| 299 | if (FSE_32bits()) |
| 300 | return (size_t)FSE_readLE32(memPtr); |
| 301 | else |
| 302 | return (size_t)FSE_readLE64(memPtr); |
| 303 | } |
| 304 | |
| 305 | |
| 306 | |
| 307 | /**************************************************************** |
| 308 | * Constants |
| 309 | *****************************************************************/ |
| 310 | #define FSE_MAX_TABLELOG (FSE_MAX_MEMORY_USAGE-2) |
| 311 | #define FSE_MAX_TABLESIZE (1U<<FSE_MAX_TABLELOG) |
| 312 | #define FSE_MAXTABLESIZE_MASK (FSE_MAX_TABLESIZE-1) |
| 313 | #define FSE_DEFAULT_TABLELOG (FSE_DEFAULT_MEMORY_USAGE-2) |
| 314 | #define FSE_MIN_TABLELOG 5 |
| 315 | |
| 316 | #define FSE_TABLELOG_ABSOLUTE_MAX 15 |
| 317 | #if FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX |
| 318 | #error "FSE_MAX_TABLELOG > FSE_TABLELOG_ABSOLUTE_MAX is not supported" |
| 319 | #endif |
| 320 | |
| 321 | |
| 322 | /**************************************************************** |
| 323 | * Error Management |
| 324 | ****************************************************************/ |
| 325 | #define FSE_STATIC_ASSERT(c) { enum { FSE_static_assert = 1/(int)(!!(c)) }; } /* use only *after* variable declarations */ |
| 326 | |
| 327 | |
| 328 | /**************************************************************** |
| 329 | * Complex types |
| 330 | ****************************************************************/ |
| 331 | typedef struct |
| 332 | { |
| 333 | int deltaFindState; |
| 334 | U32 deltaNbBits; |
| 335 | } FSE_symbolCompressionTransform; /* total 8 bytes */ |
| 336 | |
| 337 | typedef U32 DTable_max_t[FSE_DTABLE_SIZE_U32(FSE_MAX_TABLELOG)]; |
| 338 | |
| 339 | /**************************************************************** |
| 340 | * Internal functions |
| 341 | ****************************************************************/ |
| 342 | FORCE_INLINE unsigned FSE_highbit32 (U32 val) |
| 343 | { |
| 344 | # if defined(_MSC_VER) /* Visual */ |
| 345 | unsigned long r; |
| 346 | _BitScanReverse ( &r, val ); |
| 347 | return (unsigned) r; |
| 348 | # elif defined(__GNUC__) && (GCC_VERSION >= 304) /* GCC Intrinsic */ |
| 349 | return 31 - __builtin_clz (val); |
| 350 | # else /* Software version */ |
| 351 | static const unsigned DeBruijnClz[32] = { 0, 9, 1, 10, 13, 21, 2, 29, 11, 14, 16, 18, 22, 25, 3, 30, 8, 12, 20, 28, 15, 17, 24, 7, 19, 27, 23, 6, 26, 5, 4, 31 }; |
| 352 | U32 v = val; |
| 353 | unsigned r; |
| 354 | v |= v >> 1; |
| 355 | v |= v >> 2; |
| 356 | v |= v >> 4; |
| 357 | v |= v >> 8; |
| 358 | v |= v >> 16; |
| 359 | r = DeBruijnClz[ (U32) (v * 0x07C4ACDDU) >> 27]; |
| 360 | return r; |
| 361 | # endif |
| 362 | } |
| 363 | |
| 364 | |
| 365 | /**************************************************************** |
| 366 | * Templates |
| 367 | ****************************************************************/ |
| 368 | /* |
| 369 | designed to be included |
| 370 | for type-specific functions (template emulation in C) |
| 371 | Objective is to write these functions only once, for improved maintenance |
| 372 | */ |
| 373 | |
| 374 | /* safety checks */ |
| 375 | #ifndef FSE_FUNCTION_EXTENSION |
| 376 | # error "FSE_FUNCTION_EXTENSION must be defined" |
| 377 | #endif |
| 378 | #ifndef FSE_FUNCTION_TYPE |
| 379 | # error "FSE_FUNCTION_TYPE must be defined" |
| 380 | #endif |
| 381 | |
| 382 | /* Function names */ |
| 383 | #define FSE_CAT(X,Y) X##Y |
| 384 | #define FSE_FUNCTION_NAME(X,Y) FSE_CAT(X,Y) |
| 385 | #define FSE_TYPE_NAME(X,Y) FSE_CAT(X,Y) |
| 386 | |
| 387 | |
| 388 | |
| 389 | static U32 FSE_tableStep(U32 tableSize) { return (tableSize>>1) + (tableSize>>3) + 3; } |
| 390 | |
| 391 | #define FSE_DECODE_TYPE FSE_decode_t |
| 392 | |
| 393 | |
| 394 | typedef struct { |
| 395 | U16 tableLog; |
| 396 | U16 fastMode; |
| 397 | } FSE_DTableHeader; /* sizeof U32 */ |
| 398 | |
| 399 | static size_t FSE_buildDTable |
| 400 | (FSE_DTable* dt, const short* normalizedCounter, unsigned maxSymbolValue, unsigned tableLog) |
| 401 | { |
| 402 | void* ptr = dt; |
| 403 | FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr; |
| 404 | FSE_DECODE_TYPE* const tableDecode = (FSE_DECODE_TYPE*)(ptr) + 1; /* because dt is unsigned, 32-bits aligned on 32-bits */ |
| 405 | const U32 tableSize = 1 << tableLog; |
| 406 | const U32 tableMask = tableSize-1; |
| 407 | const U32 step = FSE_tableStep(tableSize); |
| 408 | U16 symbolNext[FSE_MAX_SYMBOL_VALUE+1]; |
| 409 | U32 position = 0; |
| 410 | U32 highThreshold = tableSize-1; |
| 411 | const S16 largeLimit= (S16)(1 << (tableLog-1)); |
| 412 | U32 noLarge = 1; |
| 413 | U32 s; |
| 414 | |
| 415 | /* Sanity Checks */ |
| 416 | if (maxSymbolValue > FSE_MAX_SYMBOL_VALUE) return (size_t)-FSE_ERROR_maxSymbolValue_tooLarge; |
| 417 | if (tableLog > FSE_MAX_TABLELOG) return (size_t)-FSE_ERROR_tableLog_tooLarge; |
| 418 | |
| 419 | /* Init, lay down lowprob symbols */ |
| 420 | DTableH[0].tableLog = (U16)tableLog; |
| 421 | for (s=0; s<=maxSymbolValue; s++) |
| 422 | { |
| 423 | if (normalizedCounter[s]==-1) |
| 424 | { |
| 425 | tableDecode[highThreshold--].symbol = (FSE_FUNCTION_TYPE)s; |
| 426 | symbolNext[s] = 1; |
| 427 | } |
| 428 | else |
| 429 | { |
| 430 | if (normalizedCounter[s] >= largeLimit) noLarge=0; |
| 431 | symbolNext[s] = normalizedCounter[s]; |
| 432 | } |
| 433 | } |
| 434 | |
| 435 | /* Spread symbols */ |
| 436 | for (s=0; s<=maxSymbolValue; s++) |
| 437 | { |
| 438 | int i; |
| 439 | for (i=0; i<normalizedCounter[s]; i++) |
| 440 | { |
| 441 | tableDecode[position].symbol = (FSE_FUNCTION_TYPE)s; |
| 442 | position = (position + step) & tableMask; |
| 443 | while (position > highThreshold) position = (position + step) & tableMask; /* lowprob area */ |
| 444 | } |
| 445 | } |
| 446 | |
| 447 | if (position!=0) return (size_t)-FSE_ERROR_GENERIC; /* position must reach all cells once, otherwise normalizedCounter is incorrect */ |
| 448 | |
| 449 | /* Build Decoding table */ |
| 450 | { |
| 451 | U32 i; |
| 452 | for (i=0; i<tableSize; i++) |
| 453 | { |
| 454 | FSE_FUNCTION_TYPE symbol = (FSE_FUNCTION_TYPE)(tableDecode[i].symbol); |
| 455 | U16 nextState = symbolNext[symbol]++; |
| 456 | tableDecode[i].nbBits = (BYTE) (tableLog - FSE_highbit32 ((U32)nextState) ); |
| 457 | tableDecode[i].newState = (U16) ( (nextState << tableDecode[i].nbBits) - tableSize); |
| 458 | } |
| 459 | } |
| 460 | |
| 461 | DTableH->fastMode = (U16)noLarge; |
| 462 | return 0; |
| 463 | } |
| 464 | |
| 465 | |
| 466 | /****************************************** |
| 467 | * FSE byte symbol |
| 468 | ******************************************/ |
| 469 | #ifndef FSE_COMMONDEFS_ONLY |
| 470 | |
| 471 | static unsigned FSE_isError(size_t code) { return (code > (size_t)(-FSE_ERROR_maxCode)); } |
| 472 | |
| 473 | static short FSE_abs(short a) |
| 474 | { |
| 475 | return a<0? -a : a; |
| 476 | } |
| 477 | |
| 478 | |
| 479 | /**************************************************************** |
| 480 | * Header bitstream management |
| 481 | ****************************************************************/ |
| 482 | static size_t FSE_readNCount (short* normalizedCounter, unsigned* maxSVPtr, unsigned* tableLogPtr, |
| 483 | const void* headerBuffer, size_t hbSize) |
| 484 | { |
| 485 | const BYTE* const istart = (const BYTE*) headerBuffer; |
| 486 | const BYTE* const iend = istart + hbSize; |
| 487 | const BYTE* ip = istart; |
| 488 | int nbBits; |
| 489 | int remaining; |
| 490 | int threshold; |
| 491 | U32 bitStream; |
| 492 | int bitCount; |
| 493 | unsigned charnum = 0; |
| 494 | int previous0 = 0; |
| 495 | |
| 496 | if (hbSize < 4) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 497 | bitStream = FSE_readLE32(ip); |
| 498 | nbBits = (bitStream & 0xF) + FSE_MIN_TABLELOG; /* extract tableLog */ |
| 499 | if (nbBits > FSE_TABLELOG_ABSOLUTE_MAX) return (size_t)-FSE_ERROR_tableLog_tooLarge; |
| 500 | bitStream >>= 4; |
| 501 | bitCount = 4; |
| 502 | *tableLogPtr = nbBits; |
| 503 | remaining = (1<<nbBits)+1; |
| 504 | threshold = 1<<nbBits; |
| 505 | nbBits++; |
| 506 | |
| 507 | while ((remaining>1) && (charnum<=*maxSVPtr)) |
| 508 | { |
| 509 | if (previous0) |
| 510 | { |
| 511 | unsigned n0 = charnum; |
| 512 | while ((bitStream & 0xFFFF) == 0xFFFF) |
| 513 | { |
| 514 | n0+=24; |
| 515 | if (ip < iend-5) |
| 516 | { |
| 517 | ip+=2; |
| 518 | bitStream = FSE_readLE32(ip) >> bitCount; |
| 519 | } |
| 520 | else |
| 521 | { |
| 522 | bitStream >>= 16; |
| 523 | bitCount+=16; |
| 524 | } |
| 525 | } |
| 526 | while ((bitStream & 3) == 3) |
| 527 | { |
| 528 | n0+=3; |
| 529 | bitStream>>=2; |
| 530 | bitCount+=2; |
| 531 | } |
| 532 | n0 += bitStream & 3; |
| 533 | bitCount += 2; |
| 534 | if (n0 > *maxSVPtr) return (size_t)-FSE_ERROR_maxSymbolValue_tooSmall; |
| 535 | while (charnum < n0) normalizedCounter[charnum++] = 0; |
| 536 | if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) |
| 537 | { |
| 538 | ip += bitCount>>3; |
| 539 | bitCount &= 7; |
| 540 | bitStream = FSE_readLE32(ip) >> bitCount; |
| 541 | } |
| 542 | else |
| 543 | bitStream >>= 2; |
| 544 | } |
| 545 | { |
| 546 | const short max = (short)((2*threshold-1)-remaining); |
| 547 | short count; |
| 548 | |
| 549 | if ((bitStream & (threshold-1)) < (U32)max) |
| 550 | { |
| 551 | count = (short)(bitStream & (threshold-1)); |
| 552 | bitCount += nbBits-1; |
| 553 | } |
| 554 | else |
| 555 | { |
| 556 | count = (short)(bitStream & (2*threshold-1)); |
| 557 | if (count >= threshold) count -= max; |
| 558 | bitCount += nbBits; |
| 559 | } |
| 560 | |
| 561 | count--; /* extra accuracy */ |
| 562 | remaining -= FSE_abs(count); |
| 563 | normalizedCounter[charnum++] = count; |
| 564 | previous0 = !count; |
| 565 | while (remaining < threshold) |
| 566 | { |
| 567 | nbBits--; |
| 568 | threshold >>= 1; |
| 569 | } |
| 570 | |
| 571 | { |
| 572 | if ((ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) |
| 573 | { |
| 574 | ip += bitCount>>3; |
| 575 | bitCount &= 7; |
| 576 | } |
| 577 | else |
| 578 | { |
| 579 | bitCount -= (int)(8 * (iend - 4 - ip)); |
| 580 | ip = iend - 4; |
| 581 | } |
| 582 | bitStream = FSE_readLE32(ip) >> (bitCount & 31); |
| 583 | } |
| 584 | } |
| 585 | } |
| 586 | if (remaining != 1) return (size_t)-FSE_ERROR_GENERIC; |
| 587 | *maxSVPtr = charnum-1; |
| 588 | |
| 589 | ip += (bitCount+7)>>3; |
| 590 | if ((size_t)(ip-istart) > hbSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 591 | return ip-istart; |
| 592 | } |
| 593 | |
| 594 | |
| 595 | /********************************************************* |
| 596 | * Decompression (Byte symbols) |
| 597 | *********************************************************/ |
| 598 | static size_t FSE_buildDTable_rle (FSE_DTable* dt, BYTE symbolValue) |
| 599 | { |
| 600 | void* ptr = dt; |
| 601 | FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr; |
| 602 | FSE_decode_t* const cell = (FSE_decode_t*)(ptr) + 1; /* because dt is unsigned */ |
| 603 | |
| 604 | DTableH->tableLog = 0; |
| 605 | DTableH->fastMode = 0; |
| 606 | |
| 607 | cell->newState = 0; |
| 608 | cell->symbol = symbolValue; |
| 609 | cell->nbBits = 0; |
| 610 | |
| 611 | return 0; |
| 612 | } |
| 613 | |
| 614 | |
| 615 | static size_t FSE_buildDTable_raw (FSE_DTable* dt, unsigned nbBits) |
| 616 | { |
| 617 | void* ptr = dt; |
| 618 | FSE_DTableHeader* const DTableH = (FSE_DTableHeader*)ptr; |
| 619 | FSE_decode_t* const dinfo = (FSE_decode_t*)(ptr) + 1; /* because dt is unsigned */ |
| 620 | const unsigned tableSize = 1 << nbBits; |
| 621 | const unsigned tableMask = tableSize - 1; |
| 622 | const unsigned maxSymbolValue = tableMask; |
| 623 | unsigned s; |
| 624 | |
| 625 | /* Sanity checks */ |
| 626 | if (nbBits < 1) return (size_t)-FSE_ERROR_GENERIC; /* min size */ |
| 627 | |
| 628 | /* Build Decoding Table */ |
| 629 | DTableH->tableLog = (U16)nbBits; |
| 630 | DTableH->fastMode = 1; |
| 631 | for (s=0; s<=maxSymbolValue; s++) |
| 632 | { |
| 633 | dinfo[s].newState = 0; |
| 634 | dinfo[s].symbol = (BYTE)s; |
| 635 | dinfo[s].nbBits = (BYTE)nbBits; |
| 636 | } |
| 637 | |
| 638 | return 0; |
| 639 | } |
| 640 | |
| 641 | |
| 642 | /* FSE_initDStream |
| 643 | * Initialize a FSE_DStream_t. |
| 644 | * srcBuffer must point at the beginning of an FSE block. |
| 645 | * The function result is the size of the FSE_block (== srcSize). |
| 646 | * If srcSize is too small, the function will return an errorCode; |
| 647 | */ |
| 648 | static size_t FSE_initDStream(FSE_DStream_t* bitD, const void* srcBuffer, size_t srcSize) |
| 649 | { |
| 650 | if (srcSize < 1) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 651 | |
| 652 | if (srcSize >= sizeof(size_t)) |
| 653 | { |
| 654 | U32 contain32; |
| 655 | bitD->start = (const char*)srcBuffer; |
| 656 | bitD->ptr = (const char*)srcBuffer + srcSize - sizeof(size_t); |
| 657 | bitD->bitContainer = FSE_readLEST(bitD->ptr); |
| 658 | contain32 = ((const BYTE*)srcBuffer)[srcSize-1]; |
| 659 | if (contain32 == 0) return (size_t)-FSE_ERROR_GENERIC; /* stop bit not present */ |
| 660 | bitD->bitsConsumed = 8 - FSE_highbit32(contain32); |
| 661 | } |
| 662 | else |
| 663 | { |
| 664 | U32 contain32; |
| 665 | bitD->start = (const char*)srcBuffer; |
| 666 | bitD->ptr = bitD->start; |
| 667 | bitD->bitContainer = *(const BYTE*)(bitD->start); |
| 668 | switch(srcSize) |
| 669 | { |
| 670 | case 7: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[6]) << (sizeof(size_t)*8 - 16); |
| 671 | case 6: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[5]) << (sizeof(size_t)*8 - 24); |
| 672 | case 5: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[4]) << (sizeof(size_t)*8 - 32); |
| 673 | case 4: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[3]) << 24; |
| 674 | case 3: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[2]) << 16; |
| 675 | case 2: bitD->bitContainer += (size_t)(((const BYTE*)(bitD->start))[1]) << 8; |
| 676 | default:; |
| 677 | } |
| 678 | contain32 = ((const BYTE*)srcBuffer)[srcSize-1]; |
| 679 | if (contain32 == 0) return (size_t)-FSE_ERROR_GENERIC; /* stop bit not present */ |
| 680 | bitD->bitsConsumed = 8 - FSE_highbit32(contain32); |
| 681 | bitD->bitsConsumed += (U32)(sizeof(size_t) - srcSize)*8; |
| 682 | } |
| 683 | |
| 684 | return srcSize; |
| 685 | } |
| 686 | |
| 687 | |
| 688 | /*!FSE_lookBits |
| 689 | * Provides next n bits from the bitContainer. |
| 690 | * bitContainer is not modified (bits are still present for next read/look) |
| 691 | * On 32-bits, maxNbBits==25 |
| 692 | * On 64-bits, maxNbBits==57 |
| 693 | * return : value extracted. |
| 694 | */ |
| 695 | static size_t FSE_lookBits(FSE_DStream_t* bitD, U32 nbBits) |
| 696 | { |
| 697 | const U32 bitMask = sizeof(bitD->bitContainer)*8 - 1; |
| 698 | return ((bitD->bitContainer << (bitD->bitsConsumed & bitMask)) >> 1) >> ((bitMask-nbBits) & bitMask); |
| 699 | } |
| 700 | |
| 701 | static size_t FSE_lookBitsFast(FSE_DStream_t* bitD, U32 nbBits) /* only if nbBits >= 1 !! */ |
| 702 | { |
| 703 | const U32 bitMask = sizeof(bitD->bitContainer)*8 - 1; |
| 704 | return (bitD->bitContainer << (bitD->bitsConsumed & bitMask)) >> (((bitMask+1)-nbBits) & bitMask); |
| 705 | } |
| 706 | |
| 707 | static void FSE_skipBits(FSE_DStream_t* bitD, U32 nbBits) |
| 708 | { |
| 709 | bitD->bitsConsumed += nbBits; |
| 710 | } |
| 711 | |
| 712 | |
| 713 | /*!FSE_readBits |
| 714 | * Read next n bits from the bitContainer. |
| 715 | * On 32-bits, don't read more than maxNbBits==25 |
| 716 | * On 64-bits, don't read more than maxNbBits==57 |
| 717 | * Use the fast variant *only* if n >= 1. |
| 718 | * return : value extracted. |
| 719 | */ |
| 720 | static size_t FSE_readBits(FSE_DStream_t* bitD, U32 nbBits) |
| 721 | { |
| 722 | size_t value = FSE_lookBits(bitD, nbBits); |
| 723 | FSE_skipBits(bitD, nbBits); |
| 724 | return value; |
| 725 | } |
| 726 | |
| 727 | static size_t FSE_readBitsFast(FSE_DStream_t* bitD, U32 nbBits) /* only if nbBits >= 1 !! */ |
| 728 | { |
| 729 | size_t value = FSE_lookBitsFast(bitD, nbBits); |
| 730 | FSE_skipBits(bitD, nbBits); |
| 731 | return value; |
| 732 | } |
| 733 | |
| 734 | static unsigned FSE_reloadDStream(FSE_DStream_t* bitD) |
| 735 | { |
| 736 | if (bitD->bitsConsumed > (sizeof(bitD->bitContainer)*8)) /* should never happen */ |
| 737 | return FSE_DStream_tooFar; |
| 738 | |
| 739 | if (bitD->ptr >= bitD->start + sizeof(bitD->bitContainer)) |
| 740 | { |
| 741 | bitD->ptr -= bitD->bitsConsumed >> 3; |
| 742 | bitD->bitsConsumed &= 7; |
| 743 | bitD->bitContainer = FSE_readLEST(bitD->ptr); |
| 744 | return FSE_DStream_unfinished; |
| 745 | } |
| 746 | if (bitD->ptr == bitD->start) |
| 747 | { |
| 748 | if (bitD->bitsConsumed < sizeof(bitD->bitContainer)*8) return FSE_DStream_endOfBuffer; |
| 749 | return FSE_DStream_completed; |
| 750 | } |
| 751 | { |
| 752 | U32 nbBytes = bitD->bitsConsumed >> 3; |
| 753 | U32 result = FSE_DStream_unfinished; |
| 754 | if (bitD->ptr - nbBytes < bitD->start) |
| 755 | { |
| 756 | nbBytes = (U32)(bitD->ptr - bitD->start); /* ptr > start */ |
| 757 | result = FSE_DStream_endOfBuffer; |
| 758 | } |
| 759 | bitD->ptr -= nbBytes; |
| 760 | bitD->bitsConsumed -= nbBytes*8; |
| 761 | bitD->bitContainer = FSE_readLEST(bitD->ptr); /* reminder : srcSize > sizeof(bitD) */ |
| 762 | return result; |
| 763 | } |
| 764 | } |
| 765 | |
| 766 | |
| 767 | static void FSE_initDState(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD, const FSE_DTable* dt) |
| 768 | { |
| 769 | const void* ptr = dt; |
| 770 | const FSE_DTableHeader* const DTableH = (const FSE_DTableHeader*)ptr; |
| 771 | DStatePtr->state = FSE_readBits(bitD, DTableH->tableLog); |
| 772 | FSE_reloadDStream(bitD); |
| 773 | DStatePtr->table = dt + 1; |
| 774 | } |
| 775 | |
| 776 | static BYTE FSE_decodeSymbol(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD) |
| 777 | { |
| 778 | const FSE_decode_t DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state]; |
| 779 | const U32 nbBits = DInfo.nbBits; |
| 780 | BYTE symbol = DInfo.symbol; |
| 781 | size_t lowBits = FSE_readBits(bitD, nbBits); |
| 782 | |
| 783 | DStatePtr->state = DInfo.newState + lowBits; |
| 784 | return symbol; |
| 785 | } |
| 786 | |
| 787 | static BYTE FSE_decodeSymbolFast(FSE_DState_t* DStatePtr, FSE_DStream_t* bitD) |
| 788 | { |
| 789 | const FSE_decode_t DInfo = ((const FSE_decode_t*)(DStatePtr->table))[DStatePtr->state]; |
| 790 | const U32 nbBits = DInfo.nbBits; |
| 791 | BYTE symbol = DInfo.symbol; |
| 792 | size_t lowBits = FSE_readBitsFast(bitD, nbBits); |
| 793 | |
| 794 | DStatePtr->state = DInfo.newState + lowBits; |
| 795 | return symbol; |
| 796 | } |
| 797 | |
| 798 | /* FSE_endOfDStream |
| 799 | Tells if bitD has reached end of bitStream or not */ |
| 800 | |
| 801 | static unsigned FSE_endOfDStream(const FSE_DStream_t* bitD) |
| 802 | { |
| 803 | return ((bitD->ptr == bitD->start) && (bitD->bitsConsumed == sizeof(bitD->bitContainer)*8)); |
| 804 | } |
| 805 | |
| 806 | static unsigned FSE_endOfDState(const FSE_DState_t* DStatePtr) |
| 807 | { |
| 808 | return DStatePtr->state == 0; |
| 809 | } |
| 810 | |
| 811 | |
| 812 | FORCE_INLINE size_t FSE_decompress_usingDTable_generic( |
| 813 | void* dst, size_t maxDstSize, |
| 814 | const void* cSrc, size_t cSrcSize, |
| 815 | const FSE_DTable* dt, const unsigned fast) |
| 816 | { |
| 817 | BYTE* const ostart = (BYTE*) dst; |
| 818 | BYTE* op = ostart; |
| 819 | BYTE* const omax = op + maxDstSize; |
| 820 | BYTE* const olimit = omax-3; |
| 821 | |
| 822 | FSE_DStream_t bitD; |
| 823 | FSE_DState_t state1; |
| 824 | FSE_DState_t state2; |
| 825 | size_t errorCode; |
| 826 | |
| 827 | /* Init */ |
| 828 | errorCode = FSE_initDStream(&bitD, cSrc, cSrcSize); /* replaced last arg by maxCompressed Size */ |
| 829 | if (FSE_isError(errorCode)) return errorCode; |
| 830 | |
| 831 | FSE_initDState(&state1, &bitD, dt); |
| 832 | FSE_initDState(&state2, &bitD, dt); |
| 833 | |
| 834 | #define FSE_GETSYMBOL(statePtr) fast ? FSE_decodeSymbolFast(statePtr, &bitD) : FSE_decodeSymbol(statePtr, &bitD) |
| 835 | |
| 836 | /* 4 symbols per loop */ |
| 837 | for ( ; (FSE_reloadDStream(&bitD)==FSE_DStream_unfinished) && (op<olimit) ; op+=4) |
| 838 | { |
| 839 | op[0] = FSE_GETSYMBOL(&state1); |
| 840 | |
| 841 | if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */ |
| 842 | FSE_reloadDStream(&bitD); |
| 843 | |
| 844 | op[1] = FSE_GETSYMBOL(&state2); |
| 845 | |
| 846 | if (FSE_MAX_TABLELOG*4+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */ |
| 847 | { if (FSE_reloadDStream(&bitD) > FSE_DStream_unfinished) { op+=2; break; } } |
| 848 | |
| 849 | op[2] = FSE_GETSYMBOL(&state1); |
| 850 | |
| 851 | if (FSE_MAX_TABLELOG*2+7 > sizeof(bitD.bitContainer)*8) /* This test must be static */ |
| 852 | FSE_reloadDStream(&bitD); |
| 853 | |
| 854 | op[3] = FSE_GETSYMBOL(&state2); |
| 855 | } |
| 856 | |
| 857 | /* tail */ |
| 858 | /* note : FSE_reloadDStream(&bitD) >= FSE_DStream_partiallyFilled; Ends at exactly FSE_DStream_completed */ |
| 859 | while (1) |
| 860 | { |
| 861 | if ( (FSE_reloadDStream(&bitD)>FSE_DStream_completed) || (op==omax) || (FSE_endOfDStream(&bitD) && (fast || FSE_endOfDState(&state1))) ) |
| 862 | break; |
| 863 | |
| 864 | *op++ = FSE_GETSYMBOL(&state1); |
| 865 | |
| 866 | if ( (FSE_reloadDStream(&bitD)>FSE_DStream_completed) || (op==omax) || (FSE_endOfDStream(&bitD) && (fast || FSE_endOfDState(&state2))) ) |
| 867 | break; |
| 868 | |
| 869 | *op++ = FSE_GETSYMBOL(&state2); |
| 870 | } |
| 871 | |
| 872 | /* end ? */ |
| 873 | if (FSE_endOfDStream(&bitD) && FSE_endOfDState(&state1) && FSE_endOfDState(&state2)) |
| 874 | return op-ostart; |
| 875 | |
| 876 | if (op==omax) return (size_t)-FSE_ERROR_dstSize_tooSmall; /* dst buffer is full, but cSrc unfinished */ |
| 877 | |
| 878 | return (size_t)-FSE_ERROR_corruptionDetected; |
| 879 | } |
| 880 | |
| 881 | |
| 882 | static size_t FSE_decompress_usingDTable(void* dst, size_t originalSize, |
| 883 | const void* cSrc, size_t cSrcSize, |
| 884 | const FSE_DTable* dt) |
| 885 | { |
| 886 | FSE_DTableHeader DTableH; |
| 887 | memcpy(&DTableH, dt, sizeof(DTableH)); /* memcpy() into local variable, to avoid strict aliasing warning */ |
| 888 | |
| 889 | /* select fast mode (static) */ |
| 890 | if (DTableH.fastMode) return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 1); |
| 891 | return FSE_decompress_usingDTable_generic(dst, originalSize, cSrc, cSrcSize, dt, 0); |
| 892 | } |
| 893 | |
| 894 | |
| 895 | static size_t FSE_decompress(void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize) |
| 896 | { |
| 897 | const BYTE* const istart = (const BYTE*)cSrc; |
| 898 | const BYTE* ip = istart; |
| 899 | short counting[FSE_MAX_SYMBOL_VALUE+1]; |
| 900 | DTable_max_t dt; /* Static analyzer seems unable to understand this table will be properly initialized later */ |
| 901 | unsigned tableLog; |
| 902 | unsigned maxSymbolValue = FSE_MAX_SYMBOL_VALUE; |
| 903 | size_t errorCode; |
| 904 | |
| 905 | if (cSrcSize<2) return (size_t)-FSE_ERROR_srcSize_wrong; /* too small input size */ |
| 906 | |
| 907 | /* normal FSE decoding mode */ |
| 908 | errorCode = FSE_readNCount (counting, &maxSymbolValue, &tableLog, istart, cSrcSize); |
| 909 | if (FSE_isError(errorCode)) return errorCode; |
| 910 | if (errorCode >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong; /* too small input size */ |
| 911 | ip += errorCode; |
| 912 | cSrcSize -= errorCode; |
| 913 | |
| 914 | errorCode = FSE_buildDTable (dt, counting, maxSymbolValue, tableLog); |
| 915 | if (FSE_isError(errorCode)) return errorCode; |
| 916 | |
| 917 | /* always return, even if it is an error code */ |
| 918 | return FSE_decompress_usingDTable (dst, maxDstSize, ip, cSrcSize, dt); |
| 919 | } |
| 920 | |
| 921 | |
| 922 | |
| 923 | /* ******************************************************* |
| 924 | * Huff0 : Huffman block compression |
| 925 | *********************************************************/ |
| 926 | #define HUF_MAX_SYMBOL_VALUE 255 |
| 927 | #define HUF_DEFAULT_TABLELOG 12 /* used by default, when not specified */ |
| 928 | #define HUF_MAX_TABLELOG 12 /* max possible tableLog; for allocation purpose; can be modified */ |
| 929 | #define HUF_ABSOLUTEMAX_TABLELOG 16 /* absolute limit of HUF_MAX_TABLELOG. Beyond that value, code does not work */ |
| 930 | #if (HUF_MAX_TABLELOG > HUF_ABSOLUTEMAX_TABLELOG) |
| 931 | # error "HUF_MAX_TABLELOG is too large !" |
| 932 | #endif |
| 933 | |
| 934 | typedef struct HUF_CElt_s { |
| 935 | U16 val; |
| 936 | BYTE nbBits; |
| 937 | } HUF_CElt ; |
| 938 | |
| 939 | typedef struct nodeElt_s { |
| 940 | U32 count; |
| 941 | U16 parent; |
| 942 | BYTE byte; |
| 943 | BYTE nbBits; |
| 944 | } nodeElt; |
| 945 | |
| 946 | |
| 947 | /* ******************************************************* |
| 948 | * Huff0 : Huffman block decompression |
| 949 | *********************************************************/ |
| 950 | typedef struct { |
| 951 | BYTE byte; |
| 952 | BYTE nbBits; |
| 953 | } HUF_DElt; |
| 954 | |
| 955 | static size_t HUF_readDTable (U16* DTable, const void* src, size_t srcSize) |
| 956 | { |
| 957 | BYTE huffWeight[HUF_MAX_SYMBOL_VALUE + 1]; |
| 958 | U32 rankVal[HUF_ABSOLUTEMAX_TABLELOG + 1]; /* large enough for values from 0 to 16 */ |
| 959 | U32 weightTotal; |
| 960 | U32 maxBits; |
| 961 | const BYTE* ip = (const BYTE*) src; |
| 962 | size_t iSize; |
| 963 | size_t oSize; |
| 964 | U32 n; |
| 965 | U32 nextRankStart; |
| 966 | void* ptr = DTable+1; |
| 967 | HUF_DElt* const dt = (HUF_DElt*)ptr; |
| 968 | |
| 969 | if (!srcSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 970 | iSize = ip[0]; |
| 971 | |
| 972 | FSE_STATIC_ASSERT(sizeof(HUF_DElt) == sizeof(U16)); /* if compilation fails here, assertion is false */ |
| 973 | //memset(huffWeight, 0, sizeof(huffWeight)); /* should not be necessary, but some analyzer complain ... */ |
| 974 | if (iSize >= 128) /* special header */ |
| 975 | { |
| 976 | if (iSize >= (242)) /* RLE */ |
| 977 | { |
| 978 | static int l[14] = { 1, 2, 3, 4, 7, 8, 15, 16, 31, 32, 63, 64, 127, 128 }; |
| 979 | oSize = l[iSize-242]; |
| 980 | memset(huffWeight, 1, sizeof(huffWeight)); |
| 981 | iSize = 0; |
| 982 | } |
| 983 | else /* Incompressible */ |
| 984 | { |
| 985 | oSize = iSize - 127; |
| 986 | iSize = ((oSize+1)/2); |
| 987 | if (iSize+1 > srcSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 988 | ip += 1; |
| 989 | for (n=0; n<oSize; n+=2) |
| 990 | { |
| 991 | huffWeight[n] = ip[n/2] >> 4; |
| 992 | huffWeight[n+1] = ip[n/2] & 15; |
| 993 | } |
| 994 | } |
| 995 | } |
| 996 | else /* header compressed with FSE (normal case) */ |
| 997 | { |
| 998 | if (iSize+1 > srcSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 999 | oSize = FSE_decompress(huffWeight, HUF_MAX_SYMBOL_VALUE, ip+1, iSize); /* max 255 values decoded, last one is implied */ |
| 1000 | if (FSE_isError(oSize)) return oSize; |
| 1001 | } |
| 1002 | |
| 1003 | /* collect weight stats */ |
| 1004 | memset(rankVal, 0, sizeof(rankVal)); |
| 1005 | weightTotal = 0; |
| 1006 | for (n=0; n<oSize; n++) |
| 1007 | { |
| 1008 | if (huffWeight[n] >= HUF_ABSOLUTEMAX_TABLELOG) return (size_t)-FSE_ERROR_corruptionDetected; |
| 1009 | rankVal[huffWeight[n]]++; |
| 1010 | weightTotal += (1 << huffWeight[n]) >> 1; |
| 1011 | } |
| 1012 | if (weightTotal == 0) return (size_t)-FSE_ERROR_corruptionDetected; |
| 1013 | |
| 1014 | /* get last non-null symbol weight (implied, total must be 2^n) */ |
| 1015 | maxBits = FSE_highbit32(weightTotal) + 1; |
| 1016 | if (maxBits > DTable[0]) return (size_t)-FSE_ERROR_tableLog_tooLarge; /* DTable is too small */ |
| 1017 | DTable[0] = (U16)maxBits; |
| 1018 | { |
| 1019 | U32 total = 1 << maxBits; |
| 1020 | U32 rest = total - weightTotal; |
| 1021 | U32 verif = 1 << FSE_highbit32(rest); |
| 1022 | U32 lastWeight = FSE_highbit32(rest) + 1; |
| 1023 | if (verif != rest) return (size_t)-FSE_ERROR_corruptionDetected; /* last value must be a clean power of 2 */ |
| 1024 | huffWeight[oSize] = (BYTE)lastWeight; |
| 1025 | rankVal[lastWeight]++; |
| 1026 | } |
| 1027 | |
| 1028 | /* check tree construction validity */ |
| 1029 | if ((rankVal[1] < 2) || (rankVal[1] & 1)) return (size_t)-FSE_ERROR_corruptionDetected; /* by construction : at least 2 elts of rank 1, must be even */ |
| 1030 | |
| 1031 | /* Prepare ranks */ |
| 1032 | nextRankStart = 0; |
| 1033 | for (n=1; n<=maxBits; n++) |
| 1034 | { |
| 1035 | U32 current = nextRankStart; |
| 1036 | nextRankStart += (rankVal[n] << (n-1)); |
| 1037 | rankVal[n] = current; |
| 1038 | } |
| 1039 | |
| 1040 | /* fill DTable */ |
| 1041 | for (n=0; n<=oSize; n++) |
| 1042 | { |
| 1043 | const U32 w = huffWeight[n]; |
| 1044 | const U32 length = (1 << w) >> 1; |
| 1045 | U32 i; |
| 1046 | HUF_DElt D; |
| 1047 | D.byte = (BYTE)n; D.nbBits = (BYTE)(maxBits + 1 - w); |
| 1048 | for (i = rankVal[w]; i < rankVal[w] + length; i++) |
| 1049 | dt[i] = D; |
| 1050 | rankVal[w] += length; |
| 1051 | } |
| 1052 | |
| 1053 | return iSize+1; |
| 1054 | } |
| 1055 | |
| 1056 | |
| 1057 | static BYTE HUF_decodeSymbol(FSE_DStream_t* Dstream, const HUF_DElt* dt, const U32 dtLog) |
| 1058 | { |
| 1059 | const size_t val = FSE_lookBitsFast(Dstream, dtLog); /* note : dtLog >= 1 */ |
| 1060 | const BYTE c = dt[val].byte; |
| 1061 | FSE_skipBits(Dstream, dt[val].nbBits); |
| 1062 | return c; |
| 1063 | } |
| 1064 | |
| 1065 | static size_t HUF_decompress_usingDTable( /* -3% slower when non static */ |
| 1066 | void* dst, size_t maxDstSize, |
| 1067 | const void* cSrc, size_t cSrcSize, |
| 1068 | const U16* DTable) |
| 1069 | { |
| 1070 | BYTE* const ostart = (BYTE*) dst; |
| 1071 | BYTE* op = ostart; |
| 1072 | BYTE* const omax = op + maxDstSize; |
| 1073 | BYTE* const olimit = omax-15; |
| 1074 | |
| 1075 | const void* ptr = DTable; |
| 1076 | const HUF_DElt* const dt = (const HUF_DElt*)(ptr)+1; |
| 1077 | const U32 dtLog = DTable[0]; |
| 1078 | size_t errorCode; |
| 1079 | U32 reloadStatus; |
| 1080 | |
| 1081 | /* Init */ |
| 1082 | |
| 1083 | const U16* jumpTable = (const U16*)cSrc; |
| 1084 | const size_t length1 = FSE_readLE16(jumpTable); |
| 1085 | const size_t length2 = FSE_readLE16(jumpTable+1); |
| 1086 | const size_t length3 = FSE_readLE16(jumpTable+2); |
| 1087 | const size_t length4 = cSrcSize - 6 - length1 - length2 - length3; // check coherency !! |
| 1088 | const char* const start1 = (const char*)(cSrc) + 6; |
| 1089 | const char* const start2 = start1 + length1; |
| 1090 | const char* const start3 = start2 + length2; |
| 1091 | const char* const start4 = start3 + length3; |
| 1092 | FSE_DStream_t bitD1, bitD2, bitD3, bitD4; |
| 1093 | |
| 1094 | if (length1+length2+length3+6 >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 1095 | |
| 1096 | errorCode = FSE_initDStream(&bitD1, start1, length1); |
| 1097 | if (FSE_isError(errorCode)) return errorCode; |
| 1098 | errorCode = FSE_initDStream(&bitD2, start2, length2); |
| 1099 | if (FSE_isError(errorCode)) return errorCode; |
| 1100 | errorCode = FSE_initDStream(&bitD3, start3, length3); |
| 1101 | if (FSE_isError(errorCode)) return errorCode; |
| 1102 | errorCode = FSE_initDStream(&bitD4, start4, length4); |
| 1103 | if (FSE_isError(errorCode)) return errorCode; |
| 1104 | |
| 1105 | reloadStatus=FSE_reloadDStream(&bitD2); |
| 1106 | |
| 1107 | /* 16 symbols per loop */ |
| 1108 | for ( ; (reloadStatus<FSE_DStream_completed) && (op<olimit); /* D2-3-4 are supposed to be synchronized and finish together */ |
| 1109 | op+=16, reloadStatus = FSE_reloadDStream(&bitD2) | FSE_reloadDStream(&bitD3) | FSE_reloadDStream(&bitD4), FSE_reloadDStream(&bitD1)) |
| 1110 | { |
| 1111 | #define HUF_DECODE_SYMBOL_0(n, Dstream) \ |
| 1112 | op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog); |
| 1113 | |
| 1114 | #define HUF_DECODE_SYMBOL_1(n, Dstream) \ |
| 1115 | op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog); \ |
| 1116 | if (FSE_32bits() && (HUF_MAX_TABLELOG>12)) FSE_reloadDStream(&Dstream) |
| 1117 | |
| 1118 | #define HUF_DECODE_SYMBOL_2(n, Dstream) \ |
| 1119 | op[n] = HUF_decodeSymbol(&Dstream, dt, dtLog); \ |
| 1120 | if (FSE_32bits()) FSE_reloadDStream(&Dstream) |
| 1121 | |
| 1122 | HUF_DECODE_SYMBOL_1( 0, bitD1); |
| 1123 | HUF_DECODE_SYMBOL_1( 1, bitD2); |
| 1124 | HUF_DECODE_SYMBOL_1( 2, bitD3); |
| 1125 | HUF_DECODE_SYMBOL_1( 3, bitD4); |
| 1126 | HUF_DECODE_SYMBOL_2( 4, bitD1); |
| 1127 | HUF_DECODE_SYMBOL_2( 5, bitD2); |
| 1128 | HUF_DECODE_SYMBOL_2( 6, bitD3); |
| 1129 | HUF_DECODE_SYMBOL_2( 7, bitD4); |
| 1130 | HUF_DECODE_SYMBOL_1( 8, bitD1); |
| 1131 | HUF_DECODE_SYMBOL_1( 9, bitD2); |
| 1132 | HUF_DECODE_SYMBOL_1(10, bitD3); |
| 1133 | HUF_DECODE_SYMBOL_1(11, bitD4); |
| 1134 | HUF_DECODE_SYMBOL_0(12, bitD1); |
| 1135 | HUF_DECODE_SYMBOL_0(13, bitD2); |
| 1136 | HUF_DECODE_SYMBOL_0(14, bitD3); |
| 1137 | HUF_DECODE_SYMBOL_0(15, bitD4); |
| 1138 | } |
| 1139 | |
| 1140 | if (reloadStatus!=FSE_DStream_completed) /* not complete : some bitStream might be FSE_DStream_unfinished */ |
| 1141 | return (size_t)-FSE_ERROR_corruptionDetected; |
| 1142 | |
| 1143 | /* tail */ |
| 1144 | { |
| 1145 | // bitTail = bitD1; // *much* slower : -20% !??! |
| 1146 | FSE_DStream_t bitTail; |
| 1147 | bitTail.ptr = bitD1.ptr; |
| 1148 | bitTail.bitsConsumed = bitD1.bitsConsumed; |
| 1149 | bitTail.bitContainer = bitD1.bitContainer; // required in case of FSE_DStream_endOfBuffer |
| 1150 | bitTail.start = start1; |
| 1151 | for ( ; (FSE_reloadDStream(&bitTail) < FSE_DStream_completed) && (op<omax) ; op++) |
| 1152 | { |
| 1153 | HUF_DECODE_SYMBOL_0(0, bitTail); |
| 1154 | } |
| 1155 | |
| 1156 | if (FSE_endOfDStream(&bitTail)) |
| 1157 | return op-ostart; |
| 1158 | } |
| 1159 | |
| 1160 | if (op==omax) return (size_t)-FSE_ERROR_dstSize_tooSmall; /* dst buffer is full, but cSrc unfinished */ |
| 1161 | |
| 1162 | return (size_t)-FSE_ERROR_corruptionDetected; |
| 1163 | } |
| 1164 | |
| 1165 | |
| 1166 | static size_t HUF_decompress (void* dst, size_t maxDstSize, const void* cSrc, size_t cSrcSize) |
| 1167 | { |
| 1168 | HUF_CREATE_STATIC_DTABLE(DTable, HUF_MAX_TABLELOG); |
| 1169 | const BYTE* ip = (const BYTE*) cSrc; |
| 1170 | size_t errorCode; |
| 1171 | |
| 1172 | errorCode = HUF_readDTable (DTable, cSrc, cSrcSize); |
| 1173 | if (FSE_isError(errorCode)) return errorCode; |
| 1174 | if (errorCode >= cSrcSize) return (size_t)-FSE_ERROR_srcSize_wrong; |
| 1175 | ip += errorCode; |
| 1176 | cSrcSize -= errorCode; |
| 1177 | |
| 1178 | return HUF_decompress_usingDTable (dst, maxDstSize, ip, cSrcSize, DTable); |
| 1179 | } |
| 1180 | |
| 1181 | |
| 1182 | #endif /* FSE_COMMONDEFS_ONLY */ |
| 1183 | |
| 1184 | /* |
| 1185 | zstd - standard compression library |
| 1186 | Copyright (C) 2014-2015, Yann Collet. |
| 1187 | |
| 1188 | BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php) |
| 1189 | |
| 1190 | Redistribution and use in source and binary forms, with or without |
| 1191 | modification, are permitted provided that the following conditions are |
| 1192 | met: |
| 1193 | * Redistributions of source code must retain the above copyright |
| 1194 | notice, this list of conditions and the following disclaimer. |
| 1195 | * Redistributions in binary form must reproduce the above |
| 1196 | copyright notice, this list of conditions and the following disclaimer |
| 1197 | in the documentation and/or other materials provided with the |
| 1198 | distribution. |
| 1199 | THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| 1200 | "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| 1201 | LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| 1202 | A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| 1203 | OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| 1204 | SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| 1205 | LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| 1206 | DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| 1207 | THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| 1208 | (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| 1209 | OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 1210 | |
| 1211 | You can contact the author at : |
| 1212 | - zstd source repository : https://github.com/Cyan4973/zstd |
| 1213 | - ztsd public forum : https://groups.google.com/forum/#!forum/lz4c |
| 1214 | */ |
| 1215 | |
| 1216 | /**************************************************************** |
| 1217 | * Tuning parameters |
| 1218 | *****************************************************************/ |
| 1219 | /* MEMORY_USAGE : |
| 1220 | * Memory usage formula : N->2^N Bytes (examples : 10 -> 1KB; 12 -> 4KB ; 16 -> 64KB; 20 -> 1MB; etc.) |
| 1221 | * Increasing memory usage improves compression ratio |
| 1222 | * Reduced memory usage can improve speed, due to cache effect */ |
| 1223 | #define ZSTD_MEMORY_USAGE 17 |
| 1224 | |
| 1225 | |
| 1226 | /************************************** |
| 1227 | CPU Feature Detection |
| 1228 | **************************************/ |
| 1229 | /* |
| 1230 | * Automated efficient unaligned memory access detection |
| 1231 | * Based on known hardware architectures |
| 1232 | * This list will be updated thanks to feedbacks |
| 1233 | */ |
| 1234 | #if defined(CPU_HAS_EFFICIENT_UNALIGNED_MEMORY_ACCESS) \ |
| 1235 | || defined(__ARM_FEATURE_UNALIGNED) \ |
| 1236 | || defined(__i386__) || defined(__x86_64__) \ |
| 1237 | || defined(_M_IX86) || defined(_M_X64) \ |
| 1238 | || defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_8__) \ |
| 1239 | || (defined(_M_ARM) && (_M_ARM >= 7)) |
| 1240 | # define ZSTD_UNALIGNED_ACCESS 1 |
| 1241 | #else |
| 1242 | # define ZSTD_UNALIGNED_ACCESS 0 |
| 1243 | #endif |
| 1244 | |
| 1245 | |
| 1246 | /******************************************************** |
| 1247 | * Includes |
| 1248 | *********************************************************/ |
| 1249 | #include <stdlib.h> /* calloc */ |
| 1250 | #include <string.h> /* memcpy, memmove */ |
| 1251 | #include <stdio.h> /* debug : printf */ |
| 1252 | |
| 1253 | |
| 1254 | /******************************************************** |
| 1255 | * Compiler specifics |
| 1256 | *********************************************************/ |
| 1257 | #ifdef __AVX2__ |
| 1258 | # include <immintrin.h> /* AVX2 intrinsics */ |
| 1259 | #endif |
| 1260 | |
| 1261 | #ifdef _MSC_VER /* Visual Studio */ |
| 1262 | # include <intrin.h> /* For Visual 2005 */ |
| 1263 | # pragma warning(disable : 4127) /* disable: C4127: conditional expression is constant */ |
| 1264 | # pragma warning(disable : 4324) /* disable: C4324: padded structure */ |
| 1265 | #endif |
| 1266 | |
| 1267 | |
| 1268 | #ifndef MEM_ACCESS_MODULE |
| 1269 | #define MEM_ACCESS_MODULE |
| 1270 | /******************************************************** |
| 1271 | * Basic Types |
| 1272 | *********************************************************/ |
| 1273 | #if defined (__STDC_VERSION__) && __STDC_VERSION__ >= 199901L /* C99 */ |
| 1274 | # include <stdint.h> |
| 1275 | typedef uint8_t BYTE; |
| 1276 | typedef uint16_t U16; |
| 1277 | typedef int16_t S16; |
| 1278 | typedef uint32_t U32; |
| 1279 | typedef int32_t S32; |
| 1280 | typedef uint64_t U64; |
| 1281 | #else |
| 1282 | typedef unsigned char BYTE; |
| 1283 | typedef unsigned short U16; |
| 1284 | typedef signed short S16; |
| 1285 | typedef unsigned int U32; |
| 1286 | typedef signed int S32; |
| 1287 | typedef unsigned long long U64; |
| 1288 | #endif |
| 1289 | |
| 1290 | #endif /* MEM_ACCESS_MODULE */ |
| 1291 | |
| 1292 | |
| 1293 | /******************************************************** |
| 1294 | * Constants |
| 1295 | *********************************************************/ |
| 1296 | static const U32 ZSTD_magicNumber = 0xFD2FB51E; /* 3rd version : seqNb header */ |
| 1297 | |
| 1298 | #define HASH_LOG (ZSTD_MEMORY_USAGE - 2) |
| 1299 | #define HASH_TABLESIZE (1 << HASH_LOG) |
| 1300 | #define HASH_MASK (HASH_TABLESIZE - 1) |
| 1301 | |
| 1302 | #define KNUTH 2654435761 |
| 1303 | |
| 1304 | #define BIT7 128 |
| 1305 | #define BIT6 64 |
| 1306 | #define BIT5 32 |
| 1307 | #define BIT4 16 |
| 1308 | |
| 1309 | #define KB *(1 <<10) |
| 1310 | #define MB *(1 <<20) |
| 1311 | #define GB *(1U<<30) |
| 1312 | |
| 1313 | #define BLOCKSIZE (128 KB) /* define, for static allocation */ |
| 1314 | |
| 1315 | #define WORKPLACESIZE (BLOCKSIZE*3) |
| 1316 | #define MINMATCH 4 |
| 1317 | #define MLbits 7 |
| 1318 | #define LLbits 6 |
| 1319 | #define Offbits 5 |
| 1320 | #define MaxML ((1<<MLbits )-1) |
| 1321 | #define MaxLL ((1<<LLbits )-1) |
| 1322 | #define MaxOff ((1<<Offbits)-1) |
| 1323 | #define LitFSELog 11 |
| 1324 | #define MLFSELog 10 |
| 1325 | #define LLFSELog 10 |
| 1326 | #define OffFSELog 9 |
| 1327 | #define MAX(a,b) ((a)<(b)?(b):(a)) |
| 1328 | #define MaxSeq MAX(MaxLL, MaxML) |
| 1329 | |
| 1330 | #define LITERAL_NOENTROPY 63 |
| 1331 | #define COMMAND_NOENTROPY 7 /* to remove */ |
| 1332 | |
| 1333 | static const size_t ZSTD_blockHeaderSize = 3; |
| 1334 | static const size_t ZSTD_frameHeaderSize = 4; |
| 1335 | |
| 1336 | |
| 1337 | /******************************************************** |
| 1338 | * Memory operations |
| 1339 | *********************************************************/ |
| 1340 | static unsigned ZSTD_32bits(void) { return sizeof(void*)==4; } |
| 1341 | |
| 1342 | static unsigned ZSTD_isLittleEndian(void) |
| 1343 | { |
| 1344 | const union { U32 i; BYTE c[4]; } one = { 1 }; /* don't use static : performance detrimental */ |
| 1345 | return one.c[0]; |
| 1346 | } |
| 1347 | |
| 1348 | static U16 ZSTD_read16(const void* p) { U16 r; memcpy(&r, p, sizeof(r)); return r; } |
| 1349 | |
| 1350 | static U32 ZSTD_read32(const void* p) { U32 r; memcpy(&r, p, sizeof(r)); return r; } |
| 1351 | |
| 1352 | static void ZSTD_copy4(void* dst, const void* src) { memcpy(dst, src, 4); } |
| 1353 | |
| 1354 | static void ZSTD_copy8(void* dst, const void* src) { memcpy(dst, src, 8); } |
| 1355 | |
| 1356 | #define COPY8(d,s) { ZSTD_copy8(d,s); d+=8; s+=8; } |
| 1357 | |
| 1358 | static void ZSTD_wildcopy(void* dst, const void* src, ptrdiff_t length) |
| 1359 | { |
| 1360 | const BYTE* ip = (const BYTE*)src; |
| 1361 | BYTE* op = (BYTE*)dst; |
| 1362 | BYTE* const oend = op + length; |
| 1363 | while (op < oend) COPY8(op, ip); |
| 1364 | } |
| 1365 | |
| 1366 | static U16 ZSTD_readLE16(const void* memPtr) |
| 1367 | { |
| 1368 | if (ZSTD_isLittleEndian()) return ZSTD_read16(memPtr); |
| 1369 | else |
| 1370 | { |
| 1371 | const BYTE* p = (const BYTE*)memPtr; |
| 1372 | return (U16)((U16)p[0] + ((U16)p[1]<<8)); |
| 1373 | } |
| 1374 | } |
| 1375 | |
| 1376 | |
| 1377 | static U32 ZSTD_readLE32(const void* memPtr) |
| 1378 | { |
| 1379 | if (ZSTD_isLittleEndian()) |
| 1380 | return ZSTD_read32(memPtr); |
| 1381 | else |
| 1382 | { |
| 1383 | const BYTE* p = (const BYTE*)memPtr; |
| 1384 | return (U32)((U32)p[0] + ((U32)p[1]<<8) + ((U32)p[2]<<16) + ((U32)p[3]<<24)); |
| 1385 | } |
| 1386 | } |
| 1387 | |
| 1388 | static U32 ZSTD_readBE32(const void* memPtr) |
| 1389 | { |
| 1390 | const BYTE* p = (const BYTE*)memPtr; |
| 1391 | return (U32)(((U32)p[0]<<24) + ((U32)p[1]<<16) + ((U32)p[2]<<8) + ((U32)p[3]<<0)); |
| 1392 | } |
| 1393 | |
| 1394 | |
| 1395 | /************************************** |
| 1396 | * Local structures |
| 1397 | ***************************************/ |
| 1398 | typedef struct ZSTD_Cctx_s ZSTD_Cctx; |
| 1399 | |
| 1400 | typedef enum { bt_compressed, bt_raw, bt_rle, bt_end } blockType_t; |
| 1401 | |
| 1402 | typedef struct |
| 1403 | { |
| 1404 | blockType_t blockType; |
| 1405 | U32 origSize; |
| 1406 | } blockProperties_t; |
| 1407 | |
| 1408 | typedef struct { |
| 1409 | void* buffer; |
| 1410 | U32* offsetStart; |
| 1411 | U32* offset; |
| 1412 | BYTE* offCodeStart; |
| 1413 | BYTE* offCode; |
| 1414 | BYTE* litStart; |
| 1415 | BYTE* lit; |
| 1416 | BYTE* litLengthStart; |
| 1417 | BYTE* litLength; |
| 1418 | BYTE* matchLengthStart; |
| 1419 | BYTE* matchLength; |
| 1420 | BYTE* dumpsStart; |
| 1421 | BYTE* dumps; |
| 1422 | } seqStore_t; |
| 1423 | |
| 1424 | |
| 1425 | typedef struct ZSTD_Cctx_s |
| 1426 | { |
| 1427 | const BYTE* base; |
| 1428 | U32 current; |
| 1429 | U32 nextUpdate; |
| 1430 | seqStore_t seqStore; |
| 1431 | #ifdef __AVX2__ |
| 1432 | __m256i hashTable[HASH_TABLESIZE>>3]; |
| 1433 | #else |
| 1434 | U32 hashTable[HASH_TABLESIZE]; |
| 1435 | #endif |
| 1436 | BYTE buffer[WORKPLACESIZE]; |
| 1437 | } cctxi_t; |
| 1438 | |
| 1439 | |
| 1440 | |
| 1441 | |
| 1442 | /************************************** |
| 1443 | * Error Management |
| 1444 | **************************************/ |
| 1445 | /* published entry point */ |
| 1446 | unsigned ZSTDv01_isError(size_t code) { return ERR_isError(code); } |
| 1447 | |
| 1448 | |
| 1449 | /************************************** |
| 1450 | * Tool functions |
| 1451 | **************************************/ |
| 1452 | #define ZSTD_VERSION_MAJOR 0 /* for breaking interface changes */ |
| 1453 | #define ZSTD_VERSION_MINOR 1 /* for new (non-breaking) interface capabilities */ |
| 1454 | #define ZSTD_VERSION_RELEASE 3 /* for tweaks, bug-fixes, or development */ |
| 1455 | #define ZSTD_VERSION_NUMBER (ZSTD_VERSION_MAJOR *100*100 + ZSTD_VERSION_MINOR *100 + ZSTD_VERSION_RELEASE) |
| 1456 | |
| 1457 | /************************************************************** |
| 1458 | * Decompression code |
| 1459 | **************************************************************/ |
| 1460 | |
| 1461 | size_t ZSTDv01_getcBlockSize(const void* src, size_t srcSize, blockProperties_t* bpPtr) |
| 1462 | { |
| 1463 | const BYTE* const in = (const BYTE* const)src; |
| 1464 | BYTE headerFlags; |
| 1465 | U32 cSize; |
| 1466 | |
| 1467 | if (srcSize < 3) return ERROR(srcSize_wrong); |
| 1468 | |
| 1469 | headerFlags = *in; |
| 1470 | cSize = in[2] + (in[1]<<8) + ((in[0] & 7)<<16); |
| 1471 | |
| 1472 | bpPtr->blockType = (blockType_t)(headerFlags >> 6); |
| 1473 | bpPtr->origSize = (bpPtr->blockType == bt_rle) ? cSize : 0; |
| 1474 | |
| 1475 | if (bpPtr->blockType == bt_end) return 0; |
| 1476 | if (bpPtr->blockType == bt_rle) return 1; |
| 1477 | return cSize; |
| 1478 | } |
| 1479 | |
| 1480 | |
| 1481 | static size_t ZSTD_copyUncompressedBlock(void* dst, size_t maxDstSize, const void* src, size_t srcSize) |
| 1482 | { |
| 1483 | if (srcSize > maxDstSize) return ERROR(dstSize_tooSmall); |
| 1484 | memcpy(dst, src, srcSize); |
| 1485 | return srcSize; |
| 1486 | } |
| 1487 | |
| 1488 | |
| 1489 | static size_t ZSTD_decompressLiterals(void* ctx, |
| 1490 | void* dst, size_t maxDstSize, |
| 1491 | const void* src, size_t srcSize) |
| 1492 | { |
| 1493 | BYTE* op = (BYTE*)dst; |
| 1494 | BYTE* const oend = op + maxDstSize; |
| 1495 | const BYTE* ip = (const BYTE*)src; |
| 1496 | size_t errorCode; |
| 1497 | size_t litSize; |
| 1498 | |
| 1499 | /* check : minimum 2, for litSize, +1, for content */ |
| 1500 | if (srcSize <= 3) return ERROR(corruption_detected); |
| 1501 | |
| 1502 | litSize = ip[1] + (ip[0]<<8); |
| 1503 | litSize += ((ip[-3] >> 3) & 7) << 16; // mmmmh.... |
| 1504 | op = oend - litSize; |
| 1505 | |
| 1506 | (void)ctx; |
| 1507 | if (litSize > maxDstSize) return ERROR(dstSize_tooSmall); |
| 1508 | errorCode = HUF_decompress(op, litSize, ip+2, srcSize-2); |
| 1509 | if (FSE_isError(errorCode)) return ERROR(GENERIC); |
| 1510 | return litSize; |
| 1511 | } |
| 1512 | |
| 1513 | |
| 1514 | size_t ZSTDv01_decodeLiteralsBlock(void* ctx, |
| 1515 | void* dst, size_t maxDstSize, |
| 1516 | const BYTE** litStart, size_t* litSize, |
| 1517 | const void* src, size_t srcSize) |
| 1518 | { |
| 1519 | const BYTE* const istart = (const BYTE* const)src; |
| 1520 | const BYTE* ip = istart; |
| 1521 | BYTE* const ostart = (BYTE* const)dst; |
| 1522 | BYTE* const oend = ostart + maxDstSize; |
| 1523 | blockProperties_t litbp; |
| 1524 | |
| 1525 | size_t litcSize = ZSTDv01_getcBlockSize(src, srcSize, &litbp); |
| 1526 | if (ZSTDv01_isError(litcSize)) return litcSize; |
| 1527 | if (litcSize > srcSize - ZSTD_blockHeaderSize) return ERROR(srcSize_wrong); |
| 1528 | ip += ZSTD_blockHeaderSize; |
| 1529 | |
| 1530 | switch(litbp.blockType) |
| 1531 | { |
| 1532 | case bt_raw: |
| 1533 | *litStart = ip; |
| 1534 | ip += litcSize; |
| 1535 | *litSize = litcSize; |
| 1536 | break; |
| 1537 | case bt_rle: |
| 1538 | { |
| 1539 | size_t rleSize = litbp.origSize; |
| 1540 | if (rleSize>maxDstSize) return ERROR(dstSize_tooSmall); |
| 1541 | if (!srcSize) return ERROR(srcSize_wrong); |
| 1542 | memset(oend - rleSize, *ip, rleSize); |
| 1543 | *litStart = oend - rleSize; |
| 1544 | *litSize = rleSize; |
| 1545 | ip++; |
| 1546 | break; |
| 1547 | } |
| 1548 | case bt_compressed: |
| 1549 | { |
| 1550 | size_t decodedLitSize = ZSTD_decompressLiterals(ctx, dst, maxDstSize, ip, litcSize); |
| 1551 | if (ZSTDv01_isError(decodedLitSize)) return decodedLitSize; |
| 1552 | *litStart = oend - decodedLitSize; |
| 1553 | *litSize = decodedLitSize; |
| 1554 | ip += litcSize; |
| 1555 | break; |
| 1556 | } |
| 1557 | case bt_end: |
| 1558 | default: |
| 1559 | return ERROR(GENERIC); |
| 1560 | } |
| 1561 | |
| 1562 | return ip-istart; |
| 1563 | } |
| 1564 | |
| 1565 | |
| 1566 | size_t ZSTDv01_decodeSeqHeaders(int* nbSeq, const BYTE** dumpsPtr, size_t* dumpsLengthPtr, |
| 1567 | FSE_DTable* DTableLL, FSE_DTable* DTableML, FSE_DTable* DTableOffb, |
| 1568 | const void* src, size_t srcSize) |
| 1569 | { |
| 1570 | const BYTE* const istart = (const BYTE* const)src; |
| 1571 | const BYTE* ip = istart; |
| 1572 | const BYTE* const iend = istart + srcSize; |
| 1573 | U32 LLtype, Offtype, MLtype; |
| 1574 | U32 LLlog, Offlog, MLlog; |
| 1575 | size_t dumpsLength; |
| 1576 | |
| 1577 | /* check */ |
| 1578 | if (srcSize < 5) return ERROR(srcSize_wrong); |
| 1579 | |
| 1580 | /* SeqHead */ |
| 1581 | *nbSeq = ZSTD_readLE16(ip); ip+=2; |
| 1582 | LLtype = *ip >> 6; |
| 1583 | Offtype = (*ip >> 4) & 3; |
| 1584 | MLtype = (*ip >> 2) & 3; |
| 1585 | if (*ip & 2) |
| 1586 | { |
| 1587 | dumpsLength = ip[2]; |
| 1588 | dumpsLength += ip[1] << 8; |
| 1589 | ip += 3; |
| 1590 | } |
| 1591 | else |
| 1592 | { |
| 1593 | dumpsLength = ip[1]; |
| 1594 | dumpsLength += (ip[0] & 1) << 8; |
| 1595 | ip += 2; |
| 1596 | } |
| 1597 | *dumpsPtr = ip; |
| 1598 | ip += dumpsLength; |
| 1599 | *dumpsLengthPtr = dumpsLength; |
| 1600 | |
| 1601 | /* check */ |
| 1602 | if (ip > iend-3) return ERROR(srcSize_wrong); /* min : all 3 are "raw", hence no header, but at least xxLog bits per type */ |
| 1603 | |
| 1604 | /* sequences */ |
| 1605 | { |
| 1606 | S16 norm[MaxML+1]; /* assumption : MaxML >= MaxLL and MaxOff */ |
| 1607 | size_t headerSize; |
| 1608 | |
| 1609 | /* Build DTables */ |
| 1610 | switch(LLtype) |
| 1611 | { |
| 1612 | case bt_rle : |
| 1613 | LLlog = 0; |
| 1614 | FSE_buildDTable_rle(DTableLL, *ip++); break; |
| 1615 | case bt_raw : |
| 1616 | LLlog = LLbits; |
| 1617 | FSE_buildDTable_raw(DTableLL, LLbits); break; |
| 1618 | default : |
| 1619 | { U32 max = MaxLL; |
| 1620 | headerSize = FSE_readNCount(norm, &max, &LLlog, ip, iend-ip); |
| 1621 | if (FSE_isError(headerSize)) return ERROR(GENERIC); |
| 1622 | if (LLlog > LLFSELog) return ERROR(corruption_detected); |
| 1623 | ip += headerSize; |
| 1624 | FSE_buildDTable(DTableLL, norm, max, LLlog); |
| 1625 | } } |
| 1626 | |
| 1627 | switch(Offtype) |
| 1628 | { |
| 1629 | case bt_rle : |
| 1630 | Offlog = 0; |
| 1631 | if (ip > iend-2) return ERROR(srcSize_wrong); /* min : "raw", hence no header, but at least xxLog bits */ |
| 1632 | FSE_buildDTable_rle(DTableOffb, *ip++); break; |
| 1633 | case bt_raw : |
| 1634 | Offlog = Offbits; |
| 1635 | FSE_buildDTable_raw(DTableOffb, Offbits); break; |
| 1636 | default : |
| 1637 | { U32 max = MaxOff; |
| 1638 | headerSize = FSE_readNCount(norm, &max, &Offlog, ip, iend-ip); |
| 1639 | if (FSE_isError(headerSize)) return ERROR(GENERIC); |
| 1640 | if (Offlog > OffFSELog) return ERROR(corruption_detected); |
| 1641 | ip += headerSize; |
| 1642 | FSE_buildDTable(DTableOffb, norm, max, Offlog); |
| 1643 | } } |
| 1644 | |
| 1645 | switch(MLtype) |
| 1646 | { |
| 1647 | case bt_rle : |
| 1648 | MLlog = 0; |
| 1649 | if (ip > iend-2) return ERROR(srcSize_wrong); /* min : "raw", hence no header, but at least xxLog bits */ |
| 1650 | FSE_buildDTable_rle(DTableML, *ip++); break; |
| 1651 | case bt_raw : |
| 1652 | MLlog = MLbits; |
| 1653 | FSE_buildDTable_raw(DTableML, MLbits); break; |
| 1654 | default : |
| 1655 | { U32 max = MaxML; |
| 1656 | headerSize = FSE_readNCount(norm, &max, &MLlog, ip, iend-ip); |
| 1657 | if (FSE_isError(headerSize)) return ERROR(GENERIC); |
| 1658 | if (MLlog > MLFSELog) return ERROR(corruption_detected); |
| 1659 | ip += headerSize; |
| 1660 | FSE_buildDTable(DTableML, norm, max, MLlog); |
| 1661 | } } } |
| 1662 | |
| 1663 | return ip-istart; |
| 1664 | } |
| 1665 | |
| 1666 | |
| 1667 | typedef struct { |
| 1668 | size_t litLength; |
| 1669 | size_t offset; |
| 1670 | size_t matchLength; |
| 1671 | } seq_t; |
| 1672 | |
| 1673 | typedef struct { |
| 1674 | FSE_DStream_t DStream; |
| 1675 | FSE_DState_t stateLL; |
| 1676 | FSE_DState_t stateOffb; |
| 1677 | FSE_DState_t stateML; |
| 1678 | size_t prevOffset; |
| 1679 | const BYTE* dumps; |
| 1680 | const BYTE* dumpsEnd; |
| 1681 | } seqState_t; |
| 1682 | |
| 1683 | |
| 1684 | static void ZSTD_decodeSequence(seq_t* seq, seqState_t* seqState) |
| 1685 | { |
| 1686 | size_t litLength; |
| 1687 | size_t prevOffset; |
| 1688 | size_t offset; |
| 1689 | size_t matchLength; |
| 1690 | const BYTE* dumps = seqState->dumps; |
| 1691 | const BYTE* const de = seqState->dumpsEnd; |
| 1692 | |
| 1693 | /* Literal length */ |
| 1694 | litLength = FSE_decodeSymbol(&(seqState->stateLL), &(seqState->DStream)); |
| 1695 | prevOffset = litLength ? seq->offset : seqState->prevOffset; |
| 1696 | seqState->prevOffset = seq->offset; |
| 1697 | if (litLength == MaxLL) |
| 1698 | { |
| 1699 | U32 add = dumps<de ? *dumps++ : 0; |
| 1700 | if (add < 255) litLength += add; |
| 1701 | else |
| 1702 | { |
| 1703 | if (dumps<=(de-3)) |
| 1704 | { |
| 1705 | litLength = ZSTD_readLE32(dumps) & 0xFFFFFF; /* no pb : dumps is always followed by seq tables > 1 byte */ |
| 1706 | dumps += 3; |
| 1707 | } |
| 1708 | } |
| 1709 | } |
| 1710 | |
| 1711 | /* Offset */ |
| 1712 | { |
| 1713 | U32 offsetCode, nbBits; |
| 1714 | offsetCode = FSE_decodeSymbol(&(seqState->stateOffb), &(seqState->DStream)); |
| 1715 | if (ZSTD_32bits()) FSE_reloadDStream(&(seqState->DStream)); |
| 1716 | nbBits = offsetCode - 1; |
| 1717 | if (offsetCode==0) nbBits = 0; /* cmove */ |
| 1718 | offset = ((size_t)1 << (nbBits & ((sizeof(offset)*8)-1))) + FSE_readBits(&(seqState->DStream), nbBits); |
| 1719 | if (ZSTD_32bits()) FSE_reloadDStream(&(seqState->DStream)); |
| 1720 | if (offsetCode==0) offset = prevOffset; |
| 1721 | } |
| 1722 | |
| 1723 | /* MatchLength */ |
| 1724 | matchLength = FSE_decodeSymbol(&(seqState->stateML), &(seqState->DStream)); |
| 1725 | if (matchLength == MaxML) |
| 1726 | { |
| 1727 | U32 add = dumps<de ? *dumps++ : 0; |
| 1728 | if (add < 255) matchLength += add; |
| 1729 | else |
| 1730 | { |
| 1731 | if (dumps<=(de-3)) |
| 1732 | { |
| 1733 | matchLength = ZSTD_readLE32(dumps) & 0xFFFFFF; /* no pb : dumps is always followed by seq tables > 1 byte */ |
| 1734 | dumps += 3; |
| 1735 | } |
| 1736 | } |
| 1737 | } |
| 1738 | matchLength += MINMATCH; |
| 1739 | |
| 1740 | /* save result */ |
| 1741 | seq->litLength = litLength; |
| 1742 | seq->offset = offset; |
| 1743 | seq->matchLength = matchLength; |
| 1744 | seqState->dumps = dumps; |
| 1745 | } |
| 1746 | |
| 1747 | |
| 1748 | static size_t ZSTD_execSequence(BYTE* op, |
| 1749 | seq_t sequence, |
| 1750 | const BYTE** litPtr, const BYTE* const litLimit, |
| 1751 | BYTE* const base, BYTE* const oend) |
| 1752 | { |
| 1753 | static const int dec32table[] = {0, 1, 2, 1, 4, 4, 4, 4}; /* added */ |
| 1754 | static const int dec64table[] = {8, 8, 8, 7, 8, 9,10,11}; /* substracted */ |
| 1755 | const BYTE* const ostart = op; |
| 1756 | const size_t litLength = sequence.litLength; |
| 1757 | BYTE* const endMatch = op + litLength + sequence.matchLength; /* risk : address space overflow (32-bits) */ |
| 1758 | const BYTE* const litEnd = *litPtr + litLength; |
| 1759 | |
| 1760 | /* check */ |
| 1761 | if (endMatch > oend) return ERROR(dstSize_tooSmall); /* overwrite beyond dst buffer */ |
| 1762 | if (litEnd > litLimit) return ERROR(corruption_detected); |
| 1763 | if (sequence.matchLength > (size_t)(*litPtr-op)) return ERROR(dstSize_tooSmall); /* overwrite literal segment */ |
| 1764 | |
| 1765 | /* copy Literals */ |
| 1766 | if (((size_t)(*litPtr - op) < 8) || ((size_t)(oend-litEnd) < 8) || (op+litLength > oend-8)) |
| 1767 | memmove(op, *litPtr, litLength); /* overwrite risk */ |
| 1768 | else |
| 1769 | ZSTD_wildcopy(op, *litPtr, litLength); |
| 1770 | op += litLength; |
| 1771 | *litPtr = litEnd; /* update for next sequence */ |
| 1772 | |
| 1773 | /* check : last match must be at a minimum distance of 8 from end of dest buffer */ |
| 1774 | if (oend-op < 8) return ERROR(dstSize_tooSmall); |
| 1775 | |
| 1776 | /* copy Match */ |
| 1777 | { |
| 1778 | const U32 overlapRisk = (((size_t)(litEnd - endMatch)) < 12); |
| 1779 | const BYTE* match = op - sequence.offset; /* possible underflow at op - offset ? */ |
| 1780 | size_t qutt = 12; |
| 1781 | U64 saved[2]; |
| 1782 | |
| 1783 | /* check */ |
| 1784 | if (match < base) return ERROR(corruption_detected); |
| 1785 | if (sequence.offset > (size_t)base) return ERROR(corruption_detected); |
| 1786 | |
| 1787 | /* save beginning of literal sequence, in case of write overlap */ |
| 1788 | if (overlapRisk) |
| 1789 | { |
| 1790 | if ((endMatch + qutt) > oend) qutt = oend-endMatch; |
| 1791 | memcpy(saved, endMatch, qutt); |
| 1792 | } |
| 1793 | |
| 1794 | if (sequence.offset < 8) |
| 1795 | { |
| 1796 | const int dec64 = dec64table[sequence.offset]; |
| 1797 | op[0] = match[0]; |
| 1798 | op[1] = match[1]; |
| 1799 | op[2] = match[2]; |
| 1800 | op[3] = match[3]; |
| 1801 | match += dec32table[sequence.offset]; |
| 1802 | ZSTD_copy4(op+4, match); |
| 1803 | match -= dec64; |
| 1804 | } else { ZSTD_copy8(op, match); } |
| 1805 | op += 8; match += 8; |
| 1806 | |
| 1807 | if (endMatch > oend-(16-MINMATCH)) |
| 1808 | { |
| 1809 | if (op < oend-8) |
| 1810 | { |
| 1811 | ZSTD_wildcopy(op, match, (oend-8) - op); |
| 1812 | match += (oend-8) - op; |
| 1813 | op = oend-8; |
| 1814 | } |
| 1815 | while (op<endMatch) *op++ = *match++; |
| 1816 | } |
| 1817 | else |
| 1818 | ZSTD_wildcopy(op, match, (ptrdiff_t)sequence.matchLength-8); /* works even if matchLength < 8 */ |
| 1819 | |
| 1820 | /* restore, in case of overlap */ |
| 1821 | if (overlapRisk) memcpy(endMatch, saved, qutt); |
| 1822 | } |
| 1823 | |
| 1824 | return endMatch-ostart; |
| 1825 | } |
| 1826 | |
| 1827 | typedef struct ZSTDv01_Dctx_s |
| 1828 | { |
| 1829 | U32 LLTable[FSE_DTABLE_SIZE_U32(LLFSELog)]; |
| 1830 | U32 OffTable[FSE_DTABLE_SIZE_U32(OffFSELog)]; |
| 1831 | U32 MLTable[FSE_DTABLE_SIZE_U32(MLFSELog)]; |
| 1832 | void* previousDstEnd; |
| 1833 | void* base; |
| 1834 | size_t expected; |
| 1835 | blockType_t bType; |
| 1836 | U32 phase; |
| 1837 | } dctx_t; |
| 1838 | |
| 1839 | |
| 1840 | static size_t ZSTD_decompressSequences( |
| 1841 | void* ctx, |
| 1842 | void* dst, size_t maxDstSize, |
| 1843 | const void* seqStart, size_t seqSize, |
| 1844 | const BYTE* litStart, size_t litSize) |
| 1845 | { |
| 1846 | dctx_t* dctx = (dctx_t*)ctx; |
| 1847 | const BYTE* ip = (const BYTE*)seqStart; |
| 1848 | const BYTE* const iend = ip + seqSize; |
| 1849 | BYTE* const ostart = (BYTE* const)dst; |
| 1850 | BYTE* op = ostart; |
| 1851 | BYTE* const oend = ostart + maxDstSize; |
| 1852 | size_t errorCode, dumpsLength; |
| 1853 | const BYTE* litPtr = litStart; |
| 1854 | const BYTE* const litEnd = litStart + litSize; |
| 1855 | int nbSeq; |
| 1856 | const BYTE* dumps; |
| 1857 | U32* DTableLL = dctx->LLTable; |
| 1858 | U32* DTableML = dctx->MLTable; |
| 1859 | U32* DTableOffb = dctx->OffTable; |
| 1860 | BYTE* const base = (BYTE*) (dctx->base); |
| 1861 | |
| 1862 | /* Build Decoding Tables */ |
| 1863 | errorCode = ZSTDv01_decodeSeqHeaders(&nbSeq, &dumps, &dumpsLength, |
| 1864 | DTableLL, DTableML, DTableOffb, |
| 1865 | ip, iend-ip); |
| 1866 | if (ZSTDv01_isError(errorCode)) return errorCode; |
| 1867 | ip += errorCode; |
| 1868 | |
| 1869 | /* Regen sequences */ |
| 1870 | { |
| 1871 | seq_t sequence; |
| 1872 | seqState_t seqState; |
| 1873 | |
| 1874 | memset(&sequence, 0, sizeof(sequence)); |
| 1875 | seqState.dumps = dumps; |
| 1876 | seqState.dumpsEnd = dumps + dumpsLength; |
| 1877 | seqState.prevOffset = 1; |
| 1878 | errorCode = FSE_initDStream(&(seqState.DStream), ip, iend-ip); |
| 1879 | if (FSE_isError(errorCode)) return ERROR(corruption_detected); |
| 1880 | FSE_initDState(&(seqState.stateLL), &(seqState.DStream), DTableLL); |
| 1881 | FSE_initDState(&(seqState.stateOffb), &(seqState.DStream), DTableOffb); |
| 1882 | FSE_initDState(&(seqState.stateML), &(seqState.DStream), DTableML); |
| 1883 | |
| 1884 | for ( ; (FSE_reloadDStream(&(seqState.DStream)) <= FSE_DStream_completed) && (nbSeq>0) ; ) |
| 1885 | { |
| 1886 | size_t oneSeqSize; |
| 1887 | nbSeq--; |
| 1888 | ZSTD_decodeSequence(&sequence, &seqState); |
| 1889 | oneSeqSize = ZSTD_execSequence(op, sequence, &litPtr, litEnd, base, oend); |
| 1890 | if (ZSTDv01_isError(oneSeqSize)) return oneSeqSize; |
| 1891 | op += oneSeqSize; |
| 1892 | } |
| 1893 | |
| 1894 | /* check if reached exact end */ |
| 1895 | if ( !FSE_endOfDStream(&(seqState.DStream)) ) return ERROR(corruption_detected); /* requested too much : data is corrupted */ |
| 1896 | if (nbSeq<0) return ERROR(corruption_detected); /* requested too many sequences : data is corrupted */ |
| 1897 | |
| 1898 | /* last literal segment */ |
| 1899 | { |
| 1900 | size_t lastLLSize = litEnd - litPtr; |
| 1901 | if (op+lastLLSize > oend) return ERROR(dstSize_tooSmall); |
| 1902 | if (op != litPtr) memmove(op, litPtr, lastLLSize); |
| 1903 | op += lastLLSize; |
| 1904 | } |
| 1905 | } |
| 1906 | |
| 1907 | return op-ostart; |
| 1908 | } |
| 1909 | |
| 1910 | |
| 1911 | static size_t ZSTD_decompressBlock( |
| 1912 | void* ctx, |
| 1913 | void* dst, size_t maxDstSize, |
| 1914 | const void* src, size_t srcSize) |
| 1915 | { |
| 1916 | /* blockType == blockCompressed, srcSize is trusted */ |
| 1917 | const BYTE* ip = (const BYTE*)src; |
| 1918 | const BYTE* litPtr = NULL; |
| 1919 | size_t litSize = 0; |
| 1920 | size_t errorCode; |
| 1921 | |
| 1922 | /* Decode literals sub-block */ |
| 1923 | errorCode = ZSTDv01_decodeLiteralsBlock(ctx, dst, maxDstSize, &litPtr, &litSize, src, srcSize); |
| 1924 | if (ZSTDv01_isError(errorCode)) return errorCode; |
| 1925 | ip += errorCode; |
| 1926 | srcSize -= errorCode; |
| 1927 | |
| 1928 | return ZSTD_decompressSequences(ctx, dst, maxDstSize, ip, srcSize, litPtr, litSize); |
| 1929 | } |
| 1930 | |
| 1931 | |
| 1932 | size_t ZSTDv01_decompressDCtx(void* ctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize) |
| 1933 | { |
| 1934 | const BYTE* ip = (const BYTE*)src; |
| 1935 | const BYTE* iend = ip + srcSize; |
| 1936 | BYTE* const ostart = (BYTE* const)dst; |
| 1937 | BYTE* op = ostart; |
| 1938 | BYTE* const oend = ostart + maxDstSize; |
| 1939 | size_t remainingSize = srcSize; |
| 1940 | U32 magicNumber; |
| 1941 | size_t errorCode=0; |
| 1942 | blockProperties_t blockProperties; |
| 1943 | |
| 1944 | /* Frame Header */ |
| 1945 | if (srcSize < ZSTD_frameHeaderSize+ZSTD_blockHeaderSize) return ERROR(srcSize_wrong); |
| 1946 | magicNumber = ZSTD_readBE32(src); |
| 1947 | if (magicNumber != ZSTD_magicNumber) return ERROR(prefix_unknown); |
| 1948 | ip += ZSTD_frameHeaderSize; remainingSize -= ZSTD_frameHeaderSize; |
| 1949 | |
| 1950 | /* Loop on each block */ |
| 1951 | while (1) |
| 1952 | { |
| 1953 | size_t blockSize = ZSTDv01_getcBlockSize(ip, iend-ip, &blockProperties); |
| 1954 | if (ZSTDv01_isError(blockSize)) return blockSize; |
| 1955 | |
| 1956 | ip += ZSTD_blockHeaderSize; |
| 1957 | remainingSize -= ZSTD_blockHeaderSize; |
| 1958 | if (blockSize > remainingSize) return ERROR(srcSize_wrong); |
| 1959 | |
| 1960 | switch(blockProperties.blockType) |
| 1961 | { |
| 1962 | case bt_compressed: |
| 1963 | errorCode = ZSTD_decompressBlock(ctx, op, oend-op, ip, blockSize); |
| 1964 | break; |
| 1965 | case bt_raw : |
| 1966 | errorCode = ZSTD_copyUncompressedBlock(op, oend-op, ip, blockSize); |
| 1967 | break; |
| 1968 | case bt_rle : |
| 1969 | return ERROR(GENERIC); /* not yet supported */ |
| 1970 | break; |
| 1971 | case bt_end : |
| 1972 | /* end of frame */ |
| 1973 | if (remainingSize) return ERROR(srcSize_wrong); |
| 1974 | break; |
| 1975 | default: |
| 1976 | return ERROR(GENERIC); |
| 1977 | } |
| 1978 | if (blockSize == 0) break; /* bt_end */ |
| 1979 | |
| 1980 | if (ZSTDv01_isError(errorCode)) return errorCode; |
| 1981 | op += errorCode; |
| 1982 | ip += blockSize; |
| 1983 | remainingSize -= blockSize; |
| 1984 | } |
| 1985 | |
| 1986 | return op-ostart; |
| 1987 | } |
| 1988 | |
| 1989 | size_t ZSTDv01_decompress(void* dst, size_t maxDstSize, const void* src, size_t srcSize) |
| 1990 | { |
| 1991 | dctx_t ctx; |
| 1992 | ctx.base = dst; |
| 1993 | return ZSTDv01_decompressDCtx(&ctx, dst, maxDstSize, src, srcSize); |
| 1994 | } |
| 1995 | |
| 1996 | size_t ZSTDv01_findFrameCompressedSize(const void* src, size_t srcSize) |
| 1997 | { |
| 1998 | const BYTE* ip = (const BYTE*)src; |
| 1999 | size_t remainingSize = srcSize; |
| 2000 | U32 magicNumber; |
| 2001 | blockProperties_t blockProperties; |
| 2002 | |
| 2003 | /* Frame Header */ |
| 2004 | if (srcSize < ZSTD_frameHeaderSize+ZSTD_blockHeaderSize) return ERROR(srcSize_wrong); |
| 2005 | magicNumber = ZSTD_readBE32(src); |
| 2006 | if (magicNumber != ZSTD_magicNumber) return ERROR(prefix_unknown); |
| 2007 | ip += ZSTD_frameHeaderSize; remainingSize -= ZSTD_frameHeaderSize; |
| 2008 | |
| 2009 | /* Loop on each block */ |
| 2010 | while (1) |
| 2011 | { |
| 2012 | size_t blockSize = ZSTDv01_getcBlockSize(ip, remainingSize, &blockProperties); |
| 2013 | if (ZSTDv01_isError(blockSize)) return blockSize; |
| 2014 | |
| 2015 | ip += ZSTD_blockHeaderSize; |
| 2016 | remainingSize -= ZSTD_blockHeaderSize; |
| 2017 | if (blockSize > remainingSize) return ERROR(srcSize_wrong); |
| 2018 | |
| 2019 | if (blockSize == 0) break; /* bt_end */ |
| 2020 | |
| 2021 | ip += blockSize; |
| 2022 | remainingSize -= blockSize; |
| 2023 | } |
| 2024 | |
| 2025 | return ip - (const BYTE*)src; |
| 2026 | } |
| 2027 | |
| 2028 | /******************************* |
| 2029 | * Streaming Decompression API |
| 2030 | *******************************/ |
| 2031 | |
| 2032 | size_t ZSTDv01_resetDCtx(ZSTDv01_Dctx* dctx) |
| 2033 | { |
| 2034 | dctx->expected = ZSTD_frameHeaderSize; |
| 2035 | dctx->phase = 0; |
| 2036 | dctx->previousDstEnd = NULL; |
| 2037 | dctx->base = NULL; |
| 2038 | return 0; |
| 2039 | } |
| 2040 | |
| 2041 | ZSTDv01_Dctx* ZSTDv01_createDCtx(void) |
| 2042 | { |
| 2043 | ZSTDv01_Dctx* dctx = (ZSTDv01_Dctx*)malloc(sizeof(ZSTDv01_Dctx)); |
| 2044 | if (dctx==NULL) return NULL; |
| 2045 | ZSTDv01_resetDCtx(dctx); |
| 2046 | return dctx; |
| 2047 | } |
| 2048 | |
| 2049 | size_t ZSTDv01_freeDCtx(ZSTDv01_Dctx* dctx) |
| 2050 | { |
| 2051 | free(dctx); |
| 2052 | return 0; |
| 2053 | } |
| 2054 | |
| 2055 | size_t ZSTDv01_nextSrcSizeToDecompress(ZSTDv01_Dctx* dctx) |
| 2056 | { |
| 2057 | return ((dctx_t*)dctx)->expected; |
| 2058 | } |
| 2059 | |
| 2060 | size_t ZSTDv01_decompressContinue(ZSTDv01_Dctx* dctx, void* dst, size_t maxDstSize, const void* src, size_t srcSize) |
| 2061 | { |
| 2062 | dctx_t* ctx = (dctx_t*)dctx; |
| 2063 | |
| 2064 | /* Sanity check */ |
| 2065 | if (srcSize != ctx->expected) return ERROR(srcSize_wrong); |
| 2066 | if (dst != ctx->previousDstEnd) /* not contiguous */ |
| 2067 | ctx->base = dst; |
| 2068 | |
| 2069 | /* Decompress : frame header */ |
| 2070 | if (ctx->phase == 0) |
| 2071 | { |
| 2072 | /* Check frame magic header */ |
| 2073 | U32 magicNumber = ZSTD_readBE32(src); |
| 2074 | if (magicNumber != ZSTD_magicNumber) return ERROR(prefix_unknown); |
| 2075 | ctx->phase = 1; |
| 2076 | ctx->expected = ZSTD_blockHeaderSize; |
| 2077 | return 0; |
| 2078 | } |
| 2079 | |
| 2080 | /* Decompress : block header */ |
| 2081 | if (ctx->phase == 1) |
| 2082 | { |
| 2083 | blockProperties_t bp; |
| 2084 | size_t blockSize = ZSTDv01_getcBlockSize(src, ZSTD_blockHeaderSize, &bp); |
| 2085 | if (ZSTDv01_isError(blockSize)) return blockSize; |
| 2086 | if (bp.blockType == bt_end) |
| 2087 | { |
| 2088 | ctx->expected = 0; |
| 2089 | ctx->phase = 0; |
| 2090 | } |
| 2091 | else |
| 2092 | { |
| 2093 | ctx->expected = blockSize; |
| 2094 | ctx->bType = bp.blockType; |
| 2095 | ctx->phase = 2; |
| 2096 | } |
| 2097 | |
| 2098 | return 0; |
| 2099 | } |
| 2100 | |
| 2101 | /* Decompress : block content */ |
| 2102 | { |
| 2103 | size_t rSize; |
| 2104 | switch(ctx->bType) |
| 2105 | { |
| 2106 | case bt_compressed: |
| 2107 | rSize = ZSTD_decompressBlock(ctx, dst, maxDstSize, src, srcSize); |
| 2108 | break; |
| 2109 | case bt_raw : |
| 2110 | rSize = ZSTD_copyUncompressedBlock(dst, maxDstSize, src, srcSize); |
| 2111 | break; |
| 2112 | case bt_rle : |
| 2113 | return ERROR(GENERIC); /* not yet handled */ |
| 2114 | break; |
| 2115 | case bt_end : /* should never happen (filtered at phase 1) */ |
| 2116 | rSize = 0; |
| 2117 | break; |
| 2118 | default: |
| 2119 | return ERROR(GENERIC); |
| 2120 | } |
| 2121 | ctx->phase = 1; |
| 2122 | ctx->expected = ZSTD_blockHeaderSize; |
| 2123 | ctx->previousDstEnd = (void*)( ((char*)dst) + rSize); |
| 2124 | return rSize; |
| 2125 | } |
| 2126 | |
| 2127 | } |