Lines Matching refs:U32
289 typedef uint32_t U32; typedef
299 typedef unsigned int U32; typedef
336 U32 u; in LZ4_isLittleEndian()
346 static U32 LZ4_read32(const void *memPtr) { return *(const U32 *)memPtr; } in LZ4_read32()
350 static void LZ4_write32(void *memPtr, U32 value) { *(U32 *)memPtr = value; } in LZ4_write32()
359 U32 u32;
364 static U32 LZ4_read32(const void *ptr) { return ((const unalign *)ptr)->u32; } in LZ4_read32()
368 static void LZ4_write32(void *memPtr, U32 value) { ((unalign *)memPtr)->u32 = value; } in LZ4_write32()
379 static U32 LZ4_read32(const void *memPtr) in LZ4_read32()
381 U32 val; in LZ4_read32()
398 static void LZ4_write32(void *memPtr, U32 value) in LZ4_write32()
581 _BitScanForward(&r, (U32)val); in LZ4_NbCommonBytes()
586 return (unsigned)__builtin_ctz((U32)val) >> 3; in LZ4_NbCommonBytes()
588 const U32 m = 0x01010101; in LZ4_NbCommonBytes()
742 static const U32 by32 = sizeof(val) * 4; /* 32 on 64 bits (goal), 16 on 32 bits. in LZ4_NbCommonBytes()
774 return (unsigned)__clz((U32)val) >> 3; in LZ4_NbCommonBytes()
839 static const U32 LZ4_skipTrigger = 6; /* Increase this value ==> compression run slower on incompre…
917 LZ4_FORCE_INLINE U32 LZ4_hash4(U32 sequence, tableType_t const tableType) in LZ4_hash4()
925 LZ4_FORCE_INLINE U32 LZ4_hash5(U64 sequence, tableType_t const tableType) in LZ4_hash5()
927 const U32 hashLog = (tableType == byU16) ? LZ4_HASHLOG + 1 : LZ4_HASHLOG; in LZ4_hash5()
931 return (U32)(((sequence << 24) * prime5bytes) >> (64 - hashLog)); in LZ4_hash5()
936 return (U32)(((sequence >> 24) * prime8bytes) >> (64 - hashLog)); in LZ4_hash5()
940 LZ4_FORCE_INLINE U32 LZ4_hashPosition(const void *const p, tableType_t const tableType) in LZ4_hashPosition()
947 LZ4_FORCE_INLINE void LZ4_clearHash(U32 h, void *tableBase, tableType_t const tableType) in LZ4_clearHash()
965 U32 *hashTable = (U32 *)tableBase; in LZ4_clearHash()
978 LZ4_FORCE_INLINE void LZ4_putIndexOnHash(U32 idx, U32 h, void *tableBase, tableType_t const tableTy… in LZ4_putIndexOnHash()
991 U32 *hashTable = (U32 *)tableBase; in LZ4_putIndexOnHash()
1005 LZ4_FORCE_INLINE void LZ4_putPositionOnHash(const BYTE *p, U32 h, in LZ4_putPositionOnHash()
1024 U32 *hashTable = (U32 *)tableBase; in LZ4_putPositionOnHash()
1025 hashTable[h] = (U32)(p - srcBase); in LZ4_putPositionOnHash()
1039 U32 const h = LZ4_hashPosition(p, tableType); in LZ4_putPosition()
1049 LZ4_FORCE_INLINE U32 LZ4_getIndexOnHash(U32 h, const void *tableBase, tableType_t tableType) in LZ4_getIndexOnHash()
1054 const U32 *const hashTable = (const U32 *)tableBase; in LZ4_getIndexOnHash()
1068 static const BYTE *LZ4_getPositionOnHash(U32 h, const void *tableBase, tableType_t tableType, const… in LZ4_getPositionOnHash()
1077 const U32 *const hashTable = (const U32 *)tableBase; in LZ4_getPositionOnHash()
1091 U32 const h = LZ4_hashPosition(p, tableType); in LZ4_getPosition()
1112 cctx->tableType = (U32)clearedTable; in LZ4_prepareTable()
1158 U32 const startIndex = cctx->currentOffset; in LZ4_compress_generic_validated()
1165 const U32 dictSize = in LZ4_compress_generic_validated()
1167 …const U32 dictDelta = (dictDirective == usingDictCtx) ? startIndex - dictCtx->currentOffset : 0; /… in LZ4_compress_generic_validated()
1170 U32 const prefixIdxLimit = startIndex - dictSize; /* used when dictDirective == dictSmall */ in LZ4_compress_generic_validated()
1185 U32 offset = 0; in LZ4_compress_generic_validated()
1186 U32 forwardH; in LZ4_compress_generic_validated()
1212 cctx->dictSize = (U32)inputSize; in LZ4_compress_generic_validated()
1216 cctx->dictSize += (U32)inputSize; in LZ4_compress_generic_validated()
1218 cctx->currentOffset += (U32)inputSize; in LZ4_compress_generic_validated()
1219 cctx->tableType = (U32)tableType; in LZ4_compress_generic_validated()
1244 U32 const h = forwardH; in LZ4_compress_generic_validated()
1267 U32 const h = forwardH; in LZ4_compress_generic_validated()
1268 U32 const current = (U32)(forwardIp - base); in LZ4_compress_generic_validated()
1269 U32 matchIndex = LZ4_getIndexOnHash(h, cctx->hashTable, tableType); in LZ4_compress_generic_validated()
1409 DEBUGLOG(6, " with offset=%u (same segment)", (U32)(ip - match)); in LZ4_compress_generic_validated()
1448 …U32 newMatchCode = 15 /* in token */ - 1 /* to avoid needing a zero byte */ + ((U32)(olimit - op) … in LZ4_compress_generic_validated()
1460 DEBUGLOG(5, "Clearing %u positions", (U32)(filledIp - ip)); in LZ4_compress_generic_validated()
1463 U32 const h = LZ4_hashPosition(ptr, tableType); in LZ4_compress_generic_validated()
1519 U32 const h = LZ4_hashPosition(ip, tableType); in LZ4_compress_generic_validated()
1520 U32 const current = (U32)(ip - base); in LZ4_compress_generic_validated()
1521 U32 matchIndex = LZ4_getIndexOnHash(h, cctx->hashTable, tableType); in LZ4_compress_generic_validated()
1642 if ((U32)srcSize > (U32)LZ4_MAX_INPUT_SIZE) in LZ4_compress_generic()
1938 dict->dictSize = (U32)(dictEnd - p); in LZ4_loadDict()
1939 dict->tableType = (U32)tableType; in LZ4_loadDict()
1986 U32 const delta = LZ4_dict->currentOffset - 64 KB; in LZ4_renormDictT()
2036 streamPtr->dictSize = (U32)(dictEnd - sourceEnd); in LZ4_compress_fast_continue()
2091 streamPtr->dictSize = (U32)inputSize; in LZ4_compress_fast_continue()
2114 streamPtr->dictSize = (U32)srcSize; in LZ4_compress_forceExtDict()
2131 if ((U32)dictSize > 64 KB) in LZ4_saveDict()
2135 if ((U32)dictSize > dict->dictSize) in LZ4_saveDict()
2146 dict->dictSize = (U32)dictSize; in LZ4_saveDict()
2188 U32 length = 0; in read_variable_length()
2189 U32 s; in read_variable_length()