/linux-5.19.10/lib/zstd/compress/ |
D | zstd_double_fast.c | 26 const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE; in ZSTD_fillDoubleHashTable() local 33 for (; ip + fastHashFillStep - 1 <= iend; ip += fastHashFillStep) { in ZSTD_fillDoubleHashTable() 69 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_doubleFast_generic() local 70 const BYTE* const ilimit = iend - HASH_READ_SIZE; in ZSTD_compressBlock_doubleFast_generic() 148 const BYTE* repMatchEnd = repIndex < prefixLowestIndex ? dictEnd : iend; in ZSTD_compressBlock_doubleFast_generic() 149 mLength = ZSTD_count_2segments(ip+1+4, repMatch+4, iend, repMatchEnd, prefixLowest) + 4; in ZSTD_compressBlock_doubleFast_generic() 151 ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH); in ZSTD_compressBlock_doubleFast_generic() 158 mLength = ZSTD_count(ip+1+4, ip+1+4-offset_1, iend) + 4; in ZSTD_compressBlock_doubleFast_generic() 160 ZSTD_storeSeq(seqStore, (size_t)(ip-anchor), anchor, iend, 0, mLength-MINMATCH); in ZSTD_compressBlock_doubleFast_generic() 167 mLength = ZSTD_count(ip+8, matchLong+8, iend) + 8; in ZSTD_compressBlock_doubleFast_generic() [all …]
|
D | zstd_fast.c | 25 const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE; in ZSTD_fillHashTable() local 31 for ( ; ip + fastHashFillStep < iend + 2; ip += fastHashFillStep) { in ZSTD_fillHashTable() 66 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_generic() local 67 const BYTE* const ilimit = iend - HASH_READ_SIZE; in ZSTD_compressBlock_fast_generic() 151 mLength += ZSTD_count(ip0+mLength, match0+mLength, iend); in ZSTD_compressBlock_fast_generic() 152 ZSTD_storeSeq(seqStore, (size_t)(ip0-anchor), anchor, iend, offcode, mLength-MINMATCH); in ZSTD_compressBlock_fast_generic() 166 size_t const rLength = ZSTD_count(ip0+4, ip0+4-offset_2, iend) + 4; in ZSTD_compressBlock_fast_generic() 170 … ZSTD_storeSeq(seqStore, 0 /*litLen*/, anchor, iend, 0 /*offCode*/, rLength-MINMATCH); in ZSTD_compressBlock_fast_generic() 182 return (size_t)(iend - anchor); in ZSTD_compressBlock_fast_generic() 222 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_dictMatchState_generic() local [all …]
|
D | zstd_ldm.c | 213 const BYTE* const iend = (const BYTE*)end; in ZSTD_ldm_fillFastTables() local 218 ZSTD_fillHashTable(ms, iend, ZSTD_dtlm_fast); in ZSTD_ldm_fillFastTables() 222 ZSTD_fillDoubleHashTable(ms, iend, ZSTD_dtlm_fast); in ZSTD_ldm_fillFastTables() 242 const BYTE* iend, ldmParams_t const* params) in ZSTD_ldm_fillHashTable() argument 255 while (ip < iend) { in ZSTD_ldm_fillHashTable() 260 hashed = ZSTD_ldm_gear_feed(&hashState, ip, iend - ip, splits, &numSplits); in ZSTD_ldm_fillHashTable() 313 BYTE const* const iend = istart + srcSize; in ZSTD_ldm_generateSequences_internal() local 314 BYTE const* const ilimit = iend - HASH_READ_SIZE; in ZSTD_ldm_generateSequences_internal() 326 return iend - anchor; in ZSTD_ldm_generateSequences_internal() 394 cur->offset < dictLimit ? dictEnd : iend; in ZSTD_ldm_generateSequences_internal() [all …]
|
D | zstd_lazy.c | 21 const BYTE* ip, const BYTE* iend, in ZSTD_updateDUBT() argument 39 assert(ip + 8 <= iend); /* condition for ZSTD_hashPtr */ in ZSTD_updateDUBT() 40 (void)iend; in ZSTD_updateDUBT() 78 const BYTE* const iend = (curr>=dictLimit) ? inputEnd : dictBase + dictLimit; in ZSTD_insertDUBT1() local 94 assert(ip < iend); /* condition for ZSTD_count */ in ZSTD_insertDUBT1() 113 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertDUBT1() 116 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertDUBT1() 124 if (ip+matchLength == iend) { /* equal : no way to know if inf or sup */ in ZSTD_insertDUBT1() 155 const BYTE* const ip, const BYTE* const iend, in ZSTD_DUBT_findBetterDictMatch() argument 192 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_DUBT_findBetterDictMatch() [all …]
|
D | zstd_opt.c | 383 const BYTE* const ip, const BYTE* const iend, in ZSTD_insertBt1() argument 419 assert(ip <= iend-8); /* required for h calculation */ in ZSTD_insertBt1() 452 matchLength += ZSTD_count(ip+matchLength, match+matchLength, iend); in ZSTD_insertBt1() 455 …matchLength += ZSTD_count_2segments(ip+matchLength, match+matchLength, iend, dictEnd, prefixStart); in ZSTD_insertBt1() 466 if (ip+matchLength == iend) { /* equal : no way to know if inf or sup */ in ZSTD_insertBt1() 497 const BYTE* const ip, const BYTE* const iend, in ZSTD_updateTree_internal() argument 507 U32 const forward = ZSTD_insertBt1(ms, base+idx, iend, mls, dictMode == ZSTD_extDict); in ZSTD_updateTree_internal() 512 assert((size_t)(iend - base) <= (size_t)(U32)(-1)); in ZSTD_updateTree_internal() 516 void ZSTD_updateTree(ZSTD_matchState_t* ms, const BYTE* ip, const BYTE* iend) { in ZSTD_updateTree() argument 517 ZSTD_updateTree_internal(ms, ip, iend, ms->cParams.minMatch, ZSTD_noDict); in ZSTD_updateTree() [all …]
|
D | hist.c | 73 const BYTE* const iend = ip+sourceSize; in HIST_count_parallel_wksp() local 92 while (ip < iend-15) { in HIST_count_parallel_wksp() 118 while (ip<iend) Counting1[*ip++]++; in HIST_count_parallel_wksp()
|
D | zstd_ldm.h | 26 const BYTE* iend, ldmParams_t const* params);
|
D | zstd_opt.h | 18 void ZSTD_updateTree(ZSTD_matchState_t* ms, const BYTE* ip, const BYTE* iend);
|
D | zstd_compress.c | 2760 void const* iend) in ZSTD_overflowCorrectIfNeeded() argument 2762 if (ZSTD_window_needOverflowCorrection(ms->window, iend)) { in ZSTD_overflowCorrectIfNeeded() 3057 const BYTE* const iend = ip + srcSize; in ZSTD_loadDictionaryContent() local 3060 ms->loadedDictEnd = params->forceWindow ? 0 : (U32)(iend - ms->window.base); in ZSTD_loadDictionaryContent() 3064 ls->loadedDictEnd = params->forceWindow ? 0 : (U32)(iend - ls->window.base); in ZSTD_loadDictionaryContent() 3072 while (iend - ip > HASH_READ_SIZE) { in ZSTD_loadDictionaryContent() 3073 size_t const remaining = (size_t)(iend - ip); in ZSTD_loadDictionaryContent() 3117 ms->nextToUpdate = (U32)(iend - ms->window.base); in ZSTD_loadDictionaryContent() 4099 const char* const iend = input->size != 0 ? istart + input->size : istart; in ZSTD_compressStream_generic() local 4128 && ( (size_t)(oend-op) >= ZSTD_compressBound(iend-ip) /* Enough output space */ in ZSTD_compressStream_generic() [all …]
|
D | zstd_compress_superblock.c | 721 BYTE const* const iend = ip + srcSize; in ZSTD_compressSubBlock_multi() local 778 assert(ip + decompressedSize <= iend); in ZSTD_compressSubBlock_multi() 809 if (ip < iend) { in ZSTD_compressSubBlock_multi() 810 size_t const cSize = ZSTD_noCompressBlock(op, oend - op, ip, iend - ip, lastBlock); in ZSTD_compressSubBlock_multi() 811 …GLOG(5, "ZSTD_compressSubBlock_multi last sub-block uncompressed, %zu bytes", (size_t)(iend - ip)); in ZSTD_compressSubBlock_multi()
|
D | fse_compress.c | 555 const BYTE* const iend = istart + srcSize; in FSE_compress_usingCTable_generic() local 556 const BYTE* ip=iend; in FSE_compress_usingCTable_generic()
|
D | zstd_compress_internal.h | 488 static void ZSTD_safecopyLiterals(BYTE* op, BYTE const* ip, BYTE const* const iend, BYTE const* ili… in ZSTD_safecopyLiterals() argument 489 assert(iend > ilimit_w); in ZSTD_safecopyLiterals() 495 while (ip < iend) *op++ = *ip++; in ZSTD_safecopyLiterals()
|
D | huf_compress.c | 681 const BYTE* const iend = ip + srcSize; in HUF_compress4X_usingCTable_internal() local 718 assert(ip <= iend); in HUF_compress4X_usingCTable_internal() 719 …, HUF_compress1X_usingCTable_internal(op, (size_t)(oend-op), ip, (size_t)(iend-ip), CTable, bmi2) … in HUF_compress4X_usingCTable_internal()
|
/linux-5.19.10/lib/zstd/common/ |
D | entropy_common.c | 63 const BYTE* const iend = istart + hbSize; in FSE_readNCount_body() local 108 if (LIKELY(ip <= iend-7)) { in FSE_readNCount_body() 111 bitCount -= (int)(8 * (iend - 7 - ip)); in FSE_readNCount_body() 113 ip = iend - 4; in FSE_readNCount_body() 137 if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) { in FSE_readNCount_body() 142 bitCount -= (int)(8 * (iend - 4 - ip)); in FSE_readNCount_body() 144 ip = iend - 4; in FSE_readNCount_body() 186 if (LIKELY(ip <= iend-7) || (ip + (bitCount>>3) <= iend-4)) { in FSE_readNCount_body() 190 bitCount -= (int)(8 * (iend - 4 - ip)); in FSE_readNCount_body() 192 ip = iend - 4; in FSE_readNCount_body()
|
/linux-5.19.10/arch/arm64/kernel/ |
D | head.S | 192 .macro compute_indices, vstart, vend, shift, ptrs, istart, iend, count 193 lsr \iend, \vend, \shift 196 and \iend, \iend, \istart // iend = (vend >> shift) & (ptrs - 1) 199 add \iend, \iend, \istart // iend += count * ptrs 207 sub \count, \iend, \istart 227 .macro map_memory, tbl, rtbl, vstart, vend, flags, phys, pgds, istart, iend, tmp, count, sv 232 compute_indices \vstart, \vend, #PGDIR_SHIFT, \pgds, \istart, \iend, \count 233 populate_entries \tbl, \rtbl, \istart, \iend, #PMD_TYPE_TABLE, #PAGE_SIZE, \tmp 238 compute_indices \vstart, \vend, #PUD_SHIFT, #PTRS_PER_PUD, \istart, \iend, \count 239 populate_entries \tbl, \rtbl, \istart, \iend, #PMD_TYPE_TABLE, #PAGE_SIZE, \tmp [all …]
|
/linux-5.19.10/lib/lz4/ |
D | lz4_decompress.c | 84 const BYTE * const iend = ip + srcSize; in LZ4_decompress_generic() local 98 const BYTE *const shortiend = iend - in LZ4_decompress_generic() 131 assert(!endOnInput || ip <= iend); in LZ4_decompress_generic() 195 if (unlikely(endOnInput ? ip >= iend - RUN_MASK : 0)) { in LZ4_decompress_generic() 203 ? ip < iend - RUN_MASK in LZ4_decompress_generic() 225 || (ip + length > iend - (2 + 1 + LASTLITERALS)))) in LZ4_decompress_generic() 237 && (ip + length > iend)) { in LZ4_decompress_generic() 256 && ((ip + length != iend) in LZ4_decompress_generic() 279 if (!partialDecoding || (cpy == oend) || (ip >= (iend - 2))) in LZ4_decompress_generic() 320 if ((endOnInput) && (ip > iend - LASTLITERALS)) in LZ4_decompress_generic()
|
D | lz4_compress.c | 196 const BYTE * const iend = ip + inputSize; in LZ4_compress_generic() local 197 const BYTE * const mflimit = iend - MFLIMIT; in LZ4_compress_generic() 198 const BYTE * const matchlimit = iend - LASTLITERALS; in LZ4_compress_generic() 431 size_t const lastRun = (size_t)(iend - anchor); in LZ4_compress_generic() 534 const BYTE * const iend = ip + *srcSizePtr; in LZ4_compress_destSize_generic() local 535 const BYTE * const mflimit = iend - MFLIMIT; in LZ4_compress_destSize_generic() 536 const BYTE * const matchlimit = iend - LASTLITERALS; in LZ4_compress_destSize_generic() 690 size_t lastRunSize = (size_t)(iend - anchor); in LZ4_compress_destSize_generic()
|
D | lz4hc_compress.c | 350 const BYTE * const iend = ip + inputSize; in LZ4HC_compress_generic() local 351 const BYTE * const mflimit = iend - MFLIMIT; in LZ4HC_compress_generic() 352 const BYTE * const matchlimit = (iend - LASTLITERALS); in LZ4HC_compress_generic() 556 int lastRun = (int)(iend - anchor); in LZ4HC_compress_generic() 573 LZ4_memcpy(op, anchor, iend - anchor); in LZ4HC_compress_generic() 574 op += iend - anchor; in LZ4HC_compress_generic()
|
/linux-5.19.10/lib/zstd/decompress/ |
D | zstd_decompress_block.c | 581 const BYTE* const iend = istart + srcSize; in ZSTD_decodeSeqHeaders() local 598 RETURN_ERROR_IF(ip+2 > iend, srcSize_wrong, ""); in ZSTD_decodeSeqHeaders() 602 RETURN_ERROR_IF(ip >= iend, srcSize_wrong, ""); in ZSTD_decodeSeqHeaders() 609 …RETURN_ERROR_IF(ip+1 > iend, srcSize_wrong, ""); /* minimum possible size: 1 byte for symbol encod… in ZSTD_decodeSeqHeaders() 618 ip, iend-ip, in ZSTD_decodeSeqHeaders() 630 ip, iend-ip, in ZSTD_decodeSeqHeaders() 642 ip, iend-ip, in ZSTD_decodeSeqHeaders() 1110 const BYTE* const iend = ip + seqSize; in ZSTD_decompressSequences_body() local 1129 ERR_isError(BIT_initDStream(&seqState.DStream, ip, iend-ip)), in ZSTD_decompressSequences_body() 1244 const BYTE* const iend = ip + seqSize; in ZSTD_decompressSequencesLong_body() local [all …]
|
D | zstd_decompress.c | 1816 const char* const iend = input->size != 0 ? src + input->size : src; in ZSTD_decompressStream() local 1851 DEBUGLOG(5, "stage zdss_loadHeader (srcSize : %u)", (U32)(iend - ip)); in ZSTD_decompressStream() 1862 size_t const remainingInput = (size_t)(iend-ip); in ZSTD_decompressStream() 1863 assert(iend >= ip); in ZSTD_decompressStream() 1881 size_t const cSize = ZSTD_findFrameCompressedSize(istart, (size_t)(iend-istart)); in ZSTD_decompressStream() 1882 if (cSize <= (size_t)(iend-istart)) { in ZSTD_decompressStream() 1963 … size_t const neededInSize = ZSTD_nextSrcSizeToDecompressWithInputSize(zds, (size_t)(iend - ip)); in ZSTD_decompressStream() 1970 if ((size_t)(iend-ip) >= neededInSize) { /* decode directly from src */ in ZSTD_decompressStream() 1976 if (ip==iend) { someMoreWork = 0; break; } /* no more input */ in ZSTD_decompressStream() 1986 assert(neededInSize == ZSTD_nextSrcSizeToDecompressWithInputSize(zds, iend - ip)); in ZSTD_decompressStream() [all …]
|
/linux-5.19.10/drivers/scsi/qla2xxx/ |
D | qla_sup.c | 3090 uint32_t istart, iend, iter, vend; in qla2x00_get_fcode_version() local 3099 iend = istart + 0x100; in qla2x00_get_fcode_version() 3104 while ((iter < iend) && !do_next) { in qla2x00_get_fcode_version()
|
/linux-5.19.10/drivers/media/usb/uvc/ |
D | uvc_ctrl.c | 2387 const struct uvc_control_info *iend = info + ARRAY_SIZE(uvc_ctrls); in uvc_ctrl_init_ctrl() local 2400 for (; info < iend; ++info) { in uvc_ctrl_init_ctrl()
|