/linux-6.1.9/include/uapi/linux/ |
D | cciss_defs.h | 44 #define BYTE __u8 macro 59 BYTE Dev; 60 BYTE Bus:6; 61 BYTE Mode:2; /* b00 */ 64 BYTE DevLSB; 65 BYTE DevMSB:6; 66 BYTE Mode:2; /* b01 */ 69 BYTE Dev:5; 70 BYTE Bus:3; 71 BYTE Targ:6; [all …]
|
/linux-6.1.9/lib/lz4/ |
D | lz4_compress.c | 92 const BYTE *p, in LZ4_putPositionOnHash() 96 const BYTE *srcBase) in LZ4_putPositionOnHash() 101 const BYTE **hashTable = (const BYTE **)tableBase; in LZ4_putPositionOnHash() 124 const BYTE *p, in LZ4_putPosition() 127 const BYTE *srcBase) in LZ4_putPosition() 134 static const BYTE *LZ4_getPositionOnHash( in LZ4_getPositionOnHash() 138 const BYTE *srcBase) in LZ4_getPositionOnHash() 141 const BYTE **hashTable = (const BYTE **) tableBase; in LZ4_getPositionOnHash() 160 static FORCE_INLINE const BYTE *LZ4_getPosition( in LZ4_getPosition() 161 const BYTE *p, in LZ4_getPosition() [all …]
|
D | lz4hc_compress.c | 61 static void LZ4HC_init(LZ4HC_CCtx_internal *hc4, const BYTE *start) in LZ4HC_init() 75 const BYTE *ip) in LZ4HC_Insert() 79 const BYTE * const base = hc4->base; in LZ4HC_Insert() 101 const BYTE *ip, in LZ4HC_InsertAndFindBestMatch() 102 const BYTE * const iLimit, in LZ4HC_InsertAndFindBestMatch() 103 const BYTE **matchpos, in LZ4HC_InsertAndFindBestMatch() 108 const BYTE * const base = hc4->base; in LZ4HC_InsertAndFindBestMatch() 109 const BYTE * const dictBase = hc4->dictBase; in LZ4HC_InsertAndFindBestMatch() 126 const BYTE * const match = base + matchIndex; in LZ4HC_InsertAndFindBestMatch() 139 const BYTE * const match = dictBase + matchIndex; in LZ4HC_InsertAndFindBestMatch() [all …]
|
D | lz4_decompress.c | 76 const BYTE * const lowPrefix, in LZ4_decompress_generic() 78 const BYTE * const dictStart, in LZ4_decompress_generic() 83 const BYTE *ip = (const BYTE *) src; in LZ4_decompress_generic() 84 const BYTE * const iend = ip + srcSize; in LZ4_decompress_generic() 86 BYTE *op = (BYTE *) dst; in LZ4_decompress_generic() 87 BYTE * const oend = op + outputSize; in LZ4_decompress_generic() 88 BYTE *cpy; in LZ4_decompress_generic() 90 const BYTE * const dictEnd = (const BYTE *)dictStart + dictSize; in LZ4_decompress_generic() 98 const BYTE *const shortiend = iend - in LZ4_decompress_generic() 100 const BYTE *const shortoend = oend - in LZ4_decompress_generic() [all …]
|
D | lz4defs.h | 50 typedef uint8_t BYTE; typedef 175 BYTE *d = (BYTE *)dstPtr; in LZ4_wildCopy() 176 const BYTE *s = (const BYTE *)srcPtr; in LZ4_wildCopy() 177 BYTE *const e = (BYTE *)dstEnd; in LZ4_wildCopy() 196 const BYTE *pIn, in LZ4_count() 197 const BYTE *pMatch, in LZ4_count() 198 const BYTE *pInLimit) in LZ4_count() 200 const BYTE *const pStart = pIn; in LZ4_count()
|
/linux-6.1.9/lib/zstd/compress/ |
D | zstd_compress_superblock.c | 32 BYTE hufDesBuffer[ZSTD_MAX_HUF_HEADER_SIZE]; 45 BYTE fseTablesBuffer[ZSTD_MAX_FSE_HEADERS_SIZE]; 68 BYTE* const wkspStart = (BYTE*)workspace; in ZSTD_buildSuperBlockEntropy_literal() 69 BYTE* const wkspEnd = wkspStart + wkspSize; in ZSTD_buildSuperBlockEntropy_literal() 70 BYTE* const countWkspStart = wkspStart; in ZSTD_buildSuperBlockEntropy_literal() 73 BYTE* const nodeWksp = countWkspStart + countWkspSize; in ZSTD_buildSuperBlockEntropy_literal() 101 …{ size_t const largest = HIST_count_wksp (countWksp, &maxSymbolValue, (const BYTE*)src, srcSize,… in ZSTD_buildSuperBlockEntropy_literal() 171 BYTE* const wkspStart = (BYTE*)workspace; in ZSTD_buildSuperBlockEntropy_sequences() 172 BYTE* const wkspEnd = wkspStart + wkspSize; in ZSTD_buildSuperBlockEntropy_sequences() 173 BYTE* const countWkspStart = wkspStart; in ZSTD_buildSuperBlockEntropy_sequences() [all …]
|
D | hist.c | 32 const BYTE* ip = (const BYTE*)src; in HIST_count_simple() 33 const BYTE* const end = ip + srcSize; in HIST_count_simple() 72 const BYTE* ip = (const BYTE*)source; in HIST_count_parallel_wksp() 73 const BYTE* const iend = ip+sourceSize; in HIST_count_parallel_wksp() 94 Counting1[(BYTE) c ]++; in HIST_count_parallel_wksp() 95 Counting2[(BYTE)(c>>8) ]++; in HIST_count_parallel_wksp() 96 Counting3[(BYTE)(c>>16)]++; in HIST_count_parallel_wksp() 99 Counting1[(BYTE) c ]++; in HIST_count_parallel_wksp() 100 Counting2[(BYTE)(c>>8) ]++; in HIST_count_parallel_wksp() 101 Counting3[(BYTE)(c>>16)]++; in HIST_count_parallel_wksp() [all …]
|
D | zstd_fast.c | 23 const BYTE* const base = ms->window.base; in ZSTD_fillHashTable() 24 const BYTE* ip = base + ms->nextToUpdate; in ZSTD_fillHashTable() 25 const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE; in ZSTD_fillHashTable() 57 const BYTE* const base = ms->window.base; in ZSTD_compressBlock_fast_generic() 58 const BYTE* const istart = (const BYTE*)src; in ZSTD_compressBlock_fast_generic() 60 const BYTE* ip0 = istart; in ZSTD_compressBlock_fast_generic() 61 const BYTE* ip1; in ZSTD_compressBlock_fast_generic() 62 const BYTE* anchor = istart; in ZSTD_compressBlock_fast_generic() 65 const BYTE* const prefixStart = base + prefixStartIndex; in ZSTD_compressBlock_fast_generic() 66 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_fast_generic() [all …]
|
D | zstd_ldm.c | 70 BYTE const* data, size_t size, in ZSTD_ldm_gear_feed() 157 BYTE* const pOffset = ldmState->bucketOffsets + hash; in ZSTD_ldm_insertEntry() 161 *pOffset = (BYTE)((offset + 1) & ((1u << ldmParams.bucketSizeLog) - 1)); in ZSTD_ldm_insertEntry() 170 const BYTE* pIn, const BYTE* pAnchor, in ZSTD_ldm_countBackwardsMatch() 171 const BYTE* pMatch, const BYTE* pMatchBase) in ZSTD_ldm_countBackwardsMatch() 188 const BYTE* pIn, const BYTE* pAnchor, in ZSTD_ldm_countBackwardsMatch_2segments() 189 const BYTE* pMatch, const BYTE* pMatchBase, in ZSTD_ldm_countBackwardsMatch_2segments() 190 const BYTE* pExtDictStart, const BYTE* pExtDictEnd) in ZSTD_ldm_countBackwardsMatch_2segments() 213 const BYTE* const iend = (const BYTE*)end; in ZSTD_ldm_fillFastTables() 241 ldmState_t* ldmState, const BYTE* ip, in ZSTD_ldm_fillHashTable() [all …]
|
D | zstd_lazy.c | 21 const BYTE* ip, const BYTE* iend, in ZSTD_updateDUBT() 32 const BYTE* const base = ms->window.base; in ZSTD_updateDUBT() 65 U32 curr, const BYTE* inputEnd, in ZSTD_insertDUBT1() 74 const BYTE* const base = ms->window.base; in ZSTD_insertDUBT1() 75 const BYTE* const dictBase = ms->window.dictBase; in ZSTD_insertDUBT1() 77 const BYTE* const ip = (curr>=dictLimit) ? base + curr : dictBase + curr; in ZSTD_insertDUBT1() 78 const BYTE* const iend = (curr>=dictLimit) ? inputEnd : dictBase + dictLimit; in ZSTD_insertDUBT1() 79 const BYTE* const dictEnd = dictBase + dictLimit; in ZSTD_insertDUBT1() 80 const BYTE* const prefixStart = base + dictLimit; in ZSTD_insertDUBT1() 81 const BYTE* match; in ZSTD_insertDUBT1() [all …]
|
D | zstd_double_fast.c | 24 const BYTE* const base = ms->window.base; in ZSTD_fillDoubleHashTable() 25 const BYTE* ip = base + ms->nextToUpdate; in ZSTD_fillDoubleHashTable() 26 const BYTE* const iend = ((const BYTE*)end) - HASH_READ_SIZE; in ZSTD_fillDoubleHashTable() 61 const BYTE* const base = ms->window.base; in ZSTD_compressBlock_doubleFast_generic() 62 const BYTE* const istart = (const BYTE*)src; in ZSTD_compressBlock_doubleFast_generic() 63 const BYTE* ip = istart; in ZSTD_compressBlock_doubleFast_generic() 64 const BYTE* anchor = istart; in ZSTD_compressBlock_doubleFast_generic() 68 const BYTE* const prefixLowest = base + prefixLowestIndex; in ZSTD_compressBlock_doubleFast_generic() 69 const BYTE* const iend = istart + srcSize; in ZSTD_compressBlock_doubleFast_generic() 70 const BYTE* const ilimit = iend - HASH_READ_SIZE; in ZSTD_compressBlock_doubleFast_generic() [all …]
|
D | zstd_compress_internal.h | 138 BYTE const* nextSrc; /* next block here to continue on current prefix */ 139 BYTE const* base; /* All regular indexes relative to this position */ 140 BYTE const* dictBase; /* extDict indexes relative to this position */ 181 BYTE const* split; 193 BYTE* bucketOffsets; /* Next position in bucket to insert entry */ 366 static const BYTE LL_Code[64] = { 0, 1, 2, 3, 4, 5, 6, 7, in ZSTD_LLcode() 383 … static const BYTE ML_Code[128] = { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, in ZSTD_MLcode() 441 ZSTD_memcpy((BYTE*)dst + ZSTD_blockHeaderSize, src, srcSize); in ZSTD_noCompressBlock() 445 MEM_STATIC size_t ZSTD_rleCompressBlock (void* dst, size_t dstCapacity, BYTE src, size_t srcSize, U… in ZSTD_rleCompressBlock() 447 BYTE* const op = (BYTE*)dst; in ZSTD_rleCompressBlock() [all …]
|
D | zstd_cwksp.h | 147 BYTE allocFailed; 211 ws->allocStart = (BYTE*)ws->allocStart - ((size_t)ws->allocStart & (sizeof(U32)-1)); in ZSTD_cwksp_internal_advance_phase() 235 alloc = (BYTE *)ws->allocStart - bytes; in ZSTD_cwksp_reserve_internal() 262 MEM_STATIC BYTE* ZSTD_cwksp_reserve_buffer(ZSTD_cwksp* ws, size_t bytes) { in ZSTD_cwksp_reserve_buffer() 263 return (BYTE*)ZSTD_cwksp_reserve_internal(ws, bytes, ZSTD_cwksp_alloc_buffers); in ZSTD_cwksp_reserve_buffer() 282 void* end = (BYTE *)alloc + bytes; in ZSTD_cwksp_reserve_table() 308 void* end = (BYTE*)alloc + roundedBytes; in ZSTD_cwksp_reserve_object() 359 ZSTD_memset(ws->tableValidEnd, 0, (BYTE*)ws->tableEnd - (BYTE*)ws->tableValidEnd); in ZSTD_cwksp_clean_tables() 403 ws->workspaceEnd = (BYTE*)start + size; in ZSTD_cwksp_init() 438 return (size_t)((BYTE*)ws->workspaceEnd - (BYTE*)ws->workspace); in ZSTD_cwksp_sizeof() [all …]
|
D | huf_compress.c | 69 BYTE* const ostart = (BYTE*) dst; in HUF_compressWeights() 70 BYTE* op = ostart; in HUF_compressWeights() 71 BYTE* const oend = ostart + dstSize; in HUF_compressWeights() 109 BYTE bitsToWeight[HUF_TABLELOG_MAX + 1]; /* precomputed conversion table */ 110 BYTE huffWeight[HUF_SYMBOLVALUE_MAX]; 117 BYTE* op = (BYTE*)dst; in HUF_writeCTable_wksp() 128 wksp->bitsToWeight[n] = (BYTE)(huffLog + 1 - n); in HUF_writeCTable_wksp() 135 op[0] = (BYTE)hSize; in HUF_writeCTable_wksp() 142 op[0] = (BYTE)(128 /*special case*/ + (maxSymbolValue-1)); in HUF_writeCTable_wksp() 145 op[(n/2)+1] = (BYTE)((wksp->huffWeight[n] << 4) + wksp->huffWeight[n+1]); in HUF_writeCTable_wksp() [all …]
|
D | zstd_compress_sequences.c | 73 BYTE wksp[FSE_NCOUNTBOUND]; in ZSTD_NCountCost() 244 const BYTE* codeTable, size_t nbSeq, in ZSTD_buildCTable() 249 BYTE* op = (BYTE*)dst; in ZSTD_buildCTable() 250 const BYTE* const oend = op + dstCapacity; in ZSTD_buildCTable() 255 FORWARD_IF_ERROR(FSE_buildCTable_rle(nextCTable, (BYTE)max), ""); in ZSTD_buildCTable() 290 FSE_CTable const* CTable_MatchLength, BYTE const* mlCodeTable, in ZSTD_encodeSequences_body() 291 FSE_CTable const* CTable_OffsetBits, BYTE const* ofCodeTable, in ZSTD_encodeSequences_body() 292 FSE_CTable const* CTable_LitLength, BYTE const* llCodeTable, in ZSTD_encodeSequences_body() 331 BYTE const llCode = llCodeTable[n]; in ZSTD_encodeSequences_body() 332 BYTE const ofCode = ofCodeTable[n]; in ZSTD_encodeSequences_body() [all …]
|
D | zstd_compress_literals.c | 18 BYTE* const ostart = (BYTE*)dst; in ZSTD_noCompressLiterals() 26 ostart[0] = (BYTE)((U32)set_basic + (srcSize<<3)); in ZSTD_noCompressLiterals() 45 BYTE* const ostart = (BYTE*)dst; in ZSTD_compressRleLiteralsBlock() 53 ostart[0] = (BYTE)((U32)set_rle + (srcSize<<3)); in ZSTD_compressRleLiteralsBlock() 65 ostart[flSize] = *(const BYTE*)src; in ZSTD_compressRleLiteralsBlock() 80 BYTE* const ostart = (BYTE*)dst; in ZSTD_compressLiterals() 150 ostart[4] = (BYTE)(cLitSize >> 10); in ZSTD_compressLiterals()
|
D | fse_compress.c | 194 BYTE* const ostart = (BYTE*) header; in FSE_writeNCount_generic() 195 BYTE* out = ostart; in FSE_writeNCount_generic() 196 BYTE* const oend = ostart + headerBufferSize; in FSE_writeNCount_generic() 226 out[0] = (BYTE) bitStream; in FSE_writeNCount_generic() 227 out[1] = (BYTE)(bitStream>>8); in FSE_writeNCount_generic() 241 out[0] = (BYTE)bitStream; in FSE_writeNCount_generic() 242 out[1] = (BYTE)(bitStream>>8); in FSE_writeNCount_generic() 263 out[0] = (BYTE)bitStream; in FSE_writeNCount_generic() 264 out[1] = (BYTE)(bitStream>>8); in FSE_writeNCount_generic() 277 out[0] = (BYTE)bitStream; in FSE_writeNCount_generic() [all …]
|
/linux-6.1.9/lib/zstd/decompress/ |
D | huf_decompress.c | 101 typedef struct { BYTE maxTableLog; BYTE tableType; BYTE tableLog; BYTE reserved; } DTableDesc; 116 typedef struct { BYTE byte; BYTE nbBits; } HUF_DEltX1; /* single-symbol decoding */ 122 static U64 HUF_DEltX1_set4(BYTE symbol, BYTE nbBits) { in HUF_DEltX1_set4() 137 BYTE symbols[HUF_SYMBOLVALUE_MAX + 1]; 138 BYTE huffWeight[HUF_SYMBOLVALUE_MAX + 1]; 169 dtd.tableLog = (BYTE)tableLog; in HUF_readDTableX1_wksp_bmi2() 199 wksp->symbols[wksp->rankStart[w]++] = (BYTE)(n+u); in HUF_readDTableX1_wksp_bmi2() 204 wksp->symbols[wksp->rankStart[w]++] = (BYTE)n; in HUF_readDTableX1_wksp_bmi2() 222 BYTE const nbBits = (BYTE)(tableLog + 1 - w); in HUF_readDTableX1_wksp_bmi2() 281 FORCE_INLINE_TEMPLATE BYTE [all …]
|
D | zstd_decompress_block.c | 85 { const BYTE* const istart = (const BYTE*) src; in ZSTD_decodeLiteralsBlock() 358 cell->nbAdditionalBits = (BYTE)nbAddBits; in ZSTD_buildSeqTable_rle() 378 BYTE* spread = (BYTE*)(symbolNext + MaxSeq + 1); in ZSTD_buildFSETable_body() 478 tableDecode[u].nbBits = (BYTE) (tableLog - BIT_highbit32(nextState) ); in ZSTD_buildFSETable_body() 481 tableDecode[u].nbAdditionalBits = (BYTE)nbAdditionalBits[symbol]; in ZSTD_buildFSETable_body() 541 RETURN_ERROR_IF((*(const BYTE*)src) > max, corruption_detected, ""); in ZSTD_buildSeqTable() 542 { U32 const symbol = *(const BYTE*)src; in ZSTD_buildSeqTable() 580 const BYTE* const istart = (const BYTE*)src; in ZSTD_decodeSeqHeaders() 581 const BYTE* const iend = istart + srcSize; in ZSTD_decodeSeqHeaders() 582 const BYTE* ip = istart; in ZSTD_decodeSeqHeaders() [all …]
|
/linux-6.1.9/Documentation/scsi/ |
D | arcmsr_spec.rst | 444 BYTE grsRaidSetName[16]; 448 BYTE grsDevArray[32]; 449 BYTE grsMemberDevices; 450 BYTE grsNewMemberDevices; 451 BYTE grsRaidState; 452 BYTE grsVolumes; 453 BYTE grsVolumeList[16]; 454 BYTE grsRes1; 455 BYTE grsRes2; 456 BYTE grsRes3; [all …]
|
/linux-6.1.9/lib/zstd/common/ |
D | zstd_internal.h | 265 ptrdiff_t diff = (BYTE*)dst - (const BYTE*)src; in ZSTD_wildcopy() 266 const BYTE* ip = (const BYTE*)src; in ZSTD_wildcopy() 267 BYTE* op = (BYTE*)dst; in ZSTD_wildcopy() 268 BYTE* const oend = op + length; in ZSTD_wildcopy() 342 BYTE* litStart; 343 BYTE* lit; /* ptr to end of literals */ 344 BYTE* llCode; 345 BYTE* mlCode; 346 BYTE* ofCode;
|
D | fse_decompress.c | 76 BYTE* spread = (BYTE*)(symbolNext + maxSymbolValue + 1); in FSE_buildDTable_internal() 169 tableDecode[u].nbBits = (BYTE) (tableLog - BIT_highbit32(nextState) ); in FSE_buildDTable_internal() 187 size_t FSE_buildDTable_rle (FSE_DTable* dt, BYTE symbolValue) in FSE_buildDTable_rle() 224 dinfo[s].symbol = (BYTE)s; in FSE_buildDTable_raw() 225 dinfo[s].nbBits = (BYTE)nbBits; in FSE_buildDTable_raw() 236 BYTE* const ostart = (BYTE*) dst; in FSE_decompress_usingDTable_generic() 237 BYTE* op = ostart; in FSE_decompress_usingDTable_generic() 238 BYTE* const omax = op + maxDstSize; in FSE_decompress_usingDTable_generic() 239 BYTE* const olimit = omax-3; in FSE_decompress_usingDTable_generic() 325 const BYTE* const istart = (const BYTE*)cSrc; in FSE_decompress_wksp_body() [all …]
|
D | entropy_common.c | 63 const BYTE* const istart = (const BYTE*) headerBuffer; in FSE_readNCount_body() 64 const BYTE* const iend = istart + hbSize; in FSE_readNCount_body() 65 const BYTE* ip = istart; in FSE_readNCount_body() 252 size_t HUF_readStats(BYTE* huffWeight, size_t hwSize, U32* rankStats, in HUF_readStats() 262 HUF_readStats_body(BYTE* huffWeight, size_t hwSize, U32* rankStats, in HUF_readStats_body() 269 const BYTE* ip = (const BYTE*) src; in HUF_readStats_body() 315 huffWeight[oSize] = (BYTE)lastWeight; in HUF_readStats_body() 328 static size_t HUF_readStats_body_default(BYTE* huffWeight, size_t hwSize, U32* rankStats, in HUF_readStats_body_default() 337 static TARGET_ATTRIBUTE("bmi2") size_t HUF_readStats_body_bmi2(BYTE* huffWeight, size_t hwSize, U32… in HUF_readStats_body_bmi2() 346 size_t HUF_readStats_wksp(BYTE* huffWeight, size_t hwSize, U32* rankStats, in HUF_readStats_wksp()
|
D | bitstream.h | 268 { BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1]; in BIT_initDStream() 273 bitD->bitContainer = *(const BYTE*)(bitD->start); in BIT_initDStream() 276 …case 7: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[6]) << (sizeof(bitD->bitContaine… in BIT_initDStream() 279 …case 6: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[5]) << (sizeof(bitD->bitContaine… in BIT_initDStream() 282 …case 5: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[4]) << (sizeof(bitD->bitContaine… in BIT_initDStream() 285 case 4: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[3]) << 24; in BIT_initDStream() 288 case 3: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[2]) << 16; in BIT_initDStream() 291 case 2: bitD->bitContainer += (size_t)(((const BYTE*)(srcBuffer))[1]) << 8; in BIT_initDStream() 296 { BYTE const lastByte = ((const BYTE*)srcBuffer)[srcSize-1]; in BIT_initDStream()
|
/linux-6.1.9/arch/ia64/include/asm/ |
D | module.lds.h | 10 .core.plt : { BYTE(0) } 11 .init.plt : { BYTE(0) } 12 .got : { BYTE(0) } 13 .opd : { BYTE(0) }
|