Lines Matching refs:endian
75 * By default, xxHash library provides endian-independent Hash values, based on little-endian convention.
76 * Results are therefore identical for little-endian and big-endian CPU.
77 * This comes at a performance cost for big-endian CPU, since some swapping is required to emulate little-endian format.
78 * Should endian-independence be of no importance for your application, you may set the #define below to 1,
79 * to improve speed for Big-endian CPU.
234 FORCE_INLINE U32 XXH_readLE32_align(const void* ptr, XXH_endianess endian, XXH_alignment align)
237 return endian==XXH_littleEndian ? XXH_read32(ptr) : XXH_swap32(XXH_read32(ptr));
239 return endian==XXH_littleEndian ? *(const U32*)ptr : XXH_swap32(*(const U32*)ptr);
242 FORCE_INLINE U32 XXH_readLE32(const void* ptr, XXH_endianess endian)
244 return XXH_readLE32_align(ptr, endian, XXH_unaligned);
288 #define XXH_get32bits(p) XXH_readLE32_align(p, endian, align)
292 XXH_endianess endian, XXH_alignment align)
353 XXH_endianess endian, XXH_alignment align)
388 return XXH32_finalize(h32, p, len&15, endian, align);
452 XXH32_update_endian(XXH32_state_t* state, const void* input, size_t len, XXH_endianess endian)
476 state->v1 = XXH32_round(state->v1, XXH_readLE32(p32, endian)); p32++;
477 state->v2 = XXH32_round(state->v2, XXH_readLE32(p32, endian)); p32++;
478 state->v3 = XXH32_round(state->v3, XXH_readLE32(p32, endian)); p32++;
479 state->v4 = XXH32_round(state->v4, XXH_readLE32(p32, endian));
493 v1 = XXH32_round(v1, XXH_readLE32(p, endian)); p+=4;
494 v2 = XXH32_round(v2, XXH_readLE32(p, endian)); p+=4;
495 v3 = XXH32_round(v3, XXH_readLE32(p, endian)); p+=4;
496 v4 = XXH32_round(v4, XXH_readLE32(p, endian)); p+=4;
527 XXH32_digest_endian (const XXH32_state_t* state, XXH_endianess endian)
542 return XXH32_finalize(h32, state->mem32, state->memsize, endian, XXH_aligned);
560 * The canonical representation follows human-readable write convention, aka big-endian (large digits first).
645 FORCE_INLINE U64 XXH_readLE64_align(const void* ptr, XXH_endianess endian, XXH_alignment align)
648 return endian==XXH_littleEndian ? XXH_read64(ptr) : XXH_swap64(XXH_read64(ptr));
650 return endian==XXH_littleEndian ? *(const U64*)ptr : XXH_swap64(*(const U64*)ptr);
653 FORCE_INLINE U64 XXH_readLE64(const void* ptr, XXH_endianess endian)
655 return XXH_readLE64_align(ptr, endian, XXH_unaligned);
699 #define XXH_get64bits(p) XXH_readLE64_align(p, endian, align)
703 XXH_endianess endian, XXH_alignment align)
812 XXH_endianess endian, XXH_alignment align)
851 return XXH64_finalize(h64, p, len, endian, align);
912 XXH64_update_endian (XXH64_state_t* state, const void* input, size_t len, XXH_endianess endian)
934 state->v1 = XXH64_round(state->v1, XXH_readLE64(state->mem64+0, endian));
935 state->v2 = XXH64_round(state->v2, XXH_readLE64(state->mem64+1, endian));
936 state->v3 = XXH64_round(state->v3, XXH_readLE64(state->mem64+2, endian));
937 state->v4 = XXH64_round(state->v4, XXH_readLE64(state->mem64+3, endian));
950 v1 = XXH64_round(v1, XXH_readLE64(p, endian)); p+=8;
951 v2 = XXH64_round(v2, XXH_readLE64(p, endian)); p+=8;
952 v3 = XXH64_round(v3, XXH_readLE64(p, endian)); p+=8;
953 v4 = XXH64_round(v4, XXH_readLE64(p, endian)); p+=8;
981 FORCE_INLINE U64 XXH64_digest_endian (const XXH64_state_t* state, XXH_endianess endian)
1002 return XXH64_finalize(h64, state->mem64, (size_t)state->total_len, endian, XXH_aligned);