/third_party/libunwind/libunwind/src/riscv/ |
H A D | setcontext.S | 56 LOAD t1, REG(0) 57 LOAD ra, REG(1) 58 LOAD sp, REG(2) 59 LOAD s0, REG(8) 60 LOAD s1, REG(9) 61 LOAD a1, REG(11) 62 LOAD a2, REG(12) 63 LOAD a3, REG(13) 64 LOAD a4, REG(14) 65 LOAD a [all...] |
H A D | asm.h | 27 # define LOAD lw macro 31 # define LOAD ld macro
|
/third_party/ffmpeg/libavfilter/x86/ |
H A D | vf_bwdif.asm | 60 LOAD%4 m0, [curq+t0*%5] 61 LOAD%4 m1, [curq+t1*%5] 62 LOAD%4 m2, [%2] 63 LOAD%4 m3, [%3] 70 LOAD%4 m3, [prevq+t0*%5] 71 LOAD%4 m4, [prevq+t1*%5] 79 LOAD%4 m3, [nextq+t0*%5] 80 LOAD%4 m4, [nextq+t1*%5] 88 LOAD%4 m3, [%2+t0*2*%5] 89 LOAD [all...] |
H A D | vf_yadif.asm | 83 %macro LOAD 2 91 LOAD m0, [curq+t1] 92 LOAD m1, [curq+t0] 93 LOAD m2, [%2] 94 LOAD m3, [%3] 103 LOAD m3, [prevq+t1] 104 LOAD m4, [prevq+t0] 113 LOAD m3, [nextq+t1] 114 LOAD m4, [nextq+t0] 156 LOAD m [all...] |
H A D | yadif-16.asm | 135 %macro LOAD 2 143 LOAD m0, [curq+t1] 144 LOAD m1, [curq+t0] 145 LOAD m2, [%2] 146 LOAD m3, [%3] 155 LOAD m3, [prevq+t1] 156 LOAD m4, [prevq+t0] 165 LOAD m3, [nextq+t1] 166 LOAD m4, [nextq+t0] 208 LOAD m [all...] |
H A D | yadif-10.asm | 105 %macro LOAD 2 112 LOAD m0, [curq+t1] 113 LOAD m1, [curq+t0] 114 LOAD m2, [%2] 115 LOAD m3, [%3] 124 LOAD m3, [prevq+t1] 125 LOAD m4, [prevq+t0] 133 LOAD m3, [nextq+t1] 134 LOAD m4, [nextq+t0] 173 LOAD m [all...] |
H A D | vf_removegrain.asm | 53 %macro LOAD 3 72 LOAD m1, [a1], %1 73 LOAD m2, [a2], %1 74 LOAD m3, [a3], %1 75 LOAD m4, [a4], %1 76 LOAD m0, [c], %1 77 LOAD m5, [a5], %1 78 LOAD m6, [a6], %1 79 LOAD m7, [a7], %1 80 LOAD m [all...] |
H A D | vf_hqdn3d.asm | 34 %macro LOAD 3 ; dstreg, x, bitdepth 73 LOAD pixelantd, xq, %1 77 LOAD t0d, xq+1, %1 ; skip on the last iteration to avoid overread
|
/third_party/node/deps/base64/base64/lib/arch/avx2/ |
H A D | enc_loop_asm.c | 10 #define LOAD(R0, ROUND, OFFSET) \ macro 41 LOAD("a", 0, -4) \ 60 LOAD("a", 0, -4) /* + */ \ 62 LOAD("b", 1, -4) /* | + | */ \ 64 LOAD("c", 2, -4) /* V V V */ 74 LOAD(D, (ROUND + 3), -4) /* V V V + */ \ 78 LOAD(E, (ROUND + 4), -4) /* | | | + */ \ 82 LOAD(F, (ROUND + 5), -4) /* | | | + */ \ 173 LOAD("a", 0, 0) in enc_loop_avx2()
|
/third_party/ffmpeg/libavcodec/alpha/ |
H A D | hpeldsp_alpha.c | 55 #define OP(LOAD, STORE) \ 57 STORE(LOAD(pixels), block); \ 62 #define OP_X2(LOAD, STORE) \ 66 pix1 = LOAD(pixels); \ 73 #define OP_Y2(LOAD, STORE) \ 75 uint64_t pix = LOAD(pixels); \ 80 next_pix = LOAD(pixels); \ 87 #define OP_XY2(LOAD, STORE) \ 89 uint64_t pix1 = LOAD(pixels); \ 102 npix1 = LOAD(pixel [all...] |
/third_party/node/deps/base64/base64/lib/arch/avx/ |
H A D | enc_loop_asm.c | 9 #define LOAD(R0, ROUND) \ macro 40 LOAD("a", 0) \ 59 LOAD("a", 0) /* + */ \ 61 LOAD("b", 1) /* | + | */ \ 63 LOAD("c", 2) /* V V V */ 73 LOAD(D, (ROUND + 3)) /* V V V + */ \ 77 LOAD(E, (ROUND + 4)) /* | | | + */ \ 81 LOAD(F, (ROUND + 5)) /* | | | + */ \
|
/third_party/node/deps/base64/base64/lib/arch/ssse3/ |
H A D | enc_loop_asm.c | 9 #define LOAD(R0, ROUND) \ macro 44 LOAD("a", 0) \ 63 LOAD("a", 0) /* + */ \ 65 LOAD("b", 1) /* | + | */ \ 67 LOAD("c", 2) /* V V V */ 77 LOAD(D, (ROUND + 3)) /* V V V + */ \ 81 LOAD(E, (ROUND + 4)) /* | | | + */ \ 85 LOAD(F, (ROUND + 5)) /* | | | + */ \
|
/third_party/ffmpeg/libavcodec/x86/ |
H A D | fpel.asm | 29 LOAD %3, %1 42 %define LOAD movh 46 %define LOAD movu 61 LOAD m0, [r1 + %%i] 62 LOAD m1, [r1+r2 + %%i] 63 LOAD m2, [r1+r2*2 + %%i] 64 LOAD m3, [r1+r4 + %%i]
|
H A D | rv40dsp.asm | 83 %macro LOAD 2 111 LOAD my, sixtap_filter_v 182 LOAD mx, sixtap_filter_v 247 LOAD my, sixtap_filter_hb 291 LOAD mx, sixtap_filter_hb
|
/third_party/node/deps/base64/base64/lib/arch/neon64/ |
H A D | enc_loop_asm.c | 10 #define LOAD(A, B, C) \ macro 42 LOAD("v12", "v13", "v14") \ 52 LOAD("v12", "v13", "v14") \ 59 LOAD("v2", "v3", "v4") \ 65 LOAD("v2", "v3", "v4") \
|
/third_party/skia/third_party/externals/angle2/src/compiler/translator/ |
H A D | AtomicCounterFunctionHLSL.cpp | 42 atomicMethod = AtomicCounterFunction::LOAD; in useAtomicCounterFunction() 85 case AtomicCounterFunction::LOAD: in atomicCounterFunctionHeader() 93 if (mForceResolution && atomicFunction.second != AtomicCounterFunction::LOAD) in atomicCounterFunctionHeader()
|
H A D | ImageFunctionHLSL.cpp | 53 if (imageFunction.method == ImageFunctionHLSL::ImageFunction::Method::LOAD || in OutputImageFunctionArgumentList() 195 case Method::LOAD: in name() 262 else if (method == ImageFunction::Method::LOAD) in getReturnType() 320 imageFunction.method = ImageFunction::Method::LOAD; in useImageFunction() 357 else if (imageFunction.method == ImageFunction::Method::LOAD) in imageFunctionHeader()
|
H A D | AtomicCounterFunctionHLSL.h | 36 LOAD, member in sh::final::AtomicCounterFunction
|
/third_party/node/deps/openssl/openssl/crypto/ |
H A D | mem.c | 31 # define LOAD(x) 0 macro 38 # define LOAD(x) tsan_load(&x) macro 88 *mcount = LOAD(malloc_count); in CRYPTO_get_alloc_counts() 90 *rcount = LOAD(realloc_count); in CRYPTO_get_alloc_counts() 92 *fcount = LOAD(free_count); in CRYPTO_get_alloc_counts()
|
/third_party/openssl/crypto/ |
H A D | mem.c | 31 # define LOAD(x) 0 macro 38 # define LOAD(x) tsan_load(&x) macro 88 *mcount = LOAD(malloc_count); in CRYPTO_get_alloc_counts() 90 *rcount = LOAD(realloc_count); in CRYPTO_get_alloc_counts() 92 *fcount = LOAD(free_count); in CRYPTO_get_alloc_counts()
|
/third_party/ffmpeg/libavfilter/ |
H A D | vf_hqdn3d.c | 45 #define LOAD(x) (((depth == 8 ? src[x] : AV_RN16A(src + (x) * 2)) << (16 - depth))\ macro 70 frame_ant[x] = tmp = lowpass(frame_ant[x], LOAD(x), temporal, depth); in denoise_temporal() 95 pixel_ant = LOAD(0); in denoise_spatial() 97 line_ant[x] = tmp = pixel_ant = lowpass(pixel_ant, LOAD(x), spatial, depth); in denoise_spatial() 110 pixel_ant = LOAD(0); in denoise_spatial() 113 pixel_ant = lowpass(pixel_ant, LOAD(x+1), spatial, depth); in denoise_spatial() 141 frame_ant[x] = LOAD(x); in denoise_depth()
|
/third_party/rust/crates/clang-sys/src/ |
H A D | link.rs | 10 @LOAD: 30 @LOAD: 33 link!(@LOAD: #[cfg(feature = "runtime")] fn $name($($pname: $pty), *) $(-> $ret)*); 165 $(link!(@LOAD: $(#[cfg($cfg)])* fn $name($($pname: $pty), *) $(-> $ret)*);)+
|
/third_party/node/deps/v8/src/compiler/ |
H A D | machine-operator.cc | 923 #define LOAD(Type) \ macro 956 MACHINE_TYPE_LIST(LOAD) 957 #undef LOAD macro 1338 #define LOAD(Type) \ in UnalignedLoad() macro 1342 MACHINE_TYPE_LIST(LOAD) in UnalignedLoad() 1343 #undef LOAD in UnalignedLoad() macro 1431 #define LOAD(Type) \ in Load() macro 1435 MACHINE_TYPE_LIST(LOAD) in Load() 1436 #undef LOAD in Load() macro 1447 #define LOAD(Typ in LoadImmutable() macro 1452 #undef LOAD LoadImmutable() macro 1457 #define LOAD ProtectedLoad() macro 1462 #undef LOAD ProtectedLoad() macro 1681 #define LOAD Word32AtomicLoad() macro 1689 #undef LOAD Word32AtomicLoad() macro 1798 #define LOAD Word64AtomicLoad() macro 1806 #undef LOAD Word64AtomicLoad() macro [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/AMDGPU/ |
H A D | SIMemoryLegalizer.cpp | 57 LOAD = 1u << 0, member in __anon24492::SIMemOp 1027 if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) in insertWait() 1039 if ((Op & SIMemOp::LOAD) != SIMemOp::NONE) in insertWait() 1152 SIMemOp::LOAD | SIMemOp::STORE, in expandLoad() 1160 SIMemOp::LOAD, in expandLoad() 1191 SIMemOp::LOAD | SIMemOp::STORE, in expandStore() 1228 SIMemOp::LOAD | SIMemOp::STORE, in expandAtomicFence() 1258 SIMemOp::LOAD | SIMemOp::STORE, in expandAtomicCmpxchgOrRmw() 1269 isAtomicRet(*MI) ? SIMemOp::LOAD : in expandAtomicCmpxchgOrRmw()
|
/third_party/elfutils/libcpu/ |
H A D | bpf_disasm.c | 68 #define LOAD(T) REG(1) " = *(" #T " *)(" REG(2) OFF(3) ")" macro 397 code_fmt = LOAD(u8); in bpf_disasm() 400 code_fmt = LOAD(u16); in bpf_disasm() 403 code_fmt = LOAD(u32); in bpf_disasm() 406 code_fmt = LOAD(u64); in bpf_disasm()
|