Lines Matching refs:imm32

80 		return 4; /* imm32 */
596 u32 dst_reg, const u32 imm32)
603 * Optimization: if imm32 is positive, use 'mov %eax, imm32'
604 * (which zero-extends imm32) to save 2 bytes.
606 if (sign_propagate && (s32)imm32 < 0) {
607 /* 'mov %rax, imm32' sign extends imm32 */
611 EMIT3_off32(b1, b2, add_1reg(b3, dst_reg), imm32);
616 * Optimization: if imm32 is zero, use 'xor %eax, %eax'
619 if (imm32 == 0) {
628 /* mov %eax, imm32 */
631 EMIT1_off32(add_1reg(0xB8, dst_reg), imm32);
647 * 'mov %eax, imm32' instead.
813 const s32 imm32 = insn->imm;
907 if (is_imm8(imm32))
908 EMIT3(0x83, add_1reg(b3, dst_reg), imm32);
910 EMIT1_off32(b2, imm32);
912 EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32);
918 dst_reg, imm32);
927 /* dst %= src, dst /= src, dst %= imm32, dst /= imm32 */
943 /* mov r11, imm32 */
944 EMIT3_off32(0x49, 0xC7, 0xC3, imm32);
994 emit_mov_imm32(&prog, is64, BPF_REG_0, imm32);
1030 if (imm32 == 1)
1033 EMIT3(0xC1, add_1reg(b3, dst_reg), imm32);
1079 switch (imm32) {
1111 switch (imm32) {
1168 EMIT(imm32, bpf_size_to_x86_bytes(BPF_SIZE(insn->code)));
1252 func = (u8 *) __bpf_call_base + imm32;
1257 if (!imm32 || emit_call(&prog, func, image + addrs[i - 1] + 7))
1260 if (!imm32 || emit_call(&prog, func, image + addrs[i - 1]))
1266 if (imm32)
1267 emit_bpf_tail_call_direct(&bpf_prog->aux->poke_tab[imm32 - 1],
1321 /* test dst_reg, imm32 */
1326 EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32);
1350 if (imm32 == 0) {
1365 if (is_imm8(imm32))
1366 EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32);
1368 EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32);