Lines Matching refs:DST

55 #define DST	regs[insn->dst_reg]  macro
1684 DST = DST OP (SRC & 63); \ in ___bpf_prog_run()
1687 DST = (u32) DST OP ((u32) SRC & 31); \ in ___bpf_prog_run()
1690 DST = DST OP IMM; \ in ___bpf_prog_run()
1693 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1698 DST = DST OP SRC; \ in ___bpf_prog_run()
1701 DST = (u32) DST OP (u32) SRC; \ in ___bpf_prog_run()
1704 DST = DST OP IMM; \ in ___bpf_prog_run()
1707 DST = (u32) DST OP (u32) IMM; \ in ___bpf_prog_run()
1720 DST = (u32) -DST; in ___bpf_prog_run()
1723 DST = -DST; in ___bpf_prog_run()
1726 DST = (u32) SRC; in ___bpf_prog_run()
1729 DST = (u32) IMM; in ___bpf_prog_run()
1732 DST = SRC; in ___bpf_prog_run()
1735 DST = IMM; in ___bpf_prog_run()
1738 DST = (u64) (u32) insn[0].imm | ((u64) (u32) insn[1].imm) << 32; in ___bpf_prog_run()
1742 DST = (u64) (u32) (((s32) DST) >> (SRC & 31)); in ___bpf_prog_run()
1745 DST = (u64) (u32) (((s32) DST) >> IMM); in ___bpf_prog_run()
1748 (*(s64 *) &DST) >>= (SRC & 63); in ___bpf_prog_run()
1751 (*(s64 *) &DST) >>= IMM; in ___bpf_prog_run()
1754 div64_u64_rem(DST, SRC, &AX); in ___bpf_prog_run()
1755 DST = AX; in ___bpf_prog_run()
1758 AX = (u32) DST; in ___bpf_prog_run()
1759 DST = do_div(AX, (u32) SRC); in ___bpf_prog_run()
1762 div64_u64_rem(DST, IMM, &AX); in ___bpf_prog_run()
1763 DST = AX; in ___bpf_prog_run()
1766 AX = (u32) DST; in ___bpf_prog_run()
1767 DST = do_div(AX, (u32) IMM); in ___bpf_prog_run()
1770 DST = div64_u64(DST, SRC); in ___bpf_prog_run()
1773 AX = (u32) DST; in ___bpf_prog_run()
1775 DST = (u32) AX; in ___bpf_prog_run()
1778 DST = div64_u64(DST, IMM); in ___bpf_prog_run()
1781 AX = (u32) DST; in ___bpf_prog_run()
1783 DST = (u32) AX; in ___bpf_prog_run()
1788 DST = (__force u16) cpu_to_be16(DST); in ___bpf_prog_run()
1791 DST = (__force u32) cpu_to_be32(DST); in ___bpf_prog_run()
1794 DST = (__force u64) cpu_to_be64(DST); in ___bpf_prog_run()
1801 DST = (__force u16) cpu_to_le16(DST); in ___bpf_prog_run()
1804 DST = (__force u32) cpu_to_le32(DST); in ___bpf_prog_run()
1807 DST = (__force u64) cpu_to_le64(DST); in ___bpf_prog_run()
1865 if ((SIGN##64) DST CMP_OP (SIGN##64) SRC) { \ in ___bpf_prog_run()
1871 if ((SIGN##32) DST CMP_OP (SIGN##32) SRC) { \ in ___bpf_prog_run()
1877 if ((SIGN##64) DST CMP_OP (SIGN##64) IMM) { \ in ___bpf_prog_run()
1883 if ((SIGN##32) DST CMP_OP (SIGN##32) IMM) { \ in ___bpf_prog_run()
1917 *(SIZE *)(unsigned long) (DST + insn->off) = SRC; \ in ___bpf_prog_run()
1920 *(SIZE *)(unsigned long) (DST + insn->off) = IMM; \ in ___bpf_prog_run()
1923 DST = *(SIZE *)(unsigned long) (SRC + insn->off); \ in ___bpf_prog_run()
1926 bpf_probe_read_kernel(&DST, sizeof(SIZE), \ in ___bpf_prog_run()
1928 DST = *((SIZE *)&DST); \ in ___bpf_prog_run()
1941 (DST + insn->off)); \ in ___bpf_prog_run()
1944 (DST + insn->off)); \ in ___bpf_prog_run()
1950 (atomic_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
1954 (atomic64_t *)(unsigned long) (DST + insn->off)); \ in ___bpf_prog_run()
1969 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1973 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1979 (atomic_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()
1983 (atomic64_t *)(unsigned long) (DST + insn->off), in ___bpf_prog_run()