/linux-3.4.99/net/sched/ |
D | em_cmp.c | 20 static inline int cmp_needs_transformation(struct tcf_em_cmp *cmp) in cmp_needs_transformation() argument 22 return unlikely(cmp->flags & TCF_EM_CMP_TRANS); in cmp_needs_transformation() 28 struct tcf_em_cmp *cmp = (struct tcf_em_cmp *) em->data; in em_cmp_match() local 29 unsigned char *ptr = tcf_get_base_ptr(skb, cmp->layer) + cmp->off; in em_cmp_match() 32 if (!tcf_valid_offset(skb, ptr, cmp->align)) in em_cmp_match() 35 switch (cmp->align) { in em_cmp_match() 43 if (cmp_needs_transformation(cmp)) in em_cmp_match() 53 if (cmp_needs_transformation(cmp)) in em_cmp_match() 61 if (cmp->mask) in em_cmp_match() 62 val &= cmp->mask; in em_cmp_match() [all …]
|
/linux-3.4.99/arch/sparc/lib/ |
D | strncmp_32.S | 14 cmp %o2, 3 26 cmp %o0, 0 30 cmp %o0, %g2 42 cmp %o0, 0 46 cmp %o0, %g2 58 cmp %o0, 0 62 cmp %o0, %g2 74 cmp %o0, 0 78 cmp %o0, %g2 87 cmp %o4, 0 [all …]
|
D | memscan_32.S | 18 cmp %o1, 0 31 cmp %g3, 0 79 cmp %g2, 0 88 cmp %g2, 0 96 cmp %g2, 0 104 cmp %g2, 0 118 cmp %o2, 0 126 cmp %g2, %o1
|
/linux-3.4.99/arch/arm/mach-pnx4008/include/mach/ |
D | entry-macro.S | 36 cmp \irqstat,#1<<16 40 cmp \irqstat,#1<<8 43 cmp \irqstat,#1<<4 46 cmp \irqstat,#1<<2 49 cmp \irqstat,#1<<1 57 cmp \irqnr,#1 59 cmp \irqnr,#30 93 cmp \irqstat,#1<<16 96 cmp \irqstat,#1<<8 99 cmp \irqstat,#1<<4 [all …]
|
/linux-3.4.99/arch/hexagon/lib/ |
D | memset.S | 42 p0 = cmp.eq(r2, #0) 43 p1 = cmp.gtu(r2, #7) 72 p1 = cmp.eq(r2, #1) 85 p1 = cmp.eq(r2, #2) 97 p0 = cmp.gtu(r2, #7) 98 p1 = cmp.eq(r2, #4) 104 p0 = cmp.gtu(r2, #11) 111 p1 = cmp.eq(r3, #1) 127 p1 = cmp.eq(r2, #8) 138 p1 = cmp.eq(r2, #4) [all …]
|
D | memcpy.S | 199 p2 = cmp.eq(len, #0); /* =0 */ 201 p0 = cmp.gtu(len, #23); /* %1, <24 */ 202 p1 = cmp.eq(ptr_in, ptr_out); /* attempt to overwrite self */ 206 p3 = cmp.gtu(len, #95); /* %8 < 97 */ 269 p3 = cmp.gtu(back, #8); 275 p1 = cmp.eq(prolog, #0); 281 nokernel = cmp.eq(kernel,#0); 290 p2 = cmp.eq(kernel, #1); /* skip ovr if kernel == 0 */ 324 p2 = cmp.gtu(offset, #7); 336 p0 = cmp.gt(over, #0); [all …]
|
/linux-3.4.99/arch/arm/lib/ |
D | lib1funcs.S | 67 1: cmp \divisor, #0x10000000 75 1: cmp \divisor, #0x80000000 86 1: cmp \dividend, \divisor 89 cmp \dividend, \divisor, lsr #1 92 cmp \dividend, \divisor, lsr #2 95 cmp \dividend, \divisor, lsr #3 98 cmp \dividend, #0 @ Early termination? 115 cmp \divisor, #(1 << 16) 120 cmp \divisor, #(1 << 8) 124 cmp \divisor, #(1 << 4) [all …]
|
/linux-3.4.99/arch/mips/math-emu/ |
D | sp_cmp.c | 29 int ieee754sp_cmp(ieee754sp x, ieee754sp y, int cmp, int sig) in ieee754sp_cmp() argument 43 if (cmp & IEEE754_CUN) in ieee754sp_cmp() 45 if (cmp & (IEEE754_CLT | IEEE754_CGT)) { in ieee754sp_cmp() 60 return (cmp & IEEE754_CLT) != 0; in ieee754sp_cmp() 62 return (cmp & IEEE754_CEQ) != 0; in ieee754sp_cmp() 64 return (cmp & IEEE754_CGT) != 0; in ieee754sp_cmp()
|
D | dp_cmp.c | 29 int ieee754dp_cmp(ieee754dp x, ieee754dp y, int cmp, int sig) in ieee754dp_cmp() argument 43 if (cmp & IEEE754_CUN) in ieee754dp_cmp() 45 if (cmp & (IEEE754_CLT | IEEE754_CGT)) { in ieee754dp_cmp() 60 return (cmp & IEEE754_CLT) != 0; in ieee754dp_cmp() 62 return (cmp & IEEE754_CEQ) != 0; in ieee754dp_cmp() 64 return (cmp & IEEE754_CGT) != 0; in ieee754dp_cmp()
|
/linux-3.4.99/arch/powerpc/math-emu/ |
D | fcmpu.c | 16 long cmp; in fcmpu() local 30 FP_CMP_D(cmp, A, B, 2); in fcmpu() 31 cmp = code[(cmp + 1) & 3]; in fcmpu() 34 __FPU_FPSCR |= (cmp << 12); in fcmpu() 37 *ccr |= (cmp << ((7 - crfD) << 2)); in fcmpu()
|
D | fcmpo.c | 16 long cmp; in fcmpo() local 33 FP_CMP_D(cmp, A, B, 2); in fcmpo() 34 cmp = code[(cmp + 1) & 3]; in fcmpo() 37 __FPU_FPSCR |= (cmp << 12); in fcmpo() 40 *ccr |= (cmp << ((7 - crfD) << 2)); in fcmpo()
|
/linux-3.4.99/arch/ia64/lib/ |
D | memcpy_mck.S | 92 cmp.gt p15,p0=8,in2 // check for small size 93 cmp.ne p13,p0=0,r28 // check dest alignment 94 cmp.ne p14,p0=0,r29 // check src alignment 101 cmp.le p6,p0 = 1,r30 // for .align_dest 116 cmp.lt p6,p0=2*PREFETCH_DIST,cnt 117 cmp.lt p7,p8=1,cnt 142 cmp.eq p10,p0=r29,r0 // do we really need to loop? 144 cmp.le p6,p0=8,tmp 146 cmp.le p7,p0=16,tmp 148 cmp.eq p16,p17 = r0,r0 [all …]
|
D | memset.S | 66 cmp.ne p_nz, p_zr = value, r0 // use stf.spill if value is zero 67 cmp.eq p_scr, p0 = cnt, r0 79 cmp.ne p_unalgn, p0 = tmp, r0 // 82 cmp.gt p_scr, p0 = 16, cnt // is it a minimalistic task? 116 cmp.gt p_scr, p0 = tmp, cnt // is it a minimalistic task? 135 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value 184 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching? 192 cmp.le p_scr, p0 = 8, cnt // just a few bytes left ? 205 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value 238 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching? [all …]
|
D | do_csum.S | 133 cmp.lt p0,p6=r0,len // check for zero length or negative (32bit len) 157 cmp.eq p8,p9=last,first1 // everything fits in one word ? 186 cmp.ltu p6,p0=result1[0],word1[0] // check the carry 187 cmp.eq.or.andcm p8,p0=0,count // exit if zero 8-byte 195 cmp.eq p9,p10=1,count // if (count == 1) 200 cmp.ltu p6,p0=result1[0],word1[1] 226 (ELD_1) cmp.ltu pC1[0],p0=result1[LOAD_LATENCY],word1[LOAD_LATENCY+1] 228 (ELD_1) cmp.ltu pC2[0],p0=result2[LOAD_LATENCY],word2[LOAD_LATENCY+1] 244 cmp.ltu p6,p0=result1[LOAD_LATENCY+1],carry1 245 cmp.ltu p7,p0=result2[LOAD_LATENCY+1],carry2 [all …]
|
/linux-3.4.99/arch/arm/mach-omap2/include/mach/ |
D | debug-macro.S | 42 cmp \rp, #0 @ is port configured? 53 cmp \rp, #0 @ no port configured? 55 cmp \rp, #OMAP2UART1 @ OMAP2/3/4UART1 57 cmp \rp, #OMAP2UART2 @ OMAP2/3/4UART2 59 cmp \rp, #OMAP2UART3 @ only on 24xx 61 cmp \rp, #OMAP3UART3 @ only on 34xx 63 cmp \rp, #OMAP4UART3 @ only on 44xx 65 cmp \rp, #OMAP3UART4 @ only on 36xx 67 cmp \rp, #OMAP4UART4 @ only on 44xx 69 cmp \rp, #TI81XXUART1 @ ti81Xx UART offsets different [all …]
|
/linux-3.4.99/sound/oss/ |
D | vidc_fill.S | 21 1: cmp r0, r1 28 cmp r2, r3 34 1: cmp r0, r1 42 cmp r2, r3 48 1: cmp r0, r1 54 cmp r2, r3 60 1: cmp r0, r1 68 cmp r2, r3 75 1: cmp r0, r1 81 cmp r0, r1 [all …]
|
/linux-3.4.99/arch/alpha/include/asm/ |
D | xchg.h | 139 unsigned long prev, tmp, cmp, addr64; in ____cmpxchg() local 157 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) in ____cmpxchg() 166 unsigned long prev, tmp, cmp, addr64; in ____cmpxchg() local 184 : "=&r" (prev), "=&r" (new), "=&r" (tmp), "=&r" (cmp), "=&r" (addr64) in ____cmpxchg() 193 unsigned long prev, cmp; in ____cmpxchg() local 207 : "=&r"(prev), "=&r"(cmp), "=m"(*m) in ____cmpxchg() 216 unsigned long prev, cmp; in ____cmpxchg() local 230 : "=&r"(prev), "=&r"(cmp), "=m"(*m) in ____cmpxchg()
|
/linux-3.4.99/arch/hexagon/mm/ |
D | strnlen_user.S | 52 P0 = cmp.eq(mod8,#0); 63 P0 = cmp.eq(tmp1,#0); 65 P1 = cmp.gtu(end,start); 70 P0 = cmp.eq(mod8,#0); 88 P0 = cmp.gtu(end,start); 96 P0 = cmp.eq(tmp1,#32); 108 P0 = cmp.gt(tmp1,mod8);
|
/linux-3.4.99/lib/ |
D | list_sort.c | 15 int (*cmp)(void *priv, struct list_head *a, in merge() 23 if ((*cmp)(priv, a, b) <= 0) { in merge() 44 int (*cmp)(void *priv, struct list_head *a, in merge_and_restore_back_links() 53 if ((*cmp)(priv, a, b) <= 0) { in merge_and_restore_back_links() 73 (*cmp)(priv, tail->next, tail->next); in merge_and_restore_back_links() 98 int (*cmp)(void *priv, struct list_head *a, in list_sort() 121 cur = merge(priv, cmp, part[lev], cur); in list_sort() 138 list = merge(priv, cmp, part[lev], list); in list_sort() 140 merge_and_restore_back_links(priv, cmp, head, part[max_lev], list); in list_sort() 197 static int __init cmp(void *priv, struct list_head *a, struct list_head *b) in cmp() function [all …]
|
/linux-3.4.99/net/bridge/netfilter/ |
D | ebt_among.c | 28 uint32_t cmp[2] = { 0, 0 }; in ebt_mac_wormhash_contains() local 31 memcpy(((char *) cmp) + 2, mac, 6); in ebt_mac_wormhash_contains() 37 if (cmp[1] == p->cmp[1] && cmp[0] == p->cmp[0]) in ebt_mac_wormhash_contains() 44 if (cmp[1] == p->cmp[1] && cmp[0] == p->cmp[0]) in ebt_mac_wormhash_contains()
|
/linux-3.4.99/arch/ia64/hp/sim/boot/ |
D | boot_head.S | 63 cmp.gtu p6,p7=r9,r28 /* r28 <= 255? */ 68 cmp.gtu p6,p7=r9,r28 71 static: cmp.eq p6,p7=PAL_PTCE_INFO,r28 79 1: cmp.eq p6,p7=PAL_FREQ_RATIOS,r28 86 1: cmp.eq p6,p7=PAL_RSE_INFO,r28 93 1: cmp.eq p6,p7=PAL_CACHE_FLUSH,r28 /* PAL_CACHE_FLUSH */ 111 1: cmp.eq p6,p7=PAL_PERF_MON_INFO,r28 147 1: cmp.eq p6,p7=PAL_VM_SUMMARY,r28 156 1: cmp.eq p6,p7=PAL_MEM_ATTRIB,r28
|
/linux-3.4.99/arch/sparc/include/asm/ |
D | head_64.h | 41 cmp %tmp1, %tmp2; \ 50 cmp %tmp1, %tmp2; \ 57 cmp %tmp2, CHEETAH_MANUF; \ 61 cmp %tmp2, CHEETAH_PLUS_IMPL; \ 68 cmp %tmp2, CHEETAH_MANUF; \ 72 cmp %tmp2, CHEETAH_IMPL; \
|
/linux-3.4.99/arch/x86/net/ |
D | bpf_jit.S | 34 cmp $3,%eax 51 cmp $1,%eax 66 cmp %esi,%r9d /* if (offset >= hlen) goto bpf_slow_path_byte */ 85 cmp %esi,%r9d /* if (offset >= hlen) goto bpf_slow_path_byte_msh */ 153 cmp SKF_MAX_NEG_OFF, %esi /* test range */ 163 cmp SKF_MAX_NEG_OFF, %esi 174 cmp SKF_MAX_NEG_OFF, %esi 183 cmp SKF_MAX_NEG_OFF, %esi
|
/linux-3.4.99/crypto/async_tx/ |
D | raid6test.c | 42 struct completion *cmp = param; in callback() local 44 complete(cmp); in callback() 76 struct completion cmp; in raid6_dual_recov() local 121 init_completion(&cmp); in raid6_dual_recov() 122 init_async_submit(&submit, ASYNC_TX_ACK, tx, callback, &cmp, addr_conv); in raid6_dual_recov() 126 if (wait_for_completion_timeout(&cmp, msecs_to_jiffies(3000)) == 0) in raid6_dual_recov() 164 struct completion cmp; in test() local 179 init_completion(&cmp); in test() 180 init_async_submit(&submit, ASYNC_TX_ACK, NULL, callback, &cmp, addr_conv); in test() 184 if (wait_for_completion_timeout(&cmp, msecs_to_jiffies(3000)) == 0) { in test()
|
/linux-3.4.99/arch/arm/mach-at91/ |
D | pm_slowclock.S | 55 cmp tmp2, #0 69 cmp tmp2, #0 83 cmp tmp2, #0 97 cmp tmp2, #0 128 cmp memctrl, #AT91_MEMCTRL_MC 143 cmp memctrl, #AT91_MEMCTRL_DDRSDR 153 cmp ramc1, #0 284 cmp memctrl, #AT91_MEMCTRL_MC 290 cmp memctrl, #AT91_MEMCTRL_DDRSDR 297 cmp ramc1, #0
|