/linux-2.6.39/arch/ia64/lib/ |
D | clear_user.S | 58 cmp.eq p6,p0=r0,len // check for zero length 65 (p6) br.ret.spnt.many rp 67 cmp.lt p6,p0=16,len // if len > 16 then long memset 69 (p6) br.cond.dptk .long_do_clear 102 tbit.nz p6,p0=buf,0 // odd alignment (for long_do_clear) 104 EX( .Lexit3, (p6) st1 [buf]=r0,1 ) // 1-byte aligned 105 (p6) adds len=-1,len;; // sync because buf is modified 106 tbit.nz p6,p0=buf,1 108 EX( .Lexit3, (p6) st2 [buf]=r0,2 ) // 2-byte aligned 109 (p6) adds len=-2,len;; [all …]
|
D | idiv64.S | 49 frcpa.s1 f11, p6 = f8, f9 // y0 = frcpa(b) 51 (p6) fmpy.s1 f7 = f8, f11 // q0 = a*y0 52 (p6) fnma.s1 f6 = f9, f11, f1 // e0 = -b*y0 + 1 54 (p6) fma.s1 f10 = f7, f6, f7 // q1 = q0*e0 + q0 55 (p6) fmpy.s1 f7 = f6, f6 // e1 = e0*e0 60 (p6) fma.s1 f10 = f10, f7, f10 // q2 = q1*e1 + q1 61 (p6) fma.s1 f6 = f11, f6, f11 // y1 = y0*e0 + y0 63 (p6) fma.s1 f6 = f6, f7, f6 // y2 = y1*e1 + y1 64 (p6) fnma.s1 f7 = f9, f10, f8 // r = -b*q2 + a 70 (p6) fma.s1 f11 = f7, f6, f10 // q3 = r*y2 + q2
|
D | strlen.S | 102 cmp.eq p6,p0=r0,r0 // sets p6 to true for cmp.and 117 cmp.eq.and p6,p0=8,val1 // p6 = p6 and val1==8 118 cmp.eq.and p6,p0=8,val2 // p6 = p6 and mask==8 119 (p6) br.wtop.dptk 1b // loop until p6 == 0 129 tnat.nz p6,p7=val1 // test NaT on val1 130 (p6) br.cond.spnt .recover // jump to recovery if val1 is NaT 172 cmp.eq p0,p6=r0,r0 // nullify first ld8 in loop 178 (p6) ld8 val=[base],8 // will fail if unrecoverable fault 182 cmp.eq p6,p0=8,val1 // val1==8 ? 183 (p6) br.wtop.dptk 2b // loop until p6 == 0
|
D | strlen_user.S | 104 cmp.eq p6,p0=r0,r0 // sets p6 (required because of // cmp.and) 120 cmp.eq.and p6,p0=8,val1 // p6 = p6 and val1==8 121 cmp.eq.and p6,p0=8,val2 // p6 = p6 and mask==8 122 (p6) br.wtop.dptk.few 1b // loop until p6 == 0 132 tnat.nz p6,p7=val1 // test NaT on val1 133 (p6) br.cond.spnt .recover // jump to recovery if val1 is NaT 169 cmp.eq p0,p6=r0,r0 // nullify first ld8 in loop 175 EX(.Lexit1, (p6) ld8 val=[base],8) 179 cmp.eq p6,p0=8,val1 // val1==8 ? 180 (p6) br.wtop.dptk.few 2b // loop until p6 == 0
|
D | carta_random.S | 40 cmp.gtu p6, p0 = lo, m 42 (p6) and lo = lo, m 44 (p6) add lo = 1, lo 48 cmp.gtu p6, p0 = lo, m 50 (p6) and lo = lo, m 52 (p6) add lo = 1, lo
|
D | strncpy_from_user.S | 27 cmp.eq p6,p0=r0,in2 28 (p6) br.ret.spnt.many rp 35 cmp.ne p6,p7=r8,r0 37 (p6) cmp.ne.unc p8,p0=in1,r10 40 (p6) mov r8=in2 // buffer filled up---return buffer length
|
D | memcpy_mck.S | 101 cmp.le p6,p0 = 1,r30 // for .align_dest 116 cmp.lt p6,p0=2*PREFETCH_DIST,cnt 127 (p6) br.cond.dpnt .long_copy 144 cmp.le p6,p0=8,tmp 167 EX(.ex_handler, (p6) ld8 t1=[src0]) 175 EX(.ex_handler, (p6) st8 [dst0]=t1) // store byte 1 289 cmp.lt p6,p7=blocksize,in2 292 (p6) mov in2=blocksize 318 cmp.le p6,p7=8,curlen; // have at least 8 byte remaining? 352 EX(.ex_handler, (p6) ld8 r37=[src1],8) [all …]
|
D | idiv32.S | 59 frcpa.s1 f6, p6 = f8, f9 // y0 = frcpa(b) 61 (p6) fmpy.s1 f8 = f8, f6 // q0 = a*y0 62 (p6) fnma.s1 f6 = f9, f6, f1 // e0 = -b*y0 + 1 67 (p6) fma.s1 f8 = f6, f8, f8 // q1 = e0*q0 + q0 68 (p6) fma.s1 f6 = f6, f6, f7 // e1 = e0*e0 + 2^-34 73 (p6) fma.s1 f6 = f6, f8, f8 // q2 = e1*q1 + q1
|
D | do_csum.S | 133 cmp.lt p0,p6=r0,len // check for zero length or negative (32bit len) 138 (p6) br.ret.spnt.many rp // return if zero or negative length 186 cmp.ltu p6,p0=result1[0],word1[0] // check the carry 189 (p6) adds result1[0]=1,result1[0] 200 cmp.ltu p6,p0=result1[0],word1[1] 202 (p6) adds result1[0]=1,result1[0] 244 cmp.ltu p6,p0=result1[LOAD_LATENCY+1],carry1 247 (p6) adds result1[LOAD_LATENCY+1]=1,result1[LOAD_LATENCY+1] 252 cmp.ltu p6,p0=result1[0],result2[LOAD_LATENCY+1] 254 (p6) adds result1[0]=1,result1[0]
|
D | copy_user.S | 261 SWITCH(p6, 8) 269 CASE(p6, 8) 319 tbit.nz p6,p7=src1,0 // odd alignment 332 EX(.failure_in1,(p6) ld1 val1[0]=[src1],1) // 1-byte aligned 333 (p6) adds len1=-1,len1;; 346 EX(.failure_out,(p6) st1 [dst1]=val1[0],1) 362 tbit.nz p6,p0=len1,3 389 EX(.failure_in1,(p6) ld8 val1[0]=[src1],8) // at least 8 bytes 398 EX(.failure_out, (p6) st8 [dst1]=val1[0],8) 580 cmp.ne p6,p0=dst1,enddst // Do we need to finish the tail ? [all …]
|
D | memcpy.S | 61 cmp.eq p6,p0=in2,r0 // zero length? 63 (p6) br.ret.spnt.many rp // zero length, return immediately 79 cmp.ne p6,p0=t0,r0 83 (p6) br.cond.spnt.few .memcpy_long 237 cmp.ne p6,p0=t0,r0 // is src aligned, too? 252 (p6) ld8 val[1]=[src2],8 // prime the pump...
|
D | strnlen_user.S | 35 cmp.eq p6,p0=r8,r0 36 (p6) br.cond.dpnt .Lexit
|
D | copy_page.S | 73 cmp.ltu p6,p0 = tgtf, tgt_last 90 (p6) lfetch [srcf], 64 91 (p6) lfetch [tgtf], 64
|
/linux-2.6.39/arch/ia64/kvm/ |
D | optvfault.S | 228 cmp.eq p6,p0=6,r23 230 (p6) br.cond.dpnt.many kvm_virtualization_fault_back 254 cmp.lt p6,p0=14,r18 256 (p6) mov r18=14 258 (p6) dep r19=r18,r19,2,6 260 cmp.eq p6,p0=0,r23 262 cmp.eq.or p6,p0=4,r23 265 (p6) adds r17=VMM_VCPU_META_SAVED_RR0_OFFSET,r21 269 (p6) shladd r17=r23,1,r17 271 (p6) st8 [r17]=r19 [all …]
|
D | vmm_ivt.S | 84 tbit.z p6,p7=r29,IA64_PSR_VM_BIT; \ 125 tbit.z p6,p7=r29,IA64_PSR_VM_BIT; 126 (p6) br.sptk kvm_alt_itlb_miss 139 tbit.z p6,p7=r29,IA64_PSR_VM_BIT; 140 (p6) br.sptk kvm_alt_dtlb_miss 271 tbit.z p6,p7=r29,IA64_PSR_VM_BIT 428 cmp.eq p6,p0 = 1,r30 430 (p6) add r29 = VMM_VCPU_SAVED_GP_OFFSET,r21 432 (p6) ld8 r1 = [r29] 612 cmp.eq p6,p0=EVENT_MOV_FROM_AR,r24 [all …]
|
/linux-2.6.39/arch/ia64/kernel/ |
D | ivt.S | 135 cmp.eq p6,p7=5,r17 // is IFA pointing into to region 5? 141 LOAD_PHYSICAL(p6, r19, swapper_pg_dir) // region 5 is rooted at swapper_pg_dir 143 .pred.rel "mutex", p6, p7 144 (p6) shr.u r21=r21,PGDIR_SHIFT+PAGE_SHIFT 147 (p6) dep r17=r18,r19,3,(PAGE_SHIFT-3) // r17=pgd_offset for region 5 149 cmp.eq p7,p6=0,r21 // unused address bits all zeroes? 158 (p7) cmp.eq p6,p7=r17,r0 // was pgd_present(*pgd) == NULL? 165 (p7) cmp.eq.or.andcm p6,p7=r29,r0 // was pud_present(*pud) == NULL? 174 (p7) cmp.eq.or.andcm p6,p7=r20,r0 // was pmd_present(*pmd) == NULL? 180 (p7) tbit.z p6,p7=r18,_PAGE_P_BIT // page present bit cleared? [all …]
|
D | fsys.S | 127 cmp.ne p6,p0=r18,r19 // did real_parent change? 129 (p6) br.cond.spnt.few 1b // yes -> redo the read of tgid and the check 149 tnat.z p6,p7=r32 // check argument register for being NaT 168 (p6) st8 [r18]=r32 193 tnat.nz p6,p0 = r33 // guard against NaT argument 194 (p6) br.cond.spnt.few .fail_einval 242 tnat.nz p6,p0 = r31 // guard against Nat argument 243 (p6) br.cond.spnt.few .fail_einval 257 cmp.ne p6, p0 = 0, r2 // Fallback if work is scheduled 258 (p6) br.cond.spnt.many fsys_fallback_syscall [all …]
|
D | relocate_kernel.S | 81 cmp.ltu p6,p7=r24,r19 138 tbit.z p0, p6=r30, 0;; // 0x1 dest page 139 (p6) and r17=r30, r16 140 (p6) br.cond.sptk.few .loop;; 142 tbit.z p0, p6=r30, 1;; // 0x2 indirect page 143 (p6) and in0=r30, r16 144 (p6) br.cond.sptk.few .loop;; 146 tbit.z p0, p6=r30, 2;; // 0x4 end flag 147 (p6) br.cond.sptk.few .end_loop;; 149 tbit.z p6, p0=r30, 3;; // 0x8 source page [all …]
|
D | entry.S | 74 cmp4.ge p6,p7=r8,r0 79 (p6) cmp.ne pKStk,pUStk=r0,r0 // a successful execve() lands us in user-mode... 81 (p6) mov ar.pfs=r0 // clear ar.pfs on success 122 tbit.nz p6,p0=in0,CLONE_SETTLS_BIT 125 (p6) st8 [r2]=in5 // store TLS in r16 for copy_thread() 154 tbit.nz p6,p0=in0,CLONE_SETTLS_BIT 157 (p6) st8 [r2]=in4 // store TLS in r13 (tp) 190 cmp.eq p7,p6=r25,in0 195 (p6) cmp.eq p7,p6=r26,r27 196 (p6) br.cond.dpnt .map [all …]
|
D | gate.S | 337 cmp.geu p6,p7=r19,r17 // A (sysnr > 0 && sysnr < 1024+NR_syscalls)? 344 (p6) ld8 r18=[r18] // M0|1 348 (p6) tbit.z.unc p8,p0=r18,0 // I0 (dual-issues with "mov b7=r18"!) 352 (p6) mov b7=r18 // I0 363 (p6) add r14=-8,r14 // r14 <- addr of fsys_bubble_down entry 365 (p6) ld8 r14=[r14] // r14 <- fsys_bubble_down 367 (p6) mov b7=r14 368 (p6) br.sptk.many b7 370 BRL_COND_FSYS_BUBBLE_DOWN(p6)
|
/linux-2.6.39/arch/ia64/hp/sim/boot/ |
D | boot_head.S | 63 cmp.gtu p6,p7=r9,r28 /* r28 <= 255? */ 64 (p6) br.cond.sptk.few static 68 cmp.gtu p6,p7=r9,r28 69 (p6) br.cond.sptk.few stacked 71 static: cmp.eq p6,p7=PAL_PTCE_INFO,r28 79 1: cmp.eq p6,p7=PAL_FREQ_RATIOS,r28 86 1: cmp.eq p6,p7=PAL_RSE_INFO,r28 93 1: cmp.eq p6,p7=PAL_CACHE_FLUSH,r28 /* PAL_CACHE_FLUSH */ 111 1: cmp.eq p6,p7=PAL_PERF_MON_INFO,r28 147 1: cmp.eq p6,p7=PAL_VM_SUMMARY,r28 [all …]
|
/linux-2.6.39/arch/ia64/sn/kernel/sn2/ |
D | ptc_deadlock.S | 53 cmp.ne p6,p0=zeroval,scr2 54 (p6) br.cond.sptk 5b 68 cmp.ne p6,p0=zeroval,scr2 69 (p6) br.cond.sptk 5b;; 79 cmp.ne p6,p0=zeroval,scr2 80 (p6) br.cond.sptk 5b
|
/linux-2.6.39/arch/powerpc/include/asm/ |
D | syscalls.h | 25 int __user *child_tidp, int p6, struct pt_regs *regs); 28 unsigned long p6, struct pt_regs *regs); 31 unsigned long p6, struct pt_regs *regs);
|
/linux-2.6.39/arch/ia64/xen/ |
D | xenivt.S | 48 cmp.ne p6,p0=r20,r0 // if there are pending events, 49 (p6) br.spnt.few 1b // call evtchn_do_upcall again.
|
/linux-2.6.39/arch/arm/mach-iop13xx/include/mach/ |
D | entry-macro.S | 33 mrc p6, 0, \irqnr, c3, c2, 0 @ Read IINTVEC 35 mrceq p6, 0, \irqnr, c3, c2, 0 @ Re-read on potentially bad zero
|