Lines Matching refs:p_scr
54 #define p_scr p6 // default register for same-cycle branches macro
92 cmp.eq p_scr, p0 = cnt, r0
101 (p_scr) br.ret.dpnt.many rp // return immediately if count = 0
107 cmp.gt p_scr, p0 = 16, cnt // is it a minimalistic task?
108 (p_scr) br.cond.dptk.many .move_bytes_unaligned // go move just a few (M_B_U)
141 cmp.gt p_scr, p0 = tmp, cnt // is it a minimalistic task?
144 (p_scr) br.cond.dpnt.many .fraction_of_line // go move just a few
162 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value
165 (p_scr) add loopcnt = -1, linecnt // start of stores
211 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching?
215 (p_scr) store [ptr9] = myval, 128
219 cmp.le p_scr, p0 = 8, cnt // just a few bytes left ?
220 (p_scr) br.cond.dpnt.many .fraction_of_line // Branch no. 2
236 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value
239 (p_scr) add loopcnt = -1, linecnt
269 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching?
273 (p_scr) stf.spill [ptr9] = f0, 128
277 cmp.gt p_scr, p0 = 8, cnt // just a few bytes left ?
278 (p_scr) br.cond.dpnt.many .move_bytes_from_alignment
287 cmp.eq p_scr, p0 = loopcnt, r0
289 (p_scr) br.cond.dpnt.many store_words
309 cmp.gt p_scr, p0 = 8, cnt // just a few bytes left ?
310 (p_scr) br.cond.dpnt.many .move_bytes_from_alignment // Branch
330 cmp.eq p_scr, p0 = cnt, r0
332 (p_scr) br.cond.dpnt.few .restore_and_exit
377 tbit.nz p_scr, p0 = cnt, 1 // will there be a st2 at the end ?
392 (p_scr) st2 [ptr1] = value // fill 2 (aligned) bytes