Lines Matching refs:p_scr
41 #define p_scr p6 // default register for same-cycle branches macro
67 cmp.eq p_scr, p0 = cnt, r0
76 (p_scr) br.ret.dpnt.many rp // return immediately if count = 0
82 cmp.gt p_scr, p0 = 16, cnt // is it a minimalistic task?
83 (p_scr) br.cond.dptk.many .move_bytes_unaligned // go move just a few (M_B_U)
116 cmp.gt p_scr, p0 = tmp, cnt // is it a minimalistic task?
119 (p_scr) br.cond.dpnt.many .fraction_of_line // go move just a few
135 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value
138 (p_scr) add loopcnt = -1, linecnt //
184 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching?
188 (p_scr) stf8 [ptr9] = fvalue, 128
192 cmp.le p_scr, p0 = 8, cnt // just a few bytes left ?
193 (p_scr) br.cond.dpnt.many .fraction_of_line // Branch no. 2
205 cmp.gt p_scr, p0 = PREF_AHEAD, linecnt // check against actual value
208 (p_scr) add loopcnt = -1, linecnt
238 cmp.lt p_scr, p0 = ptr9, ptr1 // do we need more prefetching?
242 (p_scr) stf.spill [ptr9] = f0, 128
246 cmp.gt p_scr, p0 = 8, cnt // just a few bytes left ?
247 (p_scr) br.cond.dpnt.many .move_bytes_from_alignment //
256 cmp.eq p_scr, p0 = loopcnt, r0
258 (p_scr) br.cond.dpnt.many .store_words
276 cmp.gt p_scr, p0 = 8, cnt // just a few bytes left ?
277 (p_scr) br.cond.dpnt.many .move_bytes_from_alignment // Branch
297 cmp.eq p_scr, p0 = cnt, r0
299 (p_scr) br.cond.dpnt.few .restore_and_exit
344 tbit.nz p_scr, p0 = cnt, 1 // will there be a st2 at the end ?
358 (p_scr) st2 [ptr1] = value // fill 2 (aligned) bytes