Lines Matching refs:insn

94 	struct sock_filter *insn;  in bpf_fill_maxinsns1()  local
98 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns1()
99 if (!insn) in bpf_fill_maxinsns1()
103 insn[i] = __BPF_STMT(BPF_RET | BPF_K, k); in bpf_fill_maxinsns1()
105 self->u.ptr.insns = insn; in bpf_fill_maxinsns1()
114 struct sock_filter *insn; in bpf_fill_maxinsns2() local
117 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns2()
118 if (!insn) in bpf_fill_maxinsns2()
122 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns2()
124 self->u.ptr.insns = insn; in bpf_fill_maxinsns2()
133 struct sock_filter *insn; in bpf_fill_maxinsns3() local
137 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns3()
138 if (!insn) in bpf_fill_maxinsns3()
146 insn[i] = __BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, k); in bpf_fill_maxinsns3()
149 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns3()
151 self->u.ptr.insns = insn; in bpf_fill_maxinsns3()
160 struct sock_filter *insn; in bpf_fill_maxinsns4() local
163 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns4()
164 if (!insn) in bpf_fill_maxinsns4()
168 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns4()
170 self->u.ptr.insns = insn; in bpf_fill_maxinsns4()
179 struct sock_filter *insn; in bpf_fill_maxinsns5() local
182 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns5()
183 if (!insn) in bpf_fill_maxinsns5()
186 insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0); in bpf_fill_maxinsns5()
189 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns5()
191 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab); in bpf_fill_maxinsns5()
193 self->u.ptr.insns = insn; in bpf_fill_maxinsns5()
202 struct sock_filter *insn; in bpf_fill_maxinsns6() local
205 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns6()
206 if (!insn) in bpf_fill_maxinsns6()
210 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns6()
213 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns6()
215 self->u.ptr.insns = insn; in bpf_fill_maxinsns6()
224 struct sock_filter *insn; in bpf_fill_maxinsns7() local
227 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns7()
228 if (!insn) in bpf_fill_maxinsns7()
232 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns7()
235 insn[len - 4] = __BPF_STMT(BPF_MISC | BPF_TAX, 0); in bpf_fill_maxinsns7()
236 insn[len - 3] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns7()
238 insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0); in bpf_fill_maxinsns7()
239 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns7()
241 self->u.ptr.insns = insn; in bpf_fill_maxinsns7()
250 struct sock_filter *insn; in bpf_fill_maxinsns8() local
253 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns8()
254 if (!insn) in bpf_fill_maxinsns8()
257 insn[0] = __BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff); in bpf_fill_maxinsns8()
260 insn[i] = __BPF_JUMP(BPF_JMP | BPF_JGT, 0xffffffff, jmp_off--, 0); in bpf_fill_maxinsns8()
262 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns8()
264 self->u.ptr.insns = insn; in bpf_fill_maxinsns8()
273 struct bpf_insn *insn; in bpf_fill_maxinsns9() local
276 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns9()
277 if (!insn) in bpf_fill_maxinsns9()
280 insn[0] = BPF_JMP_IMM(BPF_JA, 0, 0, len - 2); in bpf_fill_maxinsns9()
281 insn[1] = BPF_ALU32_IMM(BPF_MOV, R0, 0xcbababab); in bpf_fill_maxinsns9()
282 insn[2] = BPF_EXIT_INSN(); in bpf_fill_maxinsns9()
285 insn[i] = BPF_ALU32_IMM(BPF_MOV, R0, 0xfefefefe); in bpf_fill_maxinsns9()
287 insn[len - 2] = BPF_EXIT_INSN(); in bpf_fill_maxinsns9()
288 insn[len - 1] = BPF_JMP_IMM(BPF_JA, 0, 0, -(len - 1)); in bpf_fill_maxinsns9()
290 self->u.ptr.insns = insn; in bpf_fill_maxinsns9()
299 struct bpf_insn *insn; in bpf_fill_maxinsns10() local
302 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns10()
303 if (!insn) in bpf_fill_maxinsns10()
307 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 2 - 2 * i); in bpf_fill_maxinsns10()
309 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 1 - 2 * i); in bpf_fill_maxinsns10()
311 insn[hlen / 2] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen / 2 - 1); in bpf_fill_maxinsns10()
312 insn[hlen] = BPF_ALU32_IMM(BPF_MOV, R0, 0xabababac); in bpf_fill_maxinsns10()
313 insn[hlen + 1] = BPF_EXIT_INSN(); in bpf_fill_maxinsns10()
315 self->u.ptr.insns = insn; in bpf_fill_maxinsns10()
324 struct sock_filter *insn; in __bpf_fill_ja() local
328 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in __bpf_fill_ja()
329 if (!insn) in __bpf_fill_ja()
336 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, in __bpf_fill_ja()
339 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, rlen - 1 - j, in __bpf_fill_ja()
342 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xababcbac); in __bpf_fill_ja()
344 self->u.ptr.insns = insn; in __bpf_fill_ja()
365 struct sock_filter *insn; in bpf_fill_ld_abs_get_processor_id() local
368 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_ld_abs_get_processor_id()
369 if (!insn) in bpf_fill_ld_abs_get_processor_id()
373 insn[i] = __BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 0); in bpf_fill_ld_abs_get_processor_id()
374 insn[i + 1] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, in bpf_fill_ld_abs_get_processor_id()
378 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xbee); in bpf_fill_ld_abs_get_processor_id()
380 self->u.ptr.insns = insn; in bpf_fill_ld_abs_get_processor_id()
391 struct bpf_insn *insn; in bpf_fill_ld_abs_vlan_push_pop() local
394 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_ld_abs_vlan_push_pop()
395 if (!insn) in bpf_fill_ld_abs_vlan_push_pop()
398 insn[i++] = BPF_MOV64_REG(R6, R1); in bpf_fill_ld_abs_vlan_push_pop()
401 insn[i++] = BPF_LD_ABS(BPF_B, 0); in bpf_fill_ld_abs_vlan_push_pop()
402 insn[i] = BPF_JMP_IMM(BPF_JNE, R0, 0x34, len - i - 2); in bpf_fill_ld_abs_vlan_push_pop()
404 insn[i++] = BPF_MOV64_REG(R1, R6); in bpf_fill_ld_abs_vlan_push_pop()
405 insn[i++] = BPF_MOV64_IMM(R2, 1); in bpf_fill_ld_abs_vlan_push_pop()
406 insn[i++] = BPF_MOV64_IMM(R3, 2); in bpf_fill_ld_abs_vlan_push_pop()
407 insn[i++] = BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, 0, 0, in bpf_fill_ld_abs_vlan_push_pop()
409 insn[i] = BPF_JMP_IMM(BPF_JNE, R0, 0, len - i - 2); in bpf_fill_ld_abs_vlan_push_pop()
414 insn[i++] = BPF_LD_ABS(BPF_B, 0); in bpf_fill_ld_abs_vlan_push_pop()
415 insn[i] = BPF_JMP_IMM(BPF_JNE, R0, 0x34, len - i - 2); in bpf_fill_ld_abs_vlan_push_pop()
417 insn[i++] = BPF_MOV64_REG(R1, R6); in bpf_fill_ld_abs_vlan_push_pop()
418 insn[i++] = BPF_RAW_INSN(BPF_JMP | BPF_CALL, 0, 0, 0, in bpf_fill_ld_abs_vlan_push_pop()
420 insn[i] = BPF_JMP_IMM(BPF_JNE, R0, 0, len - i - 2); in bpf_fill_ld_abs_vlan_push_pop()
427 insn[i] = BPF_ALU32_IMM(BPF_MOV, R0, 0xbef); in bpf_fill_ld_abs_vlan_push_pop()
429 insn[len - 1] = BPF_EXIT_INSN(); in bpf_fill_ld_abs_vlan_push_pop()
431 self->u.ptr.insns = insn; in bpf_fill_ld_abs_vlan_push_pop()