// SPDX-License-Identifier: GPL-2.0 /* Converted from tools/testing/selftests/bpf/verifier/unpriv.c */ #include <linux/bpf.h> #include <bpf/bpf_helpers.h> #include "../../../include/linux/filter.h" #include "bpf_misc.h" #define BPF_SK_LOOKUP(func) \ /* struct bpf_sock_tuple tuple = {} */ \ "r2 = 0;" \ "*(u32*)(r10 - 8) = r2;" \ "*(u64*)(r10 - 16) = r2;" \ "*(u64*)(r10 - 24) = r2;" \ "*(u64*)(r10 - 32) = r2;" \ "*(u64*)(r10 - 40) = r2;" \ "*(u64*)(r10 - 48) = r2;" \ /* sk = func(ctx, &tuple, sizeof tuple, 0, 0) */ \ "r2 = r10;" \ "r2 += -48;" \ "r3 = %[sizeof_bpf_sock_tuple];"\ "r4 = 0;" \ "r5 = 0;" \ "call %[" #func "];" struct { __uint(type, BPF_MAP_TYPE_HASH); __uint(max_entries, 1); __type(key, long long); __type(value, long long); } map_hash_8b SEC(".maps"); void dummy_prog_42_socket(void); void dummy_prog_24_socket(void); void dummy_prog_loop1_socket(void); struct { __uint(type, BPF_MAP_TYPE_PROG_ARRAY); __uint(max_entries, 4); __uint(key_size, sizeof(int)); __array(values, void (void)); } map_prog1_socket SEC(".maps") = { .values = { [0] = (void *)&dummy_prog_42_socket, [1] = (void *)&dummy_prog_loop1_socket, [2] = (void *)&dummy_prog_24_socket, }, }; SEC("socket") __auxiliary __auxiliary_unpriv __naked void dummy_prog_42_socket(void) { asm volatile ("r0 = 42; exit;"); } SEC("socket") __auxiliary __auxiliary_unpriv __naked void dummy_prog_24_socket(void) { asm volatile ("r0 = 24; exit;"); } SEC("socket") __auxiliary __auxiliary_unpriv __naked void dummy_prog_loop1_socket(void) { asm volatile (" \ r3 = 1; \ r2 = %[map_prog1_socket] ll; \ call %[bpf_tail_call]; \ r0 = 41; \ exit; \ " : : __imm(bpf_tail_call), __imm_addr(map_prog1_socket) : __clobber_all); } SEC("socket") __description("unpriv: return pointer") __success __failure_unpriv __msg_unpriv("R0 leaks addr") __retval(POINTER_VALUE) __naked void unpriv_return_pointer(void) { asm volatile (" \ r0 = r10; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: add const to pointer") __success __success_unpriv __retval(0) __naked void unpriv_add_const_to_pointer(void) { asm volatile (" \ r1 += 8; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: add pointer to pointer") __failure __msg("R1 pointer += pointer") __failure_unpriv __naked void unpriv_add_pointer_to_pointer(void) { asm volatile (" \ r1 += r10; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: neg pointer") __success __failure_unpriv __msg_unpriv("R1 pointer arithmetic") __retval(0) __naked void unpriv_neg_pointer(void) { asm volatile (" \ r1 = -r1; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: cmp pointer with const") __success __failure_unpriv __msg_unpriv("R1 pointer comparison") __retval(0) __naked void unpriv_cmp_pointer_with_const(void) { asm volatile (" \ if r1 == 0 goto l0_%=; \ l0_%=: r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: cmp pointer with pointer") __success __failure_unpriv __msg_unpriv("R10 pointer comparison") __retval(0) __naked void unpriv_cmp_pointer_with_pointer(void) { asm volatile (" \ if r1 == r10 goto l0_%=; \ l0_%=: r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("tracepoint") __description("unpriv: check that printk is disallowed") __success __naked void check_that_printk_is_disallowed(void) { asm volatile (" \ r1 = 0; \ *(u64*)(r10 - 8) = r1; \ r1 = r10; \ r1 += -8; \ r2 = 8; \ r3 = r1; \ call %[bpf_trace_printk]; \ r0 = 0; \ exit; \ " : : __imm(bpf_trace_printk) : __clobber_all); } SEC("socket") __description("unpriv: pass pointer to helper function") __success __failure_unpriv __msg_unpriv("R4 leaks addr") __retval(0) __naked void pass_pointer_to_helper_function(void) { asm volatile (" \ r1 = 0; \ *(u64*)(r10 - 8) = r1; \ r2 = r10; \ r2 += -8; \ r1 = %[map_hash_8b] ll; \ r3 = r2; \ r4 = r2; \ call %[bpf_map_update_elem]; \ r0 = 0; \ exit; \ " : : __imm(bpf_map_update_elem), __imm_addr(map_hash_8b) : __clobber_all); } SEC("socket") __description("unpriv: indirectly pass pointer on stack to helper function") __success __failure_unpriv __msg_unpriv("invalid indirect read from stack R2 off -8+0 size 8") __retval(0) __naked void on_stack_to_helper_function(void) { asm volatile (" \ *(u64*)(r10 - 8) = r10; \ r2 = r10; \ r2 += -8; \ r1 = %[map_hash_8b] ll; \ call %[bpf_map_lookup_elem]; \ r0 = 0; \ exit; \ " : : __imm(bpf_map_lookup_elem), __imm_addr(map_hash_8b) : __clobber_all); } SEC("socket") __description("unpriv: mangle pointer on stack 1") __success __failure_unpriv __msg_unpriv("attempt to corrupt spilled") __retval(0) __naked void mangle_pointer_on_stack_1(void) { asm volatile (" \ *(u64*)(r10 - 8) = r10; \ r0 = 0; \ *(u32*)(r10 - 8) = r0; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: mangle pointer on stack 2") __success __failure_unpriv __msg_unpriv("attempt to corrupt spilled") __retval(0) __naked void mangle_pointer_on_stack_2(void) { asm volatile (" \ *(u64*)(r10 - 8) = r10; \ r0 = 0; \ *(u8*)(r10 - 1) = r0; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: read pointer from stack in small chunks") __failure __msg("invalid size") __failure_unpriv __naked void from_stack_in_small_chunks(void) { asm volatile (" \ *(u64*)(r10 - 8) = r10; \ r0 = *(u32*)(r10 - 8); \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: write pointer into ctx") __failure __msg("invalid bpf_context access") __failure_unpriv __msg_unpriv("R1 leaks addr") __naked void unpriv_write_pointer_into_ctx(void) { asm volatile (" \ *(u64*)(r1 + 0) = r1; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: spill/fill of ctx") __success __success_unpriv __retval(0) __naked void unpriv_spill_fill_of_ctx(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r1; \ r1 = *(u64*)(r6 + 0); \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("tc") __description("unpriv: spill/fill of ctx 2") __success __retval(0) __naked void spill_fill_of_ctx_2(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r1; \ r1 = *(u64*)(r6 + 0); \ call %[bpf_get_hash_recalc]; \ r0 = 0; \ exit; \ " : : __imm(bpf_get_hash_recalc) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of ctx 3") __failure __msg("R1 type=fp expected=ctx") __naked void spill_fill_of_ctx_3(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r1; \ *(u64*)(r6 + 0) = r10; \ r1 = *(u64*)(r6 + 0); \ call %[bpf_get_hash_recalc]; \ exit; \ " : : __imm(bpf_get_hash_recalc) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of ctx 4") __failure __msg("R1 type=scalar expected=ctx") __naked void spill_fill_of_ctx_4(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r1; \ r0 = 1; \ lock *(u64 *)(r10 - 8) += r0; \ r1 = *(u64*)(r6 + 0); \ call %[bpf_get_hash_recalc]; \ exit; \ " : : __imm(bpf_get_hash_recalc) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of different pointers stx") __failure __msg("same insn cannot be used with different pointers") __naked void fill_of_different_pointers_stx(void) { asm volatile (" \ r3 = 42; \ r6 = r10; \ r6 += -8; \ if r1 == 0 goto l0_%=; \ r2 = r10; \ r2 += -16; \ *(u64*)(r6 + 0) = r2; \ l0_%=: if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r1; \ l1_%=: r1 = *(u64*)(r6 + 0); \ *(u32*)(r1 + %[__sk_buff_mark]) = r3; \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)) : __clobber_all); } /* Same as above, but use BPF_ST_MEM to save 42 * instead of BPF_STX_MEM. */ SEC("tc") __description("unpriv: spill/fill of different pointers st") __failure __msg("same insn cannot be used with different pointers") __naked void fill_of_different_pointers_st(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ if r1 == 0 goto l0_%=; \ r2 = r10; \ r2 += -16; \ *(u64*)(r6 + 0) = r2; \ l0_%=: if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r1; \ l1_%=: r1 = *(u64*)(r6 + 0); \ .8byte %[st_mem]; \ r0 = 0; \ exit; \ " : : __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_insn(st_mem, BPF_ST_MEM(BPF_W, BPF_REG_1, offsetof(struct __sk_buff, mark), 42)) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of different pointers stx - ctx and sock") __failure __msg("type=ctx expected=sock") __naked void pointers_stx_ctx_and_sock(void) { asm volatile (" \ r8 = r1; \ /* struct bpf_sock *sock = bpf_sock_lookup(...); */\ " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) " r2 = r0; \ /* u64 foo; */ \ /* void *target = &foo; */ \ r6 = r10; \ r6 += -8; \ r1 = r8; \ /* if (skb == NULL) *target = sock; */ \ if r1 == 0 goto l0_%=; \ *(u64*)(r6 + 0) = r2; \ l0_%=: /* else *target = skb; */ \ if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r1; \ l1_%=: /* struct __sk_buff *skb = *target; */ \ r1 = *(u64*)(r6 + 0); \ /* skb->mark = 42; */ \ r3 = 42; \ *(u32*)(r1 + %[__sk_buff_mark]) = r3; \ /* if (sk) bpf_sk_release(sk) */ \ if r1 == 0 goto l2_%=; \ call %[bpf_sk_release]; \ l2_%=: r0 = 0; \ exit; \ " : : __imm(bpf_sk_lookup_tcp), __imm(bpf_sk_release), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of different pointers stx - leak sock") __failure //.errstr = "same insn cannot be used with different pointers", __msg("Unreleased reference") __naked void different_pointers_stx_leak_sock(void) { asm volatile (" \ r8 = r1; \ /* struct bpf_sock *sock = bpf_sock_lookup(...); */\ " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) " r2 = r0; \ /* u64 foo; */ \ /* void *target = &foo; */ \ r6 = r10; \ r6 += -8; \ r1 = r8; \ /* if (skb == NULL) *target = sock; */ \ if r1 == 0 goto l0_%=; \ *(u64*)(r6 + 0) = r2; \ l0_%=: /* else *target = skb; */ \ if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r1; \ l1_%=: /* struct __sk_buff *skb = *target; */ \ r1 = *(u64*)(r6 + 0); \ /* skb->mark = 42; */ \ r3 = 42; \ *(u32*)(r1 + %[__sk_buff_mark]) = r3; \ exit; \ " : : __imm(bpf_sk_lookup_tcp), __imm_const(__sk_buff_mark, offsetof(struct __sk_buff, mark)), __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of different pointers stx - sock and ctx (read)") __failure __msg("same insn cannot be used with different pointers") __naked void stx_sock_and_ctx_read(void) { asm volatile (" \ r8 = r1; \ /* struct bpf_sock *sock = bpf_sock_lookup(...); */\ " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) " r2 = r0; \ /* u64 foo; */ \ /* void *target = &foo; */ \ r6 = r10; \ r6 += -8; \ r1 = r8; \ /* if (skb) *target = skb */ \ if r1 == 0 goto l0_%=; \ *(u64*)(r6 + 0) = r1; \ l0_%=: /* else *target = sock */ \ if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r2; \ l1_%=: /* struct bpf_sock *sk = *target; */ \ r1 = *(u64*)(r6 + 0); \ /* if (sk) u32 foo = sk->mark; bpf_sk_release(sk); */\ if r1 == 0 goto l2_%=; \ r3 = *(u32*)(r1 + %[bpf_sock_mark]); \ call %[bpf_sk_release]; \ l2_%=: r0 = 0; \ exit; \ " : : __imm(bpf_sk_lookup_tcp), __imm(bpf_sk_release), __imm_const(bpf_sock_mark, offsetof(struct bpf_sock, mark)), __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) : __clobber_all); } SEC("tc") __description("unpriv: spill/fill of different pointers stx - sock and ctx (write)") __failure //.errstr = "same insn cannot be used with different pointers", __msg("cannot write into sock") __naked void stx_sock_and_ctx_write(void) { asm volatile (" \ r8 = r1; \ /* struct bpf_sock *sock = bpf_sock_lookup(...); */\ " BPF_SK_LOOKUP(bpf_sk_lookup_tcp) " r2 = r0; \ /* u64 foo; */ \ /* void *target = &foo; */ \ r6 = r10; \ r6 += -8; \ r1 = r8; \ /* if (skb) *target = skb */ \ if r1 == 0 goto l0_%=; \ *(u64*)(r6 + 0) = r1; \ l0_%=: /* else *target = sock */ \ if r1 != 0 goto l1_%=; \ *(u64*)(r6 + 0) = r2; \ l1_%=: /* struct bpf_sock *sk = *target; */ \ r1 = *(u64*)(r6 + 0); \ /* if (sk) sk->mark = 42; bpf_sk_release(sk); */\ if r1 == 0 goto l2_%=; \ r3 = 42; \ *(u32*)(r1 + %[bpf_sock_mark]) = r3; \ call %[bpf_sk_release]; \ l2_%=: r0 = 0; \ exit; \ " : : __imm(bpf_sk_lookup_tcp), __imm(bpf_sk_release), __imm_const(bpf_sock_mark, offsetof(struct bpf_sock, mark)), __imm_const(sizeof_bpf_sock_tuple, sizeof(struct bpf_sock_tuple)) : __clobber_all); } SEC("socket") __description("unpriv: write pointer into map elem value") __success __failure_unpriv __msg_unpriv("R0 leaks addr") __retval(0) __naked void pointer_into_map_elem_value(void) { asm volatile (" \ r1 = 0; \ *(u64*)(r10 - 8) = r1; \ r2 = r10; \ r2 += -8; \ r1 = %[map_hash_8b] ll; \ call %[bpf_map_lookup_elem]; \ if r0 == 0 goto l0_%=; \ *(u64*)(r0 + 0) = r0; \ l0_%=: exit; \ " : : __imm(bpf_map_lookup_elem), __imm_addr(map_hash_8b) : __clobber_all); } SEC("socket") __description("alu32: mov u32 const") __success __failure_unpriv __msg_unpriv("R7 invalid mem access 'scalar'") __retval(0) __naked void alu32_mov_u32_const(void) { asm volatile (" \ w7 = 0; \ w7 &= 1; \ w0 = w7; \ if r0 == 0 goto l0_%=; \ r0 = *(u64*)(r7 + 0); \ l0_%=: exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: partial copy of pointer") __success __failure_unpriv __msg_unpriv("R10 partial copy") __retval(0) __naked void unpriv_partial_copy_of_pointer(void) { asm volatile (" \ w1 = w10; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: pass pointer to tail_call") __success __failure_unpriv __msg_unpriv("R3 leaks addr into helper") __retval(0) __naked void pass_pointer_to_tail_call(void) { asm volatile (" \ r3 = r1; \ r2 = %[map_prog1_socket] ll; \ call %[bpf_tail_call]; \ r0 = 0; \ exit; \ " : : __imm(bpf_tail_call), __imm_addr(map_prog1_socket) : __clobber_all); } SEC("socket") __description("unpriv: cmp map pointer with zero") __success __failure_unpriv __msg_unpriv("R1 pointer comparison") __retval(0) __naked void cmp_map_pointer_with_zero(void) { asm volatile (" \ r1 = 0; \ r1 = %[map_hash_8b] ll; \ if r1 == 0 goto l0_%=; \ l0_%=: r0 = 0; \ exit; \ " : : __imm_addr(map_hash_8b) : __clobber_all); } SEC("socket") __description("unpriv: write into frame pointer") __failure __msg("frame pointer is read only") __failure_unpriv __naked void unpriv_write_into_frame_pointer(void) { asm volatile (" \ r10 = r1; \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: spill/fill frame pointer") __failure __msg("frame pointer is read only") __failure_unpriv __naked void unpriv_spill_fill_frame_pointer(void) { asm volatile (" \ r6 = r10; \ r6 += -8; \ *(u64*)(r6 + 0) = r10; \ r10 = *(u64*)(r6 + 0); \ r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: cmp of frame pointer") __success __failure_unpriv __msg_unpriv("R10 pointer comparison") __retval(0) __naked void unpriv_cmp_of_frame_pointer(void) { asm volatile (" \ if r10 == 0 goto l0_%=; \ l0_%=: r0 = 0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: adding of fp, reg") __success __failure_unpriv __msg_unpriv("R1 stack pointer arithmetic goes out of range") __retval(0) __naked void unpriv_adding_of_fp_reg(void) { asm volatile (" \ r0 = 0; \ r1 = 0; \ r1 += r10; \ *(u64*)(r1 - 8) = r0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: adding of fp, imm") __success __failure_unpriv __msg_unpriv("R1 stack pointer arithmetic goes out of range") __retval(0) __naked void unpriv_adding_of_fp_imm(void) { asm volatile (" \ r0 = 0; \ r1 = r10; \ r1 += 0; \ *(u64*)(r1 - 8) = r0; \ exit; \ " ::: __clobber_all); } SEC("socket") __description("unpriv: cmp of stack pointer") __success __failure_unpriv __msg_unpriv("R2 pointer comparison") __retval(0) __naked void unpriv_cmp_of_stack_pointer(void) { asm volatile (" \ r2 = r10; \ r2 += -8; \ if r2 == 0 goto l0_%=; \ l0_%=: r0 = 0; \ exit; \ " ::: __clobber_all); } char _license[] SEC("license") = "GPL";