x86/asm/bpf: Annotate callable functions
authorJosh Poimboeuf <jpoimboe@redhat.com>
Thu, 21 Jan 2016 22:49:26 +0000 (16:49 -0600)
committerIngo Molnar <mingo@kernel.org>
Wed, 24 Feb 2016 07:35:43 +0000 (08:35 +0100)
bpf_jit.S has several functions which can be called from C code.  Give
them proper ELF annotations.

Signed-off-by: Josh Poimboeuf <jpoimboe@redhat.com>
Acked-by: Alexei Starovoitov <ast@kernel.org>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Andy Lutomirski <luto@amacapital.net>
Cc: Andy Lutomirski <luto@kernel.org>
Cc: Arnaldo Carvalho de Melo <acme@kernel.org>
Cc: Bernd Petrovitsch <bernd@petrovitsch.priv.at>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Brian Gerst <brgerst@gmail.com>
Cc: Chris J Arges <chris.j.arges@canonical.com>
Cc: Denys Vlasenko <dvlasenk@redhat.com>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Jiri Slaby <jslaby@suse.cz>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Michal Marek <mmarek@suse.cz>
Cc: Namhyung Kim <namhyung@gmail.com>
Cc: Pedro Alves <palves@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: live-patching@vger.kernel.org
Cc: netdev@vger.kernel.org
Link: http://lkml.kernel.org/r/bbe1de0c299fecd4fc9a1766bae8be2647bedb01.1453405861.git.jpoimboe@redhat.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
arch/x86/net/bpf_jit.S

index 4093216b3791311c995bfb2f644deed3845ef708..eb4a3bd293eb700c5cc5795c997da357247e4409 100644 (file)
        32 /* space for rbx,r13,r14,r15 */ + \
        8 /* space for skb_copy_bits */)
 
-sk_load_word:
-       .globl  sk_load_word
+#define FUNC(name) \
+       .globl name; \
+       .type name, @function; \
+       name:
 
+FUNC(sk_load_word)
        test    %esi,%esi
        js      bpf_slow_path_word_neg
 
-sk_load_word_positive_offset:
-       .globl  sk_load_word_positive_offset
-
+FUNC(sk_load_word_positive_offset)
        mov     %r9d,%eax               # hlen
        sub     %esi,%eax               # hlen - offset
        cmp     $3,%eax
@@ -39,15 +40,11 @@ sk_load_word_positive_offset:
        bswap   %eax                    /* ntohl() */
        ret
 
-sk_load_half:
-       .globl  sk_load_half
-
+FUNC(sk_load_half)
        test    %esi,%esi
        js      bpf_slow_path_half_neg
 
-sk_load_half_positive_offset:
-       .globl  sk_load_half_positive_offset
-
+FUNC(sk_load_half_positive_offset)
        mov     %r9d,%eax
        sub     %esi,%eax               #       hlen - offset
        cmp     $1,%eax
@@ -56,15 +53,11 @@ sk_load_half_positive_offset:
        rol     $8,%ax                  # ntohs()
        ret
 
-sk_load_byte:
-       .globl  sk_load_byte
-
+FUNC(sk_load_byte)
        test    %esi,%esi
        js      bpf_slow_path_byte_neg
 
-sk_load_byte_positive_offset:
-       .globl  sk_load_byte_positive_offset
-
+FUNC(sk_load_byte_positive_offset)
        cmp     %esi,%r9d   /* if (offset >= hlen) goto bpf_slow_path_byte */
        jle     bpf_slow_path_byte
        movzbl  (SKBDATA,%rsi),%eax
@@ -120,8 +113,8 @@ bpf_slow_path_byte:
 bpf_slow_path_word_neg:
        cmp     SKF_MAX_NEG_OFF, %esi   /* test range */
        jl      bpf_error       /* offset lower -> error  */
-sk_load_word_negative_offset:
-       .globl  sk_load_word_negative_offset
+
+FUNC(sk_load_word_negative_offset)
        sk_negative_common(4)
        mov     (%rax), %eax
        bswap   %eax
@@ -130,8 +123,8 @@ sk_load_word_negative_offset:
 bpf_slow_path_half_neg:
        cmp     SKF_MAX_NEG_OFF, %esi
        jl      bpf_error
-sk_load_half_negative_offset:
-       .globl  sk_load_half_negative_offset
+
+FUNC(sk_load_half_negative_offset)
        sk_negative_common(2)
        mov     (%rax),%ax
        rol     $8,%ax
@@ -141,8 +134,8 @@ sk_load_half_negative_offset:
 bpf_slow_path_byte_neg:
        cmp     SKF_MAX_NEG_OFF, %esi
        jl      bpf_error
-sk_load_byte_negative_offset:
-       .globl  sk_load_byte_negative_offset
+
+FUNC(sk_load_byte_negative_offset)
        sk_negative_common(1)
        movzbl  (%rax), %eax
        ret
This page took 0.026816 seconds and 5 git commands to generate.