Linux 5.8-rc6
-----BEGIN PGP SIGNATURE----- iQFSBAABCAA8FiEEq68RxlopcLEwq+PEeb4+QwBBGIYFAl8UzA4eHHRvcnZhbGRz QGxpbnV4LWZvdW5kYXRpb24ub3JnAAoJEHm+PkMAQRiGQ7cH/3v+Gv+SmHJCvaT2 CSu0+7okVnYbY3UTb3hykk7/aOqb6284KjxR03r0CWFzsEsZVhC5pvvruASSiMQg Pi04sLqv6CsGLHd1n+pl4AUYEaxq6k4KS3uU3HHSWxrahDDApQoRUx2F8lpOxyj8 RiwnoO60IMPA7IFJqzcZuFqsgdxqiiYvnzT461KX8Mrw6fyMXeR2KAj2NwMX8dZN At21Sf8+LSoh6q2HnugfiUd/jR10XbfxIIx2lXgIinb15GXgWydEQVrDJ7cUV7ix Jd0S+dtOtp+lWtFHDoyjjqqsMV7+G8i/rFNZoxSkyZqsUTaKzaR6JD3moSyoYZgG 0+eXO4A= =9EpR -----END PGP SIGNATURE----- Merge 5.8-rc6 into tty-next We need the serial/tty fixes in here as well. Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
commit
6f2c6599ba
1395 changed files with 12317 additions and 6552 deletions
|
|
@ -77,6 +77,9 @@
|
|||
|
||||
#endif /* cmpxchg64_relaxed */
|
||||
|
||||
#define arch_atomic_read atomic_read
|
||||
#define arch_atomic_read_acquire atomic_read_acquire
|
||||
|
||||
#ifndef atomic_read_acquire
|
||||
static __always_inline int
|
||||
atomic_read_acquire(const atomic_t *v)
|
||||
|
|
@ -86,6 +89,9 @@ atomic_read_acquire(const atomic_t *v)
|
|||
#define atomic_read_acquire atomic_read_acquire
|
||||
#endif
|
||||
|
||||
#define arch_atomic_set atomic_set
|
||||
#define arch_atomic_set_release atomic_set_release
|
||||
|
||||
#ifndef atomic_set_release
|
||||
static __always_inline void
|
||||
atomic_set_release(atomic_t *v, int i)
|
||||
|
|
@ -95,6 +101,13 @@ atomic_set_release(atomic_t *v, int i)
|
|||
#define atomic_set_release atomic_set_release
|
||||
#endif
|
||||
|
||||
#define arch_atomic_add atomic_add
|
||||
|
||||
#define arch_atomic_add_return atomic_add_return
|
||||
#define arch_atomic_add_return_acquire atomic_add_return_acquire
|
||||
#define arch_atomic_add_return_release atomic_add_return_release
|
||||
#define arch_atomic_add_return_relaxed atomic_add_return_relaxed
|
||||
|
||||
#ifndef atomic_add_return_relaxed
|
||||
#define atomic_add_return_acquire atomic_add_return
|
||||
#define atomic_add_return_release atomic_add_return
|
||||
|
|
@ -137,6 +150,11 @@ atomic_add_return(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_add_return_relaxed */
|
||||
|
||||
#define arch_atomic_fetch_add atomic_fetch_add
|
||||
#define arch_atomic_fetch_add_acquire atomic_fetch_add_acquire
|
||||
#define arch_atomic_fetch_add_release atomic_fetch_add_release
|
||||
#define arch_atomic_fetch_add_relaxed atomic_fetch_add_relaxed
|
||||
|
||||
#ifndef atomic_fetch_add_relaxed
|
||||
#define atomic_fetch_add_acquire atomic_fetch_add
|
||||
#define atomic_fetch_add_release atomic_fetch_add
|
||||
|
|
@ -179,6 +197,13 @@ atomic_fetch_add(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_add_relaxed */
|
||||
|
||||
#define arch_atomic_sub atomic_sub
|
||||
|
||||
#define arch_atomic_sub_return atomic_sub_return
|
||||
#define arch_atomic_sub_return_acquire atomic_sub_return_acquire
|
||||
#define arch_atomic_sub_return_release atomic_sub_return_release
|
||||
#define arch_atomic_sub_return_relaxed atomic_sub_return_relaxed
|
||||
|
||||
#ifndef atomic_sub_return_relaxed
|
||||
#define atomic_sub_return_acquire atomic_sub_return
|
||||
#define atomic_sub_return_release atomic_sub_return
|
||||
|
|
@ -221,6 +246,11 @@ atomic_sub_return(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_sub_return_relaxed */
|
||||
|
||||
#define arch_atomic_fetch_sub atomic_fetch_sub
|
||||
#define arch_atomic_fetch_sub_acquire atomic_fetch_sub_acquire
|
||||
#define arch_atomic_fetch_sub_release atomic_fetch_sub_release
|
||||
#define arch_atomic_fetch_sub_relaxed atomic_fetch_sub_relaxed
|
||||
|
||||
#ifndef atomic_fetch_sub_relaxed
|
||||
#define atomic_fetch_sub_acquire atomic_fetch_sub
|
||||
#define atomic_fetch_sub_release atomic_fetch_sub
|
||||
|
|
@ -263,6 +293,8 @@ atomic_fetch_sub(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_sub_relaxed */
|
||||
|
||||
#define arch_atomic_inc atomic_inc
|
||||
|
||||
#ifndef atomic_inc
|
||||
static __always_inline void
|
||||
atomic_inc(atomic_t *v)
|
||||
|
|
@ -272,6 +304,11 @@ atomic_inc(atomic_t *v)
|
|||
#define atomic_inc atomic_inc
|
||||
#endif
|
||||
|
||||
#define arch_atomic_inc_return atomic_inc_return
|
||||
#define arch_atomic_inc_return_acquire atomic_inc_return_acquire
|
||||
#define arch_atomic_inc_return_release atomic_inc_return_release
|
||||
#define arch_atomic_inc_return_relaxed atomic_inc_return_relaxed
|
||||
|
||||
#ifndef atomic_inc_return_relaxed
|
||||
#ifdef atomic_inc_return
|
||||
#define atomic_inc_return_acquire atomic_inc_return
|
||||
|
|
@ -353,6 +390,11 @@ atomic_inc_return(atomic_t *v)
|
|||
|
||||
#endif /* atomic_inc_return_relaxed */
|
||||
|
||||
#define arch_atomic_fetch_inc atomic_fetch_inc
|
||||
#define arch_atomic_fetch_inc_acquire atomic_fetch_inc_acquire
|
||||
#define arch_atomic_fetch_inc_release atomic_fetch_inc_release
|
||||
#define arch_atomic_fetch_inc_relaxed atomic_fetch_inc_relaxed
|
||||
|
||||
#ifndef atomic_fetch_inc_relaxed
|
||||
#ifdef atomic_fetch_inc
|
||||
#define atomic_fetch_inc_acquire atomic_fetch_inc
|
||||
|
|
@ -434,6 +476,8 @@ atomic_fetch_inc(atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_inc_relaxed */
|
||||
|
||||
#define arch_atomic_dec atomic_dec
|
||||
|
||||
#ifndef atomic_dec
|
||||
static __always_inline void
|
||||
atomic_dec(atomic_t *v)
|
||||
|
|
@ -443,6 +487,11 @@ atomic_dec(atomic_t *v)
|
|||
#define atomic_dec atomic_dec
|
||||
#endif
|
||||
|
||||
#define arch_atomic_dec_return atomic_dec_return
|
||||
#define arch_atomic_dec_return_acquire atomic_dec_return_acquire
|
||||
#define arch_atomic_dec_return_release atomic_dec_return_release
|
||||
#define arch_atomic_dec_return_relaxed atomic_dec_return_relaxed
|
||||
|
||||
#ifndef atomic_dec_return_relaxed
|
||||
#ifdef atomic_dec_return
|
||||
#define atomic_dec_return_acquire atomic_dec_return
|
||||
|
|
@ -524,6 +573,11 @@ atomic_dec_return(atomic_t *v)
|
|||
|
||||
#endif /* atomic_dec_return_relaxed */
|
||||
|
||||
#define arch_atomic_fetch_dec atomic_fetch_dec
|
||||
#define arch_atomic_fetch_dec_acquire atomic_fetch_dec_acquire
|
||||
#define arch_atomic_fetch_dec_release atomic_fetch_dec_release
|
||||
#define arch_atomic_fetch_dec_relaxed atomic_fetch_dec_relaxed
|
||||
|
||||
#ifndef atomic_fetch_dec_relaxed
|
||||
#ifdef atomic_fetch_dec
|
||||
#define atomic_fetch_dec_acquire atomic_fetch_dec
|
||||
|
|
@ -605,6 +659,13 @@ atomic_fetch_dec(atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_dec_relaxed */
|
||||
|
||||
#define arch_atomic_and atomic_and
|
||||
|
||||
#define arch_atomic_fetch_and atomic_fetch_and
|
||||
#define arch_atomic_fetch_and_acquire atomic_fetch_and_acquire
|
||||
#define arch_atomic_fetch_and_release atomic_fetch_and_release
|
||||
#define arch_atomic_fetch_and_relaxed atomic_fetch_and_relaxed
|
||||
|
||||
#ifndef atomic_fetch_and_relaxed
|
||||
#define atomic_fetch_and_acquire atomic_fetch_and
|
||||
#define atomic_fetch_and_release atomic_fetch_and
|
||||
|
|
@ -647,6 +708,8 @@ atomic_fetch_and(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_and_relaxed */
|
||||
|
||||
#define arch_atomic_andnot atomic_andnot
|
||||
|
||||
#ifndef atomic_andnot
|
||||
static __always_inline void
|
||||
atomic_andnot(int i, atomic_t *v)
|
||||
|
|
@ -656,6 +719,11 @@ atomic_andnot(int i, atomic_t *v)
|
|||
#define atomic_andnot atomic_andnot
|
||||
#endif
|
||||
|
||||
#define arch_atomic_fetch_andnot atomic_fetch_andnot
|
||||
#define arch_atomic_fetch_andnot_acquire atomic_fetch_andnot_acquire
|
||||
#define arch_atomic_fetch_andnot_release atomic_fetch_andnot_release
|
||||
#define arch_atomic_fetch_andnot_relaxed atomic_fetch_andnot_relaxed
|
||||
|
||||
#ifndef atomic_fetch_andnot_relaxed
|
||||
#ifdef atomic_fetch_andnot
|
||||
#define atomic_fetch_andnot_acquire atomic_fetch_andnot
|
||||
|
|
@ -737,6 +805,13 @@ atomic_fetch_andnot(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_andnot_relaxed */
|
||||
|
||||
#define arch_atomic_or atomic_or
|
||||
|
||||
#define arch_atomic_fetch_or atomic_fetch_or
|
||||
#define arch_atomic_fetch_or_acquire atomic_fetch_or_acquire
|
||||
#define arch_atomic_fetch_or_release atomic_fetch_or_release
|
||||
#define arch_atomic_fetch_or_relaxed atomic_fetch_or_relaxed
|
||||
|
||||
#ifndef atomic_fetch_or_relaxed
|
||||
#define atomic_fetch_or_acquire atomic_fetch_or
|
||||
#define atomic_fetch_or_release atomic_fetch_or
|
||||
|
|
@ -779,6 +854,13 @@ atomic_fetch_or(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_or_relaxed */
|
||||
|
||||
#define arch_atomic_xor atomic_xor
|
||||
|
||||
#define arch_atomic_fetch_xor atomic_fetch_xor
|
||||
#define arch_atomic_fetch_xor_acquire atomic_fetch_xor_acquire
|
||||
#define arch_atomic_fetch_xor_release atomic_fetch_xor_release
|
||||
#define arch_atomic_fetch_xor_relaxed atomic_fetch_xor_relaxed
|
||||
|
||||
#ifndef atomic_fetch_xor_relaxed
|
||||
#define atomic_fetch_xor_acquire atomic_fetch_xor
|
||||
#define atomic_fetch_xor_release atomic_fetch_xor
|
||||
|
|
@ -821,6 +903,11 @@ atomic_fetch_xor(int i, atomic_t *v)
|
|||
|
||||
#endif /* atomic_fetch_xor_relaxed */
|
||||
|
||||
#define arch_atomic_xchg atomic_xchg
|
||||
#define arch_atomic_xchg_acquire atomic_xchg_acquire
|
||||
#define arch_atomic_xchg_release atomic_xchg_release
|
||||
#define arch_atomic_xchg_relaxed atomic_xchg_relaxed
|
||||
|
||||
#ifndef atomic_xchg_relaxed
|
||||
#define atomic_xchg_acquire atomic_xchg
|
||||
#define atomic_xchg_release atomic_xchg
|
||||
|
|
@ -863,6 +950,11 @@ atomic_xchg(atomic_t *v, int i)
|
|||
|
||||
#endif /* atomic_xchg_relaxed */
|
||||
|
||||
#define arch_atomic_cmpxchg atomic_cmpxchg
|
||||
#define arch_atomic_cmpxchg_acquire atomic_cmpxchg_acquire
|
||||
#define arch_atomic_cmpxchg_release atomic_cmpxchg_release
|
||||
#define arch_atomic_cmpxchg_relaxed atomic_cmpxchg_relaxed
|
||||
|
||||
#ifndef atomic_cmpxchg_relaxed
|
||||
#define atomic_cmpxchg_acquire atomic_cmpxchg
|
||||
#define atomic_cmpxchg_release atomic_cmpxchg
|
||||
|
|
@ -905,6 +997,11 @@ atomic_cmpxchg(atomic_t *v, int old, int new)
|
|||
|
||||
#endif /* atomic_cmpxchg_relaxed */
|
||||
|
||||
#define arch_atomic_try_cmpxchg atomic_try_cmpxchg
|
||||
#define arch_atomic_try_cmpxchg_acquire atomic_try_cmpxchg_acquire
|
||||
#define arch_atomic_try_cmpxchg_release atomic_try_cmpxchg_release
|
||||
#define arch_atomic_try_cmpxchg_relaxed atomic_try_cmpxchg_relaxed
|
||||
|
||||
#ifndef atomic_try_cmpxchg_relaxed
|
||||
#ifdef atomic_try_cmpxchg
|
||||
#define atomic_try_cmpxchg_acquire atomic_try_cmpxchg
|
||||
|
|
@ -1002,6 +1099,8 @@ atomic_try_cmpxchg(atomic_t *v, int *old, int new)
|
|||
|
||||
#endif /* atomic_try_cmpxchg_relaxed */
|
||||
|
||||
#define arch_atomic_sub_and_test atomic_sub_and_test
|
||||
|
||||
#ifndef atomic_sub_and_test
|
||||
/**
|
||||
* atomic_sub_and_test - subtract value from variable and test result
|
||||
|
|
@ -1020,6 +1119,8 @@ atomic_sub_and_test(int i, atomic_t *v)
|
|||
#define atomic_sub_and_test atomic_sub_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic_dec_and_test atomic_dec_and_test
|
||||
|
||||
#ifndef atomic_dec_and_test
|
||||
/**
|
||||
* atomic_dec_and_test - decrement and test
|
||||
|
|
@ -1037,6 +1138,8 @@ atomic_dec_and_test(atomic_t *v)
|
|||
#define atomic_dec_and_test atomic_dec_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic_inc_and_test atomic_inc_and_test
|
||||
|
||||
#ifndef atomic_inc_and_test
|
||||
/**
|
||||
* atomic_inc_and_test - increment and test
|
||||
|
|
@ -1054,6 +1157,8 @@ atomic_inc_and_test(atomic_t *v)
|
|||
#define atomic_inc_and_test atomic_inc_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic_add_negative atomic_add_negative
|
||||
|
||||
#ifndef atomic_add_negative
|
||||
/**
|
||||
* atomic_add_negative - add and test if negative
|
||||
|
|
@ -1072,6 +1177,8 @@ atomic_add_negative(int i, atomic_t *v)
|
|||
#define atomic_add_negative atomic_add_negative
|
||||
#endif
|
||||
|
||||
#define arch_atomic_fetch_add_unless atomic_fetch_add_unless
|
||||
|
||||
#ifndef atomic_fetch_add_unless
|
||||
/**
|
||||
* atomic_fetch_add_unless - add unless the number is already a given value
|
||||
|
|
@ -1097,6 +1204,8 @@ atomic_fetch_add_unless(atomic_t *v, int a, int u)
|
|||
#define atomic_fetch_add_unless atomic_fetch_add_unless
|
||||
#endif
|
||||
|
||||
#define arch_atomic_add_unless atomic_add_unless
|
||||
|
||||
#ifndef atomic_add_unless
|
||||
/**
|
||||
* atomic_add_unless - add unless the number is already a given value
|
||||
|
|
@ -1115,6 +1224,8 @@ atomic_add_unless(atomic_t *v, int a, int u)
|
|||
#define atomic_add_unless atomic_add_unless
|
||||
#endif
|
||||
|
||||
#define arch_atomic_inc_not_zero atomic_inc_not_zero
|
||||
|
||||
#ifndef atomic_inc_not_zero
|
||||
/**
|
||||
* atomic_inc_not_zero - increment unless the number is zero
|
||||
|
|
@ -1131,6 +1242,8 @@ atomic_inc_not_zero(atomic_t *v)
|
|||
#define atomic_inc_not_zero atomic_inc_not_zero
|
||||
#endif
|
||||
|
||||
#define arch_atomic_inc_unless_negative atomic_inc_unless_negative
|
||||
|
||||
#ifndef atomic_inc_unless_negative
|
||||
static __always_inline bool
|
||||
atomic_inc_unless_negative(atomic_t *v)
|
||||
|
|
@ -1147,6 +1260,8 @@ atomic_inc_unless_negative(atomic_t *v)
|
|||
#define atomic_inc_unless_negative atomic_inc_unless_negative
|
||||
#endif
|
||||
|
||||
#define arch_atomic_dec_unless_positive atomic_dec_unless_positive
|
||||
|
||||
#ifndef atomic_dec_unless_positive
|
||||
static __always_inline bool
|
||||
atomic_dec_unless_positive(atomic_t *v)
|
||||
|
|
@ -1163,6 +1278,8 @@ atomic_dec_unless_positive(atomic_t *v)
|
|||
#define atomic_dec_unless_positive atomic_dec_unless_positive
|
||||
#endif
|
||||
|
||||
#define arch_atomic_dec_if_positive atomic_dec_if_positive
|
||||
|
||||
#ifndef atomic_dec_if_positive
|
||||
static __always_inline int
|
||||
atomic_dec_if_positive(atomic_t *v)
|
||||
|
|
@ -1184,6 +1301,9 @@ atomic_dec_if_positive(atomic_t *v)
|
|||
#include <asm-generic/atomic64.h>
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_read atomic64_read
|
||||
#define arch_atomic64_read_acquire atomic64_read_acquire
|
||||
|
||||
#ifndef atomic64_read_acquire
|
||||
static __always_inline s64
|
||||
atomic64_read_acquire(const atomic64_t *v)
|
||||
|
|
@ -1193,6 +1313,9 @@ atomic64_read_acquire(const atomic64_t *v)
|
|||
#define atomic64_read_acquire atomic64_read_acquire
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_set atomic64_set
|
||||
#define arch_atomic64_set_release atomic64_set_release
|
||||
|
||||
#ifndef atomic64_set_release
|
||||
static __always_inline void
|
||||
atomic64_set_release(atomic64_t *v, s64 i)
|
||||
|
|
@ -1202,6 +1325,13 @@ atomic64_set_release(atomic64_t *v, s64 i)
|
|||
#define atomic64_set_release atomic64_set_release
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_add atomic64_add
|
||||
|
||||
#define arch_atomic64_add_return atomic64_add_return
|
||||
#define arch_atomic64_add_return_acquire atomic64_add_return_acquire
|
||||
#define arch_atomic64_add_return_release atomic64_add_return_release
|
||||
#define arch_atomic64_add_return_relaxed atomic64_add_return_relaxed
|
||||
|
||||
#ifndef atomic64_add_return_relaxed
|
||||
#define atomic64_add_return_acquire atomic64_add_return
|
||||
#define atomic64_add_return_release atomic64_add_return
|
||||
|
|
@ -1244,6 +1374,11 @@ atomic64_add_return(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_add_return_relaxed */
|
||||
|
||||
#define arch_atomic64_fetch_add atomic64_fetch_add
|
||||
#define arch_atomic64_fetch_add_acquire atomic64_fetch_add_acquire
|
||||
#define arch_atomic64_fetch_add_release atomic64_fetch_add_release
|
||||
#define arch_atomic64_fetch_add_relaxed atomic64_fetch_add_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_add_relaxed
|
||||
#define atomic64_fetch_add_acquire atomic64_fetch_add
|
||||
#define atomic64_fetch_add_release atomic64_fetch_add
|
||||
|
|
@ -1286,6 +1421,13 @@ atomic64_fetch_add(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_add_relaxed */
|
||||
|
||||
#define arch_atomic64_sub atomic64_sub
|
||||
|
||||
#define arch_atomic64_sub_return atomic64_sub_return
|
||||
#define arch_atomic64_sub_return_acquire atomic64_sub_return_acquire
|
||||
#define arch_atomic64_sub_return_release atomic64_sub_return_release
|
||||
#define arch_atomic64_sub_return_relaxed atomic64_sub_return_relaxed
|
||||
|
||||
#ifndef atomic64_sub_return_relaxed
|
||||
#define atomic64_sub_return_acquire atomic64_sub_return
|
||||
#define atomic64_sub_return_release atomic64_sub_return
|
||||
|
|
@ -1328,6 +1470,11 @@ atomic64_sub_return(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_sub_return_relaxed */
|
||||
|
||||
#define arch_atomic64_fetch_sub atomic64_fetch_sub
|
||||
#define arch_atomic64_fetch_sub_acquire atomic64_fetch_sub_acquire
|
||||
#define arch_atomic64_fetch_sub_release atomic64_fetch_sub_release
|
||||
#define arch_atomic64_fetch_sub_relaxed atomic64_fetch_sub_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_sub_relaxed
|
||||
#define atomic64_fetch_sub_acquire atomic64_fetch_sub
|
||||
#define atomic64_fetch_sub_release atomic64_fetch_sub
|
||||
|
|
@ -1370,6 +1517,8 @@ atomic64_fetch_sub(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_sub_relaxed */
|
||||
|
||||
#define arch_atomic64_inc atomic64_inc
|
||||
|
||||
#ifndef atomic64_inc
|
||||
static __always_inline void
|
||||
atomic64_inc(atomic64_t *v)
|
||||
|
|
@ -1379,6 +1528,11 @@ atomic64_inc(atomic64_t *v)
|
|||
#define atomic64_inc atomic64_inc
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_inc_return atomic64_inc_return
|
||||
#define arch_atomic64_inc_return_acquire atomic64_inc_return_acquire
|
||||
#define arch_atomic64_inc_return_release atomic64_inc_return_release
|
||||
#define arch_atomic64_inc_return_relaxed atomic64_inc_return_relaxed
|
||||
|
||||
#ifndef atomic64_inc_return_relaxed
|
||||
#ifdef atomic64_inc_return
|
||||
#define atomic64_inc_return_acquire atomic64_inc_return
|
||||
|
|
@ -1460,6 +1614,11 @@ atomic64_inc_return(atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_inc_return_relaxed */
|
||||
|
||||
#define arch_atomic64_fetch_inc atomic64_fetch_inc
|
||||
#define arch_atomic64_fetch_inc_acquire atomic64_fetch_inc_acquire
|
||||
#define arch_atomic64_fetch_inc_release atomic64_fetch_inc_release
|
||||
#define arch_atomic64_fetch_inc_relaxed atomic64_fetch_inc_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_inc_relaxed
|
||||
#ifdef atomic64_fetch_inc
|
||||
#define atomic64_fetch_inc_acquire atomic64_fetch_inc
|
||||
|
|
@ -1541,6 +1700,8 @@ atomic64_fetch_inc(atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_inc_relaxed */
|
||||
|
||||
#define arch_atomic64_dec atomic64_dec
|
||||
|
||||
#ifndef atomic64_dec
|
||||
static __always_inline void
|
||||
atomic64_dec(atomic64_t *v)
|
||||
|
|
@ -1550,6 +1711,11 @@ atomic64_dec(atomic64_t *v)
|
|||
#define atomic64_dec atomic64_dec
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_dec_return atomic64_dec_return
|
||||
#define arch_atomic64_dec_return_acquire atomic64_dec_return_acquire
|
||||
#define arch_atomic64_dec_return_release atomic64_dec_return_release
|
||||
#define arch_atomic64_dec_return_relaxed atomic64_dec_return_relaxed
|
||||
|
||||
#ifndef atomic64_dec_return_relaxed
|
||||
#ifdef atomic64_dec_return
|
||||
#define atomic64_dec_return_acquire atomic64_dec_return
|
||||
|
|
@ -1631,6 +1797,11 @@ atomic64_dec_return(atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_dec_return_relaxed */
|
||||
|
||||
#define arch_atomic64_fetch_dec atomic64_fetch_dec
|
||||
#define arch_atomic64_fetch_dec_acquire atomic64_fetch_dec_acquire
|
||||
#define arch_atomic64_fetch_dec_release atomic64_fetch_dec_release
|
||||
#define arch_atomic64_fetch_dec_relaxed atomic64_fetch_dec_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_dec_relaxed
|
||||
#ifdef atomic64_fetch_dec
|
||||
#define atomic64_fetch_dec_acquire atomic64_fetch_dec
|
||||
|
|
@ -1712,6 +1883,13 @@ atomic64_fetch_dec(atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_dec_relaxed */
|
||||
|
||||
#define arch_atomic64_and atomic64_and
|
||||
|
||||
#define arch_atomic64_fetch_and atomic64_fetch_and
|
||||
#define arch_atomic64_fetch_and_acquire atomic64_fetch_and_acquire
|
||||
#define arch_atomic64_fetch_and_release atomic64_fetch_and_release
|
||||
#define arch_atomic64_fetch_and_relaxed atomic64_fetch_and_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_and_relaxed
|
||||
#define atomic64_fetch_and_acquire atomic64_fetch_and
|
||||
#define atomic64_fetch_and_release atomic64_fetch_and
|
||||
|
|
@ -1754,6 +1932,8 @@ atomic64_fetch_and(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_and_relaxed */
|
||||
|
||||
#define arch_atomic64_andnot atomic64_andnot
|
||||
|
||||
#ifndef atomic64_andnot
|
||||
static __always_inline void
|
||||
atomic64_andnot(s64 i, atomic64_t *v)
|
||||
|
|
@ -1763,6 +1943,11 @@ atomic64_andnot(s64 i, atomic64_t *v)
|
|||
#define atomic64_andnot atomic64_andnot
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_fetch_andnot atomic64_fetch_andnot
|
||||
#define arch_atomic64_fetch_andnot_acquire atomic64_fetch_andnot_acquire
|
||||
#define arch_atomic64_fetch_andnot_release atomic64_fetch_andnot_release
|
||||
#define arch_atomic64_fetch_andnot_relaxed atomic64_fetch_andnot_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_andnot_relaxed
|
||||
#ifdef atomic64_fetch_andnot
|
||||
#define atomic64_fetch_andnot_acquire atomic64_fetch_andnot
|
||||
|
|
@ -1844,6 +2029,13 @@ atomic64_fetch_andnot(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_andnot_relaxed */
|
||||
|
||||
#define arch_atomic64_or atomic64_or
|
||||
|
||||
#define arch_atomic64_fetch_or atomic64_fetch_or
|
||||
#define arch_atomic64_fetch_or_acquire atomic64_fetch_or_acquire
|
||||
#define arch_atomic64_fetch_or_release atomic64_fetch_or_release
|
||||
#define arch_atomic64_fetch_or_relaxed atomic64_fetch_or_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_or_relaxed
|
||||
#define atomic64_fetch_or_acquire atomic64_fetch_or
|
||||
#define atomic64_fetch_or_release atomic64_fetch_or
|
||||
|
|
@ -1886,6 +2078,13 @@ atomic64_fetch_or(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_or_relaxed */
|
||||
|
||||
#define arch_atomic64_xor atomic64_xor
|
||||
|
||||
#define arch_atomic64_fetch_xor atomic64_fetch_xor
|
||||
#define arch_atomic64_fetch_xor_acquire atomic64_fetch_xor_acquire
|
||||
#define arch_atomic64_fetch_xor_release atomic64_fetch_xor_release
|
||||
#define arch_atomic64_fetch_xor_relaxed atomic64_fetch_xor_relaxed
|
||||
|
||||
#ifndef atomic64_fetch_xor_relaxed
|
||||
#define atomic64_fetch_xor_acquire atomic64_fetch_xor
|
||||
#define atomic64_fetch_xor_release atomic64_fetch_xor
|
||||
|
|
@ -1928,6 +2127,11 @@ atomic64_fetch_xor(s64 i, atomic64_t *v)
|
|||
|
||||
#endif /* atomic64_fetch_xor_relaxed */
|
||||
|
||||
#define arch_atomic64_xchg atomic64_xchg
|
||||
#define arch_atomic64_xchg_acquire atomic64_xchg_acquire
|
||||
#define arch_atomic64_xchg_release atomic64_xchg_release
|
||||
#define arch_atomic64_xchg_relaxed atomic64_xchg_relaxed
|
||||
|
||||
#ifndef atomic64_xchg_relaxed
|
||||
#define atomic64_xchg_acquire atomic64_xchg
|
||||
#define atomic64_xchg_release atomic64_xchg
|
||||
|
|
@ -1970,6 +2174,11 @@ atomic64_xchg(atomic64_t *v, s64 i)
|
|||
|
||||
#endif /* atomic64_xchg_relaxed */
|
||||
|
||||
#define arch_atomic64_cmpxchg atomic64_cmpxchg
|
||||
#define arch_atomic64_cmpxchg_acquire atomic64_cmpxchg_acquire
|
||||
#define arch_atomic64_cmpxchg_release atomic64_cmpxchg_release
|
||||
#define arch_atomic64_cmpxchg_relaxed atomic64_cmpxchg_relaxed
|
||||
|
||||
#ifndef atomic64_cmpxchg_relaxed
|
||||
#define atomic64_cmpxchg_acquire atomic64_cmpxchg
|
||||
#define atomic64_cmpxchg_release atomic64_cmpxchg
|
||||
|
|
@ -2012,6 +2221,11 @@ atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new)
|
|||
|
||||
#endif /* atomic64_cmpxchg_relaxed */
|
||||
|
||||
#define arch_atomic64_try_cmpxchg atomic64_try_cmpxchg
|
||||
#define arch_atomic64_try_cmpxchg_acquire atomic64_try_cmpxchg_acquire
|
||||
#define arch_atomic64_try_cmpxchg_release atomic64_try_cmpxchg_release
|
||||
#define arch_atomic64_try_cmpxchg_relaxed atomic64_try_cmpxchg_relaxed
|
||||
|
||||
#ifndef atomic64_try_cmpxchg_relaxed
|
||||
#ifdef atomic64_try_cmpxchg
|
||||
#define atomic64_try_cmpxchg_acquire atomic64_try_cmpxchg
|
||||
|
|
@ -2109,6 +2323,8 @@ atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new)
|
|||
|
||||
#endif /* atomic64_try_cmpxchg_relaxed */
|
||||
|
||||
#define arch_atomic64_sub_and_test atomic64_sub_and_test
|
||||
|
||||
#ifndef atomic64_sub_and_test
|
||||
/**
|
||||
* atomic64_sub_and_test - subtract value from variable and test result
|
||||
|
|
@ -2127,6 +2343,8 @@ atomic64_sub_and_test(s64 i, atomic64_t *v)
|
|||
#define atomic64_sub_and_test atomic64_sub_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_dec_and_test atomic64_dec_and_test
|
||||
|
||||
#ifndef atomic64_dec_and_test
|
||||
/**
|
||||
* atomic64_dec_and_test - decrement and test
|
||||
|
|
@ -2144,6 +2362,8 @@ atomic64_dec_and_test(atomic64_t *v)
|
|||
#define atomic64_dec_and_test atomic64_dec_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_inc_and_test atomic64_inc_and_test
|
||||
|
||||
#ifndef atomic64_inc_and_test
|
||||
/**
|
||||
* atomic64_inc_and_test - increment and test
|
||||
|
|
@ -2161,6 +2381,8 @@ atomic64_inc_and_test(atomic64_t *v)
|
|||
#define atomic64_inc_and_test atomic64_inc_and_test
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_add_negative atomic64_add_negative
|
||||
|
||||
#ifndef atomic64_add_negative
|
||||
/**
|
||||
* atomic64_add_negative - add and test if negative
|
||||
|
|
@ -2179,6 +2401,8 @@ atomic64_add_negative(s64 i, atomic64_t *v)
|
|||
#define atomic64_add_negative atomic64_add_negative
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_fetch_add_unless atomic64_fetch_add_unless
|
||||
|
||||
#ifndef atomic64_fetch_add_unless
|
||||
/**
|
||||
* atomic64_fetch_add_unless - add unless the number is already a given value
|
||||
|
|
@ -2204,6 +2428,8 @@ atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
|
|||
#define atomic64_fetch_add_unless atomic64_fetch_add_unless
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_add_unless atomic64_add_unless
|
||||
|
||||
#ifndef atomic64_add_unless
|
||||
/**
|
||||
* atomic64_add_unless - add unless the number is already a given value
|
||||
|
|
@ -2222,6 +2448,8 @@ atomic64_add_unless(atomic64_t *v, s64 a, s64 u)
|
|||
#define atomic64_add_unless atomic64_add_unless
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_inc_not_zero atomic64_inc_not_zero
|
||||
|
||||
#ifndef atomic64_inc_not_zero
|
||||
/**
|
||||
* atomic64_inc_not_zero - increment unless the number is zero
|
||||
|
|
@ -2238,6 +2466,8 @@ atomic64_inc_not_zero(atomic64_t *v)
|
|||
#define atomic64_inc_not_zero atomic64_inc_not_zero
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_inc_unless_negative atomic64_inc_unless_negative
|
||||
|
||||
#ifndef atomic64_inc_unless_negative
|
||||
static __always_inline bool
|
||||
atomic64_inc_unless_negative(atomic64_t *v)
|
||||
|
|
@ -2254,6 +2484,8 @@ atomic64_inc_unless_negative(atomic64_t *v)
|
|||
#define atomic64_inc_unless_negative atomic64_inc_unless_negative
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_dec_unless_positive atomic64_dec_unless_positive
|
||||
|
||||
#ifndef atomic64_dec_unless_positive
|
||||
static __always_inline bool
|
||||
atomic64_dec_unless_positive(atomic64_t *v)
|
||||
|
|
@ -2270,6 +2502,8 @@ atomic64_dec_unless_positive(atomic64_t *v)
|
|||
#define atomic64_dec_unless_positive atomic64_dec_unless_positive
|
||||
#endif
|
||||
|
||||
#define arch_atomic64_dec_if_positive atomic64_dec_if_positive
|
||||
|
||||
#ifndef atomic64_dec_if_positive
|
||||
static __always_inline s64
|
||||
atomic64_dec_if_positive(atomic64_t *v)
|
||||
|
|
@ -2288,4 +2522,4 @@ atomic64_dec_if_positive(atomic64_t *v)
|
|||
#endif
|
||||
|
||||
#endif /* _LINUX_ATOMIC_FALLBACK_H */
|
||||
// 1fac0941c79bf0ae100723cc2ac9b94061f0b67a
|
||||
// 9d95b56f98d82a2a26c7b79ccdd0c47572d50a6f
|
||||
|
|
|
|||
|
|
@ -18,8 +18,7 @@
|
|||
* position @h. For example
|
||||
* GENMASK_ULL(39, 21) gives us the 64bit vector 0x000000ffffe00000.
|
||||
*/
|
||||
#if !defined(__ASSEMBLY__) && \
|
||||
(!defined(CONFIG_CC_IS_GCC) || CONFIG_GCC_VERSION >= 49000)
|
||||
#if !defined(__ASSEMBLY__)
|
||||
#include <linux/build_bug.h>
|
||||
#define GENMASK_INPUT_CHECK(h, l) \
|
||||
(BUILD_BUG_ON_ZERO(__builtin_choose_expr( \
|
||||
|
|
|
|||
|
|
@ -590,6 +590,7 @@ struct request_queue {
|
|||
u64 write_hints[BLK_MAX_WRITE_HINTS];
|
||||
};
|
||||
|
||||
/* Keep blk_queue_flag_name[] in sync with the definitions below */
|
||||
#define QUEUE_FLAG_STOPPED 0 /* queue is stopped */
|
||||
#define QUEUE_FLAG_DYING 1 /* queue being torn down */
|
||||
#define QUEUE_FLAG_NOMERGES 3 /* disable merge attempts */
|
||||
|
|
|
|||
|
|
@ -33,7 +33,7 @@ int netns_bpf_prog_query(const union bpf_attr *attr,
|
|||
union bpf_attr __user *uattr);
|
||||
int netns_bpf_prog_attach(const union bpf_attr *attr,
|
||||
struct bpf_prog *prog);
|
||||
int netns_bpf_prog_detach(const union bpf_attr *attr);
|
||||
int netns_bpf_prog_detach(const union bpf_attr *attr, enum bpf_prog_type ptype);
|
||||
int netns_bpf_link_create(const union bpf_attr *attr,
|
||||
struct bpf_prog *prog);
|
||||
#else
|
||||
|
|
@ -49,7 +49,8 @@ static inline int netns_bpf_prog_attach(const union bpf_attr *attr,
|
|||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
static inline int netns_bpf_prog_detach(const union bpf_attr *attr)
|
||||
static inline int netns_bpf_prog_detach(const union bpf_attr *attr,
|
||||
enum bpf_prog_type ptype)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -1543,13 +1543,16 @@ static inline void bpf_map_offload_map_free(struct bpf_map *map)
|
|||
#endif /* CONFIG_NET && CONFIG_BPF_SYSCALL */
|
||||
|
||||
#if defined(CONFIG_BPF_STREAM_PARSER)
|
||||
int sock_map_prog_update(struct bpf_map *map, struct bpf_prog *prog, u32 which);
|
||||
int sock_map_prog_update(struct bpf_map *map, struct bpf_prog *prog,
|
||||
struct bpf_prog *old, u32 which);
|
||||
int sock_map_get_from_fd(const union bpf_attr *attr, struct bpf_prog *prog);
|
||||
int sock_map_prog_detach(const union bpf_attr *attr, enum bpf_prog_type ptype);
|
||||
void sock_map_unhash(struct sock *sk);
|
||||
void sock_map_close(struct sock *sk, long timeout);
|
||||
#else
|
||||
static inline int sock_map_prog_update(struct bpf_map *map,
|
||||
struct bpf_prog *prog, u32 which)
|
||||
struct bpf_prog *prog,
|
||||
struct bpf_prog *old, u32 which)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
|
@ -1559,6 +1562,12 @@ static inline int sock_map_get_from_fd(const union bpf_attr *attr,
|
|||
{
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
static inline int sock_map_prog_detach(const union bpf_attr *attr,
|
||||
enum bpf_prog_type ptype)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
#endif /* CONFIG_BPF_STREAM_PARSER */
|
||||
|
||||
#if defined(CONFIG_INET) && defined(CONFIG_BPF_SYSCALL)
|
||||
|
|
|
|||
|
|
@ -82,6 +82,11 @@ static inline bool btf_type_is_int(const struct btf_type *t)
|
|||
return BTF_INFO_KIND(t->info) == BTF_KIND_INT;
|
||||
}
|
||||
|
||||
static inline bool btf_type_is_small_int(const struct btf_type *t)
|
||||
{
|
||||
return btf_type_is_int(t) && t->size <= sizeof(u64);
|
||||
}
|
||||
|
||||
static inline bool btf_type_is_enum(const struct btf_type *t)
|
||||
{
|
||||
return BTF_INFO_KIND(t->info) == BTF_KIND_ENUM;
|
||||
|
|
|
|||
|
|
@ -790,7 +790,9 @@ struct sock_cgroup_data {
|
|||
union {
|
||||
#ifdef __LITTLE_ENDIAN
|
||||
struct {
|
||||
u8 is_data;
|
||||
u8 is_data : 1;
|
||||
u8 no_refcnt : 1;
|
||||
u8 unused : 6;
|
||||
u8 padding;
|
||||
u16 prioidx;
|
||||
u32 classid;
|
||||
|
|
@ -800,7 +802,9 @@ struct sock_cgroup_data {
|
|||
u32 classid;
|
||||
u16 prioidx;
|
||||
u8 padding;
|
||||
u8 is_data;
|
||||
u8 unused : 6;
|
||||
u8 no_refcnt : 1;
|
||||
u8 is_data : 1;
|
||||
} __packed;
|
||||
#endif
|
||||
u64 val;
|
||||
|
|
|
|||
|
|
@ -822,6 +822,7 @@ extern spinlock_t cgroup_sk_update_lock;
|
|||
|
||||
void cgroup_sk_alloc_disable(void);
|
||||
void cgroup_sk_alloc(struct sock_cgroup_data *skcd);
|
||||
void cgroup_sk_clone(struct sock_cgroup_data *skcd);
|
||||
void cgroup_sk_free(struct sock_cgroup_data *skcd);
|
||||
|
||||
static inline struct cgroup *sock_cgroup_ptr(struct sock_cgroup_data *skcd)
|
||||
|
|
@ -835,7 +836,7 @@ static inline struct cgroup *sock_cgroup_ptr(struct sock_cgroup_data *skcd)
|
|||
*/
|
||||
v = READ_ONCE(skcd->val);
|
||||
|
||||
if (v & 1)
|
||||
if (v & 3)
|
||||
return &cgrp_dfl_root.cgrp;
|
||||
|
||||
return (struct cgroup *)(unsigned long)v ?: &cgrp_dfl_root.cgrp;
|
||||
|
|
@ -847,6 +848,7 @@ static inline struct cgroup *sock_cgroup_ptr(struct sock_cgroup_data *skcd)
|
|||
#else /* CONFIG_CGROUP_DATA */
|
||||
|
||||
static inline void cgroup_sk_alloc(struct sock_cgroup_data *skcd) {}
|
||||
static inline void cgroup_sk_clone(struct sock_cgroup_data *skcd) {}
|
||||
static inline void cgroup_sk_free(struct sock_cgroup_data *skcd) {}
|
||||
|
||||
#endif /* CONFIG_CGROUP_DATA */
|
||||
|
|
|
|||
|
|
@ -33,6 +33,14 @@
|
|||
#define __no_sanitize_thread
|
||||
#endif
|
||||
|
||||
#if __has_feature(undefined_behavior_sanitizer)
|
||||
/* GCC does not have __SANITIZE_UNDEFINED__ */
|
||||
#define __no_sanitize_undefined \
|
||||
__attribute__((no_sanitize("undefined")))
|
||||
#else
|
||||
#define __no_sanitize_undefined
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Not all versions of clang implement the the type-generic versions
|
||||
* of the builtin overflow checkers. Fortunately, clang implements
|
||||
|
|
|
|||
|
|
@ -11,7 +11,7 @@
|
|||
+ __GNUC_PATCHLEVEL__)
|
||||
|
||||
/* https://gcc.gnu.org/bugzilla/show_bug.cgi?id=58145 */
|
||||
#if GCC_VERSION < 40800
|
||||
#if GCC_VERSION < 40900
|
||||
# error Sorry, your compiler is too old - please upgrade it.
|
||||
#endif
|
||||
|
||||
|
|
@ -150,6 +150,12 @@
|
|||
#define __no_sanitize_thread
|
||||
#endif
|
||||
|
||||
#if __has_attribute(__no_sanitize_undefined__)
|
||||
#define __no_sanitize_undefined __attribute__((no_sanitize_undefined))
|
||||
#else
|
||||
#define __no_sanitize_undefined
|
||||
#endif
|
||||
|
||||
#if GCC_VERSION >= 50100
|
||||
#define COMPILER_HAS_GENERIC_BUILTIN_OVERFLOW 1
|
||||
#endif
|
||||
|
|
|
|||
|
|
@ -123,7 +123,7 @@ void ftrace_likely_update(struct ftrace_likely_data *f, int val,
|
|||
#ifdef CONFIG_DEBUG_ENTRY
|
||||
/* Begin/end of an instrumentation safe region */
|
||||
#define instrumentation_begin() ({ \
|
||||
asm volatile("%c0:\n\t" \
|
||||
asm volatile("%c0: nop\n\t" \
|
||||
".pushsection .discard.instr_begin\n\t" \
|
||||
".long %c0b - .\n\t" \
|
||||
".popsection\n\t" : : "i" (__COUNTER__)); \
|
||||
|
|
|
|||
|
|
@ -40,6 +40,7 @@
|
|||
# define __GCC4_has_attribute___noclone__ 1
|
||||
# define __GCC4_has_attribute___nonstring__ 0
|
||||
# define __GCC4_has_attribute___no_sanitize_address__ (__GNUC_MINOR__ >= 8)
|
||||
# define __GCC4_has_attribute___no_sanitize_undefined__ (__GNUC_MINOR__ >= 9)
|
||||
# define __GCC4_has_attribute___fallthrough__ 0
|
||||
#endif
|
||||
|
||||
|
|
|
|||
|
|
@ -118,10 +118,6 @@ struct ftrace_likely_data {
|
|||
#define notrace __attribute__((__no_instrument_function__))
|
||||
#endif
|
||||
|
||||
/* Section for code which can't be instrumented at all */
|
||||
#define noinstr \
|
||||
noinline notrace __attribute((__section__(".noinstr.text")))
|
||||
|
||||
/*
|
||||
* it doesn't make sense on ARM (currently the only user of __naked)
|
||||
* to trace naked functions because then mcount is called without
|
||||
|
|
@ -193,16 +189,18 @@ struct ftrace_likely_data {
|
|||
|
||||
#define __no_kcsan __no_sanitize_thread
|
||||
#ifdef __SANITIZE_THREAD__
|
||||
# define __no_kcsan_or_inline __no_kcsan notrace __maybe_unused
|
||||
# define __no_sanitize_or_inline __no_kcsan_or_inline
|
||||
#else
|
||||
# define __no_kcsan_or_inline __always_inline
|
||||
# define __no_sanitize_or_inline __no_kcsan notrace __maybe_unused
|
||||
#endif
|
||||
|
||||
#ifndef __no_sanitize_or_inline
|
||||
#define __no_sanitize_or_inline __always_inline
|
||||
#endif
|
||||
|
||||
/* Section for code which can't be instrumented at all */
|
||||
#define noinstr \
|
||||
noinline notrace __attribute((__section__(".noinstr.text"))) \
|
||||
__no_kcsan __no_sanitize_address
|
||||
|
||||
#endif /* __KERNEL__ */
|
||||
|
||||
#endif /* __ASSEMBLY__ */
|
||||
|
|
@ -254,32 +252,8 @@ struct ftrace_likely_data {
|
|||
* __unqual_scalar_typeof(x) - Declare an unqualified scalar type, leaving
|
||||
* non-scalar types unchanged.
|
||||
*/
|
||||
#if (defined(CONFIG_CC_IS_GCC) && CONFIG_GCC_VERSION < 40900) || defined(__CHECKER__)
|
||||
/*
|
||||
* We build this out of a couple of helper macros in a vain attempt to
|
||||
* help you keep your lunch down while reading it.
|
||||
*/
|
||||
#define __pick_scalar_type(x, type, otherwise) \
|
||||
__builtin_choose_expr(__same_type(x, type), (type)0, otherwise)
|
||||
|
||||
/*
|
||||
* 'char' is not type-compatible with either 'signed char' or 'unsigned char',
|
||||
* so we include the naked type here as well as the signed/unsigned variants.
|
||||
*/
|
||||
#define __pick_integer_type(x, type, otherwise) \
|
||||
__pick_scalar_type(x, type, \
|
||||
__pick_scalar_type(x, unsigned type, \
|
||||
__pick_scalar_type(x, signed type, otherwise)))
|
||||
|
||||
#define __unqual_scalar_typeof(x) typeof( \
|
||||
__pick_integer_type(x, char, \
|
||||
__pick_integer_type(x, short, \
|
||||
__pick_integer_type(x, int, \
|
||||
__pick_integer_type(x, long, \
|
||||
__pick_integer_type(x, long long, x))))))
|
||||
#else
|
||||
/*
|
||||
* If supported, prefer C11 _Generic for better compile-times. As above, 'char'
|
||||
* Prefer C11 _Generic for better compile-times and simpler code. Note: 'char'
|
||||
* is not type-compatible with 'signed char', and we define a separate case.
|
||||
*/
|
||||
#define __scalar_type_to_expr_cases(type) \
|
||||
|
|
@ -295,7 +269,6 @@ struct ftrace_likely_data {
|
|||
__scalar_type_to_expr_cases(long), \
|
||||
__scalar_type_to_expr_cases(long long), \
|
||||
default: (x)))
|
||||
#endif
|
||||
|
||||
/* Is this type a native word size -- useful for atomic operations */
|
||||
#define __native_word(t) \
|
||||
|
|
|
|||
|
|
@ -433,7 +433,8 @@ enum dl_dev_state {
|
|||
* @suppliers: List of links to supplier devices.
|
||||
* @consumers: List of links to consumer devices.
|
||||
* @needs_suppliers: Hook to global list of devices waiting for suppliers.
|
||||
* @defer_sync: Hook to global list of devices that have deferred sync_state.
|
||||
* @defer_hook: Hook to global list of devices that have deferred sync_state or
|
||||
* deferred fw_devlink.
|
||||
* @need_for_probe: If needs_suppliers is on a list, this indicates if the
|
||||
* suppliers are needed for probe or not.
|
||||
* @status: Driver status information.
|
||||
|
|
@ -442,7 +443,7 @@ struct dev_links_info {
|
|||
struct list_head suppliers;
|
||||
struct list_head consumers;
|
||||
struct list_head needs_suppliers;
|
||||
struct list_head defer_sync;
|
||||
struct list_head defer_hook;
|
||||
bool need_for_probe;
|
||||
enum dl_dev_state status;
|
||||
};
|
||||
|
|
|
|||
|
|
@ -311,6 +311,7 @@ struct dma_buf {
|
|||
void *vmap_ptr;
|
||||
const char *exp_name;
|
||||
const char *name;
|
||||
spinlock_t name_lock; /* spinlock to protect name access */
|
||||
struct module *owner;
|
||||
struct list_head list_node;
|
||||
void *priv;
|
||||
|
|
|
|||
|
|
@ -69,6 +69,7 @@ static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size,
|
|||
u64 dma_direct_get_required_mask(struct device *dev);
|
||||
gfp_t dma_direct_optimal_gfp_mask(struct device *dev, u64 dma_mask,
|
||||
u64 *phys_mask);
|
||||
bool dma_coherent_ok(struct device *dev, phys_addr_t phys, size_t size);
|
||||
void *dma_direct_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
|
||||
gfp_t gfp, unsigned long attrs);
|
||||
void dma_direct_free(struct device *dev, size_t size, void *cpu_addr,
|
||||
|
|
@ -77,8 +78,6 @@ void *dma_direct_alloc_pages(struct device *dev, size_t size,
|
|||
dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs);
|
||||
void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr,
|
||||
dma_addr_t dma_addr, unsigned long attrs);
|
||||
struct page *__dma_direct_alloc_pages(struct device *dev, size_t size,
|
||||
gfp_t gfp, unsigned long attrs);
|
||||
int dma_direct_get_sgtable(struct device *dev, struct sg_table *sgt,
|
||||
void *cpu_addr, dma_addr_t dma_addr, size_t size,
|
||||
unsigned long attrs);
|
||||
|
|
@ -87,4 +86,5 @@ int dma_direct_mmap(struct device *dev, struct vm_area_struct *vma,
|
|||
void *cpu_addr, dma_addr_t dma_addr, size_t size,
|
||||
unsigned long attrs);
|
||||
int dma_direct_supported(struct device *dev, u64 mask);
|
||||
bool dma_direct_need_sync(struct device *dev, dma_addr_t dma_addr);
|
||||
#endif /* _LINUX_DMA_DIRECT_H */
|
||||
|
|
|
|||
|
|
@ -461,6 +461,7 @@ int dma_set_mask(struct device *dev, u64 mask);
|
|||
int dma_set_coherent_mask(struct device *dev, u64 mask);
|
||||
u64 dma_get_required_mask(struct device *dev);
|
||||
size_t dma_max_mapping_size(struct device *dev);
|
||||
bool dma_need_sync(struct device *dev, dma_addr_t dma_addr);
|
||||
unsigned long dma_get_merge_boundary(struct device *dev);
|
||||
#else /* CONFIG_HAS_DMA */
|
||||
static inline dma_addr_t dma_map_page_attrs(struct device *dev,
|
||||
|
|
@ -571,6 +572,10 @@ static inline size_t dma_max_mapping_size(struct device *dev)
|
|||
{
|
||||
return 0;
|
||||
}
|
||||
static inline bool dma_need_sync(struct device *dev, dma_addr_t dma_addr)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
static inline unsigned long dma_get_merge_boundary(struct device *dev)
|
||||
{
|
||||
return 0;
|
||||
|
|
|
|||
|
|
@ -350,6 +350,7 @@ void efi_native_runtime_setup(void);
|
|||
* associated with ConOut
|
||||
*/
|
||||
#define LINUX_EFI_ARM_SCREEN_INFO_TABLE_GUID EFI_GUID(0xe03fc20a, 0x85dc, 0x406e, 0xb9, 0x0e, 0x4a, 0xb5, 0x02, 0x37, 0x1d, 0x95)
|
||||
#define LINUX_EFI_ARM_CPU_STATE_TABLE_GUID EFI_GUID(0xef79e4aa, 0x3c3d, 0x4989, 0xb9, 0x02, 0x07, 0xa9, 0x43, 0xe5, 0x50, 0xd2)
|
||||
#define LINUX_EFI_LOADER_ENTRY_GUID EFI_GUID(0x4a67b082, 0x0a4c, 0x41cf, 0xb6, 0xc7, 0x44, 0x0b, 0x29, 0xbb, 0x8c, 0x4f)
|
||||
#define LINUX_EFI_RANDOM_SEED_TABLE_GUID EFI_GUID(0x1ce1e5bc, 0x7ceb, 0x42f2, 0x81, 0xe5, 0x8a, 0xad, 0xf1, 0x80, 0xf5, 0x7b)
|
||||
#define LINUX_EFI_TPM_EVENT_LOG_GUID EFI_GUID(0xb7799cb0, 0xeca2, 0x4943, 0x96, 0x67, 0x1f, 0xae, 0x07, 0xb7, 0x47, 0xfa)
|
||||
|
|
@ -1236,14 +1237,11 @@ struct linux_efi_memreserve {
|
|||
struct {
|
||||
phys_addr_t base;
|
||||
phys_addr_t size;
|
||||
} entry[0];
|
||||
} entry[];
|
||||
};
|
||||
|
||||
#define EFI_MEMRESERVE_SIZE(count) (sizeof(struct linux_efi_memreserve) + \
|
||||
(count) * sizeof(((struct linux_efi_memreserve *)0)->entry[0]))
|
||||
|
||||
#define EFI_MEMRESERVE_COUNT(size) (((size) - sizeof(struct linux_efi_memreserve)) \
|
||||
/ sizeof(((struct linux_efi_memreserve *)0)->entry[0]))
|
||||
/ sizeof_field(struct linux_efi_memreserve, entry[0]))
|
||||
|
||||
void __init efi_arch_mem_reserve(phys_addr_t addr, u64 size);
|
||||
|
||||
|
|
|
|||
|
|
@ -884,12 +884,12 @@ void bpf_jit_compile(struct bpf_prog *prog);
|
|||
bool bpf_jit_needs_zext(void);
|
||||
bool bpf_helper_changes_pkt_data(void *func);
|
||||
|
||||
static inline bool bpf_dump_raw_ok(void)
|
||||
static inline bool bpf_dump_raw_ok(const struct cred *cred)
|
||||
{
|
||||
/* Reconstruction of call-sites is dependent on kallsyms,
|
||||
* thus make dump the same restriction.
|
||||
*/
|
||||
return kallsyms_show_value() == 1;
|
||||
return kallsyms_show_value(cred);
|
||||
}
|
||||
|
||||
struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off,
|
||||
|
|
|
|||
|
|
@ -315,6 +315,7 @@ enum rw_hint {
|
|||
#define IOCB_SYNC (1 << 5)
|
||||
#define IOCB_WRITE (1 << 6)
|
||||
#define IOCB_NOWAIT (1 << 7)
|
||||
#define IOCB_NOIO (1 << 9)
|
||||
|
||||
struct kiocb {
|
||||
struct file *ki_filp;
|
||||
|
|
@ -1917,7 +1918,6 @@ ssize_t rw_copy_check_uvector(int type, const struct iovec __user * uvector,
|
|||
struct iovec *fast_pointer,
|
||||
struct iovec **ret_pointer);
|
||||
|
||||
extern ssize_t __vfs_read(struct file *, char __user *, size_t, loff_t *);
|
||||
extern ssize_t vfs_read(struct file *, char __user *, size_t, loff_t *);
|
||||
extern ssize_t vfs_write(struct file *, const char __user *, size_t, loff_t *);
|
||||
extern ssize_t vfs_readv(struct file *, const struct iovec __user *,
|
||||
|
|
@ -3033,6 +3033,7 @@ extern int kernel_read_file_from_path_initns(const char *, void **, loff_t *, lo
|
|||
extern int kernel_read_file_from_fd(int, void **, loff_t *, loff_t,
|
||||
enum kernel_read_file_id);
|
||||
extern ssize_t kernel_read(struct file *, void *, size_t, loff_t *);
|
||||
ssize_t __kernel_read(struct file *file, void *buf, size_t count, loff_t *pos);
|
||||
extern ssize_t kernel_write(struct file *, const void *, size_t, loff_t *);
|
||||
extern ssize_t __kernel_write(struct file *, const void *, size_t, loff_t *);
|
||||
extern struct file * open_exec(const char *);
|
||||
|
|
|
|||
|
|
@ -109,6 +109,7 @@ struct fs_context {
|
|||
enum fs_context_phase phase:8; /* The phase the context is in */
|
||||
bool need_free:1; /* Need to call ops->free() */
|
||||
bool global:1; /* Goes into &init_user_ns */
|
||||
bool oldapi:1; /* Coming from mount(2) */
|
||||
};
|
||||
|
||||
struct fs_context_operations {
|
||||
|
|
|
|||
|
|
@ -48,6 +48,9 @@ struct host1x_client_ops {
|
|||
* @channel: host1x channel associated with this client
|
||||
* @syncpts: array of syncpoints requested for this client
|
||||
* @num_syncpts: number of syncpoints requested for this client
|
||||
* @parent: pointer to parent structure
|
||||
* @usecount: reference count for this structure
|
||||
* @lock: mutex for mutually exclusive concurrency
|
||||
*/
|
||||
struct host1x_client {
|
||||
struct list_head list;
|
||||
|
|
|
|||
|
|
@ -3333,13 +3333,17 @@ struct ieee80211_multiple_bssid_configuration {
|
|||
#define WLAN_AKM_SUITE_TDLS SUITE(0x000FAC, 7)
|
||||
#define WLAN_AKM_SUITE_SAE SUITE(0x000FAC, 8)
|
||||
#define WLAN_AKM_SUITE_FT_OVER_SAE SUITE(0x000FAC, 9)
|
||||
#define WLAN_AKM_SUITE_AP_PEER_KEY SUITE(0x000FAC, 10)
|
||||
#define WLAN_AKM_SUITE_8021X_SUITE_B SUITE(0x000FAC, 11)
|
||||
#define WLAN_AKM_SUITE_8021X_SUITE_B_192 SUITE(0x000FAC, 12)
|
||||
#define WLAN_AKM_SUITE_FT_8021X_SHA384 SUITE(0x000FAC, 13)
|
||||
#define WLAN_AKM_SUITE_FILS_SHA256 SUITE(0x000FAC, 14)
|
||||
#define WLAN_AKM_SUITE_FILS_SHA384 SUITE(0x000FAC, 15)
|
||||
#define WLAN_AKM_SUITE_FT_FILS_SHA256 SUITE(0x000FAC, 16)
|
||||
#define WLAN_AKM_SUITE_FT_FILS_SHA384 SUITE(0x000FAC, 17)
|
||||
#define WLAN_AKM_SUITE_OWE SUITE(0x000FAC, 18)
|
||||
#define WLAN_AKM_SUITE_FT_PSK_SHA384 SUITE(0x000FAC, 19)
|
||||
#define WLAN_AKM_SUITE_PSK_SHA384 SUITE(0x000FAC, 20)
|
||||
|
||||
#define WLAN_MAX_KEY_LEN 32
|
||||
|
||||
|
|
|
|||
|
|
@ -25,6 +25,8 @@
|
|||
#define VLAN_ETH_DATA_LEN 1500 /* Max. octets in payload */
|
||||
#define VLAN_ETH_FRAME_LEN 1518 /* Max. octets in frame sans FCS */
|
||||
|
||||
#define VLAN_MAX_DEPTH 8 /* Max. number of nested VLAN tags parsed */
|
||||
|
||||
/*
|
||||
* struct vlan_hdr - vlan header
|
||||
* @h_vlan_TCI: priority and VLAN ID
|
||||
|
|
@ -577,10 +579,10 @@ static inline int vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci)
|
|||
* Returns the EtherType of the packet, regardless of whether it is
|
||||
* vlan encapsulated (normal or hardware accelerated) or not.
|
||||
*/
|
||||
static inline __be16 __vlan_get_protocol(struct sk_buff *skb, __be16 type,
|
||||
static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type,
|
||||
int *depth)
|
||||
{
|
||||
unsigned int vlan_depth = skb->mac_len;
|
||||
unsigned int vlan_depth = skb->mac_len, parse_depth = VLAN_MAX_DEPTH;
|
||||
|
||||
/* if type is 802.1Q/AD then the header should already be
|
||||
* present at mac_len - VLAN_HLEN (if mac_len > 0), or at
|
||||
|
|
@ -595,13 +597,12 @@ static inline __be16 __vlan_get_protocol(struct sk_buff *skb, __be16 type,
|
|||
vlan_depth = ETH_HLEN;
|
||||
}
|
||||
do {
|
||||
struct vlan_hdr *vh;
|
||||
struct vlan_hdr vhdr, *vh;
|
||||
|
||||
if (unlikely(!pskb_may_pull(skb,
|
||||
vlan_depth + VLAN_HLEN)))
|
||||
vh = skb_header_pointer(skb, vlan_depth, sizeof(vhdr), &vhdr);
|
||||
if (unlikely(!vh || !--parse_depth))
|
||||
return 0;
|
||||
|
||||
vh = (struct vlan_hdr *)(skb->data + vlan_depth);
|
||||
type = vh->h_vlan_encapsulated_proto;
|
||||
vlan_depth += VLAN_HLEN;
|
||||
} while (eth_type_vlan(type));
|
||||
|
|
@ -620,11 +621,25 @@ static inline __be16 __vlan_get_protocol(struct sk_buff *skb, __be16 type,
|
|||
* Returns the EtherType of the packet, regardless of whether it is
|
||||
* vlan encapsulated (normal or hardware accelerated) or not.
|
||||
*/
|
||||
static inline __be16 vlan_get_protocol(struct sk_buff *skb)
|
||||
static inline __be16 vlan_get_protocol(const struct sk_buff *skb)
|
||||
{
|
||||
return __vlan_get_protocol(skb, skb->protocol, NULL);
|
||||
}
|
||||
|
||||
/* A getter for the SKB protocol field which will handle VLAN tags consistently
|
||||
* whether VLAN acceleration is enabled or not.
|
||||
*/
|
||||
static inline __be16 skb_protocol(const struct sk_buff *skb, bool skip_vlan)
|
||||
{
|
||||
if (!skip_vlan)
|
||||
/* VLAN acceleration strips the VLAN header from the skb and
|
||||
* moves it to skb->vlan_proto
|
||||
*/
|
||||
return skb_vlan_tag_present(skb) ? skb->vlan_proto : skb->protocol;
|
||||
|
||||
return vlan_get_protocol(skb);
|
||||
}
|
||||
|
||||
static inline void vlan_set_encap_proto(struct sk_buff *skb,
|
||||
struct vlan_hdr *vhdr)
|
||||
{
|
||||
|
|
|
|||
|
|
@ -67,8 +67,15 @@ static const struct acpi_device_id elan_acpi_id[] = {
|
|||
{ "ELAN062B", 0 },
|
||||
{ "ELAN062C", 0 },
|
||||
{ "ELAN062D", 0 },
|
||||
{ "ELAN062E", 0 }, /* Lenovo V340 Whiskey Lake U */
|
||||
{ "ELAN062F", 0 }, /* Lenovo V340 Comet Lake U */
|
||||
{ "ELAN0631", 0 },
|
||||
{ "ELAN0632", 0 },
|
||||
{ "ELAN0633", 0 }, /* Lenovo S145 */
|
||||
{ "ELAN0634", 0 }, /* Lenovo V340 Ice lake */
|
||||
{ "ELAN0635", 0 }, /* Lenovo V1415-IIL */
|
||||
{ "ELAN0636", 0 }, /* Lenovo V1415-Dali */
|
||||
{ "ELAN0637", 0 }, /* Lenovo V1415-IGLR */
|
||||
{ "ELAN1000", 0 },
|
||||
{ }
|
||||
};
|
||||
|
|
|
|||
|
|
@ -41,6 +41,7 @@
|
|||
#define DMA_PTE_SNP BIT_ULL(11)
|
||||
|
||||
#define DMA_FL_PTE_PRESENT BIT_ULL(0)
|
||||
#define DMA_FL_PTE_US BIT_ULL(2)
|
||||
#define DMA_FL_PTE_XD BIT_ULL(63)
|
||||
|
||||
#define ADDR_WIDTH_5LEVEL (57)
|
||||
|
|
|
|||
|
|
@ -2,7 +2,7 @@
|
|||
#ifndef _LINUX_IRQ_WORK_H
|
||||
#define _LINUX_IRQ_WORK_H
|
||||
|
||||
#include <linux/llist.h>
|
||||
#include <linux/smp_types.h>
|
||||
|
||||
/*
|
||||
* An entry can be in one of four states:
|
||||
|
|
@ -13,24 +13,14 @@
|
|||
* busy NULL, 2 -> {free, claimed} : callback in progress, can be claimed
|
||||
*/
|
||||
|
||||
/* flags share CSD_FLAG_ space */
|
||||
|
||||
#define IRQ_WORK_PENDING BIT(0)
|
||||
#define IRQ_WORK_BUSY BIT(1)
|
||||
|
||||
/* Doesn't want IPI, wait for tick: */
|
||||
#define IRQ_WORK_LAZY BIT(2)
|
||||
/* Run hard IRQ context, even on RT */
|
||||
#define IRQ_WORK_HARD_IRQ BIT(3)
|
||||
|
||||
#define IRQ_WORK_CLAIMED (IRQ_WORK_PENDING | IRQ_WORK_BUSY)
|
||||
|
||||
/*
|
||||
* structure shares layout with single_call_data_t.
|
||||
*/
|
||||
struct irq_work {
|
||||
struct llist_node llnode;
|
||||
atomic_t flags;
|
||||
union {
|
||||
struct __call_single_node node;
|
||||
struct {
|
||||
struct llist_node llnode;
|
||||
atomic_t flags;
|
||||
};
|
||||
};
|
||||
void (*func)(struct irq_work *);
|
||||
};
|
||||
|
||||
|
|
|
|||
|
|
@ -18,6 +18,7 @@
|
|||
#define KSYM_SYMBOL_LEN (sizeof("%s+%#lx/%#lx [%s]") + (KSYM_NAME_LEN - 1) + \
|
||||
2*(BITS_PER_LONG*3/10) + (MODULE_NAME_LEN - 1) + 1)
|
||||
|
||||
struct cred;
|
||||
struct module;
|
||||
|
||||
static inline int is_kernel_inittext(unsigned long addr)
|
||||
|
|
@ -98,7 +99,7 @@ int lookup_symbol_name(unsigned long addr, char *symname);
|
|||
int lookup_symbol_attrs(unsigned long addr, unsigned long *size, unsigned long *offset, char *modname, char *name);
|
||||
|
||||
/* How and when do we show kallsyms values? */
|
||||
extern int kallsyms_show_value(void);
|
||||
extern bool kallsyms_show_value(const struct cred *cred);
|
||||
|
||||
#else /* !CONFIG_KALLSYMS */
|
||||
|
||||
|
|
@ -158,7 +159,7 @@ static inline int lookup_symbol_attrs(unsigned long addr, unsigned long *size, u
|
|||
return -ERANGE;
|
||||
}
|
||||
|
||||
static inline int kallsyms_show_value(void)
|
||||
static inline bool kallsyms_show_value(const struct cred *cred)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -176,6 +176,17 @@ kgdb_arch_handle_exception(int vector, int signo, int err_code,
|
|||
char *remcom_out_buffer,
|
||||
struct pt_regs *regs);
|
||||
|
||||
/**
|
||||
* kgdb_arch_handle_qxfer_pkt - Handle architecture specific GDB XML
|
||||
* packets.
|
||||
* @remcom_in_buffer: The buffer of the packet we have read.
|
||||
* @remcom_out_buffer: The buffer of %BUFMAX bytes to write a packet into.
|
||||
*/
|
||||
|
||||
extern void
|
||||
kgdb_arch_handle_qxfer_pkt(char *remcom_in_buffer,
|
||||
char *remcom_out_buffer);
|
||||
|
||||
/**
|
||||
* kgdb_call_nmi_hook - Call kgdb_nmicallback() on the current CPU
|
||||
* @ignored: This parameter is only here to match the prototype.
|
||||
|
|
@ -276,8 +287,7 @@ struct kgdb_arch {
|
|||
* the I/O driver.
|
||||
* @post_exception: Pointer to a function that will do any cleanup work
|
||||
* for the I/O driver.
|
||||
* @is_console: 1 if the end device is a console 0 if the I/O device is
|
||||
* not a console
|
||||
* @cons: valid if the I/O device is a console; else NULL.
|
||||
*/
|
||||
struct kgdb_io {
|
||||
const char *name;
|
||||
|
|
@ -288,7 +298,7 @@ struct kgdb_io {
|
|||
void (*deinit) (void);
|
||||
void (*pre_exception) (void);
|
||||
void (*post_exception) (void);
|
||||
int is_console;
|
||||
struct console *cons;
|
||||
};
|
||||
|
||||
extern const struct kgdb_arch arch_kgdb_ops;
|
||||
|
|
@ -315,6 +325,7 @@ extern int kgdb_hex2mem(char *buf, char *mem, int count);
|
|||
|
||||
extern int kgdb_isremovedbreak(unsigned long addr);
|
||||
extern void kgdb_schedule_breakpoint(void);
|
||||
extern int kgdb_has_hit_break(unsigned long addr);
|
||||
|
||||
extern int
|
||||
kgdb_handle_exception(int ex_vector, int signo, int err_code,
|
||||
|
|
|
|||
|
|
@ -1095,7 +1095,7 @@ extern int ata_scsi_ioctl(struct scsi_device *dev, unsigned int cmd,
|
|||
#define ATA_SCSI_COMPAT_IOCTL /* empty */
|
||||
#endif
|
||||
extern int ata_scsi_queuecmd(struct Scsi_Host *h, struct scsi_cmnd *cmd);
|
||||
#if IS_ENABLED(CONFIG_ATA)
|
||||
#if IS_REACHABLE(CONFIG_ATA)
|
||||
bool ata_scsi_dma_need_drain(struct request *rq);
|
||||
#else
|
||||
#define ata_scsi_dma_need_drain NULL
|
||||
|
|
|
|||
|
|
@ -150,7 +150,7 @@ LSM_HOOK(int, 0, inode_listsecurity, struct inode *inode, char *buffer,
|
|||
size_t buffer_size)
|
||||
LSM_HOOK(void, LSM_RET_VOID, inode_getsecid, struct inode *inode, u32 *secid)
|
||||
LSM_HOOK(int, 0, inode_copy_up, struct dentry *src, struct cred **new)
|
||||
LSM_HOOK(int, 0, inode_copy_up_xattr, const char *name)
|
||||
LSM_HOOK(int, -EOPNOTSUPP, inode_copy_up_xattr, const char *name)
|
||||
LSM_HOOK(int, 0, kernfs_init_security, struct kernfs_node *kn_dir,
|
||||
struct kernfs_node *kn)
|
||||
LSM_HOOK(int, 0, file_permission, struct file *file, int mask)
|
||||
|
|
@ -360,7 +360,7 @@ LSM_HOOK(int, 0, key_alloc, struct key *key, const struct cred *cred,
|
|||
unsigned long flags)
|
||||
LSM_HOOK(void, LSM_RET_VOID, key_free, struct key *key)
|
||||
LSM_HOOK(int, 0, key_permission, key_ref_t key_ref, const struct cred *cred,
|
||||
unsigned perm)
|
||||
enum key_need_perm need_perm)
|
||||
LSM_HOOK(int, 0, key_getsecurity, struct key *key, char **_buffer)
|
||||
#endif /* CONFIG_KEYS */
|
||||
|
||||
|
|
|
|||
|
|
@ -147,6 +147,7 @@ enum {
|
|||
MLX5_REG_MCDA = 0x9063,
|
||||
MLX5_REG_MCAM = 0x907f,
|
||||
MLX5_REG_MIRC = 0x9162,
|
||||
MLX5_REG_SBCAM = 0xB01F,
|
||||
MLX5_REG_RESOURCE_DUMP = 0xC000,
|
||||
};
|
||||
|
||||
|
|
|
|||
|
|
@ -4283,7 +4283,8 @@ struct mlx5_ifc_rst2init_qp_out_bits {
|
|||
|
||||
u8 syndrome[0x20];
|
||||
|
||||
u8 reserved_at_40[0x40];
|
||||
u8 reserved_at_40[0x20];
|
||||
u8 ece[0x20];
|
||||
};
|
||||
|
||||
struct mlx5_ifc_rst2init_qp_in_bits {
|
||||
|
|
@ -4300,7 +4301,7 @@ struct mlx5_ifc_rst2init_qp_in_bits {
|
|||
|
||||
u8 opt_param_mask[0x20];
|
||||
|
||||
u8 reserved_at_a0[0x20];
|
||||
u8 ece[0x20];
|
||||
|
||||
struct mlx5_ifc_qpc_bits qpc;
|
||||
|
||||
|
|
@ -6619,7 +6620,8 @@ struct mlx5_ifc_init2init_qp_out_bits {
|
|||
|
||||
u8 syndrome[0x20];
|
||||
|
||||
u8 reserved_at_40[0x40];
|
||||
u8 reserved_at_40[0x20];
|
||||
u8 ece[0x20];
|
||||
};
|
||||
|
||||
struct mlx5_ifc_init2init_qp_in_bits {
|
||||
|
|
@ -6636,7 +6638,7 @@ struct mlx5_ifc_init2init_qp_in_bits {
|
|||
|
||||
u8 opt_param_mask[0x20];
|
||||
|
||||
u8 reserved_at_a0[0x20];
|
||||
u8 ece[0x20];
|
||||
|
||||
struct mlx5_ifc_qpc_bits qpc;
|
||||
|
||||
|
|
@ -9958,6 +9960,34 @@ struct mlx5_ifc_pptb_reg_bits {
|
|||
u8 untagged_buff[0x4];
|
||||
};
|
||||
|
||||
struct mlx5_ifc_sbcam_reg_bits {
|
||||
u8 reserved_at_0[0x8];
|
||||
u8 feature_group[0x8];
|
||||
u8 reserved_at_10[0x8];
|
||||
u8 access_reg_group[0x8];
|
||||
|
||||
u8 reserved_at_20[0x20];
|
||||
|
||||
u8 sb_access_reg_cap_mask[4][0x20];
|
||||
|
||||
u8 reserved_at_c0[0x80];
|
||||
|
||||
u8 sb_feature_cap_mask[4][0x20];
|
||||
|
||||
u8 reserved_at_1c0[0x40];
|
||||
|
||||
u8 cap_total_buffer_size[0x20];
|
||||
|
||||
u8 cap_cell_size[0x10];
|
||||
u8 cap_max_pg_buffers[0x8];
|
||||
u8 cap_num_pool_supported[0x8];
|
||||
|
||||
u8 reserved_at_240[0x8];
|
||||
u8 cap_sbsr_stat_size[0x8];
|
||||
u8 cap_max_tclass_data[0x8];
|
||||
u8 cap_max_cpu_ingress_tclass_sb[0x8];
|
||||
};
|
||||
|
||||
struct mlx5_ifc_pbmc_reg_bits {
|
||||
u8 reserved_at_0[0x8];
|
||||
u8 local_port[0x8];
|
||||
|
|
|
|||
|
|
@ -257,8 +257,8 @@ struct lruvec {
|
|||
*/
|
||||
unsigned long anon_cost;
|
||||
unsigned long file_cost;
|
||||
/* Evictions & activations on the inactive file list */
|
||||
atomic_long_t inactive_age;
|
||||
/* Non-resident age, driven by LRU movement */
|
||||
atomic_long_t nonresident_age;
|
||||
/* Refaults at the time of last reclaim cycle */
|
||||
unsigned long refaults;
|
||||
/* Various lruvec state flags (enum lruvec_flags) */
|
||||
|
|
|
|||
|
|
@ -318,7 +318,7 @@ struct pcmcia_device_id {
|
|||
#define INPUT_DEVICE_ID_LED_MAX 0x0f
|
||||
#define INPUT_DEVICE_ID_SND_MAX 0x07
|
||||
#define INPUT_DEVICE_ID_FF_MAX 0x7f
|
||||
#define INPUT_DEVICE_ID_SW_MAX 0x0f
|
||||
#define INPUT_DEVICE_ID_SW_MAX 0x10
|
||||
#define INPUT_DEVICE_ID_PROP_MAX 0x1f
|
||||
|
||||
#define INPUT_DEVICE_ID_MATCH_BUS 1
|
||||
|
|
|
|||
|
|
@ -3157,7 +3157,7 @@ static inline int dev_recursion_level(void)
|
|||
return this_cpu_read(softnet_data.xmit.recursion);
|
||||
}
|
||||
|
||||
#define XMIT_RECURSION_LIMIT 10
|
||||
#define XMIT_RECURSION_LIMIT 8
|
||||
static inline bool dev_xmit_recursion(void)
|
||||
{
|
||||
return unlikely(__this_cpu_read(softnet_data.xmit.recursion) >
|
||||
|
|
|
|||
|
|
@ -25,6 +25,12 @@
|
|||
int ipt_register_table(struct net *net, const struct xt_table *table,
|
||||
const struct ipt_replace *repl,
|
||||
const struct nf_hook_ops *ops, struct xt_table **res);
|
||||
|
||||
void ipt_unregister_table_pre_exit(struct net *net, struct xt_table *table,
|
||||
const struct nf_hook_ops *ops);
|
||||
|
||||
void ipt_unregister_table_exit(struct net *net, struct xt_table *table);
|
||||
|
||||
void ipt_unregister_table(struct net *net, struct xt_table *table,
|
||||
const struct nf_hook_ops *ops);
|
||||
|
||||
|
|
|
|||
|
|
@ -29,6 +29,9 @@ int ip6t_register_table(struct net *net, const struct xt_table *table,
|
|||
const struct nf_hook_ops *ops, struct xt_table **res);
|
||||
void ip6t_unregister_table(struct net *net, struct xt_table *table,
|
||||
const struct nf_hook_ops *ops);
|
||||
void ip6t_unregister_table_pre_exit(struct net *net, struct xt_table *table,
|
||||
const struct nf_hook_ops *ops);
|
||||
void ip6t_unregister_table_exit(struct net *net, struct xt_table *table);
|
||||
extern unsigned int ip6t_do_table(struct sk_buff *skb,
|
||||
const struct nf_hook_state *state,
|
||||
struct xt_table *table);
|
||||
|
|
|
|||
|
|
@ -2169,12 +2169,11 @@ static inline int pci_pcie_type(const struct pci_dev *dev)
|
|||
*/
|
||||
static inline struct pci_dev *pcie_find_root_port(struct pci_dev *dev)
|
||||
{
|
||||
struct pci_dev *bridge = pci_upstream_bridge(dev);
|
||||
|
||||
while (bridge) {
|
||||
if (pci_pcie_type(bridge) == PCI_EXP_TYPE_ROOT_PORT)
|
||||
return bridge;
|
||||
bridge = pci_upstream_bridge(bridge);
|
||||
while (dev) {
|
||||
if (pci_is_pcie(dev) &&
|
||||
pci_pcie_type(dev) == PCI_EXP_TYPE_ROOT_PORT)
|
||||
return dev;
|
||||
dev = pci_upstream_bridge(dev);
|
||||
}
|
||||
|
||||
return NULL;
|
||||
|
|
|
|||
|
|
@ -1416,6 +1416,7 @@ int phy_ethtool_ksettings_set(struct phy_device *phydev,
|
|||
int phy_mii_ioctl(struct phy_device *phydev, struct ifreq *ifr, int cmd);
|
||||
int phy_do_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd);
|
||||
int phy_do_ioctl_running(struct net_device *dev, struct ifreq *ifr, int cmd);
|
||||
int phy_disable_interrupts(struct phy_device *phydev);
|
||||
void phy_request_interrupt(struct phy_device *phydev);
|
||||
void phy_free_interrupt(struct phy_device *phydev);
|
||||
void phy_print_status(struct phy_device *phydev);
|
||||
|
|
|
|||
|
|
@ -207,28 +207,34 @@ static inline u32 qed_chain_get_cons_idx_u32(struct qed_chain *p_chain)
|
|||
|
||||
static inline u16 qed_chain_get_elem_left(struct qed_chain *p_chain)
|
||||
{
|
||||
u16 elem_per_page = p_chain->elem_per_page;
|
||||
u32 prod = p_chain->u.chain16.prod_idx;
|
||||
u32 cons = p_chain->u.chain16.cons_idx;
|
||||
u16 used;
|
||||
|
||||
used = (u16) (((u32)0x10000 +
|
||||
(u32)p_chain->u.chain16.prod_idx) -
|
||||
(u32)p_chain->u.chain16.cons_idx);
|
||||
if (prod < cons)
|
||||
prod += (u32)U16_MAX + 1;
|
||||
|
||||
used = (u16)(prod - cons);
|
||||
if (p_chain->mode == QED_CHAIN_MODE_NEXT_PTR)
|
||||
used -= p_chain->u.chain16.prod_idx / p_chain->elem_per_page -
|
||||
p_chain->u.chain16.cons_idx / p_chain->elem_per_page;
|
||||
used -= prod / elem_per_page - cons / elem_per_page;
|
||||
|
||||
return (u16)(p_chain->capacity - used);
|
||||
}
|
||||
|
||||
static inline u32 qed_chain_get_elem_left_u32(struct qed_chain *p_chain)
|
||||
{
|
||||
u16 elem_per_page = p_chain->elem_per_page;
|
||||
u64 prod = p_chain->u.chain32.prod_idx;
|
||||
u64 cons = p_chain->u.chain32.cons_idx;
|
||||
u32 used;
|
||||
|
||||
used = (u32) (((u64)0x100000000ULL +
|
||||
(u64)p_chain->u.chain32.prod_idx) -
|
||||
(u64)p_chain->u.chain32.cons_idx);
|
||||
if (prod < cons)
|
||||
prod += (u64)U32_MAX + 1;
|
||||
|
||||
used = (u32)(prod - cons);
|
||||
if (p_chain->mode == QED_CHAIN_MODE_NEXT_PTR)
|
||||
used -= p_chain->u.chain32.prod_idx / p_chain->elem_per_page -
|
||||
p_chain->u.chain32.cons_idx / p_chain->elem_per_page;
|
||||
used -= (u32)(prod / elem_per_page - cons / elem_per_page);
|
||||
|
||||
return p_chain->capacity - used;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -155,7 +155,7 @@ static inline void sg_set_buf(struct scatterlist *sg, const void *buf,
|
|||
* Loop over each sg element in the given sg_table object.
|
||||
*/
|
||||
#define for_each_sgtable_sg(sgt, sg, i) \
|
||||
for_each_sg(sgt->sgl, sg, sgt->orig_nents, i)
|
||||
for_each_sg((sgt)->sgl, sg, (sgt)->orig_nents, i)
|
||||
|
||||
/*
|
||||
* Loop over each sg element in the given *DMA mapped* sg_table object.
|
||||
|
|
@ -163,7 +163,7 @@ static inline void sg_set_buf(struct scatterlist *sg, const void *buf,
|
|||
* of the each element.
|
||||
*/
|
||||
#define for_each_sgtable_dma_sg(sgt, sg, i) \
|
||||
for_each_sg(sgt->sgl, sg, sgt->nents, i)
|
||||
for_each_sg((sgt)->sgl, sg, (sgt)->nents, i)
|
||||
|
||||
/**
|
||||
* sg_chain - Chain two sglists together
|
||||
|
|
@ -451,7 +451,7 @@ sg_page_iter_dma_address(struct sg_dma_page_iter *dma_iter)
|
|||
* See also for_each_sg_page(). In each loop it operates on PAGE_SIZE unit.
|
||||
*/
|
||||
#define for_each_sgtable_page(sgt, piter, pgoffset) \
|
||||
for_each_sg_page(sgt->sgl, piter, sgt->orig_nents, pgoffset)
|
||||
for_each_sg_page((sgt)->sgl, piter, (sgt)->orig_nents, pgoffset)
|
||||
|
||||
/**
|
||||
* for_each_sgtable_dma_page - iterate over the DMA mapped sg_table object
|
||||
|
|
@ -465,7 +465,7 @@ sg_page_iter_dma_address(struct sg_dma_page_iter *dma_iter)
|
|||
* unit.
|
||||
*/
|
||||
#define for_each_sgtable_dma_page(sgt, dma_iter, pgoffset) \
|
||||
for_each_sg_dma_page(sgt->sgl, dma_iter, sgt->nents, pgoffset)
|
||||
for_each_sg_dma_page((sgt)->sgl, dma_iter, (sgt)->nents, pgoffset)
|
||||
|
||||
|
||||
/*
|
||||
|
|
|
|||
|
|
@ -114,10 +114,6 @@ struct task_group;
|
|||
|
||||
#define task_is_stopped_or_traced(task) ((task->state & (__TASK_STOPPED | __TASK_TRACED)) != 0)
|
||||
|
||||
#define task_contributes_to_load(task) ((task->state & TASK_UNINTERRUPTIBLE) != 0 && \
|
||||
(task->flags & PF_FROZEN) == 0 && \
|
||||
(task->state & TASK_NOLOAD) == 0)
|
||||
|
||||
#ifdef CONFIG_DEBUG_ATOMIC_SLEEP
|
||||
|
||||
/*
|
||||
|
|
@ -654,9 +650,8 @@ struct task_struct {
|
|||
unsigned int ptrace;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
struct llist_node wake_entry;
|
||||
unsigned int wake_entry_type;
|
||||
int on_cpu;
|
||||
struct __call_single_node wake_entry;
|
||||
#ifdef CONFIG_THREAD_INFO_IN_TASK
|
||||
/* Current CPU: */
|
||||
unsigned int cpu;
|
||||
|
|
|
|||
|
|
@ -19,6 +19,7 @@ struct task_struct;
|
|||
#define JOBCTL_TRAPPING_BIT 21 /* switching to TRACED */
|
||||
#define JOBCTL_LISTENING_BIT 22 /* ptracer is listening for events */
|
||||
#define JOBCTL_TRAP_FREEZE_BIT 23 /* trap for cgroup freezer */
|
||||
#define JOBCTL_TASK_WORK_BIT 24 /* set by TWA_SIGNAL */
|
||||
|
||||
#define JOBCTL_STOP_DEQUEUED (1UL << JOBCTL_STOP_DEQUEUED_BIT)
|
||||
#define JOBCTL_STOP_PENDING (1UL << JOBCTL_STOP_PENDING_BIT)
|
||||
|
|
@ -28,9 +29,10 @@ struct task_struct;
|
|||
#define JOBCTL_TRAPPING (1UL << JOBCTL_TRAPPING_BIT)
|
||||
#define JOBCTL_LISTENING (1UL << JOBCTL_LISTENING_BIT)
|
||||
#define JOBCTL_TRAP_FREEZE (1UL << JOBCTL_TRAP_FREEZE_BIT)
|
||||
#define JOBCTL_TASK_WORK (1UL << JOBCTL_TASK_WORK_BIT)
|
||||
|
||||
#define JOBCTL_TRAP_MASK (JOBCTL_TRAP_STOP | JOBCTL_TRAP_NOTIFY)
|
||||
#define JOBCTL_PENDING_MASK (JOBCTL_STOP_PENDING | JOBCTL_TRAP_MASK)
|
||||
#define JOBCTL_PENDING_MASK (JOBCTL_STOP_PENDING | JOBCTL_TRAP_MASK | JOBCTL_TASK_WORK)
|
||||
|
||||
extern bool task_set_jobctl_pending(struct task_struct *task, unsigned long mask);
|
||||
extern void task_clear_jobctl_trapping(struct task_struct *task);
|
||||
|
|
|
|||
|
|
@ -462,10 +462,104 @@ extern void uart_handle_cts_change(struct uart_port *uport,
|
|||
extern void uart_insert_char(struct uart_port *port, unsigned int status,
|
||||
unsigned int overrun, unsigned int ch, unsigned int flag);
|
||||
|
||||
extern int uart_handle_sysrq_char(struct uart_port *port, unsigned int ch);
|
||||
extern int uart_prepare_sysrq_char(struct uart_port *port, unsigned int ch);
|
||||
extern void uart_unlock_and_check_sysrq(struct uart_port *port, unsigned long flags);
|
||||
extern int uart_handle_break(struct uart_port *port);
|
||||
#ifdef CONFIG_MAGIC_SYSRQ_SERIAL
|
||||
#define SYSRQ_TIMEOUT (HZ * 5)
|
||||
|
||||
bool uart_try_toggle_sysrq(struct uart_port *port, unsigned int ch);
|
||||
|
||||
static inline int uart_handle_sysrq_char(struct uart_port *port, unsigned int ch)
|
||||
{
|
||||
if (!port->sysrq)
|
||||
return 0;
|
||||
|
||||
if (ch && time_before(jiffies, port->sysrq)) {
|
||||
if (sysrq_mask()) {
|
||||
handle_sysrq(ch);
|
||||
port->sysrq = 0;
|
||||
return 1;
|
||||
}
|
||||
if (uart_try_toggle_sysrq(port, ch))
|
||||
return 1;
|
||||
}
|
||||
port->sysrq = 0;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline int uart_prepare_sysrq_char(struct uart_port *port, unsigned int ch)
|
||||
{
|
||||
if (!port->sysrq)
|
||||
return 0;
|
||||
|
||||
if (ch && time_before(jiffies, port->sysrq)) {
|
||||
if (sysrq_mask()) {
|
||||
port->sysrq_ch = ch;
|
||||
port->sysrq = 0;
|
||||
return 1;
|
||||
}
|
||||
if (uart_try_toggle_sysrq(port, ch))
|
||||
return 1;
|
||||
}
|
||||
port->sysrq = 0;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline void uart_unlock_and_check_sysrq(struct uart_port *port, unsigned long irqflags)
|
||||
{
|
||||
int sysrq_ch;
|
||||
|
||||
if (!port->has_sysrq) {
|
||||
spin_unlock_irqrestore(&port->lock, irqflags);
|
||||
return;
|
||||
}
|
||||
|
||||
sysrq_ch = port->sysrq_ch;
|
||||
port->sysrq_ch = 0;
|
||||
|
||||
spin_unlock_irqrestore(&port->lock, irqflags);
|
||||
|
||||
if (sysrq_ch)
|
||||
handle_sysrq(sysrq_ch);
|
||||
}
|
||||
#else /* CONFIG_MAGIC_SYSRQ_SERIAL */
|
||||
static inline int uart_handle_sysrq_char(struct uart_port *port, unsigned int ch)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline int uart_prepare_sysrq_char(struct uart_port *port, unsigned int ch)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline void uart_unlock_and_check_sysrq(struct uart_port *port, unsigned long irqflags)
|
||||
{
|
||||
spin_unlock_irqrestore(&port->lock, irqflags);
|
||||
}
|
||||
#endif /* CONFIG_MAGIC_SYSRQ_SERIAL */
|
||||
|
||||
/*
|
||||
* We do the SysRQ and SAK checking like this...
|
||||
*/
|
||||
static inline int uart_handle_break(struct uart_port *port)
|
||||
{
|
||||
struct uart_state *state = port->state;
|
||||
|
||||
if (port->handle_break)
|
||||
port->handle_break(port);
|
||||
|
||||
#ifdef CONFIG_MAGIC_SYSRQ_SERIAL
|
||||
if (port->has_sysrq && uart_console(port)) {
|
||||
if (!port->sysrq) {
|
||||
port->sysrq = jiffies + SYSRQ_TIMEOUT;
|
||||
return 1;
|
||||
}
|
||||
port->sysrq = 0;
|
||||
}
|
||||
#endif
|
||||
if (port->flags & UPF_SAK)
|
||||
do_SAK(state->port.tty);
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* UART_ENABLE_MS - determine if port should enable modem status irqs
|
||||
|
|
|
|||
|
|
@ -430,6 +430,19 @@ static inline void psock_set_prog(struct bpf_prog **pprog,
|
|||
bpf_prog_put(prog);
|
||||
}
|
||||
|
||||
static inline int psock_replace_prog(struct bpf_prog **pprog,
|
||||
struct bpf_prog *prog,
|
||||
struct bpf_prog *old)
|
||||
{
|
||||
if (cmpxchg(pprog, old, prog) != old)
|
||||
return -ENOENT;
|
||||
|
||||
if (old)
|
||||
bpf_prog_put(old);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline void psock_progs_drop(struct sk_psock_progs *progs)
|
||||
{
|
||||
psock_set_prog(&progs->msg_parser, NULL);
|
||||
|
|
|
|||
|
|
@ -12,29 +12,22 @@
|
|||
#include <linux/list.h>
|
||||
#include <linux/cpumask.h>
|
||||
#include <linux/init.h>
|
||||
#include <linux/llist.h>
|
||||
#include <linux/smp_types.h>
|
||||
|
||||
typedef void (*smp_call_func_t)(void *info);
|
||||
typedef bool (*smp_cond_func_t)(int cpu, void *info);
|
||||
|
||||
enum {
|
||||
CSD_FLAG_LOCK = 0x01,
|
||||
|
||||
/* IRQ_WORK_flags */
|
||||
|
||||
CSD_TYPE_ASYNC = 0x00,
|
||||
CSD_TYPE_SYNC = 0x10,
|
||||
CSD_TYPE_IRQ_WORK = 0x20,
|
||||
CSD_TYPE_TTWU = 0x30,
|
||||
CSD_FLAG_TYPE_MASK = 0xF0,
|
||||
};
|
||||
|
||||
/*
|
||||
* structure shares (partial) layout with struct irq_work
|
||||
*/
|
||||
struct __call_single_data {
|
||||
struct llist_node llist;
|
||||
unsigned int flags;
|
||||
union {
|
||||
struct __call_single_node node;
|
||||
struct {
|
||||
struct llist_node llist;
|
||||
unsigned int flags;
|
||||
};
|
||||
};
|
||||
smp_call_func_t func;
|
||||
void *info;
|
||||
};
|
||||
|
|
|
|||
66
include/linux/smp_types.h
Normal file
66
include/linux/smp_types.h
Normal file
|
|
@ -0,0 +1,66 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0 */
|
||||
#ifndef __LINUX_SMP_TYPES_H
|
||||
#define __LINUX_SMP_TYPES_H
|
||||
|
||||
#include <linux/llist.h>
|
||||
|
||||
enum {
|
||||
CSD_FLAG_LOCK = 0x01,
|
||||
|
||||
IRQ_WORK_PENDING = 0x01,
|
||||
IRQ_WORK_BUSY = 0x02,
|
||||
IRQ_WORK_LAZY = 0x04, /* No IPI, wait for tick */
|
||||
IRQ_WORK_HARD_IRQ = 0x08, /* IRQ context on PREEMPT_RT */
|
||||
|
||||
IRQ_WORK_CLAIMED = (IRQ_WORK_PENDING | IRQ_WORK_BUSY),
|
||||
|
||||
CSD_TYPE_ASYNC = 0x00,
|
||||
CSD_TYPE_SYNC = 0x10,
|
||||
CSD_TYPE_IRQ_WORK = 0x20,
|
||||
CSD_TYPE_TTWU = 0x30,
|
||||
|
||||
CSD_FLAG_TYPE_MASK = 0xF0,
|
||||
};
|
||||
|
||||
/*
|
||||
* struct __call_single_node is the primary type on
|
||||
* smp.c:call_single_queue.
|
||||
*
|
||||
* flush_smp_call_function_queue() only reads the type from
|
||||
* __call_single_node::u_flags as a regular load, the above
|
||||
* (anonymous) enum defines all the bits of this word.
|
||||
*
|
||||
* Other bits are not modified until the type is known.
|
||||
*
|
||||
* CSD_TYPE_SYNC/ASYNC:
|
||||
* struct {
|
||||
* struct llist_node node;
|
||||
* unsigned int flags;
|
||||
* smp_call_func_t func;
|
||||
* void *info;
|
||||
* };
|
||||
*
|
||||
* CSD_TYPE_IRQ_WORK:
|
||||
* struct {
|
||||
* struct llist_node node;
|
||||
* atomic_t flags;
|
||||
* void (*func)(struct irq_work *);
|
||||
* };
|
||||
*
|
||||
* CSD_TYPE_TTWU:
|
||||
* struct {
|
||||
* struct llist_node node;
|
||||
* unsigned int flags;
|
||||
* };
|
||||
*
|
||||
*/
|
||||
|
||||
struct __call_single_node {
|
||||
struct llist_node llist;
|
||||
union {
|
||||
unsigned int u_flags;
|
||||
atomic_t a_flags;
|
||||
};
|
||||
};
|
||||
|
||||
#endif /* __LINUX_SMP_TYPES_H */
|
||||
|
|
@ -313,6 +313,7 @@ struct vma_swap_readahead {
|
|||
};
|
||||
|
||||
/* linux/mm/workingset.c */
|
||||
void workingset_age_nonresident(struct lruvec *lruvec, unsigned long nr_pages);
|
||||
void *workingset_eviction(struct page *page, struct mem_cgroup *target_memcg);
|
||||
void workingset_refault(struct page *page, void *shadow);
|
||||
void workingset_activation(struct page *page);
|
||||
|
|
|
|||
|
|
@ -1360,7 +1360,7 @@ static inline long ksys_lchown(const char __user *filename, uid_t user,
|
|||
|
||||
extern long do_sys_ftruncate(unsigned int fd, loff_t length, int small);
|
||||
|
||||
static inline long ksys_ftruncate(unsigned int fd, unsigned long length)
|
||||
static inline long ksys_ftruncate(unsigned int fd, loff_t length)
|
||||
{
|
||||
return do_sys_ftruncate(fd, length, 1);
|
||||
}
|
||||
|
|
|
|||
|
|
@ -13,7 +13,10 @@ init_task_work(struct callback_head *twork, task_work_func_t func)
|
|||
twork->func = func;
|
||||
}
|
||||
|
||||
int task_work_add(struct task_struct *task, struct callback_head *twork, bool);
|
||||
#define TWA_RESUME 1
|
||||
#define TWA_SIGNAL 2
|
||||
int task_work_add(struct task_struct *task, struct callback_head *twork, int);
|
||||
|
||||
struct callback_head *task_work_cancel(struct task_struct *, task_work_func_t);
|
||||
void task_work_run(void);
|
||||
|
||||
|
|
|
|||
|
|
@ -222,9 +222,9 @@ extern bool timekeeping_rtc_skipresume(void);
|
|||
|
||||
extern void timekeeping_inject_sleeptime64(const struct timespec64 *delta);
|
||||
|
||||
/*
|
||||
/**
|
||||
* struct system_time_snapshot - simultaneous raw/real time capture with
|
||||
* counter value
|
||||
* counter value
|
||||
* @cycles: Clocksource counter value to produce the system times
|
||||
* @real: Realtime system time
|
||||
* @raw: Monotonic raw system time
|
||||
|
|
@ -239,9 +239,9 @@ struct system_time_snapshot {
|
|||
u8 cs_was_changed_seq;
|
||||
};
|
||||
|
||||
/*
|
||||
/**
|
||||
* struct system_device_crosststamp - system/device cross-timestamp
|
||||
* (syncronized capture)
|
||||
* (synchronized capture)
|
||||
* @device: Device time
|
||||
* @sys_realtime: Realtime simultaneous with device time
|
||||
* @sys_monoraw: Monotonic raw simultaneous with device time
|
||||
|
|
@ -252,12 +252,12 @@ struct system_device_crosststamp {
|
|||
ktime_t sys_monoraw;
|
||||
};
|
||||
|
||||
/*
|
||||
/**
|
||||
* struct system_counterval_t - system counter value with the pointer to the
|
||||
* corresponding clocksource
|
||||
* corresponding clocksource
|
||||
* @cycles: System counter value
|
||||
* @cs: Clocksource corresponding to system counter value. Used by
|
||||
* timekeeping code to verify comparibility of two cycle values
|
||||
* timekeeping code to verify comparibility of two cycle values
|
||||
*/
|
||||
struct system_counterval_t {
|
||||
u64 cycles;
|
||||
|
|
|
|||
|
|
@ -81,6 +81,8 @@ struct tcg_efi_specid_event_algs {
|
|||
u16 digest_size;
|
||||
} __packed;
|
||||
|
||||
#define TCG_SPECID_SIG "Spec ID Event03"
|
||||
|
||||
struct tcg_efi_specid_event_head {
|
||||
u8 signature[16];
|
||||
u32 platform_class;
|
||||
|
|
@ -171,6 +173,7 @@ static inline int __calc_tpm2_event_size(struct tcg_pcr_event2_head *event,
|
|||
int i;
|
||||
int j;
|
||||
u32 count, event_type;
|
||||
const u8 zero_digest[sizeof(event_header->digest)] = {0};
|
||||
|
||||
marker = event;
|
||||
marker_start = marker;
|
||||
|
|
@ -198,10 +201,19 @@ static inline int __calc_tpm2_event_size(struct tcg_pcr_event2_head *event,
|
|||
count = READ_ONCE(event->count);
|
||||
event_type = READ_ONCE(event->event_type);
|
||||
|
||||
/* Verify that it's the log header */
|
||||
if (event_header->pcr_idx != 0 ||
|
||||
event_header->event_type != NO_ACTION ||
|
||||
memcmp(event_header->digest, zero_digest, sizeof(zero_digest))) {
|
||||
size = 0;
|
||||
goto out;
|
||||
}
|
||||
|
||||
efispecid = (struct tcg_efi_specid_event_head *)event_header->event;
|
||||
|
||||
/* Check if event is malformed. */
|
||||
if (count > efispecid->num_algs) {
|
||||
if (memcmp(efispecid->signature, TCG_SPECID_SIG,
|
||||
sizeof(TCG_SPECID_SIG)) || count > efispecid->num_algs) {
|
||||
size = 0;
|
||||
goto out;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -106,7 +106,6 @@ extern void *vzalloc(unsigned long size);
|
|||
extern void *vmalloc_user(unsigned long size);
|
||||
extern void *vmalloc_node(unsigned long size, int node);
|
||||
extern void *vzalloc_node(unsigned long size, int node);
|
||||
extern void *vmalloc_exec(unsigned long size);
|
||||
extern void *vmalloc_32(unsigned long size);
|
||||
extern void *vmalloc_32_user(unsigned long size);
|
||||
extern void *__vmalloc(unsigned long size, gfp_t gfp_mask);
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue