Merge git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net
Trivial conflict in CAN, keep the net-next + the byteswap wrapper. Conflicts: drivers/net/can/usb/gs_usb.c Signed-off-by: Jakub Kicinski <kuba@kernel.org>
This commit is contained in:
commit
5c39f26e67
387 changed files with 6423 additions and 3000 deletions
|
|
@ -8,8 +8,10 @@
|
|||
+ __clang_patchlevel__)
|
||||
|
||||
#if CLANG_VERSION < 100001
|
||||
#ifndef __BPF_TRACING__
|
||||
# error Sorry, your version of Clang is too old - please use 10.0.1 or newer.
|
||||
#endif
|
||||
#endif
|
||||
|
||||
/* Compiler specific definitions for Clang compiler */
|
||||
|
||||
|
|
|
|||
|
|
@ -50,10 +50,6 @@
|
|||
#define ZYNQMP_PM_CAPABILITY_WAKEUP 0x4U
|
||||
#define ZYNQMP_PM_CAPABILITY_UNUSABLE 0x8U
|
||||
|
||||
/* Feature check status */
|
||||
#define PM_FEATURE_INVALID -1
|
||||
#define PM_FEATURE_UNCHECKED 0
|
||||
|
||||
/*
|
||||
* Firmware FPGA Manager flags
|
||||
* XILINX_ZYNQMP_PM_FPGA_FULL: FPGA full reconfiguration
|
||||
|
|
|
|||
|
|
@ -798,7 +798,6 @@ extern int iommu_calculate_agaw(struct intel_iommu *iommu);
|
|||
extern int iommu_calculate_max_sagaw(struct intel_iommu *iommu);
|
||||
extern int dmar_disabled;
|
||||
extern int intel_iommu_enabled;
|
||||
extern int intel_iommu_tboot_noforce;
|
||||
extern int intel_iommu_gfx_mapped;
|
||||
#else
|
||||
static inline int iommu_calculate_agaw(struct intel_iommu *iommu)
|
||||
|
|
|
|||
|
|
@ -401,7 +401,7 @@ static inline void jbd_unlock_bh_journal_head(struct buffer_head *bh)
|
|||
#define JI_WAIT_DATA (1 << __JI_WAIT_DATA)
|
||||
|
||||
/**
|
||||
* struct jbd_inode - The jbd_inode type is the structure linking inodes in
|
||||
* struct jbd2_inode - The jbd_inode type is the structure linking inodes in
|
||||
* ordered mode present in a transaction so that we can sync them during commit.
|
||||
*/
|
||||
struct jbd2_inode {
|
||||
|
|
|
|||
|
|
@ -282,20 +282,6 @@ struct mem_cgroup {
|
|||
|
||||
MEMCG_PADDING(_pad1_);
|
||||
|
||||
/*
|
||||
* set > 0 if pages under this cgroup are moving to other cgroup.
|
||||
*/
|
||||
atomic_t moving_account;
|
||||
struct task_struct *move_lock_task;
|
||||
|
||||
/* Legacy local VM stats and events */
|
||||
struct memcg_vmstats_percpu __percpu *vmstats_local;
|
||||
|
||||
/* Subtree VM stats and events (batched updates) */
|
||||
struct memcg_vmstats_percpu __percpu *vmstats_percpu;
|
||||
|
||||
MEMCG_PADDING(_pad2_);
|
||||
|
||||
atomic_long_t vmstats[MEMCG_NR_STAT];
|
||||
atomic_long_t vmevents[NR_VM_EVENT_ITEMS];
|
||||
|
||||
|
|
@ -317,6 +303,20 @@ struct mem_cgroup {
|
|||
struct list_head objcg_list; /* list of inherited objcgs */
|
||||
#endif
|
||||
|
||||
MEMCG_PADDING(_pad2_);
|
||||
|
||||
/*
|
||||
* set > 0 if pages under this cgroup are moving to other cgroup.
|
||||
*/
|
||||
atomic_t moving_account;
|
||||
struct task_struct *move_lock_task;
|
||||
|
||||
/* Legacy local VM stats and events */
|
||||
struct memcg_vmstats_percpu __percpu *vmstats_local;
|
||||
|
||||
/* Subtree VM stats and events (batched updates) */
|
||||
struct memcg_vmstats_percpu __percpu *vmstats_percpu;
|
||||
|
||||
#ifdef CONFIG_CGROUP_WRITEBACK
|
||||
struct list_head cgwb_list;
|
||||
struct wb_domain cgwb_domain;
|
||||
|
|
|
|||
|
|
@ -281,20 +281,6 @@ static inline bool movable_node_is_enabled(void)
|
|||
}
|
||||
#endif /* ! CONFIG_MEMORY_HOTPLUG */
|
||||
|
||||
#ifdef CONFIG_NUMA
|
||||
extern int memory_add_physaddr_to_nid(u64 start);
|
||||
extern int phys_to_target_node(u64 start);
|
||||
#else
|
||||
static inline int memory_add_physaddr_to_nid(u64 start)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline int phys_to_target_node(u64 start)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
#endif
|
||||
|
||||
#if defined(CONFIG_MEMORY_HOTPLUG) || defined(CONFIG_DEFERRED_STRUCT_PAGE_INIT)
|
||||
/*
|
||||
* pgdat resizing functions
|
||||
|
|
|
|||
|
|
@ -3163,6 +3163,11 @@ static inline bool dev_validate_header(const struct net_device *dev,
|
|||
return false;
|
||||
}
|
||||
|
||||
static inline bool dev_has_header(const struct net_device *dev)
|
||||
{
|
||||
return dev->header_ops && dev->header_ops->create;
|
||||
}
|
||||
|
||||
typedef int gifconf_func_t(struct net_device * dev, char __user * bufptr,
|
||||
int len, int size);
|
||||
int register_gifconf(unsigned int family, gifconf_func_t *gifconf);
|
||||
|
|
|
|||
|
|
@ -21,13 +21,41 @@
|
|||
#endif
|
||||
|
||||
#ifdef CONFIG_NUMA
|
||||
#include <linux/printk.h>
|
||||
#include <asm/sparsemem.h>
|
||||
|
||||
/* Generic implementation available */
|
||||
int numa_map_to_online_node(int node);
|
||||
#else
|
||||
|
||||
#ifndef memory_add_physaddr_to_nid
|
||||
static inline int memory_add_physaddr_to_nid(u64 start)
|
||||
{
|
||||
pr_info_once("Unknown online node for memory at 0x%llx, assuming node 0\n",
|
||||
start);
|
||||
return 0;
|
||||
}
|
||||
#endif
|
||||
#ifndef phys_to_target_node
|
||||
static inline int phys_to_target_node(u64 start)
|
||||
{
|
||||
pr_info_once("Unknown target node for memory at 0x%llx, assuming node 0\n",
|
||||
start);
|
||||
return 0;
|
||||
}
|
||||
#endif
|
||||
#else /* !CONFIG_NUMA */
|
||||
static inline int numa_map_to_online_node(int node)
|
||||
{
|
||||
return NUMA_NO_NODE;
|
||||
}
|
||||
static inline int memory_add_physaddr_to_nid(u64 start)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
static inline int phys_to_target_node(u64 start)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
#endif
|
||||
|
||||
#endif /* _LINUX_NUMA_H */
|
||||
|
|
|
|||
|
|
@ -906,6 +906,8 @@ static inline unsigned int __readahead_batch(struct readahead_control *rac,
|
|||
xas_set(&xas, rac->_index);
|
||||
rcu_read_lock();
|
||||
xas_for_each(&xas, page, rac->_index + rac->_nr_pages - 1) {
|
||||
if (xas_retry(&xas, page))
|
||||
continue;
|
||||
VM_BUG_ON_PAGE(!PageLocked(page), page);
|
||||
VM_BUG_ON_PAGE(PageTail(page), page);
|
||||
array[i++] = page;
|
||||
|
|
|
|||
|
|
@ -1427,6 +1427,19 @@ typedef unsigned int pgtbl_mod_mask;
|
|||
|
||||
#endif /* !__ASSEMBLY__ */
|
||||
|
||||
#if !defined(MAX_POSSIBLE_PHYSMEM_BITS) && !defined(CONFIG_64BIT)
|
||||
#ifdef CONFIG_PHYS_ADDR_T_64BIT
|
||||
/*
|
||||
* ZSMALLOC needs to know the highest PFN on 32-bit architectures
|
||||
* with physical address space extension, but falls back to
|
||||
* BITS_PER_LONG otherwise.
|
||||
*/
|
||||
#error Missing MAX_POSSIBLE_PHYSMEM_BITS definition
|
||||
#else
|
||||
#define MAX_POSSIBLE_PHYSMEM_BITS 32
|
||||
#endif
|
||||
#endif
|
||||
|
||||
#ifndef has_transparent_hugepage
|
||||
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
||||
#define has_transparent_hugepage() 1
|
||||
|
|
|
|||
|
|
@ -50,6 +50,7 @@ struct sysc_regbits {
|
|||
s8 emufree_shift;
|
||||
};
|
||||
|
||||
#define SYSC_MODULE_QUIRK_ENA_RESETDONE BIT(25)
|
||||
#define SYSC_MODULE_QUIRK_PRUSS BIT(24)
|
||||
#define SYSC_MODULE_QUIRK_DSS_RESET BIT(23)
|
||||
#define SYSC_MODULE_QUIRK_RTC_UNLOCK BIT(22)
|
||||
|
|
|
|||
|
|
@ -552,7 +552,6 @@ struct sched_dl_entity {
|
|||
* overruns.
|
||||
*/
|
||||
unsigned int dl_throttled : 1;
|
||||
unsigned int dl_boosted : 1;
|
||||
unsigned int dl_yielded : 1;
|
||||
unsigned int dl_non_contending : 1;
|
||||
unsigned int dl_overrun : 1;
|
||||
|
|
@ -571,6 +570,15 @@ struct sched_dl_entity {
|
|||
* time.
|
||||
*/
|
||||
struct hrtimer inactive_timer;
|
||||
|
||||
#ifdef CONFIG_RT_MUTEXES
|
||||
/*
|
||||
* Priority Inheritance. When a DEADLINE scheduling entity is boosted
|
||||
* pi_se points to the donor, otherwise points to the dl_se it belongs
|
||||
* to (the original one/itself).
|
||||
*/
|
||||
struct sched_dl_entity *pi_se;
|
||||
#endif
|
||||
};
|
||||
|
||||
#ifdef CONFIG_UCLAMP_TASK
|
||||
|
|
@ -770,7 +778,6 @@ struct task_struct {
|
|||
unsigned sched_reset_on_fork:1;
|
||||
unsigned sched_contributes_to_load:1;
|
||||
unsigned sched_migrated:1;
|
||||
unsigned sched_remote_wakeup:1;
|
||||
#ifdef CONFIG_PSI
|
||||
unsigned sched_psi_wake_requeue:1;
|
||||
#endif
|
||||
|
|
@ -780,6 +787,21 @@ struct task_struct {
|
|||
|
||||
/* Unserialized, strictly 'current' */
|
||||
|
||||
/*
|
||||
* This field must not be in the scheduler word above due to wakelist
|
||||
* queueing no longer being serialized by p->on_cpu. However:
|
||||
*
|
||||
* p->XXX = X; ttwu()
|
||||
* schedule() if (p->on_rq && ..) // false
|
||||
* smp_mb__after_spinlock(); if (smp_load_acquire(&p->on_cpu) && //true
|
||||
* deactivate_task() ttwu_queue_wakelist())
|
||||
* p->on_rq = 0; p->sched_remote_wakeup = Y;
|
||||
*
|
||||
* guarantees all stores of 'current' are visible before
|
||||
* ->sched_remote_wakeup gets used, so it can be in this word.
|
||||
*/
|
||||
unsigned sched_remote_wakeup:1;
|
||||
|
||||
/* Bit to tell LSMs we're in execve(): */
|
||||
unsigned in_execve:1;
|
||||
unsigned in_iowait:1;
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue