Annotation of /trunk/kernel-magellan/patches-4.20/0112-4.20.13-all-fixes.patch
Parent Directory | Revision Log
Revision 3289 -
(hide annotations)
(download)
Mon Mar 4 10:38:38 2019 UTC (5 years, 6 months ago) by niro
File size: 283893 byte(s)
Mon Mar 4 10:38:38 2019 UTC (5 years, 6 months ago) by niro
File size: 283893 byte(s)
linux-4.20.13
1 | niro | 3289 | diff --git a/Makefile b/Makefile |
2 | index 0a92b4e116210..c83abc1e689b4 100644 | ||
3 | --- a/Makefile | ||
4 | +++ b/Makefile | ||
5 | @@ -1,7 +1,7 @@ | ||
6 | # SPDX-License-Identifier: GPL-2.0 | ||
7 | VERSION = 4 | ||
8 | PATCHLEVEL = 20 | ||
9 | -SUBLEVEL = 12 | ||
10 | +SUBLEVEL = 13 | ||
11 | EXTRAVERSION = | ||
12 | NAME = Shy Crocodile | ||
13 | |||
14 | diff --git a/arch/arc/include/asm/cache.h b/arch/arc/include/asm/cache.h | ||
15 | index f393b663413e4..2ad77fb43639c 100644 | ||
16 | --- a/arch/arc/include/asm/cache.h | ||
17 | +++ b/arch/arc/include/asm/cache.h | ||
18 | @@ -52,6 +52,17 @@ | ||
19 | #define cache_line_size() SMP_CACHE_BYTES | ||
20 | #define ARCH_DMA_MINALIGN SMP_CACHE_BYTES | ||
21 | |||
22 | +/* | ||
23 | + * Make sure slab-allocated buffers are 64-bit aligned when atomic64_t uses | ||
24 | + * ARCv2 64-bit atomics (LLOCKD/SCONDD). This guarantess runtime 64-bit | ||
25 | + * alignment for any atomic64_t embedded in buffer. | ||
26 | + * Default ARCH_SLAB_MINALIGN is __alignof__(long long) which has a relaxed | ||
27 | + * value of 4 (and not 8) in ARC ABI. | ||
28 | + */ | ||
29 | +#if defined(CONFIG_ARC_HAS_LL64) && defined(CONFIG_ARC_HAS_LLSC) | ||
30 | +#define ARCH_SLAB_MINALIGN 8 | ||
31 | +#endif | ||
32 | + | ||
33 | extern void arc_cache_init(void); | ||
34 | extern char *arc_cache_mumbojumbo(int cpu_id, char *buf, int len); | ||
35 | extern void read_decode_cache_bcr(void); | ||
36 | diff --git a/arch/arc/kernel/head.S b/arch/arc/kernel/head.S | ||
37 | index 8b90d25a15cca..1f945d0f40daa 100644 | ||
38 | --- a/arch/arc/kernel/head.S | ||
39 | +++ b/arch/arc/kernel/head.S | ||
40 | @@ -17,6 +17,7 @@ | ||
41 | #include <asm/entry.h> | ||
42 | #include <asm/arcregs.h> | ||
43 | #include <asm/cache.h> | ||
44 | +#include <asm/irqflags.h> | ||
45 | |||
46 | .macro CPU_EARLY_SETUP | ||
47 | |||
48 | @@ -47,6 +48,15 @@ | ||
49 | sr r5, [ARC_REG_DC_CTRL] | ||
50 | |||
51 | 1: | ||
52 | + | ||
53 | +#ifdef CONFIG_ISA_ARCV2 | ||
54 | + ; Unaligned access is disabled at reset, so re-enable early as | ||
55 | + ; gcc 7.3.1 (ARC GNU 2018.03) onwards generates unaligned access | ||
56 | + ; by default | ||
57 | + lr r5, [status32] | ||
58 | + bset r5, r5, STATUS_AD_BIT | ||
59 | + kflag r5 | ||
60 | +#endif | ||
61 | .endm | ||
62 | |||
63 | .section .init.text, "ax",@progbits | ||
64 | @@ -93,9 +103,9 @@ ENTRY(stext) | ||
65 | #ifdef CONFIG_ARC_UBOOT_SUPPORT | ||
66 | ; Uboot - kernel ABI | ||
67 | ; r0 = [0] No uboot interaction, [1] cmdline in r2, [2] DTB in r2 | ||
68 | - ; r1 = magic number (board identity, unused as of now | ||
69 | + ; r1 = magic number (always zero as of now) | ||
70 | ; r2 = pointer to uboot provided cmdline or external DTB in mem | ||
71 | - ; These are handled later in setup_arch() | ||
72 | + ; These are handled later in handle_uboot_args() | ||
73 | st r0, [@uboot_tag] | ||
74 | st r2, [@uboot_arg] | ||
75 | #endif | ||
76 | diff --git a/arch/arc/kernel/setup.c b/arch/arc/kernel/setup.c | ||
77 | index eea8c5ce63350..80dd1a716ca72 100644 | ||
78 | --- a/arch/arc/kernel/setup.c | ||
79 | +++ b/arch/arc/kernel/setup.c | ||
80 | @@ -452,43 +452,80 @@ void setup_processor(void) | ||
81 | arc_chk_core_config(); | ||
82 | } | ||
83 | |||
84 | -static inline int is_kernel(unsigned long addr) | ||
85 | +static inline bool uboot_arg_invalid(unsigned long addr) | ||
86 | { | ||
87 | - if (addr >= (unsigned long)_stext && addr <= (unsigned long)_end) | ||
88 | - return 1; | ||
89 | - return 0; | ||
90 | + /* | ||
91 | + * Check that it is a untranslated address (although MMU is not enabled | ||
92 | + * yet, it being a high address ensures this is not by fluke) | ||
93 | + */ | ||
94 | + if (addr < PAGE_OFFSET) | ||
95 | + return true; | ||
96 | + | ||
97 | + /* Check that address doesn't clobber resident kernel image */ | ||
98 | + return addr >= (unsigned long)_stext && addr <= (unsigned long)_end; | ||
99 | } | ||
100 | |||
101 | -void __init setup_arch(char **cmdline_p) | ||
102 | +#define IGNORE_ARGS "Ignore U-boot args: " | ||
103 | + | ||
104 | +/* uboot_tag values for U-boot - kernel ABI revision 0; see head.S */ | ||
105 | +#define UBOOT_TAG_NONE 0 | ||
106 | +#define UBOOT_TAG_CMDLINE 1 | ||
107 | +#define UBOOT_TAG_DTB 2 | ||
108 | + | ||
109 | +void __init handle_uboot_args(void) | ||
110 | { | ||
111 | + bool use_embedded_dtb = true; | ||
112 | + bool append_cmdline = false; | ||
113 | + | ||
114 | #ifdef CONFIG_ARC_UBOOT_SUPPORT | ||
115 | - /* make sure that uboot passed pointer to cmdline/dtb is valid */ | ||
116 | - if (uboot_tag && is_kernel((unsigned long)uboot_arg)) | ||
117 | - panic("Invalid uboot arg\n"); | ||
118 | + /* check that we know this tag */ | ||
119 | + if (uboot_tag != UBOOT_TAG_NONE && | ||
120 | + uboot_tag != UBOOT_TAG_CMDLINE && | ||
121 | + uboot_tag != UBOOT_TAG_DTB) { | ||
122 | + pr_warn(IGNORE_ARGS "invalid uboot tag: '%08x'\n", uboot_tag); | ||
123 | + goto ignore_uboot_args; | ||
124 | + } | ||
125 | + | ||
126 | + if (uboot_tag != UBOOT_TAG_NONE && | ||
127 | + uboot_arg_invalid((unsigned long)uboot_arg)) { | ||
128 | + pr_warn(IGNORE_ARGS "invalid uboot arg: '%px'\n", uboot_arg); | ||
129 | + goto ignore_uboot_args; | ||
130 | + } | ||
131 | + | ||
132 | + /* see if U-boot passed an external Device Tree blob */ | ||
133 | + if (uboot_tag == UBOOT_TAG_DTB) { | ||
134 | + machine_desc = setup_machine_fdt((void *)uboot_arg); | ||
135 | |||
136 | - /* See if u-boot passed an external Device Tree blob */ | ||
137 | - machine_desc = setup_machine_fdt(uboot_arg); /* uboot_tag == 2 */ | ||
138 | - if (!machine_desc) | ||
139 | + /* external Device Tree blob is invalid - use embedded one */ | ||
140 | + use_embedded_dtb = !machine_desc; | ||
141 | + } | ||
142 | + | ||
143 | + if (uboot_tag == UBOOT_TAG_CMDLINE) | ||
144 | + append_cmdline = true; | ||
145 | + | ||
146 | +ignore_uboot_args: | ||
147 | #endif | ||
148 | - { | ||
149 | - /* No, so try the embedded one */ | ||
150 | + | ||
151 | + if (use_embedded_dtb) { | ||
152 | machine_desc = setup_machine_fdt(__dtb_start); | ||
153 | if (!machine_desc) | ||
154 | panic("Embedded DT invalid\n"); | ||
155 | + } | ||
156 | |||
157 | - /* | ||
158 | - * If we are here, it is established that @uboot_arg didn't | ||
159 | - * point to DT blob. Instead if u-boot says it is cmdline, | ||
160 | - * append to embedded DT cmdline. | ||
161 | - * setup_machine_fdt() would have populated @boot_command_line | ||
162 | - */ | ||
163 | - if (uboot_tag == 1) { | ||
164 | - /* Ensure a whitespace between the 2 cmdlines */ | ||
165 | - strlcat(boot_command_line, " ", COMMAND_LINE_SIZE); | ||
166 | - strlcat(boot_command_line, uboot_arg, | ||
167 | - COMMAND_LINE_SIZE); | ||
168 | - } | ||
169 | + /* | ||
170 | + * NOTE: @boot_command_line is populated by setup_machine_fdt() so this | ||
171 | + * append processing can only happen after. | ||
172 | + */ | ||
173 | + if (append_cmdline) { | ||
174 | + /* Ensure a whitespace between the 2 cmdlines */ | ||
175 | + strlcat(boot_command_line, " ", COMMAND_LINE_SIZE); | ||
176 | + strlcat(boot_command_line, uboot_arg, COMMAND_LINE_SIZE); | ||
177 | } | ||
178 | +} | ||
179 | + | ||
180 | +void __init setup_arch(char **cmdline_p) | ||
181 | +{ | ||
182 | + handle_uboot_args(); | ||
183 | |||
184 | /* Save unparsed command line copy for /proc/cmdline */ | ||
185 | *cmdline_p = boot_command_line; | ||
186 | diff --git a/arch/arm/probes/kprobes/opt-arm.c b/arch/arm/probes/kprobes/opt-arm.c | ||
187 | index 2c118a6ab3587..0dc23fc227ed2 100644 | ||
188 | --- a/arch/arm/probes/kprobes/opt-arm.c | ||
189 | +++ b/arch/arm/probes/kprobes/opt-arm.c | ||
190 | @@ -247,7 +247,7 @@ int arch_prepare_optimized_kprobe(struct optimized_kprobe *op, struct kprobe *or | ||
191 | } | ||
192 | |||
193 | /* Copy arch-dep-instance from template. */ | ||
194 | - memcpy(code, (unsigned char *)optprobe_template_entry, | ||
195 | + memcpy(code, (unsigned long *)&optprobe_template_entry, | ||
196 | TMPL_END_IDX * sizeof(kprobe_opcode_t)); | ||
197 | |||
198 | /* Adjust buffer according to instruction. */ | ||
199 | diff --git a/arch/mips/configs/ath79_defconfig b/arch/mips/configs/ath79_defconfig | ||
200 | index 951c4231bdb85..4c47b3fd958b6 100644 | ||
201 | --- a/arch/mips/configs/ath79_defconfig | ||
202 | +++ b/arch/mips/configs/ath79_defconfig | ||
203 | @@ -71,6 +71,7 @@ CONFIG_SERIAL_8250_CONSOLE=y | ||
204 | # CONFIG_SERIAL_8250_PCI is not set | ||
205 | CONFIG_SERIAL_8250_NR_UARTS=1 | ||
206 | CONFIG_SERIAL_8250_RUNTIME_UARTS=1 | ||
207 | +CONFIG_SERIAL_OF_PLATFORM=y | ||
208 | CONFIG_SERIAL_AR933X=y | ||
209 | CONFIG_SERIAL_AR933X_CONSOLE=y | ||
210 | # CONFIG_HW_RANDOM is not set | ||
211 | diff --git a/arch/mips/jazz/jazzdma.c b/arch/mips/jazz/jazzdma.c | ||
212 | index 4c41ed0a637e5..415a08376c362 100644 | ||
213 | --- a/arch/mips/jazz/jazzdma.c | ||
214 | +++ b/arch/mips/jazz/jazzdma.c | ||
215 | @@ -74,14 +74,15 @@ static int __init vdma_init(void) | ||
216 | get_order(VDMA_PGTBL_SIZE)); | ||
217 | BUG_ON(!pgtbl); | ||
218 | dma_cache_wback_inv((unsigned long)pgtbl, VDMA_PGTBL_SIZE); | ||
219 | - pgtbl = (VDMA_PGTBL_ENTRY *)KSEG1ADDR(pgtbl); | ||
220 | + pgtbl = (VDMA_PGTBL_ENTRY *)CKSEG1ADDR((unsigned long)pgtbl); | ||
221 | |||
222 | /* | ||
223 | * Clear the R4030 translation table | ||
224 | */ | ||
225 | vdma_pgtbl_init(); | ||
226 | |||
227 | - r4030_write_reg32(JAZZ_R4030_TRSTBL_BASE, CPHYSADDR(pgtbl)); | ||
228 | + r4030_write_reg32(JAZZ_R4030_TRSTBL_BASE, | ||
229 | + CPHYSADDR((unsigned long)pgtbl)); | ||
230 | r4030_write_reg32(JAZZ_R4030_TRSTBL_LIM, VDMA_PGTBL_SIZE); | ||
231 | r4030_write_reg32(JAZZ_R4030_TRSTBL_INV, 0); | ||
232 | |||
233 | diff --git a/arch/mips/net/ebpf_jit.c b/arch/mips/net/ebpf_jit.c | ||
234 | index aeb7b1b0f2024..252c00985c973 100644 | ||
235 | --- a/arch/mips/net/ebpf_jit.c | ||
236 | +++ b/arch/mips/net/ebpf_jit.c | ||
237 | @@ -343,12 +343,15 @@ static int build_int_epilogue(struct jit_ctx *ctx, int dest_reg) | ||
238 | const struct bpf_prog *prog = ctx->skf; | ||
239 | int stack_adjust = ctx->stack_size; | ||
240 | int store_offset = stack_adjust - 8; | ||
241 | + enum reg_val_type td; | ||
242 | int r0 = MIPS_R_V0; | ||
243 | |||
244 | - if (dest_reg == MIPS_R_RA && | ||
245 | - get_reg_val_type(ctx, prog->len, BPF_REG_0) == REG_32BIT_ZERO_EX) | ||
246 | + if (dest_reg == MIPS_R_RA) { | ||
247 | /* Don't let zero extended value escape. */ | ||
248 | - emit_instr(ctx, sll, r0, r0, 0); | ||
249 | + td = get_reg_val_type(ctx, prog->len, BPF_REG_0); | ||
250 | + if (td == REG_64BIT || td == REG_32BIT_ZERO_EX) | ||
251 | + emit_instr(ctx, sll, r0, r0, 0); | ||
252 | + } | ||
253 | |||
254 | if (ctx->flags & EBPF_SAVE_RA) { | ||
255 | emit_instr(ctx, ld, MIPS_R_RA, store_offset, MIPS_R_SP); | ||
256 | diff --git a/arch/parisc/kernel/ptrace.c b/arch/parisc/kernel/ptrace.c | ||
257 | index 2582df1c529bb..0964c236e3e5a 100644 | ||
258 | --- a/arch/parisc/kernel/ptrace.c | ||
259 | +++ b/arch/parisc/kernel/ptrace.c | ||
260 | @@ -308,15 +308,29 @@ long compat_arch_ptrace(struct task_struct *child, compat_long_t request, | ||
261 | |||
262 | long do_syscall_trace_enter(struct pt_regs *regs) | ||
263 | { | ||
264 | - if (test_thread_flag(TIF_SYSCALL_TRACE) && | ||
265 | - tracehook_report_syscall_entry(regs)) { | ||
266 | + if (test_thread_flag(TIF_SYSCALL_TRACE)) { | ||
267 | + int rc = tracehook_report_syscall_entry(regs); | ||
268 | + | ||
269 | /* | ||
270 | - * Tracing decided this syscall should not happen or the | ||
271 | - * debugger stored an invalid system call number. Skip | ||
272 | - * the system call and the system call restart handling. | ||
273 | + * As tracesys_next does not set %r28 to -ENOSYS | ||
274 | + * when %r20 is set to -1, initialize it here. | ||
275 | */ | ||
276 | - regs->gr[20] = -1UL; | ||
277 | - goto out; | ||
278 | + regs->gr[28] = -ENOSYS; | ||
279 | + | ||
280 | + if (rc) { | ||
281 | + /* | ||
282 | + * A nonzero return code from | ||
283 | + * tracehook_report_syscall_entry() tells us | ||
284 | + * to prevent the syscall execution. Skip | ||
285 | + * the syscall call and the syscall restart handling. | ||
286 | + * | ||
287 | + * Note that the tracer may also just change | ||
288 | + * regs->gr[20] to an invalid syscall number, | ||
289 | + * that is handled by tracesys_next. | ||
290 | + */ | ||
291 | + regs->gr[20] = -1UL; | ||
292 | + return -1; | ||
293 | + } | ||
294 | } | ||
295 | |||
296 | /* Do the secure computing check after ptrace. */ | ||
297 | @@ -340,7 +354,6 @@ long do_syscall_trace_enter(struct pt_regs *regs) | ||
298 | regs->gr[24] & 0xffffffff, | ||
299 | regs->gr[23] & 0xffffffff); | ||
300 | |||
301 | -out: | ||
302 | /* | ||
303 | * Sign extend the syscall number to 64bit since it may have been | ||
304 | * modified by a compat ptrace call | ||
305 | diff --git a/arch/powerpc/kernel/head_8xx.S b/arch/powerpc/kernel/head_8xx.S | ||
306 | index 3b67b9533c82f..438512759e827 100644 | ||
307 | --- a/arch/powerpc/kernel/head_8xx.S | ||
308 | +++ b/arch/powerpc/kernel/head_8xx.S | ||
309 | @@ -927,11 +927,12 @@ start_here: | ||
310 | |||
311 | /* set up the PTE pointers for the Abatron bdiGDB. | ||
312 | */ | ||
313 | - tovirt(r6,r6) | ||
314 | lis r5, abatron_pteptrs@h | ||
315 | ori r5, r5, abatron_pteptrs@l | ||
316 | stw r5, 0xf0(0) /* Must match your Abatron config file */ | ||
317 | tophys(r5,r5) | ||
318 | + lis r6, swapper_pg_dir@h | ||
319 | + ori r6, r6, swapper_pg_dir@l | ||
320 | stw r6, 0(r5) | ||
321 | |||
322 | /* Now turn on the MMU for real! */ | ||
323 | diff --git a/arch/s390/kvm/vsie.c b/arch/s390/kvm/vsie.c | ||
324 | index a153257bf7d98..d62fa148558b9 100644 | ||
325 | --- a/arch/s390/kvm/vsie.c | ||
326 | +++ b/arch/s390/kvm/vsie.c | ||
327 | @@ -297,7 +297,7 @@ static int shadow_crycb(struct kvm_vcpu *vcpu, struct vsie_page *vsie_page) | ||
328 | scb_s->crycbd = 0; | ||
329 | |||
330 | apie_h = vcpu->arch.sie_block->eca & ECA_APIE; | ||
331 | - if (!apie_h && !key_msk) | ||
332 | + if (!apie_h && (!key_msk || fmt_o == CRYCB_FORMAT0)) | ||
333 | return 0; | ||
334 | |||
335 | if (!crycb_addr) | ||
336 | diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h | ||
337 | index e5c0174e330e5..5a0cbc717997b 100644 | ||
338 | --- a/arch/x86/include/asm/kvm_host.h | ||
339 | +++ b/arch/x86/include/asm/kvm_host.h | ||
340 | @@ -299,6 +299,7 @@ union kvm_mmu_extended_role { | ||
341 | unsigned int cr4_smap:1; | ||
342 | unsigned int cr4_smep:1; | ||
343 | unsigned int cr4_la57:1; | ||
344 | + unsigned int maxphyaddr:6; | ||
345 | }; | ||
346 | }; | ||
347 | |||
348 | @@ -397,6 +398,7 @@ struct kvm_mmu { | ||
349 | void (*update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, | ||
350 | u64 *spte, const void *pte); | ||
351 | hpa_t root_hpa; | ||
352 | + gpa_t root_cr3; | ||
353 | union kvm_mmu_role mmu_role; | ||
354 | u8 root_level; | ||
355 | u8 shadow_root_level; | ||
356 | diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c | ||
357 | index 7bcfa61375c09..98d13c6a64be0 100644 | ||
358 | --- a/arch/x86/kvm/cpuid.c | ||
359 | +++ b/arch/x86/kvm/cpuid.c | ||
360 | @@ -337,6 +337,7 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, | ||
361 | unsigned f_mpx = kvm_mpx_supported() ? F(MPX) : 0; | ||
362 | unsigned f_xsaves = kvm_x86_ops->xsaves_supported() ? F(XSAVES) : 0; | ||
363 | unsigned f_umip = kvm_x86_ops->umip_emulated() ? F(UMIP) : 0; | ||
364 | + unsigned f_la57 = 0; | ||
365 | |||
366 | /* cpuid 1.edx */ | ||
367 | const u32 kvm_cpuid_1_edx_x86_features = | ||
368 | @@ -491,7 +492,10 @@ static inline int __do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function, | ||
369 | // TSC_ADJUST is emulated | ||
370 | entry->ebx |= F(TSC_ADJUST); | ||
371 | entry->ecx &= kvm_cpuid_7_0_ecx_x86_features; | ||
372 | + f_la57 = entry->ecx & F(LA57); | ||
373 | cpuid_mask(&entry->ecx, CPUID_7_ECX); | ||
374 | + /* Set LA57 based on hardware capability. */ | ||
375 | + entry->ecx |= f_la57; | ||
376 | entry->ecx |= f_umip; | ||
377 | /* PKU is not yet implemented for shadow paging. */ | ||
378 | if (!tdp_enabled || !boot_cpu_has(X86_FEATURE_OSPKE)) | ||
379 | diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c | ||
380 | index 7c03c0f35444f..e763e5445e3ca 100644 | ||
381 | --- a/arch/x86/kvm/mmu.c | ||
382 | +++ b/arch/x86/kvm/mmu.c | ||
383 | @@ -3517,6 +3517,7 @@ void kvm_mmu_free_roots(struct kvm_vcpu *vcpu, struct kvm_mmu *mmu, | ||
384 | &invalid_list); | ||
385 | mmu->root_hpa = INVALID_PAGE; | ||
386 | } | ||
387 | + mmu->root_cr3 = 0; | ||
388 | } | ||
389 | |||
390 | kvm_mmu_commit_zap_page(vcpu->kvm, &invalid_list); | ||
391 | @@ -3572,6 +3573,7 @@ static int mmu_alloc_direct_roots(struct kvm_vcpu *vcpu) | ||
392 | vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->pae_root); | ||
393 | } else | ||
394 | BUG(); | ||
395 | + vcpu->arch.mmu->root_cr3 = vcpu->arch.mmu->get_cr3(vcpu); | ||
396 | |||
397 | return 0; | ||
398 | } | ||
399 | @@ -3580,10 +3582,11 @@ static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu) | ||
400 | { | ||
401 | struct kvm_mmu_page *sp; | ||
402 | u64 pdptr, pm_mask; | ||
403 | - gfn_t root_gfn; | ||
404 | + gfn_t root_gfn, root_cr3; | ||
405 | int i; | ||
406 | |||
407 | - root_gfn = vcpu->arch.mmu->get_cr3(vcpu) >> PAGE_SHIFT; | ||
408 | + root_cr3 = vcpu->arch.mmu->get_cr3(vcpu); | ||
409 | + root_gfn = root_cr3 >> PAGE_SHIFT; | ||
410 | |||
411 | if (mmu_check_root(vcpu, root_gfn)) | ||
412 | return 1; | ||
413 | @@ -3608,7 +3611,7 @@ static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu) | ||
414 | ++sp->root_count; | ||
415 | spin_unlock(&vcpu->kvm->mmu_lock); | ||
416 | vcpu->arch.mmu->root_hpa = root; | ||
417 | - return 0; | ||
418 | + goto set_root_cr3; | ||
419 | } | ||
420 | |||
421 | /* | ||
422 | @@ -3674,6 +3677,9 @@ static int mmu_alloc_shadow_roots(struct kvm_vcpu *vcpu) | ||
423 | vcpu->arch.mmu->root_hpa = __pa(vcpu->arch.mmu->lm_root); | ||
424 | } | ||
425 | |||
426 | +set_root_cr3: | ||
427 | + vcpu->arch.mmu->root_cr3 = root_cr3; | ||
428 | + | ||
429 | return 0; | ||
430 | } | ||
431 | |||
432 | @@ -4125,7 +4131,7 @@ static bool cached_root_available(struct kvm_vcpu *vcpu, gpa_t new_cr3, | ||
433 | struct kvm_mmu_root_info root; | ||
434 | struct kvm_mmu *mmu = vcpu->arch.mmu; | ||
435 | |||
436 | - root.cr3 = mmu->get_cr3(vcpu); | ||
437 | + root.cr3 = mmu->root_cr3; | ||
438 | root.hpa = mmu->root_hpa; | ||
439 | |||
440 | for (i = 0; i < KVM_MMU_NUM_PREV_ROOTS; i++) { | ||
441 | @@ -4138,6 +4144,7 @@ static bool cached_root_available(struct kvm_vcpu *vcpu, gpa_t new_cr3, | ||
442 | } | ||
443 | |||
444 | mmu->root_hpa = root.hpa; | ||
445 | + mmu->root_cr3 = root.cr3; | ||
446 | |||
447 | return i < KVM_MMU_NUM_PREV_ROOTS; | ||
448 | } | ||
449 | @@ -4731,6 +4738,7 @@ static union kvm_mmu_extended_role kvm_calc_mmu_role_ext(struct kvm_vcpu *vcpu) | ||
450 | ext.cr4_pse = !!is_pse(vcpu); | ||
451 | ext.cr4_pke = !!kvm_read_cr4_bits(vcpu, X86_CR4_PKE); | ||
452 | ext.cr4_la57 = !!kvm_read_cr4_bits(vcpu, X86_CR4_LA57); | ||
453 | + ext.maxphyaddr = cpuid_maxphyaddr(vcpu); | ||
454 | |||
455 | ext.valid = 1; | ||
456 | |||
457 | @@ -5477,11 +5485,13 @@ int kvm_mmu_create(struct kvm_vcpu *vcpu) | ||
458 | vcpu->arch.walk_mmu = &vcpu->arch.root_mmu; | ||
459 | |||
460 | vcpu->arch.root_mmu.root_hpa = INVALID_PAGE; | ||
461 | + vcpu->arch.root_mmu.root_cr3 = 0; | ||
462 | vcpu->arch.root_mmu.translate_gpa = translate_gpa; | ||
463 | for (i = 0; i < KVM_MMU_NUM_PREV_ROOTS; i++) | ||
464 | vcpu->arch.root_mmu.prev_roots[i] = KVM_MMU_ROOT_INFO_INVALID; | ||
465 | |||
466 | vcpu->arch.guest_mmu.root_hpa = INVALID_PAGE; | ||
467 | + vcpu->arch.guest_mmu.root_cr3 = 0; | ||
468 | vcpu->arch.guest_mmu.translate_gpa = translate_gpa; | ||
469 | for (i = 0; i < KVM_MMU_NUM_PREV_ROOTS; i++) | ||
470 | vcpu->arch.guest_mmu.prev_roots[i] = KVM_MMU_ROOT_INFO_INVALID; | ||
471 | diff --git a/arch/x86/xen/enlighten_pv.c b/arch/x86/xen/enlighten_pv.c | ||
472 | index 2f6787fc71066..c54a493e139a7 100644 | ||
473 | --- a/arch/x86/xen/enlighten_pv.c | ||
474 | +++ b/arch/x86/xen/enlighten_pv.c | ||
475 | @@ -898,10 +898,7 @@ static u64 xen_read_msr_safe(unsigned int msr, int *err) | ||
476 | val = native_read_msr_safe(msr, err); | ||
477 | switch (msr) { | ||
478 | case MSR_IA32_APICBASE: | ||
479 | -#ifdef CONFIG_X86_X2APIC | ||
480 | - if (!(cpuid_ecx(1) & (1 << (X86_FEATURE_X2APIC & 31)))) | ||
481 | -#endif | ||
482 | - val &= ~X2APIC_ENABLE; | ||
483 | + val &= ~X2APIC_ENABLE; | ||
484 | break; | ||
485 | } | ||
486 | return val; | ||
487 | diff --git a/drivers/acpi/bus.c b/drivers/acpi/bus.c | ||
488 | index bb3d96dea6dba..26d4164d394fb 100644 | ||
489 | --- a/drivers/acpi/bus.c | ||
490 | +++ b/drivers/acpi/bus.c | ||
491 | @@ -1054,18 +1054,6 @@ void __init acpi_early_init(void) | ||
492 | goto error0; | ||
493 | } | ||
494 | |||
495 | - /* | ||
496 | - * ACPI 2.0 requires the EC driver to be loaded and work before | ||
497 | - * the EC device is found in the namespace (i.e. before | ||
498 | - * acpi_load_tables() is called). | ||
499 | - * | ||
500 | - * This is accomplished by looking for the ECDT table, and getting | ||
501 | - * the EC parameters out of that. | ||
502 | - * | ||
503 | - * Ignore the result. Not having an ECDT is not fatal. | ||
504 | - */ | ||
505 | - status = acpi_ec_ecdt_probe(); | ||
506 | - | ||
507 | #ifdef CONFIG_X86 | ||
508 | if (!acpi_ioapic) { | ||
509 | /* compatible (0) means level (3) */ | ||
510 | @@ -1142,6 +1130,18 @@ static int __init acpi_bus_init(void) | ||
511 | goto error1; | ||
512 | } | ||
513 | |||
514 | + /* | ||
515 | + * ACPI 2.0 requires the EC driver to be loaded and work before the EC | ||
516 | + * device is found in the namespace. | ||
517 | + * | ||
518 | + * This is accomplished by looking for the ECDT table and getting the EC | ||
519 | + * parameters out of that. | ||
520 | + * | ||
521 | + * Do that before calling acpi_initialize_objects() which may trigger EC | ||
522 | + * address space accesses. | ||
523 | + */ | ||
524 | + acpi_ec_ecdt_probe(); | ||
525 | + | ||
526 | status = acpi_enable_subsystem(ACPI_NO_ACPI_ENABLE); | ||
527 | if (ACPI_FAILURE(status)) { | ||
528 | printk(KERN_ERR PREFIX | ||
529 | diff --git a/drivers/acpi/nfit/core.c b/drivers/acpi/nfit/core.c | ||
530 | index 8535e7999769b..2a2d7ec772526 100644 | ||
531 | --- a/drivers/acpi/nfit/core.c | ||
532 | +++ b/drivers/acpi/nfit/core.c | ||
533 | @@ -724,6 +724,7 @@ int nfit_get_smbios_id(u32 device_handle, u16 *flags) | ||
534 | struct acpi_nfit_memory_map *memdev; | ||
535 | struct acpi_nfit_desc *acpi_desc; | ||
536 | struct nfit_mem *nfit_mem; | ||
537 | + u16 physical_id; | ||
538 | |||
539 | mutex_lock(&acpi_desc_lock); | ||
540 | list_for_each_entry(acpi_desc, &acpi_descs, list) { | ||
541 | @@ -731,10 +732,11 @@ int nfit_get_smbios_id(u32 device_handle, u16 *flags) | ||
542 | list_for_each_entry(nfit_mem, &acpi_desc->dimms, list) { | ||
543 | memdev = __to_nfit_memdev(nfit_mem); | ||
544 | if (memdev->device_handle == device_handle) { | ||
545 | + *flags = memdev->flags; | ||
546 | + physical_id = memdev->physical_id; | ||
547 | mutex_unlock(&acpi_desc->init_mutex); | ||
548 | mutex_unlock(&acpi_desc_lock); | ||
549 | - *flags = memdev->flags; | ||
550 | - return memdev->physical_id; | ||
551 | + return physical_id; | ||
552 | } | ||
553 | } | ||
554 | mutex_unlock(&acpi_desc->init_mutex); | ||
555 | diff --git a/drivers/atm/he.c b/drivers/atm/he.c | ||
556 | index 29f102dcfec49..329ce9072ee9f 100644 | ||
557 | --- a/drivers/atm/he.c | ||
558 | +++ b/drivers/atm/he.c | ||
559 | @@ -717,7 +717,7 @@ static int he_init_cs_block_rcm(struct he_dev *he_dev) | ||
560 | instead of '/ 512', use '>> 9' to prevent a call | ||
561 | to divdu3 on x86 platforms | ||
562 | */ | ||
563 | - rate_cps = (unsigned long long) (1 << exp) * (man + 512) >> 9; | ||
564 | + rate_cps = (unsigned long long) (1UL << exp) * (man + 512) >> 9; | ||
565 | |||
566 | if (rate_cps < 10) | ||
567 | rate_cps = 10; /* 2.2.1 minimum payload rate is 10 cps */ | ||
568 | diff --git a/drivers/clk/at91/at91sam9x5.c b/drivers/clk/at91/at91sam9x5.c | ||
569 | index 2fe225a697df8..3487e03d4bc61 100644 | ||
570 | --- a/drivers/clk/at91/at91sam9x5.c | ||
571 | +++ b/drivers/clk/at91/at91sam9x5.c | ||
572 | @@ -144,8 +144,7 @@ static void __init at91sam9x5_pmc_setup(struct device_node *np, | ||
573 | return; | ||
574 | |||
575 | at91sam9x5_pmc = pmc_data_allocate(PMC_MAIN + 1, | ||
576 | - nck(at91sam9x5_systemck), | ||
577 | - nck(at91sam9x35_periphck), 0); | ||
578 | + nck(at91sam9x5_systemck), 31, 0); | ||
579 | if (!at91sam9x5_pmc) | ||
580 | return; | ||
581 | |||
582 | @@ -210,7 +209,7 @@ static void __init at91sam9x5_pmc_setup(struct device_node *np, | ||
583 | parent_names[1] = "mainck"; | ||
584 | parent_names[2] = "plladivck"; | ||
585 | parent_names[3] = "utmick"; | ||
586 | - parent_names[4] = "mck"; | ||
587 | + parent_names[4] = "masterck"; | ||
588 | for (i = 0; i < 2; i++) { | ||
589 | char name[6]; | ||
590 | |||
591 | diff --git a/drivers/clk/at91/sama5d2.c b/drivers/clk/at91/sama5d2.c | ||
592 | index d69ad96fe988b..cd0ef7274fdbf 100644 | ||
593 | --- a/drivers/clk/at91/sama5d2.c | ||
594 | +++ b/drivers/clk/at91/sama5d2.c | ||
595 | @@ -240,7 +240,7 @@ static void __init sama5d2_pmc_setup(struct device_node *np) | ||
596 | parent_names[1] = "mainck"; | ||
597 | parent_names[2] = "plladivck"; | ||
598 | parent_names[3] = "utmick"; | ||
599 | - parent_names[4] = "mck"; | ||
600 | + parent_names[4] = "masterck"; | ||
601 | for (i = 0; i < 3; i++) { | ||
602 | char name[6]; | ||
603 | |||
604 | @@ -291,7 +291,7 @@ static void __init sama5d2_pmc_setup(struct device_node *np) | ||
605 | parent_names[1] = "mainck"; | ||
606 | parent_names[2] = "plladivck"; | ||
607 | parent_names[3] = "utmick"; | ||
608 | - parent_names[4] = "mck"; | ||
609 | + parent_names[4] = "masterck"; | ||
610 | parent_names[5] = "audiopll_pmcck"; | ||
611 | for (i = 0; i < ARRAY_SIZE(sama5d2_gck); i++) { | ||
612 | hw = at91_clk_register_generated(regmap, &pmc_pcr_lock, | ||
613 | diff --git a/drivers/clk/at91/sama5d4.c b/drivers/clk/at91/sama5d4.c | ||
614 | index e358be7f6c8d5..b645a9d59cdbd 100644 | ||
615 | --- a/drivers/clk/at91/sama5d4.c | ||
616 | +++ b/drivers/clk/at91/sama5d4.c | ||
617 | @@ -207,7 +207,7 @@ static void __init sama5d4_pmc_setup(struct device_node *np) | ||
618 | parent_names[1] = "mainck"; | ||
619 | parent_names[2] = "plladivck"; | ||
620 | parent_names[3] = "utmick"; | ||
621 | - parent_names[4] = "mck"; | ||
622 | + parent_names[4] = "masterck"; | ||
623 | for (i = 0; i < 3; i++) { | ||
624 | char name[6]; | ||
625 | |||
626 | diff --git a/drivers/cpufreq/scmi-cpufreq.c b/drivers/cpufreq/scmi-cpufreq.c | ||
627 | index 242c3370544e6..9ed46d188cb5b 100644 | ||
628 | --- a/drivers/cpufreq/scmi-cpufreq.c | ||
629 | +++ b/drivers/cpufreq/scmi-cpufreq.c | ||
630 | @@ -187,8 +187,8 @@ static int scmi_cpufreq_exit(struct cpufreq_policy *policy) | ||
631 | |||
632 | cpufreq_cooling_unregister(priv->cdev); | ||
633 | dev_pm_opp_free_cpufreq_table(priv->cpu_dev, &policy->freq_table); | ||
634 | - kfree(priv); | ||
635 | dev_pm_opp_remove_all_dynamic(priv->cpu_dev); | ||
636 | + kfree(priv); | ||
637 | |||
638 | return 0; | ||
639 | } | ||
640 | diff --git a/drivers/gpio/gpio-mt7621.c b/drivers/gpio/gpio-mt7621.c | ||
641 | index 00e954f22bc92..74401e0adb29c 100644 | ||
642 | --- a/drivers/gpio/gpio-mt7621.c | ||
643 | +++ b/drivers/gpio/gpio-mt7621.c | ||
644 | @@ -30,6 +30,7 @@ | ||
645 | #define GPIO_REG_EDGE 0xA0 | ||
646 | |||
647 | struct mtk_gc { | ||
648 | + struct irq_chip irq_chip; | ||
649 | struct gpio_chip chip; | ||
650 | spinlock_t lock; | ||
651 | int bank; | ||
652 | @@ -189,13 +190,6 @@ mediatek_gpio_irq_type(struct irq_data *d, unsigned int type) | ||
653 | return 0; | ||
654 | } | ||
655 | |||
656 | -static struct irq_chip mediatek_gpio_irq_chip = { | ||
657 | - .irq_unmask = mediatek_gpio_irq_unmask, | ||
658 | - .irq_mask = mediatek_gpio_irq_mask, | ||
659 | - .irq_mask_ack = mediatek_gpio_irq_mask, | ||
660 | - .irq_set_type = mediatek_gpio_irq_type, | ||
661 | -}; | ||
662 | - | ||
663 | static int | ||
664 | mediatek_gpio_xlate(struct gpio_chip *chip, | ||
665 | const struct of_phandle_args *spec, u32 *flags) | ||
666 | @@ -254,6 +248,13 @@ mediatek_gpio_bank_probe(struct device *dev, | ||
667 | return ret; | ||
668 | } | ||
669 | |||
670 | + rg->irq_chip.name = dev_name(dev); | ||
671 | + rg->irq_chip.parent_device = dev; | ||
672 | + rg->irq_chip.irq_unmask = mediatek_gpio_irq_unmask; | ||
673 | + rg->irq_chip.irq_mask = mediatek_gpio_irq_mask; | ||
674 | + rg->irq_chip.irq_mask_ack = mediatek_gpio_irq_mask; | ||
675 | + rg->irq_chip.irq_set_type = mediatek_gpio_irq_type; | ||
676 | + | ||
677 | if (mtk->gpio_irq) { | ||
678 | /* | ||
679 | * Manually request the irq here instead of passing | ||
680 | @@ -270,14 +271,14 @@ mediatek_gpio_bank_probe(struct device *dev, | ||
681 | return ret; | ||
682 | } | ||
683 | |||
684 | - ret = gpiochip_irqchip_add(&rg->chip, &mediatek_gpio_irq_chip, | ||
685 | + ret = gpiochip_irqchip_add(&rg->chip, &rg->irq_chip, | ||
686 | 0, handle_simple_irq, IRQ_TYPE_NONE); | ||
687 | if (ret) { | ||
688 | dev_err(dev, "failed to add gpiochip_irqchip\n"); | ||
689 | return ret; | ||
690 | } | ||
691 | |||
692 | - gpiochip_set_chained_irqchip(&rg->chip, &mediatek_gpio_irq_chip, | ||
693 | + gpiochip_set_chained_irqchip(&rg->chip, &rg->irq_chip, | ||
694 | mtk->gpio_irq, NULL); | ||
695 | } | ||
696 | |||
697 | @@ -310,7 +311,6 @@ mediatek_gpio_probe(struct platform_device *pdev) | ||
698 | mtk->gpio_irq = irq_of_parse_and_map(np, 0); | ||
699 | mtk->dev = dev; | ||
700 | platform_set_drvdata(pdev, mtk); | ||
701 | - mediatek_gpio_irq_chip.name = dev_name(dev); | ||
702 | |||
703 | for (i = 0; i < MTK_BANK_CNT; i++) { | ||
704 | ret = mediatek_gpio_bank_probe(dev, np, i); | ||
705 | diff --git a/drivers/gpio/gpio-pxa.c b/drivers/gpio/gpio-pxa.c | ||
706 | index e9600b556f397..bcc6be4a5cb2e 100644 | ||
707 | --- a/drivers/gpio/gpio-pxa.c | ||
708 | +++ b/drivers/gpio/gpio-pxa.c | ||
709 | @@ -245,6 +245,7 @@ static bool pxa_gpio_has_pinctrl(void) | ||
710 | { | ||
711 | switch (gpio_type) { | ||
712 | case PXA3XX_GPIO: | ||
713 | + case MMP2_GPIO: | ||
714 | return false; | ||
715 | |||
716 | default: | ||
717 | diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c | ||
718 | index 8f3d44e5e7878..722b1421d8f39 100644 | ||
719 | --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c | ||
720 | +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c | ||
721 | @@ -212,6 +212,7 @@ int amdgpu_driver_load_kms(struct drm_device *dev, unsigned long flags) | ||
722 | } | ||
723 | |||
724 | if (amdgpu_device_is_px(dev)) { | ||
725 | + dev_pm_set_driver_flags(dev->dev, DPM_FLAG_NEVER_SKIP); | ||
726 | pm_runtime_use_autosuspend(dev->dev); | ||
727 | pm_runtime_set_autosuspend_delay(dev->dev, 5000); | ||
728 | pm_runtime_set_active(dev->dev); | ||
729 | diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | ||
730 | index 8c9abaa7601a7..62df4bd0a0fc2 100644 | ||
731 | --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | ||
732 | +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | ||
733 | @@ -637,12 +637,14 @@ void amdgpu_vm_move_to_lru_tail(struct amdgpu_device *adev, | ||
734 | struct ttm_bo_global *glob = adev->mman.bdev.glob; | ||
735 | struct amdgpu_vm_bo_base *bo_base; | ||
736 | |||
737 | +#if 0 | ||
738 | if (vm->bulk_moveable) { | ||
739 | spin_lock(&glob->lru_lock); | ||
740 | ttm_bo_bulk_move_lru_tail(&vm->lru_bulk_move); | ||
741 | spin_unlock(&glob->lru_lock); | ||
742 | return; | ||
743 | } | ||
744 | +#endif | ||
745 | |||
746 | memset(&vm->lru_bulk_move, 0, sizeof(vm->lru_bulk_move)); | ||
747 | |||
748 | diff --git a/drivers/gpu/drm/amd/amdkfd/kfd_topology.c b/drivers/gpu/drm/amd/amdkfd/kfd_topology.c | ||
749 | index e3843c5929edf..fffece5e42c56 100644 | ||
750 | --- a/drivers/gpu/drm/amd/amdkfd/kfd_topology.c | ||
751 | +++ b/drivers/gpu/drm/amd/amdkfd/kfd_topology.c | ||
752 | @@ -1074,8 +1074,6 @@ static uint32_t kfd_generate_gpu_id(struct kfd_dev *gpu) | ||
753 | * the GPU device is not already present in the topology device | ||
754 | * list then return NULL. This means a new topology device has to | ||
755 | * be created for this GPU. | ||
756 | - * TODO: Rather than assiging @gpu to first topology device withtout | ||
757 | - * gpu attached, it will better to have more stringent check. | ||
758 | */ | ||
759 | static struct kfd_topology_device *kfd_assign_gpu(struct kfd_dev *gpu) | ||
760 | { | ||
761 | @@ -1083,12 +1081,20 @@ static struct kfd_topology_device *kfd_assign_gpu(struct kfd_dev *gpu) | ||
762 | struct kfd_topology_device *out_dev = NULL; | ||
763 | |||
764 | down_write(&topology_lock); | ||
765 | - list_for_each_entry(dev, &topology_device_list, list) | ||
766 | + list_for_each_entry(dev, &topology_device_list, list) { | ||
767 | + /* Discrete GPUs need their own topology device list | ||
768 | + * entries. Don't assign them to CPU/APU nodes. | ||
769 | + */ | ||
770 | + if (!gpu->device_info->needs_iommu_device && | ||
771 | + dev->node_props.cpu_cores_count) | ||
772 | + continue; | ||
773 | + | ||
774 | if (!dev->gpu && (dev->node_props.simd_count > 0)) { | ||
775 | dev->gpu = gpu; | ||
776 | out_dev = dev; | ||
777 | break; | ||
778 | } | ||
779 | + } | ||
780 | up_write(&topology_lock); | ||
781 | return out_dev; | ||
782 | } | ||
783 | diff --git a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c | ||
784 | index 315a245aedc29..d92120b62e89f 100644 | ||
785 | --- a/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c | ||
786 | +++ b/drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c | ||
787 | @@ -704,12 +704,13 @@ static int dm_suspend(void *handle) | ||
788 | struct amdgpu_display_manager *dm = &adev->dm; | ||
789 | int ret = 0; | ||
790 | |||
791 | + WARN_ON(adev->dm.cached_state); | ||
792 | + adev->dm.cached_state = drm_atomic_helper_suspend(adev->ddev); | ||
793 | + | ||
794 | s3_handle_mst(adev->ddev, true); | ||
795 | |||
796 | amdgpu_dm_irq_suspend(adev); | ||
797 | |||
798 | - WARN_ON(adev->dm.cached_state); | ||
799 | - adev->dm.cached_state = drm_atomic_helper_suspend(adev->ddev); | ||
800 | |||
801 | dc_set_power_state(dm->dc, DC_ACPI_CM_POWER_STATE_D3); | ||
802 | |||
803 | diff --git a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c | ||
804 | index 4443a916a0fb6..e84275f15e7ad 100644 | ||
805 | --- a/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c | ||
806 | +++ b/drivers/gpu/drm/amd/display/dc/dce110/dce110_hw_sequencer.c | ||
807 | @@ -1000,7 +1000,7 @@ void dce110_enable_audio_stream(struct pipe_ctx *pipe_ctx) | ||
808 | |||
809 | pipe_ctx->stream_res.audio->funcs->az_enable(pipe_ctx->stream_res.audio); | ||
810 | |||
811 | - if (num_audio == 1 && pp_smu != NULL && pp_smu->set_pme_wa_enable != NULL) | ||
812 | + if (num_audio >= 1 && pp_smu != NULL && pp_smu->set_pme_wa_enable != NULL) | ||
813 | /*this is the first audio. apply the PME w/a in order to wake AZ from D3*/ | ||
814 | pp_smu->set_pme_wa_enable(&pp_smu->pp_smu); | ||
815 | /* un-mute audio */ | ||
816 | @@ -1017,6 +1017,8 @@ void dce110_disable_audio_stream(struct pipe_ctx *pipe_ctx, int option) | ||
817 | pipe_ctx->stream_res.stream_enc->funcs->audio_mute_control( | ||
818 | pipe_ctx->stream_res.stream_enc, true); | ||
819 | if (pipe_ctx->stream_res.audio) { | ||
820 | + struct pp_smu_funcs_rv *pp_smu = dc->res_pool->pp_smu; | ||
821 | + | ||
822 | if (option != KEEP_ACQUIRED_RESOURCE || | ||
823 | !dc->debug.az_endpoint_mute_only) { | ||
824 | /*only disalbe az_endpoint if power down or free*/ | ||
825 | @@ -1036,6 +1038,9 @@ void dce110_disable_audio_stream(struct pipe_ctx *pipe_ctx, int option) | ||
826 | update_audio_usage(&dc->current_state->res_ctx, dc->res_pool, pipe_ctx->stream_res.audio, false); | ||
827 | pipe_ctx->stream_res.audio = NULL; | ||
828 | } | ||
829 | + if (pp_smu != NULL && pp_smu->set_pme_wa_enable != NULL) | ||
830 | + /*this is the first audio. apply the PME w/a in order to wake AZ from D3*/ | ||
831 | + pp_smu->set_pme_wa_enable(&pp_smu->pp_smu); | ||
832 | |||
833 | /* TODO: notify audio driver for if audio modes list changed | ||
834 | * add audio mode list change flag */ | ||
835 | diff --git a/drivers/gpu/drm/i915/intel_fbdev.c b/drivers/gpu/drm/i915/intel_fbdev.c | ||
836 | index 9e2e998b198f0..e0c02a9889b2c 100644 | ||
837 | --- a/drivers/gpu/drm/i915/intel_fbdev.c | ||
838 | +++ b/drivers/gpu/drm/i915/intel_fbdev.c | ||
839 | @@ -336,8 +336,8 @@ static bool intel_fb_initial_config(struct drm_fb_helper *fb_helper, | ||
840 | bool *enabled, int width, int height) | ||
841 | { | ||
842 | struct drm_i915_private *dev_priv = to_i915(fb_helper->dev); | ||
843 | - unsigned long conn_configured, conn_seq, mask; | ||
844 | unsigned int count = min(fb_helper->connector_count, BITS_PER_LONG); | ||
845 | + unsigned long conn_configured, conn_seq; | ||
846 | int i, j; | ||
847 | bool *save_enabled; | ||
848 | bool fallback = true, ret = true; | ||
849 | @@ -355,10 +355,9 @@ static bool intel_fb_initial_config(struct drm_fb_helper *fb_helper, | ||
850 | drm_modeset_backoff(&ctx); | ||
851 | |||
852 | memcpy(save_enabled, enabled, count); | ||
853 | - mask = GENMASK(count - 1, 0); | ||
854 | + conn_seq = GENMASK(count - 1, 0); | ||
855 | conn_configured = 0; | ||
856 | retry: | ||
857 | - conn_seq = conn_configured; | ||
858 | for (i = 0; i < count; i++) { | ||
859 | struct drm_fb_helper_connector *fb_conn; | ||
860 | struct drm_connector *connector; | ||
861 | @@ -371,7 +370,8 @@ retry: | ||
862 | if (conn_configured & BIT(i)) | ||
863 | continue; | ||
864 | |||
865 | - if (conn_seq == 0 && !connector->has_tile) | ||
866 | + /* First pass, only consider tiled connectors */ | ||
867 | + if (conn_seq == GENMASK(count - 1, 0) && !connector->has_tile) | ||
868 | continue; | ||
869 | |||
870 | if (connector->status == connector_status_connected) | ||
871 | @@ -475,8 +475,10 @@ retry: | ||
872 | conn_configured |= BIT(i); | ||
873 | } | ||
874 | |||
875 | - if ((conn_configured & mask) != mask && conn_configured != conn_seq) | ||
876 | + if (conn_configured != conn_seq) { /* repeat until no more are found */ | ||
877 | + conn_seq = conn_configured; | ||
878 | goto retry; | ||
879 | + } | ||
880 | |||
881 | /* | ||
882 | * If the BIOS didn't enable everything it could, fall back to have the | ||
883 | diff --git a/drivers/gpu/drm/meson/meson_drv.c b/drivers/gpu/drm/meson/meson_drv.c | ||
884 | index bf5f294f172fa..611ac340fb289 100644 | ||
885 | --- a/drivers/gpu/drm/meson/meson_drv.c | ||
886 | +++ b/drivers/gpu/drm/meson/meson_drv.c | ||
887 | @@ -368,8 +368,10 @@ static int meson_probe_remote(struct platform_device *pdev, | ||
888 | remote_node = of_graph_get_remote_port_parent(ep); | ||
889 | if (!remote_node || | ||
890 | remote_node == parent || /* Ignore parent endpoint */ | ||
891 | - !of_device_is_available(remote_node)) | ||
892 | + !of_device_is_available(remote_node)) { | ||
893 | + of_node_put(remote_node); | ||
894 | continue; | ||
895 | + } | ||
896 | |||
897 | count += meson_probe_remote(pdev, match, remote, remote_node); | ||
898 | |||
899 | @@ -388,10 +390,13 @@ static int meson_drv_probe(struct platform_device *pdev) | ||
900 | |||
901 | for_each_endpoint_of_node(np, ep) { | ||
902 | remote = of_graph_get_remote_port_parent(ep); | ||
903 | - if (!remote || !of_device_is_available(remote)) | ||
904 | + if (!remote || !of_device_is_available(remote)) { | ||
905 | + of_node_put(remote); | ||
906 | continue; | ||
907 | + } | ||
908 | |||
909 | count += meson_probe_remote(pdev, &match, np, remote); | ||
910 | + of_node_put(remote); | ||
911 | } | ||
912 | |||
913 | if (count && !match) | ||
914 | diff --git a/drivers/gpu/drm/radeon/radeon_kms.c b/drivers/gpu/drm/radeon/radeon_kms.c | ||
915 | index dec1e081f5295..6a8fb6fd183c3 100644 | ||
916 | --- a/drivers/gpu/drm/radeon/radeon_kms.c | ||
917 | +++ b/drivers/gpu/drm/radeon/radeon_kms.c | ||
918 | @@ -172,6 +172,7 @@ int radeon_driver_load_kms(struct drm_device *dev, unsigned long flags) | ||
919 | } | ||
920 | |||
921 | if (radeon_is_px(dev)) { | ||
922 | + dev_pm_set_driver_flags(dev->dev, DPM_FLAG_NEVER_SKIP); | ||
923 | pm_runtime_use_autosuspend(dev->dev); | ||
924 | pm_runtime_set_autosuspend_delay(dev->dev, 5000); | ||
925 | pm_runtime_set_active(dev->dev); | ||
926 | diff --git a/drivers/gpu/drm/sun4i/sun4i_backend.c b/drivers/gpu/drm/sun4i/sun4i_backend.c | ||
927 | index bf49c55b0f2c7..9f27d5464804b 100644 | ||
928 | --- a/drivers/gpu/drm/sun4i/sun4i_backend.c | ||
929 | +++ b/drivers/gpu/drm/sun4i/sun4i_backend.c | ||
930 | @@ -704,17 +704,18 @@ static struct sun4i_frontend *sun4i_backend_find_frontend(struct sun4i_drv *drv, | ||
931 | remote = of_graph_get_remote_port_parent(ep); | ||
932 | if (!remote) | ||
933 | continue; | ||
934 | + of_node_put(remote); | ||
935 | |||
936 | /* does this node match any registered engines? */ | ||
937 | list_for_each_entry(frontend, &drv->frontend_list, list) { | ||
938 | if (remote == frontend->node) { | ||
939 | - of_node_put(remote); | ||
940 | of_node_put(port); | ||
941 | + of_node_put(ep); | ||
942 | return frontend; | ||
943 | } | ||
944 | } | ||
945 | } | ||
946 | - | ||
947 | + of_node_put(port); | ||
948 | return ERR_PTR(-EINVAL); | ||
949 | } | ||
950 | |||
951 | diff --git a/drivers/hwmon/nct6775.c b/drivers/hwmon/nct6775.c | ||
952 | index c3040079b1cb6..4adec4ab7d066 100644 | ||
953 | --- a/drivers/hwmon/nct6775.c | ||
954 | +++ b/drivers/hwmon/nct6775.c | ||
955 | @@ -44,8 +44,8 @@ | ||
956 | * nct6796d 14 7 7 2+6 0xd420 0xc1 0x5ca3 | ||
957 | * nct6797d 14 7 7 2+6 0xd450 0xc1 0x5ca3 | ||
958 | * (0xd451) | ||
959 | - * nct6798d 14 7 7 2+6 0xd458 0xc1 0x5ca3 | ||
960 | - * (0xd459) | ||
961 | + * nct6798d 14 7 7 2+6 0xd428 0xc1 0x5ca3 | ||
962 | + * (0xd429) | ||
963 | * | ||
964 | * #temp lists the number of monitored temperature sources (first value) plus | ||
965 | * the number of directly connectable temperature sensors (second value). | ||
966 | @@ -138,7 +138,7 @@ MODULE_PARM_DESC(fan_debounce, "Enable debouncing for fan RPM signal"); | ||
967 | #define SIO_NCT6795_ID 0xd350 | ||
968 | #define SIO_NCT6796_ID 0xd420 | ||
969 | #define SIO_NCT6797_ID 0xd450 | ||
970 | -#define SIO_NCT6798_ID 0xd458 | ||
971 | +#define SIO_NCT6798_ID 0xd428 | ||
972 | #define SIO_ID_MASK 0xFFF8 | ||
973 | |||
974 | enum pwm_enable { off, manual, thermal_cruise, speed_cruise, sf3, sf4 }; | ||
975 | @@ -4508,7 +4508,8 @@ static int __maybe_unused nct6775_resume(struct device *dev) | ||
976 | |||
977 | if (data->kind == nct6791 || data->kind == nct6792 || | ||
978 | data->kind == nct6793 || data->kind == nct6795 || | ||
979 | - data->kind == nct6796) | ||
980 | + data->kind == nct6796 || data->kind == nct6797 || | ||
981 | + data->kind == nct6798) | ||
982 | nct6791_enable_io_mapping(sioreg); | ||
983 | |||
984 | superio_exit(sioreg); | ||
985 | @@ -4644,7 +4645,8 @@ static int __init nct6775_find(int sioaddr, struct nct6775_sio_data *sio_data) | ||
986 | |||
987 | if (sio_data->kind == nct6791 || sio_data->kind == nct6792 || | ||
988 | sio_data->kind == nct6793 || sio_data->kind == nct6795 || | ||
989 | - sio_data->kind == nct6796) | ||
990 | + sio_data->kind == nct6796 || sio_data->kind == nct6797 || | ||
991 | + sio_data->kind == nct6798) | ||
992 | nct6791_enable_io_mapping(sioaddr); | ||
993 | |||
994 | superio_exit(sioaddr); | ||
995 | diff --git a/drivers/hwmon/tmp421.c b/drivers/hwmon/tmp421.c | ||
996 | index 8844c9565d2a4..7053be59ad2e4 100644 | ||
997 | --- a/drivers/hwmon/tmp421.c | ||
998 | +++ b/drivers/hwmon/tmp421.c | ||
999 | @@ -88,7 +88,7 @@ static const struct of_device_id tmp421_of_match[] = { | ||
1000 | .data = (void *)2 | ||
1001 | }, | ||
1002 | { | ||
1003 | - .compatible = "ti,tmp422", | ||
1004 | + .compatible = "ti,tmp442", | ||
1005 | .data = (void *)3 | ||
1006 | }, | ||
1007 | { }, | ||
1008 | diff --git a/drivers/infiniband/hw/mthca/mthca_provider.c b/drivers/infiniband/hw/mthca/mthca_provider.c | ||
1009 | index 691c6f0489386..2428c7d89c6be 100644 | ||
1010 | --- a/drivers/infiniband/hw/mthca/mthca_provider.c | ||
1011 | +++ b/drivers/infiniband/hw/mthca/mthca_provider.c | ||
1012 | @@ -533,7 +533,7 @@ static struct ib_qp *mthca_create_qp(struct ib_pd *pd, | ||
1013 | { | ||
1014 | struct mthca_ucontext *context; | ||
1015 | |||
1016 | - qp = kmalloc(sizeof *qp, GFP_KERNEL); | ||
1017 | + qp = kzalloc(sizeof(*qp), GFP_KERNEL); | ||
1018 | if (!qp) | ||
1019 | return ERR_PTR(-ENOMEM); | ||
1020 | |||
1021 | @@ -599,7 +599,7 @@ static struct ib_qp *mthca_create_qp(struct ib_pd *pd, | ||
1022 | if (pd->uobject) | ||
1023 | return ERR_PTR(-EINVAL); | ||
1024 | |||
1025 | - qp = kmalloc(sizeof (struct mthca_sqp), GFP_KERNEL); | ||
1026 | + qp = kzalloc(sizeof(struct mthca_sqp), GFP_KERNEL); | ||
1027 | if (!qp) | ||
1028 | return ERR_PTR(-ENOMEM); | ||
1029 | |||
1030 | diff --git a/drivers/infiniband/ulp/srp/ib_srp.c b/drivers/infiniband/ulp/srp/ib_srp.c | ||
1031 | index eed0eb3bb04c6..0466f2ac9ad08 100644 | ||
1032 | --- a/drivers/infiniband/ulp/srp/ib_srp.c | ||
1033 | +++ b/drivers/infiniband/ulp/srp/ib_srp.c | ||
1034 | @@ -2942,7 +2942,6 @@ static int srp_reset_device(struct scsi_cmnd *scmnd) | ||
1035 | { | ||
1036 | struct srp_target_port *target = host_to_target(scmnd->device->host); | ||
1037 | struct srp_rdma_ch *ch; | ||
1038 | - int i, j; | ||
1039 | u8 status; | ||
1040 | |||
1041 | shost_printk(KERN_ERR, target->scsi_host, "SRP reset_device called\n"); | ||
1042 | @@ -2954,15 +2953,6 @@ static int srp_reset_device(struct scsi_cmnd *scmnd) | ||
1043 | if (status) | ||
1044 | return FAILED; | ||
1045 | |||
1046 | - for (i = 0; i < target->ch_count; i++) { | ||
1047 | - ch = &target->ch[i]; | ||
1048 | - for (j = 0; j < target->req_ring_size; ++j) { | ||
1049 | - struct srp_request *req = &ch->req_ring[j]; | ||
1050 | - | ||
1051 | - srp_finish_req(ch, req, scmnd->device, DID_RESET << 16); | ||
1052 | - } | ||
1053 | - } | ||
1054 | - | ||
1055 | return SUCCESS; | ||
1056 | } | ||
1057 | |||
1058 | diff --git a/drivers/isdn/hardware/avm/b1.c b/drivers/isdn/hardware/avm/b1.c | ||
1059 | index 4ac378e489023..40ca1e8fa09fc 100644 | ||
1060 | --- a/drivers/isdn/hardware/avm/b1.c | ||
1061 | +++ b/drivers/isdn/hardware/avm/b1.c | ||
1062 | @@ -423,7 +423,7 @@ void b1_parse_version(avmctrl_info *cinfo) | ||
1063 | int i, j; | ||
1064 | |||
1065 | for (j = 0; j < AVM_MAXVERSION; j++) | ||
1066 | - cinfo->version[j] = "\0\0" + 1; | ||
1067 | + cinfo->version[j] = ""; | ||
1068 | for (i = 0, j = 0; | ||
1069 | j < AVM_MAXVERSION && i < cinfo->versionlen; | ||
1070 | j++, i += cinfo->versionbuf[i] + 1) | ||
1071 | diff --git a/drivers/isdn/i4l/isdn_tty.c b/drivers/isdn/i4l/isdn_tty.c | ||
1072 | index 1b2239c1d5694..dc1cded716c1a 100644 | ||
1073 | --- a/drivers/isdn/i4l/isdn_tty.c | ||
1074 | +++ b/drivers/isdn/i4l/isdn_tty.c | ||
1075 | @@ -1437,15 +1437,19 @@ isdn_tty_set_termios(struct tty_struct *tty, struct ktermios *old_termios) | ||
1076 | { | ||
1077 | modem_info *info = (modem_info *) tty->driver_data; | ||
1078 | |||
1079 | + mutex_lock(&modem_info_mutex); | ||
1080 | if (!old_termios) | ||
1081 | isdn_tty_change_speed(info); | ||
1082 | else { | ||
1083 | if (tty->termios.c_cflag == old_termios->c_cflag && | ||
1084 | tty->termios.c_ispeed == old_termios->c_ispeed && | ||
1085 | - tty->termios.c_ospeed == old_termios->c_ospeed) | ||
1086 | + tty->termios.c_ospeed == old_termios->c_ospeed) { | ||
1087 | + mutex_unlock(&modem_info_mutex); | ||
1088 | return; | ||
1089 | + } | ||
1090 | isdn_tty_change_speed(info); | ||
1091 | } | ||
1092 | + mutex_unlock(&modem_info_mutex); | ||
1093 | } | ||
1094 | |||
1095 | /* | ||
1096 | diff --git a/drivers/leds/leds-lp5523.c b/drivers/leds/leds-lp5523.c | ||
1097 | index a2e74feee2b2f..fd64df5a57a5e 100644 | ||
1098 | --- a/drivers/leds/leds-lp5523.c | ||
1099 | +++ b/drivers/leds/leds-lp5523.c | ||
1100 | @@ -318,7 +318,9 @@ static int lp5523_init_program_engine(struct lp55xx_chip *chip) | ||
1101 | |||
1102 | /* Let the programs run for couple of ms and check the engine status */ | ||
1103 | usleep_range(3000, 6000); | ||
1104 | - lp55xx_read(chip, LP5523_REG_STATUS, &status); | ||
1105 | + ret = lp55xx_read(chip, LP5523_REG_STATUS, &status); | ||
1106 | + if (ret) | ||
1107 | + return ret; | ||
1108 | status &= LP5523_ENG_STATUS_MASK; | ||
1109 | |||
1110 | if (status != LP5523_ENG_STATUS_MASK) { | ||
1111 | diff --git a/drivers/mfd/Kconfig b/drivers/mfd/Kconfig | ||
1112 | index 8c5dfdce4326c..f461460a2aeb9 100644 | ||
1113 | --- a/drivers/mfd/Kconfig | ||
1114 | +++ b/drivers/mfd/Kconfig | ||
1115 | @@ -102,6 +102,7 @@ config MFD_AAT2870_CORE | ||
1116 | config MFD_AT91_USART | ||
1117 | tristate "AT91 USART Driver" | ||
1118 | select MFD_CORE | ||
1119 | + depends on ARCH_AT91 || COMPILE_TEST | ||
1120 | help | ||
1121 | Select this to get support for AT91 USART IP. This is a wrapper | ||
1122 | over at91-usart-serial driver and usart-spi-driver. Only one function | ||
1123 | diff --git a/drivers/mfd/ab8500-core.c b/drivers/mfd/ab8500-core.c | ||
1124 | index 30d09d1771717..11ab17f64c649 100644 | ||
1125 | --- a/drivers/mfd/ab8500-core.c | ||
1126 | +++ b/drivers/mfd/ab8500-core.c | ||
1127 | @@ -261,7 +261,7 @@ static int get_register_interruptible(struct ab8500 *ab8500, u8 bank, | ||
1128 | mutex_unlock(&ab8500->lock); | ||
1129 | dev_vdbg(ab8500->dev, "rd: addr %#x => data %#x\n", addr, ret); | ||
1130 | |||
1131 | - return ret; | ||
1132 | + return (ret < 0) ? ret : 0; | ||
1133 | } | ||
1134 | |||
1135 | static int ab8500_get_register(struct device *dev, u8 bank, | ||
1136 | diff --git a/drivers/mfd/axp20x.c b/drivers/mfd/axp20x.c | ||
1137 | index 0be511dd93d01..f8e0fa97bb31e 100644 | ||
1138 | --- a/drivers/mfd/axp20x.c | ||
1139 | +++ b/drivers/mfd/axp20x.c | ||
1140 | @@ -640,9 +640,9 @@ static const struct mfd_cell axp221_cells[] = { | ||
1141 | |||
1142 | static const struct mfd_cell axp223_cells[] = { | ||
1143 | { | ||
1144 | - .name = "axp221-pek", | ||
1145 | - .num_resources = ARRAY_SIZE(axp22x_pek_resources), | ||
1146 | - .resources = axp22x_pek_resources, | ||
1147 | + .name = "axp221-pek", | ||
1148 | + .num_resources = ARRAY_SIZE(axp22x_pek_resources), | ||
1149 | + .resources = axp22x_pek_resources, | ||
1150 | }, { | ||
1151 | .name = "axp22x-adc", | ||
1152 | .of_compatible = "x-powers,axp221-adc", | ||
1153 | @@ -650,7 +650,7 @@ static const struct mfd_cell axp223_cells[] = { | ||
1154 | .name = "axp20x-battery-power-supply", | ||
1155 | .of_compatible = "x-powers,axp221-battery-power-supply", | ||
1156 | }, { | ||
1157 | - .name = "axp20x-regulator", | ||
1158 | + .name = "axp20x-regulator", | ||
1159 | }, { | ||
1160 | .name = "axp20x-ac-power-supply", | ||
1161 | .of_compatible = "x-powers,axp221-ac-power-supply", | ||
1162 | @@ -666,9 +666,9 @@ static const struct mfd_cell axp223_cells[] = { | ||
1163 | |||
1164 | static const struct mfd_cell axp152_cells[] = { | ||
1165 | { | ||
1166 | - .name = "axp20x-pek", | ||
1167 | - .num_resources = ARRAY_SIZE(axp152_pek_resources), | ||
1168 | - .resources = axp152_pek_resources, | ||
1169 | + .name = "axp20x-pek", | ||
1170 | + .num_resources = ARRAY_SIZE(axp152_pek_resources), | ||
1171 | + .resources = axp152_pek_resources, | ||
1172 | }, | ||
1173 | }; | ||
1174 | |||
1175 | @@ -697,87 +697,101 @@ static const struct resource axp288_charger_resources[] = { | ||
1176 | |||
1177 | static const struct mfd_cell axp288_cells[] = { | ||
1178 | { | ||
1179 | - .name = "axp288_adc", | ||
1180 | - .num_resources = ARRAY_SIZE(axp288_adc_resources), | ||
1181 | - .resources = axp288_adc_resources, | ||
1182 | - }, | ||
1183 | - { | ||
1184 | - .name = "axp288_extcon", | ||
1185 | - .num_resources = ARRAY_SIZE(axp288_extcon_resources), | ||
1186 | - .resources = axp288_extcon_resources, | ||
1187 | - }, | ||
1188 | - { | ||
1189 | - .name = "axp288_charger", | ||
1190 | - .num_resources = ARRAY_SIZE(axp288_charger_resources), | ||
1191 | - .resources = axp288_charger_resources, | ||
1192 | - }, | ||
1193 | - { | ||
1194 | - .name = "axp288_fuel_gauge", | ||
1195 | - .num_resources = ARRAY_SIZE(axp288_fuel_gauge_resources), | ||
1196 | - .resources = axp288_fuel_gauge_resources, | ||
1197 | - }, | ||
1198 | - { | ||
1199 | - .name = "axp221-pek", | ||
1200 | - .num_resources = ARRAY_SIZE(axp288_power_button_resources), | ||
1201 | - .resources = axp288_power_button_resources, | ||
1202 | - }, | ||
1203 | - { | ||
1204 | - .name = "axp288_pmic_acpi", | ||
1205 | + .name = "axp288_adc", | ||
1206 | + .num_resources = ARRAY_SIZE(axp288_adc_resources), | ||
1207 | + .resources = axp288_adc_resources, | ||
1208 | + }, { | ||
1209 | + .name = "axp288_extcon", | ||
1210 | + .num_resources = ARRAY_SIZE(axp288_extcon_resources), | ||
1211 | + .resources = axp288_extcon_resources, | ||
1212 | + }, { | ||
1213 | + .name = "axp288_charger", | ||
1214 | + .num_resources = ARRAY_SIZE(axp288_charger_resources), | ||
1215 | + .resources = axp288_charger_resources, | ||
1216 | + }, { | ||
1217 | + .name = "axp288_fuel_gauge", | ||
1218 | + .num_resources = ARRAY_SIZE(axp288_fuel_gauge_resources), | ||
1219 | + .resources = axp288_fuel_gauge_resources, | ||
1220 | + }, { | ||
1221 | + .name = "axp221-pek", | ||
1222 | + .num_resources = ARRAY_SIZE(axp288_power_button_resources), | ||
1223 | + .resources = axp288_power_button_resources, | ||
1224 | + }, { | ||
1225 | + .name = "axp288_pmic_acpi", | ||
1226 | }, | ||
1227 | }; | ||
1228 | |||
1229 | static const struct mfd_cell axp803_cells[] = { | ||
1230 | { | ||
1231 | - .name = "axp221-pek", | ||
1232 | - .num_resources = ARRAY_SIZE(axp803_pek_resources), | ||
1233 | - .resources = axp803_pek_resources, | ||
1234 | + .name = "axp221-pek", | ||
1235 | + .num_resources = ARRAY_SIZE(axp803_pek_resources), | ||
1236 | + .resources = axp803_pek_resources, | ||
1237 | + }, { | ||
1238 | + .name = "axp20x-gpio", | ||
1239 | + .of_compatible = "x-powers,axp813-gpio", | ||
1240 | + }, { | ||
1241 | + .name = "axp813-adc", | ||
1242 | + .of_compatible = "x-powers,axp813-adc", | ||
1243 | + }, { | ||
1244 | + .name = "axp20x-battery-power-supply", | ||
1245 | + .of_compatible = "x-powers,axp813-battery-power-supply", | ||
1246 | + }, { | ||
1247 | + .name = "axp20x-ac-power-supply", | ||
1248 | + .of_compatible = "x-powers,axp813-ac-power-supply", | ||
1249 | + .num_resources = ARRAY_SIZE(axp20x_ac_power_supply_resources), | ||
1250 | + .resources = axp20x_ac_power_supply_resources, | ||
1251 | }, | ||
1252 | - { .name = "axp20x-regulator" }, | ||
1253 | + { .name = "axp20x-regulator" }, | ||
1254 | }; | ||
1255 | |||
1256 | static const struct mfd_cell axp806_self_working_cells[] = { | ||
1257 | { | ||
1258 | - .name = "axp221-pek", | ||
1259 | - .num_resources = ARRAY_SIZE(axp806_pek_resources), | ||
1260 | - .resources = axp806_pek_resources, | ||
1261 | + .name = "axp221-pek", | ||
1262 | + .num_resources = ARRAY_SIZE(axp806_pek_resources), | ||
1263 | + .resources = axp806_pek_resources, | ||
1264 | }, | ||
1265 | - { .name = "axp20x-regulator" }, | ||
1266 | + { .name = "axp20x-regulator" }, | ||
1267 | }; | ||
1268 | |||
1269 | static const struct mfd_cell axp806_cells[] = { | ||
1270 | { | ||
1271 | - .id = 2, | ||
1272 | - .name = "axp20x-regulator", | ||
1273 | + .id = 2, | ||
1274 | + .name = "axp20x-regulator", | ||
1275 | }, | ||
1276 | }; | ||
1277 | |||
1278 | static const struct mfd_cell axp809_cells[] = { | ||
1279 | { | ||
1280 | - .name = "axp221-pek", | ||
1281 | - .num_resources = ARRAY_SIZE(axp809_pek_resources), | ||
1282 | - .resources = axp809_pek_resources, | ||
1283 | + .name = "axp221-pek", | ||
1284 | + .num_resources = ARRAY_SIZE(axp809_pek_resources), | ||
1285 | + .resources = axp809_pek_resources, | ||
1286 | }, { | ||
1287 | - .id = 1, | ||
1288 | - .name = "axp20x-regulator", | ||
1289 | + .id = 1, | ||
1290 | + .name = "axp20x-regulator", | ||
1291 | }, | ||
1292 | }; | ||
1293 | |||
1294 | static const struct mfd_cell axp813_cells[] = { | ||
1295 | { | ||
1296 | - .name = "axp221-pek", | ||
1297 | - .num_resources = ARRAY_SIZE(axp803_pek_resources), | ||
1298 | - .resources = axp803_pek_resources, | ||
1299 | + .name = "axp221-pek", | ||
1300 | + .num_resources = ARRAY_SIZE(axp803_pek_resources), | ||
1301 | + .resources = axp803_pek_resources, | ||
1302 | }, { | ||
1303 | - .name = "axp20x-regulator", | ||
1304 | + .name = "axp20x-regulator", | ||
1305 | }, { | ||
1306 | - .name = "axp20x-gpio", | ||
1307 | - .of_compatible = "x-powers,axp813-gpio", | ||
1308 | + .name = "axp20x-gpio", | ||
1309 | + .of_compatible = "x-powers,axp813-gpio", | ||
1310 | }, { | ||
1311 | - .name = "axp813-adc", | ||
1312 | - .of_compatible = "x-powers,axp813-adc", | ||
1313 | + .name = "axp813-adc", | ||
1314 | + .of_compatible = "x-powers,axp813-adc", | ||
1315 | }, { | ||
1316 | .name = "axp20x-battery-power-supply", | ||
1317 | .of_compatible = "x-powers,axp813-battery-power-supply", | ||
1318 | + }, { | ||
1319 | + .name = "axp20x-ac-power-supply", | ||
1320 | + .of_compatible = "x-powers,axp813-ac-power-supply", | ||
1321 | + .num_resources = ARRAY_SIZE(axp20x_ac_power_supply_resources), | ||
1322 | + .resources = axp20x_ac_power_supply_resources, | ||
1323 | }, | ||
1324 | }; | ||
1325 | |||
1326 | diff --git a/drivers/mfd/bd9571mwv.c b/drivers/mfd/bd9571mwv.c | ||
1327 | index 503979c81dae1..fab3cdc27ed64 100644 | ||
1328 | --- a/drivers/mfd/bd9571mwv.c | ||
1329 | +++ b/drivers/mfd/bd9571mwv.c | ||
1330 | @@ -59,6 +59,7 @@ static const struct regmap_access_table bd9571mwv_writable_table = { | ||
1331 | }; | ||
1332 | |||
1333 | static const struct regmap_range bd9571mwv_volatile_yes_ranges[] = { | ||
1334 | + regmap_reg_range(BD9571MWV_DVFS_MONIVDAC, BD9571MWV_DVFS_MONIVDAC), | ||
1335 | regmap_reg_range(BD9571MWV_GPIO_IN, BD9571MWV_GPIO_IN), | ||
1336 | regmap_reg_range(BD9571MWV_GPIO_INT, BD9571MWV_GPIO_INT), | ||
1337 | regmap_reg_range(BD9571MWV_INT_INTREQ, BD9571MWV_INT_INTREQ), | ||
1338 | diff --git a/drivers/mfd/cros_ec_dev.c b/drivers/mfd/cros_ec_dev.c | ||
1339 | index b99a194ce5a4a..2d0fee488c5aa 100644 | ||
1340 | --- a/drivers/mfd/cros_ec_dev.c | ||
1341 | +++ b/drivers/mfd/cros_ec_dev.c | ||
1342 | @@ -499,6 +499,7 @@ static int ec_device_remove(struct platform_device *pdev) | ||
1343 | |||
1344 | cros_ec_debugfs_remove(ec); | ||
1345 | |||
1346 | + mfd_remove_devices(ec->dev); | ||
1347 | cdev_del(&ec->cdev); | ||
1348 | device_unregister(&ec->class_dev); | ||
1349 | return 0; | ||
1350 | diff --git a/drivers/mfd/db8500-prcmu.c b/drivers/mfd/db8500-prcmu.c | ||
1351 | index 5970b8def5487..aec20e1c7d3d5 100644 | ||
1352 | --- a/drivers/mfd/db8500-prcmu.c | ||
1353 | +++ b/drivers/mfd/db8500-prcmu.c | ||
1354 | @@ -2584,7 +2584,7 @@ static struct irq_chip prcmu_irq_chip = { | ||
1355 | .irq_unmask = prcmu_irq_unmask, | ||
1356 | }; | ||
1357 | |||
1358 | -static __init char *fw_project_name(u32 project) | ||
1359 | +static char *fw_project_name(u32 project) | ||
1360 | { | ||
1361 | switch (project) { | ||
1362 | case PRCMU_FW_PROJECT_U8500: | ||
1363 | @@ -2732,7 +2732,7 @@ void __init db8500_prcmu_early_init(u32 phy_base, u32 size) | ||
1364 | INIT_WORK(&mb0_transfer.mask_work, prcmu_mask_work); | ||
1365 | } | ||
1366 | |||
1367 | -static void __init init_prcm_registers(void) | ||
1368 | +static void init_prcm_registers(void) | ||
1369 | { | ||
1370 | u32 val; | ||
1371 | |||
1372 | diff --git a/drivers/mfd/mc13xxx-core.c b/drivers/mfd/mc13xxx-core.c | ||
1373 | index f475e848252fa..d0bf50e3568d7 100644 | ||
1374 | --- a/drivers/mfd/mc13xxx-core.c | ||
1375 | +++ b/drivers/mfd/mc13xxx-core.c | ||
1376 | @@ -274,7 +274,9 @@ int mc13xxx_adc_do_conversion(struct mc13xxx *mc13xxx, unsigned int mode, | ||
1377 | |||
1378 | mc13xxx->adcflags |= MC13XXX_ADC_WORKING; | ||
1379 | |||
1380 | - mc13xxx_reg_read(mc13xxx, MC13XXX_ADC0, &old_adc0); | ||
1381 | + ret = mc13xxx_reg_read(mc13xxx, MC13XXX_ADC0, &old_adc0); | ||
1382 | + if (ret) | ||
1383 | + goto out; | ||
1384 | |||
1385 | adc0 = MC13XXX_ADC0_ADINC1 | MC13XXX_ADC0_ADINC2 | | ||
1386 | MC13XXX_ADC0_CHRGRAWDIV; | ||
1387 | diff --git a/drivers/mfd/mt6397-core.c b/drivers/mfd/mt6397-core.c | ||
1388 | index 77b64bd64df36..ab24e176ef448 100644 | ||
1389 | --- a/drivers/mfd/mt6397-core.c | ||
1390 | +++ b/drivers/mfd/mt6397-core.c | ||
1391 | @@ -329,8 +329,7 @@ static int mt6397_probe(struct platform_device *pdev) | ||
1392 | |||
1393 | default: | ||
1394 | dev_err(&pdev->dev, "unsupported chip: %d\n", id); | ||
1395 | - ret = -ENODEV; | ||
1396 | - break; | ||
1397 | + return -ENODEV; | ||
1398 | } | ||
1399 | |||
1400 | if (ret) { | ||
1401 | diff --git a/drivers/mfd/qcom_rpm.c b/drivers/mfd/qcom_rpm.c | ||
1402 | index 52fafea06067e..8d420c37b2a61 100644 | ||
1403 | --- a/drivers/mfd/qcom_rpm.c | ||
1404 | +++ b/drivers/mfd/qcom_rpm.c | ||
1405 | @@ -638,6 +638,10 @@ static int qcom_rpm_probe(struct platform_device *pdev) | ||
1406 | return -EFAULT; | ||
1407 | } | ||
1408 | |||
1409 | + writel(fw_version[0], RPM_CTRL_REG(rpm, 0)); | ||
1410 | + writel(fw_version[1], RPM_CTRL_REG(rpm, 1)); | ||
1411 | + writel(fw_version[2], RPM_CTRL_REG(rpm, 2)); | ||
1412 | + | ||
1413 | dev_info(&pdev->dev, "RPM firmware %u.%u.%u\n", fw_version[0], | ||
1414 | fw_version[1], | ||
1415 | fw_version[2]); | ||
1416 | diff --git a/drivers/mfd/ti_am335x_tscadc.c b/drivers/mfd/ti_am335x_tscadc.c | ||
1417 | index c2d47d78705b8..fd111296b9592 100644 | ||
1418 | --- a/drivers/mfd/ti_am335x_tscadc.c | ||
1419 | +++ b/drivers/mfd/ti_am335x_tscadc.c | ||
1420 | @@ -264,8 +264,9 @@ static int ti_tscadc_probe(struct platform_device *pdev) | ||
1421 | cell->pdata_size = sizeof(tscadc); | ||
1422 | } | ||
1423 | |||
1424 | - err = mfd_add_devices(&pdev->dev, pdev->id, tscadc->cells, | ||
1425 | - tscadc->used_cells, NULL, 0, NULL); | ||
1426 | + err = mfd_add_devices(&pdev->dev, PLATFORM_DEVID_AUTO, | ||
1427 | + tscadc->cells, tscadc->used_cells, NULL, | ||
1428 | + 0, NULL); | ||
1429 | if (err < 0) | ||
1430 | goto err_disable_clk; | ||
1431 | |||
1432 | diff --git a/drivers/mfd/tps65218.c b/drivers/mfd/tps65218.c | ||
1433 | index 910f569ff77c1..8bcdecf494d05 100644 | ||
1434 | --- a/drivers/mfd/tps65218.c | ||
1435 | +++ b/drivers/mfd/tps65218.c | ||
1436 | @@ -235,9 +235,9 @@ static int tps65218_probe(struct i2c_client *client, | ||
1437 | |||
1438 | mutex_init(&tps->tps_lock); | ||
1439 | |||
1440 | - ret = regmap_add_irq_chip(tps->regmap, tps->irq, | ||
1441 | - IRQF_ONESHOT, 0, &tps65218_irq_chip, | ||
1442 | - &tps->irq_data); | ||
1443 | + ret = devm_regmap_add_irq_chip(&client->dev, tps->regmap, tps->irq, | ||
1444 | + IRQF_ONESHOT, 0, &tps65218_irq_chip, | ||
1445 | + &tps->irq_data); | ||
1446 | if (ret < 0) | ||
1447 | return ret; | ||
1448 | |||
1449 | @@ -253,26 +253,9 @@ static int tps65218_probe(struct i2c_client *client, | ||
1450 | ARRAY_SIZE(tps65218_cells), NULL, 0, | ||
1451 | regmap_irq_get_domain(tps->irq_data)); | ||
1452 | |||
1453 | - if (ret < 0) | ||
1454 | - goto err_irq; | ||
1455 | - | ||
1456 | - return 0; | ||
1457 | - | ||
1458 | -err_irq: | ||
1459 | - regmap_del_irq_chip(tps->irq, tps->irq_data); | ||
1460 | - | ||
1461 | return ret; | ||
1462 | } | ||
1463 | |||
1464 | -static int tps65218_remove(struct i2c_client *client) | ||
1465 | -{ | ||
1466 | - struct tps65218 *tps = i2c_get_clientdata(client); | ||
1467 | - | ||
1468 | - regmap_del_irq_chip(tps->irq, tps->irq_data); | ||
1469 | - | ||
1470 | - return 0; | ||
1471 | -} | ||
1472 | - | ||
1473 | static const struct i2c_device_id tps65218_id_table[] = { | ||
1474 | { "tps65218", TPS65218 }, | ||
1475 | { }, | ||
1476 | @@ -285,7 +268,6 @@ static struct i2c_driver tps65218_driver = { | ||
1477 | .of_match_table = of_tps65218_match_table, | ||
1478 | }, | ||
1479 | .probe = tps65218_probe, | ||
1480 | - .remove = tps65218_remove, | ||
1481 | .id_table = tps65218_id_table, | ||
1482 | }; | ||
1483 | |||
1484 | diff --git a/drivers/mfd/twl-core.c b/drivers/mfd/twl-core.c | ||
1485 | index 4be3d239da9ec..299016bc46d90 100644 | ||
1486 | --- a/drivers/mfd/twl-core.c | ||
1487 | +++ b/drivers/mfd/twl-core.c | ||
1488 | @@ -979,7 +979,7 @@ add_children(struct twl4030_platform_data *pdata, unsigned irq_base, | ||
1489 | * letting it generate the right frequencies for USB, MADC, and | ||
1490 | * other purposes. | ||
1491 | */ | ||
1492 | -static inline int __init protect_pm_master(void) | ||
1493 | +static inline int protect_pm_master(void) | ||
1494 | { | ||
1495 | int e = 0; | ||
1496 | |||
1497 | @@ -988,7 +988,7 @@ static inline int __init protect_pm_master(void) | ||
1498 | return e; | ||
1499 | } | ||
1500 | |||
1501 | -static inline int __init unprotect_pm_master(void) | ||
1502 | +static inline int unprotect_pm_master(void) | ||
1503 | { | ||
1504 | int e = 0; | ||
1505 | |||
1506 | diff --git a/drivers/mfd/wm5110-tables.c b/drivers/mfd/wm5110-tables.c | ||
1507 | index 1ee68bd440fbc..16c6e2accfaa5 100644 | ||
1508 | --- a/drivers/mfd/wm5110-tables.c | ||
1509 | +++ b/drivers/mfd/wm5110-tables.c | ||
1510 | @@ -1618,6 +1618,7 @@ static const struct reg_default wm5110_reg_default[] = { | ||
1511 | { 0x00000ECD, 0x0000 }, /* R3789 - HPLPF4_2 */ | ||
1512 | { 0x00000EE0, 0x0000 }, /* R3808 - ASRC_ENABLE */ | ||
1513 | { 0x00000EE2, 0x0000 }, /* R3810 - ASRC_RATE1 */ | ||
1514 | + { 0x00000EE3, 0x4000 }, /* R3811 - ASRC_RATE2 */ | ||
1515 | { 0x00000EF0, 0x0000 }, /* R3824 - ISRC 1 CTRL 1 */ | ||
1516 | { 0x00000EF1, 0x0000 }, /* R3825 - ISRC 1 CTRL 2 */ | ||
1517 | { 0x00000EF2, 0x0000 }, /* R3826 - ISRC 1 CTRL 3 */ | ||
1518 | @@ -2869,6 +2870,7 @@ static bool wm5110_readable_register(struct device *dev, unsigned int reg) | ||
1519 | case ARIZONA_ASRC_ENABLE: | ||
1520 | case ARIZONA_ASRC_STATUS: | ||
1521 | case ARIZONA_ASRC_RATE1: | ||
1522 | + case ARIZONA_ASRC_RATE2: | ||
1523 | case ARIZONA_ISRC_1_CTRL_1: | ||
1524 | case ARIZONA_ISRC_1_CTRL_2: | ||
1525 | case ARIZONA_ISRC_1_CTRL_3: | ||
1526 | diff --git a/drivers/net/ethernet/amazon/ena/ena_netdev.c b/drivers/net/ethernet/amazon/ena/ena_netdev.c | ||
1527 | index a70bb1bb90e7d..a6eacf2099c30 100644 | ||
1528 | --- a/drivers/net/ethernet/amazon/ena/ena_netdev.c | ||
1529 | +++ b/drivers/net/ethernet/amazon/ena/ena_netdev.c | ||
1530 | @@ -2663,11 +2663,6 @@ static int ena_restore_device(struct ena_adapter *adapter) | ||
1531 | goto err_device_destroy; | ||
1532 | } | ||
1533 | |||
1534 | - clear_bit(ENA_FLAG_ONGOING_RESET, &adapter->flags); | ||
1535 | - /* Make sure we don't have a race with AENQ Links state handler */ | ||
1536 | - if (test_bit(ENA_FLAG_LINK_UP, &adapter->flags)) | ||
1537 | - netif_carrier_on(adapter->netdev); | ||
1538 | - | ||
1539 | rc = ena_enable_msix_and_set_admin_interrupts(adapter, | ||
1540 | adapter->num_queues); | ||
1541 | if (rc) { | ||
1542 | @@ -2684,6 +2679,11 @@ static int ena_restore_device(struct ena_adapter *adapter) | ||
1543 | } | ||
1544 | |||
1545 | set_bit(ENA_FLAG_DEVICE_RUNNING, &adapter->flags); | ||
1546 | + | ||
1547 | + clear_bit(ENA_FLAG_ONGOING_RESET, &adapter->flags); | ||
1548 | + if (test_bit(ENA_FLAG_LINK_UP, &adapter->flags)) | ||
1549 | + netif_carrier_on(adapter->netdev); | ||
1550 | + | ||
1551 | mod_timer(&adapter->timer_service, round_jiffies(jiffies + HZ)); | ||
1552 | dev_err(&pdev->dev, | ||
1553 | "Device reset completed successfully, Driver info: %s\n", | ||
1554 | diff --git a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c | ||
1555 | index 6e0f47f2c8a37..3e53be0fcd7ec 100644 | ||
1556 | --- a/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c | ||
1557 | +++ b/drivers/net/ethernet/freescale/dpaa/dpaa_eth.c | ||
1558 | @@ -2051,6 +2051,7 @@ dpaa_start_xmit(struct sk_buff *skb, struct net_device *net_dev) | ||
1559 | bool nonlinear = skb_is_nonlinear(skb); | ||
1560 | struct rtnl_link_stats64 *percpu_stats; | ||
1561 | struct dpaa_percpu_priv *percpu_priv; | ||
1562 | + struct netdev_queue *txq; | ||
1563 | struct dpaa_priv *priv; | ||
1564 | struct qm_fd fd; | ||
1565 | int offset = 0; | ||
1566 | @@ -2100,6 +2101,11 @@ dpaa_start_xmit(struct sk_buff *skb, struct net_device *net_dev) | ||
1567 | if (unlikely(err < 0)) | ||
1568 | goto skb_to_fd_failed; | ||
1569 | |||
1570 | + txq = netdev_get_tx_queue(net_dev, queue_mapping); | ||
1571 | + | ||
1572 | + /* LLTX requires to do our own update of trans_start */ | ||
1573 | + txq->trans_start = jiffies; | ||
1574 | + | ||
1575 | if (priv->tx_tstamp && skb_shinfo(skb)->tx_flags & SKBTX_HW_TSTAMP) { | ||
1576 | fd.cmd |= cpu_to_be32(FM_FD_CMD_UPD); | ||
1577 | skb_shinfo(skb)->tx_flags |= SKBTX_IN_PROGRESS; | ||
1578 | diff --git a/drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c b/drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c | ||
1579 | index ad1779fc410e6..a78bfafd212c8 100644 | ||
1580 | --- a/drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c | ||
1581 | +++ b/drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c | ||
1582 | @@ -147,12 +147,10 @@ static void hns_ae_put_handle(struct hnae_handle *handle) | ||
1583 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); | ||
1584 | int i; | ||
1585 | |||
1586 | - vf_cb->mac_cb = NULL; | ||
1587 | - | ||
1588 | - kfree(vf_cb); | ||
1589 | - | ||
1590 | for (i = 0; i < handle->q_num; i++) | ||
1591 | hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0; | ||
1592 | + | ||
1593 | + kfree(vf_cb); | ||
1594 | } | ||
1595 | |||
1596 | static int hns_ae_wait_flow_down(struct hnae_handle *handle) | ||
1597 | diff --git a/drivers/net/ethernet/mellanox/mlx4/en_rx.c b/drivers/net/ethernet/mellanox/mlx4/en_rx.c | ||
1598 | index db00bf1c23f5a..d47d4f86ac11d 100644 | ||
1599 | --- a/drivers/net/ethernet/mellanox/mlx4/en_rx.c | ||
1600 | +++ b/drivers/net/ethernet/mellanox/mlx4/en_rx.c | ||
1601 | @@ -620,6 +620,8 @@ static int get_fixed_ipv6_csum(__wsum hw_checksum, struct sk_buff *skb, | ||
1602 | } | ||
1603 | #endif | ||
1604 | |||
1605 | +#define short_frame(size) ((size) <= ETH_ZLEN + ETH_FCS_LEN) | ||
1606 | + | ||
1607 | /* We reach this function only after checking that any of | ||
1608 | * the (IPv4 | IPv6) bits are set in cqe->status. | ||
1609 | */ | ||
1610 | @@ -627,9 +629,20 @@ static int check_csum(struct mlx4_cqe *cqe, struct sk_buff *skb, void *va, | ||
1611 | netdev_features_t dev_features) | ||
1612 | { | ||
1613 | __wsum hw_checksum = 0; | ||
1614 | + void *hdr; | ||
1615 | + | ||
1616 | + /* CQE csum doesn't cover padding octets in short ethernet | ||
1617 | + * frames. And the pad field is appended prior to calculating | ||
1618 | + * and appending the FCS field. | ||
1619 | + * | ||
1620 | + * Detecting these padded frames requires to verify and parse | ||
1621 | + * IP headers, so we simply force all those small frames to skip | ||
1622 | + * checksum complete. | ||
1623 | + */ | ||
1624 | + if (short_frame(skb->len)) | ||
1625 | + return -EINVAL; | ||
1626 | |||
1627 | - void *hdr = (u8 *)va + sizeof(struct ethhdr); | ||
1628 | - | ||
1629 | + hdr = (u8 *)va + sizeof(struct ethhdr); | ||
1630 | hw_checksum = csum_unfold((__force __sum16)cqe->checksum); | ||
1631 | |||
1632 | if (cqe->vlan_my_qpn & cpu_to_be32(MLX4_CQE_CVLAN_PRESENT_MASK) && | ||
1633 | @@ -822,6 +835,11 @@ xdp_drop_no_cnt: | ||
1634 | skb_record_rx_queue(skb, cq_ring); | ||
1635 | |||
1636 | if (likely(dev->features & NETIF_F_RXCSUM)) { | ||
1637 | + /* TODO: For IP non TCP/UDP packets when csum complete is | ||
1638 | + * not an option (not supported or any other reason) we can | ||
1639 | + * actually check cqe IPOK status bit and report | ||
1640 | + * CHECKSUM_UNNECESSARY rather than CHECKSUM_NONE | ||
1641 | + */ | ||
1642 | if ((cqe->status & cpu_to_be16(MLX4_CQE_STATUS_TCP | | ||
1643 | MLX4_CQE_STATUS_UDP)) && | ||
1644 | (cqe->status & cpu_to_be16(MLX4_CQE_STATUS_IPOK)) && | ||
1645 | diff --git a/drivers/net/ethernet/mellanox/mlx4/icm.c b/drivers/net/ethernet/mellanox/mlx4/icm.c | ||
1646 | index 4b4351141b94c..76b84d08a058b 100644 | ||
1647 | --- a/drivers/net/ethernet/mellanox/mlx4/icm.c | ||
1648 | +++ b/drivers/net/ethernet/mellanox/mlx4/icm.c | ||
1649 | @@ -57,12 +57,12 @@ static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chu | ||
1650 | int i; | ||
1651 | |||
1652 | if (chunk->nsg > 0) | ||
1653 | - pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, | ||
1654 | + pci_unmap_sg(dev->persist->pdev, chunk->sg, chunk->npages, | ||
1655 | PCI_DMA_BIDIRECTIONAL); | ||
1656 | |||
1657 | for (i = 0; i < chunk->npages; ++i) | ||
1658 | - __free_pages(sg_page(&chunk->mem[i]), | ||
1659 | - get_order(chunk->mem[i].length)); | ||
1660 | + __free_pages(sg_page(&chunk->sg[i]), | ||
1661 | + get_order(chunk->sg[i].length)); | ||
1662 | } | ||
1663 | |||
1664 | static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) | ||
1665 | @@ -71,9 +71,9 @@ static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk * | ||
1666 | |||
1667 | for (i = 0; i < chunk->npages; ++i) | ||
1668 | dma_free_coherent(&dev->persist->pdev->dev, | ||
1669 | - chunk->mem[i].length, | ||
1670 | - lowmem_page_address(sg_page(&chunk->mem[i])), | ||
1671 | - sg_dma_address(&chunk->mem[i])); | ||
1672 | + chunk->buf[i].size, | ||
1673 | + chunk->buf[i].addr, | ||
1674 | + chunk->buf[i].dma_addr); | ||
1675 | } | ||
1676 | |||
1677 | void mlx4_free_icm(struct mlx4_dev *dev, struct mlx4_icm *icm, int coherent) | ||
1678 | @@ -111,22 +111,21 @@ static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, | ||
1679 | return 0; | ||
1680 | } | ||
1681 | |||
1682 | -static int mlx4_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, | ||
1683 | - int order, gfp_t gfp_mask) | ||
1684 | +static int mlx4_alloc_icm_coherent(struct device *dev, struct mlx4_icm_buf *buf, | ||
1685 | + int order, gfp_t gfp_mask) | ||
1686 | { | ||
1687 | - void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, | ||
1688 | - &sg_dma_address(mem), gfp_mask); | ||
1689 | - if (!buf) | ||
1690 | + buf->addr = dma_alloc_coherent(dev, PAGE_SIZE << order, | ||
1691 | + &buf->dma_addr, gfp_mask); | ||
1692 | + if (!buf->addr) | ||
1693 | return -ENOMEM; | ||
1694 | |||
1695 | - if (offset_in_page(buf)) { | ||
1696 | - dma_free_coherent(dev, PAGE_SIZE << order, | ||
1697 | - buf, sg_dma_address(mem)); | ||
1698 | + if (offset_in_page(buf->addr)) { | ||
1699 | + dma_free_coherent(dev, PAGE_SIZE << order, buf->addr, | ||
1700 | + buf->dma_addr); | ||
1701 | return -ENOMEM; | ||
1702 | } | ||
1703 | |||
1704 | - sg_set_buf(mem, buf, PAGE_SIZE << order); | ||
1705 | - sg_dma_len(mem) = PAGE_SIZE << order; | ||
1706 | + buf->size = PAGE_SIZE << order; | ||
1707 | return 0; | ||
1708 | } | ||
1709 | |||
1710 | @@ -159,21 +158,21 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages, | ||
1711 | |||
1712 | while (npages > 0) { | ||
1713 | if (!chunk) { | ||
1714 | - chunk = kmalloc_node(sizeof(*chunk), | ||
1715 | + chunk = kzalloc_node(sizeof(*chunk), | ||
1716 | gfp_mask & ~(__GFP_HIGHMEM | | ||
1717 | __GFP_NOWARN), | ||
1718 | dev->numa_node); | ||
1719 | if (!chunk) { | ||
1720 | - chunk = kmalloc(sizeof(*chunk), | ||
1721 | + chunk = kzalloc(sizeof(*chunk), | ||
1722 | gfp_mask & ~(__GFP_HIGHMEM | | ||
1723 | __GFP_NOWARN)); | ||
1724 | if (!chunk) | ||
1725 | goto fail; | ||
1726 | } | ||
1727 | + chunk->coherent = coherent; | ||
1728 | |||
1729 | - sg_init_table(chunk->mem, MLX4_ICM_CHUNK_LEN); | ||
1730 | - chunk->npages = 0; | ||
1731 | - chunk->nsg = 0; | ||
1732 | + if (!coherent) | ||
1733 | + sg_init_table(chunk->sg, MLX4_ICM_CHUNK_LEN); | ||
1734 | list_add_tail(&chunk->list, &icm->chunk_list); | ||
1735 | } | ||
1736 | |||
1737 | @@ -186,10 +185,10 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages, | ||
1738 | |||
1739 | if (coherent) | ||
1740 | ret = mlx4_alloc_icm_coherent(&dev->persist->pdev->dev, | ||
1741 | - &chunk->mem[chunk->npages], | ||
1742 | - cur_order, mask); | ||
1743 | + &chunk->buf[chunk->npages], | ||
1744 | + cur_order, mask); | ||
1745 | else | ||
1746 | - ret = mlx4_alloc_icm_pages(&chunk->mem[chunk->npages], | ||
1747 | + ret = mlx4_alloc_icm_pages(&chunk->sg[chunk->npages], | ||
1748 | cur_order, mask, | ||
1749 | dev->numa_node); | ||
1750 | |||
1751 | @@ -205,7 +204,7 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages, | ||
1752 | if (coherent) | ||
1753 | ++chunk->nsg; | ||
1754 | else if (chunk->npages == MLX4_ICM_CHUNK_LEN) { | ||
1755 | - chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, | ||
1756 | + chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->sg, | ||
1757 | chunk->npages, | ||
1758 | PCI_DMA_BIDIRECTIONAL); | ||
1759 | |||
1760 | @@ -220,7 +219,7 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages, | ||
1761 | } | ||
1762 | |||
1763 | if (!coherent && chunk) { | ||
1764 | - chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, | ||
1765 | + chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->sg, | ||
1766 | chunk->npages, | ||
1767 | PCI_DMA_BIDIRECTIONAL); | ||
1768 | |||
1769 | @@ -320,7 +319,7 @@ void *mlx4_table_find(struct mlx4_icm_table *table, u32 obj, | ||
1770 | u64 idx; | ||
1771 | struct mlx4_icm_chunk *chunk; | ||
1772 | struct mlx4_icm *icm; | ||
1773 | - struct page *page = NULL; | ||
1774 | + void *addr = NULL; | ||
1775 | |||
1776 | if (!table->lowmem) | ||
1777 | return NULL; | ||
1778 | @@ -336,28 +335,49 @@ void *mlx4_table_find(struct mlx4_icm_table *table, u32 obj, | ||
1779 | |||
1780 | list_for_each_entry(chunk, &icm->chunk_list, list) { | ||
1781 | for (i = 0; i < chunk->npages; ++i) { | ||
1782 | + dma_addr_t dma_addr; | ||
1783 | + size_t len; | ||
1784 | + | ||
1785 | + if (table->coherent) { | ||
1786 | + len = chunk->buf[i].size; | ||
1787 | + dma_addr = chunk->buf[i].dma_addr; | ||
1788 | + addr = chunk->buf[i].addr; | ||
1789 | + } else { | ||
1790 | + struct page *page; | ||
1791 | + | ||
1792 | + len = sg_dma_len(&chunk->sg[i]); | ||
1793 | + dma_addr = sg_dma_address(&chunk->sg[i]); | ||
1794 | + | ||
1795 | + /* XXX: we should never do this for highmem | ||
1796 | + * allocation. This function either needs | ||
1797 | + * to be split, or the kernel virtual address | ||
1798 | + * return needs to be made optional. | ||
1799 | + */ | ||
1800 | + page = sg_page(&chunk->sg[i]); | ||
1801 | + addr = lowmem_page_address(page); | ||
1802 | + } | ||
1803 | + | ||
1804 | if (dma_handle && dma_offset >= 0) { | ||
1805 | - if (sg_dma_len(&chunk->mem[i]) > dma_offset) | ||
1806 | - *dma_handle = sg_dma_address(&chunk->mem[i]) + | ||
1807 | - dma_offset; | ||
1808 | - dma_offset -= sg_dma_len(&chunk->mem[i]); | ||
1809 | + if (len > dma_offset) | ||
1810 | + *dma_handle = dma_addr + dma_offset; | ||
1811 | + dma_offset -= len; | ||
1812 | } | ||
1813 | + | ||
1814 | /* | ||
1815 | * DMA mapping can merge pages but not split them, | ||
1816 | * so if we found the page, dma_handle has already | ||
1817 | * been assigned to. | ||
1818 | */ | ||
1819 | - if (chunk->mem[i].length > offset) { | ||
1820 | - page = sg_page(&chunk->mem[i]); | ||
1821 | + if (len > offset) | ||
1822 | goto out; | ||
1823 | - } | ||
1824 | - offset -= chunk->mem[i].length; | ||
1825 | + offset -= len; | ||
1826 | } | ||
1827 | } | ||
1828 | |||
1829 | + addr = NULL; | ||
1830 | out: | ||
1831 | mutex_unlock(&table->mutex); | ||
1832 | - return page ? lowmem_page_address(page) + offset : NULL; | ||
1833 | + return addr ? addr + offset : NULL; | ||
1834 | } | ||
1835 | |||
1836 | int mlx4_table_get_range(struct mlx4_dev *dev, struct mlx4_icm_table *table, | ||
1837 | diff --git a/drivers/net/ethernet/mellanox/mlx4/icm.h b/drivers/net/ethernet/mellanox/mlx4/icm.h | ||
1838 | index c9169a490557c..d199874b1c074 100644 | ||
1839 | --- a/drivers/net/ethernet/mellanox/mlx4/icm.h | ||
1840 | +++ b/drivers/net/ethernet/mellanox/mlx4/icm.h | ||
1841 | @@ -47,11 +47,21 @@ enum { | ||
1842 | MLX4_ICM_PAGE_SIZE = 1 << MLX4_ICM_PAGE_SHIFT, | ||
1843 | }; | ||
1844 | |||
1845 | +struct mlx4_icm_buf { | ||
1846 | + void *addr; | ||
1847 | + size_t size; | ||
1848 | + dma_addr_t dma_addr; | ||
1849 | +}; | ||
1850 | + | ||
1851 | struct mlx4_icm_chunk { | ||
1852 | struct list_head list; | ||
1853 | int npages; | ||
1854 | int nsg; | ||
1855 | - struct scatterlist mem[MLX4_ICM_CHUNK_LEN]; | ||
1856 | + bool coherent; | ||
1857 | + union { | ||
1858 | + struct scatterlist sg[MLX4_ICM_CHUNK_LEN]; | ||
1859 | + struct mlx4_icm_buf buf[MLX4_ICM_CHUNK_LEN]; | ||
1860 | + }; | ||
1861 | }; | ||
1862 | |||
1863 | struct mlx4_icm { | ||
1864 | @@ -114,12 +124,18 @@ static inline void mlx4_icm_next(struct mlx4_icm_iter *iter) | ||
1865 | |||
1866 | static inline dma_addr_t mlx4_icm_addr(struct mlx4_icm_iter *iter) | ||
1867 | { | ||
1868 | - return sg_dma_address(&iter->chunk->mem[iter->page_idx]); | ||
1869 | + if (iter->chunk->coherent) | ||
1870 | + return iter->chunk->buf[iter->page_idx].dma_addr; | ||
1871 | + else | ||
1872 | + return sg_dma_address(&iter->chunk->sg[iter->page_idx]); | ||
1873 | } | ||
1874 | |||
1875 | static inline unsigned long mlx4_icm_size(struct mlx4_icm_iter *iter) | ||
1876 | { | ||
1877 | - return sg_dma_len(&iter->chunk->mem[iter->page_idx]); | ||
1878 | + if (iter->chunk->coherent) | ||
1879 | + return iter->chunk->buf[iter->page_idx].size; | ||
1880 | + else | ||
1881 | + return sg_dma_len(&iter->chunk->sg[iter->page_idx]); | ||
1882 | } | ||
1883 | |||
1884 | int mlx4_MAP_ICM_AUX(struct mlx4_dev *dev, struct mlx4_icm *icm); | ||
1885 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en.h b/drivers/net/ethernet/mellanox/mlx5/core/en.h | ||
1886 | index 1183248029264..7c72b3b5eedfa 100644 | ||
1887 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en.h | ||
1888 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en.h | ||
1889 | @@ -636,6 +636,7 @@ enum { | ||
1890 | MLX5E_STATE_ASYNC_EVENTS_ENABLED, | ||
1891 | MLX5E_STATE_OPENED, | ||
1892 | MLX5E_STATE_DESTROYING, | ||
1893 | + MLX5E_STATE_XDP_TX_ENABLED, | ||
1894 | }; | ||
1895 | |||
1896 | struct mlx5e_rqt { | ||
1897 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c | ||
1898 | index ad6d471d00dd4..4a33c9a7cac7e 100644 | ||
1899 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c | ||
1900 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c | ||
1901 | @@ -262,7 +262,8 @@ int mlx5e_xdp_xmit(struct net_device *dev, int n, struct xdp_frame **frames, | ||
1902 | int sq_num; | ||
1903 | int i; | ||
1904 | |||
1905 | - if (unlikely(!test_bit(MLX5E_STATE_OPENED, &priv->state))) | ||
1906 | + /* this flag is sufficient, no need to test internal sq state */ | ||
1907 | + if (unlikely(!mlx5e_xdp_tx_is_enabled(priv))) | ||
1908 | return -ENETDOWN; | ||
1909 | |||
1910 | if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK)) | ||
1911 | @@ -275,9 +276,6 @@ int mlx5e_xdp_xmit(struct net_device *dev, int n, struct xdp_frame **frames, | ||
1912 | |||
1913 | sq = &priv->channels.c[sq_num]->xdpsq; | ||
1914 | |||
1915 | - if (unlikely(!test_bit(MLX5E_SQ_STATE_ENABLED, &sq->state))) | ||
1916 | - return -ENETDOWN; | ||
1917 | - | ||
1918 | for (i = 0; i < n; i++) { | ||
1919 | struct xdp_frame *xdpf = frames[i]; | ||
1920 | struct mlx5e_xdp_info xdpi; | ||
1921 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h | ||
1922 | index 6dfab045925f0..4d096623178b9 100644 | ||
1923 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h | ||
1924 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h | ||
1925 | @@ -49,6 +49,23 @@ bool mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xdp_info *xdpi); | ||
1926 | int mlx5e_xdp_xmit(struct net_device *dev, int n, struct xdp_frame **frames, | ||
1927 | u32 flags); | ||
1928 | |||
1929 | +static inline void mlx5e_xdp_tx_enable(struct mlx5e_priv *priv) | ||
1930 | +{ | ||
1931 | + set_bit(MLX5E_STATE_XDP_TX_ENABLED, &priv->state); | ||
1932 | +} | ||
1933 | + | ||
1934 | +static inline void mlx5e_xdp_tx_disable(struct mlx5e_priv *priv) | ||
1935 | +{ | ||
1936 | + clear_bit(MLX5E_STATE_XDP_TX_ENABLED, &priv->state); | ||
1937 | + /* let other device's napi(s) see our new state */ | ||
1938 | + synchronize_rcu(); | ||
1939 | +} | ||
1940 | + | ||
1941 | +static inline bool mlx5e_xdp_tx_is_enabled(struct mlx5e_priv *priv) | ||
1942 | +{ | ||
1943 | + return test_bit(MLX5E_STATE_XDP_TX_ENABLED, &priv->state); | ||
1944 | +} | ||
1945 | + | ||
1946 | static inline void mlx5e_xmit_xdp_doorbell(struct mlx5e_xdpsq *sq) | ||
1947 | { | ||
1948 | struct mlx5_wq_cyc *wq = &sq->wq; | ||
1949 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c | ||
1950 | index 9577d06578398..1d66a4e22d64f 100644 | ||
1951 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en_main.c | ||
1952 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_main.c | ||
1953 | @@ -2903,6 +2903,7 @@ void mlx5e_activate_priv_channels(struct mlx5e_priv *priv) | ||
1954 | |||
1955 | mlx5e_build_tx2sq_maps(priv); | ||
1956 | mlx5e_activate_channels(&priv->channels); | ||
1957 | + mlx5e_xdp_tx_enable(priv); | ||
1958 | netif_tx_start_all_queues(priv->netdev); | ||
1959 | |||
1960 | if (MLX5_ESWITCH_MANAGER(priv->mdev)) | ||
1961 | @@ -2924,6 +2925,7 @@ void mlx5e_deactivate_priv_channels(struct mlx5e_priv *priv) | ||
1962 | */ | ||
1963 | netif_tx_stop_all_queues(priv->netdev); | ||
1964 | netif_tx_disable(priv->netdev); | ||
1965 | + mlx5e_xdp_tx_disable(priv); | ||
1966 | mlx5e_deactivate_channels(&priv->channels); | ||
1967 | } | ||
1968 | |||
1969 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c b/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c | ||
1970 | index 820fe85100b08..4dccc84fdcf2c 100644 | ||
1971 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c | ||
1972 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_rep.c | ||
1973 | @@ -143,6 +143,7 @@ static void mlx5e_rep_update_sw_counters(struct mlx5e_priv *priv) | ||
1974 | |||
1975 | s->tx_packets += sq_stats->packets; | ||
1976 | s->tx_bytes += sq_stats->bytes; | ||
1977 | + s->tx_queue_dropped += sq_stats->dropped; | ||
1978 | } | ||
1979 | } | ||
1980 | } | ||
1981 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c b/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c | ||
1982 | index 4e0151918db13..3fba80a8b436f 100644 | ||
1983 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c | ||
1984 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_tc.c | ||
1985 | @@ -98,6 +98,7 @@ struct mlx5e_tc_flow_parse_attr { | ||
1986 | struct ip_tunnel_info tun_info; | ||
1987 | struct mlx5_flow_spec spec; | ||
1988 | int num_mod_hdr_actions; | ||
1989 | + int max_mod_hdr_actions; | ||
1990 | void *mod_hdr_actions; | ||
1991 | int mirred_ifindex; | ||
1992 | }; | ||
1993 | @@ -1888,9 +1889,9 @@ static struct mlx5_fields fields[] = { | ||
1994 | OFFLOAD(UDP_DPORT, 2, udp.dest, 0), | ||
1995 | }; | ||
1996 | |||
1997 | -/* On input attr->num_mod_hdr_actions tells how many HW actions can be parsed at | ||
1998 | - * max from the SW pedit action. On success, it says how many HW actions were | ||
1999 | - * actually parsed. | ||
2000 | +/* On input attr->max_mod_hdr_actions tells how many HW actions can be parsed at | ||
2001 | + * max from the SW pedit action. On success, attr->num_mod_hdr_actions | ||
2002 | + * says how many HW actions were actually parsed. | ||
2003 | */ | ||
2004 | static int offload_pedit_fields(struct pedit_headers *masks, | ||
2005 | struct pedit_headers *vals, | ||
2006 | @@ -1914,9 +1915,11 @@ static int offload_pedit_fields(struct pedit_headers *masks, | ||
2007 | add_vals = &vals[TCA_PEDIT_KEY_EX_CMD_ADD]; | ||
2008 | |||
2009 | action_size = MLX5_UN_SZ_BYTES(set_action_in_add_action_in_auto); | ||
2010 | - action = parse_attr->mod_hdr_actions; | ||
2011 | - max_actions = parse_attr->num_mod_hdr_actions; | ||
2012 | - nactions = 0; | ||
2013 | + action = parse_attr->mod_hdr_actions + | ||
2014 | + parse_attr->num_mod_hdr_actions * action_size; | ||
2015 | + | ||
2016 | + max_actions = parse_attr->max_mod_hdr_actions; | ||
2017 | + nactions = parse_attr->num_mod_hdr_actions; | ||
2018 | |||
2019 | for (i = 0; i < ARRAY_SIZE(fields); i++) { | ||
2020 | f = &fields[i]; | ||
2021 | @@ -2027,7 +2030,7 @@ static int alloc_mod_hdr_actions(struct mlx5e_priv *priv, | ||
2022 | if (!parse_attr->mod_hdr_actions) | ||
2023 | return -ENOMEM; | ||
2024 | |||
2025 | - parse_attr->num_mod_hdr_actions = max_actions; | ||
2026 | + parse_attr->max_mod_hdr_actions = max_actions; | ||
2027 | return 0; | ||
2028 | } | ||
2029 | |||
2030 | @@ -2073,9 +2076,11 @@ static int parse_tc_pedit_action(struct mlx5e_priv *priv, | ||
2031 | goto out_err; | ||
2032 | } | ||
2033 | |||
2034 | - err = alloc_mod_hdr_actions(priv, a, namespace, parse_attr); | ||
2035 | - if (err) | ||
2036 | - goto out_err; | ||
2037 | + if (!parse_attr->mod_hdr_actions) { | ||
2038 | + err = alloc_mod_hdr_actions(priv, a, namespace, parse_attr); | ||
2039 | + if (err) | ||
2040 | + goto out_err; | ||
2041 | + } | ||
2042 | |||
2043 | err = offload_pedit_fields(masks, vals, parse_attr, extack); | ||
2044 | if (err < 0) | ||
2045 | diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c b/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c | ||
2046 | index 6dacaeba2fbff..0b03d65474e93 100644 | ||
2047 | --- a/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c | ||
2048 | +++ b/drivers/net/ethernet/mellanox/mlx5/core/en_tx.c | ||
2049 | @@ -387,8 +387,14 @@ netdev_tx_t mlx5e_sq_xmit(struct mlx5e_txqsq *sq, struct sk_buff *skb, | ||
2050 | num_wqebbs = DIV_ROUND_UP(ds_cnt, MLX5_SEND_WQEBB_NUM_DS); | ||
2051 | contig_wqebbs_room = mlx5_wq_cyc_get_contig_wqebbs(wq, pi); | ||
2052 | if (unlikely(contig_wqebbs_room < num_wqebbs)) { | ||
2053 | +#ifdef CONFIG_MLX5_EN_IPSEC | ||
2054 | + struct mlx5_wqe_eth_seg cur_eth = wqe->eth; | ||
2055 | +#endif | ||
2056 | mlx5e_fill_sq_frag_edge(sq, wq, pi, contig_wqebbs_room); | ||
2057 | mlx5e_sq_fetch_wqe(sq, &wqe, &pi); | ||
2058 | +#ifdef CONFIG_MLX5_EN_IPSEC | ||
2059 | + wqe->eth = cur_eth; | ||
2060 | +#endif | ||
2061 | } | ||
2062 | |||
2063 | /* fill wqe */ | ||
2064 | diff --git a/drivers/net/ethernet/mellanox/mlxsw/Kconfig b/drivers/net/ethernet/mellanox/mlxsw/Kconfig | ||
2065 | index 8a291eb36c64c..7338c9bac4e6a 100644 | ||
2066 | --- a/drivers/net/ethernet/mellanox/mlxsw/Kconfig | ||
2067 | +++ b/drivers/net/ethernet/mellanox/mlxsw/Kconfig | ||
2068 | @@ -78,6 +78,7 @@ config MLXSW_SPECTRUM | ||
2069 | depends on IPV6 || IPV6=n | ||
2070 | depends on NET_IPGRE || NET_IPGRE=n | ||
2071 | depends on IPV6_GRE || IPV6_GRE=n | ||
2072 | + depends on VXLAN || VXLAN=n | ||
2073 | select GENERIC_ALLOCATOR | ||
2074 | select PARMAN | ||
2075 | select MLXFW | ||
2076 | diff --git a/drivers/net/ethernet/mellanox/mlxsw/pci.c b/drivers/net/ethernet/mellanox/mlxsw/pci.c | ||
2077 | index c7901a3f2a794..a903e97793f9a 100644 | ||
2078 | --- a/drivers/net/ethernet/mellanox/mlxsw/pci.c | ||
2079 | +++ b/drivers/net/ethernet/mellanox/mlxsw/pci.c | ||
2080 | @@ -1367,10 +1367,10 @@ static int mlxsw_pci_sw_reset(struct mlxsw_pci *mlxsw_pci, | ||
2081 | u32 val = mlxsw_pci_read32(mlxsw_pci, FW_READY); | ||
2082 | |||
2083 | if ((val & MLXSW_PCI_FW_READY_MASK) == MLXSW_PCI_FW_READY_MAGIC) | ||
2084 | - break; | ||
2085 | + return 0; | ||
2086 | cond_resched(); | ||
2087 | } while (time_before(jiffies, end)); | ||
2088 | - return 0; | ||
2089 | + return -EBUSY; | ||
2090 | } | ||
2091 | |||
2092 | static int mlxsw_pci_alloc_irq_vectors(struct mlxsw_pci *mlxsw_pci) | ||
2093 | diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c | ||
2094 | index e3c6fe8b1d406..1dcf152b28138 100644 | ||
2095 | --- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c | ||
2096 | +++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c | ||
2097 | @@ -75,7 +75,15 @@ mlxsw_sp_acl_ctcam_region_entry_insert(struct mlxsw_sp *mlxsw_sp, | ||
2098 | act_set = mlxsw_afa_block_first_set(rulei->act_block); | ||
2099 | mlxsw_reg_ptce2_flex_action_set_memcpy_to(ptce2_pl, act_set); | ||
2100 | |||
2101 | - return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ptce2), ptce2_pl); | ||
2102 | + err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(ptce2), ptce2_pl); | ||
2103 | + if (err) | ||
2104 | + goto err_ptce2_write; | ||
2105 | + | ||
2106 | + return 0; | ||
2107 | + | ||
2108 | +err_ptce2_write: | ||
2109 | + cregion->ops->entry_remove(cregion, centry); | ||
2110 | + return err; | ||
2111 | } | ||
2112 | |||
2113 | static void | ||
2114 | diff --git a/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c b/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c | ||
2115 | index c514af438fc28..b606db9833e9e 100644 | ||
2116 | --- a/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c | ||
2117 | +++ b/drivers/net/ethernet/mellanox/mlxsw/spectrum_switchdev.c | ||
2118 | @@ -1219,7 +1219,7 @@ mlxsw_sp_bridge_port_fdb_flush(struct mlxsw_sp *mlxsw_sp, | ||
2119 | static enum mlxsw_reg_sfd_rec_policy mlxsw_sp_sfd_rec_policy(bool dynamic) | ||
2120 | { | ||
2121 | return dynamic ? MLXSW_REG_SFD_REC_POLICY_DYNAMIC_ENTRY_INGRESS : | ||
2122 | - MLXSW_REG_SFD_REC_POLICY_STATIC_ENTRY; | ||
2123 | + MLXSW_REG_SFD_REC_POLICY_DYNAMIC_ENTRY_MLAG; | ||
2124 | } | ||
2125 | |||
2126 | static enum mlxsw_reg_sfd_op mlxsw_sp_sfd_op(bool adding) | ||
2127 | @@ -1276,7 +1276,7 @@ out: | ||
2128 | static int __mlxsw_sp_port_fdb_uc_op(struct mlxsw_sp *mlxsw_sp, u8 local_port, | ||
2129 | const char *mac, u16 fid, bool adding, | ||
2130 | enum mlxsw_reg_sfd_rec_action action, | ||
2131 | - bool dynamic) | ||
2132 | + enum mlxsw_reg_sfd_rec_policy policy) | ||
2133 | { | ||
2134 | char *sfd_pl; | ||
2135 | u8 num_rec; | ||
2136 | @@ -1287,8 +1287,7 @@ static int __mlxsw_sp_port_fdb_uc_op(struct mlxsw_sp *mlxsw_sp, u8 local_port, | ||
2137 | return -ENOMEM; | ||
2138 | |||
2139 | mlxsw_reg_sfd_pack(sfd_pl, mlxsw_sp_sfd_op(adding), 0); | ||
2140 | - mlxsw_reg_sfd_uc_pack(sfd_pl, 0, mlxsw_sp_sfd_rec_policy(dynamic), | ||
2141 | - mac, fid, action, local_port); | ||
2142 | + mlxsw_reg_sfd_uc_pack(sfd_pl, 0, policy, mac, fid, action, local_port); | ||
2143 | num_rec = mlxsw_reg_sfd_num_rec_get(sfd_pl); | ||
2144 | err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(sfd), sfd_pl); | ||
2145 | if (err) | ||
2146 | @@ -1307,7 +1306,8 @@ static int mlxsw_sp_port_fdb_uc_op(struct mlxsw_sp *mlxsw_sp, u8 local_port, | ||
2147 | bool dynamic) | ||
2148 | { | ||
2149 | return __mlxsw_sp_port_fdb_uc_op(mlxsw_sp, local_port, mac, fid, adding, | ||
2150 | - MLXSW_REG_SFD_REC_ACTION_NOP, dynamic); | ||
2151 | + MLXSW_REG_SFD_REC_ACTION_NOP, | ||
2152 | + mlxsw_sp_sfd_rec_policy(dynamic)); | ||
2153 | } | ||
2154 | |||
2155 | int mlxsw_sp_rif_fdb_op(struct mlxsw_sp *mlxsw_sp, const char *mac, u16 fid, | ||
2156 | @@ -1315,7 +1315,7 @@ int mlxsw_sp_rif_fdb_op(struct mlxsw_sp *mlxsw_sp, const char *mac, u16 fid, | ||
2157 | { | ||
2158 | return __mlxsw_sp_port_fdb_uc_op(mlxsw_sp, 0, mac, fid, adding, | ||
2159 | MLXSW_REG_SFD_REC_ACTION_FORWARD_IP_ROUTER, | ||
2160 | - false); | ||
2161 | + MLXSW_REG_SFD_REC_POLICY_STATIC_ENTRY); | ||
2162 | } | ||
2163 | |||
2164 | static int mlxsw_sp_port_fdb_uc_lag_op(struct mlxsw_sp *mlxsw_sp, u16 lag_id, | ||
2165 | diff --git a/drivers/net/ethernet/qlogic/qed/qed_ll2.c b/drivers/net/ethernet/qlogic/qed/qed_ll2.c | ||
2166 | index c6f4bab67a5fc..9e728ec82c218 100644 | ||
2167 | --- a/drivers/net/ethernet/qlogic/qed/qed_ll2.c | ||
2168 | +++ b/drivers/net/ethernet/qlogic/qed/qed_ll2.c | ||
2169 | @@ -1603,6 +1603,10 @@ static void qed_ll2_post_rx_buffer_notify_fw(struct qed_hwfn *p_hwfn, | ||
2170 | cq_prod = qed_chain_get_prod_idx(&p_rx->rcq_chain); | ||
2171 | rx_prod.bd_prod = cpu_to_le16(bd_prod); | ||
2172 | rx_prod.cqe_prod = cpu_to_le16(cq_prod); | ||
2173 | + | ||
2174 | + /* Make sure chain element is updated before ringing the doorbell */ | ||
2175 | + dma_wmb(); | ||
2176 | + | ||
2177 | DIRECT_REG_WR(p_rx->set_prod_addr, *((u32 *)&rx_prod)); | ||
2178 | } | ||
2179 | |||
2180 | diff --git a/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c b/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c | ||
2181 | index 6c5092e7771cd..c5e25580a43fa 100644 | ||
2182 | --- a/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c | ||
2183 | +++ b/drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c | ||
2184 | @@ -263,6 +263,7 @@ static int dwxgmac2_dma_interrupt(void __iomem *ioaddr, | ||
2185 | struct stmmac_extra_stats *x, u32 chan) | ||
2186 | { | ||
2187 | u32 intr_status = readl(ioaddr + XGMAC_DMA_CH_STATUS(chan)); | ||
2188 | + u32 intr_en = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan)); | ||
2189 | int ret = 0; | ||
2190 | |||
2191 | /* ABNORMAL interrupts */ | ||
2192 | @@ -282,8 +283,7 @@ static int dwxgmac2_dma_interrupt(void __iomem *ioaddr, | ||
2193 | x->normal_irq_n++; | ||
2194 | |||
2195 | if (likely(intr_status & XGMAC_RI)) { | ||
2196 | - u32 value = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan)); | ||
2197 | - if (likely(value & XGMAC_RIE)) { | ||
2198 | + if (likely(intr_en & XGMAC_RIE)) { | ||
2199 | x->rx_normal_irq_n++; | ||
2200 | ret |= handle_rx; | ||
2201 | } | ||
2202 | @@ -295,7 +295,7 @@ static int dwxgmac2_dma_interrupt(void __iomem *ioaddr, | ||
2203 | } | ||
2204 | |||
2205 | /* Clear interrupts */ | ||
2206 | - writel(~0x0, ioaddr + XGMAC_DMA_CH_STATUS(chan)); | ||
2207 | + writel(intr_en & intr_status, ioaddr + XGMAC_DMA_CH_STATUS(chan)); | ||
2208 | |||
2209 | return ret; | ||
2210 | } | ||
2211 | diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | ||
2212 | index c4a35e932f052..5d83d6a7694b0 100644 | ||
2213 | --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | ||
2214 | +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | ||
2215 | @@ -3525,27 +3525,28 @@ static int stmmac_napi_poll(struct napi_struct *napi, int budget) | ||
2216 | struct stmmac_channel *ch = | ||
2217 | container_of(napi, struct stmmac_channel, napi); | ||
2218 | struct stmmac_priv *priv = ch->priv_data; | ||
2219 | - int work_done = 0, work_rem = budget; | ||
2220 | + int work_done, rx_done = 0, tx_done = 0; | ||
2221 | u32 chan = ch->index; | ||
2222 | |||
2223 | priv->xstats.napi_poll++; | ||
2224 | |||
2225 | - if (ch->has_tx) { | ||
2226 | - int done = stmmac_tx_clean(priv, work_rem, chan); | ||
2227 | + if (ch->has_tx) | ||
2228 | + tx_done = stmmac_tx_clean(priv, budget, chan); | ||
2229 | + if (ch->has_rx) | ||
2230 | + rx_done = stmmac_rx(priv, budget, chan); | ||
2231 | |||
2232 | - work_done += done; | ||
2233 | - work_rem -= done; | ||
2234 | - } | ||
2235 | - | ||
2236 | - if (ch->has_rx) { | ||
2237 | - int done = stmmac_rx(priv, work_rem, chan); | ||
2238 | + work_done = max(rx_done, tx_done); | ||
2239 | + work_done = min(work_done, budget); | ||
2240 | |||
2241 | - work_done += done; | ||
2242 | - work_rem -= done; | ||
2243 | - } | ||
2244 | + if (work_done < budget && napi_complete_done(napi, work_done)) { | ||
2245 | + int stat; | ||
2246 | |||
2247 | - if (work_done < budget && napi_complete_done(napi, work_done)) | ||
2248 | stmmac_enable_dma_irq(priv, priv->ioaddr, chan); | ||
2249 | + stat = stmmac_dma_interrupt_status(priv, priv->ioaddr, | ||
2250 | + &priv->xstats, chan); | ||
2251 | + if (stat && napi_reschedule(napi)) | ||
2252 | + stmmac_disable_dma_irq(priv, priv->ioaddr, chan); | ||
2253 | + } | ||
2254 | |||
2255 | return work_done; | ||
2256 | } | ||
2257 | @@ -4194,6 +4195,18 @@ static int stmmac_hw_init(struct stmmac_priv *priv) | ||
2258 | return ret; | ||
2259 | } | ||
2260 | |||
2261 | + /* Rx Watchdog is available in the COREs newer than the 3.40. | ||
2262 | + * In some case, for example on bugged HW this feature | ||
2263 | + * has to be disable and this can be done by passing the | ||
2264 | + * riwt_off field from the platform. | ||
2265 | + */ | ||
2266 | + if (((priv->synopsys_id >= DWMAC_CORE_3_50) || | ||
2267 | + (priv->plat->has_xgmac)) && (!priv->plat->riwt_off)) { | ||
2268 | + priv->use_riwt = 1; | ||
2269 | + dev_info(priv->device, | ||
2270 | + "Enable RX Mitigation via HW Watchdog Timer\n"); | ||
2271 | + } | ||
2272 | + | ||
2273 | return 0; | ||
2274 | } | ||
2275 | |||
2276 | @@ -4326,18 +4339,6 @@ int stmmac_dvr_probe(struct device *device, | ||
2277 | if (flow_ctrl) | ||
2278 | priv->flow_ctrl = FLOW_AUTO; /* RX/TX pause on */ | ||
2279 | |||
2280 | - /* Rx Watchdog is available in the COREs newer than the 3.40. | ||
2281 | - * In some case, for example on bugged HW this feature | ||
2282 | - * has to be disable and this can be done by passing the | ||
2283 | - * riwt_off field from the platform. | ||
2284 | - */ | ||
2285 | - if (((priv->synopsys_id >= DWMAC_CORE_3_50) || | ||
2286 | - (priv->plat->has_xgmac)) && (!priv->plat->riwt_off)) { | ||
2287 | - priv->use_riwt = 1; | ||
2288 | - dev_info(priv->device, | ||
2289 | - "Enable RX Mitigation via HW Watchdog Timer\n"); | ||
2290 | - } | ||
2291 | - | ||
2292 | /* Setup channels NAPI */ | ||
2293 | maxq = max(priv->plat->rx_queues_to_use, priv->plat->tx_queues_to_use); | ||
2294 | |||
2295 | diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c | ||
2296 | index c54a50dbd5ac2..d819e8eaba122 100644 | ||
2297 | --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c | ||
2298 | +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c | ||
2299 | @@ -299,7 +299,17 @@ static int stmmac_pci_probe(struct pci_dev *pdev, | ||
2300 | */ | ||
2301 | static void stmmac_pci_remove(struct pci_dev *pdev) | ||
2302 | { | ||
2303 | + int i; | ||
2304 | + | ||
2305 | stmmac_dvr_remove(&pdev->dev); | ||
2306 | + | ||
2307 | + for (i = 0; i <= PCI_STD_RESOURCE_END; i++) { | ||
2308 | + if (pci_resource_len(pdev, i) == 0) | ||
2309 | + continue; | ||
2310 | + pcim_iounmap_regions(pdev, BIT(i)); | ||
2311 | + break; | ||
2312 | + } | ||
2313 | + | ||
2314 | pci_disable_device(pdev); | ||
2315 | } | ||
2316 | |||
2317 | diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c | ||
2318 | index 531294f4978bc..58ea18af9813a 100644 | ||
2319 | --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c | ||
2320 | +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c | ||
2321 | @@ -301,6 +301,8 @@ static int tc_setup_cbs(struct stmmac_priv *priv, | ||
2322 | /* Queue 0 is not AVB capable */ | ||
2323 | if (queue <= 0 || queue >= tx_queues_count) | ||
2324 | return -EINVAL; | ||
2325 | + if (!priv->dma_cap.av) | ||
2326 | + return -EOPNOTSUPP; | ||
2327 | if (priv->speed != SPEED_100 && priv->speed != SPEED_1000) | ||
2328 | return -EOPNOTSUPP; | ||
2329 | |||
2330 | diff --git a/drivers/net/geneve.c b/drivers/net/geneve.c | ||
2331 | index a0cd1c41cf5f0..2e6e11d8cf5cb 100644 | ||
2332 | --- a/drivers/net/geneve.c | ||
2333 | +++ b/drivers/net/geneve.c | ||
2334 | @@ -1426,9 +1426,13 @@ static void geneve_link_config(struct net_device *dev, | ||
2335 | } | ||
2336 | #if IS_ENABLED(CONFIG_IPV6) | ||
2337 | case AF_INET6: { | ||
2338 | - struct rt6_info *rt = rt6_lookup(geneve->net, | ||
2339 | - &info->key.u.ipv6.dst, NULL, 0, | ||
2340 | - NULL, 0); | ||
2341 | + struct rt6_info *rt; | ||
2342 | + | ||
2343 | + if (!__in6_dev_get(dev)) | ||
2344 | + break; | ||
2345 | + | ||
2346 | + rt = rt6_lookup(geneve->net, &info->key.u.ipv6.dst, NULL, 0, | ||
2347 | + NULL, 0); | ||
2348 | |||
2349 | if (rt && rt->dst.dev) | ||
2350 | ldev_mtu = rt->dst.dev->mtu - GENEVE_IPV6_HLEN; | ||
2351 | diff --git a/drivers/net/phy/micrel.c b/drivers/net/phy/micrel.c | ||
2352 | index 51611c7a23d1c..22dfbd4c6aaf4 100644 | ||
2353 | --- a/drivers/net/phy/micrel.c | ||
2354 | +++ b/drivers/net/phy/micrel.c | ||
2355 | @@ -1076,6 +1076,7 @@ static struct phy_driver ksphy_driver[] = { | ||
2356 | .driver_data = &ksz9021_type, | ||
2357 | .probe = kszphy_probe, | ||
2358 | .config_init = ksz9031_config_init, | ||
2359 | + .soft_reset = genphy_soft_reset, | ||
2360 | .read_status = ksz9031_read_status, | ||
2361 | .ack_interrupt = kszphy_ack_interrupt, | ||
2362 | .config_intr = kszphy_config_intr, | ||
2363 | diff --git a/drivers/net/phy/phylink.c b/drivers/net/phy/phylink.c | ||
2364 | index 9b8dd0d0ee42c..b60c82065fd11 100644 | ||
2365 | --- a/drivers/net/phy/phylink.c | ||
2366 | +++ b/drivers/net/phy/phylink.c | ||
2367 | @@ -475,6 +475,17 @@ static void phylink_run_resolve(struct phylink *pl) | ||
2368 | queue_work(system_power_efficient_wq, &pl->resolve); | ||
2369 | } | ||
2370 | |||
2371 | +static void phylink_run_resolve_and_disable(struct phylink *pl, int bit) | ||
2372 | +{ | ||
2373 | + unsigned long state = pl->phylink_disable_state; | ||
2374 | + | ||
2375 | + set_bit(bit, &pl->phylink_disable_state); | ||
2376 | + if (state == 0) { | ||
2377 | + queue_work(system_power_efficient_wq, &pl->resolve); | ||
2378 | + flush_work(&pl->resolve); | ||
2379 | + } | ||
2380 | +} | ||
2381 | + | ||
2382 | static void phylink_fixed_poll(struct timer_list *t) | ||
2383 | { | ||
2384 | struct phylink *pl = container_of(t, struct phylink, link_poll); | ||
2385 | @@ -928,9 +939,7 @@ void phylink_stop(struct phylink *pl) | ||
2386 | if (pl->link_an_mode == MLO_AN_FIXED && !IS_ERR(pl->link_gpio)) | ||
2387 | del_timer_sync(&pl->link_poll); | ||
2388 | |||
2389 | - set_bit(PHYLINK_DISABLE_STOPPED, &pl->phylink_disable_state); | ||
2390 | - queue_work(system_power_efficient_wq, &pl->resolve); | ||
2391 | - flush_work(&pl->resolve); | ||
2392 | + phylink_run_resolve_and_disable(pl, PHYLINK_DISABLE_STOPPED); | ||
2393 | } | ||
2394 | EXPORT_SYMBOL_GPL(phylink_stop); | ||
2395 | |||
2396 | @@ -1637,9 +1646,7 @@ static void phylink_sfp_link_down(void *upstream) | ||
2397 | |||
2398 | ASSERT_RTNL(); | ||
2399 | |||
2400 | - set_bit(PHYLINK_DISABLE_LINK, &pl->phylink_disable_state); | ||
2401 | - queue_work(system_power_efficient_wq, &pl->resolve); | ||
2402 | - flush_work(&pl->resolve); | ||
2403 | + phylink_run_resolve_and_disable(pl, PHYLINK_DISABLE_LINK); | ||
2404 | } | ||
2405 | |||
2406 | static void phylink_sfp_link_up(void *upstream) | ||
2407 | diff --git a/drivers/net/phy/sfp-bus.c b/drivers/net/phy/sfp-bus.c | ||
2408 | index ad9db652874dc..fef701bfad62e 100644 | ||
2409 | --- a/drivers/net/phy/sfp-bus.c | ||
2410 | +++ b/drivers/net/phy/sfp-bus.c | ||
2411 | @@ -347,6 +347,7 @@ static int sfp_register_bus(struct sfp_bus *bus) | ||
2412 | return ret; | ||
2413 | } | ||
2414 | } | ||
2415 | + bus->socket_ops->attach(bus->sfp); | ||
2416 | if (bus->started) | ||
2417 | bus->socket_ops->start(bus->sfp); | ||
2418 | bus->netdev->sfp_bus = bus; | ||
2419 | @@ -362,6 +363,7 @@ static void sfp_unregister_bus(struct sfp_bus *bus) | ||
2420 | if (bus->registered) { | ||
2421 | if (bus->started) | ||
2422 | bus->socket_ops->stop(bus->sfp); | ||
2423 | + bus->socket_ops->detach(bus->sfp); | ||
2424 | if (bus->phydev && ops && ops->disconnect_phy) | ||
2425 | ops->disconnect_phy(bus->upstream); | ||
2426 | } | ||
2427 | diff --git a/drivers/net/phy/sfp.c b/drivers/net/phy/sfp.c | ||
2428 | index fd8bb998ae52d..68c8fbf099f87 100644 | ||
2429 | --- a/drivers/net/phy/sfp.c | ||
2430 | +++ b/drivers/net/phy/sfp.c | ||
2431 | @@ -184,6 +184,7 @@ struct sfp { | ||
2432 | |||
2433 | struct gpio_desc *gpio[GPIO_MAX]; | ||
2434 | |||
2435 | + bool attached; | ||
2436 | unsigned int state; | ||
2437 | struct delayed_work poll; | ||
2438 | struct delayed_work timeout; | ||
2439 | @@ -1475,7 +1476,7 @@ static void sfp_sm_event(struct sfp *sfp, unsigned int event) | ||
2440 | */ | ||
2441 | switch (sfp->sm_mod_state) { | ||
2442 | default: | ||
2443 | - if (event == SFP_E_INSERT) { | ||
2444 | + if (event == SFP_E_INSERT && sfp->attached) { | ||
2445 | sfp_module_tx_disable(sfp); | ||
2446 | sfp_sm_ins_next(sfp, SFP_MOD_PROBE, T_PROBE_INIT); | ||
2447 | } | ||
2448 | @@ -1607,6 +1608,19 @@ static void sfp_sm_event(struct sfp *sfp, unsigned int event) | ||
2449 | mutex_unlock(&sfp->sm_mutex); | ||
2450 | } | ||
2451 | |||
2452 | +static void sfp_attach(struct sfp *sfp) | ||
2453 | +{ | ||
2454 | + sfp->attached = true; | ||
2455 | + if (sfp->state & SFP_F_PRESENT) | ||
2456 | + sfp_sm_event(sfp, SFP_E_INSERT); | ||
2457 | +} | ||
2458 | + | ||
2459 | +static void sfp_detach(struct sfp *sfp) | ||
2460 | +{ | ||
2461 | + sfp->attached = false; | ||
2462 | + sfp_sm_event(sfp, SFP_E_REMOVE); | ||
2463 | +} | ||
2464 | + | ||
2465 | static void sfp_start(struct sfp *sfp) | ||
2466 | { | ||
2467 | sfp_sm_event(sfp, SFP_E_DEV_UP); | ||
2468 | @@ -1667,6 +1681,8 @@ static int sfp_module_eeprom(struct sfp *sfp, struct ethtool_eeprom *ee, | ||
2469 | } | ||
2470 | |||
2471 | static const struct sfp_socket_ops sfp_module_ops = { | ||
2472 | + .attach = sfp_attach, | ||
2473 | + .detach = sfp_detach, | ||
2474 | .start = sfp_start, | ||
2475 | .stop = sfp_stop, | ||
2476 | .module_info = sfp_module_info, | ||
2477 | @@ -1834,10 +1850,6 @@ static int sfp_probe(struct platform_device *pdev) | ||
2478 | dev_info(sfp->dev, "Host maximum power %u.%uW\n", | ||
2479 | sfp->max_power_mW / 1000, (sfp->max_power_mW / 100) % 10); | ||
2480 | |||
2481 | - sfp->sfp_bus = sfp_register_socket(sfp->dev, sfp, &sfp_module_ops); | ||
2482 | - if (!sfp->sfp_bus) | ||
2483 | - return -ENOMEM; | ||
2484 | - | ||
2485 | /* Get the initial state, and always signal TX disable, | ||
2486 | * since the network interface will not be up. | ||
2487 | */ | ||
2488 | @@ -1848,10 +1860,6 @@ static int sfp_probe(struct platform_device *pdev) | ||
2489 | sfp->state |= SFP_F_RATE_SELECT; | ||
2490 | sfp_set_state(sfp, sfp->state); | ||
2491 | sfp_module_tx_disable(sfp); | ||
2492 | - rtnl_lock(); | ||
2493 | - if (sfp->state & SFP_F_PRESENT) | ||
2494 | - sfp_sm_event(sfp, SFP_E_INSERT); | ||
2495 | - rtnl_unlock(); | ||
2496 | |||
2497 | for (i = 0; i < GPIO_MAX; i++) { | ||
2498 | if (gpio_flags[i] != GPIOD_IN || !sfp->gpio[i]) | ||
2499 | @@ -1884,6 +1892,10 @@ static int sfp_probe(struct platform_device *pdev) | ||
2500 | dev_warn(sfp->dev, | ||
2501 | "No tx_disable pin: SFP modules will always be emitting.\n"); | ||
2502 | |||
2503 | + sfp->sfp_bus = sfp_register_socket(sfp->dev, sfp, &sfp_module_ops); | ||
2504 | + if (!sfp->sfp_bus) | ||
2505 | + return -ENOMEM; | ||
2506 | + | ||
2507 | return 0; | ||
2508 | } | ||
2509 | |||
2510 | diff --git a/drivers/net/phy/sfp.h b/drivers/net/phy/sfp.h | ||
2511 | index 31b0acf337e27..64f54b0bbd8c4 100644 | ||
2512 | --- a/drivers/net/phy/sfp.h | ||
2513 | +++ b/drivers/net/phy/sfp.h | ||
2514 | @@ -7,6 +7,8 @@ | ||
2515 | struct sfp; | ||
2516 | |||
2517 | struct sfp_socket_ops { | ||
2518 | + void (*attach)(struct sfp *sfp); | ||
2519 | + void (*detach)(struct sfp *sfp); | ||
2520 | void (*start)(struct sfp *sfp); | ||
2521 | void (*stop)(struct sfp *sfp); | ||
2522 | int (*module_info)(struct sfp *sfp, struct ethtool_modinfo *modinfo); | ||
2523 | diff --git a/drivers/net/team/team.c b/drivers/net/team/team.c | ||
2524 | index 364f514d56d87..86db1205a3968 100644 | ||
2525 | --- a/drivers/net/team/team.c | ||
2526 | +++ b/drivers/net/team/team.c | ||
2527 | @@ -256,17 +256,6 @@ static void __team_option_inst_mark_removed_port(struct team *team, | ||
2528 | } | ||
2529 | } | ||
2530 | |||
2531 | -static bool __team_option_inst_tmp_find(const struct list_head *opts, | ||
2532 | - const struct team_option_inst *needle) | ||
2533 | -{ | ||
2534 | - struct team_option_inst *opt_inst; | ||
2535 | - | ||
2536 | - list_for_each_entry(opt_inst, opts, tmp_list) | ||
2537 | - if (opt_inst == needle) | ||
2538 | - return true; | ||
2539 | - return false; | ||
2540 | -} | ||
2541 | - | ||
2542 | static int __team_options_register(struct team *team, | ||
2543 | const struct team_option *option, | ||
2544 | size_t option_count) | ||
2545 | @@ -2460,7 +2449,6 @@ static int team_nl_cmd_options_set(struct sk_buff *skb, struct genl_info *info) | ||
2546 | int err = 0; | ||
2547 | int i; | ||
2548 | struct nlattr *nl_option; | ||
2549 | - LIST_HEAD(opt_inst_list); | ||
2550 | |||
2551 | rtnl_lock(); | ||
2552 | |||
2553 | @@ -2480,6 +2468,7 @@ static int team_nl_cmd_options_set(struct sk_buff *skb, struct genl_info *info) | ||
2554 | struct nlattr *opt_attrs[TEAM_ATTR_OPTION_MAX + 1]; | ||
2555 | struct nlattr *attr; | ||
2556 | struct nlattr *attr_data; | ||
2557 | + LIST_HEAD(opt_inst_list); | ||
2558 | enum team_option_type opt_type; | ||
2559 | int opt_port_ifindex = 0; /* != 0 for per-port options */ | ||
2560 | u32 opt_array_index = 0; | ||
2561 | @@ -2584,23 +2573,17 @@ static int team_nl_cmd_options_set(struct sk_buff *skb, struct genl_info *info) | ||
2562 | if (err) | ||
2563 | goto team_put; | ||
2564 | opt_inst->changed = true; | ||
2565 | - | ||
2566 | - /* dumb/evil user-space can send us duplicate opt, | ||
2567 | - * keep only the last one | ||
2568 | - */ | ||
2569 | - if (__team_option_inst_tmp_find(&opt_inst_list, | ||
2570 | - opt_inst)) | ||
2571 | - continue; | ||
2572 | - | ||
2573 | list_add(&opt_inst->tmp_list, &opt_inst_list); | ||
2574 | } | ||
2575 | if (!opt_found) { | ||
2576 | err = -ENOENT; | ||
2577 | goto team_put; | ||
2578 | } | ||
2579 | - } | ||
2580 | |||
2581 | - err = team_nl_send_event_options_get(team, &opt_inst_list); | ||
2582 | + err = team_nl_send_event_options_get(team, &opt_inst_list); | ||
2583 | + if (err) | ||
2584 | + break; | ||
2585 | + } | ||
2586 | |||
2587 | team_put: | ||
2588 | team_nl_team_put(team); | ||
2589 | diff --git a/drivers/net/wan/fsl_ucc_hdlc.c b/drivers/net/wan/fsl_ucc_hdlc.c | ||
2590 | index 4d6409605207c..af13d8cf94ad4 100644 | ||
2591 | --- a/drivers/net/wan/fsl_ucc_hdlc.c | ||
2592 | +++ b/drivers/net/wan/fsl_ucc_hdlc.c | ||
2593 | @@ -1049,6 +1049,54 @@ static const struct net_device_ops uhdlc_ops = { | ||
2594 | .ndo_tx_timeout = uhdlc_tx_timeout, | ||
2595 | }; | ||
2596 | |||
2597 | +static int hdlc_map_iomem(char *name, int init_flag, void __iomem **ptr) | ||
2598 | +{ | ||
2599 | + struct device_node *np; | ||
2600 | + struct platform_device *pdev; | ||
2601 | + struct resource *res; | ||
2602 | + static int siram_init_flag; | ||
2603 | + int ret = 0; | ||
2604 | + | ||
2605 | + np = of_find_compatible_node(NULL, NULL, name); | ||
2606 | + if (!np) | ||
2607 | + return -EINVAL; | ||
2608 | + | ||
2609 | + pdev = of_find_device_by_node(np); | ||
2610 | + if (!pdev) { | ||
2611 | + pr_err("%pOFn: failed to lookup pdev\n", np); | ||
2612 | + of_node_put(np); | ||
2613 | + return -EINVAL; | ||
2614 | + } | ||
2615 | + | ||
2616 | + of_node_put(np); | ||
2617 | + res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
2618 | + if (!res) { | ||
2619 | + ret = -EINVAL; | ||
2620 | + goto error_put_device; | ||
2621 | + } | ||
2622 | + *ptr = ioremap(res->start, resource_size(res)); | ||
2623 | + if (!*ptr) { | ||
2624 | + ret = -ENOMEM; | ||
2625 | + goto error_put_device; | ||
2626 | + } | ||
2627 | + | ||
2628 | + /* We've remapped the addresses, and we don't need the device any | ||
2629 | + * more, so we should release it. | ||
2630 | + */ | ||
2631 | + put_device(&pdev->dev); | ||
2632 | + | ||
2633 | + if (init_flag && siram_init_flag == 0) { | ||
2634 | + memset_io(*ptr, 0, resource_size(res)); | ||
2635 | + siram_init_flag = 1; | ||
2636 | + } | ||
2637 | + return 0; | ||
2638 | + | ||
2639 | +error_put_device: | ||
2640 | + put_device(&pdev->dev); | ||
2641 | + | ||
2642 | + return ret; | ||
2643 | +} | ||
2644 | + | ||
2645 | static int ucc_hdlc_probe(struct platform_device *pdev) | ||
2646 | { | ||
2647 | struct device_node *np = pdev->dev.of_node; | ||
2648 | @@ -1143,6 +1191,15 @@ static int ucc_hdlc_probe(struct platform_device *pdev) | ||
2649 | ret = ucc_of_parse_tdm(np, utdm, ut_info); | ||
2650 | if (ret) | ||
2651 | goto free_utdm; | ||
2652 | + | ||
2653 | + ret = hdlc_map_iomem("fsl,t1040-qe-si", 0, | ||
2654 | + (void __iomem **)&utdm->si_regs); | ||
2655 | + if (ret) | ||
2656 | + goto free_utdm; | ||
2657 | + ret = hdlc_map_iomem("fsl,t1040-qe-siram", 1, | ||
2658 | + (void __iomem **)&utdm->siram); | ||
2659 | + if (ret) | ||
2660 | + goto unmap_si_regs; | ||
2661 | } | ||
2662 | |||
2663 | if (of_property_read_u16(np, "fsl,hmask", &uhdlc_priv->hmask)) | ||
2664 | @@ -1151,7 +1208,7 @@ static int ucc_hdlc_probe(struct platform_device *pdev) | ||
2665 | ret = uhdlc_init(uhdlc_priv); | ||
2666 | if (ret) { | ||
2667 | dev_err(&pdev->dev, "Failed to init uhdlc\n"); | ||
2668 | - goto free_utdm; | ||
2669 | + goto undo_uhdlc_init; | ||
2670 | } | ||
2671 | |||
2672 | dev = alloc_hdlcdev(uhdlc_priv); | ||
2673 | @@ -1181,6 +1238,9 @@ static int ucc_hdlc_probe(struct platform_device *pdev) | ||
2674 | free_dev: | ||
2675 | free_netdev(dev); | ||
2676 | undo_uhdlc_init: | ||
2677 | + iounmap(utdm->siram); | ||
2678 | +unmap_si_regs: | ||
2679 | + iounmap(utdm->si_regs); | ||
2680 | free_utdm: | ||
2681 | if (uhdlc_priv->tsa) | ||
2682 | kfree(utdm); | ||
2683 | diff --git a/drivers/net/wireless/mediatek/mt76/mt76x0/usb.c b/drivers/net/wireless/mediatek/mt76/mt76x0/usb.c | ||
2684 | index ea517864186b4..76f25008491a5 100644 | ||
2685 | --- a/drivers/net/wireless/mediatek/mt76/mt76x0/usb.c | ||
2686 | +++ b/drivers/net/wireless/mediatek/mt76/mt76x0/usb.c | ||
2687 | @@ -159,39 +159,49 @@ static const struct ieee80211_ops mt76x0u_ops = { | ||
2688 | .wake_tx_queue = mt76_wake_tx_queue, | ||
2689 | }; | ||
2690 | |||
2691 | -static int mt76x0u_register_device(struct mt76x02_dev *dev) | ||
2692 | +static int mt76x0u_init_hardware(struct mt76x02_dev *dev) | ||
2693 | { | ||
2694 | - struct ieee80211_hw *hw = dev->mt76.hw; | ||
2695 | int err; | ||
2696 | |||
2697 | - err = mt76u_alloc_queues(&dev->mt76); | ||
2698 | - if (err < 0) | ||
2699 | - goto out_err; | ||
2700 | - | ||
2701 | - err = mt76u_mcu_init_rx(&dev->mt76); | ||
2702 | - if (err < 0) | ||
2703 | - goto out_err; | ||
2704 | - | ||
2705 | mt76x0_chip_onoff(dev, true, true); | ||
2706 | - if (!mt76x02_wait_for_mac(&dev->mt76)) { | ||
2707 | - err = -ETIMEDOUT; | ||
2708 | - goto out_err; | ||
2709 | - } | ||
2710 | + | ||
2711 | + if (!mt76x02_wait_for_mac(&dev->mt76)) | ||
2712 | + return -ETIMEDOUT; | ||
2713 | |||
2714 | err = mt76x0u_mcu_init(dev); | ||
2715 | if (err < 0) | ||
2716 | - goto out_err; | ||
2717 | + return err; | ||
2718 | |||
2719 | mt76x0_init_usb_dma(dev); | ||
2720 | err = mt76x0_init_hardware(dev); | ||
2721 | if (err < 0) | ||
2722 | - goto out_err; | ||
2723 | + return err; | ||
2724 | |||
2725 | mt76_rmw(dev, MT_US_CYC_CFG, MT_US_CYC_CNT, 0x1e); | ||
2726 | mt76_wr(dev, MT_TXOP_CTRL_CFG, | ||
2727 | FIELD_PREP(MT_TXOP_TRUN_EN, 0x3f) | | ||
2728 | FIELD_PREP(MT_TXOP_EXT_CCA_DLY, 0x58)); | ||
2729 | |||
2730 | + return 0; | ||
2731 | +} | ||
2732 | + | ||
2733 | +static int mt76x0u_register_device(struct mt76x02_dev *dev) | ||
2734 | +{ | ||
2735 | + struct ieee80211_hw *hw = dev->mt76.hw; | ||
2736 | + int err; | ||
2737 | + | ||
2738 | + err = mt76u_alloc_queues(&dev->mt76); | ||
2739 | + if (err < 0) | ||
2740 | + goto out_err; | ||
2741 | + | ||
2742 | + err = mt76u_mcu_init_rx(&dev->mt76); | ||
2743 | + if (err < 0) | ||
2744 | + goto out_err; | ||
2745 | + | ||
2746 | + err = mt76x0u_init_hardware(dev); | ||
2747 | + if (err < 0) | ||
2748 | + goto out_err; | ||
2749 | + | ||
2750 | err = mt76x0_register_device(dev); | ||
2751 | if (err < 0) | ||
2752 | goto out_err; | ||
2753 | @@ -300,6 +310,8 @@ static int __maybe_unused mt76x0_suspend(struct usb_interface *usb_intf, | ||
2754 | |||
2755 | mt76u_stop_queues(&dev->mt76); | ||
2756 | mt76x0u_mac_stop(dev); | ||
2757 | + clear_bit(MT76_STATE_MCU_RUNNING, &dev->mt76.state); | ||
2758 | + mt76x0_chip_onoff(dev, false, false); | ||
2759 | usb_kill_urb(usb->mcu.res.urb); | ||
2760 | |||
2761 | return 0; | ||
2762 | @@ -327,7 +339,7 @@ static int __maybe_unused mt76x0_resume(struct usb_interface *usb_intf) | ||
2763 | tasklet_enable(&usb->rx_tasklet); | ||
2764 | tasklet_enable(&usb->tx_tasklet); | ||
2765 | |||
2766 | - ret = mt76x0_init_hardware(dev); | ||
2767 | + ret = mt76x0u_init_hardware(dev); | ||
2768 | if (ret) | ||
2769 | goto err; | ||
2770 | |||
2771 | diff --git a/drivers/pinctrl/pinctrl-max77620.c b/drivers/pinctrl/pinctrl-max77620.c | ||
2772 | index a7f37063518ec..3d05bc1937d40 100644 | ||
2773 | --- a/drivers/pinctrl/pinctrl-max77620.c | ||
2774 | +++ b/drivers/pinctrl/pinctrl-max77620.c | ||
2775 | @@ -34,14 +34,12 @@ enum max77620_pin_ppdrv { | ||
2776 | MAX77620_PIN_PP_DRV, | ||
2777 | }; | ||
2778 | |||
2779 | -enum max77620_pinconf_param { | ||
2780 | - MAX77620_ACTIVE_FPS_SOURCE = PIN_CONFIG_END + 1, | ||
2781 | - MAX77620_ACTIVE_FPS_POWER_ON_SLOTS, | ||
2782 | - MAX77620_ACTIVE_FPS_POWER_DOWN_SLOTS, | ||
2783 | - MAX77620_SUSPEND_FPS_SOURCE, | ||
2784 | - MAX77620_SUSPEND_FPS_POWER_ON_SLOTS, | ||
2785 | - MAX77620_SUSPEND_FPS_POWER_DOWN_SLOTS, | ||
2786 | -}; | ||
2787 | +#define MAX77620_ACTIVE_FPS_SOURCE (PIN_CONFIG_END + 1) | ||
2788 | +#define MAX77620_ACTIVE_FPS_POWER_ON_SLOTS (PIN_CONFIG_END + 2) | ||
2789 | +#define MAX77620_ACTIVE_FPS_POWER_DOWN_SLOTS (PIN_CONFIG_END + 3) | ||
2790 | +#define MAX77620_SUSPEND_FPS_SOURCE (PIN_CONFIG_END + 4) | ||
2791 | +#define MAX77620_SUSPEND_FPS_POWER_ON_SLOTS (PIN_CONFIG_END + 5) | ||
2792 | +#define MAX77620_SUSPEND_FPS_POWER_DOWN_SLOTS (PIN_CONFIG_END + 6) | ||
2793 | |||
2794 | struct max77620_pin_function { | ||
2795 | const char *name; | ||
2796 | diff --git a/drivers/pinctrl/qcom/pinctrl-qcs404.c b/drivers/pinctrl/qcom/pinctrl-qcs404.c | ||
2797 | index 7aae52a09ff03..4ffd56ff809eb 100644 | ||
2798 | --- a/drivers/pinctrl/qcom/pinctrl-qcs404.c | ||
2799 | +++ b/drivers/pinctrl/qcom/pinctrl-qcs404.c | ||
2800 | @@ -79,7 +79,7 @@ enum { | ||
2801 | .intr_cfg_reg = 0, \ | ||
2802 | .intr_status_reg = 0, \ | ||
2803 | .intr_target_reg = 0, \ | ||
2804 | - .tile = NORTH, \ | ||
2805 | + .tile = SOUTH, \ | ||
2806 | .mux_bit = -1, \ | ||
2807 | .pull_bit = pull, \ | ||
2808 | .drv_bit = drv, \ | ||
2809 | diff --git a/drivers/scsi/cxgbi/cxgb3i/cxgb3i.c b/drivers/scsi/cxgbi/cxgb3i/cxgb3i.c | ||
2810 | index bf07735275a49..0fc382cb977bf 100644 | ||
2811 | --- a/drivers/scsi/cxgbi/cxgb3i/cxgb3i.c | ||
2812 | +++ b/drivers/scsi/cxgbi/cxgb3i/cxgb3i.c | ||
2813 | @@ -1144,7 +1144,7 @@ static void ddp_clear_map(struct cxgbi_device *cdev, struct cxgbi_ppm *ppm, | ||
2814 | } | ||
2815 | |||
2816 | static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, | ||
2817 | - unsigned int tid, int pg_idx, bool reply) | ||
2818 | + unsigned int tid, int pg_idx) | ||
2819 | { | ||
2820 | struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0, | ||
2821 | GFP_KERNEL); | ||
2822 | @@ -1160,7 +1160,7 @@ static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, | ||
2823 | req = (struct cpl_set_tcb_field *)skb->head; | ||
2824 | req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); | ||
2825 | OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, tid)); | ||
2826 | - req->reply = V_NO_REPLY(reply ? 0 : 1); | ||
2827 | + req->reply = V_NO_REPLY(1); | ||
2828 | req->cpu_idx = 0; | ||
2829 | req->word = htons(31); | ||
2830 | req->mask = cpu_to_be64(0xF0000000); | ||
2831 | @@ -1177,11 +1177,10 @@ static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, | ||
2832 | * @tid: connection id | ||
2833 | * @hcrc: header digest enabled | ||
2834 | * @dcrc: data digest enabled | ||
2835 | - * @reply: request reply from h/w | ||
2836 | * set up the iscsi digest settings for a connection identified by tid | ||
2837 | */ | ||
2838 | static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid, | ||
2839 | - int hcrc, int dcrc, int reply) | ||
2840 | + int hcrc, int dcrc) | ||
2841 | { | ||
2842 | struct sk_buff *skb = alloc_wr(sizeof(struct cpl_set_tcb_field), 0, | ||
2843 | GFP_KERNEL); | ||
2844 | @@ -1197,7 +1196,7 @@ static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid, | ||
2845 | req = (struct cpl_set_tcb_field *)skb->head; | ||
2846 | req->wr.wr_hi = htonl(V_WR_OP(FW_WROPCODE_FORWARD)); | ||
2847 | OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, tid)); | ||
2848 | - req->reply = V_NO_REPLY(reply ? 0 : 1); | ||
2849 | + req->reply = V_NO_REPLY(1); | ||
2850 | req->cpu_idx = 0; | ||
2851 | req->word = htons(31); | ||
2852 | req->mask = cpu_to_be64(0x0F000000); | ||
2853 | diff --git a/drivers/scsi/cxgbi/cxgb4i/cxgb4i.c b/drivers/scsi/cxgbi/cxgb4i/cxgb4i.c | ||
2854 | index 064ef57351828..bd6cc014cab04 100644 | ||
2855 | --- a/drivers/scsi/cxgbi/cxgb4i/cxgb4i.c | ||
2856 | +++ b/drivers/scsi/cxgbi/cxgb4i/cxgb4i.c | ||
2857 | @@ -1548,16 +1548,22 @@ static void do_set_tcb_rpl(struct cxgbi_device *cdev, struct sk_buff *skb) | ||
2858 | struct cxgbi_sock *csk; | ||
2859 | |||
2860 | csk = lookup_tid(t, tid); | ||
2861 | - if (!csk) | ||
2862 | + if (!csk) { | ||
2863 | pr_err("can't find conn. for tid %u.\n", tid); | ||
2864 | + return; | ||
2865 | + } | ||
2866 | |||
2867 | log_debug(1 << CXGBI_DBG_TOE | 1 << CXGBI_DBG_SOCK, | ||
2868 | "csk 0x%p,%u,%lx,%u, status 0x%x.\n", | ||
2869 | csk, csk->state, csk->flags, csk->tid, rpl->status); | ||
2870 | |||
2871 | - if (rpl->status != CPL_ERR_NONE) | ||
2872 | + if (rpl->status != CPL_ERR_NONE) { | ||
2873 | pr_err("csk 0x%p,%u, SET_TCB_RPL status %u.\n", | ||
2874 | csk, tid, rpl->status); | ||
2875 | + csk->err = -EINVAL; | ||
2876 | + } | ||
2877 | + | ||
2878 | + complete(&csk->cmpl); | ||
2879 | |||
2880 | __kfree_skb(skb); | ||
2881 | } | ||
2882 | @@ -1984,7 +1990,7 @@ static int ddp_set_map(struct cxgbi_ppm *ppm, struct cxgbi_sock *csk, | ||
2883 | } | ||
2884 | |||
2885 | static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, unsigned int tid, | ||
2886 | - int pg_idx, bool reply) | ||
2887 | + int pg_idx) | ||
2888 | { | ||
2889 | struct sk_buff *skb; | ||
2890 | struct cpl_set_tcb_field *req; | ||
2891 | @@ -2000,7 +2006,7 @@ static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, unsigned int tid, | ||
2892 | req = (struct cpl_set_tcb_field *)skb->head; | ||
2893 | INIT_TP_WR(req, csk->tid); | ||
2894 | OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, csk->tid)); | ||
2895 | - req->reply_ctrl = htons(NO_REPLY_V(reply) | QUEUENO_V(csk->rss_qid)); | ||
2896 | + req->reply_ctrl = htons(NO_REPLY_V(0) | QUEUENO_V(csk->rss_qid)); | ||
2897 | req->word_cookie = htons(0); | ||
2898 | req->mask = cpu_to_be64(0x3 << 8); | ||
2899 | req->val = cpu_to_be64(pg_idx << 8); | ||
2900 | @@ -2009,12 +2015,15 @@ static int ddp_setup_conn_pgidx(struct cxgbi_sock *csk, unsigned int tid, | ||
2901 | log_debug(1 << CXGBI_DBG_TOE | 1 << CXGBI_DBG_SOCK, | ||
2902 | "csk 0x%p, tid 0x%x, pg_idx %u.\n", csk, csk->tid, pg_idx); | ||
2903 | |||
2904 | + reinit_completion(&csk->cmpl); | ||
2905 | cxgb4_ofld_send(csk->cdev->ports[csk->port_id], skb); | ||
2906 | - return 0; | ||
2907 | + wait_for_completion(&csk->cmpl); | ||
2908 | + | ||
2909 | + return csk->err; | ||
2910 | } | ||
2911 | |||
2912 | static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid, | ||
2913 | - int hcrc, int dcrc, int reply) | ||
2914 | + int hcrc, int dcrc) | ||
2915 | { | ||
2916 | struct sk_buff *skb; | ||
2917 | struct cpl_set_tcb_field *req; | ||
2918 | @@ -2032,7 +2041,7 @@ static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid, | ||
2919 | req = (struct cpl_set_tcb_field *)skb->head; | ||
2920 | INIT_TP_WR(req, tid); | ||
2921 | OPCODE_TID(req) = htonl(MK_OPCODE_TID(CPL_SET_TCB_FIELD, tid)); | ||
2922 | - req->reply_ctrl = htons(NO_REPLY_V(reply) | QUEUENO_V(csk->rss_qid)); | ||
2923 | + req->reply_ctrl = htons(NO_REPLY_V(0) | QUEUENO_V(csk->rss_qid)); | ||
2924 | req->word_cookie = htons(0); | ||
2925 | req->mask = cpu_to_be64(0x3 << 4); | ||
2926 | req->val = cpu_to_be64(((hcrc ? ULP_CRC_HEADER : 0) | | ||
2927 | @@ -2042,8 +2051,11 @@ static int ddp_setup_conn_digest(struct cxgbi_sock *csk, unsigned int tid, | ||
2928 | log_debug(1 << CXGBI_DBG_TOE | 1 << CXGBI_DBG_SOCK, | ||
2929 | "csk 0x%p, tid 0x%x, crc %d,%d.\n", csk, csk->tid, hcrc, dcrc); | ||
2930 | |||
2931 | + reinit_completion(&csk->cmpl); | ||
2932 | cxgb4_ofld_send(csk->cdev->ports[csk->port_id], skb); | ||
2933 | - return 0; | ||
2934 | + wait_for_completion(&csk->cmpl); | ||
2935 | + | ||
2936 | + return csk->err; | ||
2937 | } | ||
2938 | |||
2939 | static struct cxgbi_ppm *cdev2ppm(struct cxgbi_device *cdev) | ||
2940 | diff --git a/drivers/scsi/cxgbi/libcxgbi.c b/drivers/scsi/cxgbi/libcxgbi.c | ||
2941 | index 75f876409fb9d..245742557c036 100644 | ||
2942 | --- a/drivers/scsi/cxgbi/libcxgbi.c | ||
2943 | +++ b/drivers/scsi/cxgbi/libcxgbi.c | ||
2944 | @@ -573,6 +573,7 @@ static struct cxgbi_sock *cxgbi_sock_create(struct cxgbi_device *cdev) | ||
2945 | skb_queue_head_init(&csk->receive_queue); | ||
2946 | skb_queue_head_init(&csk->write_queue); | ||
2947 | timer_setup(&csk->retry_timer, NULL, 0); | ||
2948 | + init_completion(&csk->cmpl); | ||
2949 | rwlock_init(&csk->callback_lock); | ||
2950 | csk->cdev = cdev; | ||
2951 | csk->flags = 0; | ||
2952 | @@ -2251,14 +2252,14 @@ int cxgbi_set_conn_param(struct iscsi_cls_conn *cls_conn, | ||
2953 | if (!err && conn->hdrdgst_en) | ||
2954 | err = csk->cdev->csk_ddp_setup_digest(csk, csk->tid, | ||
2955 | conn->hdrdgst_en, | ||
2956 | - conn->datadgst_en, 0); | ||
2957 | + conn->datadgst_en); | ||
2958 | break; | ||
2959 | case ISCSI_PARAM_DATADGST_EN: | ||
2960 | err = iscsi_set_param(cls_conn, param, buf, buflen); | ||
2961 | if (!err && conn->datadgst_en) | ||
2962 | err = csk->cdev->csk_ddp_setup_digest(csk, csk->tid, | ||
2963 | conn->hdrdgst_en, | ||
2964 | - conn->datadgst_en, 0); | ||
2965 | + conn->datadgst_en); | ||
2966 | break; | ||
2967 | case ISCSI_PARAM_MAX_R2T: | ||
2968 | return iscsi_tcp_set_max_r2t(conn, buf); | ||
2969 | @@ -2384,7 +2385,7 @@ int cxgbi_bind_conn(struct iscsi_cls_session *cls_session, | ||
2970 | |||
2971 | ppm = csk->cdev->cdev2ppm(csk->cdev); | ||
2972 | err = csk->cdev->csk_ddp_setup_pgidx(csk, csk->tid, | ||
2973 | - ppm->tformat.pgsz_idx_dflt, 0); | ||
2974 | + ppm->tformat.pgsz_idx_dflt); | ||
2975 | if (err < 0) | ||
2976 | return err; | ||
2977 | |||
2978 | diff --git a/drivers/scsi/cxgbi/libcxgbi.h b/drivers/scsi/cxgbi/libcxgbi.h | ||
2979 | index 5d5d8b50d8426..1917ff57651d7 100644 | ||
2980 | --- a/drivers/scsi/cxgbi/libcxgbi.h | ||
2981 | +++ b/drivers/scsi/cxgbi/libcxgbi.h | ||
2982 | @@ -149,6 +149,7 @@ struct cxgbi_sock { | ||
2983 | struct sk_buff_head receive_queue; | ||
2984 | struct sk_buff_head write_queue; | ||
2985 | struct timer_list retry_timer; | ||
2986 | + struct completion cmpl; | ||
2987 | int err; | ||
2988 | rwlock_t callback_lock; | ||
2989 | void *user_data; | ||
2990 | @@ -490,9 +491,9 @@ struct cxgbi_device { | ||
2991 | struct cxgbi_ppm *, | ||
2992 | struct cxgbi_task_tag_info *); | ||
2993 | int (*csk_ddp_setup_digest)(struct cxgbi_sock *, | ||
2994 | - unsigned int, int, int, int); | ||
2995 | + unsigned int, int, int); | ||
2996 | int (*csk_ddp_setup_pgidx)(struct cxgbi_sock *, | ||
2997 | - unsigned int, int, bool); | ||
2998 | + unsigned int, int); | ||
2999 | |||
3000 | void (*csk_release_offload_resources)(struct cxgbi_sock *); | ||
3001 | int (*csk_rx_pdu_ready)(struct cxgbi_sock *, struct sk_buff *); | ||
3002 | diff --git a/drivers/scsi/isci/init.c b/drivers/scsi/isci/init.c | ||
3003 | index 08c7b1e25fe48..dde84f7443136 100644 | ||
3004 | --- a/drivers/scsi/isci/init.c | ||
3005 | +++ b/drivers/scsi/isci/init.c | ||
3006 | @@ -588,6 +588,13 @@ static struct isci_host *isci_host_alloc(struct pci_dev *pdev, int id) | ||
3007 | shost->max_lun = ~0; | ||
3008 | shost->max_cmd_len = MAX_COMMAND_SIZE; | ||
3009 | |||
3010 | + /* turn on DIF support */ | ||
3011 | + scsi_host_set_prot(shost, | ||
3012 | + SHOST_DIF_TYPE1_PROTECTION | | ||
3013 | + SHOST_DIF_TYPE2_PROTECTION | | ||
3014 | + SHOST_DIF_TYPE3_PROTECTION); | ||
3015 | + scsi_host_set_guard(shost, SHOST_DIX_GUARD_CRC); | ||
3016 | + | ||
3017 | err = scsi_add_host(shost, &pdev->dev); | ||
3018 | if (err) | ||
3019 | goto err_shost; | ||
3020 | @@ -675,13 +682,6 @@ static int isci_pci_probe(struct pci_dev *pdev, const struct pci_device_id *id) | ||
3021 | goto err_host_alloc; | ||
3022 | } | ||
3023 | pci_info->hosts[i] = h; | ||
3024 | - | ||
3025 | - /* turn on DIF support */ | ||
3026 | - scsi_host_set_prot(to_shost(h), | ||
3027 | - SHOST_DIF_TYPE1_PROTECTION | | ||
3028 | - SHOST_DIF_TYPE2_PROTECTION | | ||
3029 | - SHOST_DIF_TYPE3_PROTECTION); | ||
3030 | - scsi_host_set_guard(to_shost(h), SHOST_DIX_GUARD_CRC); | ||
3031 | } | ||
3032 | |||
3033 | err = isci_setup_interrupts(pdev); | ||
3034 | diff --git a/drivers/scsi/qedi/qedi_iscsi.c b/drivers/scsi/qedi/qedi_iscsi.c | ||
3035 | index 2f0a4f2c5ff80..d4821b9dea45d 100644 | ||
3036 | --- a/drivers/scsi/qedi/qedi_iscsi.c | ||
3037 | +++ b/drivers/scsi/qedi/qedi_iscsi.c | ||
3038 | @@ -954,6 +954,7 @@ static int qedi_ep_poll(struct iscsi_endpoint *ep, int timeout_ms) | ||
3039 | |||
3040 | qedi_ep = ep->dd_data; | ||
3041 | if (qedi_ep->state == EP_STATE_IDLE || | ||
3042 | + qedi_ep->state == EP_STATE_OFLDCONN_NONE || | ||
3043 | qedi_ep->state == EP_STATE_OFLDCONN_FAILED) | ||
3044 | return -1; | ||
3045 | |||
3046 | @@ -1036,6 +1037,7 @@ static void qedi_ep_disconnect(struct iscsi_endpoint *ep) | ||
3047 | |||
3048 | switch (qedi_ep->state) { | ||
3049 | case EP_STATE_OFLDCONN_START: | ||
3050 | + case EP_STATE_OFLDCONN_NONE: | ||
3051 | goto ep_release_conn; | ||
3052 | case EP_STATE_OFLDCONN_FAILED: | ||
3053 | break; | ||
3054 | @@ -1226,6 +1228,7 @@ static int qedi_set_path(struct Scsi_Host *shost, struct iscsi_path *path_data) | ||
3055 | |||
3056 | if (!is_valid_ether_addr(&path_data->mac_addr[0])) { | ||
3057 | QEDI_NOTICE(&qedi->dbg_ctx, "dst mac NOT VALID\n"); | ||
3058 | + qedi_ep->state = EP_STATE_OFLDCONN_NONE; | ||
3059 | ret = -EIO; | ||
3060 | goto set_path_exit; | ||
3061 | } | ||
3062 | diff --git a/drivers/scsi/qedi/qedi_iscsi.h b/drivers/scsi/qedi/qedi_iscsi.h | ||
3063 | index 11260776212fa..892d70d545537 100644 | ||
3064 | --- a/drivers/scsi/qedi/qedi_iscsi.h | ||
3065 | +++ b/drivers/scsi/qedi/qedi_iscsi.h | ||
3066 | @@ -59,6 +59,7 @@ enum { | ||
3067 | EP_STATE_OFLDCONN_FAILED = 0x2000, | ||
3068 | EP_STATE_CONNECT_FAILED = 0x4000, | ||
3069 | EP_STATE_DISCONN_TIMEDOUT = 0x8000, | ||
3070 | + EP_STATE_OFLDCONN_NONE = 0x10000, | ||
3071 | }; | ||
3072 | |||
3073 | struct qedi_conn; | ||
3074 | diff --git a/drivers/scsi/qla1280.c b/drivers/scsi/qla1280.c | ||
3075 | index 15a50cc7e4b36..c8589926014d4 100644 | ||
3076 | --- a/drivers/scsi/qla1280.c | ||
3077 | +++ b/drivers/scsi/qla1280.c | ||
3078 | @@ -4259,7 +4259,7 @@ qla1280_probe_one(struct pci_dev *pdev, const struct pci_device_id *id) | ||
3079 | ha->devnum = devnum; /* specifies microcode load address */ | ||
3080 | |||
3081 | #ifdef QLA_64BIT_PTR | ||
3082 | - if (dma_set_mask(&ha->pdev->dev, DMA_BIT_MASK(64))) { | ||
3083 | + if (dma_set_mask_and_coherent(&ha->pdev->dev, DMA_BIT_MASK(64))) { | ||
3084 | if (dma_set_mask(&ha->pdev->dev, DMA_BIT_MASK(32))) { | ||
3085 | printk(KERN_WARNING "scsi(%li): Unable to set a " | ||
3086 | "suitable DMA mask - aborting\n", ha->host_no); | ||
3087 | diff --git a/drivers/scsi/qla4xxx/ql4_os.c b/drivers/scsi/qla4xxx/ql4_os.c | ||
3088 | index 051164f755a4c..a13396c56a6a1 100644 | ||
3089 | --- a/drivers/scsi/qla4xxx/ql4_os.c | ||
3090 | +++ b/drivers/scsi/qla4xxx/ql4_os.c | ||
3091 | @@ -7237,6 +7237,8 @@ static int qla4xxx_sysfs_ddb_tgt_create(struct scsi_qla_host *ha, | ||
3092 | |||
3093 | rc = qla4xxx_copy_from_fwddb_param(fnode_sess, fnode_conn, | ||
3094 | fw_ddb_entry); | ||
3095 | + if (rc) | ||
3096 | + goto free_sess; | ||
3097 | |||
3098 | ql4_printk(KERN_INFO, ha, "%s: sysfs entry %s created\n", | ||
3099 | __func__, fnode_sess->dev.kobj.name); | ||
3100 | diff --git a/drivers/scsi/sd_zbc.c b/drivers/scsi/sd_zbc.c | ||
3101 | index 3f81bab48ac24..7d8442c377dfa 100644 | ||
3102 | --- a/drivers/scsi/sd_zbc.c | ||
3103 | +++ b/drivers/scsi/sd_zbc.c | ||
3104 | @@ -142,10 +142,12 @@ int sd_zbc_report_zones(struct gendisk *disk, sector_t sector, | ||
3105 | return -EOPNOTSUPP; | ||
3106 | |||
3107 | /* | ||
3108 | - * Get a reply buffer for the number of requested zones plus a header. | ||
3109 | - * For ATA, buffers must be aligned to 512B. | ||
3110 | + * Get a reply buffer for the number of requested zones plus a header, | ||
3111 | + * without exceeding the device maximum command size. For ATA disks, | ||
3112 | + * buffers must be aligned to 512B. | ||
3113 | */ | ||
3114 | - buflen = roundup((nrz + 1) * 64, 512); | ||
3115 | + buflen = min(queue_max_hw_sectors(disk->queue) << 9, | ||
3116 | + roundup((nrz + 1) * 64, 512)); | ||
3117 | buf = kmalloc(buflen, gfp_mask); | ||
3118 | if (!buf) | ||
3119 | return -ENOMEM; | ||
3120 | diff --git a/drivers/scsi/ufs/ufs.h b/drivers/scsi/ufs/ufs.h | ||
3121 | index 58087d3916d05..5417ce09b1054 100644 | ||
3122 | --- a/drivers/scsi/ufs/ufs.h | ||
3123 | +++ b/drivers/scsi/ufs/ufs.h | ||
3124 | @@ -195,7 +195,7 @@ enum ufs_desc_def_size { | ||
3125 | QUERY_DESC_CONFIGURATION_DEF_SIZE = 0x90, | ||
3126 | QUERY_DESC_UNIT_DEF_SIZE = 0x23, | ||
3127 | QUERY_DESC_INTERCONNECT_DEF_SIZE = 0x06, | ||
3128 | - QUERY_DESC_GEOMETRY_DEF_SIZE = 0x44, | ||
3129 | + QUERY_DESC_GEOMETRY_DEF_SIZE = 0x48, | ||
3130 | QUERY_DESC_POWER_DEF_SIZE = 0x62, | ||
3131 | QUERY_DESC_HEALTH_DEF_SIZE = 0x25, | ||
3132 | }; | ||
3133 | diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c | ||
3134 | index 1cb35ab8a4ec2..2772ff4357fc4 100644 | ||
3135 | --- a/drivers/scsi/ufs/ufshcd.c | ||
3136 | +++ b/drivers/scsi/ufs/ufshcd.c | ||
3137 | @@ -7924,6 +7924,8 @@ out: | ||
3138 | trace_ufshcd_system_resume(dev_name(hba->dev), ret, | ||
3139 | ktime_to_us(ktime_sub(ktime_get(), start)), | ||
3140 | hba->curr_dev_pwr_mode, hba->uic_link_state); | ||
3141 | + if (!ret) | ||
3142 | + hba->is_sys_suspended = false; | ||
3143 | return ret; | ||
3144 | } | ||
3145 | EXPORT_SYMBOL(ufshcd_system_resume); | ||
3146 | diff --git a/drivers/soc/fsl/qe/qe_tdm.c b/drivers/soc/fsl/qe/qe_tdm.c | ||
3147 | index f78c34647ca2d..76480df195a87 100644 | ||
3148 | --- a/drivers/soc/fsl/qe/qe_tdm.c | ||
3149 | +++ b/drivers/soc/fsl/qe/qe_tdm.c | ||
3150 | @@ -44,10 +44,6 @@ int ucc_of_parse_tdm(struct device_node *np, struct ucc_tdm *utdm, | ||
3151 | const char *sprop; | ||
3152 | int ret = 0; | ||
3153 | u32 val; | ||
3154 | - struct resource *res; | ||
3155 | - struct device_node *np2; | ||
3156 | - static int siram_init_flag; | ||
3157 | - struct platform_device *pdev; | ||
3158 | |||
3159 | sprop = of_get_property(np, "fsl,rx-sync-clock", NULL); | ||
3160 | if (sprop) { | ||
3161 | @@ -124,57 +120,6 @@ int ucc_of_parse_tdm(struct device_node *np, struct ucc_tdm *utdm, | ||
3162 | utdm->siram_entry_id = val; | ||
3163 | |||
3164 | set_si_param(utdm, ut_info); | ||
3165 | - | ||
3166 | - np2 = of_find_compatible_node(NULL, NULL, "fsl,t1040-qe-si"); | ||
3167 | - if (!np2) | ||
3168 | - return -EINVAL; | ||
3169 | - | ||
3170 | - pdev = of_find_device_by_node(np2); | ||
3171 | - if (!pdev) { | ||
3172 | - pr_err("%pOFn: failed to lookup pdev\n", np2); | ||
3173 | - of_node_put(np2); | ||
3174 | - return -EINVAL; | ||
3175 | - } | ||
3176 | - | ||
3177 | - of_node_put(np2); | ||
3178 | - res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
3179 | - utdm->si_regs = devm_ioremap_resource(&pdev->dev, res); | ||
3180 | - if (IS_ERR(utdm->si_regs)) { | ||
3181 | - ret = PTR_ERR(utdm->si_regs); | ||
3182 | - goto err_miss_siram_property; | ||
3183 | - } | ||
3184 | - | ||
3185 | - np2 = of_find_compatible_node(NULL, NULL, "fsl,t1040-qe-siram"); | ||
3186 | - if (!np2) { | ||
3187 | - ret = -EINVAL; | ||
3188 | - goto err_miss_siram_property; | ||
3189 | - } | ||
3190 | - | ||
3191 | - pdev = of_find_device_by_node(np2); | ||
3192 | - if (!pdev) { | ||
3193 | - ret = -EINVAL; | ||
3194 | - pr_err("%pOFn: failed to lookup pdev\n", np2); | ||
3195 | - of_node_put(np2); | ||
3196 | - goto err_miss_siram_property; | ||
3197 | - } | ||
3198 | - | ||
3199 | - of_node_put(np2); | ||
3200 | - res = platform_get_resource(pdev, IORESOURCE_MEM, 0); | ||
3201 | - utdm->siram = devm_ioremap_resource(&pdev->dev, res); | ||
3202 | - if (IS_ERR(utdm->siram)) { | ||
3203 | - ret = PTR_ERR(utdm->siram); | ||
3204 | - goto err_miss_siram_property; | ||
3205 | - } | ||
3206 | - | ||
3207 | - if (siram_init_flag == 0) { | ||
3208 | - memset_io(utdm->siram, 0, resource_size(res)); | ||
3209 | - siram_init_flag = 1; | ||
3210 | - } | ||
3211 | - | ||
3212 | - return ret; | ||
3213 | - | ||
3214 | -err_miss_siram_property: | ||
3215 | - devm_iounmap(&pdev->dev, utdm->si_regs); | ||
3216 | return ret; | ||
3217 | } | ||
3218 | EXPORT_SYMBOL(ucc_of_parse_tdm); | ||
3219 | diff --git a/drivers/staging/erofs/dir.c b/drivers/staging/erofs/dir.c | ||
3220 | index d1cb0d78ab844..e44ca93dcdc68 100644 | ||
3221 | --- a/drivers/staging/erofs/dir.c | ||
3222 | +++ b/drivers/staging/erofs/dir.c | ||
3223 | @@ -53,8 +53,11 @@ static int erofs_fill_dentries(struct dir_context *ctx, | ||
3224 | strnlen(de_name, maxsize - nameoff) : | ||
3225 | le16_to_cpu(de[1].nameoff) - nameoff; | ||
3226 | |||
3227 | - /* the corrupted directory found */ | ||
3228 | - BUG_ON(de_namelen < 0); | ||
3229 | + /* a corrupted entry is found */ | ||
3230 | + if (unlikely(de_namelen < 0)) { | ||
3231 | + DBG_BUGON(1); | ||
3232 | + return -EIO; | ||
3233 | + } | ||
3234 | |||
3235 | #ifdef CONFIG_EROFS_FS_DEBUG | ||
3236 | dbg_namelen = min(EROFS_NAME_LEN - 1, de_namelen); | ||
3237 | diff --git a/drivers/staging/erofs/inode.c b/drivers/staging/erofs/inode.c | ||
3238 | index 04c61a9d7b766..d7fbf5f4600f3 100644 | ||
3239 | --- a/drivers/staging/erofs/inode.c | ||
3240 | +++ b/drivers/staging/erofs/inode.c | ||
3241 | @@ -133,7 +133,13 @@ static int fill_inline_data(struct inode *inode, void *data, | ||
3242 | return -ENOMEM; | ||
3243 | |||
3244 | m_pofs += vi->inode_isize + vi->xattr_isize; | ||
3245 | - BUG_ON(m_pofs + inode->i_size > PAGE_SIZE); | ||
3246 | + | ||
3247 | + /* inline symlink data shouldn't across page boundary as well */ | ||
3248 | + if (unlikely(m_pofs + inode->i_size > PAGE_SIZE)) { | ||
3249 | + DBG_BUGON(1); | ||
3250 | + kfree(lnk); | ||
3251 | + return -EIO; | ||
3252 | + } | ||
3253 | |||
3254 | /* get in-page inline data */ | ||
3255 | memcpy(lnk, data + m_pofs, inode->i_size); | ||
3256 | @@ -171,7 +177,7 @@ static int fill_inode(struct inode *inode, int isdir) | ||
3257 | return PTR_ERR(page); | ||
3258 | } | ||
3259 | |||
3260 | - BUG_ON(!PageUptodate(page)); | ||
3261 | + DBG_BUGON(!PageUptodate(page)); | ||
3262 | data = page_address(page); | ||
3263 | |||
3264 | err = read_inode(inode, data + ofs); | ||
3265 | diff --git a/drivers/staging/erofs/internal.h b/drivers/staging/erofs/internal.h | ||
3266 | index 3ac4599bbe011..8929443558676 100644 | ||
3267 | --- a/drivers/staging/erofs/internal.h | ||
3268 | +++ b/drivers/staging/erofs/internal.h | ||
3269 | @@ -194,50 +194,70 @@ struct erofs_workgroup { | ||
3270 | |||
3271 | #define EROFS_LOCKED_MAGIC (INT_MIN | 0xE0F510CCL) | ||
3272 | |||
3273 | -static inline bool erofs_workgroup_try_to_freeze( | ||
3274 | - struct erofs_workgroup *grp, int v) | ||
3275 | +#if defined(CONFIG_SMP) | ||
3276 | +static inline bool erofs_workgroup_try_to_freeze(struct erofs_workgroup *grp, | ||
3277 | + int val) | ||
3278 | { | ||
3279 | -#if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) | ||
3280 | - if (v != atomic_cmpxchg(&grp->refcount, | ||
3281 | - v, EROFS_LOCKED_MAGIC)) | ||
3282 | - return false; | ||
3283 | preempt_disable(); | ||
3284 | + if (val != atomic_cmpxchg(&grp->refcount, val, EROFS_LOCKED_MAGIC)) { | ||
3285 | + preempt_enable(); | ||
3286 | + return false; | ||
3287 | + } | ||
3288 | + return true; | ||
3289 | +} | ||
3290 | + | ||
3291 | +static inline void erofs_workgroup_unfreeze(struct erofs_workgroup *grp, | ||
3292 | + int orig_val) | ||
3293 | +{ | ||
3294 | + /* | ||
3295 | + * other observers should notice all modifications | ||
3296 | + * in the freezing period. | ||
3297 | + */ | ||
3298 | + smp_mb(); | ||
3299 | + atomic_set(&grp->refcount, orig_val); | ||
3300 | + preempt_enable(); | ||
3301 | +} | ||
3302 | + | ||
3303 | +static inline int erofs_wait_on_workgroup_freezed(struct erofs_workgroup *grp) | ||
3304 | +{ | ||
3305 | + return atomic_cond_read_relaxed(&grp->refcount, | ||
3306 | + VAL != EROFS_LOCKED_MAGIC); | ||
3307 | +} | ||
3308 | #else | ||
3309 | +static inline bool erofs_workgroup_try_to_freeze(struct erofs_workgroup *grp, | ||
3310 | + int val) | ||
3311 | +{ | ||
3312 | preempt_disable(); | ||
3313 | - if (atomic_read(&grp->refcount) != v) { | ||
3314 | + /* no need to spin on UP platforms, let's just disable preemption. */ | ||
3315 | + if (val != atomic_read(&grp->refcount)) { | ||
3316 | preempt_enable(); | ||
3317 | return false; | ||
3318 | } | ||
3319 | -#endif | ||
3320 | return true; | ||
3321 | } | ||
3322 | |||
3323 | -static inline void erofs_workgroup_unfreeze( | ||
3324 | - struct erofs_workgroup *grp, int v) | ||
3325 | +static inline void erofs_workgroup_unfreeze(struct erofs_workgroup *grp, | ||
3326 | + int orig_val) | ||
3327 | { | ||
3328 | -#if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) | ||
3329 | - atomic_set(&grp->refcount, v); | ||
3330 | -#endif | ||
3331 | preempt_enable(); | ||
3332 | } | ||
3333 | |||
3334 | +static inline int erofs_wait_on_workgroup_freezed(struct erofs_workgroup *grp) | ||
3335 | +{ | ||
3336 | + int v = atomic_read(&grp->refcount); | ||
3337 | + | ||
3338 | + /* workgroup is never freezed on uniprocessor systems */ | ||
3339 | + DBG_BUGON(v == EROFS_LOCKED_MAGIC); | ||
3340 | + return v; | ||
3341 | +} | ||
3342 | +#endif | ||
3343 | + | ||
3344 | static inline bool erofs_workgroup_get(struct erofs_workgroup *grp, int *ocnt) | ||
3345 | { | ||
3346 | - const int locked = (int)EROFS_LOCKED_MAGIC; | ||
3347 | int o; | ||
3348 | |||
3349 | repeat: | ||
3350 | - o = atomic_read(&grp->refcount); | ||
3351 | - | ||
3352 | - /* spin if it is temporarily locked at the reclaim path */ | ||
3353 | - if (unlikely(o == locked)) { | ||
3354 | -#if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) | ||
3355 | - do | ||
3356 | - cpu_relax(); | ||
3357 | - while (atomic_read(&grp->refcount) == locked); | ||
3358 | -#endif | ||
3359 | - goto repeat; | ||
3360 | - } | ||
3361 | + o = erofs_wait_on_workgroup_freezed(grp); | ||
3362 | |||
3363 | if (unlikely(o <= 0)) | ||
3364 | return -1; | ||
3365 | diff --git a/drivers/staging/erofs/super.c b/drivers/staging/erofs/super.c | ||
3366 | index f69e619807a17..1c2eb69682efc 100644 | ||
3367 | --- a/drivers/staging/erofs/super.c | ||
3368 | +++ b/drivers/staging/erofs/super.c | ||
3369 | @@ -40,7 +40,6 @@ static int __init erofs_init_inode_cache(void) | ||
3370 | |||
3371 | static void erofs_exit_inode_cache(void) | ||
3372 | { | ||
3373 | - BUG_ON(erofs_inode_cachep == NULL); | ||
3374 | kmem_cache_destroy(erofs_inode_cachep); | ||
3375 | } | ||
3376 | |||
3377 | @@ -303,8 +302,8 @@ static int managed_cache_releasepage(struct page *page, gfp_t gfp_mask) | ||
3378 | int ret = 1; /* 0 - busy */ | ||
3379 | struct address_space *const mapping = page->mapping; | ||
3380 | |||
3381 | - BUG_ON(!PageLocked(page)); | ||
3382 | - BUG_ON(mapping->a_ops != &managed_cache_aops); | ||
3383 | + DBG_BUGON(!PageLocked(page)); | ||
3384 | + DBG_BUGON(mapping->a_ops != &managed_cache_aops); | ||
3385 | |||
3386 | if (PagePrivate(page)) | ||
3387 | ret = erofs_try_to_free_cached_page(mapping, page); | ||
3388 | @@ -317,10 +316,10 @@ static void managed_cache_invalidatepage(struct page *page, | ||
3389 | { | ||
3390 | const unsigned int stop = length + offset; | ||
3391 | |||
3392 | - BUG_ON(!PageLocked(page)); | ||
3393 | + DBG_BUGON(!PageLocked(page)); | ||
3394 | |||
3395 | - /* Check for overflow */ | ||
3396 | - BUG_ON(stop > PAGE_SIZE || stop < length); | ||
3397 | + /* Check for potential overflow in debug mode */ | ||
3398 | + DBG_BUGON(stop > PAGE_SIZE || stop < length); | ||
3399 | |||
3400 | if (offset == 0 && stop == PAGE_SIZE) | ||
3401 | while (!managed_cache_releasepage(page, GFP_NOFS)) | ||
3402 | @@ -442,12 +441,6 @@ static int erofs_read_super(struct super_block *sb, | ||
3403 | |||
3404 | erofs_register_super(sb); | ||
3405 | |||
3406 | - /* | ||
3407 | - * We already have a positive dentry, which was instantiated | ||
3408 | - * by d_make_root. Just need to d_rehash it. | ||
3409 | - */ | ||
3410 | - d_rehash(sb->s_root); | ||
3411 | - | ||
3412 | if (!silent) | ||
3413 | infoln("mounted on %s with opts: %s.", dev_name, | ||
3414 | (char *)data); | ||
3415 | @@ -655,7 +648,7 @@ static int erofs_remount(struct super_block *sb, int *flags, char *data) | ||
3416 | unsigned int org_inject_rate = erofs_get_fault_rate(sbi); | ||
3417 | int err; | ||
3418 | |||
3419 | - BUG_ON(!sb_rdonly(sb)); | ||
3420 | + DBG_BUGON(!sb_rdonly(sb)); | ||
3421 | err = parse_options(sb, data); | ||
3422 | if (err) | ||
3423 | goto out; | ||
3424 | diff --git a/drivers/staging/erofs/unzip_pagevec.h b/drivers/staging/erofs/unzip_pagevec.h | ||
3425 | index 0956615b86f72..23856ba2742d8 100644 | ||
3426 | --- a/drivers/staging/erofs/unzip_pagevec.h | ||
3427 | +++ b/drivers/staging/erofs/unzip_pagevec.h | ||
3428 | @@ -150,7 +150,7 @@ z_erofs_pagevec_ctor_dequeue(struct z_erofs_pagevec_ctor *ctor, | ||
3429 | erofs_vtptr_t t; | ||
3430 | |||
3431 | if (unlikely(ctor->index >= ctor->nr)) { | ||
3432 | - BUG_ON(ctor->next == NULL); | ||
3433 | + DBG_BUGON(!ctor->next); | ||
3434 | z_erofs_pagevec_ctor_pagedown(ctor, true); | ||
3435 | } | ||
3436 | |||
3437 | diff --git a/drivers/staging/erofs/unzip_vle.c b/drivers/staging/erofs/unzip_vle.c | ||
3438 | index 45e88bada907f..1c4b3e0343f58 100644 | ||
3439 | --- a/drivers/staging/erofs/unzip_vle.c | ||
3440 | +++ b/drivers/staging/erofs/unzip_vle.c | ||
3441 | @@ -20,9 +20,6 @@ static struct kmem_cache *z_erofs_workgroup_cachep __read_mostly; | ||
3442 | |||
3443 | void z_erofs_exit_zip_subsystem(void) | ||
3444 | { | ||
3445 | - BUG_ON(z_erofs_workqueue == NULL); | ||
3446 | - BUG_ON(z_erofs_workgroup_cachep == NULL); | ||
3447 | - | ||
3448 | destroy_workqueue(z_erofs_workqueue); | ||
3449 | kmem_cache_destroy(z_erofs_workgroup_cachep); | ||
3450 | } | ||
3451 | @@ -366,7 +363,10 @@ z_erofs_vle_work_register(const struct z_erofs_vle_work_finder *f, | ||
3452 | struct z_erofs_vle_work *work; | ||
3453 | |||
3454 | /* if multiref is disabled, grp should never be nullptr */ | ||
3455 | - BUG_ON(grp != NULL); | ||
3456 | + if (unlikely(grp)) { | ||
3457 | + DBG_BUGON(1); | ||
3458 | + return ERR_PTR(-EINVAL); | ||
3459 | + } | ||
3460 | |||
3461 | /* no available workgroup, let's allocate one */ | ||
3462 | grp = kmem_cache_zalloc(z_erofs_workgroup_cachep, GFP_NOFS); | ||
3463 | @@ -745,7 +745,7 @@ static inline void z_erofs_vle_read_endio(struct bio *bio) | ||
3464 | bool cachemngd = false; | ||
3465 | |||
3466 | DBG_BUGON(PageUptodate(page)); | ||
3467 | - BUG_ON(page->mapping == NULL); | ||
3468 | + DBG_BUGON(!page->mapping); | ||
3469 | |||
3470 | #ifdef EROFS_FS_HAS_MANAGED_CACHE | ||
3471 | if (unlikely(mngda == NULL && !z_erofs_is_stagingpage(page))) { | ||
3472 | @@ -803,7 +803,7 @@ static int z_erofs_vle_unzip(struct super_block *sb, | ||
3473 | |||
3474 | might_sleep(); | ||
3475 | work = z_erofs_vle_grab_primary_work(grp); | ||
3476 | - BUG_ON(!READ_ONCE(work->nr_pages)); | ||
3477 | + DBG_BUGON(!READ_ONCE(work->nr_pages)); | ||
3478 | |||
3479 | mutex_lock(&work->lock); | ||
3480 | nr_pages = work->nr_pages; | ||
3481 | @@ -852,8 +852,8 @@ repeat: | ||
3482 | else | ||
3483 | pagenr = z_erofs_onlinepage_index(page); | ||
3484 | |||
3485 | - BUG_ON(pagenr >= nr_pages); | ||
3486 | - BUG_ON(pages[pagenr] != NULL); | ||
3487 | + DBG_BUGON(pagenr >= nr_pages); | ||
3488 | + DBG_BUGON(pages[pagenr]); | ||
3489 | |||
3490 | pages[pagenr] = page; | ||
3491 | } | ||
3492 | @@ -876,9 +876,8 @@ repeat: | ||
3493 | if (z_erofs_is_stagingpage(page)) | ||
3494 | continue; | ||
3495 | #ifdef EROFS_FS_HAS_MANAGED_CACHE | ||
3496 | - else if (page->mapping == mngda) { | ||
3497 | - BUG_ON(PageLocked(page)); | ||
3498 | - BUG_ON(!PageUptodate(page)); | ||
3499 | + if (page->mapping == mngda) { | ||
3500 | + DBG_BUGON(!PageUptodate(page)); | ||
3501 | continue; | ||
3502 | } | ||
3503 | #endif | ||
3504 | @@ -886,8 +885,8 @@ repeat: | ||
3505 | /* only non-head page could be reused as a compressed page */ | ||
3506 | pagenr = z_erofs_onlinepage_index(page); | ||
3507 | |||
3508 | - BUG_ON(pagenr >= nr_pages); | ||
3509 | - BUG_ON(pages[pagenr] != NULL); | ||
3510 | + DBG_BUGON(pagenr >= nr_pages); | ||
3511 | + DBG_BUGON(pages[pagenr]); | ||
3512 | ++sparsemem_pages; | ||
3513 | pages[pagenr] = page; | ||
3514 | |||
3515 | @@ -897,9 +896,6 @@ repeat: | ||
3516 | llen = (nr_pages << PAGE_SHIFT) - work->pageofs; | ||
3517 | |||
3518 | if (z_erofs_vle_workgrp_fmt(grp) == Z_EROFS_VLE_WORKGRP_FMT_PLAIN) { | ||
3519 | - /* FIXME! this should be fixed in the future */ | ||
3520 | - BUG_ON(grp->llen != llen); | ||
3521 | - | ||
3522 | err = z_erofs_vle_plain_copy(compressed_pages, clusterpages, | ||
3523 | pages, nr_pages, work->pageofs); | ||
3524 | goto out; | ||
3525 | @@ -914,10 +910,8 @@ repeat: | ||
3526 | if (err != -ENOTSUPP) | ||
3527 | goto out_percpu; | ||
3528 | |||
3529 | - if (sparsemem_pages >= nr_pages) { | ||
3530 | - BUG_ON(sparsemem_pages > nr_pages); | ||
3531 | + if (sparsemem_pages >= nr_pages) | ||
3532 | goto skip_allocpage; | ||
3533 | - } | ||
3534 | |||
3535 | for (i = 0; i < nr_pages; ++i) { | ||
3536 | if (pages[i] != NULL) | ||
3537 | @@ -1010,7 +1004,7 @@ static void z_erofs_vle_unzip_wq(struct work_struct *work) | ||
3538 | struct z_erofs_vle_unzip_io_sb, io.u.work); | ||
3539 | LIST_HEAD(page_pool); | ||
3540 | |||
3541 | - BUG_ON(iosb->io.head == Z_EROFS_VLE_WORKGRP_TAIL_CLOSED); | ||
3542 | + DBG_BUGON(iosb->io.head == Z_EROFS_VLE_WORKGRP_TAIL_CLOSED); | ||
3543 | z_erofs_vle_unzip_all(iosb->sb, &iosb->io, &page_pool); | ||
3544 | |||
3545 | put_pages_list(&page_pool); | ||
3546 | @@ -1344,7 +1338,6 @@ static int z_erofs_vle_normalaccess_readpages(struct file *filp, | ||
3547 | continue; | ||
3548 | } | ||
3549 | |||
3550 | - BUG_ON(PagePrivate(page)); | ||
3551 | set_page_private(page, (unsigned long)head); | ||
3552 | head = page; | ||
3553 | } | ||
3554 | diff --git a/drivers/staging/erofs/unzip_vle_lz4.c b/drivers/staging/erofs/unzip_vle_lz4.c | ||
3555 | index 1a428658cbea2..16ac335ee59f4 100644 | ||
3556 | --- a/drivers/staging/erofs/unzip_vle_lz4.c | ||
3557 | +++ b/drivers/staging/erofs/unzip_vle_lz4.c | ||
3558 | @@ -57,7 +57,7 @@ int z_erofs_vle_plain_copy(struct page **compressed_pages, | ||
3559 | if (compressed_pages[j] != page) | ||
3560 | continue; | ||
3561 | |||
3562 | - BUG_ON(mirrored[j]); | ||
3563 | + DBG_BUGON(mirrored[j]); | ||
3564 | memcpy(percpu_data + j * PAGE_SIZE, dst, PAGE_SIZE); | ||
3565 | mirrored[j] = true; | ||
3566 | break; | ||
3567 | diff --git a/drivers/staging/erofs/utils.c b/drivers/staging/erofs/utils.c | ||
3568 | index d2e3ace910469..b535898ca753f 100644 | ||
3569 | --- a/drivers/staging/erofs/utils.c | ||
3570 | +++ b/drivers/staging/erofs/utils.c | ||
3571 | @@ -23,9 +23,6 @@ struct page *erofs_allocpage(struct list_head *pool, gfp_t gfp) | ||
3572 | list_del(&page->lru); | ||
3573 | } else { | ||
3574 | page = alloc_pages(gfp | __GFP_NOFAIL, 0); | ||
3575 | - | ||
3576 | - BUG_ON(page == NULL); | ||
3577 | - BUG_ON(page->mapping != NULL); | ||
3578 | } | ||
3579 | return page; | ||
3580 | } | ||
3581 | @@ -58,7 +55,7 @@ repeat: | ||
3582 | /* decrease refcount added by erofs_workgroup_put */ | ||
3583 | if (unlikely(oldcount == 1)) | ||
3584 | atomic_long_dec(&erofs_global_shrink_cnt); | ||
3585 | - BUG_ON(index != grp->index); | ||
3586 | + DBG_BUGON(index != grp->index); | ||
3587 | } | ||
3588 | rcu_read_unlock(); | ||
3589 | return grp; | ||
3590 | @@ -71,8 +68,11 @@ int erofs_register_workgroup(struct super_block *sb, | ||
3591 | struct erofs_sb_info *sbi; | ||
3592 | int err; | ||
3593 | |||
3594 | - /* grp->refcount should not < 1 */ | ||
3595 | - BUG_ON(!atomic_read(&grp->refcount)); | ||
3596 | + /* grp shouldn't be broken or used before */ | ||
3597 | + if (unlikely(atomic_read(&grp->refcount) != 1)) { | ||
3598 | + DBG_BUGON(1); | ||
3599 | + return -EINVAL; | ||
3600 | + } | ||
3601 | |||
3602 | err = radix_tree_preload(GFP_NOFS); | ||
3603 | if (err) | ||
3604 | diff --git a/drivers/target/target_core_user.c b/drivers/target/target_core_user.c | ||
3605 | index 9cd404acdb82b..ac7620120491b 100644 | ||
3606 | --- a/drivers/target/target_core_user.c | ||
3607 | +++ b/drivers/target/target_core_user.c | ||
3608 | @@ -148,7 +148,7 @@ struct tcmu_dev { | ||
3609 | size_t ring_size; | ||
3610 | |||
3611 | struct mutex cmdr_lock; | ||
3612 | - struct list_head cmdr_queue; | ||
3613 | + struct list_head qfull_queue; | ||
3614 | |||
3615 | uint32_t dbi_max; | ||
3616 | uint32_t dbi_thresh; | ||
3617 | @@ -159,6 +159,7 @@ struct tcmu_dev { | ||
3618 | |||
3619 | struct timer_list cmd_timer; | ||
3620 | unsigned int cmd_time_out; | ||
3621 | + struct list_head inflight_queue; | ||
3622 | |||
3623 | struct timer_list qfull_timer; | ||
3624 | int qfull_time_out; | ||
3625 | @@ -179,7 +180,7 @@ struct tcmu_dev { | ||
3626 | struct tcmu_cmd { | ||
3627 | struct se_cmd *se_cmd; | ||
3628 | struct tcmu_dev *tcmu_dev; | ||
3629 | - struct list_head cmdr_queue_entry; | ||
3630 | + struct list_head queue_entry; | ||
3631 | |||
3632 | uint16_t cmd_id; | ||
3633 | |||
3634 | @@ -192,6 +193,7 @@ struct tcmu_cmd { | ||
3635 | unsigned long deadline; | ||
3636 | |||
3637 | #define TCMU_CMD_BIT_EXPIRED 0 | ||
3638 | +#define TCMU_CMD_BIT_INFLIGHT 1 | ||
3639 | unsigned long flags; | ||
3640 | }; | ||
3641 | /* | ||
3642 | @@ -586,7 +588,7 @@ static struct tcmu_cmd *tcmu_alloc_cmd(struct se_cmd *se_cmd) | ||
3643 | if (!tcmu_cmd) | ||
3644 | return NULL; | ||
3645 | |||
3646 | - INIT_LIST_HEAD(&tcmu_cmd->cmdr_queue_entry); | ||
3647 | + INIT_LIST_HEAD(&tcmu_cmd->queue_entry); | ||
3648 | tcmu_cmd->se_cmd = se_cmd; | ||
3649 | tcmu_cmd->tcmu_dev = udev; | ||
3650 | |||
3651 | @@ -915,11 +917,13 @@ setup_timer: | ||
3652 | return 0; | ||
3653 | |||
3654 | tcmu_cmd->deadline = round_jiffies_up(jiffies + msecs_to_jiffies(tmo)); | ||
3655 | - mod_timer(timer, tcmu_cmd->deadline); | ||
3656 | + if (!timer_pending(timer)) | ||
3657 | + mod_timer(timer, tcmu_cmd->deadline); | ||
3658 | + | ||
3659 | return 0; | ||
3660 | } | ||
3661 | |||
3662 | -static int add_to_cmdr_queue(struct tcmu_cmd *tcmu_cmd) | ||
3663 | +static int add_to_qfull_queue(struct tcmu_cmd *tcmu_cmd) | ||
3664 | { | ||
3665 | struct tcmu_dev *udev = tcmu_cmd->tcmu_dev; | ||
3666 | unsigned int tmo; | ||
3667 | @@ -942,7 +946,7 @@ static int add_to_cmdr_queue(struct tcmu_cmd *tcmu_cmd) | ||
3668 | if (ret) | ||
3669 | return ret; | ||
3670 | |||
3671 | - list_add_tail(&tcmu_cmd->cmdr_queue_entry, &udev->cmdr_queue); | ||
3672 | + list_add_tail(&tcmu_cmd->queue_entry, &udev->qfull_queue); | ||
3673 | pr_debug("adding cmd %u on dev %s to ring space wait queue\n", | ||
3674 | tcmu_cmd->cmd_id, udev->name); | ||
3675 | return 0; | ||
3676 | @@ -999,7 +1003,7 @@ static sense_reason_t queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, int *scsi_err) | ||
3677 | base_command_size = tcmu_cmd_get_base_cmd_size(tcmu_cmd->dbi_cnt); | ||
3678 | command_size = tcmu_cmd_get_cmd_size(tcmu_cmd, base_command_size); | ||
3679 | |||
3680 | - if (!list_empty(&udev->cmdr_queue)) | ||
3681 | + if (!list_empty(&udev->qfull_queue)) | ||
3682 | goto queue; | ||
3683 | |||
3684 | mb = udev->mb_addr; | ||
3685 | @@ -1096,13 +1100,16 @@ static sense_reason_t queue_cmd_ring(struct tcmu_cmd *tcmu_cmd, int *scsi_err) | ||
3686 | UPDATE_HEAD(mb->cmd_head, command_size, udev->cmdr_size); | ||
3687 | tcmu_flush_dcache_range(mb, sizeof(*mb)); | ||
3688 | |||
3689 | + list_add_tail(&tcmu_cmd->queue_entry, &udev->inflight_queue); | ||
3690 | + set_bit(TCMU_CMD_BIT_INFLIGHT, &tcmu_cmd->flags); | ||
3691 | + | ||
3692 | /* TODO: only if FLUSH and FUA? */ | ||
3693 | uio_event_notify(&udev->uio_info); | ||
3694 | |||
3695 | return 0; | ||
3696 | |||
3697 | queue: | ||
3698 | - if (add_to_cmdr_queue(tcmu_cmd)) { | ||
3699 | + if (add_to_qfull_queue(tcmu_cmd)) { | ||
3700 | *scsi_err = TCM_OUT_OF_RESOURCES; | ||
3701 | return -1; | ||
3702 | } | ||
3703 | @@ -1145,6 +1152,8 @@ static void tcmu_handle_completion(struct tcmu_cmd *cmd, struct tcmu_cmd_entry * | ||
3704 | if (test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags)) | ||
3705 | goto out; | ||
3706 | |||
3707 | + list_del_init(&cmd->queue_entry); | ||
3708 | + | ||
3709 | tcmu_cmd_reset_dbi_cur(cmd); | ||
3710 | |||
3711 | if (entry->hdr.uflags & TCMU_UFLAG_UNKNOWN_OP) { | ||
3712 | @@ -1194,9 +1203,29 @@ out: | ||
3713 | tcmu_free_cmd(cmd); | ||
3714 | } | ||
3715 | |||
3716 | +static void tcmu_set_next_deadline(struct list_head *queue, | ||
3717 | + struct timer_list *timer) | ||
3718 | +{ | ||
3719 | + struct tcmu_cmd *tcmu_cmd, *tmp_cmd; | ||
3720 | + unsigned long deadline = 0; | ||
3721 | + | ||
3722 | + list_for_each_entry_safe(tcmu_cmd, tmp_cmd, queue, queue_entry) { | ||
3723 | + if (!time_after(jiffies, tcmu_cmd->deadline)) { | ||
3724 | + deadline = tcmu_cmd->deadline; | ||
3725 | + break; | ||
3726 | + } | ||
3727 | + } | ||
3728 | + | ||
3729 | + if (deadline) | ||
3730 | + mod_timer(timer, deadline); | ||
3731 | + else | ||
3732 | + del_timer(timer); | ||
3733 | +} | ||
3734 | + | ||
3735 | static unsigned int tcmu_handle_completions(struct tcmu_dev *udev) | ||
3736 | { | ||
3737 | struct tcmu_mailbox *mb; | ||
3738 | + struct tcmu_cmd *cmd; | ||
3739 | int handled = 0; | ||
3740 | |||
3741 | if (test_bit(TCMU_DEV_BIT_BROKEN, &udev->flags)) { | ||
3742 | @@ -1210,7 +1239,6 @@ static unsigned int tcmu_handle_completions(struct tcmu_dev *udev) | ||
3743 | while (udev->cmdr_last_cleaned != READ_ONCE(mb->cmd_tail)) { | ||
3744 | |||
3745 | struct tcmu_cmd_entry *entry = (void *) mb + CMDR_OFF + udev->cmdr_last_cleaned; | ||
3746 | - struct tcmu_cmd *cmd; | ||
3747 | |||
3748 | tcmu_flush_dcache_range(entry, sizeof(*entry)); | ||
3749 | |||
3750 | @@ -1243,7 +1271,7 @@ static unsigned int tcmu_handle_completions(struct tcmu_dev *udev) | ||
3751 | /* no more pending commands */ | ||
3752 | del_timer(&udev->cmd_timer); | ||
3753 | |||
3754 | - if (list_empty(&udev->cmdr_queue)) { | ||
3755 | + if (list_empty(&udev->qfull_queue)) { | ||
3756 | /* | ||
3757 | * no more pending or waiting commands so try to | ||
3758 | * reclaim blocks if needed. | ||
3759 | @@ -1252,6 +1280,8 @@ static unsigned int tcmu_handle_completions(struct tcmu_dev *udev) | ||
3760 | tcmu_global_max_blocks) | ||
3761 | schedule_delayed_work(&tcmu_unmap_work, 0); | ||
3762 | } | ||
3763 | + } else if (udev->cmd_time_out) { | ||
3764 | + tcmu_set_next_deadline(&udev->inflight_queue, &udev->cmd_timer); | ||
3765 | } | ||
3766 | |||
3767 | return handled; | ||
3768 | @@ -1271,7 +1301,7 @@ static int tcmu_check_expired_cmd(int id, void *p, void *data) | ||
3769 | if (!time_after(jiffies, cmd->deadline)) | ||
3770 | return 0; | ||
3771 | |||
3772 | - is_running = list_empty(&cmd->cmdr_queue_entry); | ||
3773 | + is_running = test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags); | ||
3774 | se_cmd = cmd->se_cmd; | ||
3775 | |||
3776 | if (is_running) { | ||
3777 | @@ -1288,12 +1318,11 @@ static int tcmu_check_expired_cmd(int id, void *p, void *data) | ||
3778 | */ | ||
3779 | scsi_status = SAM_STAT_CHECK_CONDITION; | ||
3780 | } else { | ||
3781 | - list_del_init(&cmd->cmdr_queue_entry); | ||
3782 | - | ||
3783 | idr_remove(&udev->commands, id); | ||
3784 | tcmu_free_cmd(cmd); | ||
3785 | scsi_status = SAM_STAT_TASK_SET_FULL; | ||
3786 | } | ||
3787 | + list_del_init(&cmd->queue_entry); | ||
3788 | |||
3789 | pr_debug("Timing out cmd %u on dev %s that is %s.\n", | ||
3790 | id, udev->name, is_running ? "inflight" : "queued"); | ||
3791 | @@ -1372,7 +1401,8 @@ static struct se_device *tcmu_alloc_device(struct se_hba *hba, const char *name) | ||
3792 | |||
3793 | INIT_LIST_HEAD(&udev->node); | ||
3794 | INIT_LIST_HEAD(&udev->timedout_entry); | ||
3795 | - INIT_LIST_HEAD(&udev->cmdr_queue); | ||
3796 | + INIT_LIST_HEAD(&udev->qfull_queue); | ||
3797 | + INIT_LIST_HEAD(&udev->inflight_queue); | ||
3798 | idr_init(&udev->commands); | ||
3799 | |||
3800 | timer_setup(&udev->qfull_timer, tcmu_qfull_timedout, 0); | ||
3801 | @@ -1383,7 +1413,7 @@ static struct se_device *tcmu_alloc_device(struct se_hba *hba, const char *name) | ||
3802 | return &udev->se_dev; | ||
3803 | } | ||
3804 | |||
3805 | -static bool run_cmdr_queue(struct tcmu_dev *udev, bool fail) | ||
3806 | +static bool run_qfull_queue(struct tcmu_dev *udev, bool fail) | ||
3807 | { | ||
3808 | struct tcmu_cmd *tcmu_cmd, *tmp_cmd; | ||
3809 | LIST_HEAD(cmds); | ||
3810 | @@ -1391,15 +1421,15 @@ static bool run_cmdr_queue(struct tcmu_dev *udev, bool fail) | ||
3811 | sense_reason_t scsi_ret; | ||
3812 | int ret; | ||
3813 | |||
3814 | - if (list_empty(&udev->cmdr_queue)) | ||
3815 | + if (list_empty(&udev->qfull_queue)) | ||
3816 | return true; | ||
3817 | |||
3818 | pr_debug("running %s's cmdr queue forcefail %d\n", udev->name, fail); | ||
3819 | |||
3820 | - list_splice_init(&udev->cmdr_queue, &cmds); | ||
3821 | + list_splice_init(&udev->qfull_queue, &cmds); | ||
3822 | |||
3823 | - list_for_each_entry_safe(tcmu_cmd, tmp_cmd, &cmds, cmdr_queue_entry) { | ||
3824 | - list_del_init(&tcmu_cmd->cmdr_queue_entry); | ||
3825 | + list_for_each_entry_safe(tcmu_cmd, tmp_cmd, &cmds, queue_entry) { | ||
3826 | + list_del_init(&tcmu_cmd->queue_entry); | ||
3827 | |||
3828 | pr_debug("removing cmd %u on dev %s from queue\n", | ||
3829 | tcmu_cmd->cmd_id, udev->name); | ||
3830 | @@ -1437,14 +1467,13 @@ static bool run_cmdr_queue(struct tcmu_dev *udev, bool fail) | ||
3831 | * cmd was requeued, so just put all cmds back in | ||
3832 | * the queue | ||
3833 | */ | ||
3834 | - list_splice_tail(&cmds, &udev->cmdr_queue); | ||
3835 | + list_splice_tail(&cmds, &udev->qfull_queue); | ||
3836 | drained = false; | ||
3837 | - goto done; | ||
3838 | + break; | ||
3839 | } | ||
3840 | } | ||
3841 | - if (list_empty(&udev->cmdr_queue)) | ||
3842 | - del_timer(&udev->qfull_timer); | ||
3843 | -done: | ||
3844 | + | ||
3845 | + tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer); | ||
3846 | return drained; | ||
3847 | } | ||
3848 | |||
3849 | @@ -1454,7 +1483,7 @@ static int tcmu_irqcontrol(struct uio_info *info, s32 irq_on) | ||
3850 | |||
3851 | mutex_lock(&udev->cmdr_lock); | ||
3852 | tcmu_handle_completions(udev); | ||
3853 | - run_cmdr_queue(udev, false); | ||
3854 | + run_qfull_queue(udev, false); | ||
3855 | mutex_unlock(&udev->cmdr_lock); | ||
3856 | |||
3857 | return 0; | ||
3858 | @@ -1982,7 +2011,7 @@ static void tcmu_block_dev(struct tcmu_dev *udev) | ||
3859 | /* complete IO that has executed successfully */ | ||
3860 | tcmu_handle_completions(udev); | ||
3861 | /* fail IO waiting to be queued */ | ||
3862 | - run_cmdr_queue(udev, true); | ||
3863 | + run_qfull_queue(udev, true); | ||
3864 | |||
3865 | unlock: | ||
3866 | mutex_unlock(&udev->cmdr_lock); | ||
3867 | @@ -1997,7 +2026,7 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level) | ||
3868 | mutex_lock(&udev->cmdr_lock); | ||
3869 | |||
3870 | idr_for_each_entry(&udev->commands, cmd, i) { | ||
3871 | - if (!list_empty(&cmd->cmdr_queue_entry)) | ||
3872 | + if (!test_bit(TCMU_CMD_BIT_INFLIGHT, &cmd->flags)) | ||
3873 | continue; | ||
3874 | |||
3875 | pr_debug("removing cmd %u on dev %s from ring (is expired %d)\n", | ||
3876 | @@ -2006,6 +2035,7 @@ static void tcmu_reset_ring(struct tcmu_dev *udev, u8 err_level) | ||
3877 | |||
3878 | idr_remove(&udev->commands, i); | ||
3879 | if (!test_bit(TCMU_CMD_BIT_EXPIRED, &cmd->flags)) { | ||
3880 | + list_del_init(&cmd->queue_entry); | ||
3881 | if (err_level == 1) { | ||
3882 | /* | ||
3883 | * Userspace was not able to start the | ||
3884 | @@ -2666,6 +2696,10 @@ static void check_timedout_devices(void) | ||
3885 | |||
3886 | mutex_lock(&udev->cmdr_lock); | ||
3887 | idr_for_each(&udev->commands, tcmu_check_expired_cmd, NULL); | ||
3888 | + | ||
3889 | + tcmu_set_next_deadline(&udev->inflight_queue, &udev->cmd_timer); | ||
3890 | + tcmu_set_next_deadline(&udev->qfull_queue, &udev->qfull_timer); | ||
3891 | + | ||
3892 | mutex_unlock(&udev->cmdr_lock); | ||
3893 | |||
3894 | spin_lock_bh(&timed_out_udevs_lock); | ||
3895 | diff --git a/drivers/vhost/scsi.c b/drivers/vhost/scsi.c | ||
3896 | index 73a4adeab096b..11bd8b6422ebf 100644 | ||
3897 | --- a/drivers/vhost/scsi.c | ||
3898 | +++ b/drivers/vhost/scsi.c | ||
3899 | @@ -1132,16 +1132,18 @@ vhost_scsi_send_tmf_reject(struct vhost_scsi *vs, | ||
3900 | struct vhost_virtqueue *vq, | ||
3901 | struct vhost_scsi_ctx *vc) | ||
3902 | { | ||
3903 | - struct virtio_scsi_ctrl_tmf_resp __user *resp; | ||
3904 | struct virtio_scsi_ctrl_tmf_resp rsp; | ||
3905 | + struct iov_iter iov_iter; | ||
3906 | int ret; | ||
3907 | |||
3908 | pr_debug("%s\n", __func__); | ||
3909 | memset(&rsp, 0, sizeof(rsp)); | ||
3910 | rsp.response = VIRTIO_SCSI_S_FUNCTION_REJECTED; | ||
3911 | - resp = vq->iov[vc->out].iov_base; | ||
3912 | - ret = __copy_to_user(resp, &rsp, sizeof(rsp)); | ||
3913 | - if (!ret) | ||
3914 | + | ||
3915 | + iov_iter_init(&iov_iter, READ, &vq->iov[vc->out], vc->in, sizeof(rsp)); | ||
3916 | + | ||
3917 | + ret = copy_to_iter(&rsp, sizeof(rsp), &iov_iter); | ||
3918 | + if (likely(ret == sizeof(rsp))) | ||
3919 | vhost_add_used_and_signal(&vs->dev, vq, vc->head, 0); | ||
3920 | else | ||
3921 | pr_err("Faulted on virtio_scsi_ctrl_tmf_resp\n"); | ||
3922 | @@ -1152,16 +1154,18 @@ vhost_scsi_send_an_resp(struct vhost_scsi *vs, | ||
3923 | struct vhost_virtqueue *vq, | ||
3924 | struct vhost_scsi_ctx *vc) | ||
3925 | { | ||
3926 | - struct virtio_scsi_ctrl_an_resp __user *resp; | ||
3927 | struct virtio_scsi_ctrl_an_resp rsp; | ||
3928 | + struct iov_iter iov_iter; | ||
3929 | int ret; | ||
3930 | |||
3931 | pr_debug("%s\n", __func__); | ||
3932 | memset(&rsp, 0, sizeof(rsp)); /* event_actual = 0 */ | ||
3933 | rsp.response = VIRTIO_SCSI_S_OK; | ||
3934 | - resp = vq->iov[vc->out].iov_base; | ||
3935 | - ret = __copy_to_user(resp, &rsp, sizeof(rsp)); | ||
3936 | - if (!ret) | ||
3937 | + | ||
3938 | + iov_iter_init(&iov_iter, READ, &vq->iov[vc->out], vc->in, sizeof(rsp)); | ||
3939 | + | ||
3940 | + ret = copy_to_iter(&rsp, sizeof(rsp), &iov_iter); | ||
3941 | + if (likely(ret == sizeof(rsp))) | ||
3942 | vhost_add_used_and_signal(&vs->dev, vq, vc->head, 0); | ||
3943 | else | ||
3944 | pr_err("Faulted on virtio_scsi_ctrl_an_resp\n"); | ||
3945 | diff --git a/drivers/vhost/vhost.c b/drivers/vhost/vhost.c | ||
3946 | index 5eaeca805c95c..b214a72d5caad 100644 | ||
3947 | --- a/drivers/vhost/vhost.c | ||
3948 | +++ b/drivers/vhost/vhost.c | ||
3949 | @@ -1035,8 +1035,10 @@ ssize_t vhost_chr_write_iter(struct vhost_dev *dev, | ||
3950 | int type, ret; | ||
3951 | |||
3952 | ret = copy_from_iter(&type, sizeof(type), from); | ||
3953 | - if (ret != sizeof(type)) | ||
3954 | + if (ret != sizeof(type)) { | ||
3955 | + ret = -EINVAL; | ||
3956 | goto done; | ||
3957 | + } | ||
3958 | |||
3959 | switch (type) { | ||
3960 | case VHOST_IOTLB_MSG: | ||
3961 | @@ -1055,8 +1057,10 @@ ssize_t vhost_chr_write_iter(struct vhost_dev *dev, | ||
3962 | |||
3963 | iov_iter_advance(from, offset); | ||
3964 | ret = copy_from_iter(&msg, sizeof(msg), from); | ||
3965 | - if (ret != sizeof(msg)) | ||
3966 | + if (ret != sizeof(msg)) { | ||
3967 | + ret = -EINVAL; | ||
3968 | goto done; | ||
3969 | + } | ||
3970 | if (vhost_process_iotlb_msg(dev, &msg)) { | ||
3971 | ret = -EFAULT; | ||
3972 | goto done; | ||
3973 | diff --git a/drivers/video/backlight/pwm_bl.c b/drivers/video/backlight/pwm_bl.c | ||
3974 | index f9ef0673a083c..aded3213bfb51 100644 | ||
3975 | --- a/drivers/video/backlight/pwm_bl.c | ||
3976 | +++ b/drivers/video/backlight/pwm_bl.c | ||
3977 | @@ -268,6 +268,16 @@ static int pwm_backlight_parse_dt(struct device *dev, | ||
3978 | |||
3979 | memset(data, 0, sizeof(*data)); | ||
3980 | |||
3981 | + /* | ||
3982 | + * These values are optional and set as 0 by default, the out values | ||
3983 | + * are modified only if a valid u32 value can be decoded. | ||
3984 | + */ | ||
3985 | + of_property_read_u32(node, "post-pwm-on-delay-ms", | ||
3986 | + &data->post_pwm_on_delay); | ||
3987 | + of_property_read_u32(node, "pwm-off-delay-ms", &data->pwm_off_delay); | ||
3988 | + | ||
3989 | + data->enable_gpio = -EINVAL; | ||
3990 | + | ||
3991 | /* | ||
3992 | * Determine the number of brightness levels, if this property is not | ||
3993 | * set a default table of brightness levels will be used. | ||
3994 | @@ -380,15 +390,6 @@ static int pwm_backlight_parse_dt(struct device *dev, | ||
3995 | data->max_brightness--; | ||
3996 | } | ||
3997 | |||
3998 | - /* | ||
3999 | - * These values are optional and set as 0 by default, the out values | ||
4000 | - * are modified only if a valid u32 value can be decoded. | ||
4001 | - */ | ||
4002 | - of_property_read_u32(node, "post-pwm-on-delay-ms", | ||
4003 | - &data->post_pwm_on_delay); | ||
4004 | - of_property_read_u32(node, "pwm-off-delay-ms", &data->pwm_off_delay); | ||
4005 | - | ||
4006 | - data->enable_gpio = -EINVAL; | ||
4007 | return 0; | ||
4008 | } | ||
4009 | |||
4010 | diff --git a/drivers/watchdog/mt7621_wdt.c b/drivers/watchdog/mt7621_wdt.c | ||
4011 | index 5c4a764717c4d..81208cd3f4ecb 100644 | ||
4012 | --- a/drivers/watchdog/mt7621_wdt.c | ||
4013 | +++ b/drivers/watchdog/mt7621_wdt.c | ||
4014 | @@ -17,6 +17,7 @@ | ||
4015 | #include <linux/watchdog.h> | ||
4016 | #include <linux/moduleparam.h> | ||
4017 | #include <linux/platform_device.h> | ||
4018 | +#include <linux/mod_devicetable.h> | ||
4019 | |||
4020 | #include <asm/mach-ralink/ralink_regs.h> | ||
4021 | |||
4022 | diff --git a/drivers/watchdog/rt2880_wdt.c b/drivers/watchdog/rt2880_wdt.c | ||
4023 | index 98967f0a7d10e..db7c57d82cfdc 100644 | ||
4024 | --- a/drivers/watchdog/rt2880_wdt.c | ||
4025 | +++ b/drivers/watchdog/rt2880_wdt.c | ||
4026 | @@ -18,6 +18,7 @@ | ||
4027 | #include <linux/watchdog.h> | ||
4028 | #include <linux/moduleparam.h> | ||
4029 | #include <linux/platform_device.h> | ||
4030 | +#include <linux/mod_devicetable.h> | ||
4031 | |||
4032 | #include <asm/mach-ralink/ralink_regs.h> | ||
4033 | |||
4034 | diff --git a/drivers/xen/pvcalls-back.c b/drivers/xen/pvcalls-back.c | ||
4035 | index 2e5d845b50914..7aa64d1b119c2 100644 | ||
4036 | --- a/drivers/xen/pvcalls-back.c | ||
4037 | +++ b/drivers/xen/pvcalls-back.c | ||
4038 | @@ -160,9 +160,10 @@ static void pvcalls_conn_back_read(void *opaque) | ||
4039 | |||
4040 | /* write the data, then modify the indexes */ | ||
4041 | virt_wmb(); | ||
4042 | - if (ret < 0) | ||
4043 | + if (ret < 0) { | ||
4044 | + atomic_set(&map->read, 0); | ||
4045 | intf->in_error = ret; | ||
4046 | - else | ||
4047 | + } else | ||
4048 | intf->in_prod = prod + ret; | ||
4049 | /* update the indexes, then notify the other end */ | ||
4050 | virt_wmb(); | ||
4051 | @@ -282,13 +283,11 @@ static int pvcalls_back_socket(struct xenbus_device *dev, | ||
4052 | static void pvcalls_sk_state_change(struct sock *sock) | ||
4053 | { | ||
4054 | struct sock_mapping *map = sock->sk_user_data; | ||
4055 | - struct pvcalls_data_intf *intf; | ||
4056 | |||
4057 | if (map == NULL) | ||
4058 | return; | ||
4059 | |||
4060 | - intf = map->ring; | ||
4061 | - intf->in_error = -ENOTCONN; | ||
4062 | + atomic_inc(&map->read); | ||
4063 | notify_remote_via_irq(map->irq); | ||
4064 | } | ||
4065 | |||
4066 | diff --git a/drivers/xen/pvcalls-front.c b/drivers/xen/pvcalls-front.c | ||
4067 | index 77224d8f3e6fe..91da7e44d5d4f 100644 | ||
4068 | --- a/drivers/xen/pvcalls-front.c | ||
4069 | +++ b/drivers/xen/pvcalls-front.c | ||
4070 | @@ -31,6 +31,12 @@ | ||
4071 | #define PVCALLS_NR_RSP_PER_RING __CONST_RING_SIZE(xen_pvcalls, XEN_PAGE_SIZE) | ||
4072 | #define PVCALLS_FRONT_MAX_SPIN 5000 | ||
4073 | |||
4074 | +static struct proto pvcalls_proto = { | ||
4075 | + .name = "PVCalls", | ||
4076 | + .owner = THIS_MODULE, | ||
4077 | + .obj_size = sizeof(struct sock), | ||
4078 | +}; | ||
4079 | + | ||
4080 | struct pvcalls_bedata { | ||
4081 | struct xen_pvcalls_front_ring ring; | ||
4082 | grant_ref_t ref; | ||
4083 | @@ -335,6 +341,42 @@ int pvcalls_front_socket(struct socket *sock) | ||
4084 | return ret; | ||
4085 | } | ||
4086 | |||
4087 | +static void free_active_ring(struct sock_mapping *map) | ||
4088 | +{ | ||
4089 | + if (!map->active.ring) | ||
4090 | + return; | ||
4091 | + | ||
4092 | + free_pages((unsigned long)map->active.data.in, | ||
4093 | + map->active.ring->ring_order); | ||
4094 | + free_page((unsigned long)map->active.ring); | ||
4095 | +} | ||
4096 | + | ||
4097 | +static int alloc_active_ring(struct sock_mapping *map) | ||
4098 | +{ | ||
4099 | + void *bytes; | ||
4100 | + | ||
4101 | + map->active.ring = (struct pvcalls_data_intf *) | ||
4102 | + get_zeroed_page(GFP_KERNEL); | ||
4103 | + if (!map->active.ring) | ||
4104 | + goto out; | ||
4105 | + | ||
4106 | + map->active.ring->ring_order = PVCALLS_RING_ORDER; | ||
4107 | + bytes = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, | ||
4108 | + PVCALLS_RING_ORDER); | ||
4109 | + if (!bytes) | ||
4110 | + goto out; | ||
4111 | + | ||
4112 | + map->active.data.in = bytes; | ||
4113 | + map->active.data.out = bytes + | ||
4114 | + XEN_FLEX_RING_SIZE(PVCALLS_RING_ORDER); | ||
4115 | + | ||
4116 | + return 0; | ||
4117 | + | ||
4118 | +out: | ||
4119 | + free_active_ring(map); | ||
4120 | + return -ENOMEM; | ||
4121 | +} | ||
4122 | + | ||
4123 | static int create_active(struct sock_mapping *map, int *evtchn) | ||
4124 | { | ||
4125 | void *bytes; | ||
4126 | @@ -343,15 +385,7 @@ static int create_active(struct sock_mapping *map, int *evtchn) | ||
4127 | *evtchn = -1; | ||
4128 | init_waitqueue_head(&map->active.inflight_conn_req); | ||
4129 | |||
4130 | - map->active.ring = (struct pvcalls_data_intf *) | ||
4131 | - __get_free_page(GFP_KERNEL | __GFP_ZERO); | ||
4132 | - if (map->active.ring == NULL) | ||
4133 | - goto out_error; | ||
4134 | - map->active.ring->ring_order = PVCALLS_RING_ORDER; | ||
4135 | - bytes = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, | ||
4136 | - PVCALLS_RING_ORDER); | ||
4137 | - if (bytes == NULL) | ||
4138 | - goto out_error; | ||
4139 | + bytes = map->active.data.in; | ||
4140 | for (i = 0; i < (1 << PVCALLS_RING_ORDER); i++) | ||
4141 | map->active.ring->ref[i] = gnttab_grant_foreign_access( | ||
4142 | pvcalls_front_dev->otherend_id, | ||
4143 | @@ -361,10 +395,6 @@ static int create_active(struct sock_mapping *map, int *evtchn) | ||
4144 | pvcalls_front_dev->otherend_id, | ||
4145 | pfn_to_gfn(virt_to_pfn((void *)map->active.ring)), 0); | ||
4146 | |||
4147 | - map->active.data.in = bytes; | ||
4148 | - map->active.data.out = bytes + | ||
4149 | - XEN_FLEX_RING_SIZE(PVCALLS_RING_ORDER); | ||
4150 | - | ||
4151 | ret = xenbus_alloc_evtchn(pvcalls_front_dev, evtchn); | ||
4152 | if (ret) | ||
4153 | goto out_error; | ||
4154 | @@ -385,8 +415,6 @@ static int create_active(struct sock_mapping *map, int *evtchn) | ||
4155 | out_error: | ||
4156 | if (*evtchn >= 0) | ||
4157 | xenbus_free_evtchn(pvcalls_front_dev, *evtchn); | ||
4158 | - free_pages((unsigned long)map->active.data.in, PVCALLS_RING_ORDER); | ||
4159 | - free_page((unsigned long)map->active.ring); | ||
4160 | return ret; | ||
4161 | } | ||
4162 | |||
4163 | @@ -406,17 +434,24 @@ int pvcalls_front_connect(struct socket *sock, struct sockaddr *addr, | ||
4164 | return PTR_ERR(map); | ||
4165 | |||
4166 | bedata = dev_get_drvdata(&pvcalls_front_dev->dev); | ||
4167 | + ret = alloc_active_ring(map); | ||
4168 | + if (ret < 0) { | ||
4169 | + pvcalls_exit_sock(sock); | ||
4170 | + return ret; | ||
4171 | + } | ||
4172 | |||
4173 | spin_lock(&bedata->socket_lock); | ||
4174 | ret = get_request(bedata, &req_id); | ||
4175 | if (ret < 0) { | ||
4176 | spin_unlock(&bedata->socket_lock); | ||
4177 | + free_active_ring(map); | ||
4178 | pvcalls_exit_sock(sock); | ||
4179 | return ret; | ||
4180 | } | ||
4181 | ret = create_active(map, &evtchn); | ||
4182 | if (ret < 0) { | ||
4183 | spin_unlock(&bedata->socket_lock); | ||
4184 | + free_active_ring(map); | ||
4185 | pvcalls_exit_sock(sock); | ||
4186 | return ret; | ||
4187 | } | ||
4188 | @@ -560,15 +595,13 @@ static int __read_ring(struct pvcalls_data_intf *intf, | ||
4189 | error = intf->in_error; | ||
4190 | /* get pointers before reading from the ring */ | ||
4191 | virt_rmb(); | ||
4192 | - if (error < 0) | ||
4193 | - return error; | ||
4194 | |||
4195 | size = pvcalls_queued(prod, cons, array_size); | ||
4196 | masked_prod = pvcalls_mask(prod, array_size); | ||
4197 | masked_cons = pvcalls_mask(cons, array_size); | ||
4198 | |||
4199 | if (size == 0) | ||
4200 | - return 0; | ||
4201 | + return error ?: size; | ||
4202 | |||
4203 | if (len > size) | ||
4204 | len = size; | ||
4205 | @@ -780,25 +813,36 @@ int pvcalls_front_accept(struct socket *sock, struct socket *newsock, int flags) | ||
4206 | } | ||
4207 | } | ||
4208 | |||
4209 | - spin_lock(&bedata->socket_lock); | ||
4210 | - ret = get_request(bedata, &req_id); | ||
4211 | - if (ret < 0) { | ||
4212 | + map2 = kzalloc(sizeof(*map2), GFP_KERNEL); | ||
4213 | + if (map2 == NULL) { | ||
4214 | clear_bit(PVCALLS_FLAG_ACCEPT_INFLIGHT, | ||
4215 | (void *)&map->passive.flags); | ||
4216 | - spin_unlock(&bedata->socket_lock); | ||
4217 | + pvcalls_exit_sock(sock); | ||
4218 | + return -ENOMEM; | ||
4219 | + } | ||
4220 | + ret = alloc_active_ring(map2); | ||
4221 | + if (ret < 0) { | ||
4222 | + clear_bit(PVCALLS_FLAG_ACCEPT_INFLIGHT, | ||
4223 | + (void *)&map->passive.flags); | ||
4224 | + kfree(map2); | ||
4225 | pvcalls_exit_sock(sock); | ||
4226 | return ret; | ||
4227 | } | ||
4228 | - map2 = kzalloc(sizeof(*map2), GFP_ATOMIC); | ||
4229 | - if (map2 == NULL) { | ||
4230 | + spin_lock(&bedata->socket_lock); | ||
4231 | + ret = get_request(bedata, &req_id); | ||
4232 | + if (ret < 0) { | ||
4233 | clear_bit(PVCALLS_FLAG_ACCEPT_INFLIGHT, | ||
4234 | (void *)&map->passive.flags); | ||
4235 | spin_unlock(&bedata->socket_lock); | ||
4236 | + free_active_ring(map2); | ||
4237 | + kfree(map2); | ||
4238 | pvcalls_exit_sock(sock); | ||
4239 | - return -ENOMEM; | ||
4240 | + return ret; | ||
4241 | } | ||
4242 | + | ||
4243 | ret = create_active(map2, &evtchn); | ||
4244 | if (ret < 0) { | ||
4245 | + free_active_ring(map2); | ||
4246 | kfree(map2); | ||
4247 | clear_bit(PVCALLS_FLAG_ACCEPT_INFLIGHT, | ||
4248 | (void *)&map->passive.flags); | ||
4249 | @@ -839,7 +883,7 @@ int pvcalls_front_accept(struct socket *sock, struct socket *newsock, int flags) | ||
4250 | |||
4251 | received: | ||
4252 | map2->sock = newsock; | ||
4253 | - newsock->sk = kzalloc(sizeof(*newsock->sk), GFP_KERNEL); | ||
4254 | + newsock->sk = sk_alloc(sock_net(sock->sk), PF_INET, GFP_KERNEL, &pvcalls_proto, false); | ||
4255 | if (!newsock->sk) { | ||
4256 | bedata->rsp[req_id].req_id = PVCALLS_INVALID_ID; | ||
4257 | map->passive.inflight_req_id = PVCALLS_INVALID_ID; | ||
4258 | @@ -1032,8 +1076,8 @@ int pvcalls_front_release(struct socket *sock) | ||
4259 | spin_lock(&bedata->socket_lock); | ||
4260 | list_del(&map->list); | ||
4261 | spin_unlock(&bedata->socket_lock); | ||
4262 | - if (READ_ONCE(map->passive.inflight_req_id) != | ||
4263 | - PVCALLS_INVALID_ID) { | ||
4264 | + if (READ_ONCE(map->passive.inflight_req_id) != PVCALLS_INVALID_ID && | ||
4265 | + READ_ONCE(map->passive.inflight_req_id) != 0) { | ||
4266 | pvcalls_front_free_map(bedata, | ||
4267 | map->passive.accept_map); | ||
4268 | } | ||
4269 | diff --git a/fs/afs/flock.c b/fs/afs/flock.c | ||
4270 | index 0568fd9868210..e432bd27a2e7b 100644 | ||
4271 | --- a/fs/afs/flock.c | ||
4272 | +++ b/fs/afs/flock.c | ||
4273 | @@ -208,7 +208,7 @@ again: | ||
4274 | /* The new front of the queue now owns the state variables. */ | ||
4275 | next = list_entry(vnode->pending_locks.next, | ||
4276 | struct file_lock, fl_u.afs.link); | ||
4277 | - vnode->lock_key = afs_file_key(next->fl_file); | ||
4278 | + vnode->lock_key = key_get(afs_file_key(next->fl_file)); | ||
4279 | vnode->lock_type = (next->fl_type == F_RDLCK) ? AFS_LOCK_READ : AFS_LOCK_WRITE; | ||
4280 | vnode->lock_state = AFS_VNODE_LOCK_WAITING_FOR_CB; | ||
4281 | goto again; | ||
4282 | @@ -413,7 +413,7 @@ static void afs_dequeue_lock(struct afs_vnode *vnode, struct file_lock *fl) | ||
4283 | /* The new front of the queue now owns the state variables. */ | ||
4284 | next = list_entry(vnode->pending_locks.next, | ||
4285 | struct file_lock, fl_u.afs.link); | ||
4286 | - vnode->lock_key = afs_file_key(next->fl_file); | ||
4287 | + vnode->lock_key = key_get(afs_file_key(next->fl_file)); | ||
4288 | vnode->lock_type = (next->fl_type == F_RDLCK) ? AFS_LOCK_READ : AFS_LOCK_WRITE; | ||
4289 | vnode->lock_state = AFS_VNODE_LOCK_WAITING_FOR_CB; | ||
4290 | afs_lock_may_be_available(vnode); | ||
4291 | diff --git a/fs/afs/inode.c b/fs/afs/inode.c | ||
4292 | index 6b17d36204142..1a4ce07fb406d 100644 | ||
4293 | --- a/fs/afs/inode.c | ||
4294 | +++ b/fs/afs/inode.c | ||
4295 | @@ -414,7 +414,6 @@ int afs_validate(struct afs_vnode *vnode, struct key *key) | ||
4296 | } else if (test_bit(AFS_VNODE_DELETED, &vnode->flags)) { | ||
4297 | valid = true; | ||
4298 | } else { | ||
4299 | - vnode->cb_s_break = vnode->cb_interest->server->cb_s_break; | ||
4300 | vnode->cb_v_break = vnode->volume->cb_v_break; | ||
4301 | valid = false; | ||
4302 | } | ||
4303 | @@ -546,6 +545,8 @@ void afs_evict_inode(struct inode *inode) | ||
4304 | #endif | ||
4305 | |||
4306 | afs_put_permits(rcu_access_pointer(vnode->permit_cache)); | ||
4307 | + key_put(vnode->lock_key); | ||
4308 | + vnode->lock_key = NULL; | ||
4309 | _leave(""); | ||
4310 | } | ||
4311 | |||
4312 | diff --git a/fs/afs/protocol_yfs.h b/fs/afs/protocol_yfs.h | ||
4313 | index 07bc10f076aac..d443e2bfa0946 100644 | ||
4314 | --- a/fs/afs/protocol_yfs.h | ||
4315 | +++ b/fs/afs/protocol_yfs.h | ||
4316 | @@ -161,3 +161,14 @@ struct yfs_xdr_YFSStoreVolumeStatus { | ||
4317 | struct yfs_xdr_u64 max_quota; | ||
4318 | struct yfs_xdr_u64 file_quota; | ||
4319 | } __packed; | ||
4320 | + | ||
4321 | +enum yfs_lock_type { | ||
4322 | + yfs_LockNone = -1, | ||
4323 | + yfs_LockRead = 0, | ||
4324 | + yfs_LockWrite = 1, | ||
4325 | + yfs_LockExtend = 2, | ||
4326 | + yfs_LockRelease = 3, | ||
4327 | + yfs_LockMandatoryRead = 0x100, | ||
4328 | + yfs_LockMandatoryWrite = 0x101, | ||
4329 | + yfs_LockMandatoryExtend = 0x102, | ||
4330 | +}; | ||
4331 | diff --git a/fs/afs/rxrpc.c b/fs/afs/rxrpc.c | ||
4332 | index a7b44863d502e..2c588f9bbbda2 100644 | ||
4333 | --- a/fs/afs/rxrpc.c | ||
4334 | +++ b/fs/afs/rxrpc.c | ||
4335 | @@ -23,6 +23,7 @@ struct workqueue_struct *afs_async_calls; | ||
4336 | static void afs_wake_up_call_waiter(struct sock *, struct rxrpc_call *, unsigned long); | ||
4337 | static long afs_wait_for_call_to_complete(struct afs_call *, struct afs_addr_cursor *); | ||
4338 | static void afs_wake_up_async_call(struct sock *, struct rxrpc_call *, unsigned long); | ||
4339 | +static void afs_delete_async_call(struct work_struct *); | ||
4340 | static void afs_process_async_call(struct work_struct *); | ||
4341 | static void afs_rx_new_call(struct sock *, struct rxrpc_call *, unsigned long); | ||
4342 | static void afs_rx_discard_new_call(struct rxrpc_call *, unsigned long); | ||
4343 | @@ -203,20 +204,26 @@ void afs_put_call(struct afs_call *call) | ||
4344 | } | ||
4345 | } | ||
4346 | |||
4347 | +static struct afs_call *afs_get_call(struct afs_call *call, | ||
4348 | + enum afs_call_trace why) | ||
4349 | +{ | ||
4350 | + int u = atomic_inc_return(&call->usage); | ||
4351 | + | ||
4352 | + trace_afs_call(call, why, u, | ||
4353 | + atomic_read(&call->net->nr_outstanding_calls), | ||
4354 | + __builtin_return_address(0)); | ||
4355 | + return call; | ||
4356 | +} | ||
4357 | + | ||
4358 | /* | ||
4359 | * Queue the call for actual work. | ||
4360 | */ | ||
4361 | static void afs_queue_call_work(struct afs_call *call) | ||
4362 | { | ||
4363 | if (call->type->work) { | ||
4364 | - int u = atomic_inc_return(&call->usage); | ||
4365 | - | ||
4366 | - trace_afs_call(call, afs_call_trace_work, u, | ||
4367 | - atomic_read(&call->net->nr_outstanding_calls), | ||
4368 | - __builtin_return_address(0)); | ||
4369 | - | ||
4370 | INIT_WORK(&call->work, call->type->work); | ||
4371 | |||
4372 | + afs_get_call(call, afs_call_trace_work); | ||
4373 | if (!queue_work(afs_wq, &call->work)) | ||
4374 | afs_put_call(call); | ||
4375 | } | ||
4376 | @@ -398,6 +405,12 @@ long afs_make_call(struct afs_addr_cursor *ac, struct afs_call *call, | ||
4377 | } | ||
4378 | } | ||
4379 | |||
4380 | + /* If the call is going to be asynchronous, we need an extra ref for | ||
4381 | + * the call to hold itself so the caller need not hang on to its ref. | ||
4382 | + */ | ||
4383 | + if (call->async) | ||
4384 | + afs_get_call(call, afs_call_trace_get); | ||
4385 | + | ||
4386 | /* create a call */ | ||
4387 | rxcall = rxrpc_kernel_begin_call(call->net->socket, srx, call->key, | ||
4388 | (unsigned long)call, | ||
4389 | @@ -438,15 +451,17 @@ long afs_make_call(struct afs_addr_cursor *ac, struct afs_call *call, | ||
4390 | goto error_do_abort; | ||
4391 | } | ||
4392 | |||
4393 | - /* at this point, an async call may no longer exist as it may have | ||
4394 | - * already completed */ | ||
4395 | - if (call->async) | ||
4396 | + /* Note that at this point, we may have received the reply or an abort | ||
4397 | + * - and an asynchronous call may already have completed. | ||
4398 | + */ | ||
4399 | + if (call->async) { | ||
4400 | + afs_put_call(call); | ||
4401 | return -EINPROGRESS; | ||
4402 | + } | ||
4403 | |||
4404 | return afs_wait_for_call_to_complete(call, ac); | ||
4405 | |||
4406 | error_do_abort: | ||
4407 | - call->state = AFS_CALL_COMPLETE; | ||
4408 | if (ret != -ECONNABORTED) { | ||
4409 | rxrpc_kernel_abort_call(call->net->socket, rxcall, | ||
4410 | RX_USER_ABORT, ret, "KSD"); | ||
4411 | @@ -463,8 +478,24 @@ error_do_abort: | ||
4412 | error_kill_call: | ||
4413 | if (call->type->done) | ||
4414 | call->type->done(call); | ||
4415 | - afs_put_call(call); | ||
4416 | + | ||
4417 | + /* We need to dispose of the extra ref we grabbed for an async call. | ||
4418 | + * The call, however, might be queued on afs_async_calls and we need to | ||
4419 | + * make sure we don't get any more notifications that might requeue it. | ||
4420 | + */ | ||
4421 | + if (call->rxcall) { | ||
4422 | + rxrpc_kernel_end_call(call->net->socket, call->rxcall); | ||
4423 | + call->rxcall = NULL; | ||
4424 | + } | ||
4425 | + if (call->async) { | ||
4426 | + if (cancel_work_sync(&call->async_work)) | ||
4427 | + afs_put_call(call); | ||
4428 | + afs_put_call(call); | ||
4429 | + } | ||
4430 | + | ||
4431 | ac->error = ret; | ||
4432 | + call->state = AFS_CALL_COMPLETE; | ||
4433 | + afs_put_call(call); | ||
4434 | _leave(" = %d", ret); | ||
4435 | return ret; | ||
4436 | } | ||
4437 | diff --git a/fs/afs/yfsclient.c b/fs/afs/yfsclient.c | ||
4438 | index 12658c1363ae4..5aa57929e8c23 100644 | ||
4439 | --- a/fs/afs/yfsclient.c | ||
4440 | +++ b/fs/afs/yfsclient.c | ||
4441 | @@ -803,7 +803,7 @@ int yfs_fs_create_file(struct afs_fs_cursor *fc, | ||
4442 | bp = xdr_encode_YFSFid(bp, &vnode->fid); | ||
4443 | bp = xdr_encode_string(bp, name, namesz); | ||
4444 | bp = xdr_encode_YFSStoreStatus_mode(bp, mode); | ||
4445 | - bp = xdr_encode_u32(bp, 0); /* ViceLockType */ | ||
4446 | + bp = xdr_encode_u32(bp, yfs_LockNone); /* ViceLockType */ | ||
4447 | yfs_check_req(call, bp); | ||
4448 | |||
4449 | afs_use_fs_server(call, fc->cbi); | ||
4450 | diff --git a/fs/ceph/snap.c b/fs/ceph/snap.c | ||
4451 | index 041c27ea8de15..f74193da0e092 100644 | ||
4452 | --- a/fs/ceph/snap.c | ||
4453 | +++ b/fs/ceph/snap.c | ||
4454 | @@ -616,7 +616,8 @@ int __ceph_finish_cap_snap(struct ceph_inode_info *ci, | ||
4455 | capsnap->size); | ||
4456 | |||
4457 | spin_lock(&mdsc->snap_flush_lock); | ||
4458 | - list_add_tail(&ci->i_snap_flush_item, &mdsc->snap_flush_list); | ||
4459 | + if (list_empty(&ci->i_snap_flush_item)) | ||
4460 | + list_add_tail(&ci->i_snap_flush_item, &mdsc->snap_flush_list); | ||
4461 | spin_unlock(&mdsc->snap_flush_lock); | ||
4462 | return 1; /* caller may want to ceph_flush_snaps */ | ||
4463 | } | ||
4464 | diff --git a/fs/proc/base.c b/fs/proc/base.c | ||
4465 | index ce34654794472..bde45ca75ba3e 100644 | ||
4466 | --- a/fs/proc/base.c | ||
4467 | +++ b/fs/proc/base.c | ||
4468 | @@ -1084,10 +1084,6 @@ static int __set_oom_adj(struct file *file, int oom_adj, bool legacy) | ||
4469 | |||
4470 | task_lock(p); | ||
4471 | if (!p->vfork_done && process_shares_mm(p, mm)) { | ||
4472 | - pr_info("updating oom_score_adj for %d (%s) from %d to %d because it shares mm with %d (%s). Report if this is unexpected.\n", | ||
4473 | - task_pid_nr(p), p->comm, | ||
4474 | - p->signal->oom_score_adj, oom_adj, | ||
4475 | - task_pid_nr(task), task->comm); | ||
4476 | p->signal->oom_score_adj = oom_adj; | ||
4477 | if (!legacy && has_capability_noaudit(current, CAP_SYS_RESOURCE)) | ||
4478 | p->signal->oom_score_adj_min = (short)oom_adj; | ||
4479 | diff --git a/include/keys/user-type.h b/include/keys/user-type.h | ||
4480 | index e098cbe27db54..12babe9915944 100644 | ||
4481 | --- a/include/keys/user-type.h | ||
4482 | +++ b/include/keys/user-type.h | ||
4483 | @@ -31,7 +31,7 @@ | ||
4484 | struct user_key_payload { | ||
4485 | struct rcu_head rcu; /* RCU destructor */ | ||
4486 | unsigned short datalen; /* length of this data */ | ||
4487 | - char data[0]; /* actual data */ | ||
4488 | + char data[0] __aligned(__alignof__(u64)); /* actual data */ | ||
4489 | }; | ||
4490 | |||
4491 | extern struct key_type key_type_user; | ||
4492 | diff --git a/include/linux/compiler-clang.h b/include/linux/compiler-clang.h | ||
4493 | index 3e7dafb3ea809..7ddaeb5182e33 100644 | ||
4494 | --- a/include/linux/compiler-clang.h | ||
4495 | +++ b/include/linux/compiler-clang.h | ||
4496 | @@ -3,9 +3,8 @@ | ||
4497 | #error "Please don't include <linux/compiler-clang.h> directly, include <linux/compiler.h> instead." | ||
4498 | #endif | ||
4499 | |||
4500 | -/* Some compiler specific definitions are overwritten here | ||
4501 | - * for Clang compiler | ||
4502 | - */ | ||
4503 | +/* Compiler specific definitions for Clang compiler */ | ||
4504 | + | ||
4505 | #define uninitialized_var(x) x = *(&(x)) | ||
4506 | |||
4507 | /* same as gcc, this was present in clang-2.6 so we can assume it works | ||
4508 | diff --git a/include/linux/compiler-gcc.h b/include/linux/compiler-gcc.h | ||
4509 | index 977ddf2774f97..c61c4bb2bd15f 100644 | ||
4510 | --- a/include/linux/compiler-gcc.h | ||
4511 | +++ b/include/linux/compiler-gcc.h | ||
4512 | @@ -58,10 +58,6 @@ | ||
4513 | (typeof(ptr)) (__ptr + (off)); \ | ||
4514 | }) | ||
4515 | |||
4516 | -/* Make the optimizer believe the variable can be manipulated arbitrarily. */ | ||
4517 | -#define OPTIMIZER_HIDE_VAR(var) \ | ||
4518 | - __asm__ ("" : "=r" (var) : "0" (var)) | ||
4519 | - | ||
4520 | /* | ||
4521 | * A trick to suppress uninitialized variable warning without generating any | ||
4522 | * code | ||
4523 | diff --git a/include/linux/compiler-intel.h b/include/linux/compiler-intel.h | ||
4524 | index 517bd14e12224..b17f3cd18334d 100644 | ||
4525 | --- a/include/linux/compiler-intel.h | ||
4526 | +++ b/include/linux/compiler-intel.h | ||
4527 | @@ -5,9 +5,7 @@ | ||
4528 | |||
4529 | #ifdef __ECC | ||
4530 | |||
4531 | -/* Some compiler specific definitions are overwritten here | ||
4532 | - * for Intel ECC compiler | ||
4533 | - */ | ||
4534 | +/* Compiler specific definitions for Intel ECC compiler */ | ||
4535 | |||
4536 | #include <asm/intrinsics.h> | ||
4537 | |||
4538 | diff --git a/include/linux/compiler.h b/include/linux/compiler.h | ||
4539 | index fc5004a4b07d7..445348facea97 100644 | ||
4540 | --- a/include/linux/compiler.h | ||
4541 | +++ b/include/linux/compiler.h | ||
4542 | @@ -161,7 +161,9 @@ void ftrace_likely_update(struct ftrace_likely_data *f, int val, | ||
4543 | #endif | ||
4544 | |||
4545 | #ifndef OPTIMIZER_HIDE_VAR | ||
4546 | -#define OPTIMIZER_HIDE_VAR(var) barrier() | ||
4547 | +/* Make the optimizer believe the variable can be manipulated arbitrarily. */ | ||
4548 | +#define OPTIMIZER_HIDE_VAR(var) \ | ||
4549 | + __asm__ ("" : "=r" (var) : "0" (var)) | ||
4550 | #endif | ||
4551 | |||
4552 | /* Not-quite-unique ID. */ | ||
4553 | diff --git a/include/linux/qed/qed_chain.h b/include/linux/qed/qed_chain.h | ||
4554 | index 59ddf9af909e4..2dd0a9ed5b361 100644 | ||
4555 | --- a/include/linux/qed/qed_chain.h | ||
4556 | +++ b/include/linux/qed/qed_chain.h | ||
4557 | @@ -663,6 +663,37 @@ out: | ||
4558 | static inline void qed_chain_set_prod(struct qed_chain *p_chain, | ||
4559 | u32 prod_idx, void *p_prod_elem) | ||
4560 | { | ||
4561 | + if (p_chain->mode == QED_CHAIN_MODE_PBL) { | ||
4562 | + u32 cur_prod, page_mask, page_cnt, page_diff; | ||
4563 | + | ||
4564 | + cur_prod = is_chain_u16(p_chain) ? p_chain->u.chain16.prod_idx : | ||
4565 | + p_chain->u.chain32.prod_idx; | ||
4566 | + | ||
4567 | + /* Assume that number of elements in a page is power of 2 */ | ||
4568 | + page_mask = ~p_chain->elem_per_page_mask; | ||
4569 | + | ||
4570 | + /* Use "cur_prod - 1" and "prod_idx - 1" since producer index | ||
4571 | + * reaches the first element of next page before the page index | ||
4572 | + * is incremented. See qed_chain_produce(). | ||
4573 | + * Index wrap around is not a problem because the difference | ||
4574 | + * between current and given producer indices is always | ||
4575 | + * positive and lower than the chain's capacity. | ||
4576 | + */ | ||
4577 | + page_diff = (((cur_prod - 1) & page_mask) - | ||
4578 | + ((prod_idx - 1) & page_mask)) / | ||
4579 | + p_chain->elem_per_page; | ||
4580 | + | ||
4581 | + page_cnt = qed_chain_get_page_cnt(p_chain); | ||
4582 | + if (is_chain_u16(p_chain)) | ||
4583 | + p_chain->pbl.c.u16.prod_page_idx = | ||
4584 | + (p_chain->pbl.c.u16.prod_page_idx - | ||
4585 | + page_diff + page_cnt) % page_cnt; | ||
4586 | + else | ||
4587 | + p_chain->pbl.c.u32.prod_page_idx = | ||
4588 | + (p_chain->pbl.c.u32.prod_page_idx - | ||
4589 | + page_diff + page_cnt) % page_cnt; | ||
4590 | + } | ||
4591 | + | ||
4592 | if (is_chain_u16(p_chain)) | ||
4593 | p_chain->u.chain16.prod_idx = (u16) prod_idx; | ||
4594 | else | ||
4595 | diff --git a/include/linux/skbuff.h b/include/linux/skbuff.h | ||
4596 | index a6d820ad17f07..8e63c166765ef 100644 | ||
4597 | --- a/include/linux/skbuff.h | ||
4598 | +++ b/include/linux/skbuff.h | ||
4599 | @@ -2418,7 +2418,7 @@ static inline void skb_probe_transport_header(struct sk_buff *skb, | ||
4600 | |||
4601 | if (skb_flow_dissect_flow_keys_basic(skb, &keys, NULL, 0, 0, 0, 0)) | ||
4602 | skb_set_transport_header(skb, keys.control.thoff); | ||
4603 | - else | ||
4604 | + else if (offset_hint >= 0) | ||
4605 | skb_set_transport_header(skb, offset_hint); | ||
4606 | } | ||
4607 | |||
4608 | diff --git a/include/linux/virtio_net.h b/include/linux/virtio_net.h | ||
4609 | index cb462f9ab7dd5..e0348cb0a1dd7 100644 | ||
4610 | --- a/include/linux/virtio_net.h | ||
4611 | +++ b/include/linux/virtio_net.h | ||
4612 | @@ -57,6 +57,25 @@ static inline int virtio_net_hdr_to_skb(struct sk_buff *skb, | ||
4613 | |||
4614 | if (!skb_partial_csum_set(skb, start, off)) | ||
4615 | return -EINVAL; | ||
4616 | + } else { | ||
4617 | + /* gso packets without NEEDS_CSUM do not set transport_offset. | ||
4618 | + * probe and drop if does not match one of the above types. | ||
4619 | + */ | ||
4620 | + if (gso_type && skb->network_header) { | ||
4621 | + if (!skb->protocol) | ||
4622 | + virtio_net_hdr_set_proto(skb, hdr); | ||
4623 | +retry: | ||
4624 | + skb_probe_transport_header(skb, -1); | ||
4625 | + if (!skb_transport_header_was_set(skb)) { | ||
4626 | + /* UFO does not specify ipv4 or 6: try both */ | ||
4627 | + if (gso_type & SKB_GSO_UDP && | ||
4628 | + skb->protocol == htons(ETH_P_IP)) { | ||
4629 | + skb->protocol = htons(ETH_P_IPV6); | ||
4630 | + goto retry; | ||
4631 | + } | ||
4632 | + return -EINVAL; | ||
4633 | + } | ||
4634 | + } | ||
4635 | } | ||
4636 | |||
4637 | if (hdr->gso_type != VIRTIO_NET_HDR_GSO_NONE) { | ||
4638 | diff --git a/include/net/netfilter/nf_flow_table.h b/include/net/netfilter/nf_flow_table.h | ||
4639 | index 77e2761d4f2f9..ff4eb9869e5ba 100644 | ||
4640 | --- a/include/net/netfilter/nf_flow_table.h | ||
4641 | +++ b/include/net/netfilter/nf_flow_table.h | ||
4642 | @@ -84,7 +84,6 @@ struct flow_offload { | ||
4643 | struct nf_flow_route { | ||
4644 | struct { | ||
4645 | struct dst_entry *dst; | ||
4646 | - int ifindex; | ||
4647 | } tuple[FLOW_OFFLOAD_DIR_MAX]; | ||
4648 | }; | ||
4649 | |||
4650 | diff --git a/include/sound/soc.h b/include/sound/soc.h | ||
4651 | index 3e0ac310a3df9..e721082c84a36 100644 | ||
4652 | --- a/include/sound/soc.h | ||
4653 | +++ b/include/sound/soc.h | ||
4654 | @@ -985,6 +985,12 @@ struct snd_soc_dai_link { | ||
4655 | /* Do not create a PCM for this DAI link (Backend link) */ | ||
4656 | unsigned int ignore:1; | ||
4657 | |||
4658 | + /* | ||
4659 | + * This driver uses legacy platform naming. Set by the core, machine | ||
4660 | + * drivers should not modify this value. | ||
4661 | + */ | ||
4662 | + unsigned int legacy_platform:1; | ||
4663 | + | ||
4664 | struct list_head list; /* DAI link list of the soc card */ | ||
4665 | struct snd_soc_dobj dobj; /* For topology */ | ||
4666 | }; | ||
4667 | diff --git a/include/trace/events/afs.h b/include/trace/events/afs.h | ||
4668 | index 33d291888ba9c..e3f005eae1f76 100644 | ||
4669 | --- a/include/trace/events/afs.h | ||
4670 | +++ b/include/trace/events/afs.h | ||
4671 | @@ -25,6 +25,7 @@ | ||
4672 | enum afs_call_trace { | ||
4673 | afs_call_trace_alloc, | ||
4674 | afs_call_trace_free, | ||
4675 | + afs_call_trace_get, | ||
4676 | afs_call_trace_put, | ||
4677 | afs_call_trace_wake, | ||
4678 | afs_call_trace_work, | ||
4679 | @@ -159,6 +160,7 @@ enum afs_file_error { | ||
4680 | #define afs_call_traces \ | ||
4681 | EM(afs_call_trace_alloc, "ALLOC") \ | ||
4682 | EM(afs_call_trace_free, "FREE ") \ | ||
4683 | + EM(afs_call_trace_get, "GET ") \ | ||
4684 | EM(afs_call_trace_put, "PUT ") \ | ||
4685 | EM(afs_call_trace_wake, "WAKE ") \ | ||
4686 | E_(afs_call_trace_work, "WORK ") | ||
4687 | diff --git a/include/uapi/linux/inet_diag.h b/include/uapi/linux/inet_diag.h | ||
4688 | index 14565d703291b..e8baca85bac6a 100644 | ||
4689 | --- a/include/uapi/linux/inet_diag.h | ||
4690 | +++ b/include/uapi/linux/inet_diag.h | ||
4691 | @@ -137,15 +137,21 @@ enum { | ||
4692 | INET_DIAG_TCLASS, | ||
4693 | INET_DIAG_SKMEMINFO, | ||
4694 | INET_DIAG_SHUTDOWN, | ||
4695 | - INET_DIAG_DCTCPINFO, | ||
4696 | - INET_DIAG_PROTOCOL, /* response attribute only */ | ||
4697 | + | ||
4698 | + /* | ||
4699 | + * Next extenstions cannot be requested in struct inet_diag_req_v2: | ||
4700 | + * its field idiag_ext has only 8 bits. | ||
4701 | + */ | ||
4702 | + | ||
4703 | + INET_DIAG_DCTCPINFO, /* request as INET_DIAG_VEGASINFO */ | ||
4704 | + INET_DIAG_PROTOCOL, /* response attribute only */ | ||
4705 | INET_DIAG_SKV6ONLY, | ||
4706 | INET_DIAG_LOCALS, | ||
4707 | INET_DIAG_PEERS, | ||
4708 | INET_DIAG_PAD, | ||
4709 | - INET_DIAG_MARK, | ||
4710 | - INET_DIAG_BBRINFO, | ||
4711 | - INET_DIAG_CLASS_ID, | ||
4712 | + INET_DIAG_MARK, /* only with CAP_NET_ADMIN */ | ||
4713 | + INET_DIAG_BBRINFO, /* request as INET_DIAG_VEGASINFO */ | ||
4714 | + INET_DIAG_CLASS_ID, /* request as INET_DIAG_TCLASS */ | ||
4715 | INET_DIAG_MD5SIG, | ||
4716 | __INET_DIAG_MAX, | ||
4717 | }; | ||
4718 | diff --git a/kernel/bpf/stackmap.c b/kernel/bpf/stackmap.c | ||
4719 | index 90daf285de032..d43b145358275 100644 | ||
4720 | --- a/kernel/bpf/stackmap.c | ||
4721 | +++ b/kernel/bpf/stackmap.c | ||
4722 | @@ -180,11 +180,14 @@ static inline int stack_map_parse_build_id(void *page_addr, | ||
4723 | |||
4724 | if (nhdr->n_type == BPF_BUILD_ID && | ||
4725 | nhdr->n_namesz == sizeof("GNU") && | ||
4726 | - nhdr->n_descsz == BPF_BUILD_ID_SIZE) { | ||
4727 | + nhdr->n_descsz > 0 && | ||
4728 | + nhdr->n_descsz <= BPF_BUILD_ID_SIZE) { | ||
4729 | memcpy(build_id, | ||
4730 | note_start + note_offs + | ||
4731 | ALIGN(sizeof("GNU"), 4) + sizeof(Elf32_Nhdr), | ||
4732 | - BPF_BUILD_ID_SIZE); | ||
4733 | + nhdr->n_descsz); | ||
4734 | + memset(build_id + nhdr->n_descsz, 0, | ||
4735 | + BPF_BUILD_ID_SIZE - nhdr->n_descsz); | ||
4736 | return 0; | ||
4737 | } | ||
4738 | new_offs = note_offs + sizeof(Elf32_Nhdr) + | ||
4739 | @@ -260,7 +263,7 @@ static int stack_map_get_build_id(struct vm_area_struct *vma, | ||
4740 | return -EFAULT; /* page not mapped */ | ||
4741 | |||
4742 | ret = -EINVAL; | ||
4743 | - page_addr = page_address(page); | ||
4744 | + page_addr = kmap_atomic(page); | ||
4745 | ehdr = (Elf32_Ehdr *)page_addr; | ||
4746 | |||
4747 | /* compare magic x7f "ELF" */ | ||
4748 | @@ -276,6 +279,7 @@ static int stack_map_get_build_id(struct vm_area_struct *vma, | ||
4749 | else if (ehdr->e_ident[EI_CLASS] == ELFCLASS64) | ||
4750 | ret = stack_map_get_build_id_64(page_addr, build_id); | ||
4751 | out: | ||
4752 | + kunmap_atomic(page_addr); | ||
4753 | put_page(page); | ||
4754 | return ret; | ||
4755 | } | ||
4756 | @@ -310,6 +314,7 @@ static void stack_map_get_build_id_offset(struct bpf_stack_build_id *id_offs, | ||
4757 | for (i = 0; i < trace_nr; i++) { | ||
4758 | id_offs[i].status = BPF_STACK_BUILD_ID_IP; | ||
4759 | id_offs[i].ip = ips[i]; | ||
4760 | + memset(id_offs[i].build_id, 0, BPF_BUILD_ID_SIZE); | ||
4761 | } | ||
4762 | return; | ||
4763 | } | ||
4764 | @@ -320,6 +325,7 @@ static void stack_map_get_build_id_offset(struct bpf_stack_build_id *id_offs, | ||
4765 | /* per entry fall back to ips */ | ||
4766 | id_offs[i].status = BPF_STACK_BUILD_ID_IP; | ||
4767 | id_offs[i].ip = ips[i]; | ||
4768 | + memset(id_offs[i].build_id, 0, BPF_BUILD_ID_SIZE); | ||
4769 | continue; | ||
4770 | } | ||
4771 | id_offs[i].offset = (vma->vm_pgoff << PAGE_SHIFT) + ips[i] | ||
4772 | diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c | ||
4773 | index ff1c4b20cd0a6..b331562989bd2 100644 | ||
4774 | --- a/kernel/trace/trace.c | ||
4775 | +++ b/kernel/trace/trace.c | ||
4776 | @@ -3384,6 +3384,8 @@ static void print_func_help_header_irq(struct trace_buffer *buf, struct seq_file | ||
4777 | const char tgid_space[] = " "; | ||
4778 | const char space[] = " "; | ||
4779 | |||
4780 | + print_event_info(buf, m); | ||
4781 | + | ||
4782 | seq_printf(m, "# %s _-----=> irqs-off\n", | ||
4783 | tgid ? tgid_space : space); | ||
4784 | seq_printf(m, "# %s / _----=> need-resched\n", | ||
4785 | diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c | ||
4786 | index fec67188c4d28..3387408a15c27 100644 | ||
4787 | --- a/kernel/trace/trace_kprobe.c | ||
4788 | +++ b/kernel/trace/trace_kprobe.c | ||
4789 | @@ -878,22 +878,14 @@ static const struct file_operations kprobe_profile_ops = { | ||
4790 | static nokprobe_inline int | ||
4791 | fetch_store_strlen(unsigned long addr) | ||
4792 | { | ||
4793 | - mm_segment_t old_fs; | ||
4794 | int ret, len = 0; | ||
4795 | u8 c; | ||
4796 | |||
4797 | - old_fs = get_fs(); | ||
4798 | - set_fs(KERNEL_DS); | ||
4799 | - pagefault_disable(); | ||
4800 | - | ||
4801 | do { | ||
4802 | - ret = __copy_from_user_inatomic(&c, (u8 *)addr + len, 1); | ||
4803 | + ret = probe_mem_read(&c, (u8 *)addr + len, 1); | ||
4804 | len++; | ||
4805 | } while (c && ret == 0 && len < MAX_STRING_SIZE); | ||
4806 | |||
4807 | - pagefault_enable(); | ||
4808 | - set_fs(old_fs); | ||
4809 | - | ||
4810 | return (ret < 0) ? ret : len; | ||
4811 | } | ||
4812 | |||
4813 | diff --git a/mm/mempolicy.c b/mm/mempolicy.c | ||
4814 | index d4496d9d34f53..ee2bce59d2bff 100644 | ||
4815 | --- a/mm/mempolicy.c | ||
4816 | +++ b/mm/mempolicy.c | ||
4817 | @@ -1314,7 +1314,7 @@ static int copy_nodes_to_user(unsigned long __user *mask, unsigned long maxnode, | ||
4818 | nodemask_t *nodes) | ||
4819 | { | ||
4820 | unsigned long copy = ALIGN(maxnode-1, 64) / 8; | ||
4821 | - const int nbytes = BITS_TO_LONGS(MAX_NUMNODES) * sizeof(long); | ||
4822 | + unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); | ||
4823 | |||
4824 | if (copy > nbytes) { | ||
4825 | if (copy > PAGE_SIZE) | ||
4826 | @@ -1491,7 +1491,7 @@ static int kernel_get_mempolicy(int __user *policy, | ||
4827 | int uninitialized_var(pval); | ||
4828 | nodemask_t nodes; | ||
4829 | |||
4830 | - if (nmask != NULL && maxnode < MAX_NUMNODES) | ||
4831 | + if (nmask != NULL && maxnode < nr_node_ids) | ||
4832 | return -EINVAL; | ||
4833 | |||
4834 | err = do_get_mempolicy(&pval, &nodes, addr, flags); | ||
4835 | @@ -1527,7 +1527,7 @@ COMPAT_SYSCALL_DEFINE5(get_mempolicy, int __user *, policy, | ||
4836 | unsigned long nr_bits, alloc_size; | ||
4837 | DECLARE_BITMAP(bm, MAX_NUMNODES); | ||
4838 | |||
4839 | - nr_bits = min_t(unsigned long, maxnode-1, MAX_NUMNODES); | ||
4840 | + nr_bits = min_t(unsigned long, maxnode-1, nr_node_ids); | ||
4841 | alloc_size = ALIGN(nr_bits, BITS_PER_LONG) / 8; | ||
4842 | |||
4843 | if (nmask) | ||
4844 | diff --git a/net/batman-adv/soft-interface.c b/net/batman-adv/soft-interface.c | ||
4845 | index b85ca809e5092..ffc83bebfe403 100644 | ||
4846 | --- a/net/batman-adv/soft-interface.c | ||
4847 | +++ b/net/batman-adv/soft-interface.c | ||
4848 | @@ -227,6 +227,8 @@ static netdev_tx_t batadv_interface_tx(struct sk_buff *skb, | ||
4849 | |||
4850 | switch (ntohs(ethhdr->h_proto)) { | ||
4851 | case ETH_P_8021Q: | ||
4852 | + if (!pskb_may_pull(skb, sizeof(*vhdr))) | ||
4853 | + goto dropped; | ||
4854 | vhdr = vlan_eth_hdr(skb); | ||
4855 | |||
4856 | /* drop batman-in-batman packets to prevent loops */ | ||
4857 | diff --git a/net/bridge/br_fdb.c b/net/bridge/br_fdb.c | ||
4858 | index e56ba3912a905..8b8abf88befbd 100644 | ||
4859 | --- a/net/bridge/br_fdb.c | ||
4860 | +++ b/net/bridge/br_fdb.c | ||
4861 | @@ -1102,6 +1102,8 @@ int br_fdb_external_learn_add(struct net_bridge *br, struct net_bridge_port *p, | ||
4862 | err = -ENOMEM; | ||
4863 | goto err_unlock; | ||
4864 | } | ||
4865 | + if (swdev_notify) | ||
4866 | + fdb->added_by_user = 1; | ||
4867 | fdb->added_by_external_learn = 1; | ||
4868 | fdb_notify(br, fdb, RTM_NEWNEIGH, swdev_notify); | ||
4869 | } else { | ||
4870 | @@ -1121,6 +1123,9 @@ int br_fdb_external_learn_add(struct net_bridge *br, struct net_bridge_port *p, | ||
4871 | modified = true; | ||
4872 | } | ||
4873 | |||
4874 | + if (swdev_notify) | ||
4875 | + fdb->added_by_user = 1; | ||
4876 | + | ||
4877 | if (modified) | ||
4878 | fdb_notify(br, fdb, RTM_NEWNEIGH, swdev_notify); | ||
4879 | } | ||
4880 | diff --git a/net/bridge/br_multicast.c b/net/bridge/br_multicast.c | ||
4881 | index 6bac0d6b7b941..024139b51d3a5 100644 | ||
4882 | --- a/net/bridge/br_multicast.c | ||
4883 | +++ b/net/bridge/br_multicast.c | ||
4884 | @@ -1422,14 +1422,7 @@ static void br_multicast_query_received(struct net_bridge *br, | ||
4885 | return; | ||
4886 | |||
4887 | br_multicast_update_query_timer(br, query, max_delay); | ||
4888 | - | ||
4889 | - /* Based on RFC4541, section 2.1.1 IGMP Forwarding Rules, | ||
4890 | - * the arrival port for IGMP Queries where the source address | ||
4891 | - * is 0.0.0.0 should not be added to router port list. | ||
4892 | - */ | ||
4893 | - if ((saddr->proto == htons(ETH_P_IP) && saddr->u.ip4) || | ||
4894 | - saddr->proto == htons(ETH_P_IPV6)) | ||
4895 | - br_multicast_mark_router(br, port); | ||
4896 | + br_multicast_mark_router(br, port); | ||
4897 | } | ||
4898 | |||
4899 | static void br_ip4_multicast_query(struct net_bridge *br, | ||
4900 | diff --git a/net/ceph/messenger.c b/net/ceph/messenger.c | ||
4901 | index 664f886f464da..b59b81fc1ab66 100644 | ||
4902 | --- a/net/ceph/messenger.c | ||
4903 | +++ b/net/ceph/messenger.c | ||
4904 | @@ -2071,6 +2071,8 @@ static int process_connect(struct ceph_connection *con) | ||
4905 | dout("process_connect on %p tag %d\n", con, (int)con->in_tag); | ||
4906 | |||
4907 | if (con->auth) { | ||
4908 | + int len = le32_to_cpu(con->in_reply.authorizer_len); | ||
4909 | + | ||
4910 | /* | ||
4911 | * Any connection that defines ->get_authorizer() | ||
4912 | * should also define ->add_authorizer_challenge() and | ||
4913 | @@ -2080,8 +2082,7 @@ static int process_connect(struct ceph_connection *con) | ||
4914 | */ | ||
4915 | if (con->in_reply.tag == CEPH_MSGR_TAG_CHALLENGE_AUTHORIZER) { | ||
4916 | ret = con->ops->add_authorizer_challenge( | ||
4917 | - con, con->auth->authorizer_reply_buf, | ||
4918 | - le32_to_cpu(con->in_reply.authorizer_len)); | ||
4919 | + con, con->auth->authorizer_reply_buf, len); | ||
4920 | if (ret < 0) | ||
4921 | return ret; | ||
4922 | |||
4923 | @@ -2091,10 +2092,12 @@ static int process_connect(struct ceph_connection *con) | ||
4924 | return 0; | ||
4925 | } | ||
4926 | |||
4927 | - ret = con->ops->verify_authorizer_reply(con); | ||
4928 | - if (ret < 0) { | ||
4929 | - con->error_msg = "bad authorize reply"; | ||
4930 | - return ret; | ||
4931 | + if (len) { | ||
4932 | + ret = con->ops->verify_authorizer_reply(con); | ||
4933 | + if (ret < 0) { | ||
4934 | + con->error_msg = "bad authorize reply"; | ||
4935 | + return ret; | ||
4936 | + } | ||
4937 | } | ||
4938 | } | ||
4939 | |||
4940 | diff --git a/net/core/filter.c b/net/core/filter.c | ||
4941 | index eb0007f30142b..16350f8c8815a 100644 | ||
4942 | --- a/net/core/filter.c | ||
4943 | +++ b/net/core/filter.c | ||
4944 | @@ -3935,6 +3935,10 @@ BPF_CALL_5(bpf_setsockopt, struct bpf_sock_ops_kern *, bpf_sock, | ||
4945 | sk->sk_sndbuf = max_t(int, val * 2, SOCK_MIN_SNDBUF); | ||
4946 | break; | ||
4947 | case SO_MAX_PACING_RATE: /* 32bit version */ | ||
4948 | + if (val != ~0U) | ||
4949 | + cmpxchg(&sk->sk_pacing_status, | ||
4950 | + SK_PACING_NONE, | ||
4951 | + SK_PACING_NEEDED); | ||
4952 | sk->sk_max_pacing_rate = (val == ~0U) ? ~0UL : val; | ||
4953 | sk->sk_pacing_rate = min(sk->sk_pacing_rate, | ||
4954 | sk->sk_max_pacing_rate); | ||
4955 | @@ -3948,7 +3952,10 @@ BPF_CALL_5(bpf_setsockopt, struct bpf_sock_ops_kern *, bpf_sock, | ||
4956 | sk->sk_rcvlowat = val ? : 1; | ||
4957 | break; | ||
4958 | case SO_MARK: | ||
4959 | - sk->sk_mark = val; | ||
4960 | + if (sk->sk_mark != val) { | ||
4961 | + sk->sk_mark = val; | ||
4962 | + sk_dst_reset(sk); | ||
4963 | + } | ||
4964 | break; | ||
4965 | default: | ||
4966 | ret = -EINVAL; | ||
4967 | @@ -4019,7 +4026,7 @@ BPF_CALL_5(bpf_setsockopt, struct bpf_sock_ops_kern *, bpf_sock, | ||
4968 | /* Only some options are supported */ | ||
4969 | switch (optname) { | ||
4970 | case TCP_BPF_IW: | ||
4971 | - if (val <= 0 || tp->data_segs_out > 0) | ||
4972 | + if (val <= 0 || tp->data_segs_out > tp->syn_data) | ||
4973 | ret = -EINVAL; | ||
4974 | else | ||
4975 | tp->snd_cwnd = val; | ||
4976 | diff --git a/net/dsa/port.c b/net/dsa/port.c | ||
4977 | index ed0595459df13..792a13068c50b 100644 | ||
4978 | --- a/net/dsa/port.c | ||
4979 | +++ b/net/dsa/port.c | ||
4980 | @@ -255,7 +255,7 @@ int dsa_port_vlan_add(struct dsa_port *dp, | ||
4981 | if (netif_is_bridge_master(vlan->obj.orig_dev)) | ||
4982 | return -EOPNOTSUPP; | ||
4983 | |||
4984 | - if (br_vlan_enabled(dp->bridge_dev)) | ||
4985 | + if (dp->bridge_dev && br_vlan_enabled(dp->bridge_dev)) | ||
4986 | return dsa_port_notify(dp, DSA_NOTIFIER_VLAN_ADD, &info); | ||
4987 | |||
4988 | return 0; | ||
4989 | @@ -273,7 +273,7 @@ int dsa_port_vlan_del(struct dsa_port *dp, | ||
4990 | if (netif_is_bridge_master(vlan->obj.orig_dev)) | ||
4991 | return -EOPNOTSUPP; | ||
4992 | |||
4993 | - if (br_vlan_enabled(dp->bridge_dev)) | ||
4994 | + if (dp->bridge_dev && br_vlan_enabled(dp->bridge_dev)) | ||
4995 | return dsa_port_notify(dp, DSA_NOTIFIER_VLAN_DEL, &info); | ||
4996 | |||
4997 | return 0; | ||
4998 | diff --git a/net/ipv4/inet_diag.c b/net/ipv4/inet_diag.c | ||
4999 | index 1a4e9ff02762e..5731670c560b0 100644 | ||
5000 | --- a/net/ipv4/inet_diag.c | ||
5001 | +++ b/net/ipv4/inet_diag.c | ||
5002 | @@ -108,6 +108,7 @@ static size_t inet_sk_attr_size(struct sock *sk, | ||
5003 | + nla_total_size(1) /* INET_DIAG_TOS */ | ||
5004 | + nla_total_size(1) /* INET_DIAG_TCLASS */ | ||
5005 | + nla_total_size(4) /* INET_DIAG_MARK */ | ||
5006 | + + nla_total_size(4) /* INET_DIAG_CLASS_ID */ | ||
5007 | + nla_total_size(sizeof(struct inet_diag_meminfo)) | ||
5008 | + nla_total_size(sizeof(struct inet_diag_msg)) | ||
5009 | + nla_total_size(SK_MEMINFO_VARS * sizeof(u32)) | ||
5010 | @@ -287,12 +288,19 @@ int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, | ||
5011 | goto errout; | ||
5012 | } | ||
5013 | |||
5014 | - if (ext & (1 << (INET_DIAG_CLASS_ID - 1))) { | ||
5015 | + if (ext & (1 << (INET_DIAG_CLASS_ID - 1)) || | ||
5016 | + ext & (1 << (INET_DIAG_TCLASS - 1))) { | ||
5017 | u32 classid = 0; | ||
5018 | |||
5019 | #ifdef CONFIG_SOCK_CGROUP_DATA | ||
5020 | classid = sock_cgroup_classid(&sk->sk_cgrp_data); | ||
5021 | #endif | ||
5022 | + /* Fallback to socket priority if class id isn't set. | ||
5023 | + * Classful qdiscs use it as direct reference to class. | ||
5024 | + * For cgroup2 classid is always zero. | ||
5025 | + */ | ||
5026 | + if (!classid) | ||
5027 | + classid = sk->sk_priority; | ||
5028 | |||
5029 | if (nla_put_u32(skb, INET_DIAG_CLASS_ID, classid)) | ||
5030 | goto errout; | ||
5031 | diff --git a/net/ipv4/netfilter/ipt_CLUSTERIP.c b/net/ipv4/netfilter/ipt_CLUSTERIP.c | ||
5032 | index fb1e7f237f531..3cd237b42f446 100644 | ||
5033 | --- a/net/ipv4/netfilter/ipt_CLUSTERIP.c | ||
5034 | +++ b/net/ipv4/netfilter/ipt_CLUSTERIP.c | ||
5035 | @@ -56,7 +56,7 @@ struct clusterip_config { | ||
5036 | #endif | ||
5037 | enum clusterip_hashmode hash_mode; /* which hashing mode */ | ||
5038 | u_int32_t hash_initval; /* hash initialization */ | ||
5039 | - struct rcu_head rcu; | ||
5040 | + struct rcu_head rcu; /* for call_rcu_bh */ | ||
5041 | struct net *net; /* netns for pernet list */ | ||
5042 | char ifname[IFNAMSIZ]; /* device ifname */ | ||
5043 | }; | ||
5044 | @@ -72,6 +72,8 @@ struct clusterip_net { | ||
5045 | |||
5046 | #ifdef CONFIG_PROC_FS | ||
5047 | struct proc_dir_entry *procdir; | ||
5048 | + /* mutex protects the config->pde*/ | ||
5049 | + struct mutex mutex; | ||
5050 | #endif | ||
5051 | }; | ||
5052 | |||
5053 | @@ -118,17 +120,18 @@ clusterip_config_entry_put(struct clusterip_config *c) | ||
5054 | |||
5055 | local_bh_disable(); | ||
5056 | if (refcount_dec_and_lock(&c->entries, &cn->lock)) { | ||
5057 | + list_del_rcu(&c->list); | ||
5058 | + spin_unlock(&cn->lock); | ||
5059 | + local_bh_enable(); | ||
5060 | /* In case anyone still accesses the file, the open/close | ||
5061 | * functions are also incrementing the refcount on their own, | ||
5062 | * so it's safe to remove the entry even if it's in use. */ | ||
5063 | #ifdef CONFIG_PROC_FS | ||
5064 | + mutex_lock(&cn->mutex); | ||
5065 | if (cn->procdir) | ||
5066 | proc_remove(c->pde); | ||
5067 | + mutex_unlock(&cn->mutex); | ||
5068 | #endif | ||
5069 | - list_del_rcu(&c->list); | ||
5070 | - spin_unlock(&cn->lock); | ||
5071 | - local_bh_enable(); | ||
5072 | - | ||
5073 | return; | ||
5074 | } | ||
5075 | local_bh_enable(); | ||
5076 | @@ -278,9 +281,11 @@ clusterip_config_init(struct net *net, const struct ipt_clusterip_tgt_info *i, | ||
5077 | |||
5078 | /* create proc dir entry */ | ||
5079 | sprintf(buffer, "%pI4", &ip); | ||
5080 | + mutex_lock(&cn->mutex); | ||
5081 | c->pde = proc_create_data(buffer, 0600, | ||
5082 | cn->procdir, | ||
5083 | &clusterip_proc_fops, c); | ||
5084 | + mutex_unlock(&cn->mutex); | ||
5085 | if (!c->pde) { | ||
5086 | err = -ENOMEM; | ||
5087 | goto err; | ||
5088 | @@ -833,6 +838,7 @@ static int clusterip_net_init(struct net *net) | ||
5089 | pr_err("Unable to proc dir entry\n"); | ||
5090 | return -ENOMEM; | ||
5091 | } | ||
5092 | + mutex_init(&cn->mutex); | ||
5093 | #endif /* CONFIG_PROC_FS */ | ||
5094 | |||
5095 | return 0; | ||
5096 | @@ -841,9 +847,12 @@ static int clusterip_net_init(struct net *net) | ||
5097 | static void clusterip_net_exit(struct net *net) | ||
5098 | { | ||
5099 | struct clusterip_net *cn = clusterip_pernet(net); | ||
5100 | + | ||
5101 | #ifdef CONFIG_PROC_FS | ||
5102 | + mutex_lock(&cn->mutex); | ||
5103 | proc_remove(cn->procdir); | ||
5104 | cn->procdir = NULL; | ||
5105 | + mutex_unlock(&cn->mutex); | ||
5106 | #endif | ||
5107 | nf_unregister_net_hook(net, &cip_arp_ops); | ||
5108 | } | ||
5109 | diff --git a/net/ipv6/netfilter.c b/net/ipv6/netfilter.c | ||
5110 | index 8b075f0bc3516..6d0b1f3e927bd 100644 | ||
5111 | --- a/net/ipv6/netfilter.c | ||
5112 | +++ b/net/ipv6/netfilter.c | ||
5113 | @@ -23,9 +23,11 @@ int ip6_route_me_harder(struct net *net, struct sk_buff *skb) | ||
5114 | struct sock *sk = sk_to_full_sk(skb->sk); | ||
5115 | unsigned int hh_len; | ||
5116 | struct dst_entry *dst; | ||
5117 | + int strict = (ipv6_addr_type(&iph->daddr) & | ||
5118 | + (IPV6_ADDR_MULTICAST | IPV6_ADDR_LINKLOCAL)); | ||
5119 | struct flowi6 fl6 = { | ||
5120 | .flowi6_oif = sk && sk->sk_bound_dev_if ? sk->sk_bound_dev_if : | ||
5121 | - rt6_need_strict(&iph->daddr) ? skb_dst(skb)->dev->ifindex : 0, | ||
5122 | + strict ? skb_dst(skb)->dev->ifindex : 0, | ||
5123 | .flowi6_mark = skb->mark, | ||
5124 | .flowi6_uid = sock_net_uid(net, sk), | ||
5125 | .daddr = iph->daddr, | ||
5126 | diff --git a/net/ipv6/seg6.c b/net/ipv6/seg6.c | ||
5127 | index 8d0ba757a46ce..9b2f272ca1649 100644 | ||
5128 | --- a/net/ipv6/seg6.c | ||
5129 | +++ b/net/ipv6/seg6.c | ||
5130 | @@ -221,9 +221,7 @@ static int seg6_genl_get_tunsrc(struct sk_buff *skb, struct genl_info *info) | ||
5131 | rcu_read_unlock(); | ||
5132 | |||
5133 | genlmsg_end(msg, hdr); | ||
5134 | - genlmsg_reply(msg, info); | ||
5135 | - | ||
5136 | - return 0; | ||
5137 | + return genlmsg_reply(msg, info); | ||
5138 | |||
5139 | nla_put_failure: | ||
5140 | rcu_read_unlock(); | ||
5141 | diff --git a/net/ipv6/sit.c b/net/ipv6/sit.c | ||
5142 | index 1e03305c05492..e8a1dabef803e 100644 | ||
5143 | --- a/net/ipv6/sit.c | ||
5144 | +++ b/net/ipv6/sit.c | ||
5145 | @@ -546,7 +546,8 @@ static int ipip6_err(struct sk_buff *skb, u32 info) | ||
5146 | } | ||
5147 | |||
5148 | err = 0; | ||
5149 | - if (!ip6_err_gen_icmpv6_unreach(skb, iph->ihl * 4, type, data_len)) | ||
5150 | + if (__in6_dev_get(skb->dev) && | ||
5151 | + !ip6_err_gen_icmpv6_unreach(skb, iph->ihl * 4, type, data_len)) | ||
5152 | goto out; | ||
5153 | |||
5154 | if (t->parms.iph.daddr == 0) | ||
5155 | diff --git a/net/ipv6/udp.c b/net/ipv6/udp.c | ||
5156 | index d01ec252cb81d..848dd38a907a1 100644 | ||
5157 | --- a/net/ipv6/udp.c | ||
5158 | +++ b/net/ipv6/udp.c | ||
5159 | @@ -1322,10 +1322,7 @@ do_udp_sendmsg: | ||
5160 | ipc6.opt = opt; | ||
5161 | |||
5162 | fl6.flowi6_proto = sk->sk_protocol; | ||
5163 | - if (!ipv6_addr_any(daddr)) | ||
5164 | - fl6.daddr = *daddr; | ||
5165 | - else | ||
5166 | - fl6.daddr.s6_addr[15] = 0x1; /* :: means loopback (BSD'ism) */ | ||
5167 | + fl6.daddr = *daddr; | ||
5168 | if (ipv6_addr_any(&fl6.saddr) && !ipv6_addr_any(&np->saddr)) | ||
5169 | fl6.saddr = np->saddr; | ||
5170 | fl6.fl6_sport = inet->inet_sport; | ||
5171 | @@ -1353,6 +1350,9 @@ do_udp_sendmsg: | ||
5172 | } | ||
5173 | } | ||
5174 | |||
5175 | + if (ipv6_addr_any(&fl6.daddr)) | ||
5176 | + fl6.daddr.s6_addr[15] = 0x1; /* :: means loopback (BSD'ism) */ | ||
5177 | + | ||
5178 | final_p = fl6_update_dst(&fl6, opt, &final); | ||
5179 | if (final_p) | ||
5180 | connected = false; | ||
5181 | diff --git a/net/mac80211/cfg.c b/net/mac80211/cfg.c | ||
5182 | index 818aa00603495..517dad83c2fa7 100644 | ||
5183 | --- a/net/mac80211/cfg.c | ||
5184 | +++ b/net/mac80211/cfg.c | ||
5185 | @@ -941,6 +941,7 @@ static int ieee80211_start_ap(struct wiphy *wiphy, struct net_device *dev, | ||
5186 | BSS_CHANGED_P2P_PS | | ||
5187 | BSS_CHANGED_TXPOWER; | ||
5188 | int err; | ||
5189 | + int prev_beacon_int; | ||
5190 | |||
5191 | old = sdata_dereference(sdata->u.ap.beacon, sdata); | ||
5192 | if (old) | ||
5193 | @@ -963,6 +964,7 @@ static int ieee80211_start_ap(struct wiphy *wiphy, struct net_device *dev, | ||
5194 | |||
5195 | sdata->needed_rx_chains = sdata->local->rx_chains; | ||
5196 | |||
5197 | + prev_beacon_int = sdata->vif.bss_conf.beacon_int; | ||
5198 | sdata->vif.bss_conf.beacon_int = params->beacon_interval; | ||
5199 | |||
5200 | if (params->he_cap) | ||
5201 | @@ -974,8 +976,10 @@ static int ieee80211_start_ap(struct wiphy *wiphy, struct net_device *dev, | ||
5202 | if (!err) | ||
5203 | ieee80211_vif_copy_chanctx_to_vlans(sdata, false); | ||
5204 | mutex_unlock(&local->mtx); | ||
5205 | - if (err) | ||
5206 | + if (err) { | ||
5207 | + sdata->vif.bss_conf.beacon_int = prev_beacon_int; | ||
5208 | return err; | ||
5209 | + } | ||
5210 | |||
5211 | /* | ||
5212 | * Apply control port protocol, this allows us to | ||
5213 | diff --git a/net/mac80211/main.c b/net/mac80211/main.c | ||
5214 | index 7b8320d4a8e4b..3131356e290a0 100644 | ||
5215 | --- a/net/mac80211/main.c | ||
5216 | +++ b/net/mac80211/main.c | ||
5217 | @@ -615,13 +615,13 @@ struct ieee80211_hw *ieee80211_alloc_hw_nm(size_t priv_data_len, | ||
5218 | * We need a bit of data queued to build aggregates properly, so | ||
5219 | * instruct the TCP stack to allow more than a single ms of data | ||
5220 | * to be queued in the stack. The value is a bit-shift of 1 | ||
5221 | - * second, so 8 is ~4ms of queued data. Only affects local TCP | ||
5222 | + * second, so 7 is ~8ms of queued data. Only affects local TCP | ||
5223 | * sockets. | ||
5224 | * This is the default, anyhow - drivers may need to override it | ||
5225 | * for local reasons (longer buffers, longer completion time, or | ||
5226 | * similar). | ||
5227 | */ | ||
5228 | - local->hw.tx_sk_pacing_shift = 8; | ||
5229 | + local->hw.tx_sk_pacing_shift = 7; | ||
5230 | |||
5231 | /* set up some defaults */ | ||
5232 | local->hw.queues = 1; | ||
5233 | diff --git a/net/mac80211/mesh.h b/net/mac80211/mesh.h | ||
5234 | index 21526630bf655..e84103b405341 100644 | ||
5235 | --- a/net/mac80211/mesh.h | ||
5236 | +++ b/net/mac80211/mesh.h | ||
5237 | @@ -70,6 +70,7 @@ enum mesh_deferred_task_flags { | ||
5238 | * @dst: mesh path destination mac address | ||
5239 | * @mpp: mesh proxy mac address | ||
5240 | * @rhash: rhashtable list pointer | ||
5241 | + * @walk_list: linked list containing all mesh_path objects. | ||
5242 | * @gate_list: list pointer for known gates list | ||
5243 | * @sdata: mesh subif | ||
5244 | * @next_hop: mesh neighbor to which frames for this destination will be | ||
5245 | @@ -105,6 +106,7 @@ struct mesh_path { | ||
5246 | u8 dst[ETH_ALEN]; | ||
5247 | u8 mpp[ETH_ALEN]; /* used for MPP or MAP */ | ||
5248 | struct rhash_head rhash; | ||
5249 | + struct hlist_node walk_list; | ||
5250 | struct hlist_node gate_list; | ||
5251 | struct ieee80211_sub_if_data *sdata; | ||
5252 | struct sta_info __rcu *next_hop; | ||
5253 | @@ -133,12 +135,16 @@ struct mesh_path { | ||
5254 | * gate's mpath may or may not be resolved and active. | ||
5255 | * @gates_lock: protects updates to known_gates | ||
5256 | * @rhead: the rhashtable containing struct mesh_paths, keyed by dest addr | ||
5257 | + * @walk_head: linked list containging all mesh_path objects | ||
5258 | + * @walk_lock: lock protecting walk_head | ||
5259 | * @entries: number of entries in the table | ||
5260 | */ | ||
5261 | struct mesh_table { | ||
5262 | struct hlist_head known_gates; | ||
5263 | spinlock_t gates_lock; | ||
5264 | struct rhashtable rhead; | ||
5265 | + struct hlist_head walk_head; | ||
5266 | + spinlock_t walk_lock; | ||
5267 | atomic_t entries; /* Up to MAX_MESH_NEIGHBOURS */ | ||
5268 | }; | ||
5269 | |||
5270 | diff --git a/net/mac80211/mesh_pathtbl.c b/net/mac80211/mesh_pathtbl.c | ||
5271 | index a5125624a76dc..c3a7396fb9556 100644 | ||
5272 | --- a/net/mac80211/mesh_pathtbl.c | ||
5273 | +++ b/net/mac80211/mesh_pathtbl.c | ||
5274 | @@ -59,8 +59,10 @@ static struct mesh_table *mesh_table_alloc(void) | ||
5275 | return NULL; | ||
5276 | |||
5277 | INIT_HLIST_HEAD(&newtbl->known_gates); | ||
5278 | + INIT_HLIST_HEAD(&newtbl->walk_head); | ||
5279 | atomic_set(&newtbl->entries, 0); | ||
5280 | spin_lock_init(&newtbl->gates_lock); | ||
5281 | + spin_lock_init(&newtbl->walk_lock); | ||
5282 | |||
5283 | return newtbl; | ||
5284 | } | ||
5285 | @@ -249,28 +251,15 @@ mpp_path_lookup(struct ieee80211_sub_if_data *sdata, const u8 *dst) | ||
5286 | static struct mesh_path * | ||
5287 | __mesh_path_lookup_by_idx(struct mesh_table *tbl, int idx) | ||
5288 | { | ||
5289 | - int i = 0, ret; | ||
5290 | - struct mesh_path *mpath = NULL; | ||
5291 | - struct rhashtable_iter iter; | ||
5292 | - | ||
5293 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_ATOMIC); | ||
5294 | - if (ret) | ||
5295 | - return NULL; | ||
5296 | - | ||
5297 | - rhashtable_walk_start(&iter); | ||
5298 | + int i = 0; | ||
5299 | + struct mesh_path *mpath; | ||
5300 | |||
5301 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5302 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5303 | - continue; | ||
5304 | - if (IS_ERR(mpath)) | ||
5305 | - break; | ||
5306 | + hlist_for_each_entry_rcu(mpath, &tbl->walk_head, walk_list) { | ||
5307 | if (i++ == idx) | ||
5308 | break; | ||
5309 | } | ||
5310 | - rhashtable_walk_stop(&iter); | ||
5311 | - rhashtable_walk_exit(&iter); | ||
5312 | |||
5313 | - if (IS_ERR(mpath) || !mpath) | ||
5314 | + if (!mpath) | ||
5315 | return NULL; | ||
5316 | |||
5317 | if (mpath_expired(mpath)) { | ||
5318 | @@ -432,6 +421,7 @@ struct mesh_path *mesh_path_add(struct ieee80211_sub_if_data *sdata, | ||
5319 | return ERR_PTR(-ENOMEM); | ||
5320 | |||
5321 | tbl = sdata->u.mesh.mesh_paths; | ||
5322 | + spin_lock_bh(&tbl->walk_lock); | ||
5323 | do { | ||
5324 | ret = rhashtable_lookup_insert_fast(&tbl->rhead, | ||
5325 | &new_mpath->rhash, | ||
5326 | @@ -441,20 +431,20 @@ struct mesh_path *mesh_path_add(struct ieee80211_sub_if_data *sdata, | ||
5327 | mpath = rhashtable_lookup_fast(&tbl->rhead, | ||
5328 | dst, | ||
5329 | mesh_rht_params); | ||
5330 | - | ||
5331 | + else if (!ret) | ||
5332 | + hlist_add_head(&new_mpath->walk_list, &tbl->walk_head); | ||
5333 | } while (unlikely(ret == -EEXIST && !mpath)); | ||
5334 | + spin_unlock_bh(&tbl->walk_lock); | ||
5335 | |||
5336 | - if (ret && ret != -EEXIST) | ||
5337 | - return ERR_PTR(ret); | ||
5338 | - | ||
5339 | - /* At this point either new_mpath was added, or we found a | ||
5340 | - * matching entry already in the table; in the latter case | ||
5341 | - * free the unnecessary new entry. | ||
5342 | - */ | ||
5343 | - if (ret == -EEXIST) { | ||
5344 | + if (ret) { | ||
5345 | kfree(new_mpath); | ||
5346 | + | ||
5347 | + if (ret != -EEXIST) | ||
5348 | + return ERR_PTR(ret); | ||
5349 | + | ||
5350 | new_mpath = mpath; | ||
5351 | } | ||
5352 | + | ||
5353 | sdata->u.mesh.mesh_paths_generation++; | ||
5354 | return new_mpath; | ||
5355 | } | ||
5356 | @@ -480,9 +470,17 @@ int mpp_path_add(struct ieee80211_sub_if_data *sdata, | ||
5357 | |||
5358 | memcpy(new_mpath->mpp, mpp, ETH_ALEN); | ||
5359 | tbl = sdata->u.mesh.mpp_paths; | ||
5360 | + | ||
5361 | + spin_lock_bh(&tbl->walk_lock); | ||
5362 | ret = rhashtable_lookup_insert_fast(&tbl->rhead, | ||
5363 | &new_mpath->rhash, | ||
5364 | mesh_rht_params); | ||
5365 | + if (!ret) | ||
5366 | + hlist_add_head_rcu(&new_mpath->walk_list, &tbl->walk_head); | ||
5367 | + spin_unlock_bh(&tbl->walk_lock); | ||
5368 | + | ||
5369 | + if (ret) | ||
5370 | + kfree(new_mpath); | ||
5371 | |||
5372 | sdata->u.mesh.mpp_paths_generation++; | ||
5373 | return ret; | ||
5374 | @@ -503,20 +501,9 @@ void mesh_plink_broken(struct sta_info *sta) | ||
5375 | struct mesh_table *tbl = sdata->u.mesh.mesh_paths; | ||
5376 | static const u8 bcast[ETH_ALEN] = {0xff, 0xff, 0xff, 0xff, 0xff, 0xff}; | ||
5377 | struct mesh_path *mpath; | ||
5378 | - struct rhashtable_iter iter; | ||
5379 | - int ret; | ||
5380 | - | ||
5381 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_ATOMIC); | ||
5382 | - if (ret) | ||
5383 | - return; | ||
5384 | |||
5385 | - rhashtable_walk_start(&iter); | ||
5386 | - | ||
5387 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5388 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5389 | - continue; | ||
5390 | - if (IS_ERR(mpath)) | ||
5391 | - break; | ||
5392 | + rcu_read_lock(); | ||
5393 | + hlist_for_each_entry_rcu(mpath, &tbl->walk_head, walk_list) { | ||
5394 | if (rcu_access_pointer(mpath->next_hop) == sta && | ||
5395 | mpath->flags & MESH_PATH_ACTIVE && | ||
5396 | !(mpath->flags & MESH_PATH_FIXED)) { | ||
5397 | @@ -530,8 +517,7 @@ void mesh_plink_broken(struct sta_info *sta) | ||
5398 | WLAN_REASON_MESH_PATH_DEST_UNREACHABLE, bcast); | ||
5399 | } | ||
5400 | } | ||
5401 | - rhashtable_walk_stop(&iter); | ||
5402 | - rhashtable_walk_exit(&iter); | ||
5403 | + rcu_read_unlock(); | ||
5404 | } | ||
5405 | |||
5406 | static void mesh_path_free_rcu(struct mesh_table *tbl, | ||
5407 | @@ -551,6 +537,7 @@ static void mesh_path_free_rcu(struct mesh_table *tbl, | ||
5408 | |||
5409 | static void __mesh_path_del(struct mesh_table *tbl, struct mesh_path *mpath) | ||
5410 | { | ||
5411 | + hlist_del_rcu(&mpath->walk_list); | ||
5412 | rhashtable_remove_fast(&tbl->rhead, &mpath->rhash, mesh_rht_params); | ||
5413 | mesh_path_free_rcu(tbl, mpath); | ||
5414 | } | ||
5415 | @@ -571,27 +558,14 @@ void mesh_path_flush_by_nexthop(struct sta_info *sta) | ||
5416 | struct ieee80211_sub_if_data *sdata = sta->sdata; | ||
5417 | struct mesh_table *tbl = sdata->u.mesh.mesh_paths; | ||
5418 | struct mesh_path *mpath; | ||
5419 | - struct rhashtable_iter iter; | ||
5420 | - int ret; | ||
5421 | - | ||
5422 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_ATOMIC); | ||
5423 | - if (ret) | ||
5424 | - return; | ||
5425 | - | ||
5426 | - rhashtable_walk_start(&iter); | ||
5427 | - | ||
5428 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5429 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5430 | - continue; | ||
5431 | - if (IS_ERR(mpath)) | ||
5432 | - break; | ||
5433 | + struct hlist_node *n; | ||
5434 | |||
5435 | + spin_lock_bh(&tbl->walk_lock); | ||
5436 | + hlist_for_each_entry_safe(mpath, n, &tbl->walk_head, walk_list) { | ||
5437 | if (rcu_access_pointer(mpath->next_hop) == sta) | ||
5438 | __mesh_path_del(tbl, mpath); | ||
5439 | } | ||
5440 | - | ||
5441 | - rhashtable_walk_stop(&iter); | ||
5442 | - rhashtable_walk_exit(&iter); | ||
5443 | + spin_unlock_bh(&tbl->walk_lock); | ||
5444 | } | ||
5445 | |||
5446 | static void mpp_flush_by_proxy(struct ieee80211_sub_if_data *sdata, | ||
5447 | @@ -599,51 +573,26 @@ static void mpp_flush_by_proxy(struct ieee80211_sub_if_data *sdata, | ||
5448 | { | ||
5449 | struct mesh_table *tbl = sdata->u.mesh.mpp_paths; | ||
5450 | struct mesh_path *mpath; | ||
5451 | - struct rhashtable_iter iter; | ||
5452 | - int ret; | ||
5453 | - | ||
5454 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_ATOMIC); | ||
5455 | - if (ret) | ||
5456 | - return; | ||
5457 | - | ||
5458 | - rhashtable_walk_start(&iter); | ||
5459 | - | ||
5460 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5461 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5462 | - continue; | ||
5463 | - if (IS_ERR(mpath)) | ||
5464 | - break; | ||
5465 | + struct hlist_node *n; | ||
5466 | |||
5467 | + spin_lock_bh(&tbl->walk_lock); | ||
5468 | + hlist_for_each_entry_safe(mpath, n, &tbl->walk_head, walk_list) { | ||
5469 | if (ether_addr_equal(mpath->mpp, proxy)) | ||
5470 | __mesh_path_del(tbl, mpath); | ||
5471 | } | ||
5472 | - | ||
5473 | - rhashtable_walk_stop(&iter); | ||
5474 | - rhashtable_walk_exit(&iter); | ||
5475 | + spin_unlock_bh(&tbl->walk_lock); | ||
5476 | } | ||
5477 | |||
5478 | static void table_flush_by_iface(struct mesh_table *tbl) | ||
5479 | { | ||
5480 | struct mesh_path *mpath; | ||
5481 | - struct rhashtable_iter iter; | ||
5482 | - int ret; | ||
5483 | - | ||
5484 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_ATOMIC); | ||
5485 | - if (ret) | ||
5486 | - return; | ||
5487 | - | ||
5488 | - rhashtable_walk_start(&iter); | ||
5489 | + struct hlist_node *n; | ||
5490 | |||
5491 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5492 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5493 | - continue; | ||
5494 | - if (IS_ERR(mpath)) | ||
5495 | - break; | ||
5496 | + spin_lock_bh(&tbl->walk_lock); | ||
5497 | + hlist_for_each_entry_safe(mpath, n, &tbl->walk_head, walk_list) { | ||
5498 | __mesh_path_del(tbl, mpath); | ||
5499 | } | ||
5500 | - | ||
5501 | - rhashtable_walk_stop(&iter); | ||
5502 | - rhashtable_walk_exit(&iter); | ||
5503 | + spin_unlock_bh(&tbl->walk_lock); | ||
5504 | } | ||
5505 | |||
5506 | /** | ||
5507 | @@ -675,7 +624,7 @@ static int table_path_del(struct mesh_table *tbl, | ||
5508 | { | ||
5509 | struct mesh_path *mpath; | ||
5510 | |||
5511 | - rcu_read_lock(); | ||
5512 | + spin_lock_bh(&tbl->walk_lock); | ||
5513 | mpath = rhashtable_lookup_fast(&tbl->rhead, addr, mesh_rht_params); | ||
5514 | if (!mpath) { | ||
5515 | rcu_read_unlock(); | ||
5516 | @@ -683,7 +632,7 @@ static int table_path_del(struct mesh_table *tbl, | ||
5517 | } | ||
5518 | |||
5519 | __mesh_path_del(tbl, mpath); | ||
5520 | - rcu_read_unlock(); | ||
5521 | + spin_unlock_bh(&tbl->walk_lock); | ||
5522 | return 0; | ||
5523 | } | ||
5524 | |||
5525 | @@ -854,28 +803,16 @@ void mesh_path_tbl_expire(struct ieee80211_sub_if_data *sdata, | ||
5526 | struct mesh_table *tbl) | ||
5527 | { | ||
5528 | struct mesh_path *mpath; | ||
5529 | - struct rhashtable_iter iter; | ||
5530 | - int ret; | ||
5531 | + struct hlist_node *n; | ||
5532 | |||
5533 | - ret = rhashtable_walk_init(&tbl->rhead, &iter, GFP_KERNEL); | ||
5534 | - if (ret) | ||
5535 | - return; | ||
5536 | - | ||
5537 | - rhashtable_walk_start(&iter); | ||
5538 | - | ||
5539 | - while ((mpath = rhashtable_walk_next(&iter))) { | ||
5540 | - if (IS_ERR(mpath) && PTR_ERR(mpath) == -EAGAIN) | ||
5541 | - continue; | ||
5542 | - if (IS_ERR(mpath)) | ||
5543 | - break; | ||
5544 | + spin_lock_bh(&tbl->walk_lock); | ||
5545 | + hlist_for_each_entry_safe(mpath, n, &tbl->walk_head, walk_list) { | ||
5546 | if ((!(mpath->flags & MESH_PATH_RESOLVING)) && | ||
5547 | (!(mpath->flags & MESH_PATH_FIXED)) && | ||
5548 | time_after(jiffies, mpath->exp_time + MESH_PATH_EXPIRE)) | ||
5549 | __mesh_path_del(tbl, mpath); | ||
5550 | } | ||
5551 | - | ||
5552 | - rhashtable_walk_stop(&iter); | ||
5553 | - rhashtable_walk_exit(&iter); | ||
5554 | + spin_unlock_bh(&tbl->walk_lock); | ||
5555 | } | ||
5556 | |||
5557 | void mesh_path_expire(struct ieee80211_sub_if_data *sdata) | ||
5558 | diff --git a/net/mac80211/rx.c b/net/mac80211/rx.c | ||
5559 | index 85c365fc7a0c1..46ecc417c4210 100644 | ||
5560 | --- a/net/mac80211/rx.c | ||
5561 | +++ b/net/mac80211/rx.c | ||
5562 | @@ -2640,6 +2640,7 @@ ieee80211_rx_h_mesh_fwding(struct ieee80211_rx_data *rx) | ||
5563 | struct ieee80211_sub_if_data *sdata = rx->sdata; | ||
5564 | struct ieee80211_if_mesh *ifmsh = &sdata->u.mesh; | ||
5565 | u16 ac, q, hdrlen; | ||
5566 | + int tailroom = 0; | ||
5567 | |||
5568 | hdr = (struct ieee80211_hdr *) skb->data; | ||
5569 | hdrlen = ieee80211_hdrlen(hdr->frame_control); | ||
5570 | @@ -2726,8 +2727,12 @@ ieee80211_rx_h_mesh_fwding(struct ieee80211_rx_data *rx) | ||
5571 | if (!ifmsh->mshcfg.dot11MeshForwarding) | ||
5572 | goto out; | ||
5573 | |||
5574 | + if (sdata->crypto_tx_tailroom_needed_cnt) | ||
5575 | + tailroom = IEEE80211_ENCRYPT_TAILROOM; | ||
5576 | + | ||
5577 | fwd_skb = skb_copy_expand(skb, local->tx_headroom + | ||
5578 | - sdata->encrypt_headroom, 0, GFP_ATOMIC); | ||
5579 | + sdata->encrypt_headroom, | ||
5580 | + tailroom, GFP_ATOMIC); | ||
5581 | if (!fwd_skb) | ||
5582 | goto out; | ||
5583 | |||
5584 | diff --git a/net/netfilter/nf_flow_table_core.c b/net/netfilter/nf_flow_table_core.c | ||
5585 | index b7a4816add765..cc91b4d6aa22f 100644 | ||
5586 | --- a/net/netfilter/nf_flow_table_core.c | ||
5587 | +++ b/net/netfilter/nf_flow_table_core.c | ||
5588 | @@ -28,6 +28,7 @@ flow_offload_fill_dir(struct flow_offload *flow, struct nf_conn *ct, | ||
5589 | { | ||
5590 | struct flow_offload_tuple *ft = &flow->tuplehash[dir].tuple; | ||
5591 | struct nf_conntrack_tuple *ctt = &ct->tuplehash[dir].tuple; | ||
5592 | + struct dst_entry *other_dst = route->tuple[!dir].dst; | ||
5593 | struct dst_entry *dst = route->tuple[dir].dst; | ||
5594 | |||
5595 | ft->dir = dir; | ||
5596 | @@ -50,8 +51,8 @@ flow_offload_fill_dir(struct flow_offload *flow, struct nf_conn *ct, | ||
5597 | ft->src_port = ctt->src.u.tcp.port; | ||
5598 | ft->dst_port = ctt->dst.u.tcp.port; | ||
5599 | |||
5600 | - ft->iifidx = route->tuple[dir].ifindex; | ||
5601 | - ft->oifidx = route->tuple[!dir].ifindex; | ||
5602 | + ft->iifidx = other_dst->dev->ifindex; | ||
5603 | + ft->oifidx = dst->dev->ifindex; | ||
5604 | ft->dst_cache = dst; | ||
5605 | } | ||
5606 | |||
5607 | diff --git a/net/netfilter/nf_tables_api.c b/net/netfilter/nf_tables_api.c | ||
5608 | index 6e548d7c9f67b..5114a0d2a41eb 100644 | ||
5609 | --- a/net/netfilter/nf_tables_api.c | ||
5610 | +++ b/net/netfilter/nf_tables_api.c | ||
5611 | @@ -307,6 +307,9 @@ static int nft_delrule_by_chain(struct nft_ctx *ctx) | ||
5612 | int err; | ||
5613 | |||
5614 | list_for_each_entry(rule, &ctx->chain->rules, list) { | ||
5615 | + if (!nft_is_active_next(ctx->net, rule)) | ||
5616 | + continue; | ||
5617 | + | ||
5618 | err = nft_delrule(ctx, rule); | ||
5619 | if (err < 0) | ||
5620 | return err; | ||
5621 | @@ -4474,6 +4477,8 @@ err6: | ||
5622 | err5: | ||
5623 | kfree(trans); | ||
5624 | err4: | ||
5625 | + if (obj) | ||
5626 | + obj->use--; | ||
5627 | kfree(elem.priv); | ||
5628 | err3: | ||
5629 | if (nla[NFTA_SET_ELEM_DATA] != NULL) | ||
5630 | diff --git a/net/netfilter/nfnetlink_osf.c b/net/netfilter/nfnetlink_osf.c | ||
5631 | index 6f41dd74729d9..1f1d90c1716b5 100644 | ||
5632 | --- a/net/netfilter/nfnetlink_osf.c | ||
5633 | +++ b/net/netfilter/nfnetlink_osf.c | ||
5634 | @@ -66,6 +66,7 @@ static bool nf_osf_match_one(const struct sk_buff *skb, | ||
5635 | int ttl_check, | ||
5636 | struct nf_osf_hdr_ctx *ctx) | ||
5637 | { | ||
5638 | + const __u8 *optpinit = ctx->optp; | ||
5639 | unsigned int check_WSS = 0; | ||
5640 | int fmatch = FMATCH_WRONG; | ||
5641 | int foptsize, optnum; | ||
5642 | @@ -155,6 +156,9 @@ static bool nf_osf_match_one(const struct sk_buff *skb, | ||
5643 | } | ||
5644 | } | ||
5645 | |||
5646 | + if (fmatch != FMATCH_OK) | ||
5647 | + ctx->optp = optpinit; | ||
5648 | + | ||
5649 | return fmatch == FMATCH_OK; | ||
5650 | } | ||
5651 | |||
5652 | diff --git a/net/netfilter/nft_compat.c b/net/netfilter/nft_compat.c | ||
5653 | index 7334e0b80a5ef..c90a4640723f5 100644 | ||
5654 | --- a/net/netfilter/nft_compat.c | ||
5655 | +++ b/net/netfilter/nft_compat.c | ||
5656 | @@ -282,6 +282,7 @@ nft_target_destroy(const struct nft_ctx *ctx, const struct nft_expr *expr) | ||
5657 | { | ||
5658 | struct xt_target *target = expr->ops->data; | ||
5659 | void *info = nft_expr_priv(expr); | ||
5660 | + struct module *me = target->me; | ||
5661 | struct xt_tgdtor_param par; | ||
5662 | |||
5663 | par.net = ctx->net; | ||
5664 | @@ -292,7 +293,7 @@ nft_target_destroy(const struct nft_ctx *ctx, const struct nft_expr *expr) | ||
5665 | par.target->destroy(&par); | ||
5666 | |||
5667 | if (nft_xt_put(container_of(expr->ops, struct nft_xt, ops))) | ||
5668 | - module_put(target->me); | ||
5669 | + module_put(me); | ||
5670 | } | ||
5671 | |||
5672 | static int nft_extension_dump_info(struct sk_buff *skb, int attr, | ||
5673 | diff --git a/net/netfilter/nft_flow_offload.c b/net/netfilter/nft_flow_offload.c | ||
5674 | index 974525eb92df7..6e6b9adf7d387 100644 | ||
5675 | --- a/net/netfilter/nft_flow_offload.c | ||
5676 | +++ b/net/netfilter/nft_flow_offload.c | ||
5677 | @@ -12,6 +12,7 @@ | ||
5678 | #include <net/netfilter/nf_conntrack_core.h> | ||
5679 | #include <linux/netfilter/nf_conntrack_common.h> | ||
5680 | #include <net/netfilter/nf_flow_table.h> | ||
5681 | +#include <net/netfilter/nf_conntrack_helper.h> | ||
5682 | |||
5683 | struct nft_flow_offload { | ||
5684 | struct nft_flowtable *flowtable; | ||
5685 | @@ -29,10 +30,12 @@ static int nft_flow_route(const struct nft_pktinfo *pkt, | ||
5686 | memset(&fl, 0, sizeof(fl)); | ||
5687 | switch (nft_pf(pkt)) { | ||
5688 | case NFPROTO_IPV4: | ||
5689 | - fl.u.ip4.daddr = ct->tuplehash[!dir].tuple.dst.u3.ip; | ||
5690 | + fl.u.ip4.daddr = ct->tuplehash[dir].tuple.src.u3.ip; | ||
5691 | + fl.u.ip4.flowi4_oif = nft_in(pkt)->ifindex; | ||
5692 | break; | ||
5693 | case NFPROTO_IPV6: | ||
5694 | - fl.u.ip6.daddr = ct->tuplehash[!dir].tuple.dst.u3.in6; | ||
5695 | + fl.u.ip6.daddr = ct->tuplehash[dir].tuple.src.u3.in6; | ||
5696 | + fl.u.ip6.flowi6_oif = nft_in(pkt)->ifindex; | ||
5697 | break; | ||
5698 | } | ||
5699 | |||
5700 | @@ -41,9 +44,7 @@ static int nft_flow_route(const struct nft_pktinfo *pkt, | ||
5701 | return -ENOENT; | ||
5702 | |||
5703 | route->tuple[dir].dst = this_dst; | ||
5704 | - route->tuple[dir].ifindex = nft_in(pkt)->ifindex; | ||
5705 | route->tuple[!dir].dst = other_dst; | ||
5706 | - route->tuple[!dir].ifindex = nft_out(pkt)->ifindex; | ||
5707 | |||
5708 | return 0; | ||
5709 | } | ||
5710 | @@ -66,6 +67,7 @@ static void nft_flow_offload_eval(const struct nft_expr *expr, | ||
5711 | { | ||
5712 | struct nft_flow_offload *priv = nft_expr_priv(expr); | ||
5713 | struct nf_flowtable *flowtable = &priv->flowtable->data; | ||
5714 | + const struct nf_conn_help *help; | ||
5715 | enum ip_conntrack_info ctinfo; | ||
5716 | struct nf_flow_route route; | ||
5717 | struct flow_offload *flow; | ||
5718 | @@ -88,7 +90,8 @@ static void nft_flow_offload_eval(const struct nft_expr *expr, | ||
5719 | goto out; | ||
5720 | } | ||
5721 | |||
5722 | - if (test_bit(IPS_HELPER_BIT, &ct->status)) | ||
5723 | + help = nfct_help(ct); | ||
5724 | + if (help) | ||
5725 | goto out; | ||
5726 | |||
5727 | if (ctinfo == IP_CT_NEW || | ||
5728 | diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c | ||
5729 | index 3b1a78906bc0c..1cd1d83a4be08 100644 | ||
5730 | --- a/net/packet/af_packet.c | ||
5731 | +++ b/net/packet/af_packet.c | ||
5732 | @@ -4292,7 +4292,7 @@ static int packet_set_ring(struct sock *sk, union tpacket_req_u *req_u, | ||
5733 | rb->frames_per_block = req->tp_block_size / req->tp_frame_size; | ||
5734 | if (unlikely(rb->frames_per_block == 0)) | ||
5735 | goto out; | ||
5736 | - if (unlikely(req->tp_block_size > UINT_MAX / req->tp_block_nr)) | ||
5737 | + if (unlikely(rb->frames_per_block > UINT_MAX / req->tp_block_nr)) | ||
5738 | goto out; | ||
5739 | if (unlikely((rb->frames_per_block * req->tp_block_nr) != | ||
5740 | req->tp_frame_nr)) | ||
5741 | diff --git a/net/sched/cls_tcindex.c b/net/sched/cls_tcindex.c | ||
5742 | index 9ccc93f257db0..38bb882bb9587 100644 | ||
5743 | --- a/net/sched/cls_tcindex.c | ||
5744 | +++ b/net/sched/cls_tcindex.c | ||
5745 | @@ -48,7 +48,7 @@ struct tcindex_data { | ||
5746 | u32 hash; /* hash table size; 0 if undefined */ | ||
5747 | u32 alloc_hash; /* allocated size */ | ||
5748 | u32 fall_through; /* 0: only classify if explicit match */ | ||
5749 | - struct rcu_head rcu; | ||
5750 | + struct rcu_work rwork; | ||
5751 | }; | ||
5752 | |||
5753 | static inline int tcindex_filter_is_set(struct tcindex_filter_result *r) | ||
5754 | @@ -221,17 +221,11 @@ found: | ||
5755 | return 0; | ||
5756 | } | ||
5757 | |||
5758 | -static int tcindex_destroy_element(struct tcf_proto *tp, | ||
5759 | - void *arg, struct tcf_walker *walker) | ||
5760 | -{ | ||
5761 | - bool last; | ||
5762 | - | ||
5763 | - return tcindex_delete(tp, arg, &last, NULL); | ||
5764 | -} | ||
5765 | - | ||
5766 | -static void __tcindex_destroy(struct rcu_head *head) | ||
5767 | +static void tcindex_destroy_work(struct work_struct *work) | ||
5768 | { | ||
5769 | - struct tcindex_data *p = container_of(head, struct tcindex_data, rcu); | ||
5770 | + struct tcindex_data *p = container_of(to_rcu_work(work), | ||
5771 | + struct tcindex_data, | ||
5772 | + rwork); | ||
5773 | |||
5774 | kfree(p->perfect); | ||
5775 | kfree(p->h); | ||
5776 | @@ -258,9 +252,11 @@ static int tcindex_filter_result_init(struct tcindex_filter_result *r) | ||
5777 | return tcf_exts_init(&r->exts, TCA_TCINDEX_ACT, TCA_TCINDEX_POLICE); | ||
5778 | } | ||
5779 | |||
5780 | -static void __tcindex_partial_destroy(struct rcu_head *head) | ||
5781 | +static void tcindex_partial_destroy_work(struct work_struct *work) | ||
5782 | { | ||
5783 | - struct tcindex_data *p = container_of(head, struct tcindex_data, rcu); | ||
5784 | + struct tcindex_data *p = container_of(to_rcu_work(work), | ||
5785 | + struct tcindex_data, | ||
5786 | + rwork); | ||
5787 | |||
5788 | kfree(p->perfect); | ||
5789 | kfree(p); | ||
5790 | @@ -275,7 +271,7 @@ static void tcindex_free_perfect_hash(struct tcindex_data *cp) | ||
5791 | kfree(cp->perfect); | ||
5792 | } | ||
5793 | |||
5794 | -static int tcindex_alloc_perfect_hash(struct tcindex_data *cp) | ||
5795 | +static int tcindex_alloc_perfect_hash(struct net *net, struct tcindex_data *cp) | ||
5796 | { | ||
5797 | int i, err = 0; | ||
5798 | |||
5799 | @@ -289,6 +285,9 @@ static int tcindex_alloc_perfect_hash(struct tcindex_data *cp) | ||
5800 | TCA_TCINDEX_ACT, TCA_TCINDEX_POLICE); | ||
5801 | if (err < 0) | ||
5802 | goto errout; | ||
5803 | +#ifdef CONFIG_NET_CLS_ACT | ||
5804 | + cp->perfect[i].exts.net = net; | ||
5805 | +#endif | ||
5806 | } | ||
5807 | |||
5808 | return 0; | ||
5809 | @@ -305,9 +304,9 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5810 | struct nlattr *est, bool ovr, struct netlink_ext_ack *extack) | ||
5811 | { | ||
5812 | struct tcindex_filter_result new_filter_result, *old_r = r; | ||
5813 | - struct tcindex_filter_result cr; | ||
5814 | struct tcindex_data *cp = NULL, *oldp; | ||
5815 | struct tcindex_filter *f = NULL; /* make gcc behave */ | ||
5816 | + struct tcf_result cr = {}; | ||
5817 | int err, balloc = 0; | ||
5818 | struct tcf_exts e; | ||
5819 | |||
5820 | @@ -337,7 +336,7 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5821 | if (p->perfect) { | ||
5822 | int i; | ||
5823 | |||
5824 | - if (tcindex_alloc_perfect_hash(cp) < 0) | ||
5825 | + if (tcindex_alloc_perfect_hash(net, cp) < 0) | ||
5826 | goto errout; | ||
5827 | for (i = 0; i < cp->hash; i++) | ||
5828 | cp->perfect[i].res = p->perfect[i].res; | ||
5829 | @@ -346,13 +345,10 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5830 | cp->h = p->h; | ||
5831 | |||
5832 | err = tcindex_filter_result_init(&new_filter_result); | ||
5833 | - if (err < 0) | ||
5834 | - goto errout1; | ||
5835 | - err = tcindex_filter_result_init(&cr); | ||
5836 | if (err < 0) | ||
5837 | goto errout1; | ||
5838 | if (old_r) | ||
5839 | - cr.res = r->res; | ||
5840 | + cr = r->res; | ||
5841 | |||
5842 | if (tb[TCA_TCINDEX_HASH]) | ||
5843 | cp->hash = nla_get_u32(tb[TCA_TCINDEX_HASH]); | ||
5844 | @@ -406,7 +402,7 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5845 | err = -ENOMEM; | ||
5846 | if (!cp->perfect && !cp->h) { | ||
5847 | if (valid_perfect_hash(cp)) { | ||
5848 | - if (tcindex_alloc_perfect_hash(cp) < 0) | ||
5849 | + if (tcindex_alloc_perfect_hash(net, cp) < 0) | ||
5850 | goto errout_alloc; | ||
5851 | balloc = 1; | ||
5852 | } else { | ||
5853 | @@ -443,8 +439,8 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5854 | } | ||
5855 | |||
5856 | if (tb[TCA_TCINDEX_CLASSID]) { | ||
5857 | - cr.res.classid = nla_get_u32(tb[TCA_TCINDEX_CLASSID]); | ||
5858 | - tcf_bind_filter(tp, &cr.res, base); | ||
5859 | + cr.classid = nla_get_u32(tb[TCA_TCINDEX_CLASSID]); | ||
5860 | + tcf_bind_filter(tp, &cr, base); | ||
5861 | } | ||
5862 | |||
5863 | if (old_r && old_r != r) { | ||
5864 | @@ -456,7 +452,7 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5865 | } | ||
5866 | |||
5867 | oldp = p; | ||
5868 | - r->res = cr.res; | ||
5869 | + r->res = cr; | ||
5870 | tcf_exts_change(&r->exts, &e); | ||
5871 | |||
5872 | rcu_assign_pointer(tp->root, cp); | ||
5873 | @@ -475,10 +471,12 @@ tcindex_set_parms(struct net *net, struct tcf_proto *tp, unsigned long base, | ||
5874 | ; /* nothing */ | ||
5875 | |||
5876 | rcu_assign_pointer(*fp, f); | ||
5877 | + } else { | ||
5878 | + tcf_exts_destroy(&new_filter_result.exts); | ||
5879 | } | ||
5880 | |||
5881 | if (oldp) | ||
5882 | - call_rcu(&oldp->rcu, __tcindex_partial_destroy); | ||
5883 | + tcf_queue_work(&oldp->rwork, tcindex_partial_destroy_work); | ||
5884 | return 0; | ||
5885 | |||
5886 | errout_alloc: | ||
5887 | @@ -487,7 +485,6 @@ errout_alloc: | ||
5888 | else if (balloc == 2) | ||
5889 | kfree(cp->h); | ||
5890 | errout1: | ||
5891 | - tcf_exts_destroy(&cr.exts); | ||
5892 | tcf_exts_destroy(&new_filter_result.exts); | ||
5893 | errout: | ||
5894 | kfree(cp); | ||
5895 | @@ -562,15 +559,34 @@ static void tcindex_destroy(struct tcf_proto *tp, | ||
5896 | struct netlink_ext_ack *extack) | ||
5897 | { | ||
5898 | struct tcindex_data *p = rtnl_dereference(tp->root); | ||
5899 | - struct tcf_walker walker; | ||
5900 | + int i; | ||
5901 | |||
5902 | pr_debug("tcindex_destroy(tp %p),p %p\n", tp, p); | ||
5903 | - walker.count = 0; | ||
5904 | - walker.skip = 0; | ||
5905 | - walker.fn = tcindex_destroy_element; | ||
5906 | - tcindex_walk(tp, &walker); | ||
5907 | |||
5908 | - call_rcu(&p->rcu, __tcindex_destroy); | ||
5909 | + if (p->perfect) { | ||
5910 | + for (i = 0; i < p->hash; i++) { | ||
5911 | + struct tcindex_filter_result *r = p->perfect + i; | ||
5912 | + | ||
5913 | + tcf_unbind_filter(tp, &r->res); | ||
5914 | + if (tcf_exts_get_net(&r->exts)) | ||
5915 | + tcf_queue_work(&r->rwork, | ||
5916 | + tcindex_destroy_rexts_work); | ||
5917 | + else | ||
5918 | + __tcindex_destroy_rexts(r); | ||
5919 | + } | ||
5920 | + } | ||
5921 | + | ||
5922 | + for (i = 0; p->h && i < p->hash; i++) { | ||
5923 | + struct tcindex_filter *f, *next; | ||
5924 | + bool last; | ||
5925 | + | ||
5926 | + for (f = rtnl_dereference(p->h[i]); f; f = next) { | ||
5927 | + next = rtnl_dereference(f->next); | ||
5928 | + tcindex_delete(tp, &f->result, &last, NULL); | ||
5929 | + } | ||
5930 | + } | ||
5931 | + | ||
5932 | + tcf_queue_work(&p->rwork, tcindex_destroy_work); | ||
5933 | } | ||
5934 | |||
5935 | |||
5936 | diff --git a/net/sctp/diag.c b/net/sctp/diag.c | ||
5937 | index 078f01a8d582a..435847d98b51c 100644 | ||
5938 | --- a/net/sctp/diag.c | ||
5939 | +++ b/net/sctp/diag.c | ||
5940 | @@ -256,6 +256,7 @@ static size_t inet_assoc_attr_size(struct sctp_association *asoc) | ||
5941 | + nla_total_size(1) /* INET_DIAG_TOS */ | ||
5942 | + nla_total_size(1) /* INET_DIAG_TCLASS */ | ||
5943 | + nla_total_size(4) /* INET_DIAG_MARK */ | ||
5944 | + + nla_total_size(4) /* INET_DIAG_CLASS_ID */ | ||
5945 | + nla_total_size(addrlen * asoc->peer.transport_count) | ||
5946 | + nla_total_size(addrlen * addrcnt) | ||
5947 | + nla_total_size(sizeof(struct inet_diag_meminfo)) | ||
5948 | diff --git a/net/sctp/offload.c b/net/sctp/offload.c | ||
5949 | index 123e9f2dc2265..edfcf16e704c4 100644 | ||
5950 | --- a/net/sctp/offload.c | ||
5951 | +++ b/net/sctp/offload.c | ||
5952 | @@ -36,6 +36,7 @@ static __le32 sctp_gso_make_checksum(struct sk_buff *skb) | ||
5953 | { | ||
5954 | skb->ip_summed = CHECKSUM_NONE; | ||
5955 | skb->csum_not_inet = 0; | ||
5956 | + gso_reset_checksum(skb, ~0); | ||
5957 | return sctp_compute_cksum(skb, skb_transport_offset(skb)); | ||
5958 | } | ||
5959 | |||
5960 | diff --git a/net/sctp/stream.c b/net/sctp/stream.c | ||
5961 | index f24633114dfdf..2936ed17bf9ef 100644 | ||
5962 | --- a/net/sctp/stream.c | ||
5963 | +++ b/net/sctp/stream.c | ||
5964 | @@ -144,8 +144,10 @@ static void sctp_stream_outq_migrate(struct sctp_stream *stream, | ||
5965 | } | ||
5966 | } | ||
5967 | |||
5968 | - for (i = outcnt; i < stream->outcnt; i++) | ||
5969 | + for (i = outcnt; i < stream->outcnt; i++) { | ||
5970 | kfree(SCTP_SO(stream, i)->ext); | ||
5971 | + SCTP_SO(stream, i)->ext = NULL; | ||
5972 | + } | ||
5973 | } | ||
5974 | |||
5975 | static int sctp_stream_alloc_out(struct sctp_stream *stream, __u16 outcnt, | ||
5976 | diff --git a/net/socket.c b/net/socket.c | ||
5977 | index 334fcc617ef27..93a45f15ee40d 100644 | ||
5978 | --- a/net/socket.c | ||
5979 | +++ b/net/socket.c | ||
5980 | @@ -941,8 +941,7 @@ void dlci_ioctl_set(int (*hook) (unsigned int, void __user *)) | ||
5981 | EXPORT_SYMBOL(dlci_ioctl_set); | ||
5982 | |||
5983 | static long sock_do_ioctl(struct net *net, struct socket *sock, | ||
5984 | - unsigned int cmd, unsigned long arg, | ||
5985 | - unsigned int ifreq_size) | ||
5986 | + unsigned int cmd, unsigned long arg) | ||
5987 | { | ||
5988 | int err; | ||
5989 | void __user *argp = (void __user *)arg; | ||
5990 | @@ -968,11 +967,11 @@ static long sock_do_ioctl(struct net *net, struct socket *sock, | ||
5991 | } else { | ||
5992 | struct ifreq ifr; | ||
5993 | bool need_copyout; | ||
5994 | - if (copy_from_user(&ifr, argp, ifreq_size)) | ||
5995 | + if (copy_from_user(&ifr, argp, sizeof(struct ifreq))) | ||
5996 | return -EFAULT; | ||
5997 | err = dev_ioctl(net, cmd, &ifr, &need_copyout); | ||
5998 | if (!err && need_copyout) | ||
5999 | - if (copy_to_user(argp, &ifr, ifreq_size)) | ||
6000 | + if (copy_to_user(argp, &ifr, sizeof(struct ifreq))) | ||
6001 | return -EFAULT; | ||
6002 | } | ||
6003 | return err; | ||
6004 | @@ -1071,8 +1070,7 @@ static long sock_ioctl(struct file *file, unsigned cmd, unsigned long arg) | ||
6005 | err = open_related_ns(&net->ns, get_net_ns); | ||
6006 | break; | ||
6007 | default: | ||
6008 | - err = sock_do_ioctl(net, sock, cmd, arg, | ||
6009 | - sizeof(struct ifreq)); | ||
6010 | + err = sock_do_ioctl(net, sock, cmd, arg); | ||
6011 | break; | ||
6012 | } | ||
6013 | return err; | ||
6014 | @@ -2750,8 +2748,7 @@ static int do_siocgstamp(struct net *net, struct socket *sock, | ||
6015 | int err; | ||
6016 | |||
6017 | set_fs(KERNEL_DS); | ||
6018 | - err = sock_do_ioctl(net, sock, cmd, (unsigned long)&ktv, | ||
6019 | - sizeof(struct compat_ifreq)); | ||
6020 | + err = sock_do_ioctl(net, sock, cmd, (unsigned long)&ktv); | ||
6021 | set_fs(old_fs); | ||
6022 | if (!err) | ||
6023 | err = compat_put_timeval(&ktv, up); | ||
6024 | @@ -2767,8 +2764,7 @@ static int do_siocgstampns(struct net *net, struct socket *sock, | ||
6025 | int err; | ||
6026 | |||
6027 | set_fs(KERNEL_DS); | ||
6028 | - err = sock_do_ioctl(net, sock, cmd, (unsigned long)&kts, | ||
6029 | - sizeof(struct compat_ifreq)); | ||
6030 | + err = sock_do_ioctl(net, sock, cmd, (unsigned long)&kts); | ||
6031 | set_fs(old_fs); | ||
6032 | if (!err) | ||
6033 | err = compat_put_timespec(&kts, up); | ||
6034 | @@ -2964,6 +2960,54 @@ static int compat_ifr_data_ioctl(struct net *net, unsigned int cmd, | ||
6035 | return dev_ioctl(net, cmd, &ifreq, NULL); | ||
6036 | } | ||
6037 | |||
6038 | +static int compat_ifreq_ioctl(struct net *net, struct socket *sock, | ||
6039 | + unsigned int cmd, | ||
6040 | + struct compat_ifreq __user *uifr32) | ||
6041 | +{ | ||
6042 | + struct ifreq __user *uifr; | ||
6043 | + int err; | ||
6044 | + | ||
6045 | + /* Handle the fact that while struct ifreq has the same *layout* on | ||
6046 | + * 32/64 for everything but ifreq::ifru_ifmap and ifreq::ifru_data, | ||
6047 | + * which are handled elsewhere, it still has different *size* due to | ||
6048 | + * ifreq::ifru_ifmap (which is 16 bytes on 32 bit, 24 bytes on 64-bit, | ||
6049 | + * resulting in struct ifreq being 32 and 40 bytes respectively). | ||
6050 | + * As a result, if the struct happens to be at the end of a page and | ||
6051 | + * the next page isn't readable/writable, we get a fault. To prevent | ||
6052 | + * that, copy back and forth to the full size. | ||
6053 | + */ | ||
6054 | + | ||
6055 | + uifr = compat_alloc_user_space(sizeof(*uifr)); | ||
6056 | + if (copy_in_user(uifr, uifr32, sizeof(*uifr32))) | ||
6057 | + return -EFAULT; | ||
6058 | + | ||
6059 | + err = sock_do_ioctl(net, sock, cmd, (unsigned long)uifr); | ||
6060 | + | ||
6061 | + if (!err) { | ||
6062 | + switch (cmd) { | ||
6063 | + case SIOCGIFFLAGS: | ||
6064 | + case SIOCGIFMETRIC: | ||
6065 | + case SIOCGIFMTU: | ||
6066 | + case SIOCGIFMEM: | ||
6067 | + case SIOCGIFHWADDR: | ||
6068 | + case SIOCGIFINDEX: | ||
6069 | + case SIOCGIFADDR: | ||
6070 | + case SIOCGIFBRDADDR: | ||
6071 | + case SIOCGIFDSTADDR: | ||
6072 | + case SIOCGIFNETMASK: | ||
6073 | + case SIOCGIFPFLAGS: | ||
6074 | + case SIOCGIFTXQLEN: | ||
6075 | + case SIOCGMIIPHY: | ||
6076 | + case SIOCGMIIREG: | ||
6077 | + case SIOCGIFNAME: | ||
6078 | + if (copy_in_user(uifr32, uifr, sizeof(*uifr32))) | ||
6079 | + err = -EFAULT; | ||
6080 | + break; | ||
6081 | + } | ||
6082 | + } | ||
6083 | + return err; | ||
6084 | +} | ||
6085 | + | ||
6086 | static int compat_sioc_ifmap(struct net *net, unsigned int cmd, | ||
6087 | struct compat_ifreq __user *uifr32) | ||
6088 | { | ||
6089 | @@ -3079,8 +3123,7 @@ static int routing_ioctl(struct net *net, struct socket *sock, | ||
6090 | } | ||
6091 | |||
6092 | set_fs(KERNEL_DS); | ||
6093 | - ret = sock_do_ioctl(net, sock, cmd, (unsigned long) r, | ||
6094 | - sizeof(struct compat_ifreq)); | ||
6095 | + ret = sock_do_ioctl(net, sock, cmd, (unsigned long) r); | ||
6096 | set_fs(old_fs); | ||
6097 | |||
6098 | out: | ||
6099 | @@ -3180,21 +3223,22 @@ static int compat_sock_ioctl_trans(struct file *file, struct socket *sock, | ||
6100 | case SIOCSIFTXQLEN: | ||
6101 | case SIOCBRADDIF: | ||
6102 | case SIOCBRDELIF: | ||
6103 | + case SIOCGIFNAME: | ||
6104 | case SIOCSIFNAME: | ||
6105 | case SIOCGMIIPHY: | ||
6106 | case SIOCGMIIREG: | ||
6107 | case SIOCSMIIREG: | ||
6108 | - case SIOCSARP: | ||
6109 | - case SIOCGARP: | ||
6110 | - case SIOCDARP: | ||
6111 | - case SIOCATMARK: | ||
6112 | case SIOCBONDENSLAVE: | ||
6113 | case SIOCBONDRELEASE: | ||
6114 | case SIOCBONDSETHWADDR: | ||
6115 | case SIOCBONDCHANGEACTIVE: | ||
6116 | - case SIOCGIFNAME: | ||
6117 | - return sock_do_ioctl(net, sock, cmd, arg, | ||
6118 | - sizeof(struct compat_ifreq)); | ||
6119 | + return compat_ifreq_ioctl(net, sock, cmd, argp); | ||
6120 | + | ||
6121 | + case SIOCSARP: | ||
6122 | + case SIOCGARP: | ||
6123 | + case SIOCDARP: | ||
6124 | + case SIOCATMARK: | ||
6125 | + return sock_do_ioctl(net, sock, cmd, arg); | ||
6126 | } | ||
6127 | |||
6128 | return -ENOIOCTLCMD; | ||
6129 | diff --git a/net/sunrpc/auth_gss/auth_gss.c b/net/sunrpc/auth_gss/auth_gss.c | ||
6130 | index ba765473d1f06..efeee5586b2ac 100644 | ||
6131 | --- a/net/sunrpc/auth_gss/auth_gss.c | ||
6132 | +++ b/net/sunrpc/auth_gss/auth_gss.c | ||
6133 | @@ -1563,8 +1563,10 @@ gss_marshal(struct rpc_task *task, __be32 *p) | ||
6134 | cred_len = p++; | ||
6135 | |||
6136 | spin_lock(&ctx->gc_seq_lock); | ||
6137 | - req->rq_seqno = ctx->gc_seq++; | ||
6138 | + req->rq_seqno = (ctx->gc_seq < MAXSEQ) ? ctx->gc_seq++ : MAXSEQ; | ||
6139 | spin_unlock(&ctx->gc_seq_lock); | ||
6140 | + if (req->rq_seqno == MAXSEQ) | ||
6141 | + goto out_expired; | ||
6142 | |||
6143 | *p++ = htonl((u32) RPC_GSS_VERSION); | ||
6144 | *p++ = htonl((u32) ctx->gc_proc); | ||
6145 | @@ -1586,14 +1588,18 @@ gss_marshal(struct rpc_task *task, __be32 *p) | ||
6146 | mic.data = (u8 *)(p + 1); | ||
6147 | maj_stat = gss_get_mic(ctx->gc_gss_ctx, &verf_buf, &mic); | ||
6148 | if (maj_stat == GSS_S_CONTEXT_EXPIRED) { | ||
6149 | - clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags); | ||
6150 | + goto out_expired; | ||
6151 | } else if (maj_stat != 0) { | ||
6152 | - printk("gss_marshal: gss_get_mic FAILED (%d)\n", maj_stat); | ||
6153 | + pr_warn("gss_marshal: gss_get_mic FAILED (%d)\n", maj_stat); | ||
6154 | + task->tk_status = -EIO; | ||
6155 | goto out_put_ctx; | ||
6156 | } | ||
6157 | p = xdr_encode_opaque(p, NULL, mic.len); | ||
6158 | gss_put_ctx(ctx); | ||
6159 | return p; | ||
6160 | +out_expired: | ||
6161 | + clear_bit(RPCAUTH_CRED_UPTODATE, &cred->cr_flags); | ||
6162 | + task->tk_status = -EKEYEXPIRED; | ||
6163 | out_put_ctx: | ||
6164 | gss_put_ctx(ctx); | ||
6165 | return NULL; | ||
6166 | diff --git a/net/sunrpc/clnt.c b/net/sunrpc/clnt.c | ||
6167 | index 24cbddc44c884..1ee04e0ec4bca 100644 | ||
6168 | --- a/net/sunrpc/clnt.c | ||
6169 | +++ b/net/sunrpc/clnt.c | ||
6170 | @@ -1738,14 +1738,10 @@ rpc_xdr_encode(struct rpc_task *task) | ||
6171 | xdr_buf_init(&req->rq_rcv_buf, | ||
6172 | req->rq_rbuffer, | ||
6173 | req->rq_rcvsize); | ||
6174 | - req->rq_bytes_sent = 0; | ||
6175 | |||
6176 | p = rpc_encode_header(task); | ||
6177 | - if (p == NULL) { | ||
6178 | - printk(KERN_INFO "RPC: couldn't encode RPC header, exit EIO\n"); | ||
6179 | - rpc_exit(task, -EIO); | ||
6180 | + if (p == NULL) | ||
6181 | return; | ||
6182 | - } | ||
6183 | |||
6184 | encode = task->tk_msg.rpc_proc->p_encode; | ||
6185 | if (encode == NULL) | ||
6186 | @@ -1770,10 +1766,17 @@ call_encode(struct rpc_task *task) | ||
6187 | /* Did the encode result in an error condition? */ | ||
6188 | if (task->tk_status != 0) { | ||
6189 | /* Was the error nonfatal? */ | ||
6190 | - if (task->tk_status == -EAGAIN || task->tk_status == -ENOMEM) | ||
6191 | + switch (task->tk_status) { | ||
6192 | + case -EAGAIN: | ||
6193 | + case -ENOMEM: | ||
6194 | rpc_delay(task, HZ >> 4); | ||
6195 | - else | ||
6196 | + break; | ||
6197 | + case -EKEYEXPIRED: | ||
6198 | + task->tk_action = call_refresh; | ||
6199 | + break; | ||
6200 | + default: | ||
6201 | rpc_exit(task, task->tk_status); | ||
6202 | + } | ||
6203 | return; | ||
6204 | } | ||
6205 | |||
6206 | @@ -2335,7 +2338,8 @@ rpc_encode_header(struct rpc_task *task) | ||
6207 | *p++ = htonl(clnt->cl_vers); /* program version */ | ||
6208 | *p++ = htonl(task->tk_msg.rpc_proc->p_proc); /* procedure */ | ||
6209 | p = rpcauth_marshcred(task, p); | ||
6210 | - req->rq_slen = xdr_adjust_iovec(&req->rq_svec[0], p); | ||
6211 | + if (p) | ||
6212 | + req->rq_slen = xdr_adjust_iovec(&req->rq_svec[0], p); | ||
6213 | return p; | ||
6214 | } | ||
6215 | |||
6216 | diff --git a/net/sunrpc/xprt.c b/net/sunrpc/xprt.c | ||
6217 | index 943f08be7c387..f1ec2110efebe 100644 | ||
6218 | --- a/net/sunrpc/xprt.c | ||
6219 | +++ b/net/sunrpc/xprt.c | ||
6220 | @@ -1151,6 +1151,7 @@ xprt_request_enqueue_transmit(struct rpc_task *task) | ||
6221 | struct rpc_xprt *xprt = req->rq_xprt; | ||
6222 | |||
6223 | if (xprt_request_need_enqueue_transmit(task, req)) { | ||
6224 | + req->rq_bytes_sent = 0; | ||
6225 | spin_lock(&xprt->queue_lock); | ||
6226 | /* | ||
6227 | * Requests that carry congestion control credits are added | ||
6228 | diff --git a/net/sunrpc/xprtrdma/verbs.c b/net/sunrpc/xprtrdma/verbs.c | ||
6229 | index 919fddec01973..dffedf1df02ce 100644 | ||
6230 | --- a/net/sunrpc/xprtrdma/verbs.c | ||
6231 | +++ b/net/sunrpc/xprtrdma/verbs.c | ||
6232 | @@ -912,17 +912,13 @@ static int rpcrdma_sendctxs_create(struct rpcrdma_xprt *r_xprt) | ||
6233 | for (i = 0; i <= buf->rb_sc_last; i++) { | ||
6234 | sc = rpcrdma_sendctx_create(&r_xprt->rx_ia); | ||
6235 | if (!sc) | ||
6236 | - goto out_destroy; | ||
6237 | + return -ENOMEM; | ||
6238 | |||
6239 | sc->sc_xprt = r_xprt; | ||
6240 | buf->rb_sc_ctxs[i] = sc; | ||
6241 | } | ||
6242 | |||
6243 | return 0; | ||
6244 | - | ||
6245 | -out_destroy: | ||
6246 | - rpcrdma_sendctxs_destroy(buf); | ||
6247 | - return -ENOMEM; | ||
6248 | } | ||
6249 | |||
6250 | /* The sendctx queue is not guaranteed to have a size that is a | ||
6251 | diff --git a/net/xdp/xdp_umem.c b/net/xdp/xdp_umem.c | ||
6252 | index a264cf2accd0f..d4de871e7d4d7 100644 | ||
6253 | --- a/net/xdp/xdp_umem.c | ||
6254 | +++ b/net/xdp/xdp_umem.c | ||
6255 | @@ -41,13 +41,20 @@ void xdp_del_sk_umem(struct xdp_umem *umem, struct xdp_sock *xs) | ||
6256 | * not know if the device has more tx queues than rx, or the opposite. | ||
6257 | * This might also change during run time. | ||
6258 | */ | ||
6259 | -static void xdp_reg_umem_at_qid(struct net_device *dev, struct xdp_umem *umem, | ||
6260 | - u16 queue_id) | ||
6261 | +static int xdp_reg_umem_at_qid(struct net_device *dev, struct xdp_umem *umem, | ||
6262 | + u16 queue_id) | ||
6263 | { | ||
6264 | + if (queue_id >= max_t(unsigned int, | ||
6265 | + dev->real_num_rx_queues, | ||
6266 | + dev->real_num_tx_queues)) | ||
6267 | + return -EINVAL; | ||
6268 | + | ||
6269 | if (queue_id < dev->real_num_rx_queues) | ||
6270 | dev->_rx[queue_id].umem = umem; | ||
6271 | if (queue_id < dev->real_num_tx_queues) | ||
6272 | dev->_tx[queue_id].umem = umem; | ||
6273 | + | ||
6274 | + return 0; | ||
6275 | } | ||
6276 | |||
6277 | struct xdp_umem *xdp_get_umem_from_qid(struct net_device *dev, | ||
6278 | @@ -88,7 +95,10 @@ int xdp_umem_assign_dev(struct xdp_umem *umem, struct net_device *dev, | ||
6279 | goto out_rtnl_unlock; | ||
6280 | } | ||
6281 | |||
6282 | - xdp_reg_umem_at_qid(dev, umem, queue_id); | ||
6283 | + err = xdp_reg_umem_at_qid(dev, umem, queue_id); | ||
6284 | + if (err) | ||
6285 | + goto out_rtnl_unlock; | ||
6286 | + | ||
6287 | umem->dev = dev; | ||
6288 | umem->queue_id = queue_id; | ||
6289 | if (force_copy) | ||
6290 | diff --git a/samples/bpf/Makefile b/samples/bpf/Makefile | ||
6291 | index be0a961450bc2..f5ce993c78e42 100644 | ||
6292 | --- a/samples/bpf/Makefile | ||
6293 | +++ b/samples/bpf/Makefile | ||
6294 | @@ -273,6 +273,7 @@ $(obj)/%.o: $(src)/%.c | ||
6295 | -Wno-gnu-variable-sized-type-not-at-end \ | ||
6296 | -Wno-address-of-packed-member -Wno-tautological-compare \ | ||
6297 | -Wno-unknown-warning-option $(CLANG_ARCH_ARGS) \ | ||
6298 | + -I$(srctree)/samples/bpf/ -include asm_goto_workaround.h \ | ||
6299 | -O2 -emit-llvm -c $< -o -| $(LLC) -march=bpf $(LLC_FLAGS) -filetype=obj -o $@ | ||
6300 | ifeq ($(DWARF2BTF),y) | ||
6301 | $(BTF_PAHOLE) -J $@ | ||
6302 | diff --git a/samples/bpf/asm_goto_workaround.h b/samples/bpf/asm_goto_workaround.h | ||
6303 | new file mode 100644 | ||
6304 | index 0000000000000..5cd7c1d1a5d56 | ||
6305 | --- /dev/null | ||
6306 | +++ b/samples/bpf/asm_goto_workaround.h | ||
6307 | @@ -0,0 +1,16 @@ | ||
6308 | +/* SPDX-License-Identifier: GPL-2.0 */ | ||
6309 | +/* Copyright (c) 2019 Facebook */ | ||
6310 | +#ifndef __ASM_GOTO_WORKAROUND_H | ||
6311 | +#define __ASM_GOTO_WORKAROUND_H | ||
6312 | + | ||
6313 | +/* this will bring in asm_volatile_goto macro definition | ||
6314 | + * if enabled by compiler and config options. | ||
6315 | + */ | ||
6316 | +#include <linux/types.h> | ||
6317 | + | ||
6318 | +#ifdef asm_volatile_goto | ||
6319 | +#undef asm_volatile_goto | ||
6320 | +#define asm_volatile_goto(x...) asm volatile("invalid use of asm_volatile_goto") | ||
6321 | +#endif | ||
6322 | + | ||
6323 | +#endif | ||
6324 | diff --git a/security/keys/key.c b/security/keys/key.c | ||
6325 | index d97c9394b5dd4..249a6da4d2770 100644 | ||
6326 | --- a/security/keys/key.c | ||
6327 | +++ b/security/keys/key.c | ||
6328 | @@ -265,8 +265,8 @@ struct key *key_alloc(struct key_type *type, const char *desc, | ||
6329 | |||
6330 | spin_lock(&user->lock); | ||
6331 | if (!(flags & KEY_ALLOC_QUOTA_OVERRUN)) { | ||
6332 | - if (user->qnkeys + 1 >= maxkeys || | ||
6333 | - user->qnbytes + quotalen >= maxbytes || | ||
6334 | + if (user->qnkeys + 1 > maxkeys || | ||
6335 | + user->qnbytes + quotalen > maxbytes || | ||
6336 | user->qnbytes + quotalen < user->qnbytes) | ||
6337 | goto no_quota; | ||
6338 | } | ||
6339 | diff --git a/security/keys/keyring.c b/security/keys/keyring.c | ||
6340 | index 41bcf57e96f21..99a55145ddcd2 100644 | ||
6341 | --- a/security/keys/keyring.c | ||
6342 | +++ b/security/keys/keyring.c | ||
6343 | @@ -661,9 +661,6 @@ static bool search_nested_keyrings(struct key *keyring, | ||
6344 | BUG_ON((ctx->flags & STATE_CHECKS) == 0 || | ||
6345 | (ctx->flags & STATE_CHECKS) == STATE_CHECKS); | ||
6346 | |||
6347 | - if (ctx->index_key.description) | ||
6348 | - ctx->index_key.desc_len = strlen(ctx->index_key.description); | ||
6349 | - | ||
6350 | /* Check to see if this top-level keyring is what we are looking for | ||
6351 | * and whether it is valid or not. | ||
6352 | */ | ||
6353 | @@ -914,6 +911,7 @@ key_ref_t keyring_search(key_ref_t keyring, | ||
6354 | struct keyring_search_context ctx = { | ||
6355 | .index_key.type = type, | ||
6356 | .index_key.description = description, | ||
6357 | + .index_key.desc_len = strlen(description), | ||
6358 | .cred = current_cred(), | ||
6359 | .match_data.cmp = key_default_cmp, | ||
6360 | .match_data.raw_data = description, | ||
6361 | diff --git a/security/keys/proc.c b/security/keys/proc.c | ||
6362 | index 5af2934965d80..d38be9db2cc07 100644 | ||
6363 | --- a/security/keys/proc.c | ||
6364 | +++ b/security/keys/proc.c | ||
6365 | @@ -166,8 +166,7 @@ static int proc_keys_show(struct seq_file *m, void *v) | ||
6366 | int rc; | ||
6367 | |||
6368 | struct keyring_search_context ctx = { | ||
6369 | - .index_key.type = key->type, | ||
6370 | - .index_key.description = key->description, | ||
6371 | + .index_key = key->index_key, | ||
6372 | .cred = m->file->f_cred, | ||
6373 | .match_data.cmp = lookup_user_key_possessed, | ||
6374 | .match_data.raw_data = key, | ||
6375 | diff --git a/security/keys/request_key.c b/security/keys/request_key.c | ||
6376 | index 114f7408feee6..7385536986497 100644 | ||
6377 | --- a/security/keys/request_key.c | ||
6378 | +++ b/security/keys/request_key.c | ||
6379 | @@ -545,6 +545,7 @@ struct key *request_key_and_link(struct key_type *type, | ||
6380 | struct keyring_search_context ctx = { | ||
6381 | .index_key.type = type, | ||
6382 | .index_key.description = description, | ||
6383 | + .index_key.desc_len = strlen(description), | ||
6384 | .cred = current_cred(), | ||
6385 | .match_data.cmp = key_default_cmp, | ||
6386 | .match_data.raw_data = description, | ||
6387 | diff --git a/security/keys/request_key_auth.c b/security/keys/request_key_auth.c | ||
6388 | index 424e1d90412ea..6797843154f03 100644 | ||
6389 | --- a/security/keys/request_key_auth.c | ||
6390 | +++ b/security/keys/request_key_auth.c | ||
6391 | @@ -246,7 +246,7 @@ struct key *key_get_instantiation_authkey(key_serial_t target_id) | ||
6392 | struct key *authkey; | ||
6393 | key_ref_t authkey_ref; | ||
6394 | |||
6395 | - sprintf(description, "%x", target_id); | ||
6396 | + ctx.index_key.desc_len = sprintf(description, "%x", target_id); | ||
6397 | |||
6398 | authkey_ref = search_process_keyrings(&ctx); | ||
6399 | |||
6400 | diff --git a/sound/pci/hda/patch_realtek.c b/sound/pci/hda/patch_realtek.c | ||
6401 | index c1042df5e346e..1bddfa7dc2169 100644 | ||
6402 | --- a/sound/pci/hda/patch_realtek.c | ||
6403 | +++ b/sound/pci/hda/patch_realtek.c | ||
6404 | @@ -1855,6 +1855,8 @@ enum { | ||
6405 | ALC887_FIXUP_BASS_CHMAP, | ||
6406 | ALC1220_FIXUP_GB_DUAL_CODECS, | ||
6407 | ALC1220_FIXUP_CLEVO_P950, | ||
6408 | + ALC1220_FIXUP_SYSTEM76_ORYP5, | ||
6409 | + ALC1220_FIXUP_SYSTEM76_ORYP5_PINS, | ||
6410 | }; | ||
6411 | |||
6412 | static void alc889_fixup_coef(struct hda_codec *codec, | ||
6413 | @@ -2056,6 +2058,17 @@ static void alc1220_fixup_clevo_p950(struct hda_codec *codec, | ||
6414 | snd_hda_override_conn_list(codec, 0x1b, 1, conn1); | ||
6415 | } | ||
6416 | |||
6417 | +static void alc_fixup_headset_mode_no_hp_mic(struct hda_codec *codec, | ||
6418 | + const struct hda_fixup *fix, int action); | ||
6419 | + | ||
6420 | +static void alc1220_fixup_system76_oryp5(struct hda_codec *codec, | ||
6421 | + const struct hda_fixup *fix, | ||
6422 | + int action) | ||
6423 | +{ | ||
6424 | + alc1220_fixup_clevo_p950(codec, fix, action); | ||
6425 | + alc_fixup_headset_mode_no_hp_mic(codec, fix, action); | ||
6426 | +} | ||
6427 | + | ||
6428 | static const struct hda_fixup alc882_fixups[] = { | ||
6429 | [ALC882_FIXUP_ABIT_AW9D_MAX] = { | ||
6430 | .type = HDA_FIXUP_PINS, | ||
6431 | @@ -2300,6 +2313,19 @@ static const struct hda_fixup alc882_fixups[] = { | ||
6432 | .type = HDA_FIXUP_FUNC, | ||
6433 | .v.func = alc1220_fixup_clevo_p950, | ||
6434 | }, | ||
6435 | + [ALC1220_FIXUP_SYSTEM76_ORYP5] = { | ||
6436 | + .type = HDA_FIXUP_FUNC, | ||
6437 | + .v.func = alc1220_fixup_system76_oryp5, | ||
6438 | + }, | ||
6439 | + [ALC1220_FIXUP_SYSTEM76_ORYP5_PINS] = { | ||
6440 | + .type = HDA_FIXUP_PINS, | ||
6441 | + .v.pins = (const struct hda_pintbl[]) { | ||
6442 | + { 0x19, 0x01a1913c }, /* use as headset mic, without its own jack detect */ | ||
6443 | + {} | ||
6444 | + }, | ||
6445 | + .chained = true, | ||
6446 | + .chain_id = ALC1220_FIXUP_SYSTEM76_ORYP5, | ||
6447 | + }, | ||
6448 | }; | ||
6449 | |||
6450 | static const struct snd_pci_quirk alc882_fixup_tbl[] = { | ||
6451 | @@ -2376,6 +2402,8 @@ static const struct snd_pci_quirk alc882_fixup_tbl[] = { | ||
6452 | SND_PCI_QUIRK(0x1558, 0x9501, "Clevo P950HR", ALC1220_FIXUP_CLEVO_P950), | ||
6453 | SND_PCI_QUIRK(0x1558, 0x95e1, "Clevo P95xER", ALC1220_FIXUP_CLEVO_P950), | ||
6454 | SND_PCI_QUIRK(0x1558, 0x95e2, "Clevo P950ER", ALC1220_FIXUP_CLEVO_P950), | ||
6455 | + SND_PCI_QUIRK(0x1558, 0x96e1, "System76 Oryx Pro (oryp5)", ALC1220_FIXUP_SYSTEM76_ORYP5_PINS), | ||
6456 | + SND_PCI_QUIRK(0x1558, 0x97e1, "System76 Oryx Pro (oryp5)", ALC1220_FIXUP_SYSTEM76_ORYP5_PINS), | ||
6457 | SND_PCI_QUIRK_VENDOR(0x1558, "Clevo laptop", ALC882_FIXUP_EAPD), | ||
6458 | SND_PCI_QUIRK(0x161f, 0x2054, "Medion laptop", ALC883_FIXUP_EAPD), | ||
6459 | SND_PCI_QUIRK(0x17aa, 0x3a0d, "Lenovo Y530", ALC882_FIXUP_LENOVO_Y530), | ||
6460 | @@ -5573,6 +5601,7 @@ enum { | ||
6461 | ALC294_FIXUP_ASUS_HEADSET_MIC, | ||
6462 | ALC294_FIXUP_ASUS_SPK, | ||
6463 | ALC293_FIXUP_SYSTEM76_MIC_NO_PRESENCE, | ||
6464 | + ALC285_FIXUP_LENOVO_PC_BEEP_IN_NOISE, | ||
6465 | }; | ||
6466 | |||
6467 | static const struct hda_fixup alc269_fixups[] = { | ||
6468 | @@ -6506,6 +6535,17 @@ static const struct hda_fixup alc269_fixups[] = { | ||
6469 | .chained = true, | ||
6470 | .chain_id = ALC269_FIXUP_HEADSET_MODE_NO_HP_MIC | ||
6471 | }, | ||
6472 | + [ALC285_FIXUP_LENOVO_PC_BEEP_IN_NOISE] = { | ||
6473 | + .type = HDA_FIXUP_VERBS, | ||
6474 | + .v.verbs = (const struct hda_verb[]) { | ||
6475 | + /* Disable PCBEEP-IN passthrough */ | ||
6476 | + { 0x20, AC_VERB_SET_COEF_INDEX, 0x36 }, | ||
6477 | + { 0x20, AC_VERB_SET_PROC_COEF, 0x57d7 }, | ||
6478 | + { } | ||
6479 | + }, | ||
6480 | + .chained = true, | ||
6481 | + .chain_id = ALC285_FIXUP_LENOVO_HEADPHONE_NOISE | ||
6482 | + }, | ||
6483 | }; | ||
6484 | |||
6485 | static const struct snd_pci_quirk alc269_fixup_tbl[] = { | ||
6486 | @@ -7187,7 +7227,7 @@ static const struct snd_hda_pin_quirk alc269_pin_fixup_tbl[] = { | ||
6487 | {0x12, 0x90a60130}, | ||
6488 | {0x19, 0x03a11020}, | ||
6489 | {0x21, 0x0321101f}), | ||
6490 | - SND_HDA_PIN_QUIRK(0x10ec0285, 0x17aa, "Lenovo", ALC285_FIXUP_LENOVO_HEADPHONE_NOISE, | ||
6491 | + SND_HDA_PIN_QUIRK(0x10ec0285, 0x17aa, "Lenovo", ALC285_FIXUP_LENOVO_PC_BEEP_IN_NOISE, | ||
6492 | {0x12, 0x90a60130}, | ||
6493 | {0x14, 0x90170110}, | ||
6494 | {0x19, 0x04a11040}, | ||
6495 | diff --git a/sound/soc/soc-core.c b/sound/soc/soc-core.c | ||
6496 | index b29d0f65611eb..2d49492d60692 100644 | ||
6497 | --- a/sound/soc/soc-core.c | ||
6498 | +++ b/sound/soc/soc-core.c | ||
6499 | @@ -1034,17 +1034,18 @@ static int snd_soc_init_platform(struct snd_soc_card *card, | ||
6500 | * this function should be removed in the future | ||
6501 | */ | ||
6502 | /* convert Legacy platform link */ | ||
6503 | - if (!platform) { | ||
6504 | + if (!platform || dai_link->legacy_platform) { | ||
6505 | platform = devm_kzalloc(card->dev, | ||
6506 | sizeof(struct snd_soc_dai_link_component), | ||
6507 | GFP_KERNEL); | ||
6508 | if (!platform) | ||
6509 | return -ENOMEM; | ||
6510 | |||
6511 | - dai_link->platform = platform; | ||
6512 | - platform->name = dai_link->platform_name; | ||
6513 | - platform->of_node = dai_link->platform_of_node; | ||
6514 | - platform->dai_name = NULL; | ||
6515 | + dai_link->platform = platform; | ||
6516 | + dai_link->legacy_platform = 1; | ||
6517 | + platform->name = dai_link->platform_name; | ||
6518 | + platform->of_node = dai_link->platform_of_node; | ||
6519 | + platform->dai_name = NULL; | ||
6520 | } | ||
6521 | |||
6522 | /* if there's no platform we match on the empty platform */ | ||
6523 | diff --git a/tools/include/uapi/linux/pkt_sched.h b/tools/include/uapi/linux/pkt_sched.h | ||
6524 | new file mode 100644 | ||
6525 | index 0000000000000..0d18b1d1fbbc8 | ||
6526 | --- /dev/null | ||
6527 | +++ b/tools/include/uapi/linux/pkt_sched.h | ||
6528 | @@ -0,0 +1,1163 @@ | ||
6529 | +/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */ | ||
6530 | +#ifndef __LINUX_PKT_SCHED_H | ||
6531 | +#define __LINUX_PKT_SCHED_H | ||
6532 | + | ||
6533 | +#include <linux/types.h> | ||
6534 | + | ||
6535 | +/* Logical priority bands not depending on specific packet scheduler. | ||
6536 | + Every scheduler will map them to real traffic classes, if it has | ||
6537 | + no more precise mechanism to classify packets. | ||
6538 | + | ||
6539 | + These numbers have no special meaning, though their coincidence | ||
6540 | + with obsolete IPv6 values is not occasional :-). New IPv6 drafts | ||
6541 | + preferred full anarchy inspired by diffserv group. | ||
6542 | + | ||
6543 | + Note: TC_PRIO_BESTEFFORT does not mean that it is the most unhappy | ||
6544 | + class, actually, as rule it will be handled with more care than | ||
6545 | + filler or even bulk. | ||
6546 | + */ | ||
6547 | + | ||
6548 | +#define TC_PRIO_BESTEFFORT 0 | ||
6549 | +#define TC_PRIO_FILLER 1 | ||
6550 | +#define TC_PRIO_BULK 2 | ||
6551 | +#define TC_PRIO_INTERACTIVE_BULK 4 | ||
6552 | +#define TC_PRIO_INTERACTIVE 6 | ||
6553 | +#define TC_PRIO_CONTROL 7 | ||
6554 | + | ||
6555 | +#define TC_PRIO_MAX 15 | ||
6556 | + | ||
6557 | +/* Generic queue statistics, available for all the elements. | ||
6558 | + Particular schedulers may have also their private records. | ||
6559 | + */ | ||
6560 | + | ||
6561 | +struct tc_stats { | ||
6562 | + __u64 bytes; /* Number of enqueued bytes */ | ||
6563 | + __u32 packets; /* Number of enqueued packets */ | ||
6564 | + __u32 drops; /* Packets dropped because of lack of resources */ | ||
6565 | + __u32 overlimits; /* Number of throttle events when this | ||
6566 | + * flow goes out of allocated bandwidth */ | ||
6567 | + __u32 bps; /* Current flow byte rate */ | ||
6568 | + __u32 pps; /* Current flow packet rate */ | ||
6569 | + __u32 qlen; | ||
6570 | + __u32 backlog; | ||
6571 | +}; | ||
6572 | + | ||
6573 | +struct tc_estimator { | ||
6574 | + signed char interval; | ||
6575 | + unsigned char ewma_log; | ||
6576 | +}; | ||
6577 | + | ||
6578 | +/* "Handles" | ||
6579 | + --------- | ||
6580 | + | ||
6581 | + All the traffic control objects have 32bit identifiers, or "handles". | ||
6582 | + | ||
6583 | + They can be considered as opaque numbers from user API viewpoint, | ||
6584 | + but actually they always consist of two fields: major and | ||
6585 | + minor numbers, which are interpreted by kernel specially, | ||
6586 | + that may be used by applications, though not recommended. | ||
6587 | + | ||
6588 | + F.e. qdisc handles always have minor number equal to zero, | ||
6589 | + classes (or flows) have major equal to parent qdisc major, and | ||
6590 | + minor uniquely identifying class inside qdisc. | ||
6591 | + | ||
6592 | + Macros to manipulate handles: | ||
6593 | + */ | ||
6594 | + | ||
6595 | +#define TC_H_MAJ_MASK (0xFFFF0000U) | ||
6596 | +#define TC_H_MIN_MASK (0x0000FFFFU) | ||
6597 | +#define TC_H_MAJ(h) ((h)&TC_H_MAJ_MASK) | ||
6598 | +#define TC_H_MIN(h) ((h)&TC_H_MIN_MASK) | ||
6599 | +#define TC_H_MAKE(maj,min) (((maj)&TC_H_MAJ_MASK)|((min)&TC_H_MIN_MASK)) | ||
6600 | + | ||
6601 | +#define TC_H_UNSPEC (0U) | ||
6602 | +#define TC_H_ROOT (0xFFFFFFFFU) | ||
6603 | +#define TC_H_INGRESS (0xFFFFFFF1U) | ||
6604 | +#define TC_H_CLSACT TC_H_INGRESS | ||
6605 | + | ||
6606 | +#define TC_H_MIN_PRIORITY 0xFFE0U | ||
6607 | +#define TC_H_MIN_INGRESS 0xFFF2U | ||
6608 | +#define TC_H_MIN_EGRESS 0xFFF3U | ||
6609 | + | ||
6610 | +/* Need to corrospond to iproute2 tc/tc_core.h "enum link_layer" */ | ||
6611 | +enum tc_link_layer { | ||
6612 | + TC_LINKLAYER_UNAWARE, /* Indicate unaware old iproute2 util */ | ||
6613 | + TC_LINKLAYER_ETHERNET, | ||
6614 | + TC_LINKLAYER_ATM, | ||
6615 | +}; | ||
6616 | +#define TC_LINKLAYER_MASK 0x0F /* limit use to lower 4 bits */ | ||
6617 | + | ||
6618 | +struct tc_ratespec { | ||
6619 | + unsigned char cell_log; | ||
6620 | + __u8 linklayer; /* lower 4 bits */ | ||
6621 | + unsigned short overhead; | ||
6622 | + short cell_align; | ||
6623 | + unsigned short mpu; | ||
6624 | + __u32 rate; | ||
6625 | +}; | ||
6626 | + | ||
6627 | +#define TC_RTAB_SIZE 1024 | ||
6628 | + | ||
6629 | +struct tc_sizespec { | ||
6630 | + unsigned char cell_log; | ||
6631 | + unsigned char size_log; | ||
6632 | + short cell_align; | ||
6633 | + int overhead; | ||
6634 | + unsigned int linklayer; | ||
6635 | + unsigned int mpu; | ||
6636 | + unsigned int mtu; | ||
6637 | + unsigned int tsize; | ||
6638 | +}; | ||
6639 | + | ||
6640 | +enum { | ||
6641 | + TCA_STAB_UNSPEC, | ||
6642 | + TCA_STAB_BASE, | ||
6643 | + TCA_STAB_DATA, | ||
6644 | + __TCA_STAB_MAX | ||
6645 | +}; | ||
6646 | + | ||
6647 | +#define TCA_STAB_MAX (__TCA_STAB_MAX - 1) | ||
6648 | + | ||
6649 | +/* FIFO section */ | ||
6650 | + | ||
6651 | +struct tc_fifo_qopt { | ||
6652 | + __u32 limit; /* Queue length: bytes for bfifo, packets for pfifo */ | ||
6653 | +}; | ||
6654 | + | ||
6655 | +/* SKBPRIO section */ | ||
6656 | + | ||
6657 | +/* | ||
6658 | + * Priorities go from zero to (SKBPRIO_MAX_PRIORITY - 1). | ||
6659 | + * SKBPRIO_MAX_PRIORITY should be at least 64 in order for skbprio to be able | ||
6660 | + * to map one to one the DS field of IPV4 and IPV6 headers. | ||
6661 | + * Memory allocation grows linearly with SKBPRIO_MAX_PRIORITY. | ||
6662 | + */ | ||
6663 | + | ||
6664 | +#define SKBPRIO_MAX_PRIORITY 64 | ||
6665 | + | ||
6666 | +struct tc_skbprio_qopt { | ||
6667 | + __u32 limit; /* Queue length in packets. */ | ||
6668 | +}; | ||
6669 | + | ||
6670 | +/* PRIO section */ | ||
6671 | + | ||
6672 | +#define TCQ_PRIO_BANDS 16 | ||
6673 | +#define TCQ_MIN_PRIO_BANDS 2 | ||
6674 | + | ||
6675 | +struct tc_prio_qopt { | ||
6676 | + int bands; /* Number of bands */ | ||
6677 | + __u8 priomap[TC_PRIO_MAX+1]; /* Map: logical priority -> PRIO band */ | ||
6678 | +}; | ||
6679 | + | ||
6680 | +/* MULTIQ section */ | ||
6681 | + | ||
6682 | +struct tc_multiq_qopt { | ||
6683 | + __u16 bands; /* Number of bands */ | ||
6684 | + __u16 max_bands; /* Maximum number of queues */ | ||
6685 | +}; | ||
6686 | + | ||
6687 | +/* PLUG section */ | ||
6688 | + | ||
6689 | +#define TCQ_PLUG_BUFFER 0 | ||
6690 | +#define TCQ_PLUG_RELEASE_ONE 1 | ||
6691 | +#define TCQ_PLUG_RELEASE_INDEFINITE 2 | ||
6692 | +#define TCQ_PLUG_LIMIT 3 | ||
6693 | + | ||
6694 | +struct tc_plug_qopt { | ||
6695 | + /* TCQ_PLUG_BUFFER: Inset a plug into the queue and | ||
6696 | + * buffer any incoming packets | ||
6697 | + * TCQ_PLUG_RELEASE_ONE: Dequeue packets from queue head | ||
6698 | + * to beginning of the next plug. | ||
6699 | + * TCQ_PLUG_RELEASE_INDEFINITE: Dequeue all packets from queue. | ||
6700 | + * Stop buffering packets until the next TCQ_PLUG_BUFFER | ||
6701 | + * command is received (just act as a pass-thru queue). | ||
6702 | + * TCQ_PLUG_LIMIT: Increase/decrease queue size | ||
6703 | + */ | ||
6704 | + int action; | ||
6705 | + __u32 limit; | ||
6706 | +}; | ||
6707 | + | ||
6708 | +/* TBF section */ | ||
6709 | + | ||
6710 | +struct tc_tbf_qopt { | ||
6711 | + struct tc_ratespec rate; | ||
6712 | + struct tc_ratespec peakrate; | ||
6713 | + __u32 limit; | ||
6714 | + __u32 buffer; | ||
6715 | + __u32 mtu; | ||
6716 | +}; | ||
6717 | + | ||
6718 | +enum { | ||
6719 | + TCA_TBF_UNSPEC, | ||
6720 | + TCA_TBF_PARMS, | ||
6721 | + TCA_TBF_RTAB, | ||
6722 | + TCA_TBF_PTAB, | ||
6723 | + TCA_TBF_RATE64, | ||
6724 | + TCA_TBF_PRATE64, | ||
6725 | + TCA_TBF_BURST, | ||
6726 | + TCA_TBF_PBURST, | ||
6727 | + TCA_TBF_PAD, | ||
6728 | + __TCA_TBF_MAX, | ||
6729 | +}; | ||
6730 | + | ||
6731 | +#define TCA_TBF_MAX (__TCA_TBF_MAX - 1) | ||
6732 | + | ||
6733 | + | ||
6734 | +/* TEQL section */ | ||
6735 | + | ||
6736 | +/* TEQL does not require any parameters */ | ||
6737 | + | ||
6738 | +/* SFQ section */ | ||
6739 | + | ||
6740 | +struct tc_sfq_qopt { | ||
6741 | + unsigned quantum; /* Bytes per round allocated to flow */ | ||
6742 | + int perturb_period; /* Period of hash perturbation */ | ||
6743 | + __u32 limit; /* Maximal packets in queue */ | ||
6744 | + unsigned divisor; /* Hash divisor */ | ||
6745 | + unsigned flows; /* Maximal number of flows */ | ||
6746 | +}; | ||
6747 | + | ||
6748 | +struct tc_sfqred_stats { | ||
6749 | + __u32 prob_drop; /* Early drops, below max threshold */ | ||
6750 | + __u32 forced_drop; /* Early drops, after max threshold */ | ||
6751 | + __u32 prob_mark; /* Marked packets, below max threshold */ | ||
6752 | + __u32 forced_mark; /* Marked packets, after max threshold */ | ||
6753 | + __u32 prob_mark_head; /* Marked packets, below max threshold */ | ||
6754 | + __u32 forced_mark_head;/* Marked packets, after max threshold */ | ||
6755 | +}; | ||
6756 | + | ||
6757 | +struct tc_sfq_qopt_v1 { | ||
6758 | + struct tc_sfq_qopt v0; | ||
6759 | + unsigned int depth; /* max number of packets per flow */ | ||
6760 | + unsigned int headdrop; | ||
6761 | +/* SFQRED parameters */ | ||
6762 | + __u32 limit; /* HARD maximal flow queue length (bytes) */ | ||
6763 | + __u32 qth_min; /* Min average length threshold (bytes) */ | ||
6764 | + __u32 qth_max; /* Max average length threshold (bytes) */ | ||
6765 | + unsigned char Wlog; /* log(W) */ | ||
6766 | + unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ | ||
6767 | + unsigned char Scell_log; /* cell size for idle damping */ | ||
6768 | + unsigned char flags; | ||
6769 | + __u32 max_P; /* probability, high resolution */ | ||
6770 | +/* SFQRED stats */ | ||
6771 | + struct tc_sfqred_stats stats; | ||
6772 | +}; | ||
6773 | + | ||
6774 | + | ||
6775 | +struct tc_sfq_xstats { | ||
6776 | + __s32 allot; | ||
6777 | +}; | ||
6778 | + | ||
6779 | +/* RED section */ | ||
6780 | + | ||
6781 | +enum { | ||
6782 | + TCA_RED_UNSPEC, | ||
6783 | + TCA_RED_PARMS, | ||
6784 | + TCA_RED_STAB, | ||
6785 | + TCA_RED_MAX_P, | ||
6786 | + __TCA_RED_MAX, | ||
6787 | +}; | ||
6788 | + | ||
6789 | +#define TCA_RED_MAX (__TCA_RED_MAX - 1) | ||
6790 | + | ||
6791 | +struct tc_red_qopt { | ||
6792 | + __u32 limit; /* HARD maximal queue length (bytes) */ | ||
6793 | + __u32 qth_min; /* Min average length threshold (bytes) */ | ||
6794 | + __u32 qth_max; /* Max average length threshold (bytes) */ | ||
6795 | + unsigned char Wlog; /* log(W) */ | ||
6796 | + unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ | ||
6797 | + unsigned char Scell_log; /* cell size for idle damping */ | ||
6798 | + unsigned char flags; | ||
6799 | +#define TC_RED_ECN 1 | ||
6800 | +#define TC_RED_HARDDROP 2 | ||
6801 | +#define TC_RED_ADAPTATIVE 4 | ||
6802 | +}; | ||
6803 | + | ||
6804 | +struct tc_red_xstats { | ||
6805 | + __u32 early; /* Early drops */ | ||
6806 | + __u32 pdrop; /* Drops due to queue limits */ | ||
6807 | + __u32 other; /* Drops due to drop() calls */ | ||
6808 | + __u32 marked; /* Marked packets */ | ||
6809 | +}; | ||
6810 | + | ||
6811 | +/* GRED section */ | ||
6812 | + | ||
6813 | +#define MAX_DPs 16 | ||
6814 | + | ||
6815 | +enum { | ||
6816 | + TCA_GRED_UNSPEC, | ||
6817 | + TCA_GRED_PARMS, | ||
6818 | + TCA_GRED_STAB, | ||
6819 | + TCA_GRED_DPS, | ||
6820 | + TCA_GRED_MAX_P, | ||
6821 | + TCA_GRED_LIMIT, | ||
6822 | + TCA_GRED_VQ_LIST, /* nested TCA_GRED_VQ_ENTRY */ | ||
6823 | + __TCA_GRED_MAX, | ||
6824 | +}; | ||
6825 | + | ||
6826 | +#define TCA_GRED_MAX (__TCA_GRED_MAX - 1) | ||
6827 | + | ||
6828 | +enum { | ||
6829 | + TCA_GRED_VQ_ENTRY_UNSPEC, | ||
6830 | + TCA_GRED_VQ_ENTRY, /* nested TCA_GRED_VQ_* */ | ||
6831 | + __TCA_GRED_VQ_ENTRY_MAX, | ||
6832 | +}; | ||
6833 | +#define TCA_GRED_VQ_ENTRY_MAX (__TCA_GRED_VQ_ENTRY_MAX - 1) | ||
6834 | + | ||
6835 | +enum { | ||
6836 | + TCA_GRED_VQ_UNSPEC, | ||
6837 | + TCA_GRED_VQ_PAD, | ||
6838 | + TCA_GRED_VQ_DP, /* u32 */ | ||
6839 | + TCA_GRED_VQ_STAT_BYTES, /* u64 */ | ||
6840 | + TCA_GRED_VQ_STAT_PACKETS, /* u32 */ | ||
6841 | + TCA_GRED_VQ_STAT_BACKLOG, /* u32 */ | ||
6842 | + TCA_GRED_VQ_STAT_PROB_DROP, /* u32 */ | ||
6843 | + TCA_GRED_VQ_STAT_PROB_MARK, /* u32 */ | ||
6844 | + TCA_GRED_VQ_STAT_FORCED_DROP, /* u32 */ | ||
6845 | + TCA_GRED_VQ_STAT_FORCED_MARK, /* u32 */ | ||
6846 | + TCA_GRED_VQ_STAT_PDROP, /* u32 */ | ||
6847 | + TCA_GRED_VQ_STAT_OTHER, /* u32 */ | ||
6848 | + TCA_GRED_VQ_FLAGS, /* u32 */ | ||
6849 | + __TCA_GRED_VQ_MAX | ||
6850 | +}; | ||
6851 | + | ||
6852 | +#define TCA_GRED_VQ_MAX (__TCA_GRED_VQ_MAX - 1) | ||
6853 | + | ||
6854 | +struct tc_gred_qopt { | ||
6855 | + __u32 limit; /* HARD maximal queue length (bytes) */ | ||
6856 | + __u32 qth_min; /* Min average length threshold (bytes) */ | ||
6857 | + __u32 qth_max; /* Max average length threshold (bytes) */ | ||
6858 | + __u32 DP; /* up to 2^32 DPs */ | ||
6859 | + __u32 backlog; | ||
6860 | + __u32 qave; | ||
6861 | + __u32 forced; | ||
6862 | + __u32 early; | ||
6863 | + __u32 other; | ||
6864 | + __u32 pdrop; | ||
6865 | + __u8 Wlog; /* log(W) */ | ||
6866 | + __u8 Plog; /* log(P_max/(qth_max-qth_min)) */ | ||
6867 | + __u8 Scell_log; /* cell size for idle damping */ | ||
6868 | + __u8 prio; /* prio of this VQ */ | ||
6869 | + __u32 packets; | ||
6870 | + __u32 bytesin; | ||
6871 | +}; | ||
6872 | + | ||
6873 | +/* gred setup */ | ||
6874 | +struct tc_gred_sopt { | ||
6875 | + __u32 DPs; | ||
6876 | + __u32 def_DP; | ||
6877 | + __u8 grio; | ||
6878 | + __u8 flags; | ||
6879 | + __u16 pad1; | ||
6880 | +}; | ||
6881 | + | ||
6882 | +/* CHOKe section */ | ||
6883 | + | ||
6884 | +enum { | ||
6885 | + TCA_CHOKE_UNSPEC, | ||
6886 | + TCA_CHOKE_PARMS, | ||
6887 | + TCA_CHOKE_STAB, | ||
6888 | + TCA_CHOKE_MAX_P, | ||
6889 | + __TCA_CHOKE_MAX, | ||
6890 | +}; | ||
6891 | + | ||
6892 | +#define TCA_CHOKE_MAX (__TCA_CHOKE_MAX - 1) | ||
6893 | + | ||
6894 | +struct tc_choke_qopt { | ||
6895 | + __u32 limit; /* Hard queue length (packets) */ | ||
6896 | + __u32 qth_min; /* Min average threshold (packets) */ | ||
6897 | + __u32 qth_max; /* Max average threshold (packets) */ | ||
6898 | + unsigned char Wlog; /* log(W) */ | ||
6899 | + unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ | ||
6900 | + unsigned char Scell_log; /* cell size for idle damping */ | ||
6901 | + unsigned char flags; /* see RED flags */ | ||
6902 | +}; | ||
6903 | + | ||
6904 | +struct tc_choke_xstats { | ||
6905 | + __u32 early; /* Early drops */ | ||
6906 | + __u32 pdrop; /* Drops due to queue limits */ | ||
6907 | + __u32 other; /* Drops due to drop() calls */ | ||
6908 | + __u32 marked; /* Marked packets */ | ||
6909 | + __u32 matched; /* Drops due to flow match */ | ||
6910 | +}; | ||
6911 | + | ||
6912 | +/* HTB section */ | ||
6913 | +#define TC_HTB_NUMPRIO 8 | ||
6914 | +#define TC_HTB_MAXDEPTH 8 | ||
6915 | +#define TC_HTB_PROTOVER 3 /* the same as HTB and TC's major */ | ||
6916 | + | ||
6917 | +struct tc_htb_opt { | ||
6918 | + struct tc_ratespec rate; | ||
6919 | + struct tc_ratespec ceil; | ||
6920 | + __u32 buffer; | ||
6921 | + __u32 cbuffer; | ||
6922 | + __u32 quantum; | ||
6923 | + __u32 level; /* out only */ | ||
6924 | + __u32 prio; | ||
6925 | +}; | ||
6926 | +struct tc_htb_glob { | ||
6927 | + __u32 version; /* to match HTB/TC */ | ||
6928 | + __u32 rate2quantum; /* bps->quantum divisor */ | ||
6929 | + __u32 defcls; /* default class number */ | ||
6930 | + __u32 debug; /* debug flags */ | ||
6931 | + | ||
6932 | + /* stats */ | ||
6933 | + __u32 direct_pkts; /* count of non shaped packets */ | ||
6934 | +}; | ||
6935 | +enum { | ||
6936 | + TCA_HTB_UNSPEC, | ||
6937 | + TCA_HTB_PARMS, | ||
6938 | + TCA_HTB_INIT, | ||
6939 | + TCA_HTB_CTAB, | ||
6940 | + TCA_HTB_RTAB, | ||
6941 | + TCA_HTB_DIRECT_QLEN, | ||
6942 | + TCA_HTB_RATE64, | ||
6943 | + TCA_HTB_CEIL64, | ||
6944 | + TCA_HTB_PAD, | ||
6945 | + __TCA_HTB_MAX, | ||
6946 | +}; | ||
6947 | + | ||
6948 | +#define TCA_HTB_MAX (__TCA_HTB_MAX - 1) | ||
6949 | + | ||
6950 | +struct tc_htb_xstats { | ||
6951 | + __u32 lends; | ||
6952 | + __u32 borrows; | ||
6953 | + __u32 giants; /* unused since 'Make HTB scheduler work with TSO.' */ | ||
6954 | + __s32 tokens; | ||
6955 | + __s32 ctokens; | ||
6956 | +}; | ||
6957 | + | ||
6958 | +/* HFSC section */ | ||
6959 | + | ||
6960 | +struct tc_hfsc_qopt { | ||
6961 | + __u16 defcls; /* default class */ | ||
6962 | +}; | ||
6963 | + | ||
6964 | +struct tc_service_curve { | ||
6965 | + __u32 m1; /* slope of the first segment in bps */ | ||
6966 | + __u32 d; /* x-projection of the first segment in us */ | ||
6967 | + __u32 m2; /* slope of the second segment in bps */ | ||
6968 | +}; | ||
6969 | + | ||
6970 | +struct tc_hfsc_stats { | ||
6971 | + __u64 work; /* total work done */ | ||
6972 | + __u64 rtwork; /* work done by real-time criteria */ | ||
6973 | + __u32 period; /* current period */ | ||
6974 | + __u32 level; /* class level in hierarchy */ | ||
6975 | +}; | ||
6976 | + | ||
6977 | +enum { | ||
6978 | + TCA_HFSC_UNSPEC, | ||
6979 | + TCA_HFSC_RSC, | ||
6980 | + TCA_HFSC_FSC, | ||
6981 | + TCA_HFSC_USC, | ||
6982 | + __TCA_HFSC_MAX, | ||
6983 | +}; | ||
6984 | + | ||
6985 | +#define TCA_HFSC_MAX (__TCA_HFSC_MAX - 1) | ||
6986 | + | ||
6987 | + | ||
6988 | +/* CBQ section */ | ||
6989 | + | ||
6990 | +#define TC_CBQ_MAXPRIO 8 | ||
6991 | +#define TC_CBQ_MAXLEVEL 8 | ||
6992 | +#define TC_CBQ_DEF_EWMA 5 | ||
6993 | + | ||
6994 | +struct tc_cbq_lssopt { | ||
6995 | + unsigned char change; | ||
6996 | + unsigned char flags; | ||
6997 | +#define TCF_CBQ_LSS_BOUNDED 1 | ||
6998 | +#define TCF_CBQ_LSS_ISOLATED 2 | ||
6999 | + unsigned char ewma_log; | ||
7000 | + unsigned char level; | ||
7001 | +#define TCF_CBQ_LSS_FLAGS 1 | ||
7002 | +#define TCF_CBQ_LSS_EWMA 2 | ||
7003 | +#define TCF_CBQ_LSS_MAXIDLE 4 | ||
7004 | +#define TCF_CBQ_LSS_MINIDLE 8 | ||
7005 | +#define TCF_CBQ_LSS_OFFTIME 0x10 | ||
7006 | +#define TCF_CBQ_LSS_AVPKT 0x20 | ||
7007 | + __u32 maxidle; | ||
7008 | + __u32 minidle; | ||
7009 | + __u32 offtime; | ||
7010 | + __u32 avpkt; | ||
7011 | +}; | ||
7012 | + | ||
7013 | +struct tc_cbq_wrropt { | ||
7014 | + unsigned char flags; | ||
7015 | + unsigned char priority; | ||
7016 | + unsigned char cpriority; | ||
7017 | + unsigned char __reserved; | ||
7018 | + __u32 allot; | ||
7019 | + __u32 weight; | ||
7020 | +}; | ||
7021 | + | ||
7022 | +struct tc_cbq_ovl { | ||
7023 | + unsigned char strategy; | ||
7024 | +#define TC_CBQ_OVL_CLASSIC 0 | ||
7025 | +#define TC_CBQ_OVL_DELAY 1 | ||
7026 | +#define TC_CBQ_OVL_LOWPRIO 2 | ||
7027 | +#define TC_CBQ_OVL_DROP 3 | ||
7028 | +#define TC_CBQ_OVL_RCLASSIC 4 | ||
7029 | + unsigned char priority2; | ||
7030 | + __u16 pad; | ||
7031 | + __u32 penalty; | ||
7032 | +}; | ||
7033 | + | ||
7034 | +struct tc_cbq_police { | ||
7035 | + unsigned char police; | ||
7036 | + unsigned char __res1; | ||
7037 | + unsigned short __res2; | ||
7038 | +}; | ||
7039 | + | ||
7040 | +struct tc_cbq_fopt { | ||
7041 | + __u32 split; | ||
7042 | + __u32 defmap; | ||
7043 | + __u32 defchange; | ||
7044 | +}; | ||
7045 | + | ||
7046 | +struct tc_cbq_xstats { | ||
7047 | + __u32 borrows; | ||
7048 | + __u32 overactions; | ||
7049 | + __s32 avgidle; | ||
7050 | + __s32 undertime; | ||
7051 | +}; | ||
7052 | + | ||
7053 | +enum { | ||
7054 | + TCA_CBQ_UNSPEC, | ||
7055 | + TCA_CBQ_LSSOPT, | ||
7056 | + TCA_CBQ_WRROPT, | ||
7057 | + TCA_CBQ_FOPT, | ||
7058 | + TCA_CBQ_OVL_STRATEGY, | ||
7059 | + TCA_CBQ_RATE, | ||
7060 | + TCA_CBQ_RTAB, | ||
7061 | + TCA_CBQ_POLICE, | ||
7062 | + __TCA_CBQ_MAX, | ||
7063 | +}; | ||
7064 | + | ||
7065 | +#define TCA_CBQ_MAX (__TCA_CBQ_MAX - 1) | ||
7066 | + | ||
7067 | +/* dsmark section */ | ||
7068 | + | ||
7069 | +enum { | ||
7070 | + TCA_DSMARK_UNSPEC, | ||
7071 | + TCA_DSMARK_INDICES, | ||
7072 | + TCA_DSMARK_DEFAULT_INDEX, | ||
7073 | + TCA_DSMARK_SET_TC_INDEX, | ||
7074 | + TCA_DSMARK_MASK, | ||
7075 | + TCA_DSMARK_VALUE, | ||
7076 | + __TCA_DSMARK_MAX, | ||
7077 | +}; | ||
7078 | + | ||
7079 | +#define TCA_DSMARK_MAX (__TCA_DSMARK_MAX - 1) | ||
7080 | + | ||
7081 | +/* ATM section */ | ||
7082 | + | ||
7083 | +enum { | ||
7084 | + TCA_ATM_UNSPEC, | ||
7085 | + TCA_ATM_FD, /* file/socket descriptor */ | ||
7086 | + TCA_ATM_PTR, /* pointer to descriptor - later */ | ||
7087 | + TCA_ATM_HDR, /* LL header */ | ||
7088 | + TCA_ATM_EXCESS, /* excess traffic class (0 for CLP) */ | ||
7089 | + TCA_ATM_ADDR, /* PVC address (for output only) */ | ||
7090 | + TCA_ATM_STATE, /* VC state (ATM_VS_*; for output only) */ | ||
7091 | + __TCA_ATM_MAX, | ||
7092 | +}; | ||
7093 | + | ||
7094 | +#define TCA_ATM_MAX (__TCA_ATM_MAX - 1) | ||
7095 | + | ||
7096 | +/* Network emulator */ | ||
7097 | + | ||
7098 | +enum { | ||
7099 | + TCA_NETEM_UNSPEC, | ||
7100 | + TCA_NETEM_CORR, | ||
7101 | + TCA_NETEM_DELAY_DIST, | ||
7102 | + TCA_NETEM_REORDER, | ||
7103 | + TCA_NETEM_CORRUPT, | ||
7104 | + TCA_NETEM_LOSS, | ||
7105 | + TCA_NETEM_RATE, | ||
7106 | + TCA_NETEM_ECN, | ||
7107 | + TCA_NETEM_RATE64, | ||
7108 | + TCA_NETEM_PAD, | ||
7109 | + TCA_NETEM_LATENCY64, | ||
7110 | + TCA_NETEM_JITTER64, | ||
7111 | + TCA_NETEM_SLOT, | ||
7112 | + TCA_NETEM_SLOT_DIST, | ||
7113 | + __TCA_NETEM_MAX, | ||
7114 | +}; | ||
7115 | + | ||
7116 | +#define TCA_NETEM_MAX (__TCA_NETEM_MAX - 1) | ||
7117 | + | ||
7118 | +struct tc_netem_qopt { | ||
7119 | + __u32 latency; /* added delay (us) */ | ||
7120 | + __u32 limit; /* fifo limit (packets) */ | ||
7121 | + __u32 loss; /* random packet loss (0=none ~0=100%) */ | ||
7122 | + __u32 gap; /* re-ordering gap (0 for none) */ | ||
7123 | + __u32 duplicate; /* random packet dup (0=none ~0=100%) */ | ||
7124 | + __u32 jitter; /* random jitter in latency (us) */ | ||
7125 | +}; | ||
7126 | + | ||
7127 | +struct tc_netem_corr { | ||
7128 | + __u32 delay_corr; /* delay correlation */ | ||
7129 | + __u32 loss_corr; /* packet loss correlation */ | ||
7130 | + __u32 dup_corr; /* duplicate correlation */ | ||
7131 | +}; | ||
7132 | + | ||
7133 | +struct tc_netem_reorder { | ||
7134 | + __u32 probability; | ||
7135 | + __u32 correlation; | ||
7136 | +}; | ||
7137 | + | ||
7138 | +struct tc_netem_corrupt { | ||
7139 | + __u32 probability; | ||
7140 | + __u32 correlation; | ||
7141 | +}; | ||
7142 | + | ||
7143 | +struct tc_netem_rate { | ||
7144 | + __u32 rate; /* byte/s */ | ||
7145 | + __s32 packet_overhead; | ||
7146 | + __u32 cell_size; | ||
7147 | + __s32 cell_overhead; | ||
7148 | +}; | ||
7149 | + | ||
7150 | +struct tc_netem_slot { | ||
7151 | + __s64 min_delay; /* nsec */ | ||
7152 | + __s64 max_delay; | ||
7153 | + __s32 max_packets; | ||
7154 | + __s32 max_bytes; | ||
7155 | + __s64 dist_delay; /* nsec */ | ||
7156 | + __s64 dist_jitter; /* nsec */ | ||
7157 | +}; | ||
7158 | + | ||
7159 | +enum { | ||
7160 | + NETEM_LOSS_UNSPEC, | ||
7161 | + NETEM_LOSS_GI, /* General Intuitive - 4 state model */ | ||
7162 | + NETEM_LOSS_GE, /* Gilbert Elliot models */ | ||
7163 | + __NETEM_LOSS_MAX | ||
7164 | +}; | ||
7165 | +#define NETEM_LOSS_MAX (__NETEM_LOSS_MAX - 1) | ||
7166 | + | ||
7167 | +/* State transition probabilities for 4 state model */ | ||
7168 | +struct tc_netem_gimodel { | ||
7169 | + __u32 p13; | ||
7170 | + __u32 p31; | ||
7171 | + __u32 p32; | ||
7172 | + __u32 p14; | ||
7173 | + __u32 p23; | ||
7174 | +}; | ||
7175 | + | ||
7176 | +/* Gilbert-Elliot models */ | ||
7177 | +struct tc_netem_gemodel { | ||
7178 | + __u32 p; | ||
7179 | + __u32 r; | ||
7180 | + __u32 h; | ||
7181 | + __u32 k1; | ||
7182 | +}; | ||
7183 | + | ||
7184 | +#define NETEM_DIST_SCALE 8192 | ||
7185 | +#define NETEM_DIST_MAX 16384 | ||
7186 | + | ||
7187 | +/* DRR */ | ||
7188 | + | ||
7189 | +enum { | ||
7190 | + TCA_DRR_UNSPEC, | ||
7191 | + TCA_DRR_QUANTUM, | ||
7192 | + __TCA_DRR_MAX | ||
7193 | +}; | ||
7194 | + | ||
7195 | +#define TCA_DRR_MAX (__TCA_DRR_MAX - 1) | ||
7196 | + | ||
7197 | +struct tc_drr_stats { | ||
7198 | + __u32 deficit; | ||
7199 | +}; | ||
7200 | + | ||
7201 | +/* MQPRIO */ | ||
7202 | +#define TC_QOPT_BITMASK 15 | ||
7203 | +#define TC_QOPT_MAX_QUEUE 16 | ||
7204 | + | ||
7205 | +enum { | ||
7206 | + TC_MQPRIO_HW_OFFLOAD_NONE, /* no offload requested */ | ||
7207 | + TC_MQPRIO_HW_OFFLOAD_TCS, /* offload TCs, no queue counts */ | ||
7208 | + __TC_MQPRIO_HW_OFFLOAD_MAX | ||
7209 | +}; | ||
7210 | + | ||
7211 | +#define TC_MQPRIO_HW_OFFLOAD_MAX (__TC_MQPRIO_HW_OFFLOAD_MAX - 1) | ||
7212 | + | ||
7213 | +enum { | ||
7214 | + TC_MQPRIO_MODE_DCB, | ||
7215 | + TC_MQPRIO_MODE_CHANNEL, | ||
7216 | + __TC_MQPRIO_MODE_MAX | ||
7217 | +}; | ||
7218 | + | ||
7219 | +#define __TC_MQPRIO_MODE_MAX (__TC_MQPRIO_MODE_MAX - 1) | ||
7220 | + | ||
7221 | +enum { | ||
7222 | + TC_MQPRIO_SHAPER_DCB, | ||
7223 | + TC_MQPRIO_SHAPER_BW_RATE, /* Add new shapers below */ | ||
7224 | + __TC_MQPRIO_SHAPER_MAX | ||
7225 | +}; | ||
7226 | + | ||
7227 | +#define __TC_MQPRIO_SHAPER_MAX (__TC_MQPRIO_SHAPER_MAX - 1) | ||
7228 | + | ||
7229 | +struct tc_mqprio_qopt { | ||
7230 | + __u8 num_tc; | ||
7231 | + __u8 prio_tc_map[TC_QOPT_BITMASK + 1]; | ||
7232 | + __u8 hw; | ||
7233 | + __u16 count[TC_QOPT_MAX_QUEUE]; | ||
7234 | + __u16 offset[TC_QOPT_MAX_QUEUE]; | ||
7235 | +}; | ||
7236 | + | ||
7237 | +#define TC_MQPRIO_F_MODE 0x1 | ||
7238 | +#define TC_MQPRIO_F_SHAPER 0x2 | ||
7239 | +#define TC_MQPRIO_F_MIN_RATE 0x4 | ||
7240 | +#define TC_MQPRIO_F_MAX_RATE 0x8 | ||
7241 | + | ||
7242 | +enum { | ||
7243 | + TCA_MQPRIO_UNSPEC, | ||
7244 | + TCA_MQPRIO_MODE, | ||
7245 | + TCA_MQPRIO_SHAPER, | ||
7246 | + TCA_MQPRIO_MIN_RATE64, | ||
7247 | + TCA_MQPRIO_MAX_RATE64, | ||
7248 | + __TCA_MQPRIO_MAX, | ||
7249 | +}; | ||
7250 | + | ||
7251 | +#define TCA_MQPRIO_MAX (__TCA_MQPRIO_MAX - 1) | ||
7252 | + | ||
7253 | +/* SFB */ | ||
7254 | + | ||
7255 | +enum { | ||
7256 | + TCA_SFB_UNSPEC, | ||
7257 | + TCA_SFB_PARMS, | ||
7258 | + __TCA_SFB_MAX, | ||
7259 | +}; | ||
7260 | + | ||
7261 | +#define TCA_SFB_MAX (__TCA_SFB_MAX - 1) | ||
7262 | + | ||
7263 | +/* | ||
7264 | + * Note: increment, decrement are Q0.16 fixed-point values. | ||
7265 | + */ | ||
7266 | +struct tc_sfb_qopt { | ||
7267 | + __u32 rehash_interval; /* delay between hash move, in ms */ | ||
7268 | + __u32 warmup_time; /* double buffering warmup time in ms (warmup_time < rehash_interval) */ | ||
7269 | + __u32 max; /* max len of qlen_min */ | ||
7270 | + __u32 bin_size; /* maximum queue length per bin */ | ||
7271 | + __u32 increment; /* probability increment, (d1 in Blue) */ | ||
7272 | + __u32 decrement; /* probability decrement, (d2 in Blue) */ | ||
7273 | + __u32 limit; /* max SFB queue length */ | ||
7274 | + __u32 penalty_rate; /* inelastic flows are rate limited to 'rate' pps */ | ||
7275 | + __u32 penalty_burst; | ||
7276 | +}; | ||
7277 | + | ||
7278 | +struct tc_sfb_xstats { | ||
7279 | + __u32 earlydrop; | ||
7280 | + __u32 penaltydrop; | ||
7281 | + __u32 bucketdrop; | ||
7282 | + __u32 queuedrop; | ||
7283 | + __u32 childdrop; /* drops in child qdisc */ | ||
7284 | + __u32 marked; | ||
7285 | + __u32 maxqlen; | ||
7286 | + __u32 maxprob; | ||
7287 | + __u32 avgprob; | ||
7288 | +}; | ||
7289 | + | ||
7290 | +#define SFB_MAX_PROB 0xFFFF | ||
7291 | + | ||
7292 | +/* QFQ */ | ||
7293 | +enum { | ||
7294 | + TCA_QFQ_UNSPEC, | ||
7295 | + TCA_QFQ_WEIGHT, | ||
7296 | + TCA_QFQ_LMAX, | ||
7297 | + __TCA_QFQ_MAX | ||
7298 | +}; | ||
7299 | + | ||
7300 | +#define TCA_QFQ_MAX (__TCA_QFQ_MAX - 1) | ||
7301 | + | ||
7302 | +struct tc_qfq_stats { | ||
7303 | + __u32 weight; | ||
7304 | + __u32 lmax; | ||
7305 | +}; | ||
7306 | + | ||
7307 | +/* CODEL */ | ||
7308 | + | ||
7309 | +enum { | ||
7310 | + TCA_CODEL_UNSPEC, | ||
7311 | + TCA_CODEL_TARGET, | ||
7312 | + TCA_CODEL_LIMIT, | ||
7313 | + TCA_CODEL_INTERVAL, | ||
7314 | + TCA_CODEL_ECN, | ||
7315 | + TCA_CODEL_CE_THRESHOLD, | ||
7316 | + __TCA_CODEL_MAX | ||
7317 | +}; | ||
7318 | + | ||
7319 | +#define TCA_CODEL_MAX (__TCA_CODEL_MAX - 1) | ||
7320 | + | ||
7321 | +struct tc_codel_xstats { | ||
7322 | + __u32 maxpacket; /* largest packet we've seen so far */ | ||
7323 | + __u32 count; /* how many drops we've done since the last time we | ||
7324 | + * entered dropping state | ||
7325 | + */ | ||
7326 | + __u32 lastcount; /* count at entry to dropping state */ | ||
7327 | + __u32 ldelay; /* in-queue delay seen by most recently dequeued packet */ | ||
7328 | + __s32 drop_next; /* time to drop next packet */ | ||
7329 | + __u32 drop_overlimit; /* number of time max qdisc packet limit was hit */ | ||
7330 | + __u32 ecn_mark; /* number of packets we ECN marked instead of dropped */ | ||
7331 | + __u32 dropping; /* are we in dropping state ? */ | ||
7332 | + __u32 ce_mark; /* number of CE marked packets because of ce_threshold */ | ||
7333 | +}; | ||
7334 | + | ||
7335 | +/* FQ_CODEL */ | ||
7336 | + | ||
7337 | +enum { | ||
7338 | + TCA_FQ_CODEL_UNSPEC, | ||
7339 | + TCA_FQ_CODEL_TARGET, | ||
7340 | + TCA_FQ_CODEL_LIMIT, | ||
7341 | + TCA_FQ_CODEL_INTERVAL, | ||
7342 | + TCA_FQ_CODEL_ECN, | ||
7343 | + TCA_FQ_CODEL_FLOWS, | ||
7344 | + TCA_FQ_CODEL_QUANTUM, | ||
7345 | + TCA_FQ_CODEL_CE_THRESHOLD, | ||
7346 | + TCA_FQ_CODEL_DROP_BATCH_SIZE, | ||
7347 | + TCA_FQ_CODEL_MEMORY_LIMIT, | ||
7348 | + __TCA_FQ_CODEL_MAX | ||
7349 | +}; | ||
7350 | + | ||
7351 | +#define TCA_FQ_CODEL_MAX (__TCA_FQ_CODEL_MAX - 1) | ||
7352 | + | ||
7353 | +enum { | ||
7354 | + TCA_FQ_CODEL_XSTATS_QDISC, | ||
7355 | + TCA_FQ_CODEL_XSTATS_CLASS, | ||
7356 | +}; | ||
7357 | + | ||
7358 | +struct tc_fq_codel_qd_stats { | ||
7359 | + __u32 maxpacket; /* largest packet we've seen so far */ | ||
7360 | + __u32 drop_overlimit; /* number of time max qdisc | ||
7361 | + * packet limit was hit | ||
7362 | + */ | ||
7363 | + __u32 ecn_mark; /* number of packets we ECN marked | ||
7364 | + * instead of being dropped | ||
7365 | + */ | ||
7366 | + __u32 new_flow_count; /* number of time packets | ||
7367 | + * created a 'new flow' | ||
7368 | + */ | ||
7369 | + __u32 new_flows_len; /* count of flows in new list */ | ||
7370 | + __u32 old_flows_len; /* count of flows in old list */ | ||
7371 | + __u32 ce_mark; /* packets above ce_threshold */ | ||
7372 | + __u32 memory_usage; /* in bytes */ | ||
7373 | + __u32 drop_overmemory; | ||
7374 | +}; | ||
7375 | + | ||
7376 | +struct tc_fq_codel_cl_stats { | ||
7377 | + __s32 deficit; | ||
7378 | + __u32 ldelay; /* in-queue delay seen by most recently | ||
7379 | + * dequeued packet | ||
7380 | + */ | ||
7381 | + __u32 count; | ||
7382 | + __u32 lastcount; | ||
7383 | + __u32 dropping; | ||
7384 | + __s32 drop_next; | ||
7385 | +}; | ||
7386 | + | ||
7387 | +struct tc_fq_codel_xstats { | ||
7388 | + __u32 type; | ||
7389 | + union { | ||
7390 | + struct tc_fq_codel_qd_stats qdisc_stats; | ||
7391 | + struct tc_fq_codel_cl_stats class_stats; | ||
7392 | + }; | ||
7393 | +}; | ||
7394 | + | ||
7395 | +/* FQ */ | ||
7396 | + | ||
7397 | +enum { | ||
7398 | + TCA_FQ_UNSPEC, | ||
7399 | + | ||
7400 | + TCA_FQ_PLIMIT, /* limit of total number of packets in queue */ | ||
7401 | + | ||
7402 | + TCA_FQ_FLOW_PLIMIT, /* limit of packets per flow */ | ||
7403 | + | ||
7404 | + TCA_FQ_QUANTUM, /* RR quantum */ | ||
7405 | + | ||
7406 | + TCA_FQ_INITIAL_QUANTUM, /* RR quantum for new flow */ | ||
7407 | + | ||
7408 | + TCA_FQ_RATE_ENABLE, /* enable/disable rate limiting */ | ||
7409 | + | ||
7410 | + TCA_FQ_FLOW_DEFAULT_RATE,/* obsolete, do not use */ | ||
7411 | + | ||
7412 | + TCA_FQ_FLOW_MAX_RATE, /* per flow max rate */ | ||
7413 | + | ||
7414 | + TCA_FQ_BUCKETS_LOG, /* log2(number of buckets) */ | ||
7415 | + | ||
7416 | + TCA_FQ_FLOW_REFILL_DELAY, /* flow credit refill delay in usec */ | ||
7417 | + | ||
7418 | + TCA_FQ_ORPHAN_MASK, /* mask applied to orphaned skb hashes */ | ||
7419 | + | ||
7420 | + TCA_FQ_LOW_RATE_THRESHOLD, /* per packet delay under this rate */ | ||
7421 | + | ||
7422 | + TCA_FQ_CE_THRESHOLD, /* DCTCP-like CE-marking threshold */ | ||
7423 | + | ||
7424 | + __TCA_FQ_MAX | ||
7425 | +}; | ||
7426 | + | ||
7427 | +#define TCA_FQ_MAX (__TCA_FQ_MAX - 1) | ||
7428 | + | ||
7429 | +struct tc_fq_qd_stats { | ||
7430 | + __u64 gc_flows; | ||
7431 | + __u64 highprio_packets; | ||
7432 | + __u64 tcp_retrans; | ||
7433 | + __u64 throttled; | ||
7434 | + __u64 flows_plimit; | ||
7435 | + __u64 pkts_too_long; | ||
7436 | + __u64 allocation_errors; | ||
7437 | + __s64 time_next_delayed_flow; | ||
7438 | + __u32 flows; | ||
7439 | + __u32 inactive_flows; | ||
7440 | + __u32 throttled_flows; | ||
7441 | + __u32 unthrottle_latency_ns; | ||
7442 | + __u64 ce_mark; /* packets above ce_threshold */ | ||
7443 | +}; | ||
7444 | + | ||
7445 | +/* Heavy-Hitter Filter */ | ||
7446 | + | ||
7447 | +enum { | ||
7448 | + TCA_HHF_UNSPEC, | ||
7449 | + TCA_HHF_BACKLOG_LIMIT, | ||
7450 | + TCA_HHF_QUANTUM, | ||
7451 | + TCA_HHF_HH_FLOWS_LIMIT, | ||
7452 | + TCA_HHF_RESET_TIMEOUT, | ||
7453 | + TCA_HHF_ADMIT_BYTES, | ||
7454 | + TCA_HHF_EVICT_TIMEOUT, | ||
7455 | + TCA_HHF_NON_HH_WEIGHT, | ||
7456 | + __TCA_HHF_MAX | ||
7457 | +}; | ||
7458 | + | ||
7459 | +#define TCA_HHF_MAX (__TCA_HHF_MAX - 1) | ||
7460 | + | ||
7461 | +struct tc_hhf_xstats { | ||
7462 | + __u32 drop_overlimit; /* number of times max qdisc packet limit | ||
7463 | + * was hit | ||
7464 | + */ | ||
7465 | + __u32 hh_overlimit; /* number of times max heavy-hitters was hit */ | ||
7466 | + __u32 hh_tot_count; /* number of captured heavy-hitters so far */ | ||
7467 | + __u32 hh_cur_count; /* number of current heavy-hitters */ | ||
7468 | +}; | ||
7469 | + | ||
7470 | +/* PIE */ | ||
7471 | +enum { | ||
7472 | + TCA_PIE_UNSPEC, | ||
7473 | + TCA_PIE_TARGET, | ||
7474 | + TCA_PIE_LIMIT, | ||
7475 | + TCA_PIE_TUPDATE, | ||
7476 | + TCA_PIE_ALPHA, | ||
7477 | + TCA_PIE_BETA, | ||
7478 | + TCA_PIE_ECN, | ||
7479 | + TCA_PIE_BYTEMODE, | ||
7480 | + __TCA_PIE_MAX | ||
7481 | +}; | ||
7482 | +#define TCA_PIE_MAX (__TCA_PIE_MAX - 1) | ||
7483 | + | ||
7484 | +struct tc_pie_xstats { | ||
7485 | + __u32 prob; /* current probability */ | ||
7486 | + __u32 delay; /* current delay in ms */ | ||
7487 | + __u32 avg_dq_rate; /* current average dq_rate in bits/pie_time */ | ||
7488 | + __u32 packets_in; /* total number of packets enqueued */ | ||
7489 | + __u32 dropped; /* packets dropped due to pie_action */ | ||
7490 | + __u32 overlimit; /* dropped due to lack of space in queue */ | ||
7491 | + __u32 maxq; /* maximum queue size */ | ||
7492 | + __u32 ecn_mark; /* packets marked with ecn*/ | ||
7493 | +}; | ||
7494 | + | ||
7495 | +/* CBS */ | ||
7496 | +struct tc_cbs_qopt { | ||
7497 | + __u8 offload; | ||
7498 | + __u8 _pad[3]; | ||
7499 | + __s32 hicredit; | ||
7500 | + __s32 locredit; | ||
7501 | + __s32 idleslope; | ||
7502 | + __s32 sendslope; | ||
7503 | +}; | ||
7504 | + | ||
7505 | +enum { | ||
7506 | + TCA_CBS_UNSPEC, | ||
7507 | + TCA_CBS_PARMS, | ||
7508 | + __TCA_CBS_MAX, | ||
7509 | +}; | ||
7510 | + | ||
7511 | +#define TCA_CBS_MAX (__TCA_CBS_MAX - 1) | ||
7512 | + | ||
7513 | + | ||
7514 | +/* ETF */ | ||
7515 | +struct tc_etf_qopt { | ||
7516 | + __s32 delta; | ||
7517 | + __s32 clockid; | ||
7518 | + __u32 flags; | ||
7519 | +#define TC_ETF_DEADLINE_MODE_ON BIT(0) | ||
7520 | +#define TC_ETF_OFFLOAD_ON BIT(1) | ||
7521 | +}; | ||
7522 | + | ||
7523 | +enum { | ||
7524 | + TCA_ETF_UNSPEC, | ||
7525 | + TCA_ETF_PARMS, | ||
7526 | + __TCA_ETF_MAX, | ||
7527 | +}; | ||
7528 | + | ||
7529 | +#define TCA_ETF_MAX (__TCA_ETF_MAX - 1) | ||
7530 | + | ||
7531 | + | ||
7532 | +/* CAKE */ | ||
7533 | +enum { | ||
7534 | + TCA_CAKE_UNSPEC, | ||
7535 | + TCA_CAKE_PAD, | ||
7536 | + TCA_CAKE_BASE_RATE64, | ||
7537 | + TCA_CAKE_DIFFSERV_MODE, | ||
7538 | + TCA_CAKE_ATM, | ||
7539 | + TCA_CAKE_FLOW_MODE, | ||
7540 | + TCA_CAKE_OVERHEAD, | ||
7541 | + TCA_CAKE_RTT, | ||
7542 | + TCA_CAKE_TARGET, | ||
7543 | + TCA_CAKE_AUTORATE, | ||
7544 | + TCA_CAKE_MEMORY, | ||
7545 | + TCA_CAKE_NAT, | ||
7546 | + TCA_CAKE_RAW, | ||
7547 | + TCA_CAKE_WASH, | ||
7548 | + TCA_CAKE_MPU, | ||
7549 | + TCA_CAKE_INGRESS, | ||
7550 | + TCA_CAKE_ACK_FILTER, | ||
7551 | + TCA_CAKE_SPLIT_GSO, | ||
7552 | + __TCA_CAKE_MAX | ||
7553 | +}; | ||
7554 | +#define TCA_CAKE_MAX (__TCA_CAKE_MAX - 1) | ||
7555 | + | ||
7556 | +enum { | ||
7557 | + __TCA_CAKE_STATS_INVALID, | ||
7558 | + TCA_CAKE_STATS_PAD, | ||
7559 | + TCA_CAKE_STATS_CAPACITY_ESTIMATE64, | ||
7560 | + TCA_CAKE_STATS_MEMORY_LIMIT, | ||
7561 | + TCA_CAKE_STATS_MEMORY_USED, | ||
7562 | + TCA_CAKE_STATS_AVG_NETOFF, | ||
7563 | + TCA_CAKE_STATS_MIN_NETLEN, | ||
7564 | + TCA_CAKE_STATS_MAX_NETLEN, | ||
7565 | + TCA_CAKE_STATS_MIN_ADJLEN, | ||
7566 | + TCA_CAKE_STATS_MAX_ADJLEN, | ||
7567 | + TCA_CAKE_STATS_TIN_STATS, | ||
7568 | + TCA_CAKE_STATS_DEFICIT, | ||
7569 | + TCA_CAKE_STATS_COBALT_COUNT, | ||
7570 | + TCA_CAKE_STATS_DROPPING, | ||
7571 | + TCA_CAKE_STATS_DROP_NEXT_US, | ||
7572 | + TCA_CAKE_STATS_P_DROP, | ||
7573 | + TCA_CAKE_STATS_BLUE_TIMER_US, | ||
7574 | + __TCA_CAKE_STATS_MAX | ||
7575 | +}; | ||
7576 | +#define TCA_CAKE_STATS_MAX (__TCA_CAKE_STATS_MAX - 1) | ||
7577 | + | ||
7578 | +enum { | ||
7579 | + __TCA_CAKE_TIN_STATS_INVALID, | ||
7580 | + TCA_CAKE_TIN_STATS_PAD, | ||
7581 | + TCA_CAKE_TIN_STATS_SENT_PACKETS, | ||
7582 | + TCA_CAKE_TIN_STATS_SENT_BYTES64, | ||
7583 | + TCA_CAKE_TIN_STATS_DROPPED_PACKETS, | ||
7584 | + TCA_CAKE_TIN_STATS_DROPPED_BYTES64, | ||
7585 | + TCA_CAKE_TIN_STATS_ACKS_DROPPED_PACKETS, | ||
7586 | + TCA_CAKE_TIN_STATS_ACKS_DROPPED_BYTES64, | ||
7587 | + TCA_CAKE_TIN_STATS_ECN_MARKED_PACKETS, | ||
7588 | + TCA_CAKE_TIN_STATS_ECN_MARKED_BYTES64, | ||
7589 | + TCA_CAKE_TIN_STATS_BACKLOG_PACKETS, | ||
7590 | + TCA_CAKE_TIN_STATS_BACKLOG_BYTES, | ||
7591 | + TCA_CAKE_TIN_STATS_THRESHOLD_RATE64, | ||
7592 | + TCA_CAKE_TIN_STATS_TARGET_US, | ||
7593 | + TCA_CAKE_TIN_STATS_INTERVAL_US, | ||
7594 | + TCA_CAKE_TIN_STATS_WAY_INDIRECT_HITS, | ||
7595 | + TCA_CAKE_TIN_STATS_WAY_MISSES, | ||
7596 | + TCA_CAKE_TIN_STATS_WAY_COLLISIONS, | ||
7597 | + TCA_CAKE_TIN_STATS_PEAK_DELAY_US, | ||
7598 | + TCA_CAKE_TIN_STATS_AVG_DELAY_US, | ||
7599 | + TCA_CAKE_TIN_STATS_BASE_DELAY_US, | ||
7600 | + TCA_CAKE_TIN_STATS_SPARSE_FLOWS, | ||
7601 | + TCA_CAKE_TIN_STATS_BULK_FLOWS, | ||
7602 | + TCA_CAKE_TIN_STATS_UNRESPONSIVE_FLOWS, | ||
7603 | + TCA_CAKE_TIN_STATS_MAX_SKBLEN, | ||
7604 | + TCA_CAKE_TIN_STATS_FLOW_QUANTUM, | ||
7605 | + __TCA_CAKE_TIN_STATS_MAX | ||
7606 | +}; | ||
7607 | +#define TCA_CAKE_TIN_STATS_MAX (__TCA_CAKE_TIN_STATS_MAX - 1) | ||
7608 | +#define TC_CAKE_MAX_TINS (8) | ||
7609 | + | ||
7610 | +enum { | ||
7611 | + CAKE_FLOW_NONE = 0, | ||
7612 | + CAKE_FLOW_SRC_IP, | ||
7613 | + CAKE_FLOW_DST_IP, | ||
7614 | + CAKE_FLOW_HOSTS, /* = CAKE_FLOW_SRC_IP | CAKE_FLOW_DST_IP */ | ||
7615 | + CAKE_FLOW_FLOWS, | ||
7616 | + CAKE_FLOW_DUAL_SRC, /* = CAKE_FLOW_SRC_IP | CAKE_FLOW_FLOWS */ | ||
7617 | + CAKE_FLOW_DUAL_DST, /* = CAKE_FLOW_DST_IP | CAKE_FLOW_FLOWS */ | ||
7618 | + CAKE_FLOW_TRIPLE, /* = CAKE_FLOW_HOSTS | CAKE_FLOW_FLOWS */ | ||
7619 | + CAKE_FLOW_MAX, | ||
7620 | +}; | ||
7621 | + | ||
7622 | +enum { | ||
7623 | + CAKE_DIFFSERV_DIFFSERV3 = 0, | ||
7624 | + CAKE_DIFFSERV_DIFFSERV4, | ||
7625 | + CAKE_DIFFSERV_DIFFSERV8, | ||
7626 | + CAKE_DIFFSERV_BESTEFFORT, | ||
7627 | + CAKE_DIFFSERV_PRECEDENCE, | ||
7628 | + CAKE_DIFFSERV_MAX | ||
7629 | +}; | ||
7630 | + | ||
7631 | +enum { | ||
7632 | + CAKE_ACK_NONE = 0, | ||
7633 | + CAKE_ACK_FILTER, | ||
7634 | + CAKE_ACK_AGGRESSIVE, | ||
7635 | + CAKE_ACK_MAX | ||
7636 | +}; | ||
7637 | + | ||
7638 | +enum { | ||
7639 | + CAKE_ATM_NONE = 0, | ||
7640 | + CAKE_ATM_ATM, | ||
7641 | + CAKE_ATM_PTM, | ||
7642 | + CAKE_ATM_MAX | ||
7643 | +}; | ||
7644 | + | ||
7645 | + | ||
7646 | +/* TAPRIO */ | ||
7647 | +enum { | ||
7648 | + TC_TAPRIO_CMD_SET_GATES = 0x00, | ||
7649 | + TC_TAPRIO_CMD_SET_AND_HOLD = 0x01, | ||
7650 | + TC_TAPRIO_CMD_SET_AND_RELEASE = 0x02, | ||
7651 | +}; | ||
7652 | + | ||
7653 | +enum { | ||
7654 | + TCA_TAPRIO_SCHED_ENTRY_UNSPEC, | ||
7655 | + TCA_TAPRIO_SCHED_ENTRY_INDEX, /* u32 */ | ||
7656 | + TCA_TAPRIO_SCHED_ENTRY_CMD, /* u8 */ | ||
7657 | + TCA_TAPRIO_SCHED_ENTRY_GATE_MASK, /* u32 */ | ||
7658 | + TCA_TAPRIO_SCHED_ENTRY_INTERVAL, /* u32 */ | ||
7659 | + __TCA_TAPRIO_SCHED_ENTRY_MAX, | ||
7660 | +}; | ||
7661 | +#define TCA_TAPRIO_SCHED_ENTRY_MAX (__TCA_TAPRIO_SCHED_ENTRY_MAX - 1) | ||
7662 | + | ||
7663 | +/* The format for schedule entry list is: | ||
7664 | + * [TCA_TAPRIO_SCHED_ENTRY_LIST] | ||
7665 | + * [TCA_TAPRIO_SCHED_ENTRY] | ||
7666 | + * [TCA_TAPRIO_SCHED_ENTRY_CMD] | ||
7667 | + * [TCA_TAPRIO_SCHED_ENTRY_GATES] | ||
7668 | + * [TCA_TAPRIO_SCHED_ENTRY_INTERVAL] | ||
7669 | + */ | ||
7670 | +enum { | ||
7671 | + TCA_TAPRIO_SCHED_UNSPEC, | ||
7672 | + TCA_TAPRIO_SCHED_ENTRY, | ||
7673 | + __TCA_TAPRIO_SCHED_MAX, | ||
7674 | +}; | ||
7675 | + | ||
7676 | +#define TCA_TAPRIO_SCHED_MAX (__TCA_TAPRIO_SCHED_MAX - 1) | ||
7677 | + | ||
7678 | +enum { | ||
7679 | + TCA_TAPRIO_ATTR_UNSPEC, | ||
7680 | + TCA_TAPRIO_ATTR_PRIOMAP, /* struct tc_mqprio_qopt */ | ||
7681 | + TCA_TAPRIO_ATTR_SCHED_ENTRY_LIST, /* nested of entry */ | ||
7682 | + TCA_TAPRIO_ATTR_SCHED_BASE_TIME, /* s64 */ | ||
7683 | + TCA_TAPRIO_ATTR_SCHED_SINGLE_ENTRY, /* single entry */ | ||
7684 | + TCA_TAPRIO_ATTR_SCHED_CLOCKID, /* s32 */ | ||
7685 | + TCA_TAPRIO_PAD, | ||
7686 | + __TCA_TAPRIO_ATTR_MAX, | ||
7687 | +}; | ||
7688 | + | ||
7689 | +#define TCA_TAPRIO_ATTR_MAX (__TCA_TAPRIO_ATTR_MAX - 1) | ||
7690 | + | ||
7691 | +#endif | ||
7692 | diff --git a/tools/testing/selftests/bpf/Makefile b/tools/testing/selftests/bpf/Makefile | ||
7693 | index ecd79b7fb1073..d5e992f7c7dd5 100644 | ||
7694 | --- a/tools/testing/selftests/bpf/Makefile | ||
7695 | +++ b/tools/testing/selftests/bpf/Makefile | ||
7696 | @@ -53,7 +53,10 @@ TEST_PROGS := test_kmod.sh \ | ||
7697 | test_flow_dissector.sh \ | ||
7698 | test_xdp_vlan.sh | ||
7699 | |||
7700 | -TEST_PROGS_EXTENDED := with_addr.sh | ||
7701 | +TEST_PROGS_EXTENDED := with_addr.sh \ | ||
7702 | + with_tunnels.sh \ | ||
7703 | + tcp_client.py \ | ||
7704 | + tcp_server.py | ||
7705 | |||
7706 | # Compile but not part of 'make run_tests' | ||
7707 | TEST_GEN_PROGS_EXTENDED = test_libbpf_open test_sock_addr test_skb_cgroup_id_user \ | ||
7708 | diff --git a/tools/testing/selftests/bpf/test_progs.c b/tools/testing/selftests/bpf/test_progs.c | ||
7709 | index 6ac7232b0fdeb..3ec4ce156074c 100644 | ||
7710 | --- a/tools/testing/selftests/bpf/test_progs.c | ||
7711 | +++ b/tools/testing/selftests/bpf/test_progs.c | ||
7712 | @@ -1136,7 +1136,9 @@ static void test_stacktrace_build_id(void) | ||
7713 | int i, j; | ||
7714 | struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; | ||
7715 | int build_id_matches = 0; | ||
7716 | + int retry = 1; | ||
7717 | |||
7718 | +retry: | ||
7719 | err = bpf_prog_load(file, BPF_PROG_TYPE_TRACEPOINT, &obj, &prog_fd); | ||
7720 | if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno)) | ||
7721 | goto out; | ||
7722 | @@ -1249,6 +1251,19 @@ static void test_stacktrace_build_id(void) | ||
7723 | previous_key = key; | ||
7724 | } while (bpf_map_get_next_key(stackmap_fd, &previous_key, &key) == 0); | ||
7725 | |||
7726 | + /* stack_map_get_build_id_offset() is racy and sometimes can return | ||
7727 | + * BPF_STACK_BUILD_ID_IP instead of BPF_STACK_BUILD_ID_VALID; | ||
7728 | + * try it one more time. | ||
7729 | + */ | ||
7730 | + if (build_id_matches < 1 && retry--) { | ||
7731 | + ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE); | ||
7732 | + close(pmu_fd); | ||
7733 | + bpf_object__close(obj); | ||
7734 | + printf("%s:WARN:Didn't find expected build ID from the map, retrying\n", | ||
7735 | + __func__); | ||
7736 | + goto retry; | ||
7737 | + } | ||
7738 | + | ||
7739 | if (CHECK(build_id_matches < 1, "build id match", | ||
7740 | "Didn't find expected build ID from the map\n")) | ||
7741 | goto disable_pmu; | ||
7742 | @@ -1289,7 +1304,9 @@ static void test_stacktrace_build_id_nmi(void) | ||
7743 | int i, j; | ||
7744 | struct bpf_stack_build_id id_offs[PERF_MAX_STACK_DEPTH]; | ||
7745 | int build_id_matches = 0; | ||
7746 | + int retry = 1; | ||
7747 | |||
7748 | +retry: | ||
7749 | err = bpf_prog_load(file, BPF_PROG_TYPE_PERF_EVENT, &obj, &prog_fd); | ||
7750 | if (CHECK(err, "prog_load", "err %d errno %d\n", err, errno)) | ||
7751 | return; | ||
7752 | @@ -1384,6 +1401,19 @@ static void test_stacktrace_build_id_nmi(void) | ||
7753 | previous_key = key; | ||
7754 | } while (bpf_map_get_next_key(stackmap_fd, &previous_key, &key) == 0); | ||
7755 | |||
7756 | + /* stack_map_get_build_id_offset() is racy and sometimes can return | ||
7757 | + * BPF_STACK_BUILD_ID_IP instead of BPF_STACK_BUILD_ID_VALID; | ||
7758 | + * try it one more time. | ||
7759 | + */ | ||
7760 | + if (build_id_matches < 1 && retry--) { | ||
7761 | + ioctl(pmu_fd, PERF_EVENT_IOC_DISABLE); | ||
7762 | + close(pmu_fd); | ||
7763 | + bpf_object__close(obj); | ||
7764 | + printf("%s:WARN:Didn't find expected build ID from the map, retrying\n", | ||
7765 | + __func__); | ||
7766 | + goto retry; | ||
7767 | + } | ||
7768 | + | ||
7769 | if (CHECK(build_id_matches < 1, "build id match", | ||
7770 | "Didn't find expected build ID from the map\n")) | ||
7771 | goto disable_pmu; | ||
7772 | diff --git a/tools/testing/selftests/bpf/test_sock_addr.c b/tools/testing/selftests/bpf/test_sock_addr.c | ||
7773 | index aeeb76a54d633..e38f1cb7089d3 100644 | ||
7774 | --- a/tools/testing/selftests/bpf/test_sock_addr.c | ||
7775 | +++ b/tools/testing/selftests/bpf/test_sock_addr.c | ||
7776 | @@ -44,6 +44,7 @@ | ||
7777 | #define SERV6_V4MAPPED_IP "::ffff:192.168.0.4" | ||
7778 | #define SRC6_IP "::1" | ||
7779 | #define SRC6_REWRITE_IP "::6" | ||
7780 | +#define WILDCARD6_IP "::" | ||
7781 | #define SERV6_PORT 6060 | ||
7782 | #define SERV6_REWRITE_PORT 6666 | ||
7783 | |||
7784 | @@ -85,12 +86,14 @@ static int bind4_prog_load(const struct sock_addr_test *test); | ||
7785 | static int bind6_prog_load(const struct sock_addr_test *test); | ||
7786 | static int connect4_prog_load(const struct sock_addr_test *test); | ||
7787 | static int connect6_prog_load(const struct sock_addr_test *test); | ||
7788 | +static int sendmsg_allow_prog_load(const struct sock_addr_test *test); | ||
7789 | static int sendmsg_deny_prog_load(const struct sock_addr_test *test); | ||
7790 | static int sendmsg4_rw_asm_prog_load(const struct sock_addr_test *test); | ||
7791 | static int sendmsg4_rw_c_prog_load(const struct sock_addr_test *test); | ||
7792 | static int sendmsg6_rw_asm_prog_load(const struct sock_addr_test *test); | ||
7793 | static int sendmsg6_rw_c_prog_load(const struct sock_addr_test *test); | ||
7794 | static int sendmsg6_rw_v4mapped_prog_load(const struct sock_addr_test *test); | ||
7795 | +static int sendmsg6_rw_wildcard_prog_load(const struct sock_addr_test *test); | ||
7796 | |||
7797 | static struct sock_addr_test tests[] = { | ||
7798 | /* bind */ | ||
7799 | @@ -462,6 +465,34 @@ static struct sock_addr_test tests[] = { | ||
7800 | SRC6_REWRITE_IP, | ||
7801 | SYSCALL_ENOTSUPP, | ||
7802 | }, | ||
7803 | + { | ||
7804 | + "sendmsg6: set dst IP = [::] (BSD'ism)", | ||
7805 | + sendmsg6_rw_wildcard_prog_load, | ||
7806 | + BPF_CGROUP_UDP6_SENDMSG, | ||
7807 | + BPF_CGROUP_UDP6_SENDMSG, | ||
7808 | + AF_INET6, | ||
7809 | + SOCK_DGRAM, | ||
7810 | + SERV6_IP, | ||
7811 | + SERV6_PORT, | ||
7812 | + SERV6_REWRITE_IP, | ||
7813 | + SERV6_REWRITE_PORT, | ||
7814 | + SRC6_REWRITE_IP, | ||
7815 | + SUCCESS, | ||
7816 | + }, | ||
7817 | + { | ||
7818 | + "sendmsg6: preserve dst IP = [::] (BSD'ism)", | ||
7819 | + sendmsg_allow_prog_load, | ||
7820 | + BPF_CGROUP_UDP6_SENDMSG, | ||
7821 | + BPF_CGROUP_UDP6_SENDMSG, | ||
7822 | + AF_INET6, | ||
7823 | + SOCK_DGRAM, | ||
7824 | + WILDCARD6_IP, | ||
7825 | + SERV6_PORT, | ||
7826 | + SERV6_REWRITE_IP, | ||
7827 | + SERV6_PORT, | ||
7828 | + SRC6_IP, | ||
7829 | + SUCCESS, | ||
7830 | + }, | ||
7831 | { | ||
7832 | "sendmsg6: deny call", | ||
7833 | sendmsg_deny_prog_load, | ||
7834 | @@ -714,16 +745,27 @@ static int connect6_prog_load(const struct sock_addr_test *test) | ||
7835 | return load_path(test, CONNECT6_PROG_PATH); | ||
7836 | } | ||
7837 | |||
7838 | -static int sendmsg_deny_prog_load(const struct sock_addr_test *test) | ||
7839 | +static int sendmsg_ret_only_prog_load(const struct sock_addr_test *test, | ||
7840 | + int32_t rc) | ||
7841 | { | ||
7842 | struct bpf_insn insns[] = { | ||
7843 | - /* return 0 */ | ||
7844 | - BPF_MOV64_IMM(BPF_REG_0, 0), | ||
7845 | + /* return rc */ | ||
7846 | + BPF_MOV64_IMM(BPF_REG_0, rc), | ||
7847 | BPF_EXIT_INSN(), | ||
7848 | }; | ||
7849 | return load_insns(test, insns, sizeof(insns) / sizeof(struct bpf_insn)); | ||
7850 | } | ||
7851 | |||
7852 | +static int sendmsg_allow_prog_load(const struct sock_addr_test *test) | ||
7853 | +{ | ||
7854 | + return sendmsg_ret_only_prog_load(test, /*rc*/ 1); | ||
7855 | +} | ||
7856 | + | ||
7857 | +static int sendmsg_deny_prog_load(const struct sock_addr_test *test) | ||
7858 | +{ | ||
7859 | + return sendmsg_ret_only_prog_load(test, /*rc*/ 0); | ||
7860 | +} | ||
7861 | + | ||
7862 | static int sendmsg4_rw_asm_prog_load(const struct sock_addr_test *test) | ||
7863 | { | ||
7864 | struct sockaddr_in dst4_rw_addr; | ||
7865 | @@ -844,6 +886,11 @@ static int sendmsg6_rw_v4mapped_prog_load(const struct sock_addr_test *test) | ||
7866 | return sendmsg6_rw_dst_asm_prog_load(test, SERV6_V4MAPPED_IP); | ||
7867 | } | ||
7868 | |||
7869 | +static int sendmsg6_rw_wildcard_prog_load(const struct sock_addr_test *test) | ||
7870 | +{ | ||
7871 | + return sendmsg6_rw_dst_asm_prog_load(test, WILDCARD6_IP); | ||
7872 | +} | ||
7873 | + | ||
7874 | static int sendmsg6_rw_c_prog_load(const struct sock_addr_test *test) | ||
7875 | { | ||
7876 | return load_path(test, SENDMSG6_PROG_PATH); | ||
7877 | diff --git a/tools/testing/selftests/net/forwarding/bridge_vlan_aware.sh b/tools/testing/selftests/net/forwarding/bridge_vlan_aware.sh | ||
7878 | index d8313d0438b74..b90dff8d3a94b 100755 | ||
7879 | --- a/tools/testing/selftests/net/forwarding/bridge_vlan_aware.sh | ||
7880 | +++ b/tools/testing/selftests/net/forwarding/bridge_vlan_aware.sh | ||
7881 | @@ -1,7 +1,7 @@ | ||
7882 | #!/bin/bash | ||
7883 | # SPDX-License-Identifier: GPL-2.0 | ||
7884 | |||
7885 | -ALL_TESTS="ping_ipv4 ping_ipv6 learning flooding" | ||
7886 | +ALL_TESTS="ping_ipv4 ping_ipv6 learning flooding vlan_deletion extern_learn" | ||
7887 | NUM_NETIFS=4 | ||
7888 | CHECK_TC="yes" | ||
7889 | source lib.sh | ||
7890 | @@ -96,6 +96,51 @@ flooding() | ||
7891 | flood_test $swp2 $h1 $h2 | ||
7892 | } | ||
7893 | |||
7894 | +vlan_deletion() | ||
7895 | +{ | ||
7896 | + # Test that the deletion of a VLAN on a bridge port does not affect | ||
7897 | + # the PVID VLAN | ||
7898 | + log_info "Add and delete a VLAN on bridge port $swp1" | ||
7899 | + | ||
7900 | + bridge vlan add vid 10 dev $swp1 | ||
7901 | + bridge vlan del vid 10 dev $swp1 | ||
7902 | + | ||
7903 | + ping_ipv4 | ||
7904 | + ping_ipv6 | ||
7905 | +} | ||
7906 | + | ||
7907 | +extern_learn() | ||
7908 | +{ | ||
7909 | + local mac=de:ad:be:ef:13:37 | ||
7910 | + local ageing_time | ||
7911 | + | ||
7912 | + # Test that externally learned FDB entries can roam, but not age out | ||
7913 | + RET=0 | ||
7914 | + | ||
7915 | + bridge fdb add de:ad:be:ef:13:37 dev $swp1 master extern_learn vlan 1 | ||
7916 | + | ||
7917 | + bridge fdb show brport $swp1 | grep -q de:ad:be:ef:13:37 | ||
7918 | + check_err $? "Did not find FDB entry when should" | ||
7919 | + | ||
7920 | + # Wait for 10 seconds after the ageing time to make sure the FDB entry | ||
7921 | + # was not aged out | ||
7922 | + ageing_time=$(bridge_ageing_time_get br0) | ||
7923 | + sleep $((ageing_time + 10)) | ||
7924 | + | ||
7925 | + bridge fdb show brport $swp1 | grep -q de:ad:be:ef:13:37 | ||
7926 | + check_err $? "FDB entry was aged out when should not" | ||
7927 | + | ||
7928 | + $MZ $h2 -c 1 -p 64 -a $mac -t ip -q | ||
7929 | + | ||
7930 | + bridge fdb show brport $swp2 | grep -q de:ad:be:ef:13:37 | ||
7931 | + check_err $? "FDB entry did not roam when should" | ||
7932 | + | ||
7933 | + log_test "Externally learned FDB entry - ageing & roaming" | ||
7934 | + | ||
7935 | + bridge fdb del de:ad:be:ef:13:37 dev $swp2 master vlan 1 &> /dev/null | ||
7936 | + bridge fdb del de:ad:be:ef:13:37 dev $swp1 master vlan 1 &> /dev/null | ||
7937 | +} | ||
7938 | + | ||
7939 | trap cleanup EXIT | ||
7940 | |||
7941 | setup_prepare | ||
7942 | diff --git a/tools/testing/selftests/net/ip_defrag.c b/tools/testing/selftests/net/ip_defrag.c | ||
7943 | index 61ae2782388e9..5d56cc0838f62 100644 | ||
7944 | --- a/tools/testing/selftests/net/ip_defrag.c | ||
7945 | +++ b/tools/testing/selftests/net/ip_defrag.c | ||
7946 | @@ -203,6 +203,7 @@ static void send_udp_frags(int fd_raw, struct sockaddr *addr, | ||
7947 | { | ||
7948 | struct ip *iphdr = (struct ip *)ip_frame; | ||
7949 | struct ip6_hdr *ip6hdr = (struct ip6_hdr *)ip_frame; | ||
7950 | + const bool ipv4 = !ipv6; | ||
7951 | int res; | ||
7952 | int offset; | ||
7953 | int frag_len; | ||
7954 | @@ -239,19 +240,53 @@ static void send_udp_frags(int fd_raw, struct sockaddr *addr, | ||
7955 | iphdr->ip_sum = 0; | ||
7956 | } | ||
7957 | |||
7958 | + /* Occasionally test in-order fragments. */ | ||
7959 | + if (!cfg_overlap && (rand() % 100 < 15)) { | ||
7960 | + offset = 0; | ||
7961 | + while (offset < (UDP_HLEN + payload_len)) { | ||
7962 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7963 | + offset += max_frag_len; | ||
7964 | + } | ||
7965 | + return; | ||
7966 | + } | ||
7967 | + | ||
7968 | + /* Occasionally test IPv4 "runs" (see net/ipv4/ip_fragment.c) */ | ||
7969 | + if (ipv4 && !cfg_overlap && (rand() % 100 < 20) && | ||
7970 | + (payload_len > 9 * max_frag_len)) { | ||
7971 | + offset = 6 * max_frag_len; | ||
7972 | + while (offset < (UDP_HLEN + payload_len)) { | ||
7973 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7974 | + offset += max_frag_len; | ||
7975 | + } | ||
7976 | + offset = 3 * max_frag_len; | ||
7977 | + while (offset < 6 * max_frag_len) { | ||
7978 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7979 | + offset += max_frag_len; | ||
7980 | + } | ||
7981 | + offset = 0; | ||
7982 | + while (offset < 3 * max_frag_len) { | ||
7983 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7984 | + offset += max_frag_len; | ||
7985 | + } | ||
7986 | + return; | ||
7987 | + } | ||
7988 | + | ||
7989 | /* Odd fragments. */ | ||
7990 | offset = max_frag_len; | ||
7991 | while (offset < (UDP_HLEN + payload_len)) { | ||
7992 | send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7993 | + /* IPv4 ignores duplicates, so randomly send a duplicate. */ | ||
7994 | + if (ipv4 && (1 == rand() % 100)) | ||
7995 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
7996 | offset += 2 * max_frag_len; | ||
7997 | } | ||
7998 | |||
7999 | if (cfg_overlap) { | ||
8000 | /* Send an extra random fragment. */ | ||
8001 | - offset = rand() % (UDP_HLEN + payload_len - 1); | ||
8002 | - /* sendto() returns EINVAL if offset + frag_len is too small. */ | ||
8003 | if (ipv6) { | ||
8004 | struct ip6_frag *fraghdr = (struct ip6_frag *)(ip_frame + IP6_HLEN); | ||
8005 | + /* sendto() returns EINVAL if offset + frag_len is too small. */ | ||
8006 | + offset = rand() % (UDP_HLEN + payload_len - 1); | ||
8007 | frag_len = max_frag_len + rand() % 256; | ||
8008 | /* In IPv6 if !!(frag_len % 8), the fragment is dropped. */ | ||
8009 | frag_len &= ~0x7; | ||
8010 | @@ -259,13 +294,29 @@ static void send_udp_frags(int fd_raw, struct sockaddr *addr, | ||
8011 | ip6hdr->ip6_plen = htons(frag_len); | ||
8012 | frag_len += IP6_HLEN; | ||
8013 | } else { | ||
8014 | - frag_len = IP4_HLEN + UDP_HLEN + rand() % 256; | ||
8015 | + /* In IPv4, duplicates and some fragments completely inside | ||
8016 | + * previously sent fragments are dropped/ignored. So | ||
8017 | + * random offset and frag_len can result in a dropped | ||
8018 | + * fragment instead of a dropped queue/packet. So we | ||
8019 | + * hard-code offset and frag_len. | ||
8020 | + * | ||
8021 | + * See ade446403bfb ("net: ipv4: do not handle duplicate | ||
8022 | + * fragments as overlapping"). | ||
8023 | + */ | ||
8024 | + if (max_frag_len * 4 < payload_len || max_frag_len < 16) { | ||
8025 | + /* not enough payload to play with random offset and frag_len. */ | ||
8026 | + offset = 8; | ||
8027 | + frag_len = IP4_HLEN + UDP_HLEN + max_frag_len; | ||
8028 | + } else { | ||
8029 | + offset = rand() % (payload_len / 2); | ||
8030 | + frag_len = 2 * max_frag_len + 1 + rand() % 256; | ||
8031 | + } | ||
8032 | iphdr->ip_off = htons(offset / 8 | IP4_MF); | ||
8033 | iphdr->ip_len = htons(frag_len); | ||
8034 | } | ||
8035 | res = sendto(fd_raw, ip_frame, frag_len, 0, addr, alen); | ||
8036 | if (res < 0) | ||
8037 | - error(1, errno, "sendto overlap"); | ||
8038 | + error(1, errno, "sendto overlap: %d", frag_len); | ||
8039 | if (res != frag_len) | ||
8040 | error(1, 0, "sendto overlap: %d vs %d", (int)res, frag_len); | ||
8041 | frag_counter++; | ||
8042 | @@ -275,6 +326,9 @@ static void send_udp_frags(int fd_raw, struct sockaddr *addr, | ||
8043 | offset = 0; | ||
8044 | while (offset < (UDP_HLEN + payload_len)) { | ||
8045 | send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
8046 | + /* IPv4 ignores duplicates, so randomly send a duplicate. */ | ||
8047 | + if (ipv4 && (1 == rand() % 100)) | ||
8048 | + send_fragment(fd_raw, addr, alen, offset, ipv6); | ||
8049 | offset += 2 * max_frag_len; | ||
8050 | } | ||
8051 | } | ||
8052 | @@ -282,7 +336,11 @@ static void send_udp_frags(int fd_raw, struct sockaddr *addr, | ||
8053 | static void run_test(struct sockaddr *addr, socklen_t alen, bool ipv6) | ||
8054 | { | ||
8055 | int fd_tx_raw, fd_rx_udp; | ||
8056 | - struct timeval tv = { .tv_sec = 0, .tv_usec = 10 * 1000 }; | ||
8057 | + /* Frag queue timeout is set to one second in the calling script; | ||
8058 | + * socket timeout should be just a bit longer to avoid tests interfering | ||
8059 | + * with each other. | ||
8060 | + */ | ||
8061 | + struct timeval tv = { .tv_sec = 1, .tv_usec = 10 }; | ||
8062 | int idx; | ||
8063 | int min_frag_len = ipv6 ? 1280 : 8; | ||
8064 | |||
8065 | @@ -308,12 +366,32 @@ static void run_test(struct sockaddr *addr, socklen_t alen, bool ipv6) | ||
8066 | payload_len += (rand() % 4096)) { | ||
8067 | if (cfg_verbose) | ||
8068 | printf("payload_len: %d\n", payload_len); | ||
8069 | - max_frag_len = min_frag_len; | ||
8070 | - do { | ||
8071 | + | ||
8072 | + if (cfg_overlap) { | ||
8073 | + /* With overlaps, one send/receive pair below takes | ||
8074 | + * at least one second (== timeout) to run, so there | ||
8075 | + * is not enough test time to run a nested loop: | ||
8076 | + * the full overlap test takes 20-30 seconds. | ||
8077 | + */ | ||
8078 | + max_frag_len = min_frag_len + | ||
8079 | + rand() % (1500 - FRAG_HLEN - min_frag_len); | ||
8080 | send_udp_frags(fd_tx_raw, addr, alen, ipv6); | ||
8081 | recv_validate_udp(fd_rx_udp); | ||
8082 | - max_frag_len += 8 * (rand() % 8); | ||
8083 | - } while (max_frag_len < (1500 - FRAG_HLEN) && max_frag_len <= payload_len); | ||
8084 | + } else { | ||
8085 | + /* Without overlaps, each packet reassembly (== one | ||
8086 | + * send/receive pair below) takes very little time to | ||
8087 | + * run, so we can easily afford more thourough testing | ||
8088 | + * with a nested loop: the full non-overlap test takes | ||
8089 | + * less than one second). | ||
8090 | + */ | ||
8091 | + max_frag_len = min_frag_len; | ||
8092 | + do { | ||
8093 | + send_udp_frags(fd_tx_raw, addr, alen, ipv6); | ||
8094 | + recv_validate_udp(fd_rx_udp); | ||
8095 | + max_frag_len += 8 * (rand() % 8); | ||
8096 | + } while (max_frag_len < (1500 - FRAG_HLEN) && | ||
8097 | + max_frag_len <= payload_len); | ||
8098 | + } | ||
8099 | } | ||
8100 | |||
8101 | /* Cleanup. */ | ||
8102 | diff --git a/tools/testing/selftests/net/ip_defrag.sh b/tools/testing/selftests/net/ip_defrag.sh | ||
8103 | index f346727960449..7dd79a9efb177 100755 | ||
8104 | --- a/tools/testing/selftests/net/ip_defrag.sh | ||
8105 | +++ b/tools/testing/selftests/net/ip_defrag.sh | ||
8106 | @@ -11,10 +11,17 @@ readonly NETNS="ns-$(mktemp -u XXXXXX)" | ||
8107 | setup() { | ||
8108 | ip netns add "${NETNS}" | ||
8109 | ip -netns "${NETNS}" link set lo up | ||
8110 | + | ||
8111 | ip netns exec "${NETNS}" sysctl -w net.ipv4.ipfrag_high_thresh=9000000 >/dev/null 2>&1 | ||
8112 | ip netns exec "${NETNS}" sysctl -w net.ipv4.ipfrag_low_thresh=7000000 >/dev/null 2>&1 | ||
8113 | + ip netns exec "${NETNS}" sysctl -w net.ipv4.ipfrag_time=1 >/dev/null 2>&1 | ||
8114 | + | ||
8115 | ip netns exec "${NETNS}" sysctl -w net.ipv6.ip6frag_high_thresh=9000000 >/dev/null 2>&1 | ||
8116 | ip netns exec "${NETNS}" sysctl -w net.ipv6.ip6frag_low_thresh=7000000 >/dev/null 2>&1 | ||
8117 | + ip netns exec "${NETNS}" sysctl -w net.ipv6.ip6frag_time=1 >/dev/null 2>&1 | ||
8118 | + | ||
8119 | + # DST cache can get full with a lot of frags, with GC not keeping up with the test. | ||
8120 | + ip netns exec "${NETNS}" sysctl -w net.ipv6.route.max_size=65536 >/dev/null 2>&1 | ||
8121 | } | ||
8122 | |||
8123 | cleanup() { | ||
8124 | @@ -27,7 +34,6 @@ setup | ||
8125 | echo "ipv4 defrag" | ||
8126 | ip netns exec "${NETNS}" ./ip_defrag -4 | ||
8127 | |||
8128 | - | ||
8129 | echo "ipv4 defrag with overlaps" | ||
8130 | ip netns exec "${NETNS}" ./ip_defrag -4o | ||
8131 | |||
8132 | @@ -37,3 +43,4 @@ ip netns exec "${NETNS}" ./ip_defrag -6 | ||
8133 | echo "ipv6 defrag with overlaps" | ||
8134 | ip netns exec "${NETNS}" ./ip_defrag -6o | ||
8135 | |||
8136 | +echo "all tests done" | ||
8137 | diff --git a/tools/testing/selftests/tc-testing/tc-tests/actions/ife.json b/tools/testing/selftests/tc-testing/tc-tests/actions/ife.json | ||
8138 | index 637ea0219617f..0da3545cabdb6 100644 | ||
8139 | --- a/tools/testing/selftests/tc-testing/tc-tests/actions/ife.json | ||
8140 | +++ b/tools/testing/selftests/tc-testing/tc-tests/actions/ife.json | ||
8141 | @@ -17,7 +17,7 @@ | ||
8142 | "cmdUnderTest": "$TC actions add action ife encode allow mark pass index 2", | ||
8143 | "expExitCode": "0", | ||
8144 | "verifyCmd": "$TC actions get action ife index 2", | ||
8145 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow mark.*index 2", | ||
8146 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow mark.*index 2", | ||
8147 | "matchCount": "1", | ||
8148 | "teardown": [ | ||
8149 | "$TC actions flush action ife" | ||
8150 | @@ -41,7 +41,7 @@ | ||
8151 | "cmdUnderTest": "$TC actions add action ife encode use mark 10 pipe index 2", | ||
8152 | "expExitCode": "0", | ||
8153 | "verifyCmd": "$TC actions get action ife index 2", | ||
8154 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use mark.*index 2", | ||
8155 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use mark.*index 2", | ||
8156 | "matchCount": "1", | ||
8157 | "teardown": [ | ||
8158 | "$TC actions flush action ife" | ||
8159 | @@ -65,7 +65,7 @@ | ||
8160 | "cmdUnderTest": "$TC actions add action ife encode allow mark continue index 2", | ||
8161 | "expExitCode": "0", | ||
8162 | "verifyCmd": "$TC actions get action ife index 2", | ||
8163 | - "matchPattern": "action order [0-9]*: ife encode action continue.*type 0xED3E.*allow mark.*index 2", | ||
8164 | + "matchPattern": "action order [0-9]*: ife encode action continue.*type 0[xX]ED3E.*allow mark.*index 2", | ||
8165 | "matchCount": "1", | ||
8166 | "teardown": [ | ||
8167 | "$TC actions flush action ife" | ||
8168 | @@ -89,7 +89,7 @@ | ||
8169 | "cmdUnderTest": "$TC actions add action ife encode use mark 789 drop index 2", | ||
8170 | "expExitCode": "0", | ||
8171 | "verifyCmd": "$TC actions get action ife index 2", | ||
8172 | - "matchPattern": "action order [0-9]*: ife encode action drop.*type 0xED3E.*use mark 789.*index 2", | ||
8173 | + "matchPattern": "action order [0-9]*: ife encode action drop.*type 0[xX]ED3E.*use mark 789.*index 2", | ||
8174 | "matchCount": "1", | ||
8175 | "teardown": [ | ||
8176 | "$TC actions flush action ife" | ||
8177 | @@ -113,7 +113,7 @@ | ||
8178 | "cmdUnderTest": "$TC actions add action ife encode use mark 656768 reclassify index 2", | ||
8179 | "expExitCode": "0", | ||
8180 | "verifyCmd": "$TC actions get action ife index 2", | ||
8181 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*use mark 656768.*index 2", | ||
8182 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use mark 656768.*index 2", | ||
8183 | "matchCount": "1", | ||
8184 | "teardown": [ | ||
8185 | "$TC actions flush action ife" | ||
8186 | @@ -137,7 +137,7 @@ | ||
8187 | "cmdUnderTest": "$TC actions add action ife encode use mark 65 jump 1 index 2", | ||
8188 | "expExitCode": "0", | ||
8189 | "verifyCmd": "$TC actions get action ife index 2", | ||
8190 | - "matchPattern": "action order [0-9]*: ife encode action jump 1.*type 0xED3E.*use mark 65.*index 2", | ||
8191 | + "matchPattern": "action order [0-9]*: ife encode action jump 1.*type 0[xX]ED3E.*use mark 65.*index 2", | ||
8192 | "matchCount": "1", | ||
8193 | "teardown": [ | ||
8194 | "$TC actions flush action ife" | ||
8195 | @@ -161,7 +161,7 @@ | ||
8196 | "cmdUnderTest": "$TC actions add action ife encode use mark 4294967295 reclassify index 90", | ||
8197 | "expExitCode": "0", | ||
8198 | "verifyCmd": "$TC actions get action ife index 90", | ||
8199 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*use mark 4294967295.*index 90", | ||
8200 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use mark 4294967295.*index 90", | ||
8201 | "matchCount": "1", | ||
8202 | "teardown": [ | ||
8203 | "$TC actions flush action ife" | ||
8204 | @@ -185,7 +185,7 @@ | ||
8205 | "cmdUnderTest": "$TC actions add action ife encode use mark 4294967295999 pipe index 90", | ||
8206 | "expExitCode": "255", | ||
8207 | "verifyCmd": "$TC actions get action ife index 90", | ||
8208 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use mark 4294967295999.*index 90", | ||
8209 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use mark 4294967295999.*index 90", | ||
8210 | "matchCount": "0", | ||
8211 | "teardown": [] | ||
8212 | }, | ||
8213 | @@ -207,7 +207,7 @@ | ||
8214 | "cmdUnderTest": "$TC actions add action ife encode allow prio pass index 9", | ||
8215 | "expExitCode": "0", | ||
8216 | "verifyCmd": "$TC actions get action ife index 9", | ||
8217 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow prio.*index 9", | ||
8218 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow prio.*index 9", | ||
8219 | "matchCount": "1", | ||
8220 | "teardown": [ | ||
8221 | "$TC actions flush action ife" | ||
8222 | @@ -231,7 +231,7 @@ | ||
8223 | "cmdUnderTest": "$TC actions add action ife encode use prio 7 pipe index 9", | ||
8224 | "expExitCode": "0", | ||
8225 | "verifyCmd": "$TC actions get action ife index 9", | ||
8226 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use prio 7.*index 9", | ||
8227 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use prio 7.*index 9", | ||
8228 | "matchCount": "1", | ||
8229 | "teardown": [ | ||
8230 | "$TC actions flush action ife" | ||
8231 | @@ -255,7 +255,7 @@ | ||
8232 | "cmdUnderTest": "$TC actions add action ife encode use prio 3 continue index 9", | ||
8233 | "expExitCode": "0", | ||
8234 | "verifyCmd": "$TC actions get action ife index 9", | ||
8235 | - "matchPattern": "action order [0-9]*: ife encode action continue.*type 0xED3E.*use prio 3.*index 9", | ||
8236 | + "matchPattern": "action order [0-9]*: ife encode action continue.*type 0[xX]ED3E.*use prio 3.*index 9", | ||
8237 | "matchCount": "1", | ||
8238 | "teardown": [ | ||
8239 | "$TC actions flush action ife" | ||
8240 | @@ -279,7 +279,7 @@ | ||
8241 | "cmdUnderTest": "$TC actions add action ife encode allow prio drop index 9", | ||
8242 | "expExitCode": "0", | ||
8243 | "verifyCmd": "$TC actions get action ife index 9", | ||
8244 | - "matchPattern": "action order [0-9]*: ife encode action drop.*type 0xED3E.*allow prio.*index 9", | ||
8245 | + "matchPattern": "action order [0-9]*: ife encode action drop.*type 0[xX]ED3E.*allow prio.*index 9", | ||
8246 | "matchCount": "1", | ||
8247 | "teardown": [ | ||
8248 | "$TC actions flush action ife" | ||
8249 | @@ -303,7 +303,7 @@ | ||
8250 | "cmdUnderTest": "$TC actions add action ife encode use prio 998877 reclassify index 9", | ||
8251 | "expExitCode": "0", | ||
8252 | "verifyCmd": "$TC actions get action ife index 9", | ||
8253 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*use prio 998877.*index 9", | ||
8254 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use prio 998877.*index 9", | ||
8255 | "matchCount": "1", | ||
8256 | "teardown": [ | ||
8257 | "$TC actions flush action ife" | ||
8258 | @@ -327,7 +327,7 @@ | ||
8259 | "cmdUnderTest": "$TC actions add action ife encode use prio 998877 jump 10 index 9", | ||
8260 | "expExitCode": "0", | ||
8261 | "verifyCmd": "$TC actions get action ife index 9", | ||
8262 | - "matchPattern": "action order [0-9]*: ife encode action jump 10.*type 0xED3E.*use prio 998877.*index 9", | ||
8263 | + "matchPattern": "action order [0-9]*: ife encode action jump 10.*type 0[xX]ED3E.*use prio 998877.*index 9", | ||
8264 | "matchCount": "1", | ||
8265 | "teardown": [ | ||
8266 | "$TC actions flush action ife" | ||
8267 | @@ -351,7 +351,7 @@ | ||
8268 | "cmdUnderTest": "$TC actions add action ife encode use prio 4294967295 reclassify index 99", | ||
8269 | "expExitCode": "0", | ||
8270 | "verifyCmd": "$TC actions get action ife index 99", | ||
8271 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*use prio 4294967295.*index 99", | ||
8272 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use prio 4294967295.*index 99", | ||
8273 | "matchCount": "1", | ||
8274 | "teardown": [ | ||
8275 | "$TC actions flush action ife" | ||
8276 | @@ -375,7 +375,7 @@ | ||
8277 | "cmdUnderTest": "$TC actions add action ife encode use prio 4294967298 pipe index 99", | ||
8278 | "expExitCode": "255", | ||
8279 | "verifyCmd": "$TC actions get action ife index 99", | ||
8280 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use prio 4294967298.*index 99", | ||
8281 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use prio 4294967298.*index 99", | ||
8282 | "matchCount": "0", | ||
8283 | "teardown": [] | ||
8284 | }, | ||
8285 | @@ -397,7 +397,7 @@ | ||
8286 | "cmdUnderTest": "$TC actions add action ife encode allow tcindex pass index 1", | ||
8287 | "expExitCode": "0", | ||
8288 | "verifyCmd": "$TC actions get action ife index 1", | ||
8289 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow tcindex.*index 1", | ||
8290 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow tcindex.*index 1", | ||
8291 | "matchCount": "1", | ||
8292 | "teardown": [ | ||
8293 | "$TC actions flush action ife" | ||
8294 | @@ -421,7 +421,7 @@ | ||
8295 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 111 pipe index 1", | ||
8296 | "expExitCode": "0", | ||
8297 | "verifyCmd": "$TC actions get action ife index 1", | ||
8298 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use tcindex 111.*index 1", | ||
8299 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use tcindex 111.*index 1", | ||
8300 | "matchCount": "1", | ||
8301 | "teardown": [ | ||
8302 | "$TC actions flush action ife" | ||
8303 | @@ -445,7 +445,7 @@ | ||
8304 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 1 continue index 1", | ||
8305 | "expExitCode": "0", | ||
8306 | "verifyCmd": "$TC actions get action ife index 1", | ||
8307 | - "matchPattern": "action order [0-9]*: ife encode action continue.*type 0xED3E.*use tcindex 1.*index 1", | ||
8308 | + "matchPattern": "action order [0-9]*: ife encode action continue.*type 0[xX]ED3E.*use tcindex 1.*index 1", | ||
8309 | "matchCount": "1", | ||
8310 | "teardown": [ | ||
8311 | "$TC actions flush action ife" | ||
8312 | @@ -469,7 +469,7 @@ | ||
8313 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 1 continue index 1", | ||
8314 | "expExitCode": "0", | ||
8315 | "verifyCmd": "$TC actions get action ife index 1", | ||
8316 | - "matchPattern": "action order [0-9]*: ife encode action continue.*type 0xED3E.*use tcindex 1.*index 1", | ||
8317 | + "matchPattern": "action order [0-9]*: ife encode action continue.*type 0[xX]ED3E.*use tcindex 1.*index 1", | ||
8318 | "matchCount": "1", | ||
8319 | "teardown": [ | ||
8320 | "$TC actions flush action ife" | ||
8321 | @@ -493,7 +493,7 @@ | ||
8322 | "cmdUnderTest": "$TC actions add action ife encode allow tcindex drop index 77", | ||
8323 | "expExitCode": "0", | ||
8324 | "verifyCmd": "$TC actions get action ife index 77", | ||
8325 | - "matchPattern": "action order [0-9]*: ife encode action drop.*type 0xED3E.*allow tcindex.*index 77", | ||
8326 | + "matchPattern": "action order [0-9]*: ife encode action drop.*type 0[xX]ED3E.*allow tcindex.*index 77", | ||
8327 | "matchCount": "1", | ||
8328 | "teardown": [ | ||
8329 | "$TC actions flush action ife" | ||
8330 | @@ -517,7 +517,7 @@ | ||
8331 | "cmdUnderTest": "$TC actions add action ife encode allow tcindex reclassify index 77", | ||
8332 | "expExitCode": "0", | ||
8333 | "verifyCmd": "$TC actions get action ife index 77", | ||
8334 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*allow tcindex.*index 77", | ||
8335 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*allow tcindex.*index 77", | ||
8336 | "matchCount": "1", | ||
8337 | "teardown": [ | ||
8338 | "$TC actions flush action ife" | ||
8339 | @@ -541,7 +541,7 @@ | ||
8340 | "cmdUnderTest": "$TC actions add action ife encode allow tcindex jump 999 index 77", | ||
8341 | "expExitCode": "0", | ||
8342 | "verifyCmd": "$TC actions get action ife index 77", | ||
8343 | - "matchPattern": "action order [0-9]*: ife encode action jump 999.*type 0xED3E.*allow tcindex.*index 77", | ||
8344 | + "matchPattern": "action order [0-9]*: ife encode action jump 999.*type 0[xX]ED3E.*allow tcindex.*index 77", | ||
8345 | "matchCount": "1", | ||
8346 | "teardown": [ | ||
8347 | "$TC actions flush action ife" | ||
8348 | @@ -565,7 +565,7 @@ | ||
8349 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 65535 pass index 1", | ||
8350 | "expExitCode": "0", | ||
8351 | "verifyCmd": "$TC actions get action ife index 1", | ||
8352 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*use tcindex 65535.*index 1", | ||
8353 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*use tcindex 65535.*index 1", | ||
8354 | "matchCount": "1", | ||
8355 | "teardown": [ | ||
8356 | "$TC actions flush action ife" | ||
8357 | @@ -589,7 +589,7 @@ | ||
8358 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 65539 pipe index 1", | ||
8359 | "expExitCode": "255", | ||
8360 | "verifyCmd": "$TC actions get action ife index 1", | ||
8361 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*use tcindex 65539.*index 1", | ||
8362 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*use tcindex 65539.*index 1", | ||
8363 | "matchCount": "0", | ||
8364 | "teardown": [] | ||
8365 | }, | ||
8366 | @@ -611,7 +611,7 @@ | ||
8367 | "cmdUnderTest": "$TC actions add action ife encode allow mark src 00:11:22:33:44:55 pipe index 1", | ||
8368 | "expExitCode": "0", | ||
8369 | "verifyCmd": "$TC actions get action ife index 1", | ||
8370 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*allow mark src 00:11:22:33:44:55.*index 1", | ||
8371 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*allow mark src 00:11:22:33:44:55.*index 1", | ||
8372 | "matchCount": "1", | ||
8373 | "teardown": [ | ||
8374 | "$TC actions flush action ife" | ||
8375 | @@ -635,7 +635,7 @@ | ||
8376 | "cmdUnderTest": "$TC actions add action ife encode use prio 9876 dst 00:11:22:33:44:55 reclassify index 1", | ||
8377 | "expExitCode": "0", | ||
8378 | "verifyCmd": "$TC actions get action ife index 1", | ||
8379 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xED3E.*use prio 9876 dst 00:11:22:33:44:55.*index 1", | ||
8380 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ED3E.*use prio 9876 dst 00:11:22:33:44:55.*index 1", | ||
8381 | "matchCount": "1", | ||
8382 | "teardown": [ | ||
8383 | "$TC actions flush action ife" | ||
8384 | @@ -659,7 +659,7 @@ | ||
8385 | "cmdUnderTest": "$TC actions add action ife encode allow tcindex src 00:aa:bb:cc:dd:ee dst 00:11:22:33:44:55 pass index 11", | ||
8386 | "expExitCode": "0", | ||
8387 | "verifyCmd": "$TC actions get action ife index 11", | ||
8388 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow tcindex dst 00:11:22:33:44:55 src 00:aa:bb:cc:dd:ee .*index 11", | ||
8389 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow tcindex dst 00:11:22:33:44:55 src 00:aa:bb:cc:dd:ee .*index 11", | ||
8390 | "matchCount": "1", | ||
8391 | "teardown": [ | ||
8392 | "$TC actions flush action ife" | ||
8393 | @@ -683,7 +683,7 @@ | ||
8394 | "cmdUnderTest": "$TC actions add action ife encode use mark 7 type 0xfefe pass index 1", | ||
8395 | "expExitCode": "0", | ||
8396 | "verifyCmd": "$TC actions get action ife index 1", | ||
8397 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xFEFE.*use mark 7.*index 1", | ||
8398 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]FEFE.*use mark 7.*index 1", | ||
8399 | "matchCount": "1", | ||
8400 | "teardown": [ | ||
8401 | "$TC actions flush action ife" | ||
8402 | @@ -707,7 +707,7 @@ | ||
8403 | "cmdUnderTest": "$TC actions add action ife encode use prio 444 type 0xabba pipe index 21", | ||
8404 | "expExitCode": "0", | ||
8405 | "verifyCmd": "$TC actions get action ife index 21", | ||
8406 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xABBA.*use prio 444.*index 21", | ||
8407 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ABBA.*use prio 444.*index 21", | ||
8408 | "matchCount": "1", | ||
8409 | "teardown": [ | ||
8410 | "$TC actions flush action ife" | ||
8411 | @@ -731,7 +731,7 @@ | ||
8412 | "cmdUnderTest": "$TC actions add action ife encode use tcindex 5000 type 0xabcd reclassify index 21", | ||
8413 | "expExitCode": "0", | ||
8414 | "verifyCmd": "$TC actions get action ife index 21", | ||
8415 | - "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0xABCD.*use tcindex 5000.*index 21", | ||
8416 | + "matchPattern": "action order [0-9]*: ife encode action reclassify.*type 0[xX]ABCD.*use tcindex 5000.*index 21", | ||
8417 | "matchCount": "1", | ||
8418 | "teardown": [ | ||
8419 | "$TC actions flush action ife" | ||
8420 | @@ -739,7 +739,7 @@ | ||
8421 | }, | ||
8422 | { | ||
8423 | "id": "fac3", | ||
8424 | - "name": "Create valid ife encode action with index at 32-bit maximnum", | ||
8425 | + "name": "Create valid ife encode action with index at 32-bit maximum", | ||
8426 | "category": [ | ||
8427 | "actions", | ||
8428 | "ife" | ||
8429 | @@ -755,7 +755,7 @@ | ||
8430 | "cmdUnderTest": "$TC actions add action ife encode allow mark pass index 4294967295", | ||
8431 | "expExitCode": "0", | ||
8432 | "verifyCmd": "$TC actions get action ife index 4294967295", | ||
8433 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow mark.*index 4294967295", | ||
8434 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow mark.*index 4294967295", | ||
8435 | "matchCount": "1", | ||
8436 | "teardown": [ | ||
8437 | "$TC actions flush action ife" | ||
8438 | @@ -779,7 +779,7 @@ | ||
8439 | "cmdUnderTest": "$TC actions add action ife decode pass index 1", | ||
8440 | "expExitCode": "0", | ||
8441 | "verifyCmd": "$TC actions get action ife index 1", | ||
8442 | - "matchPattern": "action order [0-9]*: ife decode action pass.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8443 | + "matchPattern": "action order [0-9]*: ife decode action pass.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8444 | "matchCount": "1", | ||
8445 | "teardown": [ | ||
8446 | "$TC actions flush action ife" | ||
8447 | @@ -803,7 +803,7 @@ | ||
8448 | "cmdUnderTest": "$TC actions add action ife decode pipe index 1", | ||
8449 | "expExitCode": "0", | ||
8450 | "verifyCmd": "$TC actions get action ife index 1", | ||
8451 | - "matchPattern": "action order [0-9]*: ife decode action pipe.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8452 | + "matchPattern": "action order [0-9]*: ife decode action pipe.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8453 | "matchCount": "1", | ||
8454 | "teardown": [ | ||
8455 | "$TC actions flush action ife" | ||
8456 | @@ -827,7 +827,7 @@ | ||
8457 | "cmdUnderTest": "$TC actions add action ife decode continue index 1", | ||
8458 | "expExitCode": "0", | ||
8459 | "verifyCmd": "$TC actions get action ife index 1", | ||
8460 | - "matchPattern": "action order [0-9]*: ife decode action continue.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8461 | + "matchPattern": "action order [0-9]*: ife decode action continue.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8462 | "matchCount": "1", | ||
8463 | "teardown": [ | ||
8464 | "$TC actions flush action ife" | ||
8465 | @@ -851,7 +851,7 @@ | ||
8466 | "cmdUnderTest": "$TC actions add action ife decode drop index 1", | ||
8467 | "expExitCode": "0", | ||
8468 | "verifyCmd": "$TC actions get action ife index 1", | ||
8469 | - "matchPattern": "action order [0-9]*: ife decode action drop.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8470 | + "matchPattern": "action order [0-9]*: ife decode action drop.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8471 | "matchCount": "1", | ||
8472 | "teardown": [ | ||
8473 | "$TC actions flush action ife" | ||
8474 | @@ -875,7 +875,7 @@ | ||
8475 | "cmdUnderTest": "$TC actions add action ife decode reclassify index 1", | ||
8476 | "expExitCode": "0", | ||
8477 | "verifyCmd": "$TC actions get action ife index 1", | ||
8478 | - "matchPattern": "action order [0-9]*: ife decode action reclassify.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8479 | + "matchPattern": "action order [0-9]*: ife decode action reclassify.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8480 | "matchCount": "1", | ||
8481 | "teardown": [ | ||
8482 | "$TC actions flush action ife" | ||
8483 | @@ -899,7 +899,7 @@ | ||
8484 | "cmdUnderTest": "$TC actions add action ife decode jump 10 index 1", | ||
8485 | "expExitCode": "0", | ||
8486 | "verifyCmd": "$TC actions get action ife index 1", | ||
8487 | - "matchPattern": "action order [0-9]*: ife decode action jump 10.*type 0x0.*allow mark allow tcindex allow prio.*index 1", | ||
8488 | + "matchPattern": "action order [0-9]*: ife decode action jump 10.*type 0(x0)?.*allow mark allow tcindex allow prio.*index 1", | ||
8489 | "matchCount": "1", | ||
8490 | "teardown": [ | ||
8491 | "$TC actions flush action ife" | ||
8492 | @@ -923,7 +923,7 @@ | ||
8493 | "cmdUnderTest": "$TC actions add action ife encode allow mark pass index 4294967295999", | ||
8494 | "expExitCode": "255", | ||
8495 | "verifyCmd": "$TC actions get action ife index 4294967295999", | ||
8496 | - "matchPattern": "action order [0-9]*: ife encode action pass.*type 0xED3E.*allow mark.*index 4294967295999", | ||
8497 | + "matchPattern": "action order [0-9]*: ife encode action pass.*type 0[xX]ED3E.*allow mark.*index 4294967295999", | ||
8498 | "matchCount": "0", | ||
8499 | "teardown": [] | ||
8500 | }, | ||
8501 | @@ -945,7 +945,7 @@ | ||
8502 | "cmdUnderTest": "$TC actions add action ife encode allow mark kuka index 4", | ||
8503 | "expExitCode": "255", | ||
8504 | "verifyCmd": "$TC actions get action ife index 4", | ||
8505 | - "matchPattern": "action order [0-9]*: ife encode action kuka.*type 0xED3E.*allow mark.*index 4", | ||
8506 | + "matchPattern": "action order [0-9]*: ife encode action kuka.*type 0[xX]ED3E.*allow mark.*index 4", | ||
8507 | "matchCount": "0", | ||
8508 | "teardown": [] | ||
8509 | }, | ||
8510 | @@ -967,7 +967,7 @@ | ||
8511 | "cmdUnderTest": "$TC actions add action ife encode allow prio pipe index 4 cookie aabbccddeeff112233445566778800a1", | ||
8512 | "expExitCode": "0", | ||
8513 | "verifyCmd": "$TC actions get action ife index 4", | ||
8514 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*allow prio.*index 4.*cookie aabbccddeeff112233445566778800a1", | ||
8515 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*allow prio.*index 4.*cookie aabbccddeeff112233445566778800a1", | ||
8516 | "matchCount": "1", | ||
8517 | "teardown": [ | ||
8518 | "$TC actions flush action ife" | ||
8519 | @@ -991,7 +991,7 @@ | ||
8520 | "cmdUnderTest": "$TC actions add action ife encode allow foo pipe index 4", | ||
8521 | "expExitCode": "255", | ||
8522 | "verifyCmd": "$TC actions get action ife index 4", | ||
8523 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0xED3E.*allow foo.*index 4", | ||
8524 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]ED3E.*allow foo.*index 4", | ||
8525 | "matchCount": "0", | ||
8526 | "teardown": [] | ||
8527 | }, | ||
8528 | @@ -1013,7 +1013,7 @@ | ||
8529 | "cmdUnderTest": "$TC actions add action ife encode allow prio type 70000 pipe index 4", | ||
8530 | "expExitCode": "255", | ||
8531 | "verifyCmd": "$TC actions get action ife index 4", | ||
8532 | - "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0x11170.*allow prio.*index 4", | ||
8533 | + "matchPattern": "action order [0-9]*: ife encode action pipe.*type 0[xX]11170.*allow prio.*index 4", | ||
8534 | "matchCount": "0", | ||
8535 | "teardown": [] | ||
8536 | }, | ||
8537 | diff --git a/tools/testing/selftests/tc-testing/tc-tests/actions/tunnel_key.json b/tools/testing/selftests/tc-testing/tc-tests/actions/tunnel_key.json | ||
8538 | index 10b2d894e4362..e7e15a7336b6d 100644 | ||
8539 | --- a/tools/testing/selftests/tc-testing/tc-tests/actions/tunnel_key.json | ||
8540 | +++ b/tools/testing/selftests/tc-testing/tc-tests/actions/tunnel_key.json | ||
8541 | @@ -81,35 +81,6 @@ | ||
8542 | ] | ||
8543 | ] | ||
8544 | }, | ||
8545 | - { | ||
8546 | - "id": "ba4e", | ||
8547 | - "name": "Add tunnel_key set action with missing mandatory id parameter", | ||
8548 | - "category": [ | ||
8549 | - "actions", | ||
8550 | - "tunnel_key" | ||
8551 | - ], | ||
8552 | - "setup": [ | ||
8553 | - [ | ||
8554 | - "$TC actions flush action tunnel_key", | ||
8555 | - 0, | ||
8556 | - 1, | ||
8557 | - 255 | ||
8558 | - ] | ||
8559 | - ], | ||
8560 | - "cmdUnderTest": "$TC actions add action tunnel_key set src_ip 10.10.10.1 dst_ip 20.20.20.2", | ||
8561 | - "expExitCode": "255", | ||
8562 | - "verifyCmd": "$TC actions list action tunnel_key", | ||
8563 | - "matchPattern": "action order [0-9]+: tunnel_key set.*src_ip 10.10.10.1.*dst_ip 20.20.20.2", | ||
8564 | - "matchCount": "0", | ||
8565 | - "teardown": [ | ||
8566 | - [ | ||
8567 | - "$TC actions flush action tunnel_key", | ||
8568 | - 0, | ||
8569 | - 1, | ||
8570 | - 255 | ||
8571 | - ] | ||
8572 | - ] | ||
8573 | - }, | ||
8574 | { | ||
8575 | "id": "a5e0", | ||
8576 | "name": "Add tunnel_key set action with invalid src_ip parameter", | ||
8577 | @@ -634,7 +605,7 @@ | ||
8578 | "cmdUnderTest": "$TC actions add action tunnel_key set src_ip 10.10.10.1 dst_ip 10.10.10.2 id 7 index 4 cookie aa11bb22cc33dd44ee55ff66aa11b1b2", | ||
8579 | "expExitCode": "0", | ||
8580 | "verifyCmd": "$TC actions get action tunnel_key index 4", | ||
8581 | - "matchPattern": "action order [0-9]+: tunnel_key.*set.*src_ip 10.10.10.1.*dst_ip 10.10.10.2.*key_id 7.*dst_port 0.*csum pipe.*index 4 ref.*cookie aa11bb22cc33dd44ee55ff66aa11b1b2", | ||
8582 | + "matchPattern": "action order [0-9]+: tunnel_key.*set.*src_ip 10.10.10.1.*dst_ip 10.10.10.2.*key_id 7.*csum pipe.*index 4 ref.*cookie aa11bb22cc33dd44ee55ff66aa11b1b2", | ||
8583 | "matchCount": "1", | ||
8584 | "teardown": [ | ||
8585 | "$TC actions flush action tunnel_key" |