Contents of /trunk/kernel-alx/patches-4.9/0225-4.9.126-all-fixes.patch
Parent Directory | Revision Log
Revision 3224 -
(show annotations)
(download)
Tue Sep 11 10:04:15 2018 UTC (6 years ago) by niro
File size: 75480 byte(s)
Tue Sep 11 10:04:15 2018 UTC (6 years ago) by niro
File size: 75480 byte(s)
-linux-4.9.226
1 | diff --git a/Makefile b/Makefile |
2 | index aef09ca7a924..b26481fef3f0 100644 |
3 | --- a/Makefile |
4 | +++ b/Makefile |
5 | @@ -1,6 +1,6 @@ |
6 | VERSION = 4 |
7 | PATCHLEVEL = 9 |
8 | -SUBLEVEL = 125 |
9 | +SUBLEVEL = 126 |
10 | EXTRAVERSION = |
11 | NAME = Roaring Lionus |
12 | |
13 | diff --git a/arch/alpha/kernel/osf_sys.c b/arch/alpha/kernel/osf_sys.c |
14 | index 4f95577b0180..6e0d549eb3bb 100644 |
15 | --- a/arch/alpha/kernel/osf_sys.c |
16 | +++ b/arch/alpha/kernel/osf_sys.c |
17 | @@ -526,24 +526,19 @@ SYSCALL_DEFINE4(osf_mount, unsigned long, typenr, const char __user *, path, |
18 | SYSCALL_DEFINE1(osf_utsname, char __user *, name) |
19 | { |
20 | int error; |
21 | + char tmp[5 * 32]; |
22 | |
23 | down_read(&uts_sem); |
24 | - error = -EFAULT; |
25 | - if (copy_to_user(name + 0, utsname()->sysname, 32)) |
26 | - goto out; |
27 | - if (copy_to_user(name + 32, utsname()->nodename, 32)) |
28 | - goto out; |
29 | - if (copy_to_user(name + 64, utsname()->release, 32)) |
30 | - goto out; |
31 | - if (copy_to_user(name + 96, utsname()->version, 32)) |
32 | - goto out; |
33 | - if (copy_to_user(name + 128, utsname()->machine, 32)) |
34 | - goto out; |
35 | + memcpy(tmp + 0 * 32, utsname()->sysname, 32); |
36 | + memcpy(tmp + 1 * 32, utsname()->nodename, 32); |
37 | + memcpy(tmp + 2 * 32, utsname()->release, 32); |
38 | + memcpy(tmp + 3 * 32, utsname()->version, 32); |
39 | + memcpy(tmp + 4 * 32, utsname()->machine, 32); |
40 | + up_read(&uts_sem); |
41 | |
42 | - error = 0; |
43 | - out: |
44 | - up_read(&uts_sem); |
45 | - return error; |
46 | + if (copy_to_user(name, tmp, sizeof(tmp))) |
47 | + return -EFAULT; |
48 | + return 0; |
49 | } |
50 | |
51 | SYSCALL_DEFINE0(getpagesize) |
52 | @@ -561,24 +556,22 @@ SYSCALL_DEFINE0(getdtablesize) |
53 | */ |
54 | SYSCALL_DEFINE2(osf_getdomainname, char __user *, name, int, namelen) |
55 | { |
56 | - unsigned len; |
57 | - int i; |
58 | + int len, err = 0; |
59 | + char *kname; |
60 | + char tmp[32]; |
61 | |
62 | - if (!access_ok(VERIFY_WRITE, name, namelen)) |
63 | - return -EFAULT; |
64 | - |
65 | - len = namelen; |
66 | - if (len > 32) |
67 | - len = 32; |
68 | + if (namelen < 0 || namelen > 32) |
69 | + namelen = 32; |
70 | |
71 | down_read(&uts_sem); |
72 | - for (i = 0; i < len; ++i) { |
73 | - __put_user(utsname()->domainname[i], name + i); |
74 | - if (utsname()->domainname[i] == '\0') |
75 | - break; |
76 | - } |
77 | + kname = utsname()->domainname; |
78 | + len = strnlen(kname, namelen); |
79 | + len = min(len + 1, namelen); |
80 | + memcpy(tmp, kname, len); |
81 | up_read(&uts_sem); |
82 | |
83 | + if (copy_to_user(name, tmp, len)) |
84 | + return -EFAULT; |
85 | return 0; |
86 | } |
87 | |
88 | @@ -741,13 +734,14 @@ SYSCALL_DEFINE3(osf_sysinfo, int, command, char __user *, buf, long, count) |
89 | }; |
90 | unsigned long offset; |
91 | const char *res; |
92 | - long len, err = -EINVAL; |
93 | + long len; |
94 | + char tmp[__NEW_UTS_LEN + 1]; |
95 | |
96 | offset = command-1; |
97 | if (offset >= ARRAY_SIZE(sysinfo_table)) { |
98 | /* Digital UNIX has a few unpublished interfaces here */ |
99 | printk("sysinfo(%d)", command); |
100 | - goto out; |
101 | + return -EINVAL; |
102 | } |
103 | |
104 | down_read(&uts_sem); |
105 | @@ -755,13 +749,11 @@ SYSCALL_DEFINE3(osf_sysinfo, int, command, char __user *, buf, long, count) |
106 | len = strlen(res)+1; |
107 | if ((unsigned long)len > (unsigned long)count) |
108 | len = count; |
109 | - if (copy_to_user(buf, res, len)) |
110 | - err = -EFAULT; |
111 | - else |
112 | - err = 0; |
113 | + memcpy(tmp, res, len); |
114 | up_read(&uts_sem); |
115 | - out: |
116 | - return err; |
117 | + if (copy_to_user(buf, tmp, len)) |
118 | + return -EFAULT; |
119 | + return 0; |
120 | } |
121 | |
122 | SYSCALL_DEFINE5(osf_getsysinfo, unsigned long, op, void __user *, buffer, |
123 | diff --git a/arch/arm/boot/dts/tegra30-cardhu.dtsi b/arch/arm/boot/dts/tegra30-cardhu.dtsi |
124 | index f11012bb58cc..cfcf5dcb51a8 100644 |
125 | --- a/arch/arm/boot/dts/tegra30-cardhu.dtsi |
126 | +++ b/arch/arm/boot/dts/tegra30-cardhu.dtsi |
127 | @@ -205,6 +205,7 @@ |
128 | #address-cells = <1>; |
129 | #size-cells = <0>; |
130 | reg = <0x70>; |
131 | + reset-gpio = <&gpio TEGRA_GPIO(BB, 0) GPIO_ACTIVE_LOW>; |
132 | }; |
133 | }; |
134 | |
135 | diff --git a/arch/powerpc/include/asm/fadump.h b/arch/powerpc/include/asm/fadump.h |
136 | index 0031806475f0..f93238ad06bd 100644 |
137 | --- a/arch/powerpc/include/asm/fadump.h |
138 | +++ b/arch/powerpc/include/asm/fadump.h |
139 | @@ -190,9 +190,6 @@ struct fadump_crash_info_header { |
140 | struct cpumask online_mask; |
141 | }; |
142 | |
143 | -/* Crash memory ranges */ |
144 | -#define INIT_CRASHMEM_RANGES (INIT_MEMBLOCK_REGIONS + 2) |
145 | - |
146 | struct fad_crash_memory_ranges { |
147 | unsigned long long base; |
148 | unsigned long long size; |
149 | diff --git a/arch/powerpc/kernel/fadump.c b/arch/powerpc/kernel/fadump.c |
150 | index 93a6eeba3ace..e3acf5c3480e 100644 |
151 | --- a/arch/powerpc/kernel/fadump.c |
152 | +++ b/arch/powerpc/kernel/fadump.c |
153 | @@ -35,6 +35,7 @@ |
154 | #include <linux/crash_dump.h> |
155 | #include <linux/kobject.h> |
156 | #include <linux/sysfs.h> |
157 | +#include <linux/slab.h> |
158 | |
159 | #include <asm/page.h> |
160 | #include <asm/prom.h> |
161 | @@ -48,8 +49,10 @@ static struct fadump_mem_struct fdm; |
162 | static const struct fadump_mem_struct *fdm_active; |
163 | |
164 | static DEFINE_MUTEX(fadump_mutex); |
165 | -struct fad_crash_memory_ranges crash_memory_ranges[INIT_CRASHMEM_RANGES]; |
166 | +struct fad_crash_memory_ranges *crash_memory_ranges; |
167 | +int crash_memory_ranges_size; |
168 | int crash_mem_ranges; |
169 | +int max_crash_mem_ranges; |
170 | |
171 | /* Scan the Firmware Assisted dump configuration details. */ |
172 | int __init early_init_dt_scan_fw_dump(unsigned long node, |
173 | @@ -731,38 +734,88 @@ static int __init process_fadump(const struct fadump_mem_struct *fdm_active) |
174 | return 0; |
175 | } |
176 | |
177 | -static inline void fadump_add_crash_memory(unsigned long long base, |
178 | - unsigned long long end) |
179 | +static void free_crash_memory_ranges(void) |
180 | +{ |
181 | + kfree(crash_memory_ranges); |
182 | + crash_memory_ranges = NULL; |
183 | + crash_memory_ranges_size = 0; |
184 | + max_crash_mem_ranges = 0; |
185 | +} |
186 | + |
187 | +/* |
188 | + * Allocate or reallocate crash memory ranges array in incremental units |
189 | + * of PAGE_SIZE. |
190 | + */ |
191 | +static int allocate_crash_memory_ranges(void) |
192 | +{ |
193 | + struct fad_crash_memory_ranges *new_array; |
194 | + u64 new_size; |
195 | + |
196 | + new_size = crash_memory_ranges_size + PAGE_SIZE; |
197 | + pr_debug("Allocating %llu bytes of memory for crash memory ranges\n", |
198 | + new_size); |
199 | + |
200 | + new_array = krealloc(crash_memory_ranges, new_size, GFP_KERNEL); |
201 | + if (new_array == NULL) { |
202 | + pr_err("Insufficient memory for setting up crash memory ranges\n"); |
203 | + free_crash_memory_ranges(); |
204 | + return -ENOMEM; |
205 | + } |
206 | + |
207 | + crash_memory_ranges = new_array; |
208 | + crash_memory_ranges_size = new_size; |
209 | + max_crash_mem_ranges = (new_size / |
210 | + sizeof(struct fad_crash_memory_ranges)); |
211 | + return 0; |
212 | +} |
213 | + |
214 | +static inline int fadump_add_crash_memory(unsigned long long base, |
215 | + unsigned long long end) |
216 | { |
217 | if (base == end) |
218 | - return; |
219 | + return 0; |
220 | + |
221 | + if (crash_mem_ranges == max_crash_mem_ranges) { |
222 | + int ret; |
223 | + |
224 | + ret = allocate_crash_memory_ranges(); |
225 | + if (ret) |
226 | + return ret; |
227 | + } |
228 | |
229 | pr_debug("crash_memory_range[%d] [%#016llx-%#016llx], %#llx bytes\n", |
230 | crash_mem_ranges, base, end - 1, (end - base)); |
231 | crash_memory_ranges[crash_mem_ranges].base = base; |
232 | crash_memory_ranges[crash_mem_ranges].size = end - base; |
233 | crash_mem_ranges++; |
234 | + return 0; |
235 | } |
236 | |
237 | -static void fadump_exclude_reserved_area(unsigned long long start, |
238 | +static int fadump_exclude_reserved_area(unsigned long long start, |
239 | unsigned long long end) |
240 | { |
241 | unsigned long long ra_start, ra_end; |
242 | + int ret = 0; |
243 | |
244 | ra_start = fw_dump.reserve_dump_area_start; |
245 | ra_end = ra_start + fw_dump.reserve_dump_area_size; |
246 | |
247 | if ((ra_start < end) && (ra_end > start)) { |
248 | if ((start < ra_start) && (end > ra_end)) { |
249 | - fadump_add_crash_memory(start, ra_start); |
250 | - fadump_add_crash_memory(ra_end, end); |
251 | + ret = fadump_add_crash_memory(start, ra_start); |
252 | + if (ret) |
253 | + return ret; |
254 | + |
255 | + ret = fadump_add_crash_memory(ra_end, end); |
256 | } else if (start < ra_start) { |
257 | - fadump_add_crash_memory(start, ra_start); |
258 | + ret = fadump_add_crash_memory(start, ra_start); |
259 | } else if (ra_end < end) { |
260 | - fadump_add_crash_memory(ra_end, end); |
261 | + ret = fadump_add_crash_memory(ra_end, end); |
262 | } |
263 | } else |
264 | - fadump_add_crash_memory(start, end); |
265 | + ret = fadump_add_crash_memory(start, end); |
266 | + |
267 | + return ret; |
268 | } |
269 | |
270 | static int fadump_init_elfcore_header(char *bufp) |
271 | @@ -802,10 +855,11 @@ static int fadump_init_elfcore_header(char *bufp) |
272 | * Traverse through memblock structure and setup crash memory ranges. These |
273 | * ranges will be used create PT_LOAD program headers in elfcore header. |
274 | */ |
275 | -static void fadump_setup_crash_memory_ranges(void) |
276 | +static int fadump_setup_crash_memory_ranges(void) |
277 | { |
278 | struct memblock_region *reg; |
279 | unsigned long long start, end; |
280 | + int ret; |
281 | |
282 | pr_debug("Setup crash memory ranges.\n"); |
283 | crash_mem_ranges = 0; |
284 | @@ -816,7 +870,9 @@ static void fadump_setup_crash_memory_ranges(void) |
285 | * specified during fadump registration. We need to create a separate |
286 | * program header for this chunk with the correct offset. |
287 | */ |
288 | - fadump_add_crash_memory(RMA_START, fw_dump.boot_memory_size); |
289 | + ret = fadump_add_crash_memory(RMA_START, fw_dump.boot_memory_size); |
290 | + if (ret) |
291 | + return ret; |
292 | |
293 | for_each_memblock(memory, reg) { |
294 | start = (unsigned long long)reg->base; |
295 | @@ -825,8 +881,12 @@ static void fadump_setup_crash_memory_ranges(void) |
296 | start = fw_dump.boot_memory_size; |
297 | |
298 | /* add this range excluding the reserved dump area. */ |
299 | - fadump_exclude_reserved_area(start, end); |
300 | + ret = fadump_exclude_reserved_area(start, end); |
301 | + if (ret) |
302 | + return ret; |
303 | } |
304 | + |
305 | + return 0; |
306 | } |
307 | |
308 | /* |
309 | @@ -950,6 +1010,7 @@ static void register_fadump(void) |
310 | { |
311 | unsigned long addr; |
312 | void *vaddr; |
313 | + int ret; |
314 | |
315 | /* |
316 | * If no memory is reserved then we can not register for firmware- |
317 | @@ -958,7 +1019,9 @@ static void register_fadump(void) |
318 | if (!fw_dump.reserve_dump_area_size) |
319 | return; |
320 | |
321 | - fadump_setup_crash_memory_ranges(); |
322 | + ret = fadump_setup_crash_memory_ranges(); |
323 | + if (ret) |
324 | + return ret; |
325 | |
326 | addr = be64_to_cpu(fdm.rmr_region.destination_address) + be64_to_cpu(fdm.rmr_region.source_len); |
327 | /* Initialize fadump crash info header. */ |
328 | @@ -1036,6 +1099,7 @@ void fadump_cleanup(void) |
329 | } else if (fw_dump.dump_registered) { |
330 | /* Un-register Firmware-assisted dump if it was registered. */ |
331 | fadump_unregister_dump(&fdm); |
332 | + free_crash_memory_ranges(); |
333 | } |
334 | } |
335 | |
336 | diff --git a/arch/powerpc/platforms/powernv/pci-ioda.c b/arch/powerpc/platforms/powernv/pci-ioda.c |
337 | index 9ed90c502005..f52cc6fd4290 100644 |
338 | --- a/arch/powerpc/platforms/powernv/pci-ioda.c |
339 | +++ b/arch/powerpc/platforms/powernv/pci-ioda.c |
340 | @@ -3124,12 +3124,49 @@ static void pnv_pci_ioda_create_dbgfs(void) |
341 | #endif /* CONFIG_DEBUG_FS */ |
342 | } |
343 | |
344 | +static void pnv_pci_enable_bridge(struct pci_bus *bus) |
345 | +{ |
346 | + struct pci_dev *dev = bus->self; |
347 | + struct pci_bus *child; |
348 | + |
349 | + /* Empty bus ? bail */ |
350 | + if (list_empty(&bus->devices)) |
351 | + return; |
352 | + |
353 | + /* |
354 | + * If there's a bridge associated with that bus enable it. This works |
355 | + * around races in the generic code if the enabling is done during |
356 | + * parallel probing. This can be removed once those races have been |
357 | + * fixed. |
358 | + */ |
359 | + if (dev) { |
360 | + int rc = pci_enable_device(dev); |
361 | + if (rc) |
362 | + pci_err(dev, "Error enabling bridge (%d)\n", rc); |
363 | + pci_set_master(dev); |
364 | + } |
365 | + |
366 | + /* Perform the same to child busses */ |
367 | + list_for_each_entry(child, &bus->children, node) |
368 | + pnv_pci_enable_bridge(child); |
369 | +} |
370 | + |
371 | +static void pnv_pci_enable_bridges(void) |
372 | +{ |
373 | + struct pci_controller *hose; |
374 | + |
375 | + list_for_each_entry(hose, &hose_list, list_node) |
376 | + pnv_pci_enable_bridge(hose->bus); |
377 | +} |
378 | + |
379 | static void pnv_pci_ioda_fixup(void) |
380 | { |
381 | pnv_pci_ioda_setup_PEs(); |
382 | pnv_pci_ioda_setup_iommu_api(); |
383 | pnv_pci_ioda_create_dbgfs(); |
384 | |
385 | + pnv_pci_enable_bridges(); |
386 | + |
387 | #ifdef CONFIG_EEH |
388 | eeh_init(); |
389 | eeh_addr_cache_build(); |
390 | diff --git a/arch/powerpc/platforms/pseries/ras.c b/arch/powerpc/platforms/pseries/ras.c |
391 | index 904a677208d1..34989ce43147 100644 |
392 | --- a/arch/powerpc/platforms/pseries/ras.c |
393 | +++ b/arch/powerpc/platforms/pseries/ras.c |
394 | @@ -346,7 +346,7 @@ static struct rtas_error_log *fwnmi_get_errinfo(struct pt_regs *regs) |
395 | } |
396 | |
397 | savep = __va(regs->gpr[3]); |
398 | - regs->gpr[3] = savep[0]; /* restore original r3 */ |
399 | + regs->gpr[3] = be64_to_cpu(savep[0]); /* restore original r3 */ |
400 | |
401 | /* If it isn't an extended log we can use the per cpu 64bit buffer */ |
402 | h = (struct rtas_error_log *)&savep[1]; |
403 | diff --git a/arch/sparc/kernel/sys_sparc_32.c b/arch/sparc/kernel/sys_sparc_32.c |
404 | index 646988d4c1a3..740f43b9b541 100644 |
405 | --- a/arch/sparc/kernel/sys_sparc_32.c |
406 | +++ b/arch/sparc/kernel/sys_sparc_32.c |
407 | @@ -201,23 +201,27 @@ SYSCALL_DEFINE5(rt_sigaction, int, sig, |
408 | |
409 | asmlinkage long sys_getdomainname(char __user *name, int len) |
410 | { |
411 | - int nlen, err; |
412 | - |
413 | + int nlen, err; |
414 | + char tmp[__NEW_UTS_LEN + 1]; |
415 | + |
416 | if (len < 0) |
417 | return -EINVAL; |
418 | |
419 | - down_read(&uts_sem); |
420 | - |
421 | + down_read(&uts_sem); |
422 | + |
423 | nlen = strlen(utsname()->domainname) + 1; |
424 | err = -EINVAL; |
425 | if (nlen > len) |
426 | - goto out; |
427 | + goto out_unlock; |
428 | + memcpy(tmp, utsname()->domainname, nlen); |
429 | |
430 | - err = -EFAULT; |
431 | - if (!copy_to_user(name, utsname()->domainname, nlen)) |
432 | - err = 0; |
433 | + up_read(&uts_sem); |
434 | |
435 | -out: |
436 | + if (copy_to_user(name, tmp, nlen)) |
437 | + return -EFAULT; |
438 | + return 0; |
439 | + |
440 | +out_unlock: |
441 | up_read(&uts_sem); |
442 | return err; |
443 | } |
444 | diff --git a/arch/sparc/kernel/sys_sparc_64.c b/arch/sparc/kernel/sys_sparc_64.c |
445 | index 02e05e221b94..ebecbc927460 100644 |
446 | --- a/arch/sparc/kernel/sys_sparc_64.c |
447 | +++ b/arch/sparc/kernel/sys_sparc_64.c |
448 | @@ -524,23 +524,27 @@ extern void check_pending(int signum); |
449 | |
450 | SYSCALL_DEFINE2(getdomainname, char __user *, name, int, len) |
451 | { |
452 | - int nlen, err; |
453 | + int nlen, err; |
454 | + char tmp[__NEW_UTS_LEN + 1]; |
455 | |
456 | if (len < 0) |
457 | return -EINVAL; |
458 | |
459 | - down_read(&uts_sem); |
460 | - |
461 | + down_read(&uts_sem); |
462 | + |
463 | nlen = strlen(utsname()->domainname) + 1; |
464 | err = -EINVAL; |
465 | if (nlen > len) |
466 | - goto out; |
467 | + goto out_unlock; |
468 | + memcpy(tmp, utsname()->domainname, nlen); |
469 | + |
470 | + up_read(&uts_sem); |
471 | |
472 | - err = -EFAULT; |
473 | - if (!copy_to_user(name, utsname()->domainname, nlen)) |
474 | - err = 0; |
475 | + if (copy_to_user(name, tmp, nlen)) |
476 | + return -EFAULT; |
477 | + return 0; |
478 | |
479 | -out: |
480 | +out_unlock: |
481 | up_read(&uts_sem); |
482 | return err; |
483 | } |
484 | diff --git a/arch/x86/kernel/kexec-bzimage64.c b/arch/x86/kernel/kexec-bzimage64.c |
485 | index 3407b148c240..490f9be3fda2 100644 |
486 | --- a/arch/x86/kernel/kexec-bzimage64.c |
487 | +++ b/arch/x86/kernel/kexec-bzimage64.c |
488 | @@ -529,7 +529,7 @@ static int bzImage64_cleanup(void *loader_data) |
489 | static int bzImage64_verify_sig(const char *kernel, unsigned long kernel_len) |
490 | { |
491 | return verify_pefile_signature(kernel, kernel_len, |
492 | - NULL, |
493 | + VERIFY_USE_SECONDARY_KEYRING, |
494 | VERIFYING_KEXEC_PE_SIGNATURE); |
495 | } |
496 | #endif |
497 | diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c |
498 | index 8e4ac0a91309..8888d894bf39 100644 |
499 | --- a/arch/x86/kvm/vmx.c |
500 | +++ b/arch/x86/kvm/vmx.c |
501 | @@ -198,12 +198,14 @@ static enum vmx_l1d_flush_state __read_mostly vmentry_l1d_flush_param = VMENTER_ |
502 | |
503 | static const struct { |
504 | const char *option; |
505 | - enum vmx_l1d_flush_state cmd; |
506 | + bool for_parse; |
507 | } vmentry_l1d_param[] = { |
508 | - {"auto", VMENTER_L1D_FLUSH_AUTO}, |
509 | - {"never", VMENTER_L1D_FLUSH_NEVER}, |
510 | - {"cond", VMENTER_L1D_FLUSH_COND}, |
511 | - {"always", VMENTER_L1D_FLUSH_ALWAYS}, |
512 | + [VMENTER_L1D_FLUSH_AUTO] = {"auto", true}, |
513 | + [VMENTER_L1D_FLUSH_NEVER] = {"never", true}, |
514 | + [VMENTER_L1D_FLUSH_COND] = {"cond", true}, |
515 | + [VMENTER_L1D_FLUSH_ALWAYS] = {"always", true}, |
516 | + [VMENTER_L1D_FLUSH_EPT_DISABLED] = {"EPT disabled", false}, |
517 | + [VMENTER_L1D_FLUSH_NOT_REQUIRED] = {"not required", false}, |
518 | }; |
519 | |
520 | #define L1D_CACHE_ORDER 4 |
521 | @@ -287,8 +289,9 @@ static int vmentry_l1d_flush_parse(const char *s) |
522 | |
523 | if (s) { |
524 | for (i = 0; i < ARRAY_SIZE(vmentry_l1d_param); i++) { |
525 | - if (sysfs_streq(s, vmentry_l1d_param[i].option)) |
526 | - return vmentry_l1d_param[i].cmd; |
527 | + if (vmentry_l1d_param[i].for_parse && |
528 | + sysfs_streq(s, vmentry_l1d_param[i].option)) |
529 | + return i; |
530 | } |
531 | } |
532 | return -EINVAL; |
533 | @@ -298,13 +301,13 @@ static int vmentry_l1d_flush_set(const char *s, const struct kernel_param *kp) |
534 | { |
535 | int l1tf, ret; |
536 | |
537 | - if (!boot_cpu_has(X86_BUG_L1TF)) |
538 | - return 0; |
539 | - |
540 | l1tf = vmentry_l1d_flush_parse(s); |
541 | if (l1tf < 0) |
542 | return l1tf; |
543 | |
544 | + if (!boot_cpu_has(X86_BUG_L1TF)) |
545 | + return 0; |
546 | + |
547 | /* |
548 | * Has vmx_init() run already? If not then this is the pre init |
549 | * parameter parsing. In that case just store the value and let |
550 | @@ -324,6 +327,9 @@ static int vmentry_l1d_flush_set(const char *s, const struct kernel_param *kp) |
551 | |
552 | static int vmentry_l1d_flush_get(char *s, const struct kernel_param *kp) |
553 | { |
554 | + if (WARN_ON_ONCE(l1tf_vmx_mitigation >= ARRAY_SIZE(vmentry_l1d_param))) |
555 | + return sprintf(s, "???\n"); |
556 | + |
557 | return sprintf(s, "%s\n", vmentry_l1d_param[l1tf_vmx_mitigation].option); |
558 | } |
559 | |
560 | diff --git a/arch/xtensa/include/asm/cacheasm.h b/arch/xtensa/include/asm/cacheasm.h |
561 | index 2041abb10a23..34545ecfdd6b 100644 |
562 | --- a/arch/xtensa/include/asm/cacheasm.h |
563 | +++ b/arch/xtensa/include/asm/cacheasm.h |
564 | @@ -31,16 +31,32 @@ |
565 | * |
566 | */ |
567 | |
568 | - .macro __loop_cache_all ar at insn size line_width |
569 | |
570 | - movi \ar, 0 |
571 | + .macro __loop_cache_unroll ar at insn size line_width max_immed |
572 | + |
573 | + .if (1 << (\line_width)) > (\max_immed) |
574 | + .set _reps, 1 |
575 | + .elseif (2 << (\line_width)) > (\max_immed) |
576 | + .set _reps, 2 |
577 | + .else |
578 | + .set _reps, 4 |
579 | + .endif |
580 | + |
581 | + __loopi \ar, \at, \size, (_reps << (\line_width)) |
582 | + .set _index, 0 |
583 | + .rep _reps |
584 | + \insn \ar, _index << (\line_width) |
585 | + .set _index, _index + 1 |
586 | + .endr |
587 | + __endla \ar, \at, _reps << (\line_width) |
588 | + |
589 | + .endm |
590 | + |
591 | |
592 | - __loopi \ar, \at, \size, (4 << (\line_width)) |
593 | - \insn \ar, 0 << (\line_width) |
594 | - \insn \ar, 1 << (\line_width) |
595 | - \insn \ar, 2 << (\line_width) |
596 | - \insn \ar, 3 << (\line_width) |
597 | - __endla \ar, \at, 4 << (\line_width) |
598 | + .macro __loop_cache_all ar at insn size line_width max_immed |
599 | + |
600 | + movi \ar, 0 |
601 | + __loop_cache_unroll \ar, \at, \insn, \size, \line_width, \max_immed |
602 | |
603 | .endm |
604 | |
605 | @@ -57,14 +73,9 @@ |
606 | .endm |
607 | |
608 | |
609 | - .macro __loop_cache_page ar at insn line_width |
610 | + .macro __loop_cache_page ar at insn line_width max_immed |
611 | |
612 | - __loopi \ar, \at, PAGE_SIZE, 4 << (\line_width) |
613 | - \insn \ar, 0 << (\line_width) |
614 | - \insn \ar, 1 << (\line_width) |
615 | - \insn \ar, 2 << (\line_width) |
616 | - \insn \ar, 3 << (\line_width) |
617 | - __endla \ar, \at, 4 << (\line_width) |
618 | + __loop_cache_unroll \ar, \at, \insn, PAGE_SIZE, \line_width, \max_immed |
619 | |
620 | .endm |
621 | |
622 | @@ -72,7 +83,8 @@ |
623 | .macro ___unlock_dcache_all ar at |
624 | |
625 | #if XCHAL_DCACHE_LINE_LOCKABLE && XCHAL_DCACHE_SIZE |
626 | - __loop_cache_all \ar \at diu XCHAL_DCACHE_SIZE XCHAL_DCACHE_LINEWIDTH |
627 | + __loop_cache_all \ar \at diu XCHAL_DCACHE_SIZE \ |
628 | + XCHAL_DCACHE_LINEWIDTH 240 |
629 | #endif |
630 | |
631 | .endm |
632 | @@ -81,7 +93,8 @@ |
633 | .macro ___unlock_icache_all ar at |
634 | |
635 | #if XCHAL_ICACHE_LINE_LOCKABLE && XCHAL_ICACHE_SIZE |
636 | - __loop_cache_all \ar \at iiu XCHAL_ICACHE_SIZE XCHAL_ICACHE_LINEWIDTH |
637 | + __loop_cache_all \ar \at iiu XCHAL_ICACHE_SIZE \ |
638 | + XCHAL_ICACHE_LINEWIDTH 240 |
639 | #endif |
640 | |
641 | .endm |
642 | @@ -90,7 +103,8 @@ |
643 | .macro ___flush_invalidate_dcache_all ar at |
644 | |
645 | #if XCHAL_DCACHE_SIZE |
646 | - __loop_cache_all \ar \at diwbi XCHAL_DCACHE_SIZE XCHAL_DCACHE_LINEWIDTH |
647 | + __loop_cache_all \ar \at diwbi XCHAL_DCACHE_SIZE \ |
648 | + XCHAL_DCACHE_LINEWIDTH 240 |
649 | #endif |
650 | |
651 | .endm |
652 | @@ -99,7 +113,8 @@ |
653 | .macro ___flush_dcache_all ar at |
654 | |
655 | #if XCHAL_DCACHE_SIZE |
656 | - __loop_cache_all \ar \at diwb XCHAL_DCACHE_SIZE XCHAL_DCACHE_LINEWIDTH |
657 | + __loop_cache_all \ar \at diwb XCHAL_DCACHE_SIZE \ |
658 | + XCHAL_DCACHE_LINEWIDTH 240 |
659 | #endif |
660 | |
661 | .endm |
662 | @@ -108,8 +123,8 @@ |
663 | .macro ___invalidate_dcache_all ar at |
664 | |
665 | #if XCHAL_DCACHE_SIZE |
666 | - __loop_cache_all \ar \at dii __stringify(DCACHE_WAY_SIZE) \ |
667 | - XCHAL_DCACHE_LINEWIDTH |
668 | + __loop_cache_all \ar \at dii XCHAL_DCACHE_SIZE \ |
669 | + XCHAL_DCACHE_LINEWIDTH 1020 |
670 | #endif |
671 | |
672 | .endm |
673 | @@ -118,8 +133,8 @@ |
674 | .macro ___invalidate_icache_all ar at |
675 | |
676 | #if XCHAL_ICACHE_SIZE |
677 | - __loop_cache_all \ar \at iii __stringify(ICACHE_WAY_SIZE) \ |
678 | - XCHAL_ICACHE_LINEWIDTH |
679 | + __loop_cache_all \ar \at iii XCHAL_ICACHE_SIZE \ |
680 | + XCHAL_ICACHE_LINEWIDTH 1020 |
681 | #endif |
682 | |
683 | .endm |
684 | @@ -166,7 +181,7 @@ |
685 | .macro ___flush_invalidate_dcache_page ar as |
686 | |
687 | #if XCHAL_DCACHE_SIZE |
688 | - __loop_cache_page \ar \as dhwbi XCHAL_DCACHE_LINEWIDTH |
689 | + __loop_cache_page \ar \as dhwbi XCHAL_DCACHE_LINEWIDTH 1020 |
690 | #endif |
691 | |
692 | .endm |
693 | @@ -175,7 +190,7 @@ |
694 | .macro ___flush_dcache_page ar as |
695 | |
696 | #if XCHAL_DCACHE_SIZE |
697 | - __loop_cache_page \ar \as dhwb XCHAL_DCACHE_LINEWIDTH |
698 | + __loop_cache_page \ar \as dhwb XCHAL_DCACHE_LINEWIDTH 1020 |
699 | #endif |
700 | |
701 | .endm |
702 | @@ -184,7 +199,7 @@ |
703 | .macro ___invalidate_dcache_page ar as |
704 | |
705 | #if XCHAL_DCACHE_SIZE |
706 | - __loop_cache_page \ar \as dhi XCHAL_DCACHE_LINEWIDTH |
707 | + __loop_cache_page \ar \as dhi XCHAL_DCACHE_LINEWIDTH 1020 |
708 | #endif |
709 | |
710 | .endm |
711 | @@ -193,7 +208,7 @@ |
712 | .macro ___invalidate_icache_page ar as |
713 | |
714 | #if XCHAL_ICACHE_SIZE |
715 | - __loop_cache_page \ar \as ihi XCHAL_ICACHE_LINEWIDTH |
716 | + __loop_cache_page \ar \as ihi XCHAL_ICACHE_LINEWIDTH 1020 |
717 | #endif |
718 | |
719 | .endm |
720 | diff --git a/certs/system_keyring.c b/certs/system_keyring.c |
721 | index 50979d6dcecd..247665054691 100644 |
722 | --- a/certs/system_keyring.c |
723 | +++ b/certs/system_keyring.c |
724 | @@ -14,6 +14,7 @@ |
725 | #include <linux/sched.h> |
726 | #include <linux/cred.h> |
727 | #include <linux/err.h> |
728 | +#include <linux/verification.h> |
729 | #include <keys/asymmetric-type.h> |
730 | #include <keys/system_keyring.h> |
731 | #include <crypto/pkcs7.h> |
732 | @@ -207,7 +208,7 @@ int verify_pkcs7_signature(const void *data, size_t len, |
733 | |
734 | if (!trusted_keys) { |
735 | trusted_keys = builtin_trusted_keys; |
736 | - } else if (trusted_keys == (void *)1UL) { |
737 | + } else if (trusted_keys == VERIFY_USE_SECONDARY_KEYRING) { |
738 | #ifdef CONFIG_SECONDARY_TRUSTED_KEYRING |
739 | trusted_keys = secondary_trusted_keys; |
740 | #else |
741 | diff --git a/crypto/asymmetric_keys/pkcs7_key_type.c b/crypto/asymmetric_keys/pkcs7_key_type.c |
742 | index 1063b644efcd..b2aa925a84bc 100644 |
743 | --- a/crypto/asymmetric_keys/pkcs7_key_type.c |
744 | +++ b/crypto/asymmetric_keys/pkcs7_key_type.c |
745 | @@ -62,7 +62,7 @@ static int pkcs7_preparse(struct key_preparsed_payload *prep) |
746 | |
747 | return verify_pkcs7_signature(NULL, 0, |
748 | prep->data, prep->datalen, |
749 | - (void *)1UL, usage, |
750 | + VERIFY_USE_SECONDARY_KEYRING, usage, |
751 | pkcs7_view_content, prep); |
752 | } |
753 | |
754 | diff --git a/drivers/crypto/caam/jr.c b/drivers/crypto/caam/jr.c |
755 | index 9e7f28122bb7..6d475a2e298c 100644 |
756 | --- a/drivers/crypto/caam/jr.c |
757 | +++ b/drivers/crypto/caam/jr.c |
758 | @@ -189,7 +189,8 @@ static void caam_jr_dequeue(unsigned long devarg) |
759 | BUG_ON(CIRC_CNT(head, tail + i, JOBR_DEPTH) <= 0); |
760 | |
761 | /* Unmap just-run descriptor so we can post-process */ |
762 | - dma_unmap_single(dev, jrp->outring[hw_idx].desc, |
763 | + dma_unmap_single(dev, |
764 | + caam_dma_to_cpu(jrp->outring[hw_idx].desc), |
765 | jrp->entinfo[sw_idx].desc_size, |
766 | DMA_TO_DEVICE); |
767 | |
768 | diff --git a/drivers/crypto/vmx/aes_cbc.c b/drivers/crypto/vmx/aes_cbc.c |
769 | index 46131701c378..92e116365272 100644 |
770 | --- a/drivers/crypto/vmx/aes_cbc.c |
771 | +++ b/drivers/crypto/vmx/aes_cbc.c |
772 | @@ -111,24 +111,23 @@ static int p8_aes_cbc_encrypt(struct blkcipher_desc *desc, |
773 | ret = crypto_blkcipher_encrypt(&fallback_desc, dst, src, |
774 | nbytes); |
775 | } else { |
776 | - preempt_disable(); |
777 | - pagefault_disable(); |
778 | - enable_kernel_vsx(); |
779 | - |
780 | blkcipher_walk_init(&walk, dst, src, nbytes); |
781 | ret = blkcipher_walk_virt(desc, &walk); |
782 | while ((nbytes = walk.nbytes)) { |
783 | + preempt_disable(); |
784 | + pagefault_disable(); |
785 | + enable_kernel_vsx(); |
786 | aes_p8_cbc_encrypt(walk.src.virt.addr, |
787 | walk.dst.virt.addr, |
788 | nbytes & AES_BLOCK_MASK, |
789 | &ctx->enc_key, walk.iv, 1); |
790 | + disable_kernel_vsx(); |
791 | + pagefault_enable(); |
792 | + preempt_enable(); |
793 | + |
794 | nbytes &= AES_BLOCK_SIZE - 1; |
795 | ret = blkcipher_walk_done(desc, &walk, nbytes); |
796 | } |
797 | - |
798 | - disable_kernel_vsx(); |
799 | - pagefault_enable(); |
800 | - preempt_enable(); |
801 | } |
802 | |
803 | return ret; |
804 | @@ -152,24 +151,23 @@ static int p8_aes_cbc_decrypt(struct blkcipher_desc *desc, |
805 | ret = crypto_blkcipher_decrypt(&fallback_desc, dst, src, |
806 | nbytes); |
807 | } else { |
808 | - preempt_disable(); |
809 | - pagefault_disable(); |
810 | - enable_kernel_vsx(); |
811 | - |
812 | blkcipher_walk_init(&walk, dst, src, nbytes); |
813 | ret = blkcipher_walk_virt(desc, &walk); |
814 | while ((nbytes = walk.nbytes)) { |
815 | + preempt_disable(); |
816 | + pagefault_disable(); |
817 | + enable_kernel_vsx(); |
818 | aes_p8_cbc_encrypt(walk.src.virt.addr, |
819 | walk.dst.virt.addr, |
820 | nbytes & AES_BLOCK_MASK, |
821 | &ctx->dec_key, walk.iv, 0); |
822 | + disable_kernel_vsx(); |
823 | + pagefault_enable(); |
824 | + preempt_enable(); |
825 | + |
826 | nbytes &= AES_BLOCK_SIZE - 1; |
827 | ret = blkcipher_walk_done(desc, &walk, nbytes); |
828 | } |
829 | - |
830 | - disable_kernel_vsx(); |
831 | - pagefault_enable(); |
832 | - preempt_enable(); |
833 | } |
834 | |
835 | return ret; |
836 | diff --git a/drivers/crypto/vmx/aes_xts.c b/drivers/crypto/vmx/aes_xts.c |
837 | index 24353ec336c5..52e7ae05ae1f 100644 |
838 | --- a/drivers/crypto/vmx/aes_xts.c |
839 | +++ b/drivers/crypto/vmx/aes_xts.c |
840 | @@ -123,32 +123,39 @@ static int p8_aes_xts_crypt(struct blkcipher_desc *desc, |
841 | ret = enc ? crypto_blkcipher_encrypt(&fallback_desc, dst, src, nbytes) : |
842 | crypto_blkcipher_decrypt(&fallback_desc, dst, src, nbytes); |
843 | } else { |
844 | + blkcipher_walk_init(&walk, dst, src, nbytes); |
845 | + |
846 | + ret = blkcipher_walk_virt(desc, &walk); |
847 | + |
848 | preempt_disable(); |
849 | pagefault_disable(); |
850 | enable_kernel_vsx(); |
851 | |
852 | - blkcipher_walk_init(&walk, dst, src, nbytes); |
853 | - |
854 | - ret = blkcipher_walk_virt(desc, &walk); |
855 | iv = walk.iv; |
856 | memset(tweak, 0, AES_BLOCK_SIZE); |
857 | aes_p8_encrypt(iv, tweak, &ctx->tweak_key); |
858 | |
859 | + disable_kernel_vsx(); |
860 | + pagefault_enable(); |
861 | + preempt_enable(); |
862 | + |
863 | while ((nbytes = walk.nbytes)) { |
864 | + preempt_disable(); |
865 | + pagefault_disable(); |
866 | + enable_kernel_vsx(); |
867 | if (enc) |
868 | aes_p8_xts_encrypt(walk.src.virt.addr, walk.dst.virt.addr, |
869 | nbytes & AES_BLOCK_MASK, &ctx->enc_key, NULL, tweak); |
870 | else |
871 | aes_p8_xts_decrypt(walk.src.virt.addr, walk.dst.virt.addr, |
872 | nbytes & AES_BLOCK_MASK, &ctx->dec_key, NULL, tweak); |
873 | + disable_kernel_vsx(); |
874 | + pagefault_enable(); |
875 | + preempt_enable(); |
876 | |
877 | nbytes &= AES_BLOCK_SIZE - 1; |
878 | ret = blkcipher_walk_done(desc, &walk, nbytes); |
879 | } |
880 | - |
881 | - disable_kernel_vsx(); |
882 | - pagefault_enable(); |
883 | - preempt_enable(); |
884 | } |
885 | return ret; |
886 | } |
887 | diff --git a/drivers/gpu/drm/i915/i915_gem_userptr.c b/drivers/gpu/drm/i915/i915_gem_userptr.c |
888 | index c6f780f5abc9..555fd47c1831 100644 |
889 | --- a/drivers/gpu/drm/i915/i915_gem_userptr.c |
890 | +++ b/drivers/gpu/drm/i915/i915_gem_userptr.c |
891 | @@ -778,6 +778,9 @@ i915_gem_userptr_ioctl(struct drm_device *dev, void *data, struct drm_file *file |
892 | I915_USERPTR_UNSYNCHRONIZED)) |
893 | return -EINVAL; |
894 | |
895 | + if (!args->user_size) |
896 | + return -EINVAL; |
897 | + |
898 | if (offset_in_page(args->user_ptr | args->user_size)) |
899 | return -EINVAL; |
900 | |
901 | diff --git a/drivers/iio/frequency/ad9523.c b/drivers/iio/frequency/ad9523.c |
902 | index 99eba524f6dd..1642b55f70da 100644 |
903 | --- a/drivers/iio/frequency/ad9523.c |
904 | +++ b/drivers/iio/frequency/ad9523.c |
905 | @@ -508,7 +508,7 @@ static ssize_t ad9523_store(struct device *dev, |
906 | return ret; |
907 | |
908 | if (!state) |
909 | - return 0; |
910 | + return len; |
911 | |
912 | mutex_lock(&indio_dev->mlock); |
913 | switch ((u32)this_attr->address) { |
914 | @@ -642,7 +642,7 @@ static int ad9523_read_raw(struct iio_dev *indio_dev, |
915 | code = (AD9523_CLK_DIST_DIV_PHASE_REV(ret) * 3141592) / |
916 | AD9523_CLK_DIST_DIV_REV(ret); |
917 | *val = code / 1000000; |
918 | - *val2 = (code % 1000000) * 10; |
919 | + *val2 = code % 1000000; |
920 | return IIO_VAL_INT_PLUS_MICRO; |
921 | default: |
922 | return -EINVAL; |
923 | diff --git a/drivers/infiniband/sw/rxe/rxe_comp.c b/drivers/infiniband/sw/rxe/rxe_comp.c |
924 | index 6c5e29db88e3..df15b6d7b645 100644 |
925 | --- a/drivers/infiniband/sw/rxe/rxe_comp.c |
926 | +++ b/drivers/infiniband/sw/rxe/rxe_comp.c |
927 | @@ -273,6 +273,7 @@ static inline enum comp_state check_ack(struct rxe_qp *qp, |
928 | case IB_OPCODE_RC_RDMA_READ_RESPONSE_MIDDLE: |
929 | if (wqe->wr.opcode != IB_WR_RDMA_READ && |
930 | wqe->wr.opcode != IB_WR_RDMA_READ_WITH_INV) { |
931 | + wqe->status = IB_WC_FATAL_ERR; |
932 | return COMPST_ERROR; |
933 | } |
934 | reset_retry_counters(qp); |
935 | diff --git a/drivers/infiniband/ulp/srpt/ib_srpt.c b/drivers/infiniband/ulp/srpt/ib_srpt.c |
936 | index 876f438aa048..fe7c6ec67d98 100644 |
937 | --- a/drivers/infiniband/ulp/srpt/ib_srpt.c |
938 | +++ b/drivers/infiniband/ulp/srpt/ib_srpt.c |
939 | @@ -1701,8 +1701,7 @@ static bool srpt_close_ch(struct srpt_rdma_ch *ch) |
940 | int ret; |
941 | |
942 | if (!srpt_set_ch_state(ch, CH_DRAINING)) { |
943 | - pr_debug("%s-%d: already closed\n", ch->sess_name, |
944 | - ch->qp->qp_num); |
945 | + pr_debug("%s: already closed\n", ch->sess_name); |
946 | return false; |
947 | } |
948 | |
949 | diff --git a/drivers/iommu/dmar.c b/drivers/iommu/dmar.c |
950 | index 8c53748a769d..63110fbbb410 100644 |
951 | --- a/drivers/iommu/dmar.c |
952 | +++ b/drivers/iommu/dmar.c |
953 | @@ -1328,8 +1328,8 @@ void qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr, |
954 | qi_submit_sync(&desc, iommu); |
955 | } |
956 | |
957 | -void qi_flush_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 qdep, |
958 | - u64 addr, unsigned mask) |
959 | +void qi_flush_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 pfsid, |
960 | + u16 qdep, u64 addr, unsigned mask) |
961 | { |
962 | struct qi_desc desc; |
963 | |
964 | @@ -1344,7 +1344,7 @@ void qi_flush_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 qdep, |
965 | qdep = 0; |
966 | |
967 | desc.low = QI_DEV_IOTLB_SID(sid) | QI_DEV_IOTLB_QDEP(qdep) | |
968 | - QI_DIOTLB_TYPE; |
969 | + QI_DIOTLB_TYPE | QI_DEV_IOTLB_PFSID(pfsid); |
970 | |
971 | qi_submit_sync(&desc, iommu); |
972 | } |
973 | diff --git a/drivers/iommu/intel-iommu.c b/drivers/iommu/intel-iommu.c |
974 | index 1612d3a22d42..2558a381e118 100644 |
975 | --- a/drivers/iommu/intel-iommu.c |
976 | +++ b/drivers/iommu/intel-iommu.c |
977 | @@ -421,6 +421,7 @@ struct device_domain_info { |
978 | struct list_head global; /* link to global list */ |
979 | u8 bus; /* PCI bus number */ |
980 | u8 devfn; /* PCI devfn number */ |
981 | + u16 pfsid; /* SRIOV physical function source ID */ |
982 | u8 pasid_supported:3; |
983 | u8 pasid_enabled:1; |
984 | u8 pri_supported:1; |
985 | @@ -1511,6 +1512,20 @@ static void iommu_enable_dev_iotlb(struct device_domain_info *info) |
986 | return; |
987 | |
988 | pdev = to_pci_dev(info->dev); |
989 | + /* For IOMMU that supports device IOTLB throttling (DIT), we assign |
990 | + * PFSID to the invalidation desc of a VF such that IOMMU HW can gauge |
991 | + * queue depth at PF level. If DIT is not set, PFSID will be treated as |
992 | + * reserved, which should be set to 0. |
993 | + */ |
994 | + if (!ecap_dit(info->iommu->ecap)) |
995 | + info->pfsid = 0; |
996 | + else { |
997 | + struct pci_dev *pf_pdev; |
998 | + |
999 | + /* pdev will be returned if device is not a vf */ |
1000 | + pf_pdev = pci_physfn(pdev); |
1001 | + info->pfsid = PCI_DEVID(pf_pdev->bus->number, pf_pdev->devfn); |
1002 | + } |
1003 | |
1004 | #ifdef CONFIG_INTEL_IOMMU_SVM |
1005 | /* The PCIe spec, in its wisdom, declares that the behaviour of |
1006 | @@ -1576,7 +1591,8 @@ static void iommu_flush_dev_iotlb(struct dmar_domain *domain, |
1007 | |
1008 | sid = info->bus << 8 | info->devfn; |
1009 | qdep = info->ats_qdep; |
1010 | - qi_flush_dev_iotlb(info->iommu, sid, qdep, addr, mask); |
1011 | + qi_flush_dev_iotlb(info->iommu, sid, info->pfsid, |
1012 | + qdep, addr, mask); |
1013 | } |
1014 | spin_unlock_irqrestore(&device_domain_lock, flags); |
1015 | } |
1016 | diff --git a/drivers/mailbox/mailbox-xgene-slimpro.c b/drivers/mailbox/mailbox-xgene-slimpro.c |
1017 | index dd2afbca51c9..26d2f89db4d9 100644 |
1018 | --- a/drivers/mailbox/mailbox-xgene-slimpro.c |
1019 | +++ b/drivers/mailbox/mailbox-xgene-slimpro.c |
1020 | @@ -195,9 +195,9 @@ static int slimpro_mbox_probe(struct platform_device *pdev) |
1021 | platform_set_drvdata(pdev, ctx); |
1022 | |
1023 | regs = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
1024 | - mb_base = devm_ioremap(&pdev->dev, regs->start, resource_size(regs)); |
1025 | - if (!mb_base) |
1026 | - return -ENOMEM; |
1027 | + mb_base = devm_ioremap_resource(&pdev->dev, regs); |
1028 | + if (IS_ERR(mb_base)) |
1029 | + return PTR_ERR(mb_base); |
1030 | |
1031 | /* Setup mailbox links */ |
1032 | for (i = 0; i < MBOX_CNT; i++) { |
1033 | diff --git a/drivers/md/bcache/writeback.c b/drivers/md/bcache/writeback.c |
1034 | index bb7aa31c2a08..cdf388d40274 100644 |
1035 | --- a/drivers/md/bcache/writeback.c |
1036 | +++ b/drivers/md/bcache/writeback.c |
1037 | @@ -456,8 +456,10 @@ static int bch_writeback_thread(void *arg) |
1038 | * data on cache. BCACHE_DEV_DETACHING flag is set in |
1039 | * bch_cached_dev_detach(). |
1040 | */ |
1041 | - if (test_bit(BCACHE_DEV_DETACHING, &dc->disk.flags)) |
1042 | + if (test_bit(BCACHE_DEV_DETACHING, &dc->disk.flags)) { |
1043 | + up_write(&dc->writeback_lock); |
1044 | break; |
1045 | + } |
1046 | } |
1047 | |
1048 | up_write(&dc->writeback_lock); |
1049 | diff --git a/drivers/md/dm-cache-metadata.c b/drivers/md/dm-cache-metadata.c |
1050 | index 6937ca42be8c..a184c9830ca5 100644 |
1051 | --- a/drivers/md/dm-cache-metadata.c |
1052 | +++ b/drivers/md/dm-cache-metadata.c |
1053 | @@ -344,7 +344,7 @@ static int __write_initial_superblock(struct dm_cache_metadata *cmd) |
1054 | disk_super->version = cpu_to_le32(MAX_CACHE_VERSION); |
1055 | memset(disk_super->policy_name, 0, sizeof(disk_super->policy_name)); |
1056 | memset(disk_super->policy_version, 0, sizeof(disk_super->policy_version)); |
1057 | - disk_super->policy_hint_size = 0; |
1058 | + disk_super->policy_hint_size = cpu_to_le32(0); |
1059 | |
1060 | __copy_sm_root(cmd, disk_super); |
1061 | |
1062 | @@ -659,6 +659,7 @@ static int __commit_transaction(struct dm_cache_metadata *cmd, |
1063 | disk_super->policy_version[0] = cpu_to_le32(cmd->policy_version[0]); |
1064 | disk_super->policy_version[1] = cpu_to_le32(cmd->policy_version[1]); |
1065 | disk_super->policy_version[2] = cpu_to_le32(cmd->policy_version[2]); |
1066 | + disk_super->policy_hint_size = cpu_to_le32(cmd->policy_hint_size); |
1067 | |
1068 | disk_super->read_hits = cpu_to_le32(cmd->stats.read_hits); |
1069 | disk_super->read_misses = cpu_to_le32(cmd->stats.read_misses); |
1070 | diff --git a/drivers/md/dm-thin.c b/drivers/md/dm-thin.c |
1071 | index 0f0374a4ac6e..a952ad890f32 100644 |
1072 | --- a/drivers/md/dm-thin.c |
1073 | +++ b/drivers/md/dm-thin.c |
1074 | @@ -2518,6 +2518,8 @@ static void set_pool_mode(struct pool *pool, enum pool_mode new_mode) |
1075 | case PM_WRITE: |
1076 | if (old_mode != new_mode) |
1077 | notify_of_pool_mode_change(pool, "write"); |
1078 | + if (old_mode == PM_OUT_OF_DATA_SPACE) |
1079 | + cancel_delayed_work_sync(&pool->no_space_timeout); |
1080 | pool->out_of_data_space = false; |
1081 | pool->pf.error_if_no_space = pt->requested_pf.error_if_no_space; |
1082 | dm_pool_metadata_read_write(pool->pmd); |
1083 | diff --git a/drivers/mfd/hi655x-pmic.c b/drivers/mfd/hi655x-pmic.c |
1084 | index 0fc62995695b..11347a3e6d40 100644 |
1085 | --- a/drivers/mfd/hi655x-pmic.c |
1086 | +++ b/drivers/mfd/hi655x-pmic.c |
1087 | @@ -49,7 +49,7 @@ static struct regmap_config hi655x_regmap_config = { |
1088 | .reg_bits = 32, |
1089 | .reg_stride = HI655X_STRIDE, |
1090 | .val_bits = 8, |
1091 | - .max_register = HI655X_BUS_ADDR(0xFFF), |
1092 | + .max_register = HI655X_BUS_ADDR(0x400) - HI655X_STRIDE, |
1093 | }; |
1094 | |
1095 | static struct resource pwrkey_resources[] = { |
1096 | diff --git a/drivers/misc/cxl/main.c b/drivers/misc/cxl/main.c |
1097 | index cc1706a92ace..a078d49485d8 100644 |
1098 | --- a/drivers/misc/cxl/main.c |
1099 | +++ b/drivers/misc/cxl/main.c |
1100 | @@ -293,7 +293,7 @@ int cxl_adapter_context_get(struct cxl *adapter) |
1101 | int rc; |
1102 | |
1103 | rc = atomic_inc_unless_negative(&adapter->contexts_num); |
1104 | - return rc >= 0 ? 0 : -EBUSY; |
1105 | + return rc ? 0 : -EBUSY; |
1106 | } |
1107 | |
1108 | void cxl_adapter_context_put(struct cxl *adapter) |
1109 | diff --git a/drivers/misc/vmw_balloon.c b/drivers/misc/vmw_balloon.c |
1110 | index 5e047bfc0cc4..518e2dec2aa2 100644 |
1111 | --- a/drivers/misc/vmw_balloon.c |
1112 | +++ b/drivers/misc/vmw_balloon.c |
1113 | @@ -341,7 +341,13 @@ static bool vmballoon_send_start(struct vmballoon *b, unsigned long req_caps) |
1114 | success = false; |
1115 | } |
1116 | |
1117 | - if (b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) |
1118 | + /* |
1119 | + * 2MB pages are only supported with batching. If batching is for some |
1120 | + * reason disabled, do not use 2MB pages, since otherwise the legacy |
1121 | + * mechanism is used with 2MB pages, causing a failure. |
1122 | + */ |
1123 | + if ((b->capabilities & VMW_BALLOON_BATCHED_2M_CMDS) && |
1124 | + (b->capabilities & VMW_BALLOON_BATCHED_CMDS)) |
1125 | b->supported_page_sizes = 2; |
1126 | else |
1127 | b->supported_page_sizes = 1; |
1128 | @@ -450,7 +456,7 @@ static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, |
1129 | |
1130 | pfn32 = (u32)pfn; |
1131 | if (pfn32 != pfn) |
1132 | - return -1; |
1133 | + return -EINVAL; |
1134 | |
1135 | STATS_INC(b->stats.lock[false]); |
1136 | |
1137 | @@ -460,7 +466,7 @@ static int vmballoon_send_lock_page(struct vmballoon *b, unsigned long pfn, |
1138 | |
1139 | pr_debug("%s - ppn %lx, hv returns %ld\n", __func__, pfn, status); |
1140 | STATS_INC(b->stats.lock_fail[false]); |
1141 | - return 1; |
1142 | + return -EIO; |
1143 | } |
1144 | |
1145 | static int vmballoon_send_batched_lock(struct vmballoon *b, |
1146 | @@ -597,11 +603,12 @@ static int vmballoon_lock_page(struct vmballoon *b, unsigned int num_pages, |
1147 | |
1148 | locked = vmballoon_send_lock_page(b, page_to_pfn(page), &hv_status, |
1149 | target); |
1150 | - if (locked > 0) { |
1151 | + if (locked) { |
1152 | STATS_INC(b->stats.refused_alloc[false]); |
1153 | |
1154 | - if (hv_status == VMW_BALLOON_ERROR_RESET || |
1155 | - hv_status == VMW_BALLOON_ERROR_PPN_NOTNEEDED) { |
1156 | + if (locked == -EIO && |
1157 | + (hv_status == VMW_BALLOON_ERROR_RESET || |
1158 | + hv_status == VMW_BALLOON_ERROR_PPN_NOTNEEDED)) { |
1159 | vmballoon_free_page(page, false); |
1160 | return -EIO; |
1161 | } |
1162 | @@ -617,7 +624,7 @@ static int vmballoon_lock_page(struct vmballoon *b, unsigned int num_pages, |
1163 | } else { |
1164 | vmballoon_free_page(page, false); |
1165 | } |
1166 | - return -EIO; |
1167 | + return locked; |
1168 | } |
1169 | |
1170 | /* track allocated page */ |
1171 | @@ -1029,29 +1036,30 @@ static void vmballoon_vmci_cleanup(struct vmballoon *b) |
1172 | */ |
1173 | static int vmballoon_vmci_init(struct vmballoon *b) |
1174 | { |
1175 | - int error = 0; |
1176 | + unsigned long error, dummy; |
1177 | |
1178 | - if ((b->capabilities & VMW_BALLOON_SIGNALLED_WAKEUP_CMD) != 0) { |
1179 | - error = vmci_doorbell_create(&b->vmci_doorbell, |
1180 | - VMCI_FLAG_DELAYED_CB, |
1181 | - VMCI_PRIVILEGE_FLAG_RESTRICTED, |
1182 | - vmballoon_doorbell, b); |
1183 | - |
1184 | - if (error == VMCI_SUCCESS) { |
1185 | - VMWARE_BALLOON_CMD(VMCI_DOORBELL_SET, |
1186 | - b->vmci_doorbell.context, |
1187 | - b->vmci_doorbell.resource, error); |
1188 | - STATS_INC(b->stats.doorbell_set); |
1189 | - } |
1190 | - } |
1191 | + if ((b->capabilities & VMW_BALLOON_SIGNALLED_WAKEUP_CMD) == 0) |
1192 | + return 0; |
1193 | |
1194 | - if (error != 0) { |
1195 | - vmballoon_vmci_cleanup(b); |
1196 | + error = vmci_doorbell_create(&b->vmci_doorbell, VMCI_FLAG_DELAYED_CB, |
1197 | + VMCI_PRIVILEGE_FLAG_RESTRICTED, |
1198 | + vmballoon_doorbell, b); |
1199 | |
1200 | - return -EIO; |
1201 | - } |
1202 | + if (error != VMCI_SUCCESS) |
1203 | + goto fail; |
1204 | + |
1205 | + error = VMWARE_BALLOON_CMD(VMCI_DOORBELL_SET, b->vmci_doorbell.context, |
1206 | + b->vmci_doorbell.resource, dummy); |
1207 | + |
1208 | + STATS_INC(b->stats.doorbell_set); |
1209 | + |
1210 | + if (error != VMW_BALLOON_SUCCESS) |
1211 | + goto fail; |
1212 | |
1213 | return 0; |
1214 | +fail: |
1215 | + vmballoon_vmci_cleanup(b); |
1216 | + return -EIO; |
1217 | } |
1218 | |
1219 | /* |
1220 | @@ -1289,7 +1297,14 @@ static int __init vmballoon_init(void) |
1221 | |
1222 | return 0; |
1223 | } |
1224 | -module_init(vmballoon_init); |
1225 | + |
1226 | +/* |
1227 | + * Using late_initcall() instead of module_init() allows the balloon to use the |
1228 | + * VMCI doorbell even when the balloon is built into the kernel. Otherwise the |
1229 | + * VMCI is probed only after the balloon is initialized. If the balloon is used |
1230 | + * as a module, late_initcall() is equivalent to module_init(). |
1231 | + */ |
1232 | +late_initcall(vmballoon_init); |
1233 | |
1234 | static void __exit vmballoon_exit(void) |
1235 | { |
1236 | diff --git a/drivers/net/wireless/marvell/libertas/dev.h b/drivers/net/wireless/marvell/libertas/dev.h |
1237 | index edf710bc5e77..3de1457ad3a7 100644 |
1238 | --- a/drivers/net/wireless/marvell/libertas/dev.h |
1239 | +++ b/drivers/net/wireless/marvell/libertas/dev.h |
1240 | @@ -103,6 +103,7 @@ struct lbs_private { |
1241 | u8 fw_ready; |
1242 | u8 surpriseremoved; |
1243 | u8 setup_fw_on_resume; |
1244 | + u8 power_up_on_resume; |
1245 | int (*hw_host_to_card) (struct lbs_private *priv, u8 type, u8 *payload, u16 nb); |
1246 | void (*reset_card) (struct lbs_private *priv); |
1247 | int (*power_save) (struct lbs_private *priv); |
1248 | diff --git a/drivers/net/wireless/marvell/libertas/if_sdio.c b/drivers/net/wireless/marvell/libertas/if_sdio.c |
1249 | index 47f4a14c84fe..a0ae8d8763bb 100644 |
1250 | --- a/drivers/net/wireless/marvell/libertas/if_sdio.c |
1251 | +++ b/drivers/net/wireless/marvell/libertas/if_sdio.c |
1252 | @@ -1341,15 +1341,23 @@ static void if_sdio_remove(struct sdio_func *func) |
1253 | static int if_sdio_suspend(struct device *dev) |
1254 | { |
1255 | struct sdio_func *func = dev_to_sdio_func(dev); |
1256 | - int ret; |
1257 | struct if_sdio_card *card = sdio_get_drvdata(func); |
1258 | + struct lbs_private *priv = card->priv; |
1259 | + int ret; |
1260 | |
1261 | mmc_pm_flag_t flags = sdio_get_host_pm_caps(func); |
1262 | + priv->power_up_on_resume = false; |
1263 | |
1264 | /* If we're powered off anyway, just let the mmc layer remove the |
1265 | * card. */ |
1266 | - if (!lbs_iface_active(card->priv)) |
1267 | - return -ENOSYS; |
1268 | + if (!lbs_iface_active(priv)) { |
1269 | + if (priv->fw_ready) { |
1270 | + priv->power_up_on_resume = true; |
1271 | + if_sdio_power_off(card); |
1272 | + } |
1273 | + |
1274 | + return 0; |
1275 | + } |
1276 | |
1277 | dev_info(dev, "%s: suspend: PM flags = 0x%x\n", |
1278 | sdio_func_id(func), flags); |
1279 | @@ -1357,9 +1365,14 @@ static int if_sdio_suspend(struct device *dev) |
1280 | /* If we aren't being asked to wake on anything, we should bail out |
1281 | * and let the SD stack power down the card. |
1282 | */ |
1283 | - if (card->priv->wol_criteria == EHS_REMOVE_WAKEUP) { |
1284 | + if (priv->wol_criteria == EHS_REMOVE_WAKEUP) { |
1285 | dev_info(dev, "Suspend without wake params -- powering down card\n"); |
1286 | - return -ENOSYS; |
1287 | + if (priv->fw_ready) { |
1288 | + priv->power_up_on_resume = true; |
1289 | + if_sdio_power_off(card); |
1290 | + } |
1291 | + |
1292 | + return 0; |
1293 | } |
1294 | |
1295 | if (!(flags & MMC_PM_KEEP_POWER)) { |
1296 | @@ -1372,7 +1385,7 @@ static int if_sdio_suspend(struct device *dev) |
1297 | if (ret) |
1298 | return ret; |
1299 | |
1300 | - ret = lbs_suspend(card->priv); |
1301 | + ret = lbs_suspend(priv); |
1302 | if (ret) |
1303 | return ret; |
1304 | |
1305 | @@ -1387,6 +1400,11 @@ static int if_sdio_resume(struct device *dev) |
1306 | |
1307 | dev_info(dev, "%s: resume: we're back\n", sdio_func_id(func)); |
1308 | |
1309 | + if (card->priv->power_up_on_resume) { |
1310 | + if_sdio_power_on(card); |
1311 | + wait_event(card->pwron_waitq, card->priv->fw_ready); |
1312 | + } |
1313 | + |
1314 | ret = lbs_resume(card->priv); |
1315 | |
1316 | return ret; |
1317 | diff --git a/drivers/nvdimm/bus.c b/drivers/nvdimm/bus.c |
1318 | index c1a65ce31243..de6d3b749c60 100644 |
1319 | --- a/drivers/nvdimm/bus.c |
1320 | +++ b/drivers/nvdimm/bus.c |
1321 | @@ -748,9 +748,9 @@ u32 nd_cmd_out_size(struct nvdimm *nvdimm, int cmd, |
1322 | * overshoots the remainder by 4 bytes, assume it was |
1323 | * including 'status'. |
1324 | */ |
1325 | - if (out_field[1] - 8 == remainder) |
1326 | + if (out_field[1] - 4 == remainder) |
1327 | return remainder; |
1328 | - return out_field[1] - 4; |
1329 | + return out_field[1] - 8; |
1330 | } else if (cmd == ND_CMD_CALL) { |
1331 | struct nd_cmd_pkg *pkg = (struct nd_cmd_pkg *) in_field; |
1332 | |
1333 | diff --git a/drivers/pwm/pwm-tiehrpwm.c b/drivers/pwm/pwm-tiehrpwm.c |
1334 | index b5c6b0636893..c0e06f0c19d1 100644 |
1335 | --- a/drivers/pwm/pwm-tiehrpwm.c |
1336 | +++ b/drivers/pwm/pwm-tiehrpwm.c |
1337 | @@ -382,6 +382,8 @@ static void ehrpwm_pwm_disable(struct pwm_chip *chip, struct pwm_device *pwm) |
1338 | aqcsfrc_mask = AQCSFRC_CSFA_MASK; |
1339 | } |
1340 | |
1341 | + /* Update shadow register first before modifying active register */ |
1342 | + ehrpwm_modify(pc->mmio_base, AQCSFRC, aqcsfrc_mask, aqcsfrc_val); |
1343 | /* |
1344 | * Changes to immediate action on Action Qualifier. This puts |
1345 | * Action Qualifier control on PWM output from next TBCLK |
1346 | diff --git a/drivers/rtc/rtc-omap.c b/drivers/rtc/rtc-omap.c |
1347 | index 51e52446eacb..bd5ca548c265 100644 |
1348 | --- a/drivers/rtc/rtc-omap.c |
1349 | +++ b/drivers/rtc/rtc-omap.c |
1350 | @@ -817,13 +817,6 @@ static int omap_rtc_probe(struct platform_device *pdev) |
1351 | goto err; |
1352 | } |
1353 | |
1354 | - if (rtc->is_pmic_controller) { |
1355 | - if (!pm_power_off) { |
1356 | - omap_rtc_power_off_rtc = rtc; |
1357 | - pm_power_off = omap_rtc_power_off; |
1358 | - } |
1359 | - } |
1360 | - |
1361 | /* Support ext_wakeup pinconf */ |
1362 | rtc_pinctrl_desc.name = dev_name(&pdev->dev); |
1363 | |
1364 | @@ -833,6 +826,13 @@ static int omap_rtc_probe(struct platform_device *pdev) |
1365 | return PTR_ERR(rtc->pctldev); |
1366 | } |
1367 | |
1368 | + if (rtc->is_pmic_controller) { |
1369 | + if (!pm_power_off) { |
1370 | + omap_rtc_power_off_rtc = rtc; |
1371 | + pm_power_off = omap_rtc_power_off; |
1372 | + } |
1373 | + } |
1374 | + |
1375 | return 0; |
1376 | |
1377 | err: |
1378 | diff --git a/drivers/spi/spi-davinci.c b/drivers/spi/spi-davinci.c |
1379 | index 0d8f43a17edb..1905d20c229f 100644 |
1380 | --- a/drivers/spi/spi-davinci.c |
1381 | +++ b/drivers/spi/spi-davinci.c |
1382 | @@ -215,7 +215,7 @@ static void davinci_spi_chipselect(struct spi_device *spi, int value) |
1383 | pdata = &dspi->pdata; |
1384 | |
1385 | /* program delay transfers if tx_delay is non zero */ |
1386 | - if (spicfg->wdelay) |
1387 | + if (spicfg && spicfg->wdelay) |
1388 | spidat1 |= SPIDAT1_WDEL; |
1389 | |
1390 | /* |
1391 | diff --git a/drivers/spi/spi-fsl-dspi.c b/drivers/spi/spi-fsl-dspi.c |
1392 | index a67b0ff6a362..db3b6e9151a8 100644 |
1393 | --- a/drivers/spi/spi-fsl-dspi.c |
1394 | +++ b/drivers/spi/spi-fsl-dspi.c |
1395 | @@ -715,30 +715,30 @@ static int dspi_probe(struct platform_device *pdev) |
1396 | return PTR_ERR(dspi->regmap); |
1397 | } |
1398 | |
1399 | + dspi->clk = devm_clk_get(&pdev->dev, "dspi"); |
1400 | + if (IS_ERR(dspi->clk)) { |
1401 | + ret = PTR_ERR(dspi->clk); |
1402 | + dev_err(&pdev->dev, "unable to get clock\n"); |
1403 | + goto out_master_put; |
1404 | + } |
1405 | + ret = clk_prepare_enable(dspi->clk); |
1406 | + if (ret) |
1407 | + goto out_master_put; |
1408 | + |
1409 | dspi_init(dspi); |
1410 | dspi->irq = platform_get_irq(pdev, 0); |
1411 | if (dspi->irq < 0) { |
1412 | dev_err(&pdev->dev, "can't get platform irq\n"); |
1413 | ret = dspi->irq; |
1414 | - goto out_master_put; |
1415 | + goto out_clk_put; |
1416 | } |
1417 | |
1418 | ret = devm_request_irq(&pdev->dev, dspi->irq, dspi_interrupt, 0, |
1419 | pdev->name, dspi); |
1420 | if (ret < 0) { |
1421 | dev_err(&pdev->dev, "Unable to attach DSPI interrupt\n"); |
1422 | - goto out_master_put; |
1423 | - } |
1424 | - |
1425 | - dspi->clk = devm_clk_get(&pdev->dev, "dspi"); |
1426 | - if (IS_ERR(dspi->clk)) { |
1427 | - ret = PTR_ERR(dspi->clk); |
1428 | - dev_err(&pdev->dev, "unable to get clock\n"); |
1429 | - goto out_master_put; |
1430 | + goto out_clk_put; |
1431 | } |
1432 | - ret = clk_prepare_enable(dspi->clk); |
1433 | - if (ret) |
1434 | - goto out_master_put; |
1435 | |
1436 | master->max_speed_hz = |
1437 | clk_get_rate(dspi->clk) / dspi->devtype_data->max_clock_factor; |
1438 | diff --git a/drivers/tty/serial/serial_core.c b/drivers/tty/serial/serial_core.c |
1439 | index 6cbd46c565f8..53e6db8b0330 100644 |
1440 | --- a/drivers/tty/serial/serial_core.c |
1441 | +++ b/drivers/tty/serial/serial_core.c |
1442 | @@ -172,6 +172,7 @@ static int uart_port_startup(struct tty_struct *tty, struct uart_state *state, |
1443 | { |
1444 | struct uart_port *uport = uart_port_check(state); |
1445 | unsigned long page; |
1446 | + unsigned long flags = 0; |
1447 | int retval = 0; |
1448 | |
1449 | if (uport->type == PORT_UNKNOWN) |
1450 | @@ -186,15 +187,18 @@ static int uart_port_startup(struct tty_struct *tty, struct uart_state *state, |
1451 | * Initialise and allocate the transmit and temporary |
1452 | * buffer. |
1453 | */ |
1454 | - if (!state->xmit.buf) { |
1455 | - /* This is protected by the per port mutex */ |
1456 | - page = get_zeroed_page(GFP_KERNEL); |
1457 | - if (!page) |
1458 | - return -ENOMEM; |
1459 | + page = get_zeroed_page(GFP_KERNEL); |
1460 | + if (!page) |
1461 | + return -ENOMEM; |
1462 | |
1463 | + uart_port_lock(state, flags); |
1464 | + if (!state->xmit.buf) { |
1465 | state->xmit.buf = (unsigned char *) page; |
1466 | uart_circ_clear(&state->xmit); |
1467 | + } else { |
1468 | + free_page(page); |
1469 | } |
1470 | + uart_port_unlock(uport, flags); |
1471 | |
1472 | retval = uport->ops->startup(uport); |
1473 | if (retval == 0) { |
1474 | @@ -253,6 +257,7 @@ static void uart_shutdown(struct tty_struct *tty, struct uart_state *state) |
1475 | { |
1476 | struct uart_port *uport = uart_port_check(state); |
1477 | struct tty_port *port = &state->port; |
1478 | + unsigned long flags = 0; |
1479 | |
1480 | /* |
1481 | * Set the TTY IO error marker |
1482 | @@ -285,10 +290,12 @@ static void uart_shutdown(struct tty_struct *tty, struct uart_state *state) |
1483 | /* |
1484 | * Free the transmit buffer page. |
1485 | */ |
1486 | + uart_port_lock(state, flags); |
1487 | if (state->xmit.buf) { |
1488 | free_page((unsigned long)state->xmit.buf); |
1489 | state->xmit.buf = NULL; |
1490 | } |
1491 | + uart_port_unlock(uport, flags); |
1492 | } |
1493 | |
1494 | /** |
1495 | diff --git a/drivers/video/fbdev/core/fbmem.c b/drivers/video/fbdev/core/fbmem.c |
1496 | index 76c1ad96fb37..74273bc7ca9a 100644 |
1497 | --- a/drivers/video/fbdev/core/fbmem.c |
1498 | +++ b/drivers/video/fbdev/core/fbmem.c |
1499 | @@ -1695,12 +1695,12 @@ static int do_register_framebuffer(struct fb_info *fb_info) |
1500 | return 0; |
1501 | } |
1502 | |
1503 | -static int do_unregister_framebuffer(struct fb_info *fb_info) |
1504 | +static int unbind_console(struct fb_info *fb_info) |
1505 | { |
1506 | struct fb_event event; |
1507 | - int i, ret = 0; |
1508 | + int ret; |
1509 | + int i = fb_info->node; |
1510 | |
1511 | - i = fb_info->node; |
1512 | if (i < 0 || i >= FB_MAX || registered_fb[i] != fb_info) |
1513 | return -EINVAL; |
1514 | |
1515 | @@ -1715,17 +1715,29 @@ static int do_unregister_framebuffer(struct fb_info *fb_info) |
1516 | unlock_fb_info(fb_info); |
1517 | console_unlock(); |
1518 | |
1519 | + return ret; |
1520 | +} |
1521 | + |
1522 | +static int __unlink_framebuffer(struct fb_info *fb_info); |
1523 | + |
1524 | +static int do_unregister_framebuffer(struct fb_info *fb_info) |
1525 | +{ |
1526 | + struct fb_event event; |
1527 | + int ret; |
1528 | + |
1529 | + ret = unbind_console(fb_info); |
1530 | + |
1531 | if (ret) |
1532 | return -EINVAL; |
1533 | |
1534 | pm_vt_switch_unregister(fb_info->dev); |
1535 | |
1536 | - unlink_framebuffer(fb_info); |
1537 | + __unlink_framebuffer(fb_info); |
1538 | if (fb_info->pixmap.addr && |
1539 | (fb_info->pixmap.flags & FB_PIXMAP_DEFAULT)) |
1540 | kfree(fb_info->pixmap.addr); |
1541 | fb_destroy_modelist(&fb_info->modelist); |
1542 | - registered_fb[i] = NULL; |
1543 | + registered_fb[fb_info->node] = NULL; |
1544 | num_registered_fb--; |
1545 | fb_cleanup_device(fb_info); |
1546 | event.info = fb_info; |
1547 | @@ -1738,7 +1750,7 @@ static int do_unregister_framebuffer(struct fb_info *fb_info) |
1548 | return 0; |
1549 | } |
1550 | |
1551 | -int unlink_framebuffer(struct fb_info *fb_info) |
1552 | +static int __unlink_framebuffer(struct fb_info *fb_info) |
1553 | { |
1554 | int i; |
1555 | |
1556 | @@ -1750,6 +1762,20 @@ int unlink_framebuffer(struct fb_info *fb_info) |
1557 | device_destroy(fb_class, MKDEV(FB_MAJOR, i)); |
1558 | fb_info->dev = NULL; |
1559 | } |
1560 | + |
1561 | + return 0; |
1562 | +} |
1563 | + |
1564 | +int unlink_framebuffer(struct fb_info *fb_info) |
1565 | +{ |
1566 | + int ret; |
1567 | + |
1568 | + ret = __unlink_framebuffer(fb_info); |
1569 | + if (ret) |
1570 | + return ret; |
1571 | + |
1572 | + unbind_console(fb_info); |
1573 | + |
1574 | return 0; |
1575 | } |
1576 | EXPORT_SYMBOL(unlink_framebuffer); |
1577 | diff --git a/fs/9p/xattr.c b/fs/9p/xattr.c |
1578 | index f329eee6dc93..352abc39e891 100644 |
1579 | --- a/fs/9p/xattr.c |
1580 | +++ b/fs/9p/xattr.c |
1581 | @@ -105,7 +105,7 @@ int v9fs_fid_xattr_set(struct p9_fid *fid, const char *name, |
1582 | { |
1583 | struct kvec kvec = {.iov_base = (void *)value, .iov_len = value_len}; |
1584 | struct iov_iter from; |
1585 | - int retval; |
1586 | + int retval, err; |
1587 | |
1588 | iov_iter_kvec(&from, WRITE | ITER_KVEC, &kvec, 1, value_len); |
1589 | |
1590 | @@ -126,7 +126,9 @@ int v9fs_fid_xattr_set(struct p9_fid *fid, const char *name, |
1591 | retval); |
1592 | else |
1593 | p9_client_write(fid, 0, &from, &retval); |
1594 | - p9_client_clunk(fid); |
1595 | + err = p9_client_clunk(fid); |
1596 | + if (!retval && err) |
1597 | + retval = err; |
1598 | return retval; |
1599 | } |
1600 | |
1601 | diff --git a/fs/nfs/blocklayout/dev.c b/fs/nfs/blocklayout/dev.c |
1602 | index a69ef4e9c24c..d6e4191276c0 100644 |
1603 | --- a/fs/nfs/blocklayout/dev.c |
1604 | +++ b/fs/nfs/blocklayout/dev.c |
1605 | @@ -203,7 +203,7 @@ static bool bl_map_stripe(struct pnfs_block_dev *dev, u64 offset, |
1606 | chunk = div_u64(offset, dev->chunk_size); |
1607 | div_u64_rem(chunk, dev->nr_children, &chunk_idx); |
1608 | |
1609 | - if (chunk_idx > dev->nr_children) { |
1610 | + if (chunk_idx >= dev->nr_children) { |
1611 | dprintk("%s: invalid chunk idx %d (%lld/%lld)\n", |
1612 | __func__, chunk_idx, offset, dev->chunk_size); |
1613 | /* error, should not happen */ |
1614 | diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c |
1615 | index cf5fdc25289a..e7ca62a86dab 100644 |
1616 | --- a/fs/nfs/nfs4proc.c |
1617 | +++ b/fs/nfs/nfs4proc.c |
1618 | @@ -541,8 +541,15 @@ nfs4_async_handle_exception(struct rpc_task *task, struct nfs_server *server, |
1619 | ret = -EIO; |
1620 | return ret; |
1621 | out_retry: |
1622 | - if (ret == 0) |
1623 | + if (ret == 0) { |
1624 | exception->retry = 1; |
1625 | + /* |
1626 | + * For NFS4ERR_MOVED, the client transport will need to |
1627 | + * be recomputed after migration recovery has completed. |
1628 | + */ |
1629 | + if (errorcode == -NFS4ERR_MOVED) |
1630 | + rpc_task_release_transport(task); |
1631 | + } |
1632 | return ret; |
1633 | } |
1634 | |
1635 | diff --git a/fs/quota/quota.c b/fs/quota/quota.c |
1636 | index 2d445425aad7..a2329f7ec638 100644 |
1637 | --- a/fs/quota/quota.c |
1638 | +++ b/fs/quota/quota.c |
1639 | @@ -17,6 +17,7 @@ |
1640 | #include <linux/quotaops.h> |
1641 | #include <linux/types.h> |
1642 | #include <linux/writeback.h> |
1643 | +#include <linux/nospec.h> |
1644 | |
1645 | static int check_quotactl_permission(struct super_block *sb, int type, int cmd, |
1646 | qid_t id) |
1647 | @@ -706,6 +707,7 @@ static int do_quotactl(struct super_block *sb, int type, int cmd, qid_t id, |
1648 | |
1649 | if (type >= (XQM_COMMAND(cmd) ? XQM_MAXQUOTAS : MAXQUOTAS)) |
1650 | return -EINVAL; |
1651 | + type = array_index_nospec(type, MAXQUOTAS); |
1652 | /* |
1653 | * Quota not supported on this fs? Check this before s_quota_types |
1654 | * since they needn't be set if quota is not supported at all. |
1655 | diff --git a/fs/ubifs/journal.c b/fs/ubifs/journal.c |
1656 | index 504658fd0d08..f8ce849e90d1 100644 |
1657 | --- a/fs/ubifs/journal.c |
1658 | +++ b/fs/ubifs/journal.c |
1659 | @@ -661,6 +661,11 @@ int ubifs_jnl_update(struct ubifs_info *c, const struct inode *dir, |
1660 | spin_lock(&ui->ui_lock); |
1661 | ui->synced_i_size = ui->ui_size; |
1662 | spin_unlock(&ui->ui_lock); |
1663 | + if (xent) { |
1664 | + spin_lock(&host_ui->ui_lock); |
1665 | + host_ui->synced_i_size = host_ui->ui_size; |
1666 | + spin_unlock(&host_ui->ui_lock); |
1667 | + } |
1668 | mark_inode_clean(c, ui); |
1669 | mark_inode_clean(c, host_ui); |
1670 | return 0; |
1671 | @@ -1265,7 +1270,7 @@ static int recomp_data_node(const struct ubifs_info *c, |
1672 | int err, len, compr_type, out_len; |
1673 | |
1674 | out_len = le32_to_cpu(dn->size); |
1675 | - buf = kmalloc_array(out_len, WORST_COMPR_FACTOR, GFP_NOFS); |
1676 | + buf = kmalloc(out_len * WORST_COMPR_FACTOR, GFP_NOFS); |
1677 | if (!buf) |
1678 | return -ENOMEM; |
1679 | |
1680 | @@ -1344,7 +1349,16 @@ int ubifs_jnl_truncate(struct ubifs_info *c, const struct inode *inode, |
1681 | else if (err) |
1682 | goto out_free; |
1683 | else { |
1684 | - if (le32_to_cpu(dn->size) <= dlen) |
1685 | + int dn_len = le32_to_cpu(dn->size); |
1686 | + |
1687 | + if (dn_len <= 0 || dn_len > UBIFS_BLOCK_SIZE) { |
1688 | + ubifs_err(c, "bad data node (block %u, inode %lu)", |
1689 | + blk, inode->i_ino); |
1690 | + ubifs_dump_node(c, dn); |
1691 | + goto out_free; |
1692 | + } |
1693 | + |
1694 | + if (dn_len <= dlen) |
1695 | dlen = 0; /* Nothing to do */ |
1696 | else { |
1697 | int compr_type = le16_to_cpu(dn->compr_type); |
1698 | diff --git a/fs/ubifs/lprops.c b/fs/ubifs/lprops.c |
1699 | index 6c3a1abd0e22..780a436d8c45 100644 |
1700 | --- a/fs/ubifs/lprops.c |
1701 | +++ b/fs/ubifs/lprops.c |
1702 | @@ -1091,10 +1091,6 @@ static int scan_check_cb(struct ubifs_info *c, |
1703 | } |
1704 | } |
1705 | |
1706 | - buf = __vmalloc(c->leb_size, GFP_NOFS, PAGE_KERNEL); |
1707 | - if (!buf) |
1708 | - return -ENOMEM; |
1709 | - |
1710 | /* |
1711 | * After an unclean unmount, empty and freeable LEBs |
1712 | * may contain garbage - do not scan them. |
1713 | @@ -1113,6 +1109,10 @@ static int scan_check_cb(struct ubifs_info *c, |
1714 | return LPT_SCAN_CONTINUE; |
1715 | } |
1716 | |
1717 | + buf = __vmalloc(c->leb_size, GFP_NOFS, PAGE_KERNEL); |
1718 | + if (!buf) |
1719 | + return -ENOMEM; |
1720 | + |
1721 | sleb = ubifs_scan(c, lnum, 0, buf, 0); |
1722 | if (IS_ERR(sleb)) { |
1723 | ret = PTR_ERR(sleb); |
1724 | diff --git a/fs/xattr.c b/fs/xattr.c |
1725 | index 932b9061a3a2..093998872329 100644 |
1726 | --- a/fs/xattr.c |
1727 | +++ b/fs/xattr.c |
1728 | @@ -540,7 +540,7 @@ getxattr(struct dentry *d, const char __user *name, void __user *value, |
1729 | if (error > 0) { |
1730 | if ((strcmp(kname, XATTR_NAME_POSIX_ACL_ACCESS) == 0) || |
1731 | (strcmp(kname, XATTR_NAME_POSIX_ACL_DEFAULT) == 0)) |
1732 | - posix_acl_fix_xattr_to_user(kvalue, size); |
1733 | + posix_acl_fix_xattr_to_user(kvalue, error); |
1734 | if (size && copy_to_user(value, kvalue, error)) |
1735 | error = -EFAULT; |
1736 | } else if (error == -ERANGE && size >= XATTR_SIZE_MAX) { |
1737 | diff --git a/include/linux/intel-iommu.h b/include/linux/intel-iommu.h |
1738 | index 23e129ef6726..e353f6600b0b 100644 |
1739 | --- a/include/linux/intel-iommu.h |
1740 | +++ b/include/linux/intel-iommu.h |
1741 | @@ -125,6 +125,7 @@ static inline void dmar_writeq(void __iomem *addr, u64 val) |
1742 | * Extended Capability Register |
1743 | */ |
1744 | |
1745 | +#define ecap_dit(e) ((e >> 41) & 0x1) |
1746 | #define ecap_pasid(e) ((e >> 40) & 0x1) |
1747 | #define ecap_pss(e) ((e >> 35) & 0x1f) |
1748 | #define ecap_eafs(e) ((e >> 34) & 0x1) |
1749 | @@ -294,6 +295,7 @@ enum { |
1750 | #define QI_DEV_IOTLB_SID(sid) ((u64)((sid) & 0xffff) << 32) |
1751 | #define QI_DEV_IOTLB_QDEP(qdep) (((qdep) & 0x1f) << 16) |
1752 | #define QI_DEV_IOTLB_ADDR(addr) ((u64)(addr) & VTD_PAGE_MASK) |
1753 | +#define QI_DEV_IOTLB_PFSID(pfsid) (((u64)(pfsid & 0xf) << 12) | ((u64)(pfsid & 0xfff) << 52)) |
1754 | #define QI_DEV_IOTLB_SIZE 1 |
1755 | #define QI_DEV_IOTLB_MAX_INVS 32 |
1756 | |
1757 | @@ -318,6 +320,7 @@ enum { |
1758 | #define QI_DEV_EIOTLB_PASID(p) (((u64)p) << 32) |
1759 | #define QI_DEV_EIOTLB_SID(sid) ((u64)((sid) & 0xffff) << 16) |
1760 | #define QI_DEV_EIOTLB_QDEP(qd) ((u64)((qd) & 0x1f) << 4) |
1761 | +#define QI_DEV_EIOTLB_PFSID(pfsid) (((u64)(pfsid & 0xf) << 12) | ((u64)(pfsid & 0xfff) << 52)) |
1762 | #define QI_DEV_EIOTLB_MAX_INVS 32 |
1763 | |
1764 | #define QI_PGRP_IDX(idx) (((u64)(idx)) << 55) |
1765 | @@ -463,9 +466,8 @@ extern void qi_flush_context(struct intel_iommu *iommu, u16 did, u16 sid, |
1766 | u8 fm, u64 type); |
1767 | extern void qi_flush_iotlb(struct intel_iommu *iommu, u16 did, u64 addr, |
1768 | unsigned int size_order, u64 type); |
1769 | -extern void qi_flush_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 qdep, |
1770 | - u64 addr, unsigned mask); |
1771 | - |
1772 | +extern void qi_flush_dev_iotlb(struct intel_iommu *iommu, u16 sid, u16 pfsid, |
1773 | + u16 qdep, u64 addr, unsigned mask); |
1774 | extern int qi_submit_sync(struct qi_desc *desc, struct intel_iommu *iommu); |
1775 | |
1776 | extern int dmar_ir_support(void); |
1777 | diff --git a/include/linux/pci.h b/include/linux/pci.h |
1778 | index 78c9f4a91d94..534cb43e8635 100644 |
1779 | --- a/include/linux/pci.h |
1780 | +++ b/include/linux/pci.h |
1781 | @@ -2151,4 +2151,16 @@ static inline bool pci_ari_enabled(struct pci_bus *bus) |
1782 | /* provide the legacy pci_dma_* API */ |
1783 | #include <linux/pci-dma-compat.h> |
1784 | |
1785 | +#define pci_printk(level, pdev, fmt, arg...) \ |
1786 | + dev_printk(level, &(pdev)->dev, fmt, ##arg) |
1787 | + |
1788 | +#define pci_emerg(pdev, fmt, arg...) dev_emerg(&(pdev)->dev, fmt, ##arg) |
1789 | +#define pci_alert(pdev, fmt, arg...) dev_alert(&(pdev)->dev, fmt, ##arg) |
1790 | +#define pci_crit(pdev, fmt, arg...) dev_crit(&(pdev)->dev, fmt, ##arg) |
1791 | +#define pci_err(pdev, fmt, arg...) dev_err(&(pdev)->dev, fmt, ##arg) |
1792 | +#define pci_warn(pdev, fmt, arg...) dev_warn(&(pdev)->dev, fmt, ##arg) |
1793 | +#define pci_notice(pdev, fmt, arg...) dev_notice(&(pdev)->dev, fmt, ##arg) |
1794 | +#define pci_info(pdev, fmt, arg...) dev_info(&(pdev)->dev, fmt, ##arg) |
1795 | +#define pci_dbg(pdev, fmt, arg...) dev_dbg(&(pdev)->dev, fmt, ##arg) |
1796 | + |
1797 | #endif /* LINUX_PCI_H */ |
1798 | diff --git a/include/linux/sunrpc/clnt.h b/include/linux/sunrpc/clnt.h |
1799 | index 333ad11b3dd9..44161a4c124c 100644 |
1800 | --- a/include/linux/sunrpc/clnt.h |
1801 | +++ b/include/linux/sunrpc/clnt.h |
1802 | @@ -155,6 +155,7 @@ int rpc_switch_client_transport(struct rpc_clnt *, |
1803 | |
1804 | void rpc_shutdown_client(struct rpc_clnt *); |
1805 | void rpc_release_client(struct rpc_clnt *); |
1806 | +void rpc_task_release_transport(struct rpc_task *); |
1807 | void rpc_task_release_client(struct rpc_task *); |
1808 | |
1809 | int rpcb_create_local(struct net *); |
1810 | diff --git a/include/linux/verification.h b/include/linux/verification.h |
1811 | index a10549a6c7cd..cfa4730d607a 100644 |
1812 | --- a/include/linux/verification.h |
1813 | +++ b/include/linux/verification.h |
1814 | @@ -12,6 +12,12 @@ |
1815 | #ifndef _LINUX_VERIFICATION_H |
1816 | #define _LINUX_VERIFICATION_H |
1817 | |
1818 | +/* |
1819 | + * Indicate that both builtin trusted keys and secondary trusted keys |
1820 | + * should be used. |
1821 | + */ |
1822 | +#define VERIFY_USE_SECONDARY_KEYRING ((struct key *)1UL) |
1823 | + |
1824 | /* |
1825 | * The use to which an asymmetric key is being put. |
1826 | */ |
1827 | diff --git a/include/video/udlfb.h b/include/video/udlfb.h |
1828 | index f9466fa54ba4..2ad9a6d37ff4 100644 |
1829 | --- a/include/video/udlfb.h |
1830 | +++ b/include/video/udlfb.h |
1831 | @@ -87,7 +87,7 @@ struct dlfb_data { |
1832 | #define MIN_RAW_PIX_BYTES 2 |
1833 | #define MIN_RAW_CMD_BYTES (RAW_HEADER_BYTES + MIN_RAW_PIX_BYTES) |
1834 | |
1835 | -#define DL_DEFIO_WRITE_DELAY 5 /* fb_deferred_io.delay in jiffies */ |
1836 | +#define DL_DEFIO_WRITE_DELAY msecs_to_jiffies(HZ <= 300 ? 4 : 10) /* optimal value for 720p video */ |
1837 | #define DL_DEFIO_WRITE_DISABLE (HZ*60) /* "disable" with long delay */ |
1838 | |
1839 | /* remove these once align.h patch is taken into kernel */ |
1840 | diff --git a/kernel/power/Kconfig b/kernel/power/Kconfig |
1841 | index e8517b63eb37..dd2b5a4d89a5 100644 |
1842 | --- a/kernel/power/Kconfig |
1843 | +++ b/kernel/power/Kconfig |
1844 | @@ -105,6 +105,7 @@ config PM_SLEEP |
1845 | def_bool y |
1846 | depends on SUSPEND || HIBERNATE_CALLBACKS |
1847 | select PM |
1848 | + select SRCU |
1849 | |
1850 | config PM_SLEEP_SMP |
1851 | def_bool y |
1852 | diff --git a/kernel/printk/nmi.c b/kernel/printk/nmi.c |
1853 | index 5fa65aa904d3..2c3e7f024c15 100644 |
1854 | --- a/kernel/printk/nmi.c |
1855 | +++ b/kernel/printk/nmi.c |
1856 | @@ -260,12 +260,12 @@ void __init printk_nmi_init(void) |
1857 | printk_nmi_flush(); |
1858 | } |
1859 | |
1860 | -void printk_nmi_enter(void) |
1861 | +void notrace printk_nmi_enter(void) |
1862 | { |
1863 | this_cpu_write(printk_func, vprintk_nmi); |
1864 | } |
1865 | |
1866 | -void printk_nmi_exit(void) |
1867 | +void notrace printk_nmi_exit(void) |
1868 | { |
1869 | this_cpu_write(printk_func, vprintk_default); |
1870 | } |
1871 | diff --git a/kernel/sys.c b/kernel/sys.c |
1872 | index b13b530b5e0f..6c4e9b533258 100644 |
1873 | --- a/kernel/sys.c |
1874 | +++ b/kernel/sys.c |
1875 | @@ -1142,18 +1142,19 @@ static int override_release(char __user *release, size_t len) |
1876 | |
1877 | SYSCALL_DEFINE1(newuname, struct new_utsname __user *, name) |
1878 | { |
1879 | - int errno = 0; |
1880 | + struct new_utsname tmp; |
1881 | |
1882 | down_read(&uts_sem); |
1883 | - if (copy_to_user(name, utsname(), sizeof *name)) |
1884 | - errno = -EFAULT; |
1885 | + memcpy(&tmp, utsname(), sizeof(tmp)); |
1886 | up_read(&uts_sem); |
1887 | + if (copy_to_user(name, &tmp, sizeof(tmp))) |
1888 | + return -EFAULT; |
1889 | |
1890 | - if (!errno && override_release(name->release, sizeof(name->release))) |
1891 | - errno = -EFAULT; |
1892 | - if (!errno && override_architecture(name)) |
1893 | - errno = -EFAULT; |
1894 | - return errno; |
1895 | + if (override_release(name->release, sizeof(name->release))) |
1896 | + return -EFAULT; |
1897 | + if (override_architecture(name)) |
1898 | + return -EFAULT; |
1899 | + return 0; |
1900 | } |
1901 | |
1902 | #ifdef __ARCH_WANT_SYS_OLD_UNAME |
1903 | @@ -1162,55 +1163,46 @@ SYSCALL_DEFINE1(newuname, struct new_utsname __user *, name) |
1904 | */ |
1905 | SYSCALL_DEFINE1(uname, struct old_utsname __user *, name) |
1906 | { |
1907 | - int error = 0; |
1908 | + struct old_utsname tmp; |
1909 | |
1910 | if (!name) |
1911 | return -EFAULT; |
1912 | |
1913 | down_read(&uts_sem); |
1914 | - if (copy_to_user(name, utsname(), sizeof(*name))) |
1915 | - error = -EFAULT; |
1916 | + memcpy(&tmp, utsname(), sizeof(tmp)); |
1917 | up_read(&uts_sem); |
1918 | + if (copy_to_user(name, &tmp, sizeof(tmp))) |
1919 | + return -EFAULT; |
1920 | |
1921 | - if (!error && override_release(name->release, sizeof(name->release))) |
1922 | - error = -EFAULT; |
1923 | - if (!error && override_architecture(name)) |
1924 | - error = -EFAULT; |
1925 | - return error; |
1926 | + if (override_release(name->release, sizeof(name->release))) |
1927 | + return -EFAULT; |
1928 | + if (override_architecture(name)) |
1929 | + return -EFAULT; |
1930 | + return 0; |
1931 | } |
1932 | |
1933 | SYSCALL_DEFINE1(olduname, struct oldold_utsname __user *, name) |
1934 | { |
1935 | - int error; |
1936 | + struct oldold_utsname tmp = {}; |
1937 | |
1938 | if (!name) |
1939 | return -EFAULT; |
1940 | - if (!access_ok(VERIFY_WRITE, name, sizeof(struct oldold_utsname))) |
1941 | - return -EFAULT; |
1942 | |
1943 | down_read(&uts_sem); |
1944 | - error = __copy_to_user(&name->sysname, &utsname()->sysname, |
1945 | - __OLD_UTS_LEN); |
1946 | - error |= __put_user(0, name->sysname + __OLD_UTS_LEN); |
1947 | - error |= __copy_to_user(&name->nodename, &utsname()->nodename, |
1948 | - __OLD_UTS_LEN); |
1949 | - error |= __put_user(0, name->nodename + __OLD_UTS_LEN); |
1950 | - error |= __copy_to_user(&name->release, &utsname()->release, |
1951 | - __OLD_UTS_LEN); |
1952 | - error |= __put_user(0, name->release + __OLD_UTS_LEN); |
1953 | - error |= __copy_to_user(&name->version, &utsname()->version, |
1954 | - __OLD_UTS_LEN); |
1955 | - error |= __put_user(0, name->version + __OLD_UTS_LEN); |
1956 | - error |= __copy_to_user(&name->machine, &utsname()->machine, |
1957 | - __OLD_UTS_LEN); |
1958 | - error |= __put_user(0, name->machine + __OLD_UTS_LEN); |
1959 | + memcpy(&tmp.sysname, &utsname()->sysname, __OLD_UTS_LEN); |
1960 | + memcpy(&tmp.nodename, &utsname()->nodename, __OLD_UTS_LEN); |
1961 | + memcpy(&tmp.release, &utsname()->release, __OLD_UTS_LEN); |
1962 | + memcpy(&tmp.version, &utsname()->version, __OLD_UTS_LEN); |
1963 | + memcpy(&tmp.machine, &utsname()->machine, __OLD_UTS_LEN); |
1964 | up_read(&uts_sem); |
1965 | + if (copy_to_user(name, &tmp, sizeof(tmp))) |
1966 | + return -EFAULT; |
1967 | |
1968 | - if (!error && override_architecture(name)) |
1969 | - error = -EFAULT; |
1970 | - if (!error && override_release(name->release, sizeof(name->release))) |
1971 | - error = -EFAULT; |
1972 | - return error ? -EFAULT : 0; |
1973 | + if (override_architecture(name)) |
1974 | + return -EFAULT; |
1975 | + if (override_release(name->release, sizeof(name->release))) |
1976 | + return -EFAULT; |
1977 | + return 0; |
1978 | } |
1979 | #endif |
1980 | |
1981 | @@ -1224,17 +1216,18 @@ SYSCALL_DEFINE2(sethostname, char __user *, name, int, len) |
1982 | |
1983 | if (len < 0 || len > __NEW_UTS_LEN) |
1984 | return -EINVAL; |
1985 | - down_write(&uts_sem); |
1986 | errno = -EFAULT; |
1987 | if (!copy_from_user(tmp, name, len)) { |
1988 | - struct new_utsname *u = utsname(); |
1989 | + struct new_utsname *u; |
1990 | |
1991 | + down_write(&uts_sem); |
1992 | + u = utsname(); |
1993 | memcpy(u->nodename, tmp, len); |
1994 | memset(u->nodename + len, 0, sizeof(u->nodename) - len); |
1995 | errno = 0; |
1996 | uts_proc_notify(UTS_PROC_HOSTNAME); |
1997 | + up_write(&uts_sem); |
1998 | } |
1999 | - up_write(&uts_sem); |
2000 | return errno; |
2001 | } |
2002 | |
2003 | @@ -1242,8 +1235,9 @@ SYSCALL_DEFINE2(sethostname, char __user *, name, int, len) |
2004 | |
2005 | SYSCALL_DEFINE2(gethostname, char __user *, name, int, len) |
2006 | { |
2007 | - int i, errno; |
2008 | + int i; |
2009 | struct new_utsname *u; |
2010 | + char tmp[__NEW_UTS_LEN + 1]; |
2011 | |
2012 | if (len < 0) |
2013 | return -EINVAL; |
2014 | @@ -1252,11 +1246,11 @@ SYSCALL_DEFINE2(gethostname, char __user *, name, int, len) |
2015 | i = 1 + strlen(u->nodename); |
2016 | if (i > len) |
2017 | i = len; |
2018 | - errno = 0; |
2019 | - if (copy_to_user(name, u->nodename, i)) |
2020 | - errno = -EFAULT; |
2021 | + memcpy(tmp, u->nodename, i); |
2022 | up_read(&uts_sem); |
2023 | - return errno; |
2024 | + if (copy_to_user(name, tmp, i)) |
2025 | + return -EFAULT; |
2026 | + return 0; |
2027 | } |
2028 | |
2029 | #endif |
2030 | @@ -1275,17 +1269,18 @@ SYSCALL_DEFINE2(setdomainname, char __user *, name, int, len) |
2031 | if (len < 0 || len > __NEW_UTS_LEN) |
2032 | return -EINVAL; |
2033 | |
2034 | - down_write(&uts_sem); |
2035 | errno = -EFAULT; |
2036 | if (!copy_from_user(tmp, name, len)) { |
2037 | - struct new_utsname *u = utsname(); |
2038 | + struct new_utsname *u; |
2039 | |
2040 | + down_write(&uts_sem); |
2041 | + u = utsname(); |
2042 | memcpy(u->domainname, tmp, len); |
2043 | memset(u->domainname + len, 0, sizeof(u->domainname) - len); |
2044 | errno = 0; |
2045 | uts_proc_notify(UTS_PROC_DOMAINNAME); |
2046 | + up_write(&uts_sem); |
2047 | } |
2048 | - up_write(&uts_sem); |
2049 | return errno; |
2050 | } |
2051 | |
2052 | diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c |
2053 | index 4e17d55ba127..bfa8bb3a6e19 100644 |
2054 | --- a/kernel/trace/blktrace.c |
2055 | +++ b/kernel/trace/blktrace.c |
2056 | @@ -1720,6 +1720,10 @@ static ssize_t sysfs_blk_trace_attr_store(struct device *dev, |
2057 | mutex_lock(&bdev->bd_mutex); |
2058 | |
2059 | if (attr == &dev_attr_enable) { |
2060 | + if (!!value == !!q->blk_trace) { |
2061 | + ret = 0; |
2062 | + goto out_unlock_bdev; |
2063 | + } |
2064 | if (value) |
2065 | ret = blk_trace_setup_queue(q, bdev); |
2066 | else |
2067 | diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c |
2068 | index 148c2210a2b8..a47339b156ce 100644 |
2069 | --- a/kernel/trace/trace.c |
2070 | +++ b/kernel/trace/trace.c |
2071 | @@ -6920,7 +6920,9 @@ rb_simple_write(struct file *filp, const char __user *ubuf, |
2072 | |
2073 | if (buffer) { |
2074 | mutex_lock(&trace_types_lock); |
2075 | - if (val) { |
2076 | + if (!!val == tracer_tracing_is_on(tr)) { |
2077 | + val = 0; /* do nothing */ |
2078 | + } else if (val) { |
2079 | tracer_tracing_on(tr); |
2080 | if (tr->current_trace->start) |
2081 | tr->current_trace->start(tr); |
2082 | diff --git a/kernel/trace/trace_uprobe.c b/kernel/trace/trace_uprobe.c |
2083 | index 788262984818..f0ab801a6437 100644 |
2084 | --- a/kernel/trace/trace_uprobe.c |
2085 | +++ b/kernel/trace/trace_uprobe.c |
2086 | @@ -969,7 +969,7 @@ probe_event_disable(struct trace_uprobe *tu, struct trace_event_file *file) |
2087 | |
2088 | list_del_rcu(&link->list); |
2089 | /* synchronize with u{,ret}probe_trace_func */ |
2090 | - synchronize_sched(); |
2091 | + synchronize_rcu(); |
2092 | kfree(link); |
2093 | |
2094 | if (!list_empty(&tu->tp.files)) |
2095 | diff --git a/kernel/user_namespace.c b/kernel/user_namespace.c |
2096 | index 86b7854fec8e..f789bbba9b8e 100644 |
2097 | --- a/kernel/user_namespace.c |
2098 | +++ b/kernel/user_namespace.c |
2099 | @@ -649,7 +649,16 @@ static ssize_t map_write(struct file *file, const char __user *buf, |
2100 | unsigned idx; |
2101 | struct uid_gid_extent *extent = NULL; |
2102 | char *kbuf = NULL, *pos, *next_line; |
2103 | - ssize_t ret = -EINVAL; |
2104 | + ssize_t ret; |
2105 | + |
2106 | + /* Only allow < page size writes at the beginning of the file */ |
2107 | + if ((*ppos != 0) || (count >= PAGE_SIZE)) |
2108 | + return -EINVAL; |
2109 | + |
2110 | + /* Slurp in the user data */ |
2111 | + kbuf = memdup_user_nul(buf, count); |
2112 | + if (IS_ERR(kbuf)) |
2113 | + return PTR_ERR(kbuf); |
2114 | |
2115 | /* |
2116 | * The userns_state_mutex serializes all writes to any given map. |
2117 | @@ -683,19 +692,6 @@ static ssize_t map_write(struct file *file, const char __user *buf, |
2118 | if (cap_valid(cap_setid) && !file_ns_capable(file, ns, CAP_SYS_ADMIN)) |
2119 | goto out; |
2120 | |
2121 | - /* Only allow < page size writes at the beginning of the file */ |
2122 | - ret = -EINVAL; |
2123 | - if ((*ppos != 0) || (count >= PAGE_SIZE)) |
2124 | - goto out; |
2125 | - |
2126 | - /* Slurp in the user data */ |
2127 | - kbuf = memdup_user_nul(buf, count); |
2128 | - if (IS_ERR(kbuf)) { |
2129 | - ret = PTR_ERR(kbuf); |
2130 | - kbuf = NULL; |
2131 | - goto out; |
2132 | - } |
2133 | - |
2134 | /* Parse the user data */ |
2135 | ret = -EINVAL; |
2136 | pos = kbuf; |
2137 | diff --git a/kernel/utsname_sysctl.c b/kernel/utsname_sysctl.c |
2138 | index c8eac43267e9..d2b3b2973456 100644 |
2139 | --- a/kernel/utsname_sysctl.c |
2140 | +++ b/kernel/utsname_sysctl.c |
2141 | @@ -17,7 +17,7 @@ |
2142 | |
2143 | #ifdef CONFIG_PROC_SYSCTL |
2144 | |
2145 | -static void *get_uts(struct ctl_table *table, int write) |
2146 | +static void *get_uts(struct ctl_table *table) |
2147 | { |
2148 | char *which = table->data; |
2149 | struct uts_namespace *uts_ns; |
2150 | @@ -25,21 +25,9 @@ static void *get_uts(struct ctl_table *table, int write) |
2151 | uts_ns = current->nsproxy->uts_ns; |
2152 | which = (which - (char *)&init_uts_ns) + (char *)uts_ns; |
2153 | |
2154 | - if (!write) |
2155 | - down_read(&uts_sem); |
2156 | - else |
2157 | - down_write(&uts_sem); |
2158 | return which; |
2159 | } |
2160 | |
2161 | -static void put_uts(struct ctl_table *table, int write, void *which) |
2162 | -{ |
2163 | - if (!write) |
2164 | - up_read(&uts_sem); |
2165 | - else |
2166 | - up_write(&uts_sem); |
2167 | -} |
2168 | - |
2169 | /* |
2170 | * Special case of dostring for the UTS structure. This has locks |
2171 | * to observe. Should this be in kernel/sys.c ???? |
2172 | @@ -49,13 +37,34 @@ static int proc_do_uts_string(struct ctl_table *table, int write, |
2173 | { |
2174 | struct ctl_table uts_table; |
2175 | int r; |
2176 | + char tmp_data[__NEW_UTS_LEN + 1]; |
2177 | + |
2178 | memcpy(&uts_table, table, sizeof(uts_table)); |
2179 | - uts_table.data = get_uts(table, write); |
2180 | + uts_table.data = tmp_data; |
2181 | + |
2182 | + /* |
2183 | + * Buffer the value in tmp_data so that proc_dostring() can be called |
2184 | + * without holding any locks. |
2185 | + * We also need to read the original value in the write==1 case to |
2186 | + * support partial writes. |
2187 | + */ |
2188 | + down_read(&uts_sem); |
2189 | + memcpy(tmp_data, get_uts(table), sizeof(tmp_data)); |
2190 | + up_read(&uts_sem); |
2191 | r = proc_dostring(&uts_table, write, buffer, lenp, ppos); |
2192 | - put_uts(table, write, uts_table.data); |
2193 | |
2194 | - if (write) |
2195 | + if (write) { |
2196 | + /* |
2197 | + * Write back the new value. |
2198 | + * Note that, since we dropped uts_sem, the result can |
2199 | + * theoretically be incorrect if there are two parallel writes |
2200 | + * at non-zero offsets to the same sysctl. |
2201 | + */ |
2202 | + down_write(&uts_sem); |
2203 | + memcpy(get_uts(table), tmp_data, sizeof(tmp_data)); |
2204 | + up_write(&uts_sem); |
2205 | proc_sys_poll_notify(table->poll); |
2206 | + } |
2207 | |
2208 | return r; |
2209 | } |
2210 | diff --git a/mm/memory.c b/mm/memory.c |
2211 | index 0ff735601654..f3fef1df7402 100644 |
2212 | --- a/mm/memory.c |
2213 | +++ b/mm/memory.c |
2214 | @@ -373,15 +373,6 @@ void tlb_remove_table(struct mmu_gather *tlb, void *table) |
2215 | { |
2216 | struct mmu_table_batch **batch = &tlb->batch; |
2217 | |
2218 | - /* |
2219 | - * When there's less then two users of this mm there cannot be a |
2220 | - * concurrent page-table walk. |
2221 | - */ |
2222 | - if (atomic_read(&tlb->mm->mm_users) < 2) { |
2223 | - __tlb_remove_table(table); |
2224 | - return; |
2225 | - } |
2226 | - |
2227 | if (*batch == NULL) { |
2228 | *batch = (struct mmu_table_batch *)__get_free_page(GFP_NOWAIT | __GFP_NOWARN); |
2229 | if (*batch == NULL) { |
2230 | diff --git a/net/9p/client.c b/net/9p/client.c |
2231 | index 1fd60190177e..98d299ea52ee 100644 |
2232 | --- a/net/9p/client.c |
2233 | +++ b/net/9p/client.c |
2234 | @@ -931,7 +931,7 @@ static int p9_client_version(struct p9_client *c) |
2235 | { |
2236 | int err = 0; |
2237 | struct p9_req_t *req; |
2238 | - char *version; |
2239 | + char *version = NULL; |
2240 | int msize; |
2241 | |
2242 | p9_debug(P9_DEBUG_9P, ">>> TVERSION msize %d protocol %d\n", |
2243 | diff --git a/net/9p/trans_fd.c b/net/9p/trans_fd.c |
2244 | index 7bc2208b6cc4..2b543532e2f1 100644 |
2245 | --- a/net/9p/trans_fd.c |
2246 | +++ b/net/9p/trans_fd.c |
2247 | @@ -181,6 +181,8 @@ static void p9_mux_poll_stop(struct p9_conn *m) |
2248 | spin_lock_irqsave(&p9_poll_lock, flags); |
2249 | list_del_init(&m->poll_pending_link); |
2250 | spin_unlock_irqrestore(&p9_poll_lock, flags); |
2251 | + |
2252 | + flush_work(&p9_poll_work); |
2253 | } |
2254 | |
2255 | /** |
2256 | @@ -937,7 +939,7 @@ p9_fd_create_tcp(struct p9_client *client, const char *addr, char *args) |
2257 | if (err < 0) |
2258 | return err; |
2259 | |
2260 | - if (valid_ipaddr4(addr) < 0) |
2261 | + if (addr == NULL || valid_ipaddr4(addr) < 0) |
2262 | return -EINVAL; |
2263 | |
2264 | csocket = NULL; |
2265 | @@ -985,6 +987,9 @@ p9_fd_create_unix(struct p9_client *client, const char *addr, char *args) |
2266 | |
2267 | csocket = NULL; |
2268 | |
2269 | + if (addr == NULL) |
2270 | + return -EINVAL; |
2271 | + |
2272 | if (strlen(addr) >= UNIX_PATH_MAX) { |
2273 | pr_err("%s (%d): address too long: %s\n", |
2274 | __func__, task_pid_nr(current), addr); |
2275 | diff --git a/net/9p/trans_rdma.c b/net/9p/trans_rdma.c |
2276 | index 553ed4ecb6a0..5a2ad4707463 100644 |
2277 | --- a/net/9p/trans_rdma.c |
2278 | +++ b/net/9p/trans_rdma.c |
2279 | @@ -622,6 +622,9 @@ rdma_create_trans(struct p9_client *client, const char *addr, char *args) |
2280 | struct rdma_conn_param conn_param; |
2281 | struct ib_qp_init_attr qp_attr; |
2282 | |
2283 | + if (addr == NULL) |
2284 | + return -EINVAL; |
2285 | + |
2286 | /* Parse the transport specific mount options */ |
2287 | err = parse_opts(args, &opts); |
2288 | if (err < 0) |
2289 | diff --git a/net/9p/trans_virtio.c b/net/9p/trans_virtio.c |
2290 | index 3aa5a93ad107..da0d3b257459 100644 |
2291 | --- a/net/9p/trans_virtio.c |
2292 | +++ b/net/9p/trans_virtio.c |
2293 | @@ -189,7 +189,7 @@ static int pack_sg_list(struct scatterlist *sg, int start, |
2294 | s = rest_of_page(data); |
2295 | if (s > count) |
2296 | s = count; |
2297 | - BUG_ON(index > limit); |
2298 | + BUG_ON(index >= limit); |
2299 | /* Make sure we don't terminate early. */ |
2300 | sg_unmark_end(&sg[index]); |
2301 | sg_set_buf(&sg[index++], data, s); |
2302 | @@ -234,6 +234,7 @@ pack_sg_list_p(struct scatterlist *sg, int start, int limit, |
2303 | s = PAGE_SIZE - data_off; |
2304 | if (s > count) |
2305 | s = count; |
2306 | + BUG_ON(index >= limit); |
2307 | /* Make sure we don't terminate early. */ |
2308 | sg_unmark_end(&sg[index]); |
2309 | sg_set_page(&sg[index++], pdata[i++], s, data_off); |
2310 | @@ -406,6 +407,7 @@ p9_virtio_zc_request(struct p9_client *client, struct p9_req_t *req, |
2311 | p9_debug(P9_DEBUG_TRANS, "virtio request\n"); |
2312 | |
2313 | if (uodata) { |
2314 | + __le32 sz; |
2315 | int n = p9_get_mapped_pages(chan, &out_pages, uodata, |
2316 | outlen, &offs, &need_drop); |
2317 | if (n < 0) |
2318 | @@ -416,6 +418,12 @@ p9_virtio_zc_request(struct p9_client *client, struct p9_req_t *req, |
2319 | memcpy(&req->tc->sdata[req->tc->size - 4], &v, 4); |
2320 | outlen = n; |
2321 | } |
2322 | + /* The size field of the message must include the length of the |
2323 | + * header and the length of the data. We didn't actually know |
2324 | + * the length of the data until this point so add it in now. |
2325 | + */ |
2326 | + sz = cpu_to_le32(req->tc->size + outlen); |
2327 | + memcpy(&req->tc->sdata[0], &sz, sizeof(sz)); |
2328 | } else if (uidata) { |
2329 | int n = p9_get_mapped_pages(chan, &in_pages, uidata, |
2330 | inlen, &offs, &need_drop); |
2331 | @@ -643,6 +651,9 @@ p9_virtio_create(struct p9_client *client, const char *devname, char *args) |
2332 | int ret = -ENOENT; |
2333 | int found = 0; |
2334 | |
2335 | + if (devname == NULL) |
2336 | + return -EINVAL; |
2337 | + |
2338 | mutex_lock(&virtio_9p_lock); |
2339 | list_for_each_entry(chan, &virtio_chan_list, chan_list) { |
2340 | if (!strncmp(devname, chan->tag, chan->tag_len) && |
2341 | diff --git a/net/ieee802154/6lowpan/tx.c b/net/ieee802154/6lowpan/tx.c |
2342 | index dbb476d7d38f..50ed47559bb7 100644 |
2343 | --- a/net/ieee802154/6lowpan/tx.c |
2344 | +++ b/net/ieee802154/6lowpan/tx.c |
2345 | @@ -266,9 +266,24 @@ netdev_tx_t lowpan_xmit(struct sk_buff *skb, struct net_device *ldev) |
2346 | /* We must take a copy of the skb before we modify/replace the ipv6 |
2347 | * header as the header could be used elsewhere |
2348 | */ |
2349 | - skb = skb_unshare(skb, GFP_ATOMIC); |
2350 | - if (!skb) |
2351 | - return NET_XMIT_DROP; |
2352 | + if (unlikely(skb_headroom(skb) < ldev->needed_headroom || |
2353 | + skb_tailroom(skb) < ldev->needed_tailroom)) { |
2354 | + struct sk_buff *nskb; |
2355 | + |
2356 | + nskb = skb_copy_expand(skb, ldev->needed_headroom, |
2357 | + ldev->needed_tailroom, GFP_ATOMIC); |
2358 | + if (likely(nskb)) { |
2359 | + consume_skb(skb); |
2360 | + skb = nskb; |
2361 | + } else { |
2362 | + kfree_skb(skb); |
2363 | + return NET_XMIT_DROP; |
2364 | + } |
2365 | + } else { |
2366 | + skb = skb_unshare(skb, GFP_ATOMIC); |
2367 | + if (!skb) |
2368 | + return NET_XMIT_DROP; |
2369 | + } |
2370 | |
2371 | ret = lowpan_header(skb, ldev, &dgram_size, &dgram_offset); |
2372 | if (ret < 0) { |
2373 | diff --git a/net/mac802154/tx.c b/net/mac802154/tx.c |
2374 | index 7e253455f9dd..bcd1a5e6ebf4 100644 |
2375 | --- a/net/mac802154/tx.c |
2376 | +++ b/net/mac802154/tx.c |
2377 | @@ -63,8 +63,21 @@ ieee802154_tx(struct ieee802154_local *local, struct sk_buff *skb) |
2378 | int ret; |
2379 | |
2380 | if (!(local->hw.flags & IEEE802154_HW_TX_OMIT_CKSUM)) { |
2381 | - u16 crc = crc_ccitt(0, skb->data, skb->len); |
2382 | + struct sk_buff *nskb; |
2383 | + u16 crc; |
2384 | + |
2385 | + if (unlikely(skb_tailroom(skb) < IEEE802154_FCS_LEN)) { |
2386 | + nskb = skb_copy_expand(skb, 0, IEEE802154_FCS_LEN, |
2387 | + GFP_ATOMIC); |
2388 | + if (likely(nskb)) { |
2389 | + consume_skb(skb); |
2390 | + skb = nskb; |
2391 | + } else { |
2392 | + goto err_tx; |
2393 | + } |
2394 | + } |
2395 | |
2396 | + crc = crc_ccitt(0, skb->data, skb->len); |
2397 | put_unaligned_le16(crc, skb_put(skb, 2)); |
2398 | } |
2399 | |
2400 | diff --git a/net/sunrpc/clnt.c b/net/sunrpc/clnt.c |
2401 | index b2ae4f150ec6..244eac1bd648 100644 |
2402 | --- a/net/sunrpc/clnt.c |
2403 | +++ b/net/sunrpc/clnt.c |
2404 | @@ -965,10 +965,20 @@ out: |
2405 | } |
2406 | EXPORT_SYMBOL_GPL(rpc_bind_new_program); |
2407 | |
2408 | +void rpc_task_release_transport(struct rpc_task *task) |
2409 | +{ |
2410 | + struct rpc_xprt *xprt = task->tk_xprt; |
2411 | + |
2412 | + if (xprt) { |
2413 | + task->tk_xprt = NULL; |
2414 | + xprt_put(xprt); |
2415 | + } |
2416 | +} |
2417 | +EXPORT_SYMBOL_GPL(rpc_task_release_transport); |
2418 | + |
2419 | void rpc_task_release_client(struct rpc_task *task) |
2420 | { |
2421 | struct rpc_clnt *clnt = task->tk_client; |
2422 | - struct rpc_xprt *xprt = task->tk_xprt; |
2423 | |
2424 | if (clnt != NULL) { |
2425 | /* Remove from client task list */ |
2426 | @@ -979,12 +989,14 @@ void rpc_task_release_client(struct rpc_task *task) |
2427 | |
2428 | rpc_release_client(clnt); |
2429 | } |
2430 | + rpc_task_release_transport(task); |
2431 | +} |
2432 | |
2433 | - if (xprt != NULL) { |
2434 | - task->tk_xprt = NULL; |
2435 | - |
2436 | - xprt_put(xprt); |
2437 | - } |
2438 | +static |
2439 | +void rpc_task_set_transport(struct rpc_task *task, struct rpc_clnt *clnt) |
2440 | +{ |
2441 | + if (!task->tk_xprt) |
2442 | + task->tk_xprt = xprt_iter_get_next(&clnt->cl_xpi); |
2443 | } |
2444 | |
2445 | static |
2446 | @@ -992,8 +1004,7 @@ void rpc_task_set_client(struct rpc_task *task, struct rpc_clnt *clnt) |
2447 | { |
2448 | |
2449 | if (clnt != NULL) { |
2450 | - if (task->tk_xprt == NULL) |
2451 | - task->tk_xprt = xprt_iter_get_next(&clnt->cl_xpi); |
2452 | + rpc_task_set_transport(task, clnt); |
2453 | task->tk_client = clnt; |
2454 | atomic_inc(&clnt->cl_count); |
2455 | if (clnt->cl_softrtry) |
2456 | @@ -1550,6 +1561,7 @@ call_start(struct rpc_task *task) |
2457 | task->tk_msg.rpc_proc->p_count++; |
2458 | clnt->cl_stats->rpccnt++; |
2459 | task->tk_action = call_reserve; |
2460 | + rpc_task_set_transport(task, clnt); |
2461 | } |
2462 | |
2463 | /* |
2464 | diff --git a/tools/perf/util/auxtrace.c b/tools/perf/util/auxtrace.c |
2465 | index 78bd632f144d..29d015e2d900 100644 |
2466 | --- a/tools/perf/util/auxtrace.c |
2467 | +++ b/tools/perf/util/auxtrace.c |
2468 | @@ -195,6 +195,9 @@ static int auxtrace_queues__grow(struct auxtrace_queues *queues, |
2469 | for (i = 0; i < queues->nr_queues; i++) { |
2470 | list_splice_tail(&queues->queue_array[i].head, |
2471 | &queue_array[i].head); |
2472 | + queue_array[i].tid = queues->queue_array[i].tid; |
2473 | + queue_array[i].cpu = queues->queue_array[i].cpu; |
2474 | + queue_array[i].set = queues->queue_array[i].set; |
2475 | queue_array[i].priv = queues->queue_array[i].priv; |
2476 | } |
2477 |