Annotation of /trunk/kernel-magellan/patches-3.12/0107-3.12.8-all-fixes.patch
Parent Directory | Revision Log
Revision 2388 -
(hide annotations)
(download)
Tue Jan 28 08:29:18 2014 UTC (10 years, 8 months ago) by niro
File size: 87546 byte(s)
Tue Jan 28 08:29:18 2014 UTC (10 years, 8 months ago) by niro
File size: 87546 byte(s)
-linux-3.12.8
1 | niro | 2388 | diff --git a/Documentation/devicetree/bindings/clock/exynos5250-clock.txt b/Documentation/devicetree/bindings/clock/exynos5250-clock.txt |
2 | index 24765c146e31..5883ec878ddd 100644 | ||
3 | --- a/Documentation/devicetree/bindings/clock/exynos5250-clock.txt | ||
4 | +++ b/Documentation/devicetree/bindings/clock/exynos5250-clock.txt | ||
5 | @@ -159,6 +159,8 @@ clock which they consume. | ||
6 | mixer 343 | ||
7 | hdmi 344 | ||
8 | g2d 345 | ||
9 | + mdma0 346 | ||
10 | + smmu_mdma0 347 | ||
11 | |||
12 | |||
13 | [Clock Muxes] | ||
14 | diff --git a/Documentation/networking/packet_mmap.txt b/Documentation/networking/packet_mmap.txt | ||
15 | index c01223628a87..8e48e3b14227 100644 | ||
16 | --- a/Documentation/networking/packet_mmap.txt | ||
17 | +++ b/Documentation/networking/packet_mmap.txt | ||
18 | @@ -123,6 +123,16 @@ Transmission process is similar to capture as shown below. | ||
19 | [shutdown] close() --------> destruction of the transmission socket and | ||
20 | deallocation of all associated resources. | ||
21 | |||
22 | +Socket creation and destruction is also straight forward, and is done | ||
23 | +the same way as in capturing described in the previous paragraph: | ||
24 | + | ||
25 | + int fd = socket(PF_PACKET, mode, 0); | ||
26 | + | ||
27 | +The protocol can optionally be 0 in case we only want to transmit | ||
28 | +via this socket, which avoids an expensive call to packet_rcv(). | ||
29 | +In this case, you also need to bind(2) the TX_RING with sll_protocol = 0 | ||
30 | +set. Otherwise, htons(ETH_P_ALL) or any other protocol, for example. | ||
31 | + | ||
32 | Binding the socket to your network interface is mandatory (with zero copy) to | ||
33 | know the header size of frames used in the circular buffer. | ||
34 | |||
35 | diff --git a/Makefile b/Makefile | ||
36 | index c2f0b7985b41..5d0ec13bb77d 100644 | ||
37 | --- a/Makefile | ||
38 | +++ b/Makefile | ||
39 | @@ -1,6 +1,6 @@ | ||
40 | VERSION = 3 | ||
41 | PATCHLEVEL = 12 | ||
42 | -SUBLEVEL = 7 | ||
43 | +SUBLEVEL = 8 | ||
44 | EXTRAVERSION = | ||
45 | NAME = One Giant Leap for Frogkind | ||
46 | |||
47 | diff --git a/arch/arm/boot/dts/exynos5250.dtsi b/arch/arm/boot/dts/exynos5250.dtsi | ||
48 | index bbac42a78ce5..6a1117e481b1 100644 | ||
49 | --- a/arch/arm/boot/dts/exynos5250.dtsi | ||
50 | +++ b/arch/arm/boot/dts/exynos5250.dtsi | ||
51 | @@ -556,7 +556,7 @@ | ||
52 | compatible = "arm,pl330", "arm,primecell"; | ||
53 | reg = <0x10800000 0x1000>; | ||
54 | interrupts = <0 33 0>; | ||
55 | - clocks = <&clock 271>; | ||
56 | + clocks = <&clock 346>; | ||
57 | clock-names = "apb_pclk"; | ||
58 | #dma-cells = <1>; | ||
59 | #dma-channels = <8>; | ||
60 | diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c | ||
61 | index 65ed63f68ef8..1f735aafd5ec 100644 | ||
62 | --- a/arch/arm/kernel/traps.c | ||
63 | +++ b/arch/arm/kernel/traps.c | ||
64 | @@ -35,7 +35,13 @@ | ||
65 | #include <asm/tls.h> | ||
66 | #include <asm/system_misc.h> | ||
67 | |||
68 | -static const char *handler[]= { "prefetch abort", "data abort", "address exception", "interrupt" }; | ||
69 | +static const char *handler[]= { | ||
70 | + "prefetch abort", | ||
71 | + "data abort", | ||
72 | + "address exception", | ||
73 | + "interrupt", | ||
74 | + "undefined instruction", | ||
75 | +}; | ||
76 | |||
77 | void *vectors_page; | ||
78 | |||
79 | diff --git a/arch/arm/mach-footbridge/dc21285-timer.c b/arch/arm/mach-footbridge/dc21285-timer.c | ||
80 | index 9ee78f7b4990..782f6c71fa0a 100644 | ||
81 | --- a/arch/arm/mach-footbridge/dc21285-timer.c | ||
82 | +++ b/arch/arm/mach-footbridge/dc21285-timer.c | ||
83 | @@ -96,11 +96,12 @@ static struct irqaction footbridge_timer_irq = { | ||
84 | void __init footbridge_timer_init(void) | ||
85 | { | ||
86 | struct clock_event_device *ce = &ckevt_dc21285; | ||
87 | + unsigned rate = DIV_ROUND_CLOSEST(mem_fclk_21285, 16); | ||
88 | |||
89 | - clocksource_register_hz(&cksrc_dc21285, (mem_fclk_21285 + 8) / 16); | ||
90 | + clocksource_register_hz(&cksrc_dc21285, rate); | ||
91 | |||
92 | setup_irq(ce->irq, &footbridge_timer_irq); | ||
93 | |||
94 | ce->cpumask = cpumask_of(smp_processor_id()); | ||
95 | - clockevents_config_and_register(ce, mem_fclk_21285, 0x4, 0xffffff); | ||
96 | + clockevents_config_and_register(ce, rate, 0x4, 0xffffff); | ||
97 | } | ||
98 | diff --git a/arch/arm/mach-shmobile/board-armadillo800eva.c b/arch/arm/mach-shmobile/board-armadillo800eva.c | ||
99 | index 7f8f6076d360..07b91832bd2c 100644 | ||
100 | --- a/arch/arm/mach-shmobile/board-armadillo800eva.c | ||
101 | +++ b/arch/arm/mach-shmobile/board-armadillo800eva.c | ||
102 | @@ -482,7 +482,7 @@ static struct platform_device lcdc0_device = { | ||
103 | .id = 0, | ||
104 | .dev = { | ||
105 | .platform_data = &lcdc0_info, | ||
106 | - .coherent_dma_mask = ~0, | ||
107 | + .coherent_dma_mask = DMA_BIT_MASK(32), | ||
108 | }, | ||
109 | }; | ||
110 | |||
111 | @@ -579,7 +579,7 @@ static struct platform_device hdmi_lcdc_device = { | ||
112 | .id = 1, | ||
113 | .dev = { | ||
114 | .platform_data = &hdmi_lcdc_info, | ||
115 | - .coherent_dma_mask = ~0, | ||
116 | + .coherent_dma_mask = DMA_BIT_MASK(32), | ||
117 | }, | ||
118 | }; | ||
119 | |||
120 | diff --git a/arch/arm/mach-shmobile/board-kzm9g.c b/arch/arm/mach-shmobile/board-kzm9g.c | ||
121 | index f1994968d303..5cc96d004b4f 100644 | ||
122 | --- a/arch/arm/mach-shmobile/board-kzm9g.c | ||
123 | +++ b/arch/arm/mach-shmobile/board-kzm9g.c | ||
124 | @@ -334,7 +334,7 @@ static struct platform_device lcdc_device = { | ||
125 | .resource = lcdc_resources, | ||
126 | .dev = { | ||
127 | .platform_data = &lcdc_info, | ||
128 | - .coherent_dma_mask = ~0, | ||
129 | + .coherent_dma_mask = DMA_BIT_MASK(32), | ||
130 | }, | ||
131 | }; | ||
132 | |||
133 | diff --git a/arch/arm/mach-shmobile/board-mackerel.c b/arch/arm/mach-shmobile/board-mackerel.c | ||
134 | index af06753eb809..e721d2ccceae 100644 | ||
135 | --- a/arch/arm/mach-shmobile/board-mackerel.c | ||
136 | +++ b/arch/arm/mach-shmobile/board-mackerel.c | ||
137 | @@ -409,7 +409,7 @@ static struct platform_device lcdc_device = { | ||
138 | .resource = lcdc_resources, | ||
139 | .dev = { | ||
140 | .platform_data = &lcdc_info, | ||
141 | - .coherent_dma_mask = ~0, | ||
142 | + .coherent_dma_mask = DMA_BIT_MASK(32), | ||
143 | }, | ||
144 | }; | ||
145 | |||
146 | @@ -499,7 +499,7 @@ static struct platform_device hdmi_lcdc_device = { | ||
147 | .id = 1, | ||
148 | .dev = { | ||
149 | .platform_data = &hdmi_lcdc_info, | ||
150 | - .coherent_dma_mask = ~0, | ||
151 | + .coherent_dma_mask = DMA_BIT_MASK(32), | ||
152 | }, | ||
153 | }; | ||
154 | |||
155 | diff --git a/arch/arm/mm/flush.c b/arch/arm/mm/flush.c | ||
156 | index 6d5ba9afb16a..3387e60e4ea3 100644 | ||
157 | --- a/arch/arm/mm/flush.c | ||
158 | +++ b/arch/arm/mm/flush.c | ||
159 | @@ -175,16 +175,16 @@ void __flush_dcache_page(struct address_space *mapping, struct page *page) | ||
160 | unsigned long i; | ||
161 | if (cache_is_vipt_nonaliasing()) { | ||
162 | for (i = 0; i < (1 << compound_order(page)); i++) { | ||
163 | - void *addr = kmap_atomic(page); | ||
164 | + void *addr = kmap_atomic(page + i); | ||
165 | __cpuc_flush_dcache_area(addr, PAGE_SIZE); | ||
166 | kunmap_atomic(addr); | ||
167 | } | ||
168 | } else { | ||
169 | for (i = 0; i < (1 << compound_order(page)); i++) { | ||
170 | - void *addr = kmap_high_get(page); | ||
171 | + void *addr = kmap_high_get(page + i); | ||
172 | if (addr) { | ||
173 | __cpuc_flush_dcache_area(addr, PAGE_SIZE); | ||
174 | - kunmap_high(page); | ||
175 | + kunmap_high(page + i); | ||
176 | } | ||
177 | } | ||
178 | } | ||
179 | diff --git a/arch/parisc/include/asm/cacheflush.h b/arch/parisc/include/asm/cacheflush.h | ||
180 | index f0e2784e7cca..2f9b751878ba 100644 | ||
181 | --- a/arch/parisc/include/asm/cacheflush.h | ||
182 | +++ b/arch/parisc/include/asm/cacheflush.h | ||
183 | @@ -125,42 +125,38 @@ flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma | ||
184 | void mark_rodata_ro(void); | ||
185 | #endif | ||
186 | |||
187 | -#ifdef CONFIG_PA8X00 | ||
188 | -/* Only pa8800, pa8900 needs this */ | ||
189 | - | ||
190 | #include <asm/kmap_types.h> | ||
191 | |||
192 | #define ARCH_HAS_KMAP | ||
193 | |||
194 | -void kunmap_parisc(void *addr); | ||
195 | - | ||
196 | static inline void *kmap(struct page *page) | ||
197 | { | ||
198 | might_sleep(); | ||
199 | + flush_dcache_page(page); | ||
200 | return page_address(page); | ||
201 | } | ||
202 | |||
203 | static inline void kunmap(struct page *page) | ||
204 | { | ||
205 | - kunmap_parisc(page_address(page)); | ||
206 | + flush_kernel_dcache_page_addr(page_address(page)); | ||
207 | } | ||
208 | |||
209 | static inline void *kmap_atomic(struct page *page) | ||
210 | { | ||
211 | pagefault_disable(); | ||
212 | + flush_dcache_page(page); | ||
213 | return page_address(page); | ||
214 | } | ||
215 | |||
216 | static inline void __kunmap_atomic(void *addr) | ||
217 | { | ||
218 | - kunmap_parisc(addr); | ||
219 | + flush_kernel_dcache_page_addr(addr); | ||
220 | pagefault_enable(); | ||
221 | } | ||
222 | |||
223 | #define kmap_atomic_prot(page, prot) kmap_atomic(page) | ||
224 | #define kmap_atomic_pfn(pfn) kmap_atomic(pfn_to_page(pfn)) | ||
225 | #define kmap_atomic_to_page(ptr) virt_to_page(ptr) | ||
226 | -#endif | ||
227 | |||
228 | #endif /* _PARISC_CACHEFLUSH_H */ | ||
229 | |||
230 | diff --git a/arch/parisc/include/asm/page.h b/arch/parisc/include/asm/page.h | ||
231 | index b7adb2ac049c..c53fc63149e8 100644 | ||
232 | --- a/arch/parisc/include/asm/page.h | ||
233 | +++ b/arch/parisc/include/asm/page.h | ||
234 | @@ -28,9 +28,8 @@ struct page; | ||
235 | |||
236 | void clear_page_asm(void *page); | ||
237 | void copy_page_asm(void *to, void *from); | ||
238 | -void clear_user_page(void *vto, unsigned long vaddr, struct page *pg); | ||
239 | -void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, | ||
240 | - struct page *pg); | ||
241 | +#define clear_user_page(vto, vaddr, page) clear_page_asm(vto) | ||
242 | +#define copy_user_page(vto, vfrom, vaddr, page) copy_page_asm(vto, vfrom) | ||
243 | |||
244 | /* #define CONFIG_PARISC_TMPALIAS */ | ||
245 | |||
246 | diff --git a/arch/parisc/kernel/cache.c b/arch/parisc/kernel/cache.c | ||
247 | index c035673209f7..a72545554a31 100644 | ||
248 | --- a/arch/parisc/kernel/cache.c | ||
249 | +++ b/arch/parisc/kernel/cache.c | ||
250 | @@ -388,41 +388,6 @@ void flush_kernel_dcache_page_addr(void *addr) | ||
251 | } | ||
252 | EXPORT_SYMBOL(flush_kernel_dcache_page_addr); | ||
253 | |||
254 | -void clear_user_page(void *vto, unsigned long vaddr, struct page *page) | ||
255 | -{ | ||
256 | - clear_page_asm(vto); | ||
257 | - if (!parisc_requires_coherency()) | ||
258 | - flush_kernel_dcache_page_asm(vto); | ||
259 | -} | ||
260 | -EXPORT_SYMBOL(clear_user_page); | ||
261 | - | ||
262 | -void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, | ||
263 | - struct page *pg) | ||
264 | -{ | ||
265 | - /* Copy using kernel mapping. No coherency is needed | ||
266 | - (all in kmap/kunmap) on machines that don't support | ||
267 | - non-equivalent aliasing. However, the `from' page | ||
268 | - needs to be flushed before it can be accessed through | ||
269 | - the kernel mapping. */ | ||
270 | - preempt_disable(); | ||
271 | - flush_dcache_page_asm(__pa(vfrom), vaddr); | ||
272 | - preempt_enable(); | ||
273 | - copy_page_asm(vto, vfrom); | ||
274 | - if (!parisc_requires_coherency()) | ||
275 | - flush_kernel_dcache_page_asm(vto); | ||
276 | -} | ||
277 | -EXPORT_SYMBOL(copy_user_page); | ||
278 | - | ||
279 | -#ifdef CONFIG_PA8X00 | ||
280 | - | ||
281 | -void kunmap_parisc(void *addr) | ||
282 | -{ | ||
283 | - if (parisc_requires_coherency()) | ||
284 | - flush_kernel_dcache_page_addr(addr); | ||
285 | -} | ||
286 | -EXPORT_SYMBOL(kunmap_parisc); | ||
287 | -#endif | ||
288 | - | ||
289 | void purge_tlb_entries(struct mm_struct *mm, unsigned long addr) | ||
290 | { | ||
291 | unsigned long flags; | ||
292 | diff --git a/arch/x86/include/asm/fpu-internal.h b/arch/x86/include/asm/fpu-internal.h | ||
293 | index 4d0bda7b11e3..5be9f879957f 100644 | ||
294 | --- a/arch/x86/include/asm/fpu-internal.h | ||
295 | +++ b/arch/x86/include/asm/fpu-internal.h | ||
296 | @@ -293,12 +293,13 @@ static inline int restore_fpu_checking(struct task_struct *tsk) | ||
297 | /* AMD K7/K8 CPUs don't save/restore FDP/FIP/FOP unless an exception | ||
298 | is pending. Clear the x87 state here by setting it to fixed | ||
299 | values. "m" is a random variable that should be in L1 */ | ||
300 | - alternative_input( | ||
301 | - ASM_NOP8 ASM_NOP2, | ||
302 | - "emms\n\t" /* clear stack tags */ | ||
303 | - "fildl %P[addr]", /* set F?P to defined value */ | ||
304 | - X86_FEATURE_FXSAVE_LEAK, | ||
305 | - [addr] "m" (tsk->thread.fpu.has_fpu)); | ||
306 | + if (unlikely(static_cpu_has(X86_FEATURE_FXSAVE_LEAK))) { | ||
307 | + asm volatile( | ||
308 | + "fnclex\n\t" | ||
309 | + "emms\n\t" | ||
310 | + "fildl %P[addr]" /* set F?P to defined value */ | ||
311 | + : : [addr] "m" (tsk->thread.fpu.has_fpu)); | ||
312 | + } | ||
313 | |||
314 | return fpu_restore_checking(&tsk->thread.fpu); | ||
315 | } | ||
316 | diff --git a/drivers/acpi/battery.c b/drivers/acpi/battery.c | ||
317 | index 2c9958cd7a43..ffa5af4c221a 100644 | ||
318 | --- a/drivers/acpi/battery.c | ||
319 | +++ b/drivers/acpi/battery.c | ||
320 | @@ -68,6 +68,7 @@ MODULE_AUTHOR("Alexey Starikovskiy <astarikovskiy@suse.de>"); | ||
321 | MODULE_DESCRIPTION("ACPI Battery Driver"); | ||
322 | MODULE_LICENSE("GPL"); | ||
323 | |||
324 | +static int battery_bix_broken_package; | ||
325 | static unsigned int cache_time = 1000; | ||
326 | module_param(cache_time, uint, 0644); | ||
327 | MODULE_PARM_DESC(cache_time, "cache time in milliseconds"); | ||
328 | @@ -443,7 +444,12 @@ static int acpi_battery_get_info(struct acpi_battery *battery) | ||
329 | ACPI_EXCEPTION((AE_INFO, status, "Evaluating %s", name)); | ||
330 | return -ENODEV; | ||
331 | } | ||
332 | - if (test_bit(ACPI_BATTERY_XINFO_PRESENT, &battery->flags)) | ||
333 | + | ||
334 | + if (battery_bix_broken_package) | ||
335 | + result = extract_package(battery, buffer.pointer, | ||
336 | + extended_info_offsets + 1, | ||
337 | + ARRAY_SIZE(extended_info_offsets) - 1); | ||
338 | + else if (test_bit(ACPI_BATTERY_XINFO_PRESENT, &battery->flags)) | ||
339 | result = extract_package(battery, buffer.pointer, | ||
340 | extended_info_offsets, | ||
341 | ARRAY_SIZE(extended_info_offsets)); | ||
342 | @@ -1054,6 +1060,17 @@ static int battery_notify(struct notifier_block *nb, | ||
343 | return 0; | ||
344 | } | ||
345 | |||
346 | +static struct dmi_system_id bat_dmi_table[] = { | ||
347 | + { | ||
348 | + .ident = "NEC LZ750/LS", | ||
349 | + .matches = { | ||
350 | + DMI_MATCH(DMI_SYS_VENDOR, "NEC"), | ||
351 | + DMI_MATCH(DMI_PRODUCT_NAME, "PC-LZ750LS"), | ||
352 | + }, | ||
353 | + }, | ||
354 | + {}, | ||
355 | +}; | ||
356 | + | ||
357 | static int acpi_battery_add(struct acpi_device *device) | ||
358 | { | ||
359 | int result = 0; | ||
360 | @@ -1163,6 +1180,8 @@ static void __init acpi_battery_init_async(void *unused, async_cookie_t cookie) | ||
361 | if (!acpi_battery_dir) | ||
362 | return; | ||
363 | #endif | ||
364 | + if (dmi_check_system(bat_dmi_table)) | ||
365 | + battery_bix_broken_package = 1; | ||
366 | if (acpi_bus_register_driver(&acpi_battery_driver) < 0) { | ||
367 | #ifdef CONFIG_ACPI_PROCFS_POWER | ||
368 | acpi_unlock_battery_dir(acpi_battery_dir); | ||
369 | diff --git a/drivers/ata/ahci.c b/drivers/ata/ahci.c | ||
370 | index 14f1e9506338..cfc6073c0487 100644 | ||
371 | --- a/drivers/ata/ahci.c | ||
372 | +++ b/drivers/ata/ahci.c | ||
373 | @@ -427,6 +427,9 @@ static const struct pci_device_id ahci_pci_tbl[] = { | ||
374 | .driver_data = board_ahci_yes_fbs }, /* 88se9128 */ | ||
375 | { PCI_DEVICE(PCI_VENDOR_ID_MARVELL_EXT, 0x9125), | ||
376 | .driver_data = board_ahci_yes_fbs }, /* 88se9125 */ | ||
377 | + { PCI_DEVICE_SUB(PCI_VENDOR_ID_MARVELL_EXT, 0x9178, | ||
378 | + PCI_VENDOR_ID_MARVELL_EXT, 0x9170), | ||
379 | + .driver_data = board_ahci_yes_fbs }, /* 88se9170 */ | ||
380 | { PCI_DEVICE(PCI_VENDOR_ID_MARVELL_EXT, 0x917a), | ||
381 | .driver_data = board_ahci_yes_fbs }, /* 88se9172 */ | ||
382 | { PCI_DEVICE(PCI_VENDOR_ID_MARVELL_EXT, 0x9172), | ||
383 | diff --git a/drivers/char/tpm/tpm_ppi.c b/drivers/char/tpm/tpm_ppi.c | ||
384 | index 2168d15bc728..57a818b2b5f2 100644 | ||
385 | --- a/drivers/char/tpm/tpm_ppi.c | ||
386 | +++ b/drivers/char/tpm/tpm_ppi.c | ||
387 | @@ -27,15 +27,18 @@ static char *tpm_device_name = "TPM"; | ||
388 | static acpi_status ppi_callback(acpi_handle handle, u32 level, void *context, | ||
389 | void **return_value) | ||
390 | { | ||
391 | - acpi_status status; | ||
392 | + acpi_status status = AE_OK; | ||
393 | struct acpi_buffer buffer = { ACPI_ALLOCATE_BUFFER, NULL }; | ||
394 | - status = acpi_get_name(handle, ACPI_FULL_PATHNAME, &buffer); | ||
395 | - if (strstr(buffer.pointer, context) != NULL) { | ||
396 | - *return_value = handle; | ||
397 | + | ||
398 | + if (ACPI_SUCCESS(acpi_get_name(handle, ACPI_FULL_PATHNAME, &buffer))) { | ||
399 | + if (strstr(buffer.pointer, context) != NULL) { | ||
400 | + *return_value = handle; | ||
401 | + status = AE_CTRL_TERMINATE; | ||
402 | + } | ||
403 | kfree(buffer.pointer); | ||
404 | - return AE_CTRL_TERMINATE; | ||
405 | } | ||
406 | - return AE_OK; | ||
407 | + | ||
408 | + return status; | ||
409 | } | ||
410 | |||
411 | static inline void ppi_assign_params(union acpi_object params[4], | ||
412 | diff --git a/drivers/clk/clk-divider.c b/drivers/clk/clk-divider.c | ||
413 | index 8d3009e44fba..5543b7df8e16 100644 | ||
414 | --- a/drivers/clk/clk-divider.c | ||
415 | +++ b/drivers/clk/clk-divider.c | ||
416 | @@ -87,7 +87,7 @@ static unsigned int _get_table_val(const struct clk_div_table *table, | ||
417 | return 0; | ||
418 | } | ||
419 | |||
420 | -static unsigned int _get_val(struct clk_divider *divider, u8 div) | ||
421 | +static unsigned int _get_val(struct clk_divider *divider, unsigned int div) | ||
422 | { | ||
423 | if (divider->flags & CLK_DIVIDER_ONE_BASED) | ||
424 | return div; | ||
425 | diff --git a/drivers/clk/samsung/clk-exynos4.c b/drivers/clk/samsung/clk-exynos4.c | ||
426 | index ad5ff50c5f28..1a7c1b929c69 100644 | ||
427 | --- a/drivers/clk/samsung/clk-exynos4.c | ||
428 | +++ b/drivers/clk/samsung/clk-exynos4.c | ||
429 | @@ -39,7 +39,7 @@ | ||
430 | #define SRC_TOP1 0xc214 | ||
431 | #define SRC_CAM 0xc220 | ||
432 | #define SRC_TV 0xc224 | ||
433 | -#define SRC_MFC 0xcc28 | ||
434 | +#define SRC_MFC 0xc228 | ||
435 | #define SRC_G3D 0xc22c | ||
436 | #define E4210_SRC_IMAGE 0xc230 | ||
437 | #define SRC_LCD0 0xc234 | ||
438 | diff --git a/drivers/clk/samsung/clk-exynos5250.c b/drivers/clk/samsung/clk-exynos5250.c | ||
439 | index adf32343c9f9..e52359cf9b6f 100644 | ||
440 | --- a/drivers/clk/samsung/clk-exynos5250.c | ||
441 | +++ b/drivers/clk/samsung/clk-exynos5250.c | ||
442 | @@ -25,6 +25,7 @@ | ||
443 | #define MPLL_LOCK 0x4000 | ||
444 | #define MPLL_CON0 0x4100 | ||
445 | #define SRC_CORE1 0x4204 | ||
446 | +#define GATE_IP_ACP 0x8800 | ||
447 | #define CPLL_LOCK 0x10020 | ||
448 | #define EPLL_LOCK 0x10030 | ||
449 | #define VPLL_LOCK 0x10040 | ||
450 | @@ -75,7 +76,6 @@ | ||
451 | #define SRC_CDREX 0x20200 | ||
452 | #define PLL_DIV2_SEL 0x20a24 | ||
453 | #define GATE_IP_DISP1 0x10928 | ||
454 | -#define GATE_IP_ACP 0x10000 | ||
455 | |||
456 | /* list of PLLs to be registered */ | ||
457 | enum exynos5250_plls { | ||
458 | @@ -120,7 +120,8 @@ enum exynos5250_clks { | ||
459 | spi2, i2s1, i2s2, pcm1, pcm2, pwm, spdif, ac97, hsi2c0, hsi2c1, hsi2c2, | ||
460 | hsi2c3, chipid, sysreg, pmu, cmu_top, cmu_core, cmu_mem, tzpc0, tzpc1, | ||
461 | tzpc2, tzpc3, tzpc4, tzpc5, tzpc6, tzpc7, tzpc8, tzpc9, hdmi_cec, mct, | ||
462 | - wdt, rtc, tmu, fimd1, mie1, dsim0, dp, mixer, hdmi, g2d, | ||
463 | + wdt, rtc, tmu, fimd1, mie1, dsim0, dp, mixer, hdmi, g2d, mdma0, | ||
464 | + smmu_mdma0, | ||
465 | |||
466 | /* mux clocks */ | ||
467 | mout_hdmi = 1024, | ||
468 | @@ -354,8 +355,8 @@ static struct samsung_gate_clock exynos5250_gate_clks[] __initdata = { | ||
469 | GATE(smmu_gscl2, "smmu_gscl2", "aclk266", GATE_IP_GSCL, 9, 0, 0), | ||
470 | GATE(smmu_gscl3, "smmu_gscl3", "aclk266", GATE_IP_GSCL, 10, 0, 0), | ||
471 | GATE(mfc, "mfc", "aclk333", GATE_IP_MFC, 0, 0, 0), | ||
472 | - GATE(smmu_mfcl, "smmu_mfcl", "aclk333", GATE_IP_MFC, 1, 0, 0), | ||
473 | - GATE(smmu_mfcr, "smmu_mfcr", "aclk333", GATE_IP_MFC, 2, 0, 0), | ||
474 | + GATE(smmu_mfcl, "smmu_mfcl", "aclk333", GATE_IP_MFC, 2, 0, 0), | ||
475 | + GATE(smmu_mfcr, "smmu_mfcr", "aclk333", GATE_IP_MFC, 1, 0, 0), | ||
476 | GATE(rotator, "rotator", "aclk266", GATE_IP_GEN, 1, 0, 0), | ||
477 | GATE(jpeg, "jpeg", "aclk166", GATE_IP_GEN, 2, 0, 0), | ||
478 | GATE(mdma1, "mdma1", "aclk266", GATE_IP_GEN, 4, 0, 0), | ||
479 | @@ -406,7 +407,8 @@ static struct samsung_gate_clock exynos5250_gate_clks[] __initdata = { | ||
480 | GATE(hsi2c2, "hsi2c2", "aclk66", GATE_IP_PERIC, 30, 0, 0), | ||
481 | GATE(hsi2c3, "hsi2c3", "aclk66", GATE_IP_PERIC, 31, 0, 0), | ||
482 | GATE(chipid, "chipid", "aclk66", GATE_IP_PERIS, 0, 0, 0), | ||
483 | - GATE(sysreg, "sysreg", "aclk66", GATE_IP_PERIS, 1, 0, 0), | ||
484 | + GATE(sysreg, "sysreg", "aclk66", | ||
485 | + GATE_IP_PERIS, 1, CLK_IGNORE_UNUSED, 0), | ||
486 | GATE(pmu, "pmu", "aclk66", GATE_IP_PERIS, 2, CLK_IGNORE_UNUSED, 0), | ||
487 | GATE(tzpc0, "tzpc0", "aclk66", GATE_IP_PERIS, 6, 0, 0), | ||
488 | GATE(tzpc1, "tzpc1", "aclk66", GATE_IP_PERIS, 7, 0, 0), | ||
489 | @@ -492,6 +494,8 @@ static struct samsung_gate_clock exynos5250_gate_clks[] __initdata = { | ||
490 | GATE(mixer, "mixer", "mout_aclk200_disp1", GATE_IP_DISP1, 5, 0, 0), | ||
491 | GATE(hdmi, "hdmi", "mout_aclk200_disp1", GATE_IP_DISP1, 6, 0, 0), | ||
492 | GATE(g2d, "g2d", "aclk200", GATE_IP_ACP, 3, 0, 0), | ||
493 | + GATE(mdma0, "mdma0", "aclk266", GATE_IP_ACP, 1, 0, 0), | ||
494 | + GATE(smmu_mdma0, "smmu_mdma0", "aclk266", GATE_IP_ACP, 5, 0, 0), | ||
495 | }; | ||
496 | |||
497 | static struct samsung_pll_rate_table vpll_24mhz_tbl[] __initdata = { | ||
498 | diff --git a/drivers/cpufreq/intel_pstate.c b/drivers/cpufreq/intel_pstate.c | ||
499 | index 99d8ab548a34..d5dc567efd96 100644 | ||
500 | --- a/drivers/cpufreq/intel_pstate.c | ||
501 | +++ b/drivers/cpufreq/intel_pstate.c | ||
502 | @@ -518,7 +518,8 @@ static void intel_pstate_timer_func(unsigned long __data) | ||
503 | } | ||
504 | |||
505 | #define ICPU(model, policy) \ | ||
506 | - { X86_VENDOR_INTEL, 6, model, X86_FEATURE_ANY, (unsigned long)&policy } | ||
507 | + { X86_VENDOR_INTEL, 6, model, X86_FEATURE_APERFMPERF,\ | ||
508 | + (unsigned long)&policy } | ||
509 | |||
510 | static const struct x86_cpu_id intel_pstate_cpu_ids[] = { | ||
511 | ICPU(0x2a, default_policy), | ||
512 | diff --git a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c | ||
513 | index 57cda2a1437b..3dc7a997b795 100644 | ||
514 | --- a/drivers/gpu/drm/nouveau/core/subdev/bios/init.c | ||
515 | +++ b/drivers/gpu/drm/nouveau/core/subdev/bios/init.c | ||
516 | @@ -1294,7 +1294,11 @@ init_jump(struct nvbios_init *init) | ||
517 | u16 offset = nv_ro16(bios, init->offset + 1); | ||
518 | |||
519 | trace("JUMP\t0x%04x\n", offset); | ||
520 | - init->offset = offset; | ||
521 | + | ||
522 | + if (init_exec(init)) | ||
523 | + init->offset = offset; | ||
524 | + else | ||
525 | + init->offset += 3; | ||
526 | } | ||
527 | |||
528 | /** | ||
529 | diff --git a/drivers/leds/leds-lp5521.c b/drivers/leds/leds-lp5521.c | ||
530 | index 05188351711d..a97263e902ff 100644 | ||
531 | --- a/drivers/leds/leds-lp5521.c | ||
532 | +++ b/drivers/leds/leds-lp5521.c | ||
533 | @@ -244,18 +244,12 @@ static int lp5521_update_program_memory(struct lp55xx_chip *chip, | ||
534 | if (i % 2) | ||
535 | goto err; | ||
536 | |||
537 | - mutex_lock(&chip->lock); | ||
538 | - | ||
539 | for (i = 0; i < LP5521_PROGRAM_LENGTH; i++) { | ||
540 | ret = lp55xx_write(chip, addr[idx] + i, pattern[i]); | ||
541 | - if (ret) { | ||
542 | - mutex_unlock(&chip->lock); | ||
543 | + if (ret) | ||
544 | return -EINVAL; | ||
545 | - } | ||
546 | } | ||
547 | |||
548 | - mutex_unlock(&chip->lock); | ||
549 | - | ||
550 | return size; | ||
551 | |||
552 | err: | ||
553 | @@ -427,15 +421,17 @@ static ssize_t store_engine_load(struct device *dev, | ||
554 | { | ||
555 | struct lp55xx_led *led = i2c_get_clientdata(to_i2c_client(dev)); | ||
556 | struct lp55xx_chip *chip = led->chip; | ||
557 | + int ret; | ||
558 | |||
559 | mutex_lock(&chip->lock); | ||
560 | |||
561 | chip->engine_idx = nr; | ||
562 | lp5521_load_engine(chip); | ||
563 | + ret = lp5521_update_program_memory(chip, buf, len); | ||
564 | |||
565 | mutex_unlock(&chip->lock); | ||
566 | |||
567 | - return lp5521_update_program_memory(chip, buf, len); | ||
568 | + return ret; | ||
569 | } | ||
570 | store_load(1) | ||
571 | store_load(2) | ||
572 | diff --git a/drivers/leds/leds-lp5523.c b/drivers/leds/leds-lp5523.c | ||
573 | index fe3bcbb5747f..f382fc1e57a0 100644 | ||
574 | --- a/drivers/leds/leds-lp5523.c | ||
575 | +++ b/drivers/leds/leds-lp5523.c | ||
576 | @@ -336,18 +336,12 @@ static int lp5523_update_program_memory(struct lp55xx_chip *chip, | ||
577 | if (i % 2) | ||
578 | goto err; | ||
579 | |||
580 | - mutex_lock(&chip->lock); | ||
581 | - | ||
582 | for (i = 0; i < LP5523_PROGRAM_LENGTH; i++) { | ||
583 | ret = lp55xx_write(chip, LP5523_REG_PROG_MEM + i, pattern[i]); | ||
584 | - if (ret) { | ||
585 | - mutex_unlock(&chip->lock); | ||
586 | + if (ret) | ||
587 | return -EINVAL; | ||
588 | - } | ||
589 | } | ||
590 | |||
591 | - mutex_unlock(&chip->lock); | ||
592 | - | ||
593 | return size; | ||
594 | |||
595 | err: | ||
596 | @@ -547,15 +541,17 @@ static ssize_t store_engine_load(struct device *dev, | ||
597 | { | ||
598 | struct lp55xx_led *led = i2c_get_clientdata(to_i2c_client(dev)); | ||
599 | struct lp55xx_chip *chip = led->chip; | ||
600 | + int ret; | ||
601 | |||
602 | mutex_lock(&chip->lock); | ||
603 | |||
604 | chip->engine_idx = nr; | ||
605 | lp5523_load_engine_and_select_page(chip); | ||
606 | + ret = lp5523_update_program_memory(chip, buf, len); | ||
607 | |||
608 | mutex_unlock(&chip->lock); | ||
609 | |||
610 | - return lp5523_update_program_memory(chip, buf, len); | ||
611 | + return ret; | ||
612 | } | ||
613 | store_load(1) | ||
614 | store_load(2) | ||
615 | diff --git a/drivers/mfd/rtsx_pcr.c b/drivers/mfd/rtsx_pcr.c | ||
616 | index e6ae7720f9e1..6ed83feb0c52 100644 | ||
617 | --- a/drivers/mfd/rtsx_pcr.c | ||
618 | +++ b/drivers/mfd/rtsx_pcr.c | ||
619 | @@ -1230,8 +1230,14 @@ static void rtsx_pci_remove(struct pci_dev *pcidev) | ||
620 | |||
621 | pcr->remove_pci = true; | ||
622 | |||
623 | - cancel_delayed_work(&pcr->carddet_work); | ||
624 | - cancel_delayed_work(&pcr->idle_work); | ||
625 | + /* Disable interrupts at the pcr level */ | ||
626 | + spin_lock_irq(&pcr->lock); | ||
627 | + rtsx_pci_writel(pcr, RTSX_BIER, 0); | ||
628 | + pcr->bier = 0; | ||
629 | + spin_unlock_irq(&pcr->lock); | ||
630 | + | ||
631 | + cancel_delayed_work_sync(&pcr->carddet_work); | ||
632 | + cancel_delayed_work_sync(&pcr->idle_work); | ||
633 | |||
634 | mfd_remove_devices(&pcidev->dev); | ||
635 | |||
636 | diff --git a/drivers/net/ethernet/arc/emac_main.c b/drivers/net/ethernet/arc/emac_main.c | ||
637 | index 9e1601487263..4fc96d6c6ac0 100644 | ||
638 | --- a/drivers/net/ethernet/arc/emac_main.c | ||
639 | +++ b/drivers/net/ethernet/arc/emac_main.c | ||
640 | @@ -565,6 +565,8 @@ static int arc_emac_tx(struct sk_buff *skb, struct net_device *ndev) | ||
641 | /* Make sure pointer to data buffer is set */ | ||
642 | wmb(); | ||
643 | |||
644 | + skb_tx_timestamp(skb); | ||
645 | + | ||
646 | *info = cpu_to_le32(FOR_EMAC | FIRST_OR_LAST_MASK | len); | ||
647 | |||
648 | /* Increment index to point to the next BD */ | ||
649 | @@ -579,8 +581,6 @@ static int arc_emac_tx(struct sk_buff *skb, struct net_device *ndev) | ||
650 | |||
651 | arc_reg_set(priv, R_STATUS, TXPL_MASK); | ||
652 | |||
653 | - skb_tx_timestamp(skb); | ||
654 | - | ||
655 | return NETDEV_TX_OK; | ||
656 | } | ||
657 | |||
658 | diff --git a/drivers/net/ethernet/broadcom/tg3.c b/drivers/net/ethernet/broadcom/tg3.c | ||
659 | index a82229fe1c7f..3ff1f272c6c8 100644 | ||
660 | --- a/drivers/net/ethernet/broadcom/tg3.c | ||
661 | +++ b/drivers/net/ethernet/broadcom/tg3.c | ||
662 | @@ -16485,6 +16485,9 @@ static int tg3_get_invariants(struct tg3 *tp, const struct pci_device_id *ent) | ||
663 | /* Clear this out for sanity. */ | ||
664 | tw32(TG3PCI_MEM_WIN_BASE_ADDR, 0); | ||
665 | |||
666 | + /* Clear TG3PCI_REG_BASE_ADDR to prevent hangs. */ | ||
667 | + tw32(TG3PCI_REG_BASE_ADDR, 0); | ||
668 | + | ||
669 | pci_read_config_dword(tp->pdev, TG3PCI_PCISTATE, | ||
670 | &pci_state_reg); | ||
671 | if ((pci_state_reg & PCISTATE_CONV_PCI_MODE) == 0 && | ||
672 | diff --git a/drivers/net/ethernet/freescale/fec_main.c b/drivers/net/ethernet/freescale/fec_main.c | ||
673 | index b2793b91cc55..63090c0ddeb9 100644 | ||
674 | --- a/drivers/net/ethernet/freescale/fec_main.c | ||
675 | +++ b/drivers/net/ethernet/freescale/fec_main.c | ||
676 | @@ -425,6 +425,8 @@ fec_enet_start_xmit(struct sk_buff *skb, struct net_device *ndev) | ||
677 | /* If this was the last BD in the ring, start at the beginning again. */ | ||
678 | bdp = fec_enet_get_nextdesc(bdp, fep); | ||
679 | |||
680 | + skb_tx_timestamp(skb); | ||
681 | + | ||
682 | fep->cur_tx = bdp; | ||
683 | |||
684 | if (fep->cur_tx == fep->dirty_tx) | ||
685 | @@ -433,8 +435,6 @@ fec_enet_start_xmit(struct sk_buff *skb, struct net_device *ndev) | ||
686 | /* Trigger transmission start */ | ||
687 | writel(0, fep->hwp + FEC_X_DES_ACTIVE); | ||
688 | |||
689 | - skb_tx_timestamp(skb); | ||
690 | - | ||
691 | return NETDEV_TX_OK; | ||
692 | } | ||
693 | |||
694 | diff --git a/drivers/net/ethernet/ibm/ehea/ehea_main.c b/drivers/net/ethernet/ibm/ehea/ehea_main.c | ||
695 | index 2d1c6bdd3618..7628e0fd8455 100644 | ||
696 | --- a/drivers/net/ethernet/ibm/ehea/ehea_main.c | ||
697 | +++ b/drivers/net/ethernet/ibm/ehea/ehea_main.c | ||
698 | @@ -3033,7 +3033,7 @@ static struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter, | ||
699 | |||
700 | dev->hw_features = NETIF_F_SG | NETIF_F_TSO | | ||
701 | NETIF_F_IP_CSUM | NETIF_F_HW_VLAN_CTAG_TX; | ||
702 | - dev->features = NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_TSO | | ||
703 | + dev->features = NETIF_F_SG | NETIF_F_TSO | | ||
704 | NETIF_F_HIGHDMA | NETIF_F_IP_CSUM | | ||
705 | NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | | ||
706 | NETIF_F_HW_VLAN_CTAG_FILTER | NETIF_F_RXCSUM; | ||
707 | diff --git a/drivers/net/ethernet/sfc/efx.c b/drivers/net/ethernet/sfc/efx.c | ||
708 | index 07c9bc4c61bc..f59a0b6f1ae5 100644 | ||
709 | --- a/drivers/net/ethernet/sfc/efx.c | ||
710 | +++ b/drivers/net/ethernet/sfc/efx.c | ||
711 | @@ -585,7 +585,7 @@ static void efx_start_datapath(struct efx_nic *efx) | ||
712 | EFX_MAX_FRAME_LEN(efx->net_dev->mtu) + | ||
713 | efx->type->rx_buffer_padding); | ||
714 | rx_buf_len = (sizeof(struct efx_rx_page_state) + | ||
715 | - NET_IP_ALIGN + efx->rx_dma_len); | ||
716 | + efx->rx_ip_align + efx->rx_dma_len); | ||
717 | if (rx_buf_len <= PAGE_SIZE) { | ||
718 | efx->rx_scatter = efx->type->always_rx_scatter; | ||
719 | efx->rx_buffer_order = 0; | ||
720 | @@ -645,6 +645,8 @@ static void efx_start_datapath(struct efx_nic *efx) | ||
721 | WARN_ON(channel->rx_pkt_n_frags); | ||
722 | } | ||
723 | |||
724 | + efx_ptp_start_datapath(efx); | ||
725 | + | ||
726 | if (netif_device_present(efx->net_dev)) | ||
727 | netif_tx_wake_all_queues(efx->net_dev); | ||
728 | } | ||
729 | @@ -659,6 +661,8 @@ static void efx_stop_datapath(struct efx_nic *efx) | ||
730 | EFX_ASSERT_RESET_SERIALISED(efx); | ||
731 | BUG_ON(efx->port_enabled); | ||
732 | |||
733 | + efx_ptp_stop_datapath(efx); | ||
734 | + | ||
735 | /* Stop RX refill */ | ||
736 | efx_for_each_channel(channel, efx) { | ||
737 | efx_for_each_channel_rx_queue(rx_queue, channel) | ||
738 | @@ -2550,6 +2554,8 @@ static int efx_init_struct(struct efx_nic *efx, | ||
739 | |||
740 | efx->net_dev = net_dev; | ||
741 | efx->rx_prefix_size = efx->type->rx_prefix_size; | ||
742 | + efx->rx_ip_align = | ||
743 | + NET_IP_ALIGN ? (efx->rx_prefix_size + NET_IP_ALIGN) % 4 : 0; | ||
744 | efx->rx_packet_hash_offset = | ||
745 | efx->type->rx_hash_offset - efx->type->rx_prefix_size; | ||
746 | spin_lock_init(&efx->stats_lock); | ||
747 | diff --git a/drivers/net/ethernet/sfc/mcdi.c b/drivers/net/ethernet/sfc/mcdi.c | ||
748 | index 366c8e3e3784..4b0bd8a1514d 100644 | ||
749 | --- a/drivers/net/ethernet/sfc/mcdi.c | ||
750 | +++ b/drivers/net/ethernet/sfc/mcdi.c | ||
751 | @@ -50,6 +50,7 @@ struct efx_mcdi_async_param { | ||
752 | static void efx_mcdi_timeout_async(unsigned long context); | ||
753 | static int efx_mcdi_drv_attach(struct efx_nic *efx, bool driver_operating, | ||
754 | bool *was_attached_out); | ||
755 | +static bool efx_mcdi_poll_once(struct efx_nic *efx); | ||
756 | |||
757 | static inline struct efx_mcdi_iface *efx_mcdi(struct efx_nic *efx) | ||
758 | { | ||
759 | @@ -237,6 +238,21 @@ static void efx_mcdi_read_response_header(struct efx_nic *efx) | ||
760 | } | ||
761 | } | ||
762 | |||
763 | +static bool efx_mcdi_poll_once(struct efx_nic *efx) | ||
764 | +{ | ||
765 | + struct efx_mcdi_iface *mcdi = efx_mcdi(efx); | ||
766 | + | ||
767 | + rmb(); | ||
768 | + if (!efx->type->mcdi_poll_response(efx)) | ||
769 | + return false; | ||
770 | + | ||
771 | + spin_lock_bh(&mcdi->iface_lock); | ||
772 | + efx_mcdi_read_response_header(efx); | ||
773 | + spin_unlock_bh(&mcdi->iface_lock); | ||
774 | + | ||
775 | + return true; | ||
776 | +} | ||
777 | + | ||
778 | static int efx_mcdi_poll(struct efx_nic *efx) | ||
779 | { | ||
780 | struct efx_mcdi_iface *mcdi = efx_mcdi(efx); | ||
781 | @@ -272,18 +288,13 @@ static int efx_mcdi_poll(struct efx_nic *efx) | ||
782 | |||
783 | time = jiffies; | ||
784 | |||
785 | - rmb(); | ||
786 | - if (efx->type->mcdi_poll_response(efx)) | ||
787 | + if (efx_mcdi_poll_once(efx)) | ||
788 | break; | ||
789 | |||
790 | if (time_after(time, finish)) | ||
791 | return -ETIMEDOUT; | ||
792 | } | ||
793 | |||
794 | - spin_lock_bh(&mcdi->iface_lock); | ||
795 | - efx_mcdi_read_response_header(efx); | ||
796 | - spin_unlock_bh(&mcdi->iface_lock); | ||
797 | - | ||
798 | /* Return rc=0 like wait_event_timeout() */ | ||
799 | return 0; | ||
800 | } | ||
801 | @@ -619,6 +630,16 @@ int efx_mcdi_rpc_finish(struct efx_nic *efx, unsigned cmd, size_t inlen, | ||
802 | rc = efx_mcdi_await_completion(efx); | ||
803 | |||
804 | if (rc != 0) { | ||
805 | + netif_err(efx, hw, efx->net_dev, | ||
806 | + "MC command 0x%x inlen %d mode %d timed out\n", | ||
807 | + cmd, (int)inlen, mcdi->mode); | ||
808 | + | ||
809 | + if (mcdi->mode == MCDI_MODE_EVENTS && efx_mcdi_poll_once(efx)) { | ||
810 | + netif_err(efx, hw, efx->net_dev, | ||
811 | + "MCDI request was completed without an event\n"); | ||
812 | + rc = 0; | ||
813 | + } | ||
814 | + | ||
815 | /* Close the race with efx_mcdi_ev_cpl() executing just too late | ||
816 | * and completing a request we've just cancelled, by ensuring | ||
817 | * that the seqno check therein fails. | ||
818 | @@ -627,11 +648,9 @@ int efx_mcdi_rpc_finish(struct efx_nic *efx, unsigned cmd, size_t inlen, | ||
819 | ++mcdi->seqno; | ||
820 | ++mcdi->credits; | ||
821 | spin_unlock_bh(&mcdi->iface_lock); | ||
822 | + } | ||
823 | |||
824 | - netif_err(efx, hw, efx->net_dev, | ||
825 | - "MC command 0x%x inlen %d mode %d timed out\n", | ||
826 | - cmd, (int)inlen, mcdi->mode); | ||
827 | - } else { | ||
828 | + if (rc == 0) { | ||
829 | size_t hdr_len, data_len; | ||
830 | |||
831 | /* At the very least we need a memory barrier here to ensure | ||
832 | diff --git a/drivers/net/ethernet/sfc/net_driver.h b/drivers/net/ethernet/sfc/net_driver.h | ||
833 | index b172ed133055..2aeb45167511 100644 | ||
834 | --- a/drivers/net/ethernet/sfc/net_driver.h | ||
835 | +++ b/drivers/net/ethernet/sfc/net_driver.h | ||
836 | @@ -673,6 +673,8 @@ struct vfdi_status; | ||
837 | * @n_channels: Number of channels in use | ||
838 | * @n_rx_channels: Number of channels used for RX (= number of RX queues) | ||
839 | * @n_tx_channels: Number of channels used for TX | ||
840 | + * @rx_ip_align: RX DMA address offset to have IP header aligned in | ||
841 | + * in accordance with NET_IP_ALIGN | ||
842 | * @rx_dma_len: Current maximum RX DMA length | ||
843 | * @rx_buffer_order: Order (log2) of number of pages for each RX buffer | ||
844 | * @rx_buffer_truesize: Amortised allocation size of an RX buffer, | ||
845 | @@ -806,6 +808,7 @@ struct efx_nic { | ||
846 | unsigned rss_spread; | ||
847 | unsigned tx_channel_offset; | ||
848 | unsigned n_tx_channels; | ||
849 | + unsigned int rx_ip_align; | ||
850 | unsigned int rx_dma_len; | ||
851 | unsigned int rx_buffer_order; | ||
852 | unsigned int rx_buffer_truesize; | ||
853 | diff --git a/drivers/net/ethernet/sfc/nic.h b/drivers/net/ethernet/sfc/nic.h | ||
854 | index 890bbbe8320e..3d713b518847 100644 | ||
855 | --- a/drivers/net/ethernet/sfc/nic.h | ||
856 | +++ b/drivers/net/ethernet/sfc/nic.h | ||
857 | @@ -528,6 +528,8 @@ extern void efx_ptp_get_ts_info(struct efx_nic *efx, | ||
858 | extern bool efx_ptp_is_ptp_tx(struct efx_nic *efx, struct sk_buff *skb); | ||
859 | extern int efx_ptp_tx(struct efx_nic *efx, struct sk_buff *skb); | ||
860 | extern void efx_ptp_event(struct efx_nic *efx, efx_qword_t *ev); | ||
861 | +void efx_ptp_start_datapath(struct efx_nic *efx); | ||
862 | +void efx_ptp_stop_datapath(struct efx_nic *efx); | ||
863 | |||
864 | extern const struct efx_nic_type falcon_a1_nic_type; | ||
865 | extern const struct efx_nic_type falcon_b0_nic_type; | ||
866 | diff --git a/drivers/net/ethernet/sfc/ptp.c b/drivers/net/ethernet/sfc/ptp.c | ||
867 | index 03acf57df045..3dd39dcfe36b 100644 | ||
868 | --- a/drivers/net/ethernet/sfc/ptp.c | ||
869 | +++ b/drivers/net/ethernet/sfc/ptp.c | ||
870 | @@ -220,6 +220,7 @@ struct efx_ptp_timeset { | ||
871 | * @evt_list: List of MC receive events awaiting packets | ||
872 | * @evt_free_list: List of free events | ||
873 | * @evt_lock: Lock for manipulating evt_list and evt_free_list | ||
874 | + * @evt_overflow: Boolean indicating that event list has overflowed | ||
875 | * @rx_evts: Instantiated events (on evt_list and evt_free_list) | ||
876 | * @workwq: Work queue for processing pending PTP operations | ||
877 | * @work: Work task | ||
878 | @@ -270,6 +271,7 @@ struct efx_ptp_data { | ||
879 | struct list_head evt_list; | ||
880 | struct list_head evt_free_list; | ||
881 | spinlock_t evt_lock; | ||
882 | + bool evt_overflow; | ||
883 | struct efx_ptp_event_rx rx_evts[MAX_RECEIVE_EVENTS]; | ||
884 | struct workqueue_struct *workwq; | ||
885 | struct work_struct work; | ||
886 | @@ -635,6 +637,11 @@ static void efx_ptp_drop_time_expired_events(struct efx_nic *efx) | ||
887 | } | ||
888 | } | ||
889 | } | ||
890 | + /* If the event overflow flag is set and the event list is now empty | ||
891 | + * clear the flag to re-enable the overflow warning message. | ||
892 | + */ | ||
893 | + if (ptp->evt_overflow && list_empty(&ptp->evt_list)) | ||
894 | + ptp->evt_overflow = false; | ||
895 | spin_unlock_bh(&ptp->evt_lock); | ||
896 | } | ||
897 | |||
898 | @@ -676,6 +683,11 @@ static enum ptp_packet_state efx_ptp_match_rx(struct efx_nic *efx, | ||
899 | break; | ||
900 | } | ||
901 | } | ||
902 | + /* If the event overflow flag is set and the event list is now empty | ||
903 | + * clear the flag to re-enable the overflow warning message. | ||
904 | + */ | ||
905 | + if (ptp->evt_overflow && list_empty(&ptp->evt_list)) | ||
906 | + ptp->evt_overflow = false; | ||
907 | spin_unlock_bh(&ptp->evt_lock); | ||
908 | |||
909 | return rc; | ||
910 | @@ -705,8 +717,9 @@ static bool efx_ptp_process_events(struct efx_nic *efx, struct sk_buff_head *q) | ||
911 | __skb_queue_tail(q, skb); | ||
912 | } else if (time_after(jiffies, match->expiry)) { | ||
913 | match->state = PTP_PACKET_STATE_TIMED_OUT; | ||
914 | - netif_warn(efx, rx_err, efx->net_dev, | ||
915 | - "PTP packet - no timestamp seen\n"); | ||
916 | + if (net_ratelimit()) | ||
917 | + netif_warn(efx, rx_err, efx->net_dev, | ||
918 | + "PTP packet - no timestamp seen\n"); | ||
919 | __skb_queue_tail(q, skb); | ||
920 | } else { | ||
921 | /* Replace unprocessed entry and stop */ | ||
922 | @@ -788,9 +801,14 @@ fail: | ||
923 | static int efx_ptp_stop(struct efx_nic *efx) | ||
924 | { | ||
925 | struct efx_ptp_data *ptp = efx->ptp_data; | ||
926 | - int rc = efx_ptp_disable(efx); | ||
927 | struct list_head *cursor; | ||
928 | struct list_head *next; | ||
929 | + int rc; | ||
930 | + | ||
931 | + if (ptp == NULL) | ||
932 | + return 0; | ||
933 | + | ||
934 | + rc = efx_ptp_disable(efx); | ||
935 | |||
936 | if (ptp->rxfilter_installed) { | ||
937 | efx_filter_remove_id_safe(efx, EFX_FILTER_PRI_REQUIRED, | ||
938 | @@ -809,11 +827,19 @@ static int efx_ptp_stop(struct efx_nic *efx) | ||
939 | list_for_each_safe(cursor, next, &efx->ptp_data->evt_list) { | ||
940 | list_move(cursor, &efx->ptp_data->evt_free_list); | ||
941 | } | ||
942 | + ptp->evt_overflow = false; | ||
943 | spin_unlock_bh(&efx->ptp_data->evt_lock); | ||
944 | |||
945 | return rc; | ||
946 | } | ||
947 | |||
948 | +static int efx_ptp_restart(struct efx_nic *efx) | ||
949 | +{ | ||
950 | + if (efx->ptp_data && efx->ptp_data->enabled) | ||
951 | + return efx_ptp_start(efx); | ||
952 | + return 0; | ||
953 | +} | ||
954 | + | ||
955 | static void efx_ptp_pps_worker(struct work_struct *work) | ||
956 | { | ||
957 | struct efx_ptp_data *ptp = | ||
958 | @@ -901,6 +927,7 @@ static int efx_ptp_probe_channel(struct efx_channel *channel) | ||
959 | spin_lock_init(&ptp->evt_lock); | ||
960 | for (pos = 0; pos < MAX_RECEIVE_EVENTS; pos++) | ||
961 | list_add(&ptp->rx_evts[pos].link, &ptp->evt_free_list); | ||
962 | + ptp->evt_overflow = false; | ||
963 | |||
964 | ptp->phc_clock_info.owner = THIS_MODULE; | ||
965 | snprintf(ptp->phc_clock_info.name, | ||
966 | @@ -989,7 +1016,11 @@ bool efx_ptp_is_ptp_tx(struct efx_nic *efx, struct sk_buff *skb) | ||
967 | skb->len >= PTP_MIN_LENGTH && | ||
968 | skb->len <= MC_CMD_PTP_IN_TRANSMIT_PACKET_MAXNUM && | ||
969 | likely(skb->protocol == htons(ETH_P_IP)) && | ||
970 | + skb_transport_header_was_set(skb) && | ||
971 | + skb_network_header_len(skb) >= sizeof(struct iphdr) && | ||
972 | ip_hdr(skb)->protocol == IPPROTO_UDP && | ||
973 | + skb_headlen(skb) >= | ||
974 | + skb_transport_offset(skb) + sizeof(struct udphdr) && | ||
975 | udp_hdr(skb)->dest == htons(PTP_EVENT_PORT); | ||
976 | } | ||
977 | |||
978 | @@ -1106,7 +1137,7 @@ static int efx_ptp_change_mode(struct efx_nic *efx, bool enable_wanted, | ||
979 | { | ||
980 | if ((enable_wanted != efx->ptp_data->enabled) || | ||
981 | (enable_wanted && (efx->ptp_data->mode != new_mode))) { | ||
982 | - int rc; | ||
983 | + int rc = 0; | ||
984 | |||
985 | if (enable_wanted) { | ||
986 | /* Change of mode requires disable */ | ||
987 | @@ -1123,7 +1154,8 @@ static int efx_ptp_change_mode(struct efx_nic *efx, bool enable_wanted, | ||
988 | * succeed. | ||
989 | */ | ||
990 | efx->ptp_data->mode = new_mode; | ||
991 | - rc = efx_ptp_start(efx); | ||
992 | + if (netif_running(efx->net_dev)) | ||
993 | + rc = efx_ptp_start(efx); | ||
994 | if (rc == 0) { | ||
995 | rc = efx_ptp_synchronize(efx, | ||
996 | PTP_SYNC_ATTEMPTS * 2); | ||
997 | @@ -1295,8 +1327,13 @@ static void ptp_event_rx(struct efx_nic *efx, struct efx_ptp_data *ptp) | ||
998 | list_add_tail(&evt->link, &ptp->evt_list); | ||
999 | |||
1000 | queue_work(ptp->workwq, &ptp->work); | ||
1001 | - } else { | ||
1002 | - netif_err(efx, rx_err, efx->net_dev, "No free PTP event"); | ||
1003 | + } else if (!ptp->evt_overflow) { | ||
1004 | + /* Log a warning message and set the event overflow flag. | ||
1005 | + * The message won't be logged again until the event queue | ||
1006 | + * becomes empty. | ||
1007 | + */ | ||
1008 | + netif_err(efx, rx_err, efx->net_dev, "PTP event queue overflow\n"); | ||
1009 | + ptp->evt_overflow = true; | ||
1010 | } | ||
1011 | spin_unlock_bh(&ptp->evt_lock); | ||
1012 | } | ||
1013 | @@ -1389,7 +1426,7 @@ static int efx_phc_adjfreq(struct ptp_clock_info *ptp, s32 delta) | ||
1014 | if (rc != 0) | ||
1015 | return rc; | ||
1016 | |||
1017 | - ptp_data->current_adjfreq = delta; | ||
1018 | + ptp_data->current_adjfreq = adjustment_ns; | ||
1019 | return 0; | ||
1020 | } | ||
1021 | |||
1022 | @@ -1404,7 +1441,7 @@ static int efx_phc_adjtime(struct ptp_clock_info *ptp, s64 delta) | ||
1023 | |||
1024 | MCDI_SET_DWORD(inbuf, PTP_IN_OP, MC_CMD_PTP_OP_ADJUST); | ||
1025 | MCDI_SET_DWORD(inbuf, PTP_IN_PERIPH_ID, 0); | ||
1026 | - MCDI_SET_QWORD(inbuf, PTP_IN_ADJUST_FREQ, 0); | ||
1027 | + MCDI_SET_QWORD(inbuf, PTP_IN_ADJUST_FREQ, ptp_data->current_adjfreq); | ||
1028 | MCDI_SET_DWORD(inbuf, PTP_IN_ADJUST_SECONDS, (u32)delta_ts.tv_sec); | ||
1029 | MCDI_SET_DWORD(inbuf, PTP_IN_ADJUST_NANOSECONDS, (u32)delta_ts.tv_nsec); | ||
1030 | return efx_mcdi_rpc(efx, MC_CMD_PTP, inbuf, sizeof(inbuf), | ||
1031 | @@ -1491,3 +1528,14 @@ void efx_ptp_probe(struct efx_nic *efx) | ||
1032 | efx->extra_channel_type[EFX_EXTRA_CHANNEL_PTP] = | ||
1033 | &efx_ptp_channel_type; | ||
1034 | } | ||
1035 | + | ||
1036 | +void efx_ptp_start_datapath(struct efx_nic *efx) | ||
1037 | +{ | ||
1038 | + if (efx_ptp_restart(efx)) | ||
1039 | + netif_err(efx, drv, efx->net_dev, "Failed to restart PTP.\n"); | ||
1040 | +} | ||
1041 | + | ||
1042 | +void efx_ptp_stop_datapath(struct efx_nic *efx) | ||
1043 | +{ | ||
1044 | + efx_ptp_stop(efx); | ||
1045 | +} | ||
1046 | diff --git a/drivers/net/ethernet/sfc/rx.c b/drivers/net/ethernet/sfc/rx.c | ||
1047 | index 4a596725023f..f18d5864bfa7 100644 | ||
1048 | --- a/drivers/net/ethernet/sfc/rx.c | ||
1049 | +++ b/drivers/net/ethernet/sfc/rx.c | ||
1050 | @@ -93,7 +93,7 @@ static inline void efx_sync_rx_buffer(struct efx_nic *efx, | ||
1051 | |||
1052 | void efx_rx_config_page_split(struct efx_nic *efx) | ||
1053 | { | ||
1054 | - efx->rx_page_buf_step = ALIGN(efx->rx_dma_len + NET_IP_ALIGN, | ||
1055 | + efx->rx_page_buf_step = ALIGN(efx->rx_dma_len + efx->rx_ip_align, | ||
1056 | EFX_RX_BUF_ALIGNMENT); | ||
1057 | efx->rx_bufs_per_page = efx->rx_buffer_order ? 1 : | ||
1058 | ((PAGE_SIZE - sizeof(struct efx_rx_page_state)) / | ||
1059 | @@ -188,9 +188,9 @@ static int efx_init_rx_buffers(struct efx_rx_queue *rx_queue) | ||
1060 | do { | ||
1061 | index = rx_queue->added_count & rx_queue->ptr_mask; | ||
1062 | rx_buf = efx_rx_buffer(rx_queue, index); | ||
1063 | - rx_buf->dma_addr = dma_addr + NET_IP_ALIGN; | ||
1064 | + rx_buf->dma_addr = dma_addr + efx->rx_ip_align; | ||
1065 | rx_buf->page = page; | ||
1066 | - rx_buf->page_offset = page_offset + NET_IP_ALIGN; | ||
1067 | + rx_buf->page_offset = page_offset + efx->rx_ip_align; | ||
1068 | rx_buf->len = efx->rx_dma_len; | ||
1069 | rx_buf->flags = 0; | ||
1070 | ++rx_queue->added_count; | ||
1071 | diff --git a/drivers/net/ethernet/tehuti/tehuti.c b/drivers/net/ethernet/tehuti/tehuti.c | ||
1072 | index 571452e786d5..61a1540f1347 100644 | ||
1073 | --- a/drivers/net/ethernet/tehuti/tehuti.c | ||
1074 | +++ b/drivers/net/ethernet/tehuti/tehuti.c | ||
1075 | @@ -2019,7 +2019,6 @@ bdx_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | ||
1076 | ndev->features = NETIF_F_IP_CSUM | NETIF_F_SG | NETIF_F_TSO | ||
1077 | | NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | | ||
1078 | NETIF_F_HW_VLAN_CTAG_FILTER | NETIF_F_RXCSUM | ||
1079 | - /*| NETIF_F_FRAGLIST */ | ||
1080 | ; | ||
1081 | ndev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG | | ||
1082 | NETIF_F_TSO | NETIF_F_HW_VLAN_CTAG_TX; | ||
1083 | diff --git a/drivers/net/ethernet/xilinx/ll_temac_main.c b/drivers/net/ethernet/xilinx/ll_temac_main.c | ||
1084 | index 0029148077a9..def50bd53124 100644 | ||
1085 | --- a/drivers/net/ethernet/xilinx/ll_temac_main.c | ||
1086 | +++ b/drivers/net/ethernet/xilinx/ll_temac_main.c | ||
1087 | @@ -1016,7 +1016,7 @@ static int temac_of_probe(struct platform_device *op) | ||
1088 | platform_set_drvdata(op, ndev); | ||
1089 | SET_NETDEV_DEV(ndev, &op->dev); | ||
1090 | ndev->flags &= ~IFF_MULTICAST; /* clear multicast */ | ||
1091 | - ndev->features = NETIF_F_SG | NETIF_F_FRAGLIST; | ||
1092 | + ndev->features = NETIF_F_SG; | ||
1093 | ndev->netdev_ops = &temac_netdev_ops; | ||
1094 | ndev->ethtool_ops = &temac_ethtool_ops; | ||
1095 | #if 0 | ||
1096 | diff --git a/drivers/net/ethernet/xilinx/xilinx_axienet_main.c b/drivers/net/ethernet/xilinx/xilinx_axienet_main.c | ||
1097 | index b2ff038d6d20..f9293da19e26 100644 | ||
1098 | --- a/drivers/net/ethernet/xilinx/xilinx_axienet_main.c | ||
1099 | +++ b/drivers/net/ethernet/xilinx/xilinx_axienet_main.c | ||
1100 | @@ -1486,7 +1486,7 @@ static int axienet_of_probe(struct platform_device *op) | ||
1101 | |||
1102 | SET_NETDEV_DEV(ndev, &op->dev); | ||
1103 | ndev->flags &= ~IFF_MULTICAST; /* clear multicast */ | ||
1104 | - ndev->features = NETIF_F_SG | NETIF_F_FRAGLIST; | ||
1105 | + ndev->features = NETIF_F_SG; | ||
1106 | ndev->netdev_ops = &axienet_netdev_ops; | ||
1107 | ndev->ethtool_ops = &axienet_ethtool_ops; | ||
1108 | |||
1109 | diff --git a/drivers/net/hamradio/hdlcdrv.c b/drivers/net/hamradio/hdlcdrv.c | ||
1110 | index 3169252613fa..5d78c1d08abd 100644 | ||
1111 | --- a/drivers/net/hamradio/hdlcdrv.c | ||
1112 | +++ b/drivers/net/hamradio/hdlcdrv.c | ||
1113 | @@ -571,6 +571,8 @@ static int hdlcdrv_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) | ||
1114 | case HDLCDRVCTL_CALIBRATE: | ||
1115 | if(!capable(CAP_SYS_RAWIO)) | ||
1116 | return -EPERM; | ||
1117 | + if (bi.data.calibrate > INT_MAX / s->par.bitrate) | ||
1118 | + return -EINVAL; | ||
1119 | s->hdlctx.calibrate = bi.data.calibrate * s->par.bitrate / 16; | ||
1120 | return 0; | ||
1121 | |||
1122 | diff --git a/drivers/net/hamradio/yam.c b/drivers/net/hamradio/yam.c | ||
1123 | index 5af1c3e5032a..b7e967540509 100644 | ||
1124 | --- a/drivers/net/hamradio/yam.c | ||
1125 | +++ b/drivers/net/hamradio/yam.c | ||
1126 | @@ -1057,6 +1057,7 @@ static int yam_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) | ||
1127 | break; | ||
1128 | |||
1129 | case SIOCYAMGCFG: | ||
1130 | + memset(&yi, 0, sizeof(yi)); | ||
1131 | yi.cfg.mask = 0xffffffff; | ||
1132 | yi.cfg.iobase = yp->iobase; | ||
1133 | yi.cfg.irq = yp->irq; | ||
1134 | diff --git a/drivers/net/hyperv/netvsc_drv.c b/drivers/net/hyperv/netvsc_drv.c | ||
1135 | index 524f713f6017..f8135725bcf6 100644 | ||
1136 | --- a/drivers/net/hyperv/netvsc_drv.c | ||
1137 | +++ b/drivers/net/hyperv/netvsc_drv.c | ||
1138 | @@ -327,7 +327,6 @@ static int netvsc_change_mtu(struct net_device *ndev, int mtu) | ||
1139 | return -EINVAL; | ||
1140 | |||
1141 | nvdev->start_remove = true; | ||
1142 | - cancel_delayed_work_sync(&ndevctx->dwork); | ||
1143 | cancel_work_sync(&ndevctx->work); | ||
1144 | netif_tx_disable(ndev); | ||
1145 | rndis_filter_device_remove(hdev); | ||
1146 | diff --git a/drivers/net/macvtap.c b/drivers/net/macvtap.c | ||
1147 | index dc76670c2f2a..5895e4dbbf2a 100644 | ||
1148 | --- a/drivers/net/macvtap.c | ||
1149 | +++ b/drivers/net/macvtap.c | ||
1150 | @@ -767,11 +767,10 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q, | ||
1151 | const struct sk_buff *skb, | ||
1152 | const struct iovec *iv, int len) | ||
1153 | { | ||
1154 | - struct macvlan_dev *vlan; | ||
1155 | int ret; | ||
1156 | int vnet_hdr_len = 0; | ||
1157 | int vlan_offset = 0; | ||
1158 | - int copied; | ||
1159 | + int copied, total; | ||
1160 | |||
1161 | if (q->flags & IFF_VNET_HDR) { | ||
1162 | struct virtio_net_hdr vnet_hdr; | ||
1163 | @@ -786,7 +785,8 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q, | ||
1164 | if (memcpy_toiovecend(iv, (void *)&vnet_hdr, 0, sizeof(vnet_hdr))) | ||
1165 | return -EFAULT; | ||
1166 | } | ||
1167 | - copied = vnet_hdr_len; | ||
1168 | + total = copied = vnet_hdr_len; | ||
1169 | + total += skb->len; | ||
1170 | |||
1171 | if (!vlan_tx_tag_present(skb)) | ||
1172 | len = min_t(int, skb->len, len); | ||
1173 | @@ -801,6 +801,7 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q, | ||
1174 | |||
1175 | vlan_offset = offsetof(struct vlan_ethhdr, h_vlan_proto); | ||
1176 | len = min_t(int, skb->len + VLAN_HLEN, len); | ||
1177 | + total += VLAN_HLEN; | ||
1178 | |||
1179 | copy = min_t(int, vlan_offset, len); | ||
1180 | ret = skb_copy_datagram_const_iovec(skb, 0, iv, copied, copy); | ||
1181 | @@ -818,19 +819,9 @@ static ssize_t macvtap_put_user(struct macvtap_queue *q, | ||
1182 | } | ||
1183 | |||
1184 | ret = skb_copy_datagram_const_iovec(skb, vlan_offset, iv, copied, len); | ||
1185 | - copied += len; | ||
1186 | |||
1187 | done: | ||
1188 | - rcu_read_lock(); | ||
1189 | - vlan = rcu_dereference(q->vlan); | ||
1190 | - if (vlan) { | ||
1191 | - preempt_disable(); | ||
1192 | - macvlan_count_rx(vlan, copied - vnet_hdr_len, ret == 0, 0); | ||
1193 | - preempt_enable(); | ||
1194 | - } | ||
1195 | - rcu_read_unlock(); | ||
1196 | - | ||
1197 | - return ret ? ret : copied; | ||
1198 | + return ret ? ret : total; | ||
1199 | } | ||
1200 | |||
1201 | static ssize_t macvtap_do_read(struct macvtap_queue *q, struct kiocb *iocb, | ||
1202 | @@ -885,7 +876,9 @@ static ssize_t macvtap_aio_read(struct kiocb *iocb, const struct iovec *iv, | ||
1203 | } | ||
1204 | |||
1205 | ret = macvtap_do_read(q, iocb, iv, len, file->f_flags & O_NONBLOCK); | ||
1206 | - ret = min_t(ssize_t, ret, len); /* XXX copied from tun.c. Why? */ | ||
1207 | + ret = min_t(ssize_t, ret, len); | ||
1208 | + if (ret > 0) | ||
1209 | + iocb->ki_pos = ret; | ||
1210 | out: | ||
1211 | return ret; | ||
1212 | } | ||
1213 | diff --git a/drivers/net/tun.c b/drivers/net/tun.c | ||
1214 | index 782e38bfc1ee..7c8343a4f918 100644 | ||
1215 | --- a/drivers/net/tun.c | ||
1216 | +++ b/drivers/net/tun.c | ||
1217 | @@ -1184,7 +1184,7 @@ static ssize_t tun_put_user(struct tun_struct *tun, | ||
1218 | { | ||
1219 | struct tun_pi pi = { 0, skb->protocol }; | ||
1220 | ssize_t total = 0; | ||
1221 | - int vlan_offset = 0; | ||
1222 | + int vlan_offset = 0, copied; | ||
1223 | |||
1224 | if (!(tun->flags & TUN_NO_PI)) { | ||
1225 | if ((len -= sizeof(pi)) < 0) | ||
1226 | @@ -1248,6 +1248,8 @@ static ssize_t tun_put_user(struct tun_struct *tun, | ||
1227 | total += tun->vnet_hdr_sz; | ||
1228 | } | ||
1229 | |||
1230 | + copied = total; | ||
1231 | + total += skb->len; | ||
1232 | if (!vlan_tx_tag_present(skb)) { | ||
1233 | len = min_t(int, skb->len, len); | ||
1234 | } else { | ||
1235 | @@ -1262,24 +1264,24 @@ static ssize_t tun_put_user(struct tun_struct *tun, | ||
1236 | |||
1237 | vlan_offset = offsetof(struct vlan_ethhdr, h_vlan_proto); | ||
1238 | len = min_t(int, skb->len + VLAN_HLEN, len); | ||
1239 | + total += VLAN_HLEN; | ||
1240 | |||
1241 | copy = min_t(int, vlan_offset, len); | ||
1242 | - ret = skb_copy_datagram_const_iovec(skb, 0, iv, total, copy); | ||
1243 | + ret = skb_copy_datagram_const_iovec(skb, 0, iv, copied, copy); | ||
1244 | len -= copy; | ||
1245 | - total += copy; | ||
1246 | + copied += copy; | ||
1247 | if (ret || !len) | ||
1248 | goto done; | ||
1249 | |||
1250 | copy = min_t(int, sizeof(veth), len); | ||
1251 | - ret = memcpy_toiovecend(iv, (void *)&veth, total, copy); | ||
1252 | + ret = memcpy_toiovecend(iv, (void *)&veth, copied, copy); | ||
1253 | len -= copy; | ||
1254 | - total += copy; | ||
1255 | + copied += copy; | ||
1256 | if (ret || !len) | ||
1257 | goto done; | ||
1258 | } | ||
1259 | |||
1260 | - skb_copy_datagram_const_iovec(skb, vlan_offset, iv, total, len); | ||
1261 | - total += len; | ||
1262 | + skb_copy_datagram_const_iovec(skb, vlan_offset, iv, copied, len); | ||
1263 | |||
1264 | done: | ||
1265 | tun->dev->stats.tx_packets++; | ||
1266 | @@ -1356,6 +1358,8 @@ static ssize_t tun_chr_aio_read(struct kiocb *iocb, const struct iovec *iv, | ||
1267 | ret = tun_do_read(tun, tfile, iocb, iv, len, | ||
1268 | file->f_flags & O_NONBLOCK); | ||
1269 | ret = min_t(ssize_t, ret, len); | ||
1270 | + if (ret > 0) | ||
1271 | + iocb->ki_pos = ret; | ||
1272 | out: | ||
1273 | tun_put(tun); | ||
1274 | return ret; | ||
1275 | diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c | ||
1276 | index bbc9cb84ec1f..8065066a6230 100644 | ||
1277 | --- a/drivers/net/virtio_net.c | ||
1278 | +++ b/drivers/net/virtio_net.c | ||
1279 | @@ -297,26 +297,61 @@ static struct sk_buff *page_to_skb(struct receive_queue *rq, | ||
1280 | return skb; | ||
1281 | } | ||
1282 | |||
1283 | -static int receive_mergeable(struct receive_queue *rq, struct sk_buff *skb) | ||
1284 | +static struct sk_buff *receive_small(void *buf, unsigned int len) | ||
1285 | { | ||
1286 | - struct skb_vnet_hdr *hdr = skb_vnet_hdr(skb); | ||
1287 | - struct page *page; | ||
1288 | - int num_buf, i, len; | ||
1289 | + struct sk_buff * skb = buf; | ||
1290 | + | ||
1291 | + len -= sizeof(struct virtio_net_hdr); | ||
1292 | + skb_trim(skb, len); | ||
1293 | + | ||
1294 | + return skb; | ||
1295 | +} | ||
1296 | + | ||
1297 | +static struct sk_buff *receive_big(struct net_device *dev, | ||
1298 | + struct receive_queue *rq, | ||
1299 | + void *buf) | ||
1300 | +{ | ||
1301 | + struct page *page = buf; | ||
1302 | + struct sk_buff *skb = page_to_skb(rq, page, 0); | ||
1303 | + | ||
1304 | + if (unlikely(!skb)) | ||
1305 | + goto err; | ||
1306 | + | ||
1307 | + return skb; | ||
1308 | + | ||
1309 | +err: | ||
1310 | + dev->stats.rx_dropped++; | ||
1311 | + give_pages(rq, page); | ||
1312 | + return NULL; | ||
1313 | +} | ||
1314 | + | ||
1315 | +static struct sk_buff *receive_mergeable(struct net_device *dev, | ||
1316 | + struct receive_queue *rq, | ||
1317 | + void *buf, | ||
1318 | + unsigned int len) | ||
1319 | +{ | ||
1320 | + struct skb_vnet_hdr *hdr = page_address(buf); | ||
1321 | + int num_buf = hdr->mhdr.num_buffers; | ||
1322 | + struct page *page = buf; | ||
1323 | + struct sk_buff *skb = page_to_skb(rq, page, len); | ||
1324 | + int i; | ||
1325 | + | ||
1326 | + if (unlikely(!skb)) | ||
1327 | + goto err_skb; | ||
1328 | |||
1329 | - num_buf = hdr->mhdr.num_buffers; | ||
1330 | while (--num_buf) { | ||
1331 | i = skb_shinfo(skb)->nr_frags; | ||
1332 | if (i >= MAX_SKB_FRAGS) { | ||
1333 | pr_debug("%s: packet too long\n", skb->dev->name); | ||
1334 | skb->dev->stats.rx_length_errors++; | ||
1335 | - return -EINVAL; | ||
1336 | + goto err_frags; | ||
1337 | } | ||
1338 | page = virtqueue_get_buf(rq->vq, &len); | ||
1339 | if (!page) { | ||
1340 | - pr_debug("%s: rx error: %d buffers missing\n", | ||
1341 | - skb->dev->name, hdr->mhdr.num_buffers); | ||
1342 | - skb->dev->stats.rx_length_errors++; | ||
1343 | - return -EINVAL; | ||
1344 | + pr_debug("%s: rx error: %d buffers %d missing\n", | ||
1345 | + dev->name, hdr->mhdr.num_buffers, num_buf); | ||
1346 | + dev->stats.rx_length_errors++; | ||
1347 | + goto err_buf; | ||
1348 | } | ||
1349 | |||
1350 | if (len > PAGE_SIZE) | ||
1351 | @@ -326,7 +361,26 @@ static int receive_mergeable(struct receive_queue *rq, struct sk_buff *skb) | ||
1352 | |||
1353 | --rq->num; | ||
1354 | } | ||
1355 | - return 0; | ||
1356 | + return skb; | ||
1357 | +err_skb: | ||
1358 | + give_pages(rq, page); | ||
1359 | + while (--num_buf) { | ||
1360 | +err_frags: | ||
1361 | + buf = virtqueue_get_buf(rq->vq, &len); | ||
1362 | + if (unlikely(!buf)) { | ||
1363 | + pr_debug("%s: rx error: %d buffers missing\n", | ||
1364 | + dev->name, num_buf); | ||
1365 | + dev->stats.rx_length_errors++; | ||
1366 | + break; | ||
1367 | + } | ||
1368 | + page = buf; | ||
1369 | + give_pages(rq, page); | ||
1370 | + --rq->num; | ||
1371 | + } | ||
1372 | +err_buf: | ||
1373 | + dev->stats.rx_dropped++; | ||
1374 | + dev_kfree_skb(skb); | ||
1375 | + return NULL; | ||
1376 | } | ||
1377 | |||
1378 | static void receive_buf(struct receive_queue *rq, void *buf, unsigned int len) | ||
1379 | @@ -335,7 +389,6 @@ static void receive_buf(struct receive_queue *rq, void *buf, unsigned int len) | ||
1380 | struct net_device *dev = vi->dev; | ||
1381 | struct virtnet_stats *stats = this_cpu_ptr(vi->stats); | ||
1382 | struct sk_buff *skb; | ||
1383 | - struct page *page; | ||
1384 | struct skb_vnet_hdr *hdr; | ||
1385 | |||
1386 | if (unlikely(len < sizeof(struct virtio_net_hdr) + ETH_HLEN)) { | ||
1387 | @@ -347,25 +400,15 @@ static void receive_buf(struct receive_queue *rq, void *buf, unsigned int len) | ||
1388 | dev_kfree_skb(buf); | ||
1389 | return; | ||
1390 | } | ||
1391 | + if (vi->mergeable_rx_bufs) | ||
1392 | + skb = receive_mergeable(dev, rq, buf, len); | ||
1393 | + else if (vi->big_packets) | ||
1394 | + skb = receive_big(dev, rq, buf); | ||
1395 | + else | ||
1396 | + skb = receive_small(buf, len); | ||
1397 | |||
1398 | - if (!vi->mergeable_rx_bufs && !vi->big_packets) { | ||
1399 | - skb = buf; | ||
1400 | - len -= sizeof(struct virtio_net_hdr); | ||
1401 | - skb_trim(skb, len); | ||
1402 | - } else { | ||
1403 | - page = buf; | ||
1404 | - skb = page_to_skb(rq, page, len); | ||
1405 | - if (unlikely(!skb)) { | ||
1406 | - dev->stats.rx_dropped++; | ||
1407 | - give_pages(rq, page); | ||
1408 | - return; | ||
1409 | - } | ||
1410 | - if (vi->mergeable_rx_bufs) | ||
1411 | - if (receive_mergeable(rq, skb)) { | ||
1412 | - dev_kfree_skb(skb); | ||
1413 | - return; | ||
1414 | - } | ||
1415 | - } | ||
1416 | + if (unlikely(!skb)) | ||
1417 | + return; | ||
1418 | |||
1419 | hdr = skb_vnet_hdr(skb); | ||
1420 | |||
1421 | @@ -1307,6 +1350,11 @@ static void virtnet_config_changed(struct virtio_device *vdev) | ||
1422 | |||
1423 | static void virtnet_free_queues(struct virtnet_info *vi) | ||
1424 | { | ||
1425 | + int i; | ||
1426 | + | ||
1427 | + for (i = 0; i < vi->max_queue_pairs; i++) | ||
1428 | + netif_napi_del(&vi->rq[i].napi); | ||
1429 | + | ||
1430 | kfree(vi->rq); | ||
1431 | kfree(vi->sq); | ||
1432 | } | ||
1433 | @@ -1724,16 +1772,17 @@ static int virtnet_restore(struct virtio_device *vdev) | ||
1434 | if (err) | ||
1435 | return err; | ||
1436 | |||
1437 | - if (netif_running(vi->dev)) | ||
1438 | + if (netif_running(vi->dev)) { | ||
1439 | + for (i = 0; i < vi->curr_queue_pairs; i++) | ||
1440 | + if (!try_fill_recv(&vi->rq[i], GFP_KERNEL)) | ||
1441 | + schedule_delayed_work(&vi->refill, 0); | ||
1442 | + | ||
1443 | for (i = 0; i < vi->max_queue_pairs; i++) | ||
1444 | virtnet_napi_enable(&vi->rq[i]); | ||
1445 | + } | ||
1446 | |||
1447 | netif_device_attach(vi->dev); | ||
1448 | |||
1449 | - for (i = 0; i < vi->curr_queue_pairs; i++) | ||
1450 | - if (!try_fill_recv(&vi->rq[i], GFP_KERNEL)) | ||
1451 | - schedule_delayed_work(&vi->refill, 0); | ||
1452 | - | ||
1453 | mutex_lock(&vi->config_lock); | ||
1454 | vi->config_enable = true; | ||
1455 | mutex_unlock(&vi->config_lock); | ||
1456 | diff --git a/drivers/net/vxlan.c b/drivers/net/vxlan.c | ||
1457 | index 2ef5b6219f3f..146236891889 100644 | ||
1458 | --- a/drivers/net/vxlan.c | ||
1459 | +++ b/drivers/net/vxlan.c | ||
1460 | @@ -1672,7 +1672,7 @@ static void vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev, | ||
1461 | netdev_dbg(dev, "circular route to %pI4\n", | ||
1462 | &dst->sin.sin_addr.s_addr); | ||
1463 | dev->stats.collisions++; | ||
1464 | - goto tx_error; | ||
1465 | + goto rt_tx_error; | ||
1466 | } | ||
1467 | |||
1468 | /* Bypass encapsulation if the destination is local */ | ||
1469 | diff --git a/drivers/scsi/sd.c b/drivers/scsi/sd.c | ||
1470 | index 2634d691ec17..dbc024bd4adf 100644 | ||
1471 | --- a/drivers/scsi/sd.c | ||
1472 | +++ b/drivers/scsi/sd.c | ||
1473 | @@ -2645,13 +2645,16 @@ static void sd_read_write_same(struct scsi_disk *sdkp, unsigned char *buffer) | ||
1474 | } | ||
1475 | |||
1476 | if (scsi_report_opcode(sdev, buffer, SD_BUF_SIZE, INQUIRY) < 0) { | ||
1477 | + /* too large values might cause issues with arcmsr */ | ||
1478 | + int vpd_buf_len = 64; | ||
1479 | + | ||
1480 | sdev->no_report_opcodes = 1; | ||
1481 | |||
1482 | /* Disable WRITE SAME if REPORT SUPPORTED OPERATION | ||
1483 | * CODES is unsupported and the device has an ATA | ||
1484 | * Information VPD page (SAT). | ||
1485 | */ | ||
1486 | - if (!scsi_get_vpd_page(sdev, 0x89, buffer, SD_BUF_SIZE)) | ||
1487 | + if (!scsi_get_vpd_page(sdev, 0x89, buffer, vpd_buf_len)) | ||
1488 | sdev->no_write_same = 1; | ||
1489 | } | ||
1490 | |||
1491 | diff --git a/include/linux/net.h b/include/linux/net.h | ||
1492 | index 8bd9d926b3cf..41103f84527e 100644 | ||
1493 | --- a/include/linux/net.h | ||
1494 | +++ b/include/linux/net.h | ||
1495 | @@ -180,7 +180,7 @@ struct proto_ops { | ||
1496 | int offset, size_t size, int flags); | ||
1497 | ssize_t (*splice_read)(struct socket *sock, loff_t *ppos, | ||
1498 | struct pipe_inode_info *pipe, size_t len, unsigned int flags); | ||
1499 | - void (*set_peek_off)(struct sock *sk, int val); | ||
1500 | + int (*set_peek_off)(struct sock *sk, int val); | ||
1501 | }; | ||
1502 | |||
1503 | #define DECLARE_SOCKADDR(type, dst, src) \ | ||
1504 | diff --git a/include/linux/netdevice.h b/include/linux/netdevice.h | ||
1505 | index 25f5d2d11e7c..21eae43348fb 100644 | ||
1506 | --- a/include/linux/netdevice.h | ||
1507 | +++ b/include/linux/netdevice.h | ||
1508 | @@ -1872,6 +1872,15 @@ static inline int dev_parse_header(const struct sk_buff *skb, | ||
1509 | return dev->header_ops->parse(skb, haddr); | ||
1510 | } | ||
1511 | |||
1512 | +static inline int dev_rebuild_header(struct sk_buff *skb) | ||
1513 | +{ | ||
1514 | + const struct net_device *dev = skb->dev; | ||
1515 | + | ||
1516 | + if (!dev->header_ops || !dev->header_ops->rebuild) | ||
1517 | + return 0; | ||
1518 | + return dev->header_ops->rebuild(skb); | ||
1519 | +} | ||
1520 | + | ||
1521 | typedef int gifconf_func_t(struct net_device * dev, char __user * bufptr, int len); | ||
1522 | extern int register_gifconf(unsigned int family, gifconf_func_t * gifconf); | ||
1523 | static inline int unregister_gifconf(unsigned int family) | ||
1524 | @@ -2945,6 +2954,19 @@ static inline void netif_set_gso_max_size(struct net_device *dev, | ||
1525 | dev->gso_max_size = size; | ||
1526 | } | ||
1527 | |||
1528 | +static inline void skb_gso_error_unwind(struct sk_buff *skb, __be16 protocol, | ||
1529 | + int pulled_hlen, u16 mac_offset, | ||
1530 | + int mac_len) | ||
1531 | +{ | ||
1532 | + skb->protocol = protocol; | ||
1533 | + skb->encapsulation = 1; | ||
1534 | + skb_push(skb, pulled_hlen); | ||
1535 | + skb_reset_transport_header(skb); | ||
1536 | + skb->mac_header = mac_offset; | ||
1537 | + skb->network_header = skb->mac_header + mac_len; | ||
1538 | + skb->mac_len = mac_len; | ||
1539 | +} | ||
1540 | + | ||
1541 | static inline bool netif_is_bond_master(struct net_device *dev) | ||
1542 | { | ||
1543 | return dev->flags & IFF_MASTER && dev->priv_flags & IFF_BONDING; | ||
1544 | diff --git a/include/linux/skbuff.h b/include/linux/skbuff.h | ||
1545 | index f66f346dd164..efa1649a822a 100644 | ||
1546 | --- a/include/linux/skbuff.h | ||
1547 | +++ b/include/linux/skbuff.h | ||
1548 | @@ -1638,6 +1638,11 @@ static inline void skb_set_mac_header(struct sk_buff *skb, const int offset) | ||
1549 | skb->mac_header += offset; | ||
1550 | } | ||
1551 | |||
1552 | +static inline void skb_pop_mac_header(struct sk_buff *skb) | ||
1553 | +{ | ||
1554 | + skb->mac_header = skb->network_header; | ||
1555 | +} | ||
1556 | + | ||
1557 | static inline void skb_probe_transport_header(struct sk_buff *skb, | ||
1558 | const int offset_hint) | ||
1559 | { | ||
1560 | diff --git a/kernel/sched/core.c b/kernel/sched/core.c | ||
1561 | index 5ac63c9a995a..ceae65e69a64 100644 | ||
1562 | --- a/kernel/sched/core.c | ||
1563 | +++ b/kernel/sched/core.c | ||
1564 | @@ -7277,7 +7277,12 @@ static int tg_set_cfs_bandwidth(struct task_group *tg, u64 period, u64 quota) | ||
1565 | |||
1566 | runtime_enabled = quota != RUNTIME_INF; | ||
1567 | runtime_was_enabled = cfs_b->quota != RUNTIME_INF; | ||
1568 | - account_cfs_bandwidth_used(runtime_enabled, runtime_was_enabled); | ||
1569 | + /* | ||
1570 | + * If we need to toggle cfs_bandwidth_used, off->on must occur | ||
1571 | + * before making related changes, and on->off must occur afterwards | ||
1572 | + */ | ||
1573 | + if (runtime_enabled && !runtime_was_enabled) | ||
1574 | + cfs_bandwidth_usage_inc(); | ||
1575 | raw_spin_lock_irq(&cfs_b->lock); | ||
1576 | cfs_b->period = ns_to_ktime(period); | ||
1577 | cfs_b->quota = quota; | ||
1578 | @@ -7303,6 +7308,8 @@ static int tg_set_cfs_bandwidth(struct task_group *tg, u64 period, u64 quota) | ||
1579 | unthrottle_cfs_rq(cfs_rq); | ||
1580 | raw_spin_unlock_irq(&rq->lock); | ||
1581 | } | ||
1582 | + if (runtime_was_enabled && !runtime_enabled) | ||
1583 | + cfs_bandwidth_usage_dec(); | ||
1584 | out_unlock: | ||
1585 | mutex_unlock(&cfs_constraints_mutex); | ||
1586 | |||
1587 | diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c | ||
1588 | index 7765ad82736a..411732334906 100644 | ||
1589 | --- a/kernel/sched/fair.c | ||
1590 | +++ b/kernel/sched/fair.c | ||
1591 | @@ -2077,13 +2077,14 @@ static inline bool cfs_bandwidth_used(void) | ||
1592 | return static_key_false(&__cfs_bandwidth_used); | ||
1593 | } | ||
1594 | |||
1595 | -void account_cfs_bandwidth_used(int enabled, int was_enabled) | ||
1596 | +void cfs_bandwidth_usage_inc(void) | ||
1597 | { | ||
1598 | - /* only need to count groups transitioning between enabled/!enabled */ | ||
1599 | - if (enabled && !was_enabled) | ||
1600 | - static_key_slow_inc(&__cfs_bandwidth_used); | ||
1601 | - else if (!enabled && was_enabled) | ||
1602 | - static_key_slow_dec(&__cfs_bandwidth_used); | ||
1603 | + static_key_slow_inc(&__cfs_bandwidth_used); | ||
1604 | +} | ||
1605 | + | ||
1606 | +void cfs_bandwidth_usage_dec(void) | ||
1607 | +{ | ||
1608 | + static_key_slow_dec(&__cfs_bandwidth_used); | ||
1609 | } | ||
1610 | #else /* HAVE_JUMP_LABEL */ | ||
1611 | static bool cfs_bandwidth_used(void) | ||
1612 | @@ -2091,7 +2092,8 @@ static bool cfs_bandwidth_used(void) | ||
1613 | return true; | ||
1614 | } | ||
1615 | |||
1616 | -void account_cfs_bandwidth_used(int enabled, int was_enabled) {} | ||
1617 | +void cfs_bandwidth_usage_inc(void) {} | ||
1618 | +void cfs_bandwidth_usage_dec(void) {} | ||
1619 | #endif /* HAVE_JUMP_LABEL */ | ||
1620 | |||
1621 | /* | ||
1622 | @@ -2457,6 +2459,13 @@ static int do_sched_cfs_period_timer(struct cfs_bandwidth *cfs_b, int overrun) | ||
1623 | if (idle) | ||
1624 | goto out_unlock; | ||
1625 | |||
1626 | + /* | ||
1627 | + * if we have relooped after returning idle once, we need to update our | ||
1628 | + * status as actually running, so that other cpus doing | ||
1629 | + * __start_cfs_bandwidth will stop trying to cancel us. | ||
1630 | + */ | ||
1631 | + cfs_b->timer_active = 1; | ||
1632 | + | ||
1633 | __refill_cfs_bandwidth_runtime(cfs_b); | ||
1634 | |||
1635 | if (!throttled) { | ||
1636 | @@ -2517,7 +2526,13 @@ static const u64 min_bandwidth_expiration = 2 * NSEC_PER_MSEC; | ||
1637 | /* how long we wait to gather additional slack before distributing */ | ||
1638 | static const u64 cfs_bandwidth_slack_period = 5 * NSEC_PER_MSEC; | ||
1639 | |||
1640 | -/* are we near the end of the current quota period? */ | ||
1641 | +/* | ||
1642 | + * Are we near the end of the current quota period? | ||
1643 | + * | ||
1644 | + * Requires cfs_b->lock for hrtimer_expires_remaining to be safe against the | ||
1645 | + * hrtimer base being cleared by __hrtimer_start_range_ns. In the case of | ||
1646 | + * migrate_hrtimers, base is never cleared, so we are fine. | ||
1647 | + */ | ||
1648 | static int runtime_refresh_within(struct cfs_bandwidth *cfs_b, u64 min_expire) | ||
1649 | { | ||
1650 | struct hrtimer *refresh_timer = &cfs_b->period_timer; | ||
1651 | @@ -2593,10 +2608,12 @@ static void do_sched_cfs_slack_timer(struct cfs_bandwidth *cfs_b) | ||
1652 | u64 expires; | ||
1653 | |||
1654 | /* confirm we're still not at a refresh boundary */ | ||
1655 | - if (runtime_refresh_within(cfs_b, min_bandwidth_expiration)) | ||
1656 | + raw_spin_lock(&cfs_b->lock); | ||
1657 | + if (runtime_refresh_within(cfs_b, min_bandwidth_expiration)) { | ||
1658 | + raw_spin_unlock(&cfs_b->lock); | ||
1659 | return; | ||
1660 | + } | ||
1661 | |||
1662 | - raw_spin_lock(&cfs_b->lock); | ||
1663 | if (cfs_b->quota != RUNTIME_INF && cfs_b->runtime > slice) { | ||
1664 | runtime = cfs_b->runtime; | ||
1665 | cfs_b->runtime = 0; | ||
1666 | @@ -2717,11 +2734,11 @@ void __start_cfs_bandwidth(struct cfs_bandwidth *cfs_b) | ||
1667 | * (timer_active==0 becomes visible before the hrtimer call-back | ||
1668 | * terminates). In either case we ensure that it's re-programmed | ||
1669 | */ | ||
1670 | - while (unlikely(hrtimer_active(&cfs_b->period_timer))) { | ||
1671 | + while (unlikely(hrtimer_active(&cfs_b->period_timer)) && | ||
1672 | + hrtimer_try_to_cancel(&cfs_b->period_timer) < 0) { | ||
1673 | + /* bounce the lock to allow do_sched_cfs_period_timer to run */ | ||
1674 | raw_spin_unlock(&cfs_b->lock); | ||
1675 | - /* ensure cfs_b->lock is available while we wait */ | ||
1676 | - hrtimer_cancel(&cfs_b->period_timer); | ||
1677 | - | ||
1678 | + cpu_relax(); | ||
1679 | raw_spin_lock(&cfs_b->lock); | ||
1680 | /* if someone else restarted the timer then we're done */ | ||
1681 | if (cfs_b->timer_active) | ||
1682 | @@ -6223,7 +6240,8 @@ void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, | ||
1683 | se->cfs_rq = parent->my_q; | ||
1684 | |||
1685 | se->my_q = cfs_rq; | ||
1686 | - update_load_set(&se->load, 0); | ||
1687 | + /* guarantee group entities always have weight */ | ||
1688 | + update_load_set(&se->load, NICE_0_LOAD); | ||
1689 | se->parent = parent; | ||
1690 | } | ||
1691 | |||
1692 | diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h | ||
1693 | index b3c5653e1dca..a6208afd80e7 100644 | ||
1694 | --- a/kernel/sched/sched.h | ||
1695 | +++ b/kernel/sched/sched.h | ||
1696 | @@ -1305,7 +1305,8 @@ extern void print_rt_stats(struct seq_file *m, int cpu); | ||
1697 | extern void init_cfs_rq(struct cfs_rq *cfs_rq); | ||
1698 | extern void init_rt_rq(struct rt_rq *rt_rq, struct rq *rq); | ||
1699 | |||
1700 | -extern void account_cfs_bandwidth_used(int enabled, int was_enabled); | ||
1701 | +extern void cfs_bandwidth_usage_inc(void); | ||
1702 | +extern void cfs_bandwidth_usage_dec(void); | ||
1703 | |||
1704 | #ifdef CONFIG_NO_HZ_COMMON | ||
1705 | enum rq_nohz_flag_bits { | ||
1706 | diff --git a/net/8021q/vlan_dev.c b/net/8021q/vlan_dev.c | ||
1707 | index 09bf1c38805b..edf44d079da7 100644 | ||
1708 | --- a/net/8021q/vlan_dev.c | ||
1709 | +++ b/net/8021q/vlan_dev.c | ||
1710 | @@ -549,6 +549,23 @@ static const struct header_ops vlan_header_ops = { | ||
1711 | .parse = eth_header_parse, | ||
1712 | }; | ||
1713 | |||
1714 | +static int vlan_passthru_hard_header(struct sk_buff *skb, struct net_device *dev, | ||
1715 | + unsigned short type, | ||
1716 | + const void *daddr, const void *saddr, | ||
1717 | + unsigned int len) | ||
1718 | +{ | ||
1719 | + struct vlan_dev_priv *vlan = vlan_dev_priv(dev); | ||
1720 | + struct net_device *real_dev = vlan->real_dev; | ||
1721 | + | ||
1722 | + return dev_hard_header(skb, real_dev, type, daddr, saddr, len); | ||
1723 | +} | ||
1724 | + | ||
1725 | +static const struct header_ops vlan_passthru_header_ops = { | ||
1726 | + .create = vlan_passthru_hard_header, | ||
1727 | + .rebuild = dev_rebuild_header, | ||
1728 | + .parse = eth_header_parse, | ||
1729 | +}; | ||
1730 | + | ||
1731 | static struct device_type vlan_type = { | ||
1732 | .name = "vlan", | ||
1733 | }; | ||
1734 | @@ -592,7 +609,7 @@ static int vlan_dev_init(struct net_device *dev) | ||
1735 | |||
1736 | dev->needed_headroom = real_dev->needed_headroom; | ||
1737 | if (real_dev->features & NETIF_F_HW_VLAN_CTAG_TX) { | ||
1738 | - dev->header_ops = real_dev->header_ops; | ||
1739 | + dev->header_ops = &vlan_passthru_header_ops; | ||
1740 | dev->hard_header_len = real_dev->hard_header_len; | ||
1741 | } else { | ||
1742 | dev->header_ops = &vlan_header_ops; | ||
1743 | diff --git a/net/bridge/br_multicast.c b/net/bridge/br_multicast.c | ||
1744 | index 8b0b610ca2c9..1b148a3affa7 100644 | ||
1745 | --- a/net/bridge/br_multicast.c | ||
1746 | +++ b/net/bridge/br_multicast.c | ||
1747 | @@ -2004,7 +2004,7 @@ int br_multicast_set_hash_max(struct net_bridge *br, unsigned long val) | ||
1748 | u32 old; | ||
1749 | struct net_bridge_mdb_htable *mdb; | ||
1750 | |||
1751 | - spin_lock(&br->multicast_lock); | ||
1752 | + spin_lock_bh(&br->multicast_lock); | ||
1753 | if (!netif_running(br->dev)) | ||
1754 | goto unlock; | ||
1755 | |||
1756 | @@ -2036,7 +2036,7 @@ rollback: | ||
1757 | } | ||
1758 | |||
1759 | unlock: | ||
1760 | - spin_unlock(&br->multicast_lock); | ||
1761 | + spin_unlock_bh(&br->multicast_lock); | ||
1762 | |||
1763 | return err; | ||
1764 | } | ||
1765 | diff --git a/net/bridge/br_private.h b/net/bridge/br_private.h | ||
1766 | index e14c33b42f75..9a63c4206e4a 100644 | ||
1767 | --- a/net/bridge/br_private.h | ||
1768 | +++ b/net/bridge/br_private.h | ||
1769 | @@ -442,6 +442,16 @@ extern netdev_features_t br_features_recompute(struct net_bridge *br, | ||
1770 | extern int br_handle_frame_finish(struct sk_buff *skb); | ||
1771 | extern rx_handler_result_t br_handle_frame(struct sk_buff **pskb); | ||
1772 | |||
1773 | +static inline bool br_rx_handler_check_rcu(const struct net_device *dev) | ||
1774 | +{ | ||
1775 | + return rcu_dereference(dev->rx_handler) == br_handle_frame; | ||
1776 | +} | ||
1777 | + | ||
1778 | +static inline struct net_bridge_port *br_port_get_check_rcu(const struct net_device *dev) | ||
1779 | +{ | ||
1780 | + return br_rx_handler_check_rcu(dev) ? br_port_get_rcu(dev) : NULL; | ||
1781 | +} | ||
1782 | + | ||
1783 | /* br_ioctl.c */ | ||
1784 | extern int br_dev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); | ||
1785 | extern int br_ioctl_deviceless_stub(struct net *net, unsigned int cmd, void __user *arg); | ||
1786 | diff --git a/net/bridge/br_stp_bpdu.c b/net/bridge/br_stp_bpdu.c | ||
1787 | index 8660ea3be705..bdb459d21ad8 100644 | ||
1788 | --- a/net/bridge/br_stp_bpdu.c | ||
1789 | +++ b/net/bridge/br_stp_bpdu.c | ||
1790 | @@ -153,7 +153,7 @@ void br_stp_rcv(const struct stp_proto *proto, struct sk_buff *skb, | ||
1791 | if (buf[0] != 0 || buf[1] != 0 || buf[2] != 0) | ||
1792 | goto err; | ||
1793 | |||
1794 | - p = br_port_get_rcu(dev); | ||
1795 | + p = br_port_get_check_rcu(dev); | ||
1796 | if (!p) | ||
1797 | goto err; | ||
1798 | |||
1799 | diff --git a/net/core/drop_monitor.c b/net/core/drop_monitor.c | ||
1800 | index 5e78d44333b9..f27d126239b1 100644 | ||
1801 | --- a/net/core/drop_monitor.c | ||
1802 | +++ b/net/core/drop_monitor.c | ||
1803 | @@ -64,7 +64,6 @@ static struct genl_family net_drop_monitor_family = { | ||
1804 | .hdrsize = 0, | ||
1805 | .name = "NET_DM", | ||
1806 | .version = 2, | ||
1807 | - .maxattr = NET_DM_CMD_MAX, | ||
1808 | }; | ||
1809 | |||
1810 | static DEFINE_PER_CPU(struct per_cpu_dm_data, dm_cpu_data); | ||
1811 | diff --git a/net/core/neighbour.c b/net/core/neighbour.c | ||
1812 | index 6072610a8672..11af243bf92f 100644 | ||
1813 | --- a/net/core/neighbour.c | ||
1814 | +++ b/net/core/neighbour.c | ||
1815 | @@ -1274,7 +1274,7 @@ int neigh_compat_output(struct neighbour *neigh, struct sk_buff *skb) | ||
1816 | |||
1817 | if (dev_hard_header(skb, dev, ntohs(skb->protocol), NULL, NULL, | ||
1818 | skb->len) < 0 && | ||
1819 | - dev->header_ops->rebuild(skb)) | ||
1820 | + dev_rebuild_header(skb)) | ||
1821 | return 0; | ||
1822 | |||
1823 | return dev_queue_xmit(skb); | ||
1824 | diff --git a/net/core/netpoll.c b/net/core/netpoll.c | ||
1825 | index fc75c9e461b8..0c1482c6ff98 100644 | ||
1826 | --- a/net/core/netpoll.c | ||
1827 | +++ b/net/core/netpoll.c | ||
1828 | @@ -386,8 +386,14 @@ void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb, | ||
1829 | !vlan_hw_offload_capable(netif_skb_features(skb), | ||
1830 | skb->vlan_proto)) { | ||
1831 | skb = __vlan_put_tag(skb, skb->vlan_proto, vlan_tx_tag_get(skb)); | ||
1832 | - if (unlikely(!skb)) | ||
1833 | - break; | ||
1834 | + if (unlikely(!skb)) { | ||
1835 | + /* This is actually a packet drop, but we | ||
1836 | + * don't want the code at the end of this | ||
1837 | + * function to try and re-queue a NULL skb. | ||
1838 | + */ | ||
1839 | + status = NETDEV_TX_OK; | ||
1840 | + goto unlock_txq; | ||
1841 | + } | ||
1842 | skb->vlan_tci = 0; | ||
1843 | } | ||
1844 | |||
1845 | @@ -395,6 +401,7 @@ void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb, | ||
1846 | if (status == NETDEV_TX_OK) | ||
1847 | txq_trans_update(txq); | ||
1848 | } | ||
1849 | + unlock_txq: | ||
1850 | __netif_tx_unlock(txq); | ||
1851 | |||
1852 | if (status == NETDEV_TX_OK) | ||
1853 | diff --git a/net/core/skbuff.c b/net/core/skbuff.c | ||
1854 | index c28c7fed0d0b..743e6ebf5f9f 100644 | ||
1855 | --- a/net/core/skbuff.c | ||
1856 | +++ b/net/core/skbuff.c | ||
1857 | @@ -3541,6 +3541,7 @@ void skb_scrub_packet(struct sk_buff *skb, bool xnet) | ||
1858 | skb->tstamp.tv64 = 0; | ||
1859 | skb->pkt_type = PACKET_HOST; | ||
1860 | skb->skb_iif = 0; | ||
1861 | + skb->local_df = 0; | ||
1862 | skb_dst_drop(skb); | ||
1863 | skb->mark = 0; | ||
1864 | secpath_reset(skb); | ||
1865 | diff --git a/net/core/sock.c b/net/core/sock.c | ||
1866 | index 0b39e7ae4383..5cec994ee2f3 100644 | ||
1867 | --- a/net/core/sock.c | ||
1868 | +++ b/net/core/sock.c | ||
1869 | @@ -888,7 +888,7 @@ set_rcvbuf: | ||
1870 | |||
1871 | case SO_PEEK_OFF: | ||
1872 | if (sock->ops->set_peek_off) | ||
1873 | - sock->ops->set_peek_off(sk, val); | ||
1874 | + ret = sock->ops->set_peek_off(sk, val); | ||
1875 | else | ||
1876 | ret = -EOPNOTSUPP; | ||
1877 | break; | ||
1878 | diff --git a/net/ipv4/fib_rules.c b/net/ipv4/fib_rules.c | ||
1879 | index 523be38e37de..f2e15738534d 100644 | ||
1880 | --- a/net/ipv4/fib_rules.c | ||
1881 | +++ b/net/ipv4/fib_rules.c | ||
1882 | @@ -104,7 +104,10 @@ errout: | ||
1883 | static bool fib4_rule_suppress(struct fib_rule *rule, struct fib_lookup_arg *arg) | ||
1884 | { | ||
1885 | struct fib_result *result = (struct fib_result *) arg->result; | ||
1886 | - struct net_device *dev = result->fi->fib_dev; | ||
1887 | + struct net_device *dev = NULL; | ||
1888 | + | ||
1889 | + if (result->fi) | ||
1890 | + dev = result->fi->fib_dev; | ||
1891 | |||
1892 | /* do not accept result if the route does | ||
1893 | * not meet the required prefix length | ||
1894 | diff --git a/net/ipv4/gre_offload.c b/net/ipv4/gre_offload.c | ||
1895 | index 55e6bfb3a289..db98705905f7 100644 | ||
1896 | --- a/net/ipv4/gre_offload.c | ||
1897 | +++ b/net/ipv4/gre_offload.c | ||
1898 | @@ -28,6 +28,7 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb, | ||
1899 | netdev_features_t enc_features; | ||
1900 | int ghl = GRE_HEADER_SECTION; | ||
1901 | struct gre_base_hdr *greh; | ||
1902 | + u16 mac_offset = skb->mac_header; | ||
1903 | int mac_len = skb->mac_len; | ||
1904 | __be16 protocol = skb->protocol; | ||
1905 | int tnl_hlen; | ||
1906 | @@ -57,13 +58,13 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb, | ||
1907 | } else | ||
1908 | csum = false; | ||
1909 | |||
1910 | + if (unlikely(!pskb_may_pull(skb, ghl))) | ||
1911 | + goto out; | ||
1912 | + | ||
1913 | /* setup inner skb. */ | ||
1914 | skb->protocol = greh->protocol; | ||
1915 | skb->encapsulation = 0; | ||
1916 | |||
1917 | - if (unlikely(!pskb_may_pull(skb, ghl))) | ||
1918 | - goto out; | ||
1919 | - | ||
1920 | __skb_pull(skb, ghl); | ||
1921 | skb_reset_mac_header(skb); | ||
1922 | skb_set_network_header(skb, skb_inner_network_offset(skb)); | ||
1923 | @@ -72,8 +73,10 @@ static struct sk_buff *gre_gso_segment(struct sk_buff *skb, | ||
1924 | /* segment inner packet. */ | ||
1925 | enc_features = skb->dev->hw_enc_features & netif_skb_features(skb); | ||
1926 | segs = skb_mac_gso_segment(skb, enc_features); | ||
1927 | - if (!segs || IS_ERR(segs)) | ||
1928 | + if (!segs || IS_ERR(segs)) { | ||
1929 | + skb_gso_error_unwind(skb, protocol, ghl, mac_offset, mac_len); | ||
1930 | goto out; | ||
1931 | + } | ||
1932 | |||
1933 | skb = segs; | ||
1934 | tnl_hlen = skb_tnl_header_len(skb); | ||
1935 | diff --git a/net/ipv4/inet_diag.c b/net/ipv4/inet_diag.c | ||
1936 | index 5f648751fce2..31cf54d18221 100644 | ||
1937 | --- a/net/ipv4/inet_diag.c | ||
1938 | +++ b/net/ipv4/inet_diag.c | ||
1939 | @@ -106,6 +106,10 @@ int inet_sk_diag_fill(struct sock *sk, struct inet_connection_sock *icsk, | ||
1940 | |||
1941 | r->id.idiag_sport = inet->inet_sport; | ||
1942 | r->id.idiag_dport = inet->inet_dport; | ||
1943 | + | ||
1944 | + memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src)); | ||
1945 | + memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst)); | ||
1946 | + | ||
1947 | r->id.idiag_src[0] = inet->inet_rcv_saddr; | ||
1948 | r->id.idiag_dst[0] = inet->inet_daddr; | ||
1949 | |||
1950 | @@ -240,12 +244,19 @@ static int inet_twsk_diag_fill(struct inet_timewait_sock *tw, | ||
1951 | |||
1952 | r->idiag_family = tw->tw_family; | ||
1953 | r->idiag_retrans = 0; | ||
1954 | + | ||
1955 | r->id.idiag_if = tw->tw_bound_dev_if; | ||
1956 | sock_diag_save_cookie(tw, r->id.idiag_cookie); | ||
1957 | + | ||
1958 | r->id.idiag_sport = tw->tw_sport; | ||
1959 | r->id.idiag_dport = tw->tw_dport; | ||
1960 | + | ||
1961 | + memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src)); | ||
1962 | + memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst)); | ||
1963 | + | ||
1964 | r->id.idiag_src[0] = tw->tw_rcv_saddr; | ||
1965 | r->id.idiag_dst[0] = tw->tw_daddr; | ||
1966 | + | ||
1967 | r->idiag_state = tw->tw_substate; | ||
1968 | r->idiag_timer = 3; | ||
1969 | r->idiag_expires = DIV_ROUND_UP(tmo * 1000, HZ); | ||
1970 | @@ -732,8 +743,13 @@ static int inet_diag_fill_req(struct sk_buff *skb, struct sock *sk, | ||
1971 | |||
1972 | r->id.idiag_sport = inet->inet_sport; | ||
1973 | r->id.idiag_dport = ireq->rmt_port; | ||
1974 | + | ||
1975 | + memset(&r->id.idiag_src, 0, sizeof(r->id.idiag_src)); | ||
1976 | + memset(&r->id.idiag_dst, 0, sizeof(r->id.idiag_dst)); | ||
1977 | + | ||
1978 | r->id.idiag_src[0] = ireq->loc_addr; | ||
1979 | r->id.idiag_dst[0] = ireq->rmt_addr; | ||
1980 | + | ||
1981 | r->idiag_expires = jiffies_to_msecs(tmo); | ||
1982 | r->idiag_rqueue = 0; | ||
1983 | r->idiag_wqueue = 0; | ||
1984 | diff --git a/net/ipv4/ip_gre.c b/net/ipv4/ip_gre.c | ||
1985 | index d7aea4c5b940..e560ef34cf4b 100644 | ||
1986 | --- a/net/ipv4/ip_gre.c | ||
1987 | +++ b/net/ipv4/ip_gre.c | ||
1988 | @@ -217,6 +217,7 @@ static int ipgre_rcv(struct sk_buff *skb, const struct tnl_ptk_info *tpi) | ||
1989 | iph->saddr, iph->daddr, tpi->key); | ||
1990 | |||
1991 | if (tunnel) { | ||
1992 | + skb_pop_mac_header(skb); | ||
1993 | ip_tunnel_rcv(tunnel, skb, tpi, log_ecn_error); | ||
1994 | return PACKET_RCVD; | ||
1995 | } | ||
1996 | diff --git a/net/ipv4/udp.c b/net/ipv4/udp.c | ||
1997 | index 5e2c2f1a075d..6ca990726d5b 100644 | ||
1998 | --- a/net/ipv4/udp.c | ||
1999 | +++ b/net/ipv4/udp.c | ||
2000 | @@ -2294,6 +2294,7 @@ struct sk_buff *skb_udp_tunnel_segment(struct sk_buff *skb, | ||
2001 | netdev_features_t features) | ||
2002 | { | ||
2003 | struct sk_buff *segs = ERR_PTR(-EINVAL); | ||
2004 | + u16 mac_offset = skb->mac_header; | ||
2005 | int mac_len = skb->mac_len; | ||
2006 | int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb); | ||
2007 | __be16 protocol = skb->protocol; | ||
2008 | @@ -2313,8 +2314,11 @@ struct sk_buff *skb_udp_tunnel_segment(struct sk_buff *skb, | ||
2009 | /* segment inner packet. */ | ||
2010 | enc_features = skb->dev->hw_enc_features & netif_skb_features(skb); | ||
2011 | segs = skb_mac_gso_segment(skb, enc_features); | ||
2012 | - if (!segs || IS_ERR(segs)) | ||
2013 | + if (!segs || IS_ERR(segs)) { | ||
2014 | + skb_gso_error_unwind(skb, protocol, tnl_hlen, mac_offset, | ||
2015 | + mac_len); | ||
2016 | goto out; | ||
2017 | + } | ||
2018 | |||
2019 | outer_hlen = skb_tnl_header_len(skb); | ||
2020 | skb = segs; | ||
2021 | diff --git a/net/ipv4/udp_offload.c b/net/ipv4/udp_offload.c | ||
2022 | index f35eccaa855e..6b809e4bf1ed 100644 | ||
2023 | --- a/net/ipv4/udp_offload.c | ||
2024 | +++ b/net/ipv4/udp_offload.c | ||
2025 | @@ -41,6 +41,14 @@ static struct sk_buff *udp4_ufo_fragment(struct sk_buff *skb, | ||
2026 | { | ||
2027 | struct sk_buff *segs = ERR_PTR(-EINVAL); | ||
2028 | unsigned int mss; | ||
2029 | + int offset; | ||
2030 | + __wsum csum; | ||
2031 | + | ||
2032 | + if (skb->encapsulation && | ||
2033 | + skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL) { | ||
2034 | + segs = skb_udp_tunnel_segment(skb, features); | ||
2035 | + goto out; | ||
2036 | + } | ||
2037 | |||
2038 | mss = skb_shinfo(skb)->gso_size; | ||
2039 | if (unlikely(skb->len <= mss)) | ||
2040 | @@ -62,27 +70,20 @@ static struct sk_buff *udp4_ufo_fragment(struct sk_buff *skb, | ||
2041 | goto out; | ||
2042 | } | ||
2043 | |||
2044 | + /* Do software UFO. Complete and fill in the UDP checksum as | ||
2045 | + * HW cannot do checksum of UDP packets sent as multiple | ||
2046 | + * IP fragments. | ||
2047 | + */ | ||
2048 | + offset = skb_checksum_start_offset(skb); | ||
2049 | + csum = skb_checksum(skb, offset, skb->len - offset, 0); | ||
2050 | + offset += skb->csum_offset; | ||
2051 | + *(__sum16 *)(skb->data + offset) = csum_fold(csum); | ||
2052 | + skb->ip_summed = CHECKSUM_NONE; | ||
2053 | + | ||
2054 | /* Fragment the skb. IP headers of the fragments are updated in | ||
2055 | * inet_gso_segment() | ||
2056 | */ | ||
2057 | - if (skb->encapsulation && skb_shinfo(skb)->gso_type & SKB_GSO_UDP_TUNNEL) | ||
2058 | - segs = skb_udp_tunnel_segment(skb, features); | ||
2059 | - else { | ||
2060 | - int offset; | ||
2061 | - __wsum csum; | ||
2062 | - | ||
2063 | - /* Do software UFO. Complete and fill in the UDP checksum as | ||
2064 | - * HW cannot do checksum of UDP packets sent as multiple | ||
2065 | - * IP fragments. | ||
2066 | - */ | ||
2067 | - offset = skb_checksum_start_offset(skb); | ||
2068 | - csum = skb_checksum(skb, offset, skb->len - offset, 0); | ||
2069 | - offset += skb->csum_offset; | ||
2070 | - *(__sum16 *)(skb->data + offset) = csum_fold(csum); | ||
2071 | - skb->ip_summed = CHECKSUM_NONE; | ||
2072 | - | ||
2073 | - segs = skb_segment(skb, features); | ||
2074 | - } | ||
2075 | + segs = skb_segment(skb, features); | ||
2076 | out: | ||
2077 | return segs; | ||
2078 | } | ||
2079 | diff --git a/net/ipv6/fib6_rules.c b/net/ipv6/fib6_rules.c | ||
2080 | index e27591635f92..3fd0a578329e 100644 | ||
2081 | --- a/net/ipv6/fib6_rules.c | ||
2082 | +++ b/net/ipv6/fib6_rules.c | ||
2083 | @@ -122,7 +122,11 @@ out: | ||
2084 | static bool fib6_rule_suppress(struct fib_rule *rule, struct fib_lookup_arg *arg) | ||
2085 | { | ||
2086 | struct rt6_info *rt = (struct rt6_info *) arg->result; | ||
2087 | - struct net_device *dev = rt->rt6i_idev->dev; | ||
2088 | + struct net_device *dev = NULL; | ||
2089 | + | ||
2090 | + if (rt->rt6i_idev) | ||
2091 | + dev = rt->rt6i_idev->dev; | ||
2092 | + | ||
2093 | /* do not accept result if the route does | ||
2094 | * not meet the required prefix length | ||
2095 | */ | ||
2096 | diff --git a/net/ipv6/route.c b/net/ipv6/route.c | ||
2097 | index 77308af056bc..0accb1321dd6 100644 | ||
2098 | --- a/net/ipv6/route.c | ||
2099 | +++ b/net/ipv6/route.c | ||
2100 | @@ -84,6 +84,8 @@ static int ip6_dst_gc(struct dst_ops *ops); | ||
2101 | |||
2102 | static int ip6_pkt_discard(struct sk_buff *skb); | ||
2103 | static int ip6_pkt_discard_out(struct sk_buff *skb); | ||
2104 | +static int ip6_pkt_prohibit(struct sk_buff *skb); | ||
2105 | +static int ip6_pkt_prohibit_out(struct sk_buff *skb); | ||
2106 | static void ip6_link_failure(struct sk_buff *skb); | ||
2107 | static void ip6_rt_update_pmtu(struct dst_entry *dst, struct sock *sk, | ||
2108 | struct sk_buff *skb, u32 mtu); | ||
2109 | @@ -234,9 +236,6 @@ static const struct rt6_info ip6_null_entry_template = { | ||
2110 | |||
2111 | #ifdef CONFIG_IPV6_MULTIPLE_TABLES | ||
2112 | |||
2113 | -static int ip6_pkt_prohibit(struct sk_buff *skb); | ||
2114 | -static int ip6_pkt_prohibit_out(struct sk_buff *skb); | ||
2115 | - | ||
2116 | static const struct rt6_info ip6_prohibit_entry_template = { | ||
2117 | .dst = { | ||
2118 | .__refcnt = ATOMIC_INIT(1), | ||
2119 | @@ -1570,21 +1569,24 @@ int ip6_route_add(struct fib6_config *cfg) | ||
2120 | goto out; | ||
2121 | } | ||
2122 | } | ||
2123 | - rt->dst.output = ip6_pkt_discard_out; | ||
2124 | - rt->dst.input = ip6_pkt_discard; | ||
2125 | rt->rt6i_flags = RTF_REJECT|RTF_NONEXTHOP; | ||
2126 | switch (cfg->fc_type) { | ||
2127 | case RTN_BLACKHOLE: | ||
2128 | rt->dst.error = -EINVAL; | ||
2129 | + rt->dst.output = dst_discard; | ||
2130 | + rt->dst.input = dst_discard; | ||
2131 | break; | ||
2132 | case RTN_PROHIBIT: | ||
2133 | rt->dst.error = -EACCES; | ||
2134 | + rt->dst.output = ip6_pkt_prohibit_out; | ||
2135 | + rt->dst.input = ip6_pkt_prohibit; | ||
2136 | break; | ||
2137 | case RTN_THROW: | ||
2138 | - rt->dst.error = -EAGAIN; | ||
2139 | - break; | ||
2140 | default: | ||
2141 | - rt->dst.error = -ENETUNREACH; | ||
2142 | + rt->dst.error = (cfg->fc_type == RTN_THROW) ? -EAGAIN | ||
2143 | + : -ENETUNREACH; | ||
2144 | + rt->dst.output = ip6_pkt_discard_out; | ||
2145 | + rt->dst.input = ip6_pkt_discard; | ||
2146 | break; | ||
2147 | } | ||
2148 | goto install_route; | ||
2149 | @@ -1908,9 +1910,7 @@ static struct rt6_info *ip6_rt_copy(struct rt6_info *ort, | ||
2150 | else | ||
2151 | rt->rt6i_gateway = *dest; | ||
2152 | rt->rt6i_flags = ort->rt6i_flags; | ||
2153 | - if ((ort->rt6i_flags & (RTF_DEFAULT | RTF_ADDRCONF)) == | ||
2154 | - (RTF_DEFAULT | RTF_ADDRCONF)) | ||
2155 | - rt6_set_from(rt, ort); | ||
2156 | + rt6_set_from(rt, ort); | ||
2157 | rt->rt6i_metric = 0; | ||
2158 | |||
2159 | #ifdef CONFIG_IPV6_SUBTREES | ||
2160 | @@ -2149,8 +2149,6 @@ static int ip6_pkt_discard_out(struct sk_buff *skb) | ||
2161 | return ip6_pkt_drop(skb, ICMPV6_NOROUTE, IPSTATS_MIB_OUTNOROUTES); | ||
2162 | } | ||
2163 | |||
2164 | -#ifdef CONFIG_IPV6_MULTIPLE_TABLES | ||
2165 | - | ||
2166 | static int ip6_pkt_prohibit(struct sk_buff *skb) | ||
2167 | { | ||
2168 | return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_INNOROUTES); | ||
2169 | @@ -2162,8 +2160,6 @@ static int ip6_pkt_prohibit_out(struct sk_buff *skb) | ||
2170 | return ip6_pkt_drop(skb, ICMPV6_ADM_PROHIBITED, IPSTATS_MIB_OUTNOROUTES); | ||
2171 | } | ||
2172 | |||
2173 | -#endif | ||
2174 | - | ||
2175 | /* | ||
2176 | * Allocate a dst for local (unicast / anycast) address. | ||
2177 | */ | ||
2178 | @@ -2173,12 +2169,10 @@ struct rt6_info *addrconf_dst_alloc(struct inet6_dev *idev, | ||
2179 | bool anycast) | ||
2180 | { | ||
2181 | struct net *net = dev_net(idev->dev); | ||
2182 | - struct rt6_info *rt = ip6_dst_alloc(net, net->loopback_dev, 0, NULL); | ||
2183 | - | ||
2184 | - if (!rt) { | ||
2185 | - net_warn_ratelimited("Maximum number of routes reached, consider increasing route/max_size\n"); | ||
2186 | + struct rt6_info *rt = ip6_dst_alloc(net, net->loopback_dev, | ||
2187 | + DST_NOCOUNT, NULL); | ||
2188 | + if (!rt) | ||
2189 | return ERR_PTR(-ENOMEM); | ||
2190 | - } | ||
2191 | |||
2192 | in6_dev_hold(idev); | ||
2193 | |||
2194 | diff --git a/net/ipv6/udp_offload.c b/net/ipv6/udp_offload.c | ||
2195 | index 34c6fff3ae84..06556d6e1a4d 100644 | ||
2196 | --- a/net/ipv6/udp_offload.c | ||
2197 | +++ b/net/ipv6/udp_offload.c | ||
2198 | @@ -88,7 +88,7 @@ static struct sk_buff *udp6_ufo_fragment(struct sk_buff *skb, | ||
2199 | |||
2200 | /* Check if there is enough headroom to insert fragment header. */ | ||
2201 | tnl_hlen = skb_tnl_header_len(skb); | ||
2202 | - if (skb->mac_header < (tnl_hlen + frag_hdr_sz)) { | ||
2203 | + if (skb_mac_header(skb) < skb->head + tnl_hlen + frag_hdr_sz) { | ||
2204 | if (gso_pskb_expand_head(skb, tnl_hlen + frag_hdr_sz)) | ||
2205 | goto out; | ||
2206 | } | ||
2207 | diff --git a/net/llc/af_llc.c b/net/llc/af_llc.c | ||
2208 | index 7b01b9f5846c..c71b699eb555 100644 | ||
2209 | --- a/net/llc/af_llc.c | ||
2210 | +++ b/net/llc/af_llc.c | ||
2211 | @@ -715,7 +715,7 @@ static int llc_ui_recvmsg(struct kiocb *iocb, struct socket *sock, | ||
2212 | unsigned long cpu_flags; | ||
2213 | size_t copied = 0; | ||
2214 | u32 peek_seq = 0; | ||
2215 | - u32 *seq; | ||
2216 | + u32 *seq, skb_len; | ||
2217 | unsigned long used; | ||
2218 | int target; /* Read at least this many bytes */ | ||
2219 | long timeo; | ||
2220 | @@ -812,6 +812,7 @@ static int llc_ui_recvmsg(struct kiocb *iocb, struct socket *sock, | ||
2221 | } | ||
2222 | continue; | ||
2223 | found_ok_skb: | ||
2224 | + skb_len = skb->len; | ||
2225 | /* Ok so how much can we use? */ | ||
2226 | used = skb->len - offset; | ||
2227 | if (len < used) | ||
2228 | @@ -844,7 +845,7 @@ static int llc_ui_recvmsg(struct kiocb *iocb, struct socket *sock, | ||
2229 | } | ||
2230 | |||
2231 | /* Partial read */ | ||
2232 | - if (used + offset < skb->len) | ||
2233 | + if (used + offset < skb_len) | ||
2234 | continue; | ||
2235 | } while (len > 0); | ||
2236 | |||
2237 | diff --git a/net/mac80211/tx.c b/net/mac80211/tx.c | ||
2238 | index 70b5a05c0a4e..1eb2b78e927b 100644 | ||
2239 | --- a/net/mac80211/tx.c | ||
2240 | +++ b/net/mac80211/tx.c | ||
2241 | @@ -463,7 +463,6 @@ ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx) | ||
2242 | { | ||
2243 | struct sta_info *sta = tx->sta; | ||
2244 | struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); | ||
2245 | - struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; | ||
2246 | struct ieee80211_local *local = tx->local; | ||
2247 | |||
2248 | if (unlikely(!sta)) | ||
2249 | @@ -474,15 +473,6 @@ ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx) | ||
2250 | !(info->flags & IEEE80211_TX_CTL_NO_PS_BUFFER))) { | ||
2251 | int ac = skb_get_queue_mapping(tx->skb); | ||
2252 | |||
2253 | - /* only deauth, disassoc and action are bufferable MMPDUs */ | ||
2254 | - if (ieee80211_is_mgmt(hdr->frame_control) && | ||
2255 | - !ieee80211_is_deauth(hdr->frame_control) && | ||
2256 | - !ieee80211_is_disassoc(hdr->frame_control) && | ||
2257 | - !ieee80211_is_action(hdr->frame_control)) { | ||
2258 | - info->flags |= IEEE80211_TX_CTL_NO_PS_BUFFER; | ||
2259 | - return TX_CONTINUE; | ||
2260 | - } | ||
2261 | - | ||
2262 | ps_dbg(sta->sdata, "STA %pM aid %d: PS buffer for AC %d\n", | ||
2263 | sta->sta.addr, sta->sta.aid, ac); | ||
2264 | if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER) | ||
2265 | @@ -525,9 +515,22 @@ ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx) | ||
2266 | static ieee80211_tx_result debug_noinline | ||
2267 | ieee80211_tx_h_ps_buf(struct ieee80211_tx_data *tx) | ||
2268 | { | ||
2269 | + struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb); | ||
2270 | + struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data; | ||
2271 | + | ||
2272 | if (unlikely(tx->flags & IEEE80211_TX_PS_BUFFERED)) | ||
2273 | return TX_CONTINUE; | ||
2274 | |||
2275 | + /* only deauth, disassoc and action are bufferable MMPDUs */ | ||
2276 | + if (ieee80211_is_mgmt(hdr->frame_control) && | ||
2277 | + !ieee80211_is_deauth(hdr->frame_control) && | ||
2278 | + !ieee80211_is_disassoc(hdr->frame_control) && | ||
2279 | + !ieee80211_is_action(hdr->frame_control)) { | ||
2280 | + if (tx->flags & IEEE80211_TX_UNICAST) | ||
2281 | + info->flags |= IEEE80211_TX_CTL_NO_PS_BUFFER; | ||
2282 | + return TX_CONTINUE; | ||
2283 | + } | ||
2284 | + | ||
2285 | if (tx->flags & IEEE80211_TX_UNICAST) | ||
2286 | return ieee80211_tx_h_unicast_ps_buf(tx); | ||
2287 | else | ||
2288 | diff --git a/net/netfilter/nf_conntrack_seqadj.c b/net/netfilter/nf_conntrack_seqadj.c | ||
2289 | index 5f9bfd060dea..17c1bcb182c6 100644 | ||
2290 | --- a/net/netfilter/nf_conntrack_seqadj.c | ||
2291 | +++ b/net/netfilter/nf_conntrack_seqadj.c | ||
2292 | @@ -41,8 +41,8 @@ int nf_ct_seqadj_set(struct nf_conn *ct, enum ip_conntrack_info ctinfo, | ||
2293 | spin_lock_bh(&ct->lock); | ||
2294 | this_way = &seqadj->seq[dir]; | ||
2295 | if (this_way->offset_before == this_way->offset_after || | ||
2296 | - before(this_way->correction_pos, seq)) { | ||
2297 | - this_way->correction_pos = seq; | ||
2298 | + before(this_way->correction_pos, ntohl(seq))) { | ||
2299 | + this_way->correction_pos = ntohl(seq); | ||
2300 | this_way->offset_before = this_way->offset_after; | ||
2301 | this_way->offset_after += off; | ||
2302 | } | ||
2303 | diff --git a/net/netfilter/nf_nat_irc.c b/net/netfilter/nf_nat_irc.c | ||
2304 | index f02b3605823e..1fb2258c3535 100644 | ||
2305 | --- a/net/netfilter/nf_nat_irc.c | ||
2306 | +++ b/net/netfilter/nf_nat_irc.c | ||
2307 | @@ -34,10 +34,14 @@ static unsigned int help(struct sk_buff *skb, | ||
2308 | struct nf_conntrack_expect *exp) | ||
2309 | { | ||
2310 | char buffer[sizeof("4294967296 65635")]; | ||
2311 | + struct nf_conn *ct = exp->master; | ||
2312 | + union nf_inet_addr newaddr; | ||
2313 | u_int16_t port; | ||
2314 | unsigned int ret; | ||
2315 | |||
2316 | /* Reply comes from server. */ | ||
2317 | + newaddr = ct->tuplehash[IP_CT_DIR_REPLY].tuple.dst.u3; | ||
2318 | + | ||
2319 | exp->saved_proto.tcp.port = exp->tuple.dst.u.tcp.port; | ||
2320 | exp->dir = IP_CT_DIR_REPLY; | ||
2321 | exp->expectfn = nf_nat_follow_master; | ||
2322 | @@ -57,17 +61,35 @@ static unsigned int help(struct sk_buff *skb, | ||
2323 | } | ||
2324 | |||
2325 | if (port == 0) { | ||
2326 | - nf_ct_helper_log(skb, exp->master, "all ports in use"); | ||
2327 | + nf_ct_helper_log(skb, ct, "all ports in use"); | ||
2328 | return NF_DROP; | ||
2329 | } | ||
2330 | |||
2331 | - ret = nf_nat_mangle_tcp_packet(skb, exp->master, ctinfo, | ||
2332 | - protoff, matchoff, matchlen, buffer, | ||
2333 | - strlen(buffer)); | ||
2334 | + /* strlen("\1DCC CHAT chat AAAAAAAA P\1\n")=27 | ||
2335 | + * strlen("\1DCC SCHAT chat AAAAAAAA P\1\n")=28 | ||
2336 | + * strlen("\1DCC SEND F AAAAAAAA P S\1\n")=26 | ||
2337 | + * strlen("\1DCC MOVE F AAAAAAAA P S\1\n")=26 | ||
2338 | + * strlen("\1DCC TSEND F AAAAAAAA P S\1\n")=27 | ||
2339 | + * | ||
2340 | + * AAAAAAAAA: bound addr (1.0.0.0==16777216, min 8 digits, | ||
2341 | + * 255.255.255.255==4294967296, 10 digits) | ||
2342 | + * P: bound port (min 1 d, max 5d (65635)) | ||
2343 | + * F: filename (min 1 d ) | ||
2344 | + * S: size (min 1 d ) | ||
2345 | + * 0x01, \n: terminators | ||
2346 | + */ | ||
2347 | + /* AAA = "us", ie. where server normally talks to. */ | ||
2348 | + snprintf(buffer, sizeof(buffer), "%u %u", ntohl(newaddr.ip), port); | ||
2349 | + pr_debug("nf_nat_irc: inserting '%s' == %pI4, port %u\n", | ||
2350 | + buffer, &newaddr.ip, port); | ||
2351 | + | ||
2352 | + ret = nf_nat_mangle_tcp_packet(skb, ct, ctinfo, protoff, matchoff, | ||
2353 | + matchlen, buffer, strlen(buffer)); | ||
2354 | if (ret != NF_ACCEPT) { | ||
2355 | - nf_ct_helper_log(skb, exp->master, "cannot mangle packet"); | ||
2356 | + nf_ct_helper_log(skb, ct, "cannot mangle packet"); | ||
2357 | nf_ct_unexpect_related(exp); | ||
2358 | } | ||
2359 | + | ||
2360 | return ret; | ||
2361 | } | ||
2362 | |||
2363 | diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c | ||
2364 | index ba2548bd85bf..88cfbc189558 100644 | ||
2365 | --- a/net/packet/af_packet.c | ||
2366 | +++ b/net/packet/af_packet.c | ||
2367 | @@ -237,6 +237,30 @@ struct packet_skb_cb { | ||
2368 | static void __fanout_unlink(struct sock *sk, struct packet_sock *po); | ||
2369 | static void __fanout_link(struct sock *sk, struct packet_sock *po); | ||
2370 | |||
2371 | +static struct net_device *packet_cached_dev_get(struct packet_sock *po) | ||
2372 | +{ | ||
2373 | + struct net_device *dev; | ||
2374 | + | ||
2375 | + rcu_read_lock(); | ||
2376 | + dev = rcu_dereference(po->cached_dev); | ||
2377 | + if (likely(dev)) | ||
2378 | + dev_hold(dev); | ||
2379 | + rcu_read_unlock(); | ||
2380 | + | ||
2381 | + return dev; | ||
2382 | +} | ||
2383 | + | ||
2384 | +static void packet_cached_dev_assign(struct packet_sock *po, | ||
2385 | + struct net_device *dev) | ||
2386 | +{ | ||
2387 | + rcu_assign_pointer(po->cached_dev, dev); | ||
2388 | +} | ||
2389 | + | ||
2390 | +static void packet_cached_dev_reset(struct packet_sock *po) | ||
2391 | +{ | ||
2392 | + RCU_INIT_POINTER(po->cached_dev, NULL); | ||
2393 | +} | ||
2394 | + | ||
2395 | /* register_prot_hook must be invoked with the po->bind_lock held, | ||
2396 | * or from a context in which asynchronous accesses to the packet | ||
2397 | * socket is not possible (packet_create()). | ||
2398 | @@ -246,12 +270,10 @@ static void register_prot_hook(struct sock *sk) | ||
2399 | struct packet_sock *po = pkt_sk(sk); | ||
2400 | |||
2401 | if (!po->running) { | ||
2402 | - if (po->fanout) { | ||
2403 | + if (po->fanout) | ||
2404 | __fanout_link(sk, po); | ||
2405 | - } else { | ||
2406 | + else | ||
2407 | dev_add_pack(&po->prot_hook); | ||
2408 | - rcu_assign_pointer(po->cached_dev, po->prot_hook.dev); | ||
2409 | - } | ||
2410 | |||
2411 | sock_hold(sk); | ||
2412 | po->running = 1; | ||
2413 | @@ -270,12 +292,11 @@ static void __unregister_prot_hook(struct sock *sk, bool sync) | ||
2414 | struct packet_sock *po = pkt_sk(sk); | ||
2415 | |||
2416 | po->running = 0; | ||
2417 | - if (po->fanout) { | ||
2418 | + | ||
2419 | + if (po->fanout) | ||
2420 | __fanout_unlink(sk, po); | ||
2421 | - } else { | ||
2422 | + else | ||
2423 | __dev_remove_pack(&po->prot_hook); | ||
2424 | - RCU_INIT_POINTER(po->cached_dev, NULL); | ||
2425 | - } | ||
2426 | |||
2427 | __sock_put(sk); | ||
2428 | |||
2429 | @@ -2059,19 +2080,6 @@ static int tpacket_fill_skb(struct packet_sock *po, struct sk_buff *skb, | ||
2430 | return tp_len; | ||
2431 | } | ||
2432 | |||
2433 | -static struct net_device *packet_cached_dev_get(struct packet_sock *po) | ||
2434 | -{ | ||
2435 | - struct net_device *dev; | ||
2436 | - | ||
2437 | - rcu_read_lock(); | ||
2438 | - dev = rcu_dereference(po->cached_dev); | ||
2439 | - if (dev) | ||
2440 | - dev_hold(dev); | ||
2441 | - rcu_read_unlock(); | ||
2442 | - | ||
2443 | - return dev; | ||
2444 | -} | ||
2445 | - | ||
2446 | static int tpacket_snd(struct packet_sock *po, struct msghdr *msg) | ||
2447 | { | ||
2448 | struct sk_buff *skb; | ||
2449 | @@ -2088,7 +2096,7 @@ static int tpacket_snd(struct packet_sock *po, struct msghdr *msg) | ||
2450 | |||
2451 | mutex_lock(&po->pg_vec_lock); | ||
2452 | |||
2453 | - if (saddr == NULL) { | ||
2454 | + if (likely(saddr == NULL)) { | ||
2455 | dev = packet_cached_dev_get(po); | ||
2456 | proto = po->num; | ||
2457 | addr = NULL; | ||
2458 | @@ -2242,7 +2250,7 @@ static int packet_snd(struct socket *sock, | ||
2459 | * Get and verify the address. | ||
2460 | */ | ||
2461 | |||
2462 | - if (saddr == NULL) { | ||
2463 | + if (likely(saddr == NULL)) { | ||
2464 | dev = packet_cached_dev_get(po); | ||
2465 | proto = po->num; | ||
2466 | addr = NULL; | ||
2467 | @@ -2451,6 +2459,8 @@ static int packet_release(struct socket *sock) | ||
2468 | |||
2469 | spin_lock(&po->bind_lock); | ||
2470 | unregister_prot_hook(sk, false); | ||
2471 | + packet_cached_dev_reset(po); | ||
2472 | + | ||
2473 | if (po->prot_hook.dev) { | ||
2474 | dev_put(po->prot_hook.dev); | ||
2475 | po->prot_hook.dev = NULL; | ||
2476 | @@ -2506,14 +2516,17 @@ static int packet_do_bind(struct sock *sk, struct net_device *dev, __be16 protoc | ||
2477 | |||
2478 | spin_lock(&po->bind_lock); | ||
2479 | unregister_prot_hook(sk, true); | ||
2480 | + | ||
2481 | po->num = protocol; | ||
2482 | po->prot_hook.type = protocol; | ||
2483 | if (po->prot_hook.dev) | ||
2484 | dev_put(po->prot_hook.dev); | ||
2485 | - po->prot_hook.dev = dev; | ||
2486 | |||
2487 | + po->prot_hook.dev = dev; | ||
2488 | po->ifindex = dev ? dev->ifindex : 0; | ||
2489 | |||
2490 | + packet_cached_dev_assign(po, dev); | ||
2491 | + | ||
2492 | if (protocol == 0) | ||
2493 | goto out_unlock; | ||
2494 | |||
2495 | @@ -2626,7 +2639,8 @@ static int packet_create(struct net *net, struct socket *sock, int protocol, | ||
2496 | po = pkt_sk(sk); | ||
2497 | sk->sk_family = PF_PACKET; | ||
2498 | po->num = proto; | ||
2499 | - RCU_INIT_POINTER(po->cached_dev, NULL); | ||
2500 | + | ||
2501 | + packet_cached_dev_reset(po); | ||
2502 | |||
2503 | sk->sk_destruct = packet_sock_destruct; | ||
2504 | sk_refcnt_debug_inc(sk); | ||
2505 | @@ -3337,6 +3351,7 @@ static int packet_notifier(struct notifier_block *this, | ||
2506 | sk->sk_error_report(sk); | ||
2507 | } | ||
2508 | if (msg == NETDEV_UNREGISTER) { | ||
2509 | + packet_cached_dev_reset(po); | ||
2510 | po->ifindex = -1; | ||
2511 | if (po->prot_hook.dev) | ||
2512 | dev_put(po->prot_hook.dev); | ||
2513 | diff --git a/net/rds/ib.c b/net/rds/ib.c | ||
2514 | index b4c8b0022fee..ba2dffeff608 100644 | ||
2515 | --- a/net/rds/ib.c | ||
2516 | +++ b/net/rds/ib.c | ||
2517 | @@ -338,7 +338,8 @@ static int rds_ib_laddr_check(__be32 addr) | ||
2518 | ret = rdma_bind_addr(cm_id, (struct sockaddr *)&sin); | ||
2519 | /* due to this, we will claim to support iWARP devices unless we | ||
2520 | check node_type. */ | ||
2521 | - if (ret || cm_id->device->node_type != RDMA_NODE_IB_CA) | ||
2522 | + if (ret || !cm_id->device || | ||
2523 | + cm_id->device->node_type != RDMA_NODE_IB_CA) | ||
2524 | ret = -EADDRNOTAVAIL; | ||
2525 | |||
2526 | rdsdebug("addr %pI4 ret %d node type %d\n", | ||
2527 | diff --git a/net/rds/ib_send.c b/net/rds/ib_send.c | ||
2528 | index e59094981175..37be6e226d1b 100644 | ||
2529 | --- a/net/rds/ib_send.c | ||
2530 | +++ b/net/rds/ib_send.c | ||
2531 | @@ -552,9 +552,8 @@ int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm, | ||
2532 | && rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) { | ||
2533 | rds_cong_map_updated(conn->c_fcong, ~(u64) 0); | ||
2534 | scat = &rm->data.op_sg[sg]; | ||
2535 | - ret = sizeof(struct rds_header) + RDS_CONG_MAP_BYTES; | ||
2536 | - ret = min_t(int, ret, scat->length - conn->c_xmit_data_off); | ||
2537 | - return ret; | ||
2538 | + ret = max_t(int, RDS_CONG_MAP_BYTES, scat->length); | ||
2539 | + return sizeof(struct rds_header) + ret; | ||
2540 | } | ||
2541 | |||
2542 | /* FIXME we may overallocate here */ | ||
2543 | diff --git a/net/rose/af_rose.c b/net/rose/af_rose.c | ||
2544 | index 33af77246bfe..62ced6516c58 100644 | ||
2545 | --- a/net/rose/af_rose.c | ||
2546 | +++ b/net/rose/af_rose.c | ||
2547 | @@ -1253,6 +1253,7 @@ static int rose_recvmsg(struct kiocb *iocb, struct socket *sock, | ||
2548 | |||
2549 | if (msg->msg_name) { | ||
2550 | struct sockaddr_rose *srose; | ||
2551 | + struct full_sockaddr_rose *full_srose = msg->msg_name; | ||
2552 | |||
2553 | memset(msg->msg_name, 0, sizeof(struct full_sockaddr_rose)); | ||
2554 | srose = msg->msg_name; | ||
2555 | @@ -1260,18 +1261,9 @@ static int rose_recvmsg(struct kiocb *iocb, struct socket *sock, | ||
2556 | srose->srose_addr = rose->dest_addr; | ||
2557 | srose->srose_call = rose->dest_call; | ||
2558 | srose->srose_ndigis = rose->dest_ndigis; | ||
2559 | - if (msg->msg_namelen >= sizeof(struct full_sockaddr_rose)) { | ||
2560 | - struct full_sockaddr_rose *full_srose = (struct full_sockaddr_rose *)msg->msg_name; | ||
2561 | - for (n = 0 ; n < rose->dest_ndigis ; n++) | ||
2562 | - full_srose->srose_digis[n] = rose->dest_digis[n]; | ||
2563 | - msg->msg_namelen = sizeof(struct full_sockaddr_rose); | ||
2564 | - } else { | ||
2565 | - if (rose->dest_ndigis >= 1) { | ||
2566 | - srose->srose_ndigis = 1; | ||
2567 | - srose->srose_digi = rose->dest_digis[0]; | ||
2568 | - } | ||
2569 | - msg->msg_namelen = sizeof(struct sockaddr_rose); | ||
2570 | - } | ||
2571 | + for (n = 0 ; n < rose->dest_ndigis ; n++) | ||
2572 | + full_srose->srose_digis[n] = rose->dest_digis[n]; | ||
2573 | + msg->msg_namelen = sizeof(struct full_sockaddr_rose); | ||
2574 | } | ||
2575 | |||
2576 | skb_free_datagram(sk, skb); | ||
2577 | diff --git a/net/unix/af_unix.c b/net/unix/af_unix.c | ||
2578 | index 01625ccc3ae6..a427623ee574 100644 | ||
2579 | --- a/net/unix/af_unix.c | ||
2580 | +++ b/net/unix/af_unix.c | ||
2581 | @@ -530,13 +530,17 @@ static int unix_seqpacket_sendmsg(struct kiocb *, struct socket *, | ||
2582 | static int unix_seqpacket_recvmsg(struct kiocb *, struct socket *, | ||
2583 | struct msghdr *, size_t, int); | ||
2584 | |||
2585 | -static void unix_set_peek_off(struct sock *sk, int val) | ||
2586 | +static int unix_set_peek_off(struct sock *sk, int val) | ||
2587 | { | ||
2588 | struct unix_sock *u = unix_sk(sk); | ||
2589 | |||
2590 | - mutex_lock(&u->readlock); | ||
2591 | + if (mutex_lock_interruptible(&u->readlock)) | ||
2592 | + return -EINTR; | ||
2593 | + | ||
2594 | sk->sk_peek_off = val; | ||
2595 | mutex_unlock(&u->readlock); | ||
2596 | + | ||
2597 | + return 0; | ||
2598 | } | ||
2599 | |||
2600 | |||
2601 | @@ -714,7 +718,9 @@ static int unix_autobind(struct socket *sock) | ||
2602 | int err; | ||
2603 | unsigned int retries = 0; | ||
2604 | |||
2605 | - mutex_lock(&u->readlock); | ||
2606 | + err = mutex_lock_interruptible(&u->readlock); | ||
2607 | + if (err) | ||
2608 | + return err; | ||
2609 | |||
2610 | err = 0; | ||
2611 | if (u->addr) | ||
2612 | @@ -873,7 +879,9 @@ static int unix_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | ||
2613 | goto out; | ||
2614 | addr_len = err; | ||
2615 | |||
2616 | - mutex_lock(&u->readlock); | ||
2617 | + err = mutex_lock_interruptible(&u->readlock); | ||
2618 | + if (err) | ||
2619 | + goto out; | ||
2620 | |||
2621 | err = -EINVAL; | ||
2622 | if (u->addr) |