Contents of /trunk/kernel-alx/patches-4.4/0145-4.4.46-all-fixes.patch
Parent Directory | Revision Log
Revision 2881 -
(show annotations)
(download)
Mon Mar 27 13:49:20 2017 UTC (7 years, 6 months ago) by niro
File size: 28783 byte(s)
Mon Mar 27 13:49:20 2017 UTC (7 years, 6 months ago) by niro
File size: 28783 byte(s)
linux-4.4.46
1 | diff --git a/Makefile b/Makefile |
2 | index a3dfc73da722..2dd5cb2fe182 100644 |
3 | --- a/Makefile |
4 | +++ b/Makefile |
5 | @@ -1,6 +1,6 @@ |
6 | VERSION = 4 |
7 | PATCHLEVEL = 4 |
8 | -SUBLEVEL = 45 |
9 | +SUBLEVEL = 46 |
10 | EXTRAVERSION = |
11 | NAME = Blurry Fish Butt |
12 | |
13 | diff --git a/arch/arc/include/asm/delay.h b/arch/arc/include/asm/delay.h |
14 | index a36e8601114d..d5da2115d78a 100644 |
15 | --- a/arch/arc/include/asm/delay.h |
16 | +++ b/arch/arc/include/asm/delay.h |
17 | @@ -26,7 +26,9 @@ static inline void __delay(unsigned long loops) |
18 | " lp 1f \n" |
19 | " nop \n" |
20 | "1: \n" |
21 | - : : "r"(loops)); |
22 | + : |
23 | + : "r"(loops) |
24 | + : "lp_count"); |
25 | } |
26 | |
27 | extern void __bad_udelay(void); |
28 | diff --git a/arch/arc/kernel/unaligned.c b/arch/arc/kernel/unaligned.c |
29 | index abd961f3e763..91ebe382147f 100644 |
30 | --- a/arch/arc/kernel/unaligned.c |
31 | +++ b/arch/arc/kernel/unaligned.c |
32 | @@ -241,8 +241,9 @@ int misaligned_fixup(unsigned long address, struct pt_regs *regs, |
33 | if (state.fault) |
34 | goto fault; |
35 | |
36 | + /* clear any remanants of delay slot */ |
37 | if (delay_mode(regs)) { |
38 | - regs->ret = regs->bta; |
39 | + regs->ret = regs->bta ~1U; |
40 | regs->status32 &= ~STATUS_DE_MASK; |
41 | } else { |
42 | regs->ret += state.instr_len; |
43 | diff --git a/arch/parisc/include/asm/bitops.h b/arch/parisc/include/asm/bitops.h |
44 | index 3f9406d9b9d6..da87943328a5 100644 |
45 | --- a/arch/parisc/include/asm/bitops.h |
46 | +++ b/arch/parisc/include/asm/bitops.h |
47 | @@ -6,7 +6,7 @@ |
48 | #endif |
49 | |
50 | #include <linux/compiler.h> |
51 | -#include <asm/types.h> /* for BITS_PER_LONG/SHIFT_PER_LONG */ |
52 | +#include <asm/types.h> |
53 | #include <asm/byteorder.h> |
54 | #include <asm/barrier.h> |
55 | #include <linux/atomic.h> |
56 | @@ -17,6 +17,12 @@ |
57 | * to include/asm-i386/bitops.h or kerneldoc |
58 | */ |
59 | |
60 | +#if __BITS_PER_LONG == 64 |
61 | +#define SHIFT_PER_LONG 6 |
62 | +#else |
63 | +#define SHIFT_PER_LONG 5 |
64 | +#endif |
65 | + |
66 | #define CHOP_SHIFTCOUNT(x) (((unsigned long) (x)) & (BITS_PER_LONG - 1)) |
67 | |
68 | |
69 | diff --git a/arch/parisc/include/uapi/asm/bitsperlong.h b/arch/parisc/include/uapi/asm/bitsperlong.h |
70 | index e0a23c7bdd43..07fa7e50bdc0 100644 |
71 | --- a/arch/parisc/include/uapi/asm/bitsperlong.h |
72 | +++ b/arch/parisc/include/uapi/asm/bitsperlong.h |
73 | @@ -3,10 +3,8 @@ |
74 | |
75 | #if defined(__LP64__) |
76 | #define __BITS_PER_LONG 64 |
77 | -#define SHIFT_PER_LONG 6 |
78 | #else |
79 | #define __BITS_PER_LONG 32 |
80 | -#define SHIFT_PER_LONG 5 |
81 | #endif |
82 | |
83 | #include <asm-generic/bitsperlong.h> |
84 | diff --git a/arch/parisc/include/uapi/asm/swab.h b/arch/parisc/include/uapi/asm/swab.h |
85 | index e78403b129ef..928e1bbac98f 100644 |
86 | --- a/arch/parisc/include/uapi/asm/swab.h |
87 | +++ b/arch/parisc/include/uapi/asm/swab.h |
88 | @@ -1,6 +1,7 @@ |
89 | #ifndef _PARISC_SWAB_H |
90 | #define _PARISC_SWAB_H |
91 | |
92 | +#include <asm/bitsperlong.h> |
93 | #include <linux/types.h> |
94 | #include <linux/compiler.h> |
95 | |
96 | @@ -38,7 +39,7 @@ static inline __attribute_const__ __u32 __arch_swab32(__u32 x) |
97 | } |
98 | #define __arch_swab32 __arch_swab32 |
99 | |
100 | -#if BITS_PER_LONG > 32 |
101 | +#if __BITS_PER_LONG > 32 |
102 | /* |
103 | ** From "PA-RISC 2.0 Architecture", HP Professional Books. |
104 | ** See Appendix I page 8 , "Endian Byte Swapping". |
105 | @@ -61,6 +62,6 @@ static inline __attribute_const__ __u64 __arch_swab64(__u64 x) |
106 | return x; |
107 | } |
108 | #define __arch_swab64 __arch_swab64 |
109 | -#endif /* BITS_PER_LONG > 32 */ |
110 | +#endif /* __BITS_PER_LONG > 32 */ |
111 | |
112 | #endif /* _PARISC_SWAB_H */ |
113 | diff --git a/arch/s390/kernel/ptrace.c b/arch/s390/kernel/ptrace.c |
114 | index 01c37b36caf9..02bd587b610b 100644 |
115 | --- a/arch/s390/kernel/ptrace.c |
116 | +++ b/arch/s390/kernel/ptrace.c |
117 | @@ -963,6 +963,11 @@ static int s390_fpregs_set(struct task_struct *target, |
118 | if (target == current) |
119 | save_fpu_regs(); |
120 | |
121 | + if (MACHINE_HAS_VX) |
122 | + convert_vx_to_fp(fprs, target->thread.fpu.vxrs); |
123 | + else |
124 | + memcpy(&fprs, target->thread.fpu.fprs, sizeof(fprs)); |
125 | + |
126 | /* If setting FPC, must validate it first. */ |
127 | if (count > 0 && pos < offsetof(s390_fp_regs, fprs)) { |
128 | u32 ufpc[2] = { target->thread.fpu.fpc, 0 }; |
129 | @@ -1067,6 +1072,9 @@ static int s390_vxrs_low_set(struct task_struct *target, |
130 | if (target == current) |
131 | save_fpu_regs(); |
132 | |
133 | + for (i = 0; i < __NUM_VXRS_LOW; i++) |
134 | + vxrs[i] = *((__u64 *)(target->thread.fpu.vxrs + i) + 1); |
135 | + |
136 | rc = user_regset_copyin(&pos, &count, &kbuf, &ubuf, vxrs, 0, -1); |
137 | if (rc == 0) |
138 | for (i = 0; i < __NUM_VXRS_LOW; i++) |
139 | diff --git a/arch/tile/kernel/ptrace.c b/arch/tile/kernel/ptrace.c |
140 | index bdc126faf741..6239aa155f6d 100644 |
141 | --- a/arch/tile/kernel/ptrace.c |
142 | +++ b/arch/tile/kernel/ptrace.c |
143 | @@ -111,7 +111,7 @@ static int tile_gpr_set(struct task_struct *target, |
144 | const void *kbuf, const void __user *ubuf) |
145 | { |
146 | int ret; |
147 | - struct pt_regs regs; |
148 | + struct pt_regs regs = *task_pt_regs(target); |
149 | |
150 | ret = user_regset_copyin(&pos, &count, &kbuf, &ubuf, ®s, 0, |
151 | sizeof(regs)); |
152 | diff --git a/drivers/gpu/drm/drm_modes.c b/drivers/gpu/drm/drm_modes.c |
153 | index 39e30abddf08..71a10f08522e 100644 |
154 | --- a/drivers/gpu/drm/drm_modes.c |
155 | +++ b/drivers/gpu/drm/drm_modes.c |
156 | @@ -1401,6 +1401,13 @@ drm_mode_create_from_cmdline_mode(struct drm_device *dev, |
157 | return NULL; |
158 | |
159 | mode->type |= DRM_MODE_TYPE_USERDEF; |
160 | + /* fix up 1368x768: GFT/CVT can't express 1366 width due to alignment */ |
161 | + if (cmd->xres == 1366 && mode->hdisplay == 1368) { |
162 | + mode->hdisplay = 1366; |
163 | + mode->hsync_start--; |
164 | + mode->hsync_end--; |
165 | + drm_mode_set_name(mode); |
166 | + } |
167 | drm_mode_set_crtcinfo(mode, CRTC_INTERLACE_HALVE_V); |
168 | return mode; |
169 | } |
170 | diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c |
171 | index 97d1ed20418b..63fea6a2869c 100644 |
172 | --- a/drivers/gpu/drm/i915/intel_crt.c |
173 | +++ b/drivers/gpu/drm/i915/intel_crt.c |
174 | @@ -445,6 +445,7 @@ static bool intel_crt_detect_ddc(struct drm_connector *connector) |
175 | struct drm_i915_private *dev_priv = crt->base.base.dev->dev_private; |
176 | struct edid *edid; |
177 | struct i2c_adapter *i2c; |
178 | + bool ret = false; |
179 | |
180 | BUG_ON(crt->base.type != INTEL_OUTPUT_ANALOG); |
181 | |
182 | @@ -461,17 +462,17 @@ static bool intel_crt_detect_ddc(struct drm_connector *connector) |
183 | */ |
184 | if (!is_digital) { |
185 | DRM_DEBUG_KMS("CRT detected via DDC:0x50 [EDID]\n"); |
186 | - return true; |
187 | + ret = true; |
188 | + } else { |
189 | + DRM_DEBUG_KMS("CRT not detected via DDC:0x50 [EDID reports a digital panel]\n"); |
190 | } |
191 | - |
192 | - DRM_DEBUG_KMS("CRT not detected via DDC:0x50 [EDID reports a digital panel]\n"); |
193 | } else { |
194 | DRM_DEBUG_KMS("CRT not detected via DDC:0x50 [no valid EDID found]\n"); |
195 | } |
196 | |
197 | kfree(edid); |
198 | |
199 | - return false; |
200 | + return ret; |
201 | } |
202 | |
203 | static enum drm_connector_status |
204 | diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c |
205 | index 17a15c56028c..c9dcad6a53bf 100644 |
206 | --- a/drivers/infiniband/core/cma.c |
207 | +++ b/drivers/infiniband/core/cma.c |
208 | @@ -2578,7 +2578,8 @@ static int cma_bind_addr(struct rdma_cm_id *id, struct sockaddr *src_addr, |
209 | if (!src_addr || !src_addr->sa_family) { |
210 | src_addr = (struct sockaddr *) &id->route.addr.src_addr; |
211 | src_addr->sa_family = dst_addr->sa_family; |
212 | - if (dst_addr->sa_family == AF_INET6) { |
213 | + if (IS_ENABLED(CONFIG_IPV6) && |
214 | + dst_addr->sa_family == AF_INET6) { |
215 | struct sockaddr_in6 *src_addr6 = (struct sockaddr_in6 *) src_addr; |
216 | struct sockaddr_in6 *dst_addr6 = (struct sockaddr_in6 *) dst_addr; |
217 | src_addr6->sin6_scope_id = dst_addr6->sin6_scope_id; |
218 | diff --git a/drivers/infiniband/core/umem.c b/drivers/infiniband/core/umem.c |
219 | index 04f3c0db9126..0ae337bec4f2 100644 |
220 | --- a/drivers/infiniband/core/umem.c |
221 | +++ b/drivers/infiniband/core/umem.c |
222 | @@ -134,6 +134,7 @@ struct ib_umem *ib_umem_get(struct ib_ucontext *context, unsigned long addr, |
223 | IB_ACCESS_REMOTE_ATOMIC | IB_ACCESS_MW_BIND)); |
224 | |
225 | if (access & IB_ACCESS_ON_DEMAND) { |
226 | + put_pid(umem->pid); |
227 | ret = ib_umem_odp_get(context, umem); |
228 | if (ret) { |
229 | kfree(umem); |
230 | @@ -149,6 +150,7 @@ struct ib_umem *ib_umem_get(struct ib_ucontext *context, unsigned long addr, |
231 | |
232 | page_list = (struct page **) __get_free_page(GFP_KERNEL); |
233 | if (!page_list) { |
234 | + put_pid(umem->pid); |
235 | kfree(umem); |
236 | return ERR_PTR(-ENOMEM); |
237 | } |
238 | diff --git a/drivers/infiniband/ulp/ipoib/ipoib.h b/drivers/infiniband/ulp/ipoib/ipoib.h |
239 | index 69a151ae8261..07cfcc326863 100644 |
240 | --- a/drivers/infiniband/ulp/ipoib/ipoib.h |
241 | +++ b/drivers/infiniband/ulp/ipoib/ipoib.h |
242 | @@ -63,6 +63,8 @@ enum ipoib_flush_level { |
243 | |
244 | enum { |
245 | IPOIB_ENCAP_LEN = 4, |
246 | + IPOIB_PSEUDO_LEN = 20, |
247 | + IPOIB_HARD_LEN = IPOIB_ENCAP_LEN + IPOIB_PSEUDO_LEN, |
248 | |
249 | IPOIB_UD_HEAD_SIZE = IB_GRH_BYTES + IPOIB_ENCAP_LEN, |
250 | IPOIB_UD_RX_SG = 2, /* max buffer needed for 4K mtu */ |
251 | @@ -131,15 +133,21 @@ struct ipoib_header { |
252 | u16 reserved; |
253 | }; |
254 | |
255 | -struct ipoib_cb { |
256 | - struct qdisc_skb_cb qdisc_cb; |
257 | - u8 hwaddr[INFINIBAND_ALEN]; |
258 | +struct ipoib_pseudo_header { |
259 | + u8 hwaddr[INFINIBAND_ALEN]; |
260 | }; |
261 | |
262 | -static inline struct ipoib_cb *ipoib_skb_cb(const struct sk_buff *skb) |
263 | +static inline void skb_add_pseudo_hdr(struct sk_buff *skb) |
264 | { |
265 | - BUILD_BUG_ON(sizeof(skb->cb) < sizeof(struct ipoib_cb)); |
266 | - return (struct ipoib_cb *)skb->cb; |
267 | + char *data = skb_push(skb, IPOIB_PSEUDO_LEN); |
268 | + |
269 | + /* |
270 | + * only the ipoib header is present now, make room for a dummy |
271 | + * pseudo header and set skb field accordingly |
272 | + */ |
273 | + memset(data, 0, IPOIB_PSEUDO_LEN); |
274 | + skb_reset_mac_header(skb); |
275 | + skb_pull(skb, IPOIB_HARD_LEN); |
276 | } |
277 | |
278 | /* Used for all multicast joins (broadcast, IPv4 mcast and IPv6 mcast) */ |
279 | diff --git a/drivers/infiniband/ulp/ipoib/ipoib_cm.c b/drivers/infiniband/ulp/ipoib/ipoib_cm.c |
280 | index de5e2b01ab05..3ba7de5f9379 100644 |
281 | --- a/drivers/infiniband/ulp/ipoib/ipoib_cm.c |
282 | +++ b/drivers/infiniband/ulp/ipoib/ipoib_cm.c |
283 | @@ -63,6 +63,8 @@ MODULE_PARM_DESC(cm_data_debug_level, |
284 | #define IPOIB_CM_RX_DELAY (3 * 256 * HZ) |
285 | #define IPOIB_CM_RX_UPDATE_MASK (0x3) |
286 | |
287 | +#define IPOIB_CM_RX_RESERVE (ALIGN(IPOIB_HARD_LEN, 16) - IPOIB_ENCAP_LEN) |
288 | + |
289 | static struct ib_qp_attr ipoib_cm_err_attr = { |
290 | .qp_state = IB_QPS_ERR |
291 | }; |
292 | @@ -147,15 +149,15 @@ static struct sk_buff *ipoib_cm_alloc_rx_skb(struct net_device *dev, |
293 | struct sk_buff *skb; |
294 | int i; |
295 | |
296 | - skb = dev_alloc_skb(IPOIB_CM_HEAD_SIZE + 12); |
297 | + skb = dev_alloc_skb(ALIGN(IPOIB_CM_HEAD_SIZE + IPOIB_PSEUDO_LEN, 16)); |
298 | if (unlikely(!skb)) |
299 | return NULL; |
300 | |
301 | /* |
302 | - * IPoIB adds a 4 byte header. So we need 12 more bytes to align the |
303 | + * IPoIB adds a IPOIB_ENCAP_LEN byte header, this will align the |
304 | * IP header to a multiple of 16. |
305 | */ |
306 | - skb_reserve(skb, 12); |
307 | + skb_reserve(skb, IPOIB_CM_RX_RESERVE); |
308 | |
309 | mapping[0] = ib_dma_map_single(priv->ca, skb->data, IPOIB_CM_HEAD_SIZE, |
310 | DMA_FROM_DEVICE); |
311 | @@ -624,9 +626,9 @@ void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) |
312 | if (wc->byte_len < IPOIB_CM_COPYBREAK) { |
313 | int dlen = wc->byte_len; |
314 | |
315 | - small_skb = dev_alloc_skb(dlen + 12); |
316 | + small_skb = dev_alloc_skb(dlen + IPOIB_CM_RX_RESERVE); |
317 | if (small_skb) { |
318 | - skb_reserve(small_skb, 12); |
319 | + skb_reserve(small_skb, IPOIB_CM_RX_RESERVE); |
320 | ib_dma_sync_single_for_cpu(priv->ca, rx_ring[wr_id].mapping[0], |
321 | dlen, DMA_FROM_DEVICE); |
322 | skb_copy_from_linear_data(skb, small_skb->data, dlen); |
323 | @@ -663,8 +665,7 @@ void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) |
324 | |
325 | copied: |
326 | skb->protocol = ((struct ipoib_header *) skb->data)->proto; |
327 | - skb_reset_mac_header(skb); |
328 | - skb_pull(skb, IPOIB_ENCAP_LEN); |
329 | + skb_add_pseudo_hdr(skb); |
330 | |
331 | ++dev->stats.rx_packets; |
332 | dev->stats.rx_bytes += skb->len; |
333 | diff --git a/drivers/infiniband/ulp/ipoib/ipoib_ib.c b/drivers/infiniband/ulp/ipoib/ipoib_ib.c |
334 | index 85de078fb0ce..8f8c3af9f4e8 100644 |
335 | --- a/drivers/infiniband/ulp/ipoib/ipoib_ib.c |
336 | +++ b/drivers/infiniband/ulp/ipoib/ipoib_ib.c |
337 | @@ -130,16 +130,15 @@ static struct sk_buff *ipoib_alloc_rx_skb(struct net_device *dev, int id) |
338 | |
339 | buf_size = IPOIB_UD_BUF_SIZE(priv->max_ib_mtu); |
340 | |
341 | - skb = dev_alloc_skb(buf_size + IPOIB_ENCAP_LEN); |
342 | + skb = dev_alloc_skb(buf_size + IPOIB_HARD_LEN); |
343 | if (unlikely(!skb)) |
344 | return NULL; |
345 | |
346 | /* |
347 | - * IB will leave a 40 byte gap for a GRH and IPoIB adds a 4 byte |
348 | - * header. So we need 4 more bytes to get to 48 and align the |
349 | - * IP header to a multiple of 16. |
350 | + * the IP header will be at IPOIP_HARD_LEN + IB_GRH_BYTES, that is |
351 | + * 64 bytes aligned |
352 | */ |
353 | - skb_reserve(skb, 4); |
354 | + skb_reserve(skb, sizeof(struct ipoib_pseudo_header)); |
355 | |
356 | mapping = priv->rx_ring[id].mapping; |
357 | mapping[0] = ib_dma_map_single(priv->ca, skb->data, buf_size, |
358 | @@ -242,8 +241,7 @@ static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) |
359 | skb_pull(skb, IB_GRH_BYTES); |
360 | |
361 | skb->protocol = ((struct ipoib_header *) skb->data)->proto; |
362 | - skb_reset_mac_header(skb); |
363 | - skb_pull(skb, IPOIB_ENCAP_LEN); |
364 | + skb_add_pseudo_hdr(skb); |
365 | |
366 | ++dev->stats.rx_packets; |
367 | dev->stats.rx_bytes += skb->len; |
368 | diff --git a/drivers/infiniband/ulp/ipoib/ipoib_main.c b/drivers/infiniband/ulp/ipoib/ipoib_main.c |
369 | index 5f7681b975d0..8a4d10452d61 100644 |
370 | --- a/drivers/infiniband/ulp/ipoib/ipoib_main.c |
371 | +++ b/drivers/infiniband/ulp/ipoib/ipoib_main.c |
372 | @@ -850,9 +850,12 @@ static void neigh_add_path(struct sk_buff *skb, u8 *daddr, |
373 | ipoib_neigh_free(neigh); |
374 | goto err_drop; |
375 | } |
376 | - if (skb_queue_len(&neigh->queue) < IPOIB_MAX_PATH_REC_QUEUE) |
377 | + if (skb_queue_len(&neigh->queue) < |
378 | + IPOIB_MAX_PATH_REC_QUEUE) { |
379 | + /* put pseudoheader back on for next time */ |
380 | + skb_push(skb, IPOIB_PSEUDO_LEN); |
381 | __skb_queue_tail(&neigh->queue, skb); |
382 | - else { |
383 | + } else { |
384 | ipoib_warn(priv, "queue length limit %d. Packet drop.\n", |
385 | skb_queue_len(&neigh->queue)); |
386 | goto err_drop; |
387 | @@ -889,7 +892,7 @@ err_drop: |
388 | } |
389 | |
390 | static void unicast_arp_send(struct sk_buff *skb, struct net_device *dev, |
391 | - struct ipoib_cb *cb) |
392 | + struct ipoib_pseudo_header *phdr) |
393 | { |
394 | struct ipoib_dev_priv *priv = netdev_priv(dev); |
395 | struct ipoib_path *path; |
396 | @@ -897,16 +900,18 @@ static void unicast_arp_send(struct sk_buff *skb, struct net_device *dev, |
397 | |
398 | spin_lock_irqsave(&priv->lock, flags); |
399 | |
400 | - path = __path_find(dev, cb->hwaddr + 4); |
401 | + path = __path_find(dev, phdr->hwaddr + 4); |
402 | if (!path || !path->valid) { |
403 | int new_path = 0; |
404 | |
405 | if (!path) { |
406 | - path = path_rec_create(dev, cb->hwaddr + 4); |
407 | + path = path_rec_create(dev, phdr->hwaddr + 4); |
408 | new_path = 1; |
409 | } |
410 | if (path) { |
411 | if (skb_queue_len(&path->queue) < IPOIB_MAX_PATH_REC_QUEUE) { |
412 | + /* put pseudoheader back on for next time */ |
413 | + skb_push(skb, IPOIB_PSEUDO_LEN); |
414 | __skb_queue_tail(&path->queue, skb); |
415 | } else { |
416 | ++dev->stats.tx_dropped; |
417 | @@ -934,10 +939,12 @@ static void unicast_arp_send(struct sk_buff *skb, struct net_device *dev, |
418 | be16_to_cpu(path->pathrec.dlid)); |
419 | |
420 | spin_unlock_irqrestore(&priv->lock, flags); |
421 | - ipoib_send(dev, skb, path->ah, IPOIB_QPN(cb->hwaddr)); |
422 | + ipoib_send(dev, skb, path->ah, IPOIB_QPN(phdr->hwaddr)); |
423 | return; |
424 | } else if ((path->query || !path_rec_start(dev, path)) && |
425 | skb_queue_len(&path->queue) < IPOIB_MAX_PATH_REC_QUEUE) { |
426 | + /* put pseudoheader back on for next time */ |
427 | + skb_push(skb, IPOIB_PSEUDO_LEN); |
428 | __skb_queue_tail(&path->queue, skb); |
429 | } else { |
430 | ++dev->stats.tx_dropped; |
431 | @@ -951,13 +958,15 @@ static int ipoib_start_xmit(struct sk_buff *skb, struct net_device *dev) |
432 | { |
433 | struct ipoib_dev_priv *priv = netdev_priv(dev); |
434 | struct ipoib_neigh *neigh; |
435 | - struct ipoib_cb *cb = ipoib_skb_cb(skb); |
436 | + struct ipoib_pseudo_header *phdr; |
437 | struct ipoib_header *header; |
438 | unsigned long flags; |
439 | |
440 | + phdr = (struct ipoib_pseudo_header *) skb->data; |
441 | + skb_pull(skb, sizeof(*phdr)); |
442 | header = (struct ipoib_header *) skb->data; |
443 | |
444 | - if (unlikely(cb->hwaddr[4] == 0xff)) { |
445 | + if (unlikely(phdr->hwaddr[4] == 0xff)) { |
446 | /* multicast, arrange "if" according to probability */ |
447 | if ((header->proto != htons(ETH_P_IP)) && |
448 | (header->proto != htons(ETH_P_IPV6)) && |
449 | @@ -970,13 +979,13 @@ static int ipoib_start_xmit(struct sk_buff *skb, struct net_device *dev) |
450 | return NETDEV_TX_OK; |
451 | } |
452 | /* Add in the P_Key for multicast*/ |
453 | - cb->hwaddr[8] = (priv->pkey >> 8) & 0xff; |
454 | - cb->hwaddr[9] = priv->pkey & 0xff; |
455 | + phdr->hwaddr[8] = (priv->pkey >> 8) & 0xff; |
456 | + phdr->hwaddr[9] = priv->pkey & 0xff; |
457 | |
458 | - neigh = ipoib_neigh_get(dev, cb->hwaddr); |
459 | + neigh = ipoib_neigh_get(dev, phdr->hwaddr); |
460 | if (likely(neigh)) |
461 | goto send_using_neigh; |
462 | - ipoib_mcast_send(dev, cb->hwaddr, skb); |
463 | + ipoib_mcast_send(dev, phdr->hwaddr, skb); |
464 | return NETDEV_TX_OK; |
465 | } |
466 | |
467 | @@ -985,16 +994,16 @@ static int ipoib_start_xmit(struct sk_buff *skb, struct net_device *dev) |
468 | case htons(ETH_P_IP): |
469 | case htons(ETH_P_IPV6): |
470 | case htons(ETH_P_TIPC): |
471 | - neigh = ipoib_neigh_get(dev, cb->hwaddr); |
472 | + neigh = ipoib_neigh_get(dev, phdr->hwaddr); |
473 | if (unlikely(!neigh)) { |
474 | - neigh_add_path(skb, cb->hwaddr, dev); |
475 | + neigh_add_path(skb, phdr->hwaddr, dev); |
476 | return NETDEV_TX_OK; |
477 | } |
478 | break; |
479 | case htons(ETH_P_ARP): |
480 | case htons(ETH_P_RARP): |
481 | /* for unicast ARP and RARP should always perform path find */ |
482 | - unicast_arp_send(skb, dev, cb); |
483 | + unicast_arp_send(skb, dev, phdr); |
484 | return NETDEV_TX_OK; |
485 | default: |
486 | /* ethertype not supported by IPoIB */ |
487 | @@ -1011,11 +1020,13 @@ send_using_neigh: |
488 | goto unref; |
489 | } |
490 | } else if (neigh->ah) { |
491 | - ipoib_send(dev, skb, neigh->ah, IPOIB_QPN(cb->hwaddr)); |
492 | + ipoib_send(dev, skb, neigh->ah, IPOIB_QPN(phdr->hwaddr)); |
493 | goto unref; |
494 | } |
495 | |
496 | if (skb_queue_len(&neigh->queue) < IPOIB_MAX_PATH_REC_QUEUE) { |
497 | + /* put pseudoheader back on for next time */ |
498 | + skb_push(skb, sizeof(*phdr)); |
499 | spin_lock_irqsave(&priv->lock, flags); |
500 | __skb_queue_tail(&neigh->queue, skb); |
501 | spin_unlock_irqrestore(&priv->lock, flags); |
502 | @@ -1047,8 +1058,8 @@ static int ipoib_hard_header(struct sk_buff *skb, |
503 | unsigned short type, |
504 | const void *daddr, const void *saddr, unsigned len) |
505 | { |
506 | + struct ipoib_pseudo_header *phdr; |
507 | struct ipoib_header *header; |
508 | - struct ipoib_cb *cb = ipoib_skb_cb(skb); |
509 | |
510 | header = (struct ipoib_header *) skb_push(skb, sizeof *header); |
511 | |
512 | @@ -1057,12 +1068,13 @@ static int ipoib_hard_header(struct sk_buff *skb, |
513 | |
514 | /* |
515 | * we don't rely on dst_entry structure, always stuff the |
516 | - * destination address into skb->cb so we can figure out where |
517 | + * destination address into skb hard header so we can figure out where |
518 | * to send the packet later. |
519 | */ |
520 | - memcpy(cb->hwaddr, daddr, INFINIBAND_ALEN); |
521 | + phdr = (struct ipoib_pseudo_header *) skb_push(skb, sizeof(*phdr)); |
522 | + memcpy(phdr->hwaddr, daddr, INFINIBAND_ALEN); |
523 | |
524 | - return sizeof *header; |
525 | + return IPOIB_HARD_LEN; |
526 | } |
527 | |
528 | static void ipoib_set_mcast_list(struct net_device *dev) |
529 | @@ -1638,7 +1650,7 @@ void ipoib_setup(struct net_device *dev) |
530 | |
531 | dev->flags |= IFF_BROADCAST | IFF_MULTICAST; |
532 | |
533 | - dev->hard_header_len = IPOIB_ENCAP_LEN; |
534 | + dev->hard_header_len = IPOIB_HARD_LEN; |
535 | dev->addr_len = INFINIBAND_ALEN; |
536 | dev->type = ARPHRD_INFINIBAND; |
537 | dev->tx_queue_len = ipoib_sendq_size * 2; |
538 | diff --git a/drivers/infiniband/ulp/ipoib/ipoib_multicast.c b/drivers/infiniband/ulp/ipoib/ipoib_multicast.c |
539 | index 8ec99bdea76b..5580ab0b5781 100644 |
540 | --- a/drivers/infiniband/ulp/ipoib/ipoib_multicast.c |
541 | +++ b/drivers/infiniband/ulp/ipoib/ipoib_multicast.c |
542 | @@ -756,9 +756,11 @@ void ipoib_mcast_send(struct net_device *dev, u8 *daddr, struct sk_buff *skb) |
543 | __ipoib_mcast_add(dev, mcast); |
544 | list_add_tail(&mcast->list, &priv->multicast_list); |
545 | } |
546 | - if (skb_queue_len(&mcast->pkt_queue) < IPOIB_MAX_MCAST_QUEUE) |
547 | + if (skb_queue_len(&mcast->pkt_queue) < IPOIB_MAX_MCAST_QUEUE) { |
548 | + /* put pseudoheader back on for next time */ |
549 | + skb_push(skb, sizeof(struct ipoib_pseudo_header)); |
550 | skb_queue_tail(&mcast->pkt_queue, skb); |
551 | - else { |
552 | + } else { |
553 | ++dev->stats.tx_dropped; |
554 | dev_kfree_skb_any(skb); |
555 | } |
556 | diff --git a/drivers/isdn/hardware/eicon/message.c b/drivers/isdn/hardware/eicon/message.c |
557 | index d7c286656a25..7b4ddf0a39ec 100644 |
558 | --- a/drivers/isdn/hardware/eicon/message.c |
559 | +++ b/drivers/isdn/hardware/eicon/message.c |
560 | @@ -11304,7 +11304,8 @@ static void mixer_notify_update(PLCI *plci, byte others) |
561 | ((CAPI_MSG *) msg)->header.ncci = 0; |
562 | ((CAPI_MSG *) msg)->info.facility_req.Selector = SELECTOR_LINE_INTERCONNECT; |
563 | ((CAPI_MSG *) msg)->info.facility_req.structs[0] = 3; |
564 | - PUT_WORD(&(((CAPI_MSG *) msg)->info.facility_req.structs[1]), LI_REQ_SILENT_UPDATE); |
565 | + ((CAPI_MSG *) msg)->info.facility_req.structs[1] = LI_REQ_SILENT_UPDATE & 0xff; |
566 | + ((CAPI_MSG *) msg)->info.facility_req.structs[2] = LI_REQ_SILENT_UPDATE >> 8; |
567 | ((CAPI_MSG *) msg)->info.facility_req.structs[3] = 0; |
568 | w = api_put(notify_plci->appl, (CAPI_MSG *) msg); |
569 | if (w != _QUEUE_FULL) |
570 | diff --git a/drivers/media/i2c/Kconfig b/drivers/media/i2c/Kconfig |
571 | index 521bbf1b29bc..670240c0ece8 100644 |
572 | --- a/drivers/media/i2c/Kconfig |
573 | +++ b/drivers/media/i2c/Kconfig |
574 | @@ -607,6 +607,7 @@ config VIDEO_S5K6A3 |
575 | config VIDEO_S5K4ECGX |
576 | tristate "Samsung S5K4ECGX sensor support" |
577 | depends on I2C && VIDEO_V4L2 && VIDEO_V4L2_SUBDEV_API |
578 | + select CRC32 |
579 | ---help--- |
580 | This is a V4L2 sensor-level driver for Samsung S5K4ECGX 5M |
581 | camera sensor with an embedded SoC image signal processor. |
582 | diff --git a/drivers/net/can/c_can/c_can_pci.c b/drivers/net/can/c_can/c_can_pci.c |
583 | index 7be393c96b1a..cf7c18947189 100644 |
584 | --- a/drivers/net/can/c_can/c_can_pci.c |
585 | +++ b/drivers/net/can/c_can/c_can_pci.c |
586 | @@ -161,6 +161,7 @@ static int c_can_pci_probe(struct pci_dev *pdev, |
587 | |
588 | dev->irq = pdev->irq; |
589 | priv->base = addr; |
590 | + priv->device = &pdev->dev; |
591 | |
592 | if (!c_can_pci_data->freq) { |
593 | dev_err(&pdev->dev, "no clock frequency defined\n"); |
594 | diff --git a/drivers/net/can/ti_hecc.c b/drivers/net/can/ti_hecc.c |
595 | index 680d1ff07a55..6749b1829469 100644 |
596 | --- a/drivers/net/can/ti_hecc.c |
597 | +++ b/drivers/net/can/ti_hecc.c |
598 | @@ -948,7 +948,12 @@ static int ti_hecc_probe(struct platform_device *pdev) |
599 | netif_napi_add(ndev, &priv->napi, ti_hecc_rx_poll, |
600 | HECC_DEF_NAPI_WEIGHT); |
601 | |
602 | - clk_enable(priv->clk); |
603 | + err = clk_prepare_enable(priv->clk); |
604 | + if (err) { |
605 | + dev_err(&pdev->dev, "clk_prepare_enable() failed\n"); |
606 | + goto probe_exit_clk; |
607 | + } |
608 | + |
609 | err = register_candev(ndev); |
610 | if (err) { |
611 | dev_err(&pdev->dev, "register_candev() failed\n"); |
612 | @@ -981,7 +986,7 @@ static int ti_hecc_remove(struct platform_device *pdev) |
613 | struct ti_hecc_priv *priv = netdev_priv(ndev); |
614 | |
615 | unregister_candev(ndev); |
616 | - clk_disable(priv->clk); |
617 | + clk_disable_unprepare(priv->clk); |
618 | clk_put(priv->clk); |
619 | res = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
620 | iounmap(priv->base); |
621 | @@ -1006,7 +1011,7 @@ static int ti_hecc_suspend(struct platform_device *pdev, pm_message_t state) |
622 | hecc_set_bit(priv, HECC_CANMC, HECC_CANMC_PDR); |
623 | priv->can.state = CAN_STATE_SLEEPING; |
624 | |
625 | - clk_disable(priv->clk); |
626 | + clk_disable_unprepare(priv->clk); |
627 | |
628 | return 0; |
629 | } |
630 | @@ -1015,8 +1020,11 @@ static int ti_hecc_resume(struct platform_device *pdev) |
631 | { |
632 | struct net_device *dev = platform_get_drvdata(pdev); |
633 | struct ti_hecc_priv *priv = netdev_priv(dev); |
634 | + int err; |
635 | |
636 | - clk_enable(priv->clk); |
637 | + err = clk_prepare_enable(priv->clk); |
638 | + if (err) |
639 | + return err; |
640 | |
641 | hecc_clear_bit(priv, HECC_CANMC, HECC_CANMC_PDR); |
642 | priv->can.state = CAN_STATE_ERROR_ACTIVE; |
643 | diff --git a/drivers/pinctrl/intel/pinctrl-broxton.c b/drivers/pinctrl/intel/pinctrl-broxton.c |
644 | index 5979d38c46b2..7329500943a3 100644 |
645 | --- a/drivers/pinctrl/intel/pinctrl-broxton.c |
646 | +++ b/drivers/pinctrl/intel/pinctrl-broxton.c |
647 | @@ -19,7 +19,7 @@ |
648 | |
649 | #define BXT_PAD_OWN 0x020 |
650 | #define BXT_HOSTSW_OWN 0x080 |
651 | -#define BXT_PADCFGLOCK 0x090 |
652 | +#define BXT_PADCFGLOCK 0x060 |
653 | #define BXT_GPI_IE 0x110 |
654 | |
655 | #define BXT_COMMUNITY(s, e) \ |
656 | diff --git a/drivers/platform/x86/intel_mid_powerbtn.c b/drivers/platform/x86/intel_mid_powerbtn.c |
657 | index 1fc0de870ff8..361770568ad0 100644 |
658 | --- a/drivers/platform/x86/intel_mid_powerbtn.c |
659 | +++ b/drivers/platform/x86/intel_mid_powerbtn.c |
660 | @@ -77,7 +77,7 @@ static int mfld_pb_probe(struct platform_device *pdev) |
661 | |
662 | input_set_capability(input, EV_KEY, KEY_POWER); |
663 | |
664 | - error = request_threaded_irq(irq, NULL, mfld_pb_isr, 0, |
665 | + error = request_threaded_irq(irq, NULL, mfld_pb_isr, IRQF_ONESHOT, |
666 | DRIVER_NAME, input); |
667 | if (error) { |
668 | dev_err(&pdev->dev, "Unable to request irq %d for mfld power" |
669 | diff --git a/drivers/video/fbdev/core/fbcmap.c b/drivers/video/fbdev/core/fbcmap.c |
670 | index f89245b8ba8e..68a113594808 100644 |
671 | --- a/drivers/video/fbdev/core/fbcmap.c |
672 | +++ b/drivers/video/fbdev/core/fbcmap.c |
673 | @@ -163,17 +163,18 @@ void fb_dealloc_cmap(struct fb_cmap *cmap) |
674 | |
675 | int fb_copy_cmap(const struct fb_cmap *from, struct fb_cmap *to) |
676 | { |
677 | - int tooff = 0, fromoff = 0; |
678 | - int size; |
679 | + unsigned int tooff = 0, fromoff = 0; |
680 | + size_t size; |
681 | |
682 | if (to->start > from->start) |
683 | fromoff = to->start - from->start; |
684 | else |
685 | tooff = from->start - to->start; |
686 | - size = to->len - tooff; |
687 | - if (size > (int) (from->len - fromoff)) |
688 | - size = from->len - fromoff; |
689 | - if (size <= 0) |
690 | + if (fromoff >= from->len || tooff >= to->len) |
691 | + return -EINVAL; |
692 | + |
693 | + size = min_t(size_t, to->len - tooff, from->len - fromoff); |
694 | + if (size == 0) |
695 | return -EINVAL; |
696 | size *= sizeof(u16); |
697 | |
698 | @@ -187,17 +188,18 @@ int fb_copy_cmap(const struct fb_cmap *from, struct fb_cmap *to) |
699 | |
700 | int fb_cmap_to_user(const struct fb_cmap *from, struct fb_cmap_user *to) |
701 | { |
702 | - int tooff = 0, fromoff = 0; |
703 | - int size; |
704 | + unsigned int tooff = 0, fromoff = 0; |
705 | + size_t size; |
706 | |
707 | if (to->start > from->start) |
708 | fromoff = to->start - from->start; |
709 | else |
710 | tooff = from->start - to->start; |
711 | - size = to->len - tooff; |
712 | - if (size > (int) (from->len - fromoff)) |
713 | - size = from->len - fromoff; |
714 | - if (size <= 0) |
715 | + if (fromoff >= from->len || tooff >= to->len) |
716 | + return -EINVAL; |
717 | + |
718 | + size = min_t(size_t, to->len - tooff, from->len - fromoff); |
719 | + if (size == 0) |
720 | return -EINVAL; |
721 | size *= sizeof(u16); |
722 | |
723 | diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c |
724 | index 3c69299c01ab..9a524e763c3e 100644 |
725 | --- a/fs/nfs/nfs4proc.c |
726 | +++ b/fs/nfs/nfs4proc.c |
727 | @@ -2422,7 +2422,8 @@ static inline void nfs4_exclusive_attrset(struct nfs4_opendata *opendata, |
728 | sattr->ia_valid |= ATTR_MTIME; |
729 | |
730 | /* Except MODE, it seems harmless of setting twice. */ |
731 | - if ((attrset[1] & FATTR4_WORD1_MODE)) |
732 | + if (opendata->o_arg.createmode != NFS4_CREATE_EXCLUSIVE && |
733 | + attrset[1] & FATTR4_WORD1_MODE) |
734 | sattr->ia_valid &= ~ATTR_MODE; |
735 | |
736 | if (attrset[2] & FATTR4_WORD2_SECURITY_LABEL) |
737 | diff --git a/include/linux/nfs4.h b/include/linux/nfs4.h |
738 | index e7e78537aea2..63a817631f06 100644 |
739 | --- a/include/linux/nfs4.h |
740 | +++ b/include/linux/nfs4.h |
741 | @@ -266,7 +266,7 @@ enum nfsstat4 { |
742 | |
743 | static inline bool seqid_mutating_err(u32 err) |
744 | { |
745 | - /* rfc 3530 section 8.1.5: */ |
746 | + /* See RFC 7530, section 9.1.7 */ |
747 | switch (err) { |
748 | case NFS4ERR_STALE_CLIENTID: |
749 | case NFS4ERR_STALE_STATEID: |
750 | @@ -275,6 +275,7 @@ static inline bool seqid_mutating_err(u32 err) |
751 | case NFS4ERR_BADXDR: |
752 | case NFS4ERR_RESOURCE: |
753 | case NFS4ERR_NOFILEHANDLE: |
754 | + case NFS4ERR_MOVED: |
755 | return false; |
756 | }; |
757 | return true; |
758 | diff --git a/include/linux/sunrpc/clnt.h b/include/linux/sunrpc/clnt.h |
759 | index 9b6027c51736..316a5525b730 100644 |
760 | --- a/include/linux/sunrpc/clnt.h |
761 | +++ b/include/linux/sunrpc/clnt.h |
762 | @@ -180,5 +180,6 @@ const char *rpc_peeraddr2str(struct rpc_clnt *, enum rpc_display_format_t); |
763 | int rpc_localaddr(struct rpc_clnt *, struct sockaddr *, size_t); |
764 | |
765 | const char *rpc_proc_name(const struct rpc_task *task); |
766 | +void rpc_cleanup_clids(void); |
767 | #endif /* __KERNEL__ */ |
768 | #endif /* _LINUX_SUNRPC_CLNT_H */ |
769 | diff --git a/kernel/sysctl.c b/kernel/sysctl.c |
770 | index 999e025bf68e..2f0d157258a2 100644 |
771 | --- a/kernel/sysctl.c |
772 | +++ b/kernel/sysctl.c |
773 | @@ -2414,6 +2414,7 @@ static int __do_proc_doulongvec_minmax(void *data, struct ctl_table *table, int |
774 | break; |
775 | if (neg) |
776 | continue; |
777 | + val = convmul * val / convdiv; |
778 | if ((min && val < *min) || (max && val > *max)) |
779 | continue; |
780 | *i = val; |
781 | diff --git a/mm/memcontrol.c b/mm/memcontrol.c |
782 | index 5d9c8a3136bc..43eefe9d834c 100644 |
783 | --- a/mm/memcontrol.c |
784 | +++ b/mm/memcontrol.c |
785 | @@ -4496,9 +4496,9 @@ static int mem_cgroup_do_precharge(unsigned long count) |
786 | return ret; |
787 | } |
788 | |
789 | - /* Try charges one by one with reclaim */ |
790 | + /* Try charges one by one with reclaim, but do not retry */ |
791 | while (count--) { |
792 | - ret = try_charge(mc.to, GFP_KERNEL & ~__GFP_NORETRY, 1); |
793 | + ret = try_charge(mc.to, GFP_KERNEL | __GFP_NORETRY, 1); |
794 | if (ret) |
795 | return ret; |
796 | mc.precharge++; |
797 | diff --git a/mm/mempolicy.c b/mm/mempolicy.c |
798 | index 87a177917cb2..a4217fe60dff 100644 |
799 | --- a/mm/mempolicy.c |
800 | +++ b/mm/mempolicy.c |
801 | @@ -2006,8 +2006,8 @@ retry_cpuset: |
802 | |
803 | nmask = policy_nodemask(gfp, pol); |
804 | zl = policy_zonelist(gfp, pol, node); |
805 | - mpol_cond_put(pol); |
806 | page = __alloc_pages_nodemask(gfp, order, zl, nmask); |
807 | + mpol_cond_put(pol); |
808 | out: |
809 | if (unlikely(!page && read_mems_allowed_retry(cpuset_mems_cookie))) |
810 | goto retry_cpuset; |
811 | diff --git a/net/sunrpc/clnt.c b/net/sunrpc/clnt.c |
812 | index 7a93922457ff..f28aeb2cfd32 100644 |
813 | --- a/net/sunrpc/clnt.c |
814 | +++ b/net/sunrpc/clnt.c |
815 | @@ -337,6 +337,11 @@ out: |
816 | |
817 | static DEFINE_IDA(rpc_clids); |
818 | |
819 | +void rpc_cleanup_clids(void) |
820 | +{ |
821 | + ida_destroy(&rpc_clids); |
822 | +} |
823 | + |
824 | static int rpc_alloc_clid(struct rpc_clnt *clnt) |
825 | { |
826 | int clid; |
827 | diff --git a/net/sunrpc/sunrpc_syms.c b/net/sunrpc/sunrpc_syms.c |
828 | index ee5d3d253102..3142f38d1104 100644 |
829 | --- a/net/sunrpc/sunrpc_syms.c |
830 | +++ b/net/sunrpc/sunrpc_syms.c |
831 | @@ -119,6 +119,7 @@ out: |
832 | static void __exit |
833 | cleanup_sunrpc(void) |
834 | { |
835 | + rpc_cleanup_clids(); |
836 | rpcauth_remove_module(); |
837 | cleanup_socket_xprt(); |
838 | svc_cleanup_xprt_sock(); |