Magellan Linux

Contents of /trunk/kernel-magellan/patches-4.8/0103-4.8.4-all-fixes.patch

Parent Directory Parent Directory | Revision Log Revision Log


Revision 2843 - (show annotations) (download)
Tue Nov 22 13:19:27 2016 UTC (7 years, 5 months ago) by niro
File size: 73868 byte(s)
-linux-4.8.4
1 diff --git a/MAINTAINERS b/MAINTAINERS
2 index f593300e310b..babaf8261941 100644
3 --- a/MAINTAINERS
4 +++ b/MAINTAINERS
5 @@ -12951,11 +12951,10 @@ F: arch/x86/xen/*swiotlb*
6 F: drivers/xen/*swiotlb*
7
8 XFS FILESYSTEM
9 -P: Silicon Graphics Inc
10 M: Dave Chinner <david@fromorbit.com>
11 -M: xfs@oss.sgi.com
12 -L: xfs@oss.sgi.com
13 -W: http://oss.sgi.com/projects/xfs
14 +M: linux-xfs@vger.kernel.org
15 +L: linux-xfs@vger.kernel.org
16 +W: http://xfs.org/
17 T: git git://git.kernel.org/pub/scm/linux/kernel/git/dgc/linux-xfs.git
18 S: Supported
19 F: Documentation/filesystems/xfs.txt
20 diff --git a/Makefile b/Makefile
21 index 42eb45c86a42..82a36ab540a4 100644
22 --- a/Makefile
23 +++ b/Makefile
24 @@ -1,6 +1,6 @@
25 VERSION = 4
26 PATCHLEVEL = 8
27 -SUBLEVEL = 3
28 +SUBLEVEL = 4
29 EXTRAVERSION =
30 NAME = Psychotic Stoned Sheep
31
32 diff --git a/arch/arc/include/asm/irqflags-arcv2.h b/arch/arc/include/asm/irqflags-arcv2.h
33 index d1ec7f6b31e0..e880dfa3fcd3 100644
34 --- a/arch/arc/include/asm/irqflags-arcv2.h
35 +++ b/arch/arc/include/asm/irqflags-arcv2.h
36 @@ -112,7 +112,7 @@ static inline long arch_local_save_flags(void)
37 */
38 temp = (1 << 5) |
39 ((!!(temp & STATUS_IE_MASK)) << CLRI_STATUS_IE_BIT) |
40 - (temp & CLRI_STATUS_E_MASK);
41 + ((temp >> 1) & CLRI_STATUS_E_MASK);
42 return temp;
43 }
44
45 diff --git a/arch/arc/kernel/intc-arcv2.c b/arch/arc/kernel/intc-arcv2.c
46 index 6c24faf48b16..62b59409a5d9 100644
47 --- a/arch/arc/kernel/intc-arcv2.c
48 +++ b/arch/arc/kernel/intc-arcv2.c
49 @@ -74,7 +74,7 @@ void arc_init_IRQ(void)
50 tmp = read_aux_reg(0xa);
51 tmp |= STATUS_AD_MASK | (irq_prio << 1);
52 tmp &= ~STATUS_IE_MASK;
53 - asm volatile("flag %0 \n"::"r"(tmp));
54 + asm volatile("kflag %0 \n"::"r"(tmp));
55 }
56
57 static void arcv2_irq_mask(struct irq_data *data)
58 diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c
59 index cc2f6dbd4303..5e24d880306c 100644
60 --- a/block/cfq-iosched.c
61 +++ b/block/cfq-iosched.c
62 @@ -3042,7 +3042,6 @@ static struct request *cfq_check_fifo(struct cfq_queue *cfqq)
63 if (ktime_get_ns() < rq->fifo_time)
64 rq = NULL;
65
66 - cfq_log_cfqq(cfqq->cfqd, cfqq, "fifo=%p", rq);
67 return rq;
68 }
69
70 @@ -3420,6 +3419,9 @@ static bool cfq_may_dispatch(struct cfq_data *cfqd, struct cfq_queue *cfqq)
71 {
72 unsigned int max_dispatch;
73
74 + if (cfq_cfqq_must_dispatch(cfqq))
75 + return true;
76 +
77 /*
78 * Drain async requests before we start sync IO
79 */
80 @@ -3511,15 +3513,20 @@ static bool cfq_dispatch_request(struct cfq_data *cfqd, struct cfq_queue *cfqq)
81
82 BUG_ON(RB_EMPTY_ROOT(&cfqq->sort_list));
83
84 + rq = cfq_check_fifo(cfqq);
85 + if (rq)
86 + cfq_mark_cfqq_must_dispatch(cfqq);
87 +
88 if (!cfq_may_dispatch(cfqd, cfqq))
89 return false;
90
91 /*
92 * follow expired path, else get first next available
93 */
94 - rq = cfq_check_fifo(cfqq);
95 if (!rq)
96 rq = cfqq->next_rq;
97 + else
98 + cfq_log_cfqq(cfqq->cfqd, cfqq, "fifo=%p", rq);
99
100 /*
101 * insert request into driver dispatch list
102 @@ -3989,7 +3996,7 @@ cfq_should_preempt(struct cfq_data *cfqd, struct cfq_queue *new_cfqq,
103 * if the new request is sync, but the currently running queue is
104 * not, let the sync request have priority.
105 */
106 - if (rq_is_sync(rq) && !cfq_cfqq_sync(cfqq))
107 + if (rq_is_sync(rq) && !cfq_cfqq_sync(cfqq) && !cfq_cfqq_must_dispatch(cfqq))
108 return true;
109
110 /*
111 diff --git a/crypto/async_tx/async_pq.c b/crypto/async_tx/async_pq.c
112 index 08b3ac68952b..f83de99d7d71 100644
113 --- a/crypto/async_tx/async_pq.c
114 +++ b/crypto/async_tx/async_pq.c
115 @@ -368,8 +368,6 @@ async_syndrome_val(struct page **blocks, unsigned int offset, int disks,
116
117 dma_set_unmap(tx, unmap);
118 async_tx_submit(chan, tx, submit);
119 -
120 - return tx;
121 } else {
122 struct page *p_src = P(blocks, disks);
123 struct page *q_src = Q(blocks, disks);
124 @@ -424,9 +422,11 @@ async_syndrome_val(struct page **blocks, unsigned int offset, int disks,
125 submit->cb_param = cb_param_orig;
126 submit->flags = flags_orig;
127 async_tx_sync_epilog(submit);
128 -
129 - return NULL;
130 + tx = NULL;
131 }
132 + dmaengine_unmap_put(unmap);
133 +
134 + return tx;
135 }
136 EXPORT_SYMBOL_GPL(async_syndrome_val);
137
138 diff --git a/crypto/ghash-generic.c b/crypto/ghash-generic.c
139 index bac70995e064..12ad3e3a84e3 100644
140 --- a/crypto/ghash-generic.c
141 +++ b/crypto/ghash-generic.c
142 @@ -14,24 +14,13 @@
143
144 #include <crypto/algapi.h>
145 #include <crypto/gf128mul.h>
146 +#include <crypto/ghash.h>
147 #include <crypto/internal/hash.h>
148 #include <linux/crypto.h>
149 #include <linux/init.h>
150 #include <linux/kernel.h>
151 #include <linux/module.h>
152
153 -#define GHASH_BLOCK_SIZE 16
154 -#define GHASH_DIGEST_SIZE 16
155 -
156 -struct ghash_ctx {
157 - struct gf128mul_4k *gf128;
158 -};
159 -
160 -struct ghash_desc_ctx {
161 - u8 buffer[GHASH_BLOCK_SIZE];
162 - u32 bytes;
163 -};
164 -
165 static int ghash_init(struct shash_desc *desc)
166 {
167 struct ghash_desc_ctx *dctx = shash_desc_ctx(desc);
168 diff --git a/drivers/acpi/nfit/core.c b/drivers/acpi/nfit/core.c
169 index e1d5ea6d5e40..2accf784534e 100644
170 --- a/drivers/acpi/nfit/core.c
171 +++ b/drivers/acpi/nfit/core.c
172 @@ -2689,6 +2689,9 @@ static void acpi_nfit_notify(struct acpi_device *adev, u32 event)
173
174 dev_dbg(dev, "%s: event: %d\n", __func__, event);
175
176 + if (event != NFIT_NOTIFY_UPDATE)
177 + return;
178 +
179 device_lock(dev);
180 if (!dev->driver) {
181 /* dev->driver may be null if we're being removed */
182 diff --git a/drivers/acpi/nfit/nfit.h b/drivers/acpi/nfit/nfit.h
183 index e894ded24d99..51d23f130d86 100644
184 --- a/drivers/acpi/nfit/nfit.h
185 +++ b/drivers/acpi/nfit/nfit.h
186 @@ -78,6 +78,10 @@ enum {
187 NFIT_ARS_TIMEOUT = 90,
188 };
189
190 +enum nfit_root_notifiers {
191 + NFIT_NOTIFY_UPDATE = 0x80,
192 +};
193 +
194 struct nfit_spa {
195 struct list_head list;
196 struct nd_region *nd_region;
197 diff --git a/drivers/base/dma-mapping.c b/drivers/base/dma-mapping.c
198 index d799662f19eb..261420ddfe66 100644
199 --- a/drivers/base/dma-mapping.c
200 +++ b/drivers/base/dma-mapping.c
201 @@ -334,7 +334,7 @@ void dma_common_free_remap(void *cpu_addr, size_t size, unsigned long vm_flags)
202 return;
203 }
204
205 - unmap_kernel_range((unsigned long)cpu_addr, size);
206 + unmap_kernel_range((unsigned long)cpu_addr, PAGE_ALIGN(size));
207 vunmap(cpu_addr);
208 }
209 #endif
210 diff --git a/drivers/clk/mvebu/cp110-system-controller.c b/drivers/clk/mvebu/cp110-system-controller.c
211 index 7fa42d6b2b92..f2303da7fda7 100644
212 --- a/drivers/clk/mvebu/cp110-system-controller.c
213 +++ b/drivers/clk/mvebu/cp110-system-controller.c
214 @@ -81,13 +81,6 @@ enum {
215 #define CP110_GATE_EIP150 25
216 #define CP110_GATE_EIP197 26
217
218 -static struct clk *cp110_clks[CP110_CLK_NUM];
219 -
220 -static struct clk_onecell_data cp110_clk_data = {
221 - .clks = cp110_clks,
222 - .clk_num = CP110_CLK_NUM,
223 -};
224 -
225 struct cp110_gate_clk {
226 struct clk_hw hw;
227 struct regmap *regmap;
228 @@ -142,6 +135,8 @@ static struct clk *cp110_register_gate(const char *name,
229 if (!gate)
230 return ERR_PTR(-ENOMEM);
231
232 + memset(&init, 0, sizeof(init));
233 +
234 init.name = name;
235 init.ops = &cp110_gate_ops;
236 init.parent_names = &parent_name;
237 @@ -194,7 +189,8 @@ static int cp110_syscon_clk_probe(struct platform_device *pdev)
238 struct regmap *regmap;
239 struct device_node *np = pdev->dev.of_node;
240 const char *ppv2_name, *apll_name, *core_name, *eip_name, *nand_name;
241 - struct clk *clk;
242 + struct clk_onecell_data *cp110_clk_data;
243 + struct clk *clk, **cp110_clks;
244 u32 nand_clk_ctrl;
245 int i, ret;
246
247 @@ -207,6 +203,20 @@ static int cp110_syscon_clk_probe(struct platform_device *pdev)
248 if (ret)
249 return ret;
250
251 + cp110_clks = devm_kcalloc(&pdev->dev, sizeof(struct clk *),
252 + CP110_CLK_NUM, GFP_KERNEL);
253 + if (!cp110_clks)
254 + return -ENOMEM;
255 +
256 + cp110_clk_data = devm_kzalloc(&pdev->dev,
257 + sizeof(*cp110_clk_data),
258 + GFP_KERNEL);
259 + if (!cp110_clk_data)
260 + return -ENOMEM;
261 +
262 + cp110_clk_data->clks = cp110_clks;
263 + cp110_clk_data->clk_num = CP110_CLK_NUM;
264 +
265 /* Register the APLL which is the root of the clk tree */
266 of_property_read_string_index(np, "core-clock-output-names",
267 CP110_CORE_APLL, &apll_name);
268 @@ -334,10 +344,12 @@ static int cp110_syscon_clk_probe(struct platform_device *pdev)
269 cp110_clks[CP110_MAX_CORE_CLOCKS + i] = clk;
270 }
271
272 - ret = of_clk_add_provider(np, cp110_of_clk_get, &cp110_clk_data);
273 + ret = of_clk_add_provider(np, cp110_of_clk_get, cp110_clk_data);
274 if (ret)
275 goto fail_clk_add;
276
277 + platform_set_drvdata(pdev, cp110_clks);
278 +
279 return 0;
280
281 fail_clk_add:
282 @@ -364,6 +376,7 @@ fail0:
283
284 static int cp110_syscon_clk_remove(struct platform_device *pdev)
285 {
286 + struct clk **cp110_clks = platform_get_drvdata(pdev);
287 int i;
288
289 of_clk_del_provider(pdev->dev.of_node);
290 diff --git a/drivers/crypto/vmx/ghash.c b/drivers/crypto/vmx/ghash.c
291 index 6c999cb01b80..27a94a119009 100644
292 --- a/drivers/crypto/vmx/ghash.c
293 +++ b/drivers/crypto/vmx/ghash.c
294 @@ -26,16 +26,13 @@
295 #include <linux/hardirq.h>
296 #include <asm/switch_to.h>
297 #include <crypto/aes.h>
298 +#include <crypto/ghash.h>
299 #include <crypto/scatterwalk.h>
300 #include <crypto/internal/hash.h>
301 #include <crypto/b128ops.h>
302
303 #define IN_INTERRUPT in_interrupt()
304
305 -#define GHASH_BLOCK_SIZE (16)
306 -#define GHASH_DIGEST_SIZE (16)
307 -#define GHASH_KEY_LEN (16)
308 -
309 void gcm_init_p8(u128 htable[16], const u64 Xi[2]);
310 void gcm_gmult_p8(u64 Xi[2], const u128 htable[16]);
311 void gcm_ghash_p8(u64 Xi[2], const u128 htable[16],
312 @@ -55,16 +52,11 @@ struct p8_ghash_desc_ctx {
313
314 static int p8_ghash_init_tfm(struct crypto_tfm *tfm)
315 {
316 - const char *alg;
317 + const char *alg = "ghash-generic";
318 struct crypto_shash *fallback;
319 struct crypto_shash *shash_tfm = __crypto_shash_cast(tfm);
320 struct p8_ghash_ctx *ctx = crypto_tfm_ctx(tfm);
321
322 - if (!(alg = crypto_tfm_alg_name(tfm))) {
323 - printk(KERN_ERR "Failed to get algorithm name.\n");
324 - return -ENOENT;
325 - }
326 -
327 fallback = crypto_alloc_shash(alg, 0, CRYPTO_ALG_NEED_FALLBACK);
328 if (IS_ERR(fallback)) {
329 printk(KERN_ERR
330 @@ -78,10 +70,18 @@ static int p8_ghash_init_tfm(struct crypto_tfm *tfm)
331 crypto_shash_set_flags(fallback,
332 crypto_shash_get_flags((struct crypto_shash
333 *) tfm));
334 - ctx->fallback = fallback;
335
336 - shash_tfm->descsize = sizeof(struct p8_ghash_desc_ctx)
337 - + crypto_shash_descsize(fallback);
338 + /* Check if the descsize defined in the algorithm is still enough. */
339 + if (shash_tfm->descsize < sizeof(struct p8_ghash_desc_ctx)
340 + + crypto_shash_descsize(fallback)) {
341 + printk(KERN_ERR
342 + "Desc size of the fallback implementation (%s) does not match the expected value: %lu vs %u\n",
343 + alg,
344 + shash_tfm->descsize - sizeof(struct p8_ghash_desc_ctx),
345 + crypto_shash_descsize(fallback));
346 + return -EINVAL;
347 + }
348 + ctx->fallback = fallback;
349
350 return 0;
351 }
352 @@ -113,7 +113,7 @@ static int p8_ghash_setkey(struct crypto_shash *tfm, const u8 *key,
353 {
354 struct p8_ghash_ctx *ctx = crypto_tfm_ctx(crypto_shash_tfm(tfm));
355
356 - if (keylen != GHASH_KEY_LEN)
357 + if (keylen != GHASH_BLOCK_SIZE)
358 return -EINVAL;
359
360 preempt_disable();
361 @@ -211,7 +211,8 @@ struct shash_alg p8_ghash_alg = {
362 .update = p8_ghash_update,
363 .final = p8_ghash_final,
364 .setkey = p8_ghash_setkey,
365 - .descsize = sizeof(struct p8_ghash_desc_ctx),
366 + .descsize = sizeof(struct p8_ghash_desc_ctx)
367 + + sizeof(struct ghash_desc_ctx),
368 .base = {
369 .cra_name = "ghash",
370 .cra_driver_name = "p8_ghash",
371 diff --git a/drivers/gpu/drm/virtio/virtgpu_drm_bus.c b/drivers/gpu/drm/virtio/virtgpu_drm_bus.c
372 index 7f0e93f87a55..88a39165edd5 100644
373 --- a/drivers/gpu/drm/virtio/virtgpu_drm_bus.c
374 +++ b/drivers/gpu/drm/virtio/virtgpu_drm_bus.c
375 @@ -27,6 +27,16 @@
376
377 #include "virtgpu_drv.h"
378
379 +int drm_virtio_set_busid(struct drm_device *dev, struct drm_master *master)
380 +{
381 + struct pci_dev *pdev = dev->pdev;
382 +
383 + if (pdev) {
384 + return drm_pci_set_busid(dev, master);
385 + }
386 + return 0;
387 +}
388 +
389 static void virtio_pci_kick_out_firmware_fb(struct pci_dev *pci_dev)
390 {
391 struct apertures_struct *ap;
392 diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.c b/drivers/gpu/drm/virtio/virtgpu_drv.c
393 index c13f70cfc461..5820b7020ae5 100644
394 --- a/drivers/gpu/drm/virtio/virtgpu_drv.c
395 +++ b/drivers/gpu/drm/virtio/virtgpu_drv.c
396 @@ -117,6 +117,7 @@ static const struct file_operations virtio_gpu_driver_fops = {
397
398 static struct drm_driver driver = {
399 .driver_features = DRIVER_MODESET | DRIVER_GEM | DRIVER_PRIME | DRIVER_RENDER | DRIVER_ATOMIC,
400 + .set_busid = drm_virtio_set_busid,
401 .load = virtio_gpu_driver_load,
402 .unload = virtio_gpu_driver_unload,
403 .open = virtio_gpu_driver_open,
404 diff --git a/drivers/gpu/drm/virtio/virtgpu_drv.h b/drivers/gpu/drm/virtio/virtgpu_drv.h
405 index b18ef3111f0c..acf556a35cb2 100644
406 --- a/drivers/gpu/drm/virtio/virtgpu_drv.h
407 +++ b/drivers/gpu/drm/virtio/virtgpu_drv.h
408 @@ -49,6 +49,7 @@
409 #define DRIVER_PATCHLEVEL 1
410
411 /* virtgpu_drm_bus.c */
412 +int drm_virtio_set_busid(struct drm_device *dev, struct drm_master *master);
413 int drm_virtio_init(struct drm_driver *driver, struct virtio_device *vdev);
414
415 struct virtio_gpu_object {
416 diff --git a/drivers/infiniband/hw/hfi1/rc.c b/drivers/infiniband/hw/hfi1/rc.c
417 index 5da190e6011b..bcf76c33726b 100644
418 --- a/drivers/infiniband/hw/hfi1/rc.c
419 +++ b/drivers/infiniband/hw/hfi1/rc.c
420 @@ -932,8 +932,10 @@ void hfi1_send_rc_ack(struct hfi1_ctxtdata *rcd, struct rvt_qp *qp,
421 return;
422
423 queue_ack:
424 - this_cpu_inc(*ibp->rvp.rc_qacks);
425 spin_lock_irqsave(&qp->s_lock, flags);
426 + if (!(ib_rvt_state_ops[qp->state] & RVT_PROCESS_RECV_OK))
427 + goto unlock;
428 + this_cpu_inc(*ibp->rvp.rc_qacks);
429 qp->s_flags |= RVT_S_ACK_PENDING | RVT_S_RESP_PENDING;
430 qp->s_nak_state = qp->r_nak_state;
431 qp->s_ack_psn = qp->r_ack_psn;
432 @@ -942,6 +944,7 @@ queue_ack:
433
434 /* Schedule the send tasklet. */
435 hfi1_schedule_send(qp);
436 +unlock:
437 spin_unlock_irqrestore(&qp->s_lock, flags);
438 }
439
440 diff --git a/drivers/misc/mei/amthif.c b/drivers/misc/mei/amthif.c
441 index a039a5df6f21..fd9271bc1a11 100644
442 --- a/drivers/misc/mei/amthif.c
443 +++ b/drivers/misc/mei/amthif.c
444 @@ -67,8 +67,12 @@ int mei_amthif_host_init(struct mei_device *dev, struct mei_me_client *me_cl)
445 struct mei_cl *cl = &dev->iamthif_cl;
446 int ret;
447
448 - if (mei_cl_is_connected(cl))
449 - return 0;
450 + mutex_lock(&dev->device_lock);
451 +
452 + if (mei_cl_is_connected(cl)) {
453 + ret = 0;
454 + goto out;
455 + }
456
457 dev->iamthif_state = MEI_IAMTHIF_IDLE;
458
459 @@ -77,11 +81,13 @@ int mei_amthif_host_init(struct mei_device *dev, struct mei_me_client *me_cl)
460 ret = mei_cl_link(cl);
461 if (ret < 0) {
462 dev_err(dev->dev, "amthif: failed cl_link %d\n", ret);
463 - return ret;
464 + goto out;
465 }
466
467 ret = mei_cl_connect(cl, me_cl, NULL);
468
469 +out:
470 + mutex_unlock(&dev->device_lock);
471 return ret;
472 }
473
474 diff --git a/drivers/misc/mei/bus.c b/drivers/misc/mei/bus.c
475 index 1f33fea9299f..e094df3cf2d5 100644
476 --- a/drivers/misc/mei/bus.c
477 +++ b/drivers/misc/mei/bus.c
478 @@ -983,12 +983,10 @@ void mei_cl_bus_rescan_work(struct work_struct *work)
479 container_of(work, struct mei_device, bus_rescan_work);
480 struct mei_me_client *me_cl;
481
482 - mutex_lock(&bus->device_lock);
483 me_cl = mei_me_cl_by_uuid(bus, &mei_amthif_guid);
484 if (me_cl)
485 mei_amthif_host_init(bus, me_cl);
486 mei_me_cl_put(me_cl);
487 - mutex_unlock(&bus->device_lock);
488
489 mei_cl_bus_rescan(bus);
490 }
491 diff --git a/drivers/net/ethernet/intel/i40e/i40e_main.c b/drivers/net/ethernet/intel/i40e/i40e_main.c
492 index d0b3a1bb82ca..dad15b6c66dd 100644
493 --- a/drivers/net/ethernet/intel/i40e/i40e_main.c
494 +++ b/drivers/net/ethernet/intel/i40e/i40e_main.c
495 @@ -11360,6 +11360,12 @@ static pci_ers_result_t i40e_pci_error_detected(struct pci_dev *pdev,
496
497 dev_info(&pdev->dev, "%s: error %d\n", __func__, error);
498
499 + if (!pf) {
500 + dev_info(&pdev->dev,
501 + "Cannot recover - error happened during device probe\n");
502 + return PCI_ERS_RESULT_DISCONNECT;
503 + }
504 +
505 /* shutdown all operations */
506 if (!test_bit(__I40E_SUSPENDED, &pf->state)) {
507 rtnl_lock();
508 diff --git a/drivers/net/wireless/ath/carl9170/debug.c b/drivers/net/wireless/ath/carl9170/debug.c
509 index 6808db433283..ec3a64e5d2bb 100644
510 --- a/drivers/net/wireless/ath/carl9170/debug.c
511 +++ b/drivers/net/wireless/ath/carl9170/debug.c
512 @@ -75,7 +75,8 @@ static ssize_t carl9170_debugfs_read(struct file *file, char __user *userbuf,
513
514 if (!ar)
515 return -ENODEV;
516 - dfops = container_of(file->f_op, struct carl9170_debugfs_fops, fops);
517 + dfops = container_of(debugfs_real_fops(file),
518 + struct carl9170_debugfs_fops, fops);
519
520 if (!dfops->read)
521 return -ENOSYS;
522 @@ -127,7 +128,8 @@ static ssize_t carl9170_debugfs_write(struct file *file,
523
524 if (!ar)
525 return -ENODEV;
526 - dfops = container_of(file->f_op, struct carl9170_debugfs_fops, fops);
527 + dfops = container_of(debugfs_real_fops(file),
528 + struct carl9170_debugfs_fops, fops);
529
530 if (!dfops->write)
531 return -ENOSYS;
532 diff --git a/drivers/net/wireless/broadcom/b43/debugfs.c b/drivers/net/wireless/broadcom/b43/debugfs.c
533 index b4bcd94aff6c..77046384dd80 100644
534 --- a/drivers/net/wireless/broadcom/b43/debugfs.c
535 +++ b/drivers/net/wireless/broadcom/b43/debugfs.c
536 @@ -524,7 +524,8 @@ static ssize_t b43_debugfs_read(struct file *file, char __user *userbuf,
537 goto out_unlock;
538 }
539
540 - dfops = container_of(file->f_op, struct b43_debugfs_fops, fops);
541 + dfops = container_of(debugfs_real_fops(file),
542 + struct b43_debugfs_fops, fops);
543 if (!dfops->read) {
544 err = -ENOSYS;
545 goto out_unlock;
546 @@ -585,7 +586,8 @@ static ssize_t b43_debugfs_write(struct file *file,
547 goto out_unlock;
548 }
549
550 - dfops = container_of(file->f_op, struct b43_debugfs_fops, fops);
551 + dfops = container_of(debugfs_real_fops(file),
552 + struct b43_debugfs_fops, fops);
553 if (!dfops->write) {
554 err = -ENOSYS;
555 goto out_unlock;
556 diff --git a/drivers/net/wireless/broadcom/b43legacy/debugfs.c b/drivers/net/wireless/broadcom/b43legacy/debugfs.c
557 index 090910ea259e..82ef56ed7ca1 100644
558 --- a/drivers/net/wireless/broadcom/b43legacy/debugfs.c
559 +++ b/drivers/net/wireless/broadcom/b43legacy/debugfs.c
560 @@ -221,7 +221,8 @@ static ssize_t b43legacy_debugfs_read(struct file *file, char __user *userbuf,
561 goto out_unlock;
562 }
563
564 - dfops = container_of(file->f_op, struct b43legacy_debugfs_fops, fops);
565 + dfops = container_of(debugfs_real_fops(file),
566 + struct b43legacy_debugfs_fops, fops);
567 if (!dfops->read) {
568 err = -ENOSYS;
569 goto out_unlock;
570 @@ -287,7 +288,8 @@ static ssize_t b43legacy_debugfs_write(struct file *file,
571 goto out_unlock;
572 }
573
574 - dfops = container_of(file->f_op, struct b43legacy_debugfs_fops, fops);
575 + dfops = container_of(debugfs_real_fops(file),
576 + struct b43legacy_debugfs_fops, fops);
577 if (!dfops->write) {
578 err = -ENOSYS;
579 goto out_unlock;
580 diff --git a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
581 index b8aec5e5ef93..abaf003a5b39 100644
582 --- a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
583 +++ b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
584 @@ -2533,7 +2533,7 @@ static void brcmf_fill_bss_param(struct brcmf_if *ifp, struct station_info *si)
585 WL_BSS_INFO_MAX);
586 if (err) {
587 brcmf_err("Failed to get bss info (%d)\n", err);
588 - return;
589 + goto out_kfree;
590 }
591 si->filled |= BIT(NL80211_STA_INFO_BSS_PARAM);
592 si->bss_param.beacon_interval = le16_to_cpu(buf->bss_le.beacon_period);
593 @@ -2545,6 +2545,9 @@ static void brcmf_fill_bss_param(struct brcmf_if *ifp, struct station_info *si)
594 si->bss_param.flags |= BSS_PARAM_FLAGS_SHORT_PREAMBLE;
595 if (capability & WLAN_CAPABILITY_SHORT_SLOT_TIME)
596 si->bss_param.flags |= BSS_PARAM_FLAGS_SHORT_SLOT_TIME;
597 +
598 +out_kfree:
599 + kfree(buf);
600 }
601
602 static s32
603 @@ -3884,11 +3887,11 @@ brcmf_cfg80211_del_pmksa(struct wiphy *wiphy, struct net_device *ndev,
604 if (!check_vif_up(ifp->vif))
605 return -EIO;
606
607 - brcmf_dbg(CONN, "del_pmksa - PMK bssid = %pM\n", &pmksa->bssid);
608 + brcmf_dbg(CONN, "del_pmksa - PMK bssid = %pM\n", pmksa->bssid);
609
610 npmk = le32_to_cpu(cfg->pmk_list.npmk);
611 for (i = 0; i < npmk; i++)
612 - if (!memcmp(&pmksa->bssid, &pmk[i].bssid, ETH_ALEN))
613 + if (!memcmp(pmksa->bssid, pmk[i].bssid, ETH_ALEN))
614 break;
615
616 if ((npmk > 0) && (i < npmk)) {
617 diff --git a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
618 index 7e269f9aa607..63664442e687 100644
619 --- a/drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
620 +++ b/drivers/net/wireless/broadcom/brcm80211/brcmfmac/flowring.c
621 @@ -234,13 +234,20 @@ static void brcmf_flowring_block(struct brcmf_flowring *flow, u16 flowid,
622
623 void brcmf_flowring_delete(struct brcmf_flowring *flow, u16 flowid)
624 {
625 + struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev);
626 struct brcmf_flowring_ring *ring;
627 + struct brcmf_if *ifp;
628 u16 hash_idx;
629 + u8 ifidx;
630 struct sk_buff *skb;
631
632 ring = flow->rings[flowid];
633 if (!ring)
634 return;
635 +
636 + ifidx = brcmf_flowring_ifidx_get(flow, flowid);
637 + ifp = brcmf_get_ifp(bus_if->drvr, ifidx);
638 +
639 brcmf_flowring_block(flow, flowid, false);
640 hash_idx = ring->hash_id;
641 flow->hash[hash_idx].ifidx = BRCMF_FLOWRING_INVALID_IFIDX;
642 @@ -249,7 +256,7 @@ void brcmf_flowring_delete(struct brcmf_flowring *flow, u16 flowid)
643
644 skb = skb_dequeue(&ring->skblist);
645 while (skb) {
646 - brcmu_pkt_buf_free_skb(skb);
647 + brcmf_txfinalize(ifp, skb, false);
648 skb = skb_dequeue(&ring->skblist);
649 }
650
651 diff --git a/drivers/scsi/arcmsr/arcmsr_hba.c b/drivers/scsi/arcmsr/arcmsr_hba.c
652 index 7640498964a5..3d53d636b17b 100644
653 --- a/drivers/scsi/arcmsr/arcmsr_hba.c
654 +++ b/drivers/scsi/arcmsr/arcmsr_hba.c
655 @@ -2388,15 +2388,23 @@ static int arcmsr_iop_message_xfer(struct AdapterControlBlock *acb,
656 }
657 case ARCMSR_MESSAGE_WRITE_WQBUFFER: {
658 unsigned char *ver_addr;
659 - int32_t user_len, cnt2end;
660 + uint32_t user_len;
661 + int32_t cnt2end;
662 uint8_t *pQbuffer, *ptmpuserbuffer;
663 +
664 + user_len = pcmdmessagefld->cmdmessage.Length;
665 + if (user_len > ARCMSR_API_DATA_BUFLEN) {
666 + retvalue = ARCMSR_MESSAGE_FAIL;
667 + goto message_out;
668 + }
669 +
670 ver_addr = kmalloc(ARCMSR_API_DATA_BUFLEN, GFP_ATOMIC);
671 if (!ver_addr) {
672 retvalue = ARCMSR_MESSAGE_FAIL;
673 goto message_out;
674 }
675 ptmpuserbuffer = ver_addr;
676 - user_len = pcmdmessagefld->cmdmessage.Length;
677 +
678 memcpy(ptmpuserbuffer,
679 pcmdmessagefld->messagedatabuffer, user_len);
680 spin_lock_irqsave(&acb->wqbuffer_lock, flags);
681 diff --git a/drivers/scsi/ibmvscsi/ibmvfc.c b/drivers/scsi/ibmvscsi/ibmvfc.c
682 index ab67ec4b6bd6..79c9860a165f 100644
683 --- a/drivers/scsi/ibmvscsi/ibmvfc.c
684 +++ b/drivers/scsi/ibmvscsi/ibmvfc.c
685 @@ -717,7 +717,6 @@ static int ibmvfc_reset_crq(struct ibmvfc_host *vhost)
686 spin_lock_irqsave(vhost->host->host_lock, flags);
687 vhost->state = IBMVFC_NO_CRQ;
688 vhost->logged_in = 0;
689 - ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_NONE);
690
691 /* Clean out the queue */
692 memset(crq->msgs, 0, PAGE_SIZE);
693 diff --git a/drivers/tty/serial/8250/8250_dw.c b/drivers/tty/serial/8250/8250_dw.c
694 index e19969614203..b022f5a01e63 100644
695 --- a/drivers/tty/serial/8250/8250_dw.c
696 +++ b/drivers/tty/serial/8250/8250_dw.c
697 @@ -462,7 +462,7 @@ static int dw8250_probe(struct platform_device *pdev)
698 }
699
700 data->pclk = devm_clk_get(&pdev->dev, "apb_pclk");
701 - if (IS_ERR(data->clk) && PTR_ERR(data->clk) == -EPROBE_DEFER) {
702 + if (IS_ERR(data->pclk) && PTR_ERR(data->pclk) == -EPROBE_DEFER) {
703 err = -EPROBE_DEFER;
704 goto err_clk;
705 }
706 diff --git a/drivers/tty/serial/8250/8250_port.c b/drivers/tty/serial/8250/8250_port.c
707 index bdfa659b9606..858a54633664 100644
708 --- a/drivers/tty/serial/8250/8250_port.c
709 +++ b/drivers/tty/serial/8250/8250_port.c
710 @@ -1414,12 +1414,8 @@ static void __do_stop_tx_rs485(struct uart_8250_port *p)
711 if (!(p->port.rs485.flags & SER_RS485_RX_DURING_TX)) {
712 serial8250_clear_fifos(p);
713
714 - serial8250_rpm_get(p);
715 -
716 p->ier |= UART_IER_RLSI | UART_IER_RDI;
717 serial_port_out(&p->port, UART_IER, p->ier);
718 -
719 - serial8250_rpm_put(p);
720 }
721 }
722
723 @@ -1429,6 +1425,7 @@ static void serial8250_em485_handle_stop_tx(unsigned long arg)
724 struct uart_8250_em485 *em485 = p->em485;
725 unsigned long flags;
726
727 + serial8250_rpm_get(p);
728 spin_lock_irqsave(&p->port.lock, flags);
729 if (em485 &&
730 em485->active_timer == &em485->stop_tx_timer) {
731 @@ -1436,6 +1433,7 @@ static void serial8250_em485_handle_stop_tx(unsigned long arg)
732 em485->active_timer = NULL;
733 }
734 spin_unlock_irqrestore(&p->port.lock, flags);
735 + serial8250_rpm_put(p);
736 }
737
738 static void __stop_tx_rs485(struct uart_8250_port *p)
739 @@ -1475,7 +1473,7 @@ static inline void __stop_tx(struct uart_8250_port *p)
740 unsigned char lsr = serial_in(p, UART_LSR);
741 /*
742 * To provide required timeing and allow FIFO transfer,
743 - * __stop_tx_rs485 must be called only when both FIFO and
744 + * __stop_tx_rs485() must be called only when both FIFO and
745 * shift register are empty. It is for device driver to enable
746 * interrupt on TEMT.
747 */
748 @@ -1484,9 +1482,10 @@ static inline void __stop_tx(struct uart_8250_port *p)
749
750 del_timer(&em485->start_tx_timer);
751 em485->active_timer = NULL;
752 +
753 + __stop_tx_rs485(p);
754 }
755 __do_stop_tx(p);
756 - __stop_tx_rs485(p);
757 }
758
759 static void serial8250_stop_tx(struct uart_port *port)
760 diff --git a/drivers/tty/serial/atmel_serial.c b/drivers/tty/serial/atmel_serial.c
761 index 2eaa18ddef61..8bbde52db376 100644
762 --- a/drivers/tty/serial/atmel_serial.c
763 +++ b/drivers/tty/serial/atmel_serial.c
764 @@ -1929,6 +1929,9 @@ static void atmel_shutdown(struct uart_port *port)
765 {
766 struct atmel_uart_port *atmel_port = to_atmel_uart_port(port);
767
768 + /* Disable modem control lines interrupts */
769 + atmel_disable_ms(port);
770 +
771 /* Disable interrupts at device level */
772 atmel_uart_writel(port, ATMEL_US_IDR, -1);
773
774 @@ -1979,8 +1982,6 @@ static void atmel_shutdown(struct uart_port *port)
775 */
776 free_irq(port->irq, port);
777
778 - atmel_port->ms_irq_enabled = false;
779 -
780 atmel_flush_buffer(port);
781 }
782
783 diff --git a/drivers/tty/serial/imx.c b/drivers/tty/serial/imx.c
784 index 0df2b1c091ae..615c0279a1a6 100644
785 --- a/drivers/tty/serial/imx.c
786 +++ b/drivers/tty/serial/imx.c
787 @@ -740,12 +740,13 @@ static unsigned int imx_get_hwmctrl(struct imx_port *sport)
788 {
789 unsigned int tmp = TIOCM_DSR;
790 unsigned usr1 = readl(sport->port.membase + USR1);
791 + unsigned usr2 = readl(sport->port.membase + USR2);
792
793 if (usr1 & USR1_RTSS)
794 tmp |= TIOCM_CTS;
795
796 /* in DCE mode DCDIN is always 0 */
797 - if (!(usr1 & USR2_DCDIN))
798 + if (!(usr2 & USR2_DCDIN))
799 tmp |= TIOCM_CAR;
800
801 if (sport->dte_mode)
802 diff --git a/fs/attr.c b/fs/attr.c
803 index 42bb42bb3c72..3c42cab06b5d 100644
804 --- a/fs/attr.c
805 +++ b/fs/attr.c
806 @@ -202,6 +202,21 @@ int notify_change(struct dentry * dentry, struct iattr * attr, struct inode **de
807 return -EPERM;
808 }
809
810 + /*
811 + * If utimes(2) and friends are called with times == NULL (or both
812 + * times are UTIME_NOW), then we need to check for write permission
813 + */
814 + if (ia_valid & ATTR_TOUCH) {
815 + if (IS_IMMUTABLE(inode))
816 + return -EPERM;
817 +
818 + if (!inode_owner_or_capable(inode)) {
819 + error = inode_permission(inode, MAY_WRITE);
820 + if (error)
821 + return error;
822 + }
823 + }
824 +
825 if ((ia_valid & ATTR_MODE)) {
826 umode_t amode = attr->ia_mode;
827 /* Flag setting protected by i_mutex */
828 diff --git a/fs/autofs4/waitq.c b/fs/autofs4/waitq.c
829 index 431fd7ee3488..e44271dfceb6 100644
830 --- a/fs/autofs4/waitq.c
831 +++ b/fs/autofs4/waitq.c
832 @@ -431,8 +431,8 @@ int autofs4_wait(struct autofs_sb_info *sbi,
833 memcpy(&wq->name, &qstr, sizeof(struct qstr));
834 wq->dev = autofs4_get_dev(sbi);
835 wq->ino = autofs4_get_ino(sbi);
836 - wq->uid = current_uid();
837 - wq->gid = current_gid();
838 + wq->uid = current_real_cred()->uid;
839 + wq->gid = current_real_cred()->gid;
840 wq->pid = pid;
841 wq->tgid = tgid;
842 wq->status = -EINTR; /* Status return if interrupted */
843 diff --git a/fs/btrfs/compression.c b/fs/btrfs/compression.c
844 index 029db6e1105c..60a850ee8c78 100644
845 --- a/fs/btrfs/compression.c
846 +++ b/fs/btrfs/compression.c
847 @@ -698,7 +698,7 @@ int btrfs_submit_compressed_read(struct inode *inode, struct bio *bio,
848
849 ret = btrfs_map_bio(root, comp_bio, mirror_num, 0);
850 if (ret) {
851 - bio->bi_error = ret;
852 + comp_bio->bi_error = ret;
853 bio_endio(comp_bio);
854 }
855
856 @@ -728,7 +728,7 @@ int btrfs_submit_compressed_read(struct inode *inode, struct bio *bio,
857
858 ret = btrfs_map_bio(root, comp_bio, mirror_num, 0);
859 if (ret) {
860 - bio->bi_error = ret;
861 + comp_bio->bi_error = ret;
862 bio_endio(comp_bio);
863 }
864
865 diff --git a/fs/btrfs/ctree.h b/fs/btrfs/ctree.h
866 index 33fe03551105..791e47ce9d27 100644
867 --- a/fs/btrfs/ctree.h
868 +++ b/fs/btrfs/ctree.h
869 @@ -251,7 +251,8 @@ struct btrfs_super_block {
870 #define BTRFS_FEATURE_COMPAT_SAFE_CLEAR 0ULL
871
872 #define BTRFS_FEATURE_COMPAT_RO_SUPP \
873 - (BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE)
874 + (BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE | \
875 + BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID)
876
877 #define BTRFS_FEATURE_COMPAT_RO_SAFE_SET 0ULL
878 #define BTRFS_FEATURE_COMPAT_RO_SAFE_CLEAR 0ULL
879 diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c
880 index 54bc8c7c6bcd..3dede6d53bad 100644
881 --- a/fs/btrfs/disk-io.c
882 +++ b/fs/btrfs/disk-io.c
883 @@ -2566,6 +2566,7 @@ int open_ctree(struct super_block *sb,
884 int num_backups_tried = 0;
885 int backup_index = 0;
886 int max_active;
887 + int clear_free_space_tree = 0;
888
889 tree_root = fs_info->tree_root = btrfs_alloc_root(fs_info, GFP_KERNEL);
890 chunk_root = fs_info->chunk_root = btrfs_alloc_root(fs_info, GFP_KERNEL);
891 @@ -3129,6 +3130,26 @@ retry_root_backup:
892 if (sb->s_flags & MS_RDONLY)
893 return 0;
894
895 + if (btrfs_test_opt(fs_info, CLEAR_CACHE) &&
896 + btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
897 + clear_free_space_tree = 1;
898 + } else if (btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE) &&
899 + !btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID)) {
900 + btrfs_warn(fs_info, "free space tree is invalid");
901 + clear_free_space_tree = 1;
902 + }
903 +
904 + if (clear_free_space_tree) {
905 + btrfs_info(fs_info, "clearing free space tree");
906 + ret = btrfs_clear_free_space_tree(fs_info);
907 + if (ret) {
908 + btrfs_warn(fs_info,
909 + "failed to clear free space tree: %d", ret);
910 + close_ctree(tree_root);
911 + return ret;
912 + }
913 + }
914 +
915 if (btrfs_test_opt(tree_root->fs_info, FREE_SPACE_TREE) &&
916 !btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
917 btrfs_info(fs_info, "creating free space tree");
918 @@ -3166,18 +3187,6 @@ retry_root_backup:
919
920 btrfs_qgroup_rescan_resume(fs_info);
921
922 - if (btrfs_test_opt(tree_root->fs_info, CLEAR_CACHE) &&
923 - btrfs_fs_compat_ro(fs_info, FREE_SPACE_TREE)) {
924 - btrfs_info(fs_info, "clearing free space tree");
925 - ret = btrfs_clear_free_space_tree(fs_info);
926 - if (ret) {
927 - btrfs_warn(fs_info,
928 - "failed to clear free space tree: %d", ret);
929 - close_ctree(tree_root);
930 - return ret;
931 - }
932 - }
933 -
934 if (!fs_info->uuid_root) {
935 btrfs_info(fs_info, "creating UUID tree");
936 ret = btrfs_create_uuid_tree(fs_info);
937 diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c
938 index 44fe66b53c8b..c3ec30dea9a5 100644
939 --- a/fs/btrfs/extent_io.c
940 +++ b/fs/btrfs/extent_io.c
941 @@ -5524,17 +5524,45 @@ void copy_extent_buffer(struct extent_buffer *dst, struct extent_buffer *src,
942 }
943 }
944
945 -/*
946 - * The extent buffer bitmap operations are done with byte granularity because
947 - * bitmap items are not guaranteed to be aligned to a word and therefore a
948 - * single word in a bitmap may straddle two pages in the extent buffer.
949 - */
950 -#define BIT_BYTE(nr) ((nr) / BITS_PER_BYTE)
951 -#define BYTE_MASK ((1 << BITS_PER_BYTE) - 1)
952 -#define BITMAP_FIRST_BYTE_MASK(start) \
953 - ((BYTE_MASK << ((start) & (BITS_PER_BYTE - 1))) & BYTE_MASK)
954 -#define BITMAP_LAST_BYTE_MASK(nbits) \
955 - (BYTE_MASK >> (-(nbits) & (BITS_PER_BYTE - 1)))
956 +void le_bitmap_set(u8 *map, unsigned int start, int len)
957 +{
958 + u8 *p = map + BIT_BYTE(start);
959 + const unsigned int size = start + len;
960 + int bits_to_set = BITS_PER_BYTE - (start % BITS_PER_BYTE);
961 + u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(start);
962 +
963 + while (len - bits_to_set >= 0) {
964 + *p |= mask_to_set;
965 + len -= bits_to_set;
966 + bits_to_set = BITS_PER_BYTE;
967 + mask_to_set = ~(u8)0;
968 + p++;
969 + }
970 + if (len) {
971 + mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
972 + *p |= mask_to_set;
973 + }
974 +}
975 +
976 +void le_bitmap_clear(u8 *map, unsigned int start, int len)
977 +{
978 + u8 *p = map + BIT_BYTE(start);
979 + const unsigned int size = start + len;
980 + int bits_to_clear = BITS_PER_BYTE - (start % BITS_PER_BYTE);
981 + u8 mask_to_clear = BITMAP_FIRST_BYTE_MASK(start);
982 +
983 + while (len - bits_to_clear >= 0) {
984 + *p &= ~mask_to_clear;
985 + len -= bits_to_clear;
986 + bits_to_clear = BITS_PER_BYTE;
987 + mask_to_clear = ~(u8)0;
988 + p++;
989 + }
990 + if (len) {
991 + mask_to_clear &= BITMAP_LAST_BYTE_MASK(size);
992 + *p &= ~mask_to_clear;
993 + }
994 +}
995
996 /*
997 * eb_bitmap_offset() - calculate the page and offset of the byte containing the
998 @@ -5578,7 +5606,7 @@ static inline void eb_bitmap_offset(struct extent_buffer *eb,
999 int extent_buffer_test_bit(struct extent_buffer *eb, unsigned long start,
1000 unsigned long nr)
1001 {
1002 - char *kaddr;
1003 + u8 *kaddr;
1004 struct page *page;
1005 unsigned long i;
1006 size_t offset;
1007 @@ -5600,13 +5628,13 @@ int extent_buffer_test_bit(struct extent_buffer *eb, unsigned long start,
1008 void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
1009 unsigned long pos, unsigned long len)
1010 {
1011 - char *kaddr;
1012 + u8 *kaddr;
1013 struct page *page;
1014 unsigned long i;
1015 size_t offset;
1016 const unsigned int size = pos + len;
1017 int bits_to_set = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
1018 - unsigned int mask_to_set = BITMAP_FIRST_BYTE_MASK(pos);
1019 + u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(pos);
1020
1021 eb_bitmap_offset(eb, start, pos, &i, &offset);
1022 page = eb->pages[i];
1023 @@ -5617,7 +5645,7 @@ void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
1024 kaddr[offset] |= mask_to_set;
1025 len -= bits_to_set;
1026 bits_to_set = BITS_PER_BYTE;
1027 - mask_to_set = ~0U;
1028 + mask_to_set = ~(u8)0;
1029 if (++offset >= PAGE_SIZE && len > 0) {
1030 offset = 0;
1031 page = eb->pages[++i];
1032 @@ -5642,13 +5670,13 @@ void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
1033 void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
1034 unsigned long pos, unsigned long len)
1035 {
1036 - char *kaddr;
1037 + u8 *kaddr;
1038 struct page *page;
1039 unsigned long i;
1040 size_t offset;
1041 const unsigned int size = pos + len;
1042 int bits_to_clear = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
1043 - unsigned int mask_to_clear = BITMAP_FIRST_BYTE_MASK(pos);
1044 + u8 mask_to_clear = BITMAP_FIRST_BYTE_MASK(pos);
1045
1046 eb_bitmap_offset(eb, start, pos, &i, &offset);
1047 page = eb->pages[i];
1048 @@ -5659,7 +5687,7 @@ void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
1049 kaddr[offset] &= ~mask_to_clear;
1050 len -= bits_to_clear;
1051 bits_to_clear = BITS_PER_BYTE;
1052 - mask_to_clear = ~0U;
1053 + mask_to_clear = ~(u8)0;
1054 if (++offset >= PAGE_SIZE && len > 0) {
1055 offset = 0;
1056 page = eb->pages[++i];
1057 diff --git a/fs/btrfs/extent_io.h b/fs/btrfs/extent_io.h
1058 index 28cd88fccc7e..1cf4e4226fc8 100644
1059 --- a/fs/btrfs/extent_io.h
1060 +++ b/fs/btrfs/extent_io.h
1061 @@ -59,6 +59,28 @@
1062 */
1063 #define EXTENT_PAGE_PRIVATE 1
1064
1065 +/*
1066 + * The extent buffer bitmap operations are done with byte granularity instead of
1067 + * word granularity for two reasons:
1068 + * 1. The bitmaps must be little-endian on disk.
1069 + * 2. Bitmap items are not guaranteed to be aligned to a word and therefore a
1070 + * single word in a bitmap may straddle two pages in the extent buffer.
1071 + */
1072 +#define BIT_BYTE(nr) ((nr) / BITS_PER_BYTE)
1073 +#define BYTE_MASK ((1 << BITS_PER_BYTE) - 1)
1074 +#define BITMAP_FIRST_BYTE_MASK(start) \
1075 + ((BYTE_MASK << ((start) & (BITS_PER_BYTE - 1))) & BYTE_MASK)
1076 +#define BITMAP_LAST_BYTE_MASK(nbits) \
1077 + (BYTE_MASK >> (-(nbits) & (BITS_PER_BYTE - 1)))
1078 +
1079 +static inline int le_test_bit(int nr, const u8 *addr)
1080 +{
1081 + return 1U & (addr[BIT_BYTE(nr)] >> (nr & (BITS_PER_BYTE-1)));
1082 +}
1083 +
1084 +extern void le_bitmap_set(u8 *map, unsigned int start, int len);
1085 +extern void le_bitmap_clear(u8 *map, unsigned int start, int len);
1086 +
1087 struct extent_state;
1088 struct btrfs_root;
1089 struct btrfs_io_bio;
1090 diff --git a/fs/btrfs/free-space-tree.c b/fs/btrfs/free-space-tree.c
1091 index 87e7e3d3e676..ea605ffd0e03 100644
1092 --- a/fs/btrfs/free-space-tree.c
1093 +++ b/fs/btrfs/free-space-tree.c
1094 @@ -151,7 +151,7 @@ static inline u32 free_space_bitmap_size(u64 size, u32 sectorsize)
1095 return DIV_ROUND_UP((u32)div_u64(size, sectorsize), BITS_PER_BYTE);
1096 }
1097
1098 -static unsigned long *alloc_bitmap(u32 bitmap_size)
1099 +static u8 *alloc_bitmap(u32 bitmap_size)
1100 {
1101 void *mem;
1102
1103 @@ -180,8 +180,7 @@ int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
1104 struct btrfs_free_space_info *info;
1105 struct btrfs_key key, found_key;
1106 struct extent_buffer *leaf;
1107 - unsigned long *bitmap;
1108 - char *bitmap_cursor;
1109 + u8 *bitmap, *bitmap_cursor;
1110 u64 start, end;
1111 u64 bitmap_range, i;
1112 u32 bitmap_size, flags, expected_extent_count;
1113 @@ -231,7 +230,7 @@ int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
1114 block_group->sectorsize);
1115 last = div_u64(found_key.objectid + found_key.offset - start,
1116 block_group->sectorsize);
1117 - bitmap_set(bitmap, first, last - first);
1118 + le_bitmap_set(bitmap, first, last - first);
1119
1120 extent_count++;
1121 nr++;
1122 @@ -269,7 +268,7 @@ int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
1123 goto out;
1124 }
1125
1126 - bitmap_cursor = (char *)bitmap;
1127 + bitmap_cursor = bitmap;
1128 bitmap_range = block_group->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
1129 i = start;
1130 while (i < end) {
1131 @@ -318,7 +317,7 @@ int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
1132 struct btrfs_free_space_info *info;
1133 struct btrfs_key key, found_key;
1134 struct extent_buffer *leaf;
1135 - unsigned long *bitmap;
1136 + u8 *bitmap;
1137 u64 start, end;
1138 /* Initialize to silence GCC. */
1139 u64 extent_start = 0;
1140 @@ -362,7 +361,7 @@ int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
1141 break;
1142 } else if (found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
1143 unsigned long ptr;
1144 - char *bitmap_cursor;
1145 + u8 *bitmap_cursor;
1146 u32 bitmap_pos, data_size;
1147
1148 ASSERT(found_key.objectid >= start);
1149 @@ -372,7 +371,7 @@ int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
1150 bitmap_pos = div_u64(found_key.objectid - start,
1151 block_group->sectorsize *
1152 BITS_PER_BYTE);
1153 - bitmap_cursor = ((char *)bitmap) + bitmap_pos;
1154 + bitmap_cursor = bitmap + bitmap_pos;
1155 data_size = free_space_bitmap_size(found_key.offset,
1156 block_group->sectorsize);
1157
1158 @@ -409,7 +408,7 @@ int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
1159 offset = start;
1160 bitnr = 0;
1161 while (offset < end) {
1162 - bit = !!test_bit(bitnr, bitmap);
1163 + bit = !!le_test_bit(bitnr, bitmap);
1164 if (prev_bit == 0 && bit == 1) {
1165 extent_start = offset;
1166 } else if (prev_bit == 1 && bit == 0) {
1167 @@ -1183,6 +1182,7 @@ int btrfs_create_free_space_tree(struct btrfs_fs_info *fs_info)
1168 }
1169
1170 btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE);
1171 + btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID);
1172 fs_info->creating_free_space_tree = 0;
1173
1174 ret = btrfs_commit_transaction(trans, tree_root);
1175 @@ -1251,6 +1251,7 @@ int btrfs_clear_free_space_tree(struct btrfs_fs_info *fs_info)
1176 return PTR_ERR(trans);
1177
1178 btrfs_clear_fs_compat_ro(fs_info, FREE_SPACE_TREE);
1179 + btrfs_clear_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID);
1180 fs_info->free_space_root = NULL;
1181
1182 ret = clear_free_space_tree(trans, free_space_root);
1183 diff --git a/fs/cachefiles/interface.c b/fs/cachefiles/interface.c
1184 index ce5f345d70f5..e7f16a77a22a 100644
1185 --- a/fs/cachefiles/interface.c
1186 +++ b/fs/cachefiles/interface.c
1187 @@ -253,6 +253,8 @@ static void cachefiles_drop_object(struct fscache_object *_object)
1188 struct cachefiles_object *object;
1189 struct cachefiles_cache *cache;
1190 const struct cred *saved_cred;
1191 + struct inode *inode;
1192 + blkcnt_t i_blocks = 0;
1193
1194 ASSERT(_object);
1195
1196 @@ -279,6 +281,10 @@ static void cachefiles_drop_object(struct fscache_object *_object)
1197 _object != cache->cache.fsdef
1198 ) {
1199 _debug("- retire object OBJ%x", object->fscache.debug_id);
1200 + inode = d_backing_inode(object->dentry);
1201 + if (inode)
1202 + i_blocks = inode->i_blocks;
1203 +
1204 cachefiles_begin_secure(cache, &saved_cred);
1205 cachefiles_delete_object(cache, object);
1206 cachefiles_end_secure(cache, saved_cred);
1207 @@ -292,7 +298,7 @@ static void cachefiles_drop_object(struct fscache_object *_object)
1208
1209 /* note that the object is now inactive */
1210 if (test_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags))
1211 - cachefiles_mark_object_inactive(cache, object);
1212 + cachefiles_mark_object_inactive(cache, object, i_blocks);
1213
1214 dput(object->dentry);
1215 object->dentry = NULL;
1216 diff --git a/fs/cachefiles/internal.h b/fs/cachefiles/internal.h
1217 index 2fcde1a34b7c..cd1effee8a49 100644
1218 --- a/fs/cachefiles/internal.h
1219 +++ b/fs/cachefiles/internal.h
1220 @@ -160,7 +160,8 @@ extern char *cachefiles_cook_key(const u8 *raw, int keylen, uint8_t type);
1221 * namei.c
1222 */
1223 extern void cachefiles_mark_object_inactive(struct cachefiles_cache *cache,
1224 - struct cachefiles_object *object);
1225 + struct cachefiles_object *object,
1226 + blkcnt_t i_blocks);
1227 extern int cachefiles_delete_object(struct cachefiles_cache *cache,
1228 struct cachefiles_object *object);
1229 extern int cachefiles_walk_to_object(struct cachefiles_object *parent,
1230 diff --git a/fs/cachefiles/namei.c b/fs/cachefiles/namei.c
1231 index 3f7c2cd41f8f..c6ee4b5fb7e6 100644
1232 --- a/fs/cachefiles/namei.c
1233 +++ b/fs/cachefiles/namei.c
1234 @@ -261,10 +261,9 @@ requeue:
1235 * Mark an object as being inactive.
1236 */
1237 void cachefiles_mark_object_inactive(struct cachefiles_cache *cache,
1238 - struct cachefiles_object *object)
1239 + struct cachefiles_object *object,
1240 + blkcnt_t i_blocks)
1241 {
1242 - blkcnt_t i_blocks = d_backing_inode(object->dentry)->i_blocks;
1243 -
1244 write_lock(&cache->active_lock);
1245 rb_erase(&object->active_node, &cache->active_nodes);
1246 clear_bit(CACHEFILES_OBJECT_ACTIVE, &object->flags);
1247 @@ -707,7 +706,8 @@ mark_active_timed_out:
1248
1249 check_error:
1250 _debug("check error %d", ret);
1251 - cachefiles_mark_object_inactive(cache, object);
1252 + cachefiles_mark_object_inactive(
1253 + cache, object, d_backing_inode(object->dentry)->i_blocks);
1254 release_dentry:
1255 dput(object->dentry);
1256 object->dentry = NULL;
1257 diff --git a/fs/debugfs/file.c b/fs/debugfs/file.c
1258 index 592059f88e04..309f4e9b2419 100644
1259 --- a/fs/debugfs/file.c
1260 +++ b/fs/debugfs/file.c
1261 @@ -97,9 +97,6 @@ EXPORT_SYMBOL_GPL(debugfs_use_file_finish);
1262
1263 #define F_DENTRY(filp) ((filp)->f_path.dentry)
1264
1265 -#define REAL_FOPS_DEREF(dentry) \
1266 - ((const struct file_operations *)(dentry)->d_fsdata)
1267 -
1268 static int open_proxy_open(struct inode *inode, struct file *filp)
1269 {
1270 const struct dentry *dentry = F_DENTRY(filp);
1271 @@ -112,7 +109,7 @@ static int open_proxy_open(struct inode *inode, struct file *filp)
1272 goto out;
1273 }
1274
1275 - real_fops = REAL_FOPS_DEREF(dentry);
1276 + real_fops = debugfs_real_fops(filp);
1277 real_fops = fops_get(real_fops);
1278 if (!real_fops) {
1279 /* Huh? Module did not clean up after itself at exit? */
1280 @@ -143,7 +140,7 @@ static ret_type full_proxy_ ## name(proto) \
1281 { \
1282 const struct dentry *dentry = F_DENTRY(filp); \
1283 const struct file_operations *real_fops = \
1284 - REAL_FOPS_DEREF(dentry); \
1285 + debugfs_real_fops(filp); \
1286 int srcu_idx; \
1287 ret_type r; \
1288 \
1289 @@ -176,7 +173,7 @@ static unsigned int full_proxy_poll(struct file *filp,
1290 struct poll_table_struct *wait)
1291 {
1292 const struct dentry *dentry = F_DENTRY(filp);
1293 - const struct file_operations *real_fops = REAL_FOPS_DEREF(dentry);
1294 + const struct file_operations *real_fops = debugfs_real_fops(filp);
1295 int srcu_idx;
1296 unsigned int r = 0;
1297
1298 @@ -193,7 +190,7 @@ static unsigned int full_proxy_poll(struct file *filp,
1299 static int full_proxy_release(struct inode *inode, struct file *filp)
1300 {
1301 const struct dentry *dentry = F_DENTRY(filp);
1302 - const struct file_operations *real_fops = REAL_FOPS_DEREF(dentry);
1303 + const struct file_operations *real_fops = debugfs_real_fops(filp);
1304 const struct file_operations *proxy_fops = filp->f_op;
1305 int r = 0;
1306
1307 @@ -241,7 +238,7 @@ static int full_proxy_open(struct inode *inode, struct file *filp)
1308 goto out;
1309 }
1310
1311 - real_fops = REAL_FOPS_DEREF(dentry);
1312 + real_fops = debugfs_real_fops(filp);
1313 real_fops = fops_get(real_fops);
1314 if (!real_fops) {
1315 /* Huh? Module did not cleanup after itself at exit? */
1316 diff --git a/fs/dlm/lowcomms.c b/fs/dlm/lowcomms.c
1317 index 963016c8f3d1..609998de533e 100644
1318 --- a/fs/dlm/lowcomms.c
1319 +++ b/fs/dlm/lowcomms.c
1320 @@ -1656,16 +1656,12 @@ void dlm_lowcomms_stop(void)
1321 mutex_lock(&connections_lock);
1322 dlm_allow_conn = 0;
1323 foreach_conn(stop_conn);
1324 + clean_writequeues();
1325 + foreach_conn(free_conn);
1326 mutex_unlock(&connections_lock);
1327
1328 work_stop();
1329
1330 - mutex_lock(&connections_lock);
1331 - clean_writequeues();
1332 -
1333 - foreach_conn(free_conn);
1334 -
1335 - mutex_unlock(&connections_lock);
1336 kmem_cache_destroy(con_cache);
1337 }
1338
1339 diff --git a/fs/ext4/extents.c b/fs/ext4/extents.c
1340 index d7ccb7f51dfc..7f69347bd5a5 100644
1341 --- a/fs/ext4/extents.c
1342 +++ b/fs/ext4/extents.c
1343 @@ -5734,6 +5734,9 @@ int ext4_insert_range(struct inode *inode, loff_t offset, loff_t len)
1344 up_write(&EXT4_I(inode)->i_data_sem);
1345 goto out_stop;
1346 }
1347 + } else {
1348 + ext4_ext_drop_refs(path);
1349 + kfree(path);
1350 }
1351
1352 ret = ext4_es_remove_extent(inode, offset_lblk,
1353 diff --git a/fs/ext4/inode.c b/fs/ext4/inode.c
1354 index c6ea25a190f8..f4cdc647ecfc 100644
1355 --- a/fs/ext4/inode.c
1356 +++ b/fs/ext4/inode.c
1357 @@ -647,11 +647,19 @@ found:
1358 /*
1359 * We have to zeroout blocks before inserting them into extent
1360 * status tree. Otherwise someone could look them up there and
1361 - * use them before they are really zeroed.
1362 + * use them before they are really zeroed. We also have to
1363 + * unmap metadata before zeroing as otherwise writeback can
1364 + * overwrite zeros with stale data from block device.
1365 */
1366 if (flags & EXT4_GET_BLOCKS_ZERO &&
1367 map->m_flags & EXT4_MAP_MAPPED &&
1368 map->m_flags & EXT4_MAP_NEW) {
1369 + ext4_lblk_t i;
1370 +
1371 + for (i = 0; i < map->m_len; i++) {
1372 + unmap_underlying_metadata(inode->i_sb->s_bdev,
1373 + map->m_pblk + i);
1374 + }
1375 ret = ext4_issue_zeroout(inode, map->m_lblk,
1376 map->m_pblk, map->m_len);
1377 if (ret) {
1378 @@ -1649,6 +1657,8 @@ static void mpage_release_unused_pages(struct mpage_da_data *mpd,
1379 BUG_ON(!PageLocked(page));
1380 BUG_ON(PageWriteback(page));
1381 if (invalidate) {
1382 + if (page_mapped(page))
1383 + clear_page_dirty_for_io(page);
1384 block_invalidatepage(page, 0, PAGE_SIZE);
1385 ClearPageUptodate(page);
1386 }
1387 @@ -3890,7 +3900,7 @@ int ext4_update_disksize_before_punch(struct inode *inode, loff_t offset,
1388 }
1389
1390 /*
1391 - * ext4_punch_hole: punches a hole in a file by releaseing the blocks
1392 + * ext4_punch_hole: punches a hole in a file by releasing the blocks
1393 * associated with the given offset and length
1394 *
1395 * @inode: File inode
1396 @@ -3919,7 +3929,7 @@ int ext4_punch_hole(struct inode *inode, loff_t offset, loff_t length)
1397 * Write out all dirty pages to avoid race conditions
1398 * Then release them.
1399 */
1400 - if (mapping->nrpages && mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) {
1401 + if (mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) {
1402 ret = filemap_write_and_wait_range(mapping, offset,
1403 offset + length - 1);
1404 if (ret)
1405 @@ -4814,14 +4824,14 @@ static int ext4_do_update_inode(handle_t *handle,
1406 * Fix up interoperability with old kernels. Otherwise, old inodes get
1407 * re-used with the upper 16 bits of the uid/gid intact
1408 */
1409 - if (!ei->i_dtime) {
1410 + if (ei->i_dtime && list_empty(&ei->i_orphan)) {
1411 + raw_inode->i_uid_high = 0;
1412 + raw_inode->i_gid_high = 0;
1413 + } else {
1414 raw_inode->i_uid_high =
1415 cpu_to_le16(high_16_bits(i_uid));
1416 raw_inode->i_gid_high =
1417 cpu_to_le16(high_16_bits(i_gid));
1418 - } else {
1419 - raw_inode->i_uid_high = 0;
1420 - raw_inode->i_gid_high = 0;
1421 }
1422 } else {
1423 raw_inode->i_uid_low = cpu_to_le16(fs_high2lowuid(i_uid));
1424 diff --git a/fs/ext4/move_extent.c b/fs/ext4/move_extent.c
1425 index a920c5d29fac..6fc14def0c70 100644
1426 --- a/fs/ext4/move_extent.c
1427 +++ b/fs/ext4/move_extent.c
1428 @@ -598,6 +598,13 @@ ext4_move_extents(struct file *o_filp, struct file *d_filp, __u64 orig_blk,
1429 return -EOPNOTSUPP;
1430 }
1431
1432 + if (ext4_encrypted_inode(orig_inode) ||
1433 + ext4_encrypted_inode(donor_inode)) {
1434 + ext4_msg(orig_inode->i_sb, KERN_ERR,
1435 + "Online defrag not supported for encrypted files");
1436 + return -EOPNOTSUPP;
1437 + }
1438 +
1439 /* Protect orig and donor inodes against a truncate */
1440 lock_two_nondirectories(orig_inode, donor_inode);
1441
1442 diff --git a/fs/ext4/namei.c b/fs/ext4/namei.c
1443 index 34c0142caf6a..7e2f8c3c11ce 100644
1444 --- a/fs/ext4/namei.c
1445 +++ b/fs/ext4/namei.c
1446 @@ -2044,33 +2044,31 @@ static int make_indexed_dir(handle_t *handle, struct ext4_filename *fname,
1447 frame->entries = entries;
1448 frame->at = entries;
1449 frame->bh = bh;
1450 - bh = bh2;
1451
1452 retval = ext4_handle_dirty_dx_node(handle, dir, frame->bh);
1453 if (retval)
1454 goto out_frames;
1455 - retval = ext4_handle_dirty_dirent_node(handle, dir, bh);
1456 + retval = ext4_handle_dirty_dirent_node(handle, dir, bh2);
1457 if (retval)
1458 goto out_frames;
1459
1460 - de = do_split(handle,dir, &bh, frame, &fname->hinfo);
1461 + de = do_split(handle,dir, &bh2, frame, &fname->hinfo);
1462 if (IS_ERR(de)) {
1463 retval = PTR_ERR(de);
1464 goto out_frames;
1465 }
1466 - dx_release(frames);
1467
1468 - retval = add_dirent_to_buf(handle, fname, dir, inode, de, bh);
1469 - brelse(bh);
1470 - return retval;
1471 + retval = add_dirent_to_buf(handle, fname, dir, inode, de, bh2);
1472 out_frames:
1473 /*
1474 * Even if the block split failed, we have to properly write
1475 * out all the changes we did so far. Otherwise we can end up
1476 * with corrupted filesystem.
1477 */
1478 - ext4_mark_inode_dirty(handle, dir);
1479 + if (retval)
1480 + ext4_mark_inode_dirty(handle, dir);
1481 dx_release(frames);
1482 + brelse(bh2);
1483 return retval;
1484 }
1485
1486 diff --git a/fs/ext4/symlink.c b/fs/ext4/symlink.c
1487 index 4d83d9e05f2e..04a7850a0d45 100644
1488 --- a/fs/ext4/symlink.c
1489 +++ b/fs/ext4/symlink.c
1490 @@ -65,13 +65,12 @@ static const char *ext4_encrypted_get_link(struct dentry *dentry,
1491 res = fscrypt_fname_alloc_buffer(inode, cstr.len, &pstr);
1492 if (res)
1493 goto errout;
1494 + paddr = pstr.name;
1495
1496 res = fscrypt_fname_disk_to_usr(inode, 0, 0, &cstr, &pstr);
1497 if (res < 0)
1498 goto errout;
1499
1500 - paddr = pstr.name;
1501 -
1502 /* Null-terminate the name */
1503 if (res <= pstr.len)
1504 paddr[res] = '\0';
1505 diff --git a/fs/fuse/dir.c b/fs/fuse/dir.c
1506 index c47b7780ce37..4ff9251e9d3a 100644
1507 --- a/fs/fuse/dir.c
1508 +++ b/fs/fuse/dir.c
1509 @@ -1702,14 +1702,46 @@ error:
1510 static int fuse_setattr(struct dentry *entry, struct iattr *attr)
1511 {
1512 struct inode *inode = d_inode(entry);
1513 + struct file *file = (attr->ia_valid & ATTR_FILE) ? attr->ia_file : NULL;
1514 + int ret;
1515
1516 if (!fuse_allow_current_process(get_fuse_conn(inode)))
1517 return -EACCES;
1518
1519 - if (attr->ia_valid & ATTR_FILE)
1520 - return fuse_do_setattr(inode, attr, attr->ia_file);
1521 - else
1522 - return fuse_do_setattr(inode, attr, NULL);
1523 + if (attr->ia_valid & (ATTR_KILL_SUID | ATTR_KILL_SGID)) {
1524 + int kill;
1525 +
1526 + attr->ia_valid &= ~(ATTR_KILL_SUID | ATTR_KILL_SGID |
1527 + ATTR_MODE);
1528 + /*
1529 + * ia_mode calculation may have used stale i_mode. Refresh and
1530 + * recalculate.
1531 + */
1532 + ret = fuse_do_getattr(inode, NULL, file);
1533 + if (ret)
1534 + return ret;
1535 +
1536 + attr->ia_mode = inode->i_mode;
1537 + kill = should_remove_suid(entry);
1538 + if (kill & ATTR_KILL_SUID) {
1539 + attr->ia_valid |= ATTR_MODE;
1540 + attr->ia_mode &= ~S_ISUID;
1541 + }
1542 + if (kill & ATTR_KILL_SGID) {
1543 + attr->ia_valid |= ATTR_MODE;
1544 + attr->ia_mode &= ~S_ISGID;
1545 + }
1546 + }
1547 + if (!attr->ia_valid)
1548 + return 0;
1549 +
1550 + ret = fuse_do_setattr(inode, attr, file);
1551 + if (!ret) {
1552 + /* Directory mode changed, may need to revalidate access */
1553 + if (d_is_dir(entry) && (attr->ia_valid & ATTR_MODE))
1554 + fuse_invalidate_entry_cache(entry);
1555 + }
1556 + return ret;
1557 }
1558
1559 static int fuse_getattr(struct vfsmount *mnt, struct dentry *entry,
1560 @@ -1801,6 +1833,23 @@ static ssize_t fuse_getxattr(struct dentry *entry, struct inode *inode,
1561 return ret;
1562 }
1563
1564 +static int fuse_verify_xattr_list(char *list, size_t size)
1565 +{
1566 + size_t origsize = size;
1567 +
1568 + while (size) {
1569 + size_t thislen = strnlen(list, size);
1570 +
1571 + if (!thislen || thislen == size)
1572 + return -EIO;
1573 +
1574 + size -= thislen + 1;
1575 + list += thislen + 1;
1576 + }
1577 +
1578 + return origsize;
1579 +}
1580 +
1581 static ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size)
1582 {
1583 struct inode *inode = d_inode(entry);
1584 @@ -1836,6 +1885,8 @@ static ssize_t fuse_listxattr(struct dentry *entry, char *list, size_t size)
1585 ret = fuse_simple_request(fc, &args);
1586 if (!ret && !size)
1587 ret = outarg.size;
1588 + if (ret > 0 && size)
1589 + ret = fuse_verify_xattr_list(list, ret);
1590 if (ret == -ENOSYS) {
1591 fc->no_listxattr = 1;
1592 ret = -EOPNOTSUPP;
1593 diff --git a/fs/jbd2/transaction.c b/fs/jbd2/transaction.c
1594 index b5bc3e249163..3d8246a9faa4 100644
1595 --- a/fs/jbd2/transaction.c
1596 +++ b/fs/jbd2/transaction.c
1597 @@ -159,6 +159,7 @@ static void wait_transaction_locked(journal_t *journal)
1598 read_unlock(&journal->j_state_lock);
1599 if (need_to_start)
1600 jbd2_log_start_commit(journal, tid);
1601 + jbd2_might_wait_for_commit(journal);
1602 schedule();
1603 finish_wait(&journal->j_wait_transaction_locked, &wait);
1604 }
1605 @@ -182,8 +183,6 @@ static int add_transaction_credits(journal_t *journal, int blocks,
1606 int needed;
1607 int total = blocks + rsv_blocks;
1608
1609 - jbd2_might_wait_for_commit(journal);
1610 -
1611 /*
1612 * If the current transaction is locked down for commit, wait
1613 * for the lock to be released.
1614 @@ -214,6 +213,7 @@ static int add_transaction_credits(journal_t *journal, int blocks,
1615 if (atomic_read(&journal->j_reserved_credits) + total >
1616 journal->j_max_transaction_buffers) {
1617 read_unlock(&journal->j_state_lock);
1618 + jbd2_might_wait_for_commit(journal);
1619 wait_event(journal->j_wait_reserved,
1620 atomic_read(&journal->j_reserved_credits) + total <=
1621 journal->j_max_transaction_buffers);
1622 @@ -238,6 +238,7 @@ static int add_transaction_credits(journal_t *journal, int blocks,
1623 if (jbd2_log_space_left(journal) < jbd2_space_needed(journal)) {
1624 atomic_sub(total, &t->t_outstanding_credits);
1625 read_unlock(&journal->j_state_lock);
1626 + jbd2_might_wait_for_commit(journal);
1627 write_lock(&journal->j_state_lock);
1628 if (jbd2_log_space_left(journal) < jbd2_space_needed(journal))
1629 __jbd2_log_wait_for_space(journal);
1630 @@ -255,6 +256,7 @@ static int add_transaction_credits(journal_t *journal, int blocks,
1631 sub_reserved_credits(journal, rsv_blocks);
1632 atomic_sub(total, &t->t_outstanding_credits);
1633 read_unlock(&journal->j_state_lock);
1634 + jbd2_might_wait_for_commit(journal);
1635 wait_event(journal->j_wait_reserved,
1636 atomic_read(&journal->j_reserved_credits) + rsv_blocks
1637 <= journal->j_max_transaction_buffers / 2);
1638 diff --git a/fs/reiserfs/super.c b/fs/reiserfs/super.c
1639 index 7a4a85a6821e..74d5ddd26296 100644
1640 --- a/fs/reiserfs/super.c
1641 +++ b/fs/reiserfs/super.c
1642 @@ -190,7 +190,15 @@ static int remove_save_link_only(struct super_block *s,
1643 static int reiserfs_quota_on_mount(struct super_block *, int);
1644 #endif
1645
1646 -/* look for uncompleted unlinks and truncates and complete them */
1647 +/*
1648 + * Look for uncompleted unlinks and truncates and complete them
1649 + *
1650 + * Called with superblock write locked. If quotas are enabled, we have to
1651 + * release/retake lest we call dquot_quota_on_mount(), proceed to
1652 + * schedule_on_each_cpu() in invalidate_bdev() and deadlock waiting for the per
1653 + * cpu worklets to complete flush_async_commits() that in turn wait for the
1654 + * superblock write lock.
1655 + */
1656 static int finish_unfinished(struct super_block *s)
1657 {
1658 INITIALIZE_PATH(path);
1659 @@ -237,7 +245,9 @@ static int finish_unfinished(struct super_block *s)
1660 quota_enabled[i] = 0;
1661 continue;
1662 }
1663 + reiserfs_write_unlock(s);
1664 ret = reiserfs_quota_on_mount(s, i);
1665 + reiserfs_write_lock(s);
1666 if (ret < 0)
1667 reiserfs_warning(s, "reiserfs-2500",
1668 "cannot turn on journaled "
1669 diff --git a/fs/utimes.c b/fs/utimes.c
1670 index 794f5f5b1fb5..ba54b9e648c9 100644
1671 --- a/fs/utimes.c
1672 +++ b/fs/utimes.c
1673 @@ -87,21 +87,7 @@ static int utimes_common(struct path *path, struct timespec *times)
1674 */
1675 newattrs.ia_valid |= ATTR_TIMES_SET;
1676 } else {
1677 - /*
1678 - * If times is NULL (or both times are UTIME_NOW),
1679 - * then we need to check permissions, because
1680 - * inode_change_ok() won't do it.
1681 - */
1682 - error = -EPERM;
1683 - if (IS_IMMUTABLE(inode))
1684 - goto mnt_drop_write_and_out;
1685 -
1686 - error = -EACCES;
1687 - if (!inode_owner_or_capable(inode)) {
1688 - error = inode_permission(inode, MAY_WRITE);
1689 - if (error)
1690 - goto mnt_drop_write_and_out;
1691 - }
1692 + newattrs.ia_valid |= ATTR_TOUCH;
1693 }
1694 retry_deleg:
1695 inode_lock(inode);
1696 @@ -113,7 +99,6 @@ retry_deleg:
1697 goto retry_deleg;
1698 }
1699
1700 -mnt_drop_write_and_out:
1701 mnt_drop_write(path->mnt);
1702 out:
1703 return error;
1704 diff --git a/include/crypto/ghash.h b/include/crypto/ghash.h
1705 new file mode 100644
1706 index 000000000000..2a61c9bbab8f
1707 --- /dev/null
1708 +++ b/include/crypto/ghash.h
1709 @@ -0,0 +1,23 @@
1710 +/*
1711 + * Common values for GHASH algorithms
1712 + */
1713 +
1714 +#ifndef __CRYPTO_GHASH_H__
1715 +#define __CRYPTO_GHASH_H__
1716 +
1717 +#include <linux/types.h>
1718 +#include <crypto/gf128mul.h>
1719 +
1720 +#define GHASH_BLOCK_SIZE 16
1721 +#define GHASH_DIGEST_SIZE 16
1722 +
1723 +struct ghash_ctx {
1724 + struct gf128mul_4k *gf128;
1725 +};
1726 +
1727 +struct ghash_desc_ctx {
1728 + u8 buffer[GHASH_BLOCK_SIZE];
1729 + u32 bytes;
1730 +};
1731 +
1732 +#endif
1733 diff --git a/include/linux/debugfs.h b/include/linux/debugfs.h
1734 index 1438e2322d5c..4d3f0d1aec73 100644
1735 --- a/include/linux/debugfs.h
1736 +++ b/include/linux/debugfs.h
1737 @@ -45,6 +45,23 @@ extern struct dentry *arch_debugfs_dir;
1738
1739 extern struct srcu_struct debugfs_srcu;
1740
1741 +/**
1742 + * debugfs_real_fops - getter for the real file operation
1743 + * @filp: a pointer to a struct file
1744 + *
1745 + * Must only be called under the protection established by
1746 + * debugfs_use_file_start().
1747 + */
1748 +static inline const struct file_operations *debugfs_real_fops(struct file *filp)
1749 + __must_hold(&debugfs_srcu)
1750 +{
1751 + /*
1752 + * Neither the pointer to the struct file_operations, nor its
1753 + * contents ever change -- srcu_dereference() is not needed here.
1754 + */
1755 + return filp->f_path.dentry->d_fsdata;
1756 +}
1757 +
1758 #if defined(CONFIG_DEBUG_FS)
1759
1760 struct dentry *debugfs_create_file(const char *name, umode_t mode,
1761 diff --git a/include/linux/fs.h b/include/linux/fs.h
1762 index 901e25d495cc..7c391366fb43 100644
1763 --- a/include/linux/fs.h
1764 +++ b/include/linux/fs.h
1765 @@ -224,6 +224,7 @@ typedef int (dio_iodone_t)(struct kiocb *iocb, loff_t offset,
1766 #define ATTR_KILL_PRIV (1 << 14)
1767 #define ATTR_OPEN (1 << 15) /* Truncating from open(O_TRUNC) */
1768 #define ATTR_TIMES_SET (1 << 16)
1769 +#define ATTR_TOUCH (1 << 17)
1770
1771 /*
1772 * Whiteout is represented by a char device. The following constants define the
1773 diff --git a/include/linux/radix-tree.h b/include/linux/radix-tree.h
1774 index 4c45105dece3..52b97db93830 100644
1775 --- a/include/linux/radix-tree.h
1776 +++ b/include/linux/radix-tree.h
1777 @@ -280,9 +280,9 @@ bool __radix_tree_delete_node(struct radix_tree_root *root,
1778 struct radix_tree_node *node);
1779 void *radix_tree_delete_item(struct radix_tree_root *, unsigned long, void *);
1780 void *radix_tree_delete(struct radix_tree_root *, unsigned long);
1781 -struct radix_tree_node *radix_tree_replace_clear_tags(
1782 - struct radix_tree_root *root,
1783 - unsigned long index, void *entry);
1784 +void radix_tree_clear_tags(struct radix_tree_root *root,
1785 + struct radix_tree_node *node,
1786 + void **slot);
1787 unsigned int radix_tree_gang_lookup(struct radix_tree_root *root,
1788 void **results, unsigned long first_index,
1789 unsigned int max_items);
1790 diff --git a/include/linux/sem.h b/include/linux/sem.h
1791 index 976ce3a19f1b..d0efd6e6c20a 100644
1792 --- a/include/linux/sem.h
1793 +++ b/include/linux/sem.h
1794 @@ -21,6 +21,7 @@ struct sem_array {
1795 struct list_head list_id; /* undo requests on this array */
1796 int sem_nsems; /* no. of semaphores in array */
1797 int complex_count; /* pending complex operations */
1798 + bool complex_mode; /* no parallel simple ops */
1799 };
1800
1801 #ifdef CONFIG_SYSVIPC
1802 diff --git a/include/uapi/linux/btrfs.h b/include/uapi/linux/btrfs.h
1803 index ac5eacd3055b..db4c253f8011 100644
1804 --- a/include/uapi/linux/btrfs.h
1805 +++ b/include/uapi/linux/btrfs.h
1806 @@ -239,7 +239,17 @@ struct btrfs_ioctl_fs_info_args {
1807 * Used by:
1808 * struct btrfs_ioctl_feature_flags
1809 */
1810 -#define BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE (1ULL << 0)
1811 +#define BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE (1ULL << 0)
1812 +/*
1813 + * Older kernels (< 4.9) on big-endian systems produced broken free space tree
1814 + * bitmaps, and btrfs-progs also used to corrupt the free space tree (versions
1815 + * < 4.7.3). If this bit is clear, then the free space tree cannot be trusted.
1816 + * btrfs-progs can also intentionally clear this bit to ask the kernel to
1817 + * rebuild the free space tree, however this might not work on older kernels
1818 + * that do not know about this bit. If not sure, clear the cache manually on
1819 + * first mount when booting older kernel versions.
1820 + */
1821 +#define BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID (1ULL << 1)
1822
1823 #define BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF (1ULL << 0)
1824 #define BTRFS_FEATURE_INCOMPAT_DEFAULT_SUBVOL (1ULL << 1)
1825 diff --git a/ipc/sem.c b/ipc/sem.c
1826 index 7c9d4f7683c0..5e318c5f749d 100644
1827 --- a/ipc/sem.c
1828 +++ b/ipc/sem.c
1829 @@ -162,14 +162,21 @@ static int sysvipc_sem_proc_show(struct seq_file *s, void *it);
1830
1831 /*
1832 * Locking:
1833 + * a) global sem_lock() for read/write
1834 * sem_undo.id_next,
1835 * sem_array.complex_count,
1836 - * sem_array.pending{_alter,_cont},
1837 - * sem_array.sem_undo: global sem_lock() for read/write
1838 - * sem_undo.proc_next: only "current" is allowed to read/write that field.
1839 + * sem_array.complex_mode
1840 + * sem_array.pending{_alter,_const},
1841 + * sem_array.sem_undo
1842 *
1843 + * b) global or semaphore sem_lock() for read/write:
1844 * sem_array.sem_base[i].pending_{const,alter}:
1845 - * global or semaphore sem_lock() for read/write
1846 + * sem_array.complex_mode (for read)
1847 + *
1848 + * c) special:
1849 + * sem_undo_list.list_proc:
1850 + * * undo_list->lock for write
1851 + * * rcu for read
1852 */
1853
1854 #define sc_semmsl sem_ctls[0]
1855 @@ -260,30 +267,61 @@ static void sem_rcu_free(struct rcu_head *head)
1856 }
1857
1858 /*
1859 - * Wait until all currently ongoing simple ops have completed.
1860 + * Enter the mode suitable for non-simple operations:
1861 * Caller must own sem_perm.lock.
1862 - * New simple ops cannot start, because simple ops first check
1863 - * that sem_perm.lock is free.
1864 - * that a) sem_perm.lock is free and b) complex_count is 0.
1865 */
1866 -static void sem_wait_array(struct sem_array *sma)
1867 +static void complexmode_enter(struct sem_array *sma)
1868 {
1869 int i;
1870 struct sem *sem;
1871
1872 - if (sma->complex_count) {
1873 - /* The thread that increased sma->complex_count waited on
1874 - * all sem->lock locks. Thus we don't need to wait again.
1875 - */
1876 + if (sma->complex_mode) {
1877 + /* We are already in complex_mode. Nothing to do */
1878 return;
1879 }
1880
1881 + /* We need a full barrier after seting complex_mode:
1882 + * The write to complex_mode must be visible
1883 + * before we read the first sem->lock spinlock state.
1884 + */
1885 + smp_store_mb(sma->complex_mode, true);
1886 +
1887 for (i = 0; i < sma->sem_nsems; i++) {
1888 sem = sma->sem_base + i;
1889 spin_unlock_wait(&sem->lock);
1890 }
1891 + /*
1892 + * spin_unlock_wait() is not a memory barriers, it is only a
1893 + * control barrier. The code must pair with spin_unlock(&sem->lock),
1894 + * thus just the control barrier is insufficient.
1895 + *
1896 + * smp_rmb() is sufficient, as writes cannot pass the control barrier.
1897 + */
1898 + smp_rmb();
1899 +}
1900 +
1901 +/*
1902 + * Try to leave the mode that disallows simple operations:
1903 + * Caller must own sem_perm.lock.
1904 + */
1905 +static void complexmode_tryleave(struct sem_array *sma)
1906 +{
1907 + if (sma->complex_count) {
1908 + /* Complex ops are sleeping.
1909 + * We must stay in complex mode
1910 + */
1911 + return;
1912 + }
1913 + /*
1914 + * Immediately after setting complex_mode to false,
1915 + * a simple op can start. Thus: all memory writes
1916 + * performed by the current operation must be visible
1917 + * before we set complex_mode to false.
1918 + */
1919 + smp_store_release(&sma->complex_mode, false);
1920 }
1921
1922 +#define SEM_GLOBAL_LOCK (-1)
1923 /*
1924 * If the request contains only one semaphore operation, and there are
1925 * no complex transactions pending, lock only the semaphore involved.
1926 @@ -300,56 +338,42 @@ static inline int sem_lock(struct sem_array *sma, struct sembuf *sops,
1927 /* Complex operation - acquire a full lock */
1928 ipc_lock_object(&sma->sem_perm);
1929
1930 - /* And wait until all simple ops that are processed
1931 - * right now have dropped their locks.
1932 - */
1933 - sem_wait_array(sma);
1934 - return -1;
1935 + /* Prevent parallel simple ops */
1936 + complexmode_enter(sma);
1937 + return SEM_GLOBAL_LOCK;
1938 }
1939
1940 /*
1941 * Only one semaphore affected - try to optimize locking.
1942 - * The rules are:
1943 - * - optimized locking is possible if no complex operation
1944 - * is either enqueued or processed right now.
1945 - * - The test for enqueued complex ops is simple:
1946 - * sma->complex_count != 0
1947 - * - Testing for complex ops that are processed right now is
1948 - * a bit more difficult. Complex ops acquire the full lock
1949 - * and first wait that the running simple ops have completed.
1950 - * (see above)
1951 - * Thus: If we own a simple lock and the global lock is free
1952 - * and complex_count is now 0, then it will stay 0 and
1953 - * thus just locking sem->lock is sufficient.
1954 + * Optimized locking is possible if no complex operation
1955 + * is either enqueued or processed right now.
1956 + *
1957 + * Both facts are tracked by complex_mode.
1958 */
1959 sem = sma->sem_base + sops->sem_num;
1960
1961 - if (sma->complex_count == 0) {
1962 + /*
1963 + * Initial check for complex_mode. Just an optimization,
1964 + * no locking, no memory barrier.
1965 + */
1966 + if (!sma->complex_mode) {
1967 /*
1968 * It appears that no complex operation is around.
1969 * Acquire the per-semaphore lock.
1970 */
1971 spin_lock(&sem->lock);
1972
1973 - /* Then check that the global lock is free */
1974 - if (!spin_is_locked(&sma->sem_perm.lock)) {
1975 - /*
1976 - * We need a memory barrier with acquire semantics,
1977 - * otherwise we can race with another thread that does:
1978 - * complex_count++;
1979 - * spin_unlock(sem_perm.lock);
1980 - */
1981 - smp_acquire__after_ctrl_dep();
1982 + /*
1983 + * See 51d7d5205d33
1984 + * ("powerpc: Add smp_mb() to arch_spin_is_locked()"):
1985 + * A full barrier is required: the write of sem->lock
1986 + * must be visible before the read is executed
1987 + */
1988 + smp_mb();
1989
1990 - /*
1991 - * Now repeat the test of complex_count:
1992 - * It can't change anymore until we drop sem->lock.
1993 - * Thus: if is now 0, then it will stay 0.
1994 - */
1995 - if (sma->complex_count == 0) {
1996 - /* fast path successful! */
1997 - return sops->sem_num;
1998 - }
1999 + if (!smp_load_acquire(&sma->complex_mode)) {
2000 + /* fast path successful! */
2001 + return sops->sem_num;
2002 }
2003 spin_unlock(&sem->lock);
2004 }
2005 @@ -369,15 +393,16 @@ static inline int sem_lock(struct sem_array *sma, struct sembuf *sops,
2006 /* Not a false alarm, thus complete the sequence for a
2007 * full lock.
2008 */
2009 - sem_wait_array(sma);
2010 - return -1;
2011 + complexmode_enter(sma);
2012 + return SEM_GLOBAL_LOCK;
2013 }
2014 }
2015
2016 static inline void sem_unlock(struct sem_array *sma, int locknum)
2017 {
2018 - if (locknum == -1) {
2019 + if (locknum == SEM_GLOBAL_LOCK) {
2020 unmerge_queues(sma);
2021 + complexmode_tryleave(sma);
2022 ipc_unlock_object(&sma->sem_perm);
2023 } else {
2024 struct sem *sem = sma->sem_base + locknum;
2025 @@ -529,6 +554,7 @@ static int newary(struct ipc_namespace *ns, struct ipc_params *params)
2026 }
2027
2028 sma->complex_count = 0;
2029 + sma->complex_mode = true; /* dropped by sem_unlock below */
2030 INIT_LIST_HEAD(&sma->pending_alter);
2031 INIT_LIST_HEAD(&sma->pending_const);
2032 INIT_LIST_HEAD(&sma->list_id);
2033 @@ -2184,10 +2210,10 @@ static int sysvipc_sem_proc_show(struct seq_file *s, void *it)
2034 /*
2035 * The proc interface isn't aware of sem_lock(), it calls
2036 * ipc_lock_object() directly (in sysvipc_find_ipc).
2037 - * In order to stay compatible with sem_lock(), we must wait until
2038 - * all simple semop() calls have left their critical regions.
2039 + * In order to stay compatible with sem_lock(), we must
2040 + * enter / leave complex_mode.
2041 */
2042 - sem_wait_array(sma);
2043 + complexmode_enter(sma);
2044
2045 sem_otime = get_semotime(sma);
2046
2047 @@ -2204,6 +2230,8 @@ static int sysvipc_sem_proc_show(struct seq_file *s, void *it)
2048 sem_otime,
2049 sma->sem_ctime);
2050
2051 + complexmode_tryleave(sma);
2052 +
2053 return 0;
2054 }
2055 #endif
2056 diff --git a/lib/radix-tree.c b/lib/radix-tree.c
2057 index 91f0727e3cad..8e6d552c40dd 100644
2058 --- a/lib/radix-tree.c
2059 +++ b/lib/radix-tree.c
2060 @@ -1583,15 +1583,10 @@ void *radix_tree_delete(struct radix_tree_root *root, unsigned long index)
2061 }
2062 EXPORT_SYMBOL(radix_tree_delete);
2063
2064 -struct radix_tree_node *radix_tree_replace_clear_tags(
2065 - struct radix_tree_root *root,
2066 - unsigned long index, void *entry)
2067 +void radix_tree_clear_tags(struct radix_tree_root *root,
2068 + struct radix_tree_node *node,
2069 + void **slot)
2070 {
2071 - struct radix_tree_node *node;
2072 - void **slot;
2073 -
2074 - __radix_tree_lookup(root, index, &node, &slot);
2075 -
2076 if (node) {
2077 unsigned int tag, offset = get_slot_offset(node, slot);
2078 for (tag = 0; tag < RADIX_TREE_MAX_TAGS; tag++)
2079 @@ -1600,9 +1595,6 @@ struct radix_tree_node *radix_tree_replace_clear_tags(
2080 /* Clear root node tags */
2081 root->gfp_mask &= __GFP_BITS_MASK;
2082 }
2083 -
2084 - radix_tree_replace_slot(slot, entry);
2085 - return node;
2086 }
2087
2088 /**
2089 diff --git a/mm/filemap.c b/mm/filemap.c
2090 index 2d0986a64f1f..ced9ef6c06b0 100644
2091 --- a/mm/filemap.c
2092 +++ b/mm/filemap.c
2093 @@ -169,33 +169,35 @@ static int page_cache_tree_insert(struct address_space *mapping,
2094 static void page_cache_tree_delete(struct address_space *mapping,
2095 struct page *page, void *shadow)
2096 {
2097 - struct radix_tree_node *node;
2098 int i, nr = PageHuge(page) ? 1 : hpage_nr_pages(page);
2099
2100 VM_BUG_ON_PAGE(!PageLocked(page), page);
2101 VM_BUG_ON_PAGE(PageTail(page), page);
2102 VM_BUG_ON_PAGE(nr != 1 && shadow, page);
2103
2104 - if (shadow) {
2105 - mapping->nrexceptional += nr;
2106 - /*
2107 - * Make sure the nrexceptional update is committed before
2108 - * the nrpages update so that final truncate racing
2109 - * with reclaim does not see both counters 0 at the
2110 - * same time and miss a shadow entry.
2111 - */
2112 - smp_wmb();
2113 - }
2114 - mapping->nrpages -= nr;
2115 -
2116 for (i = 0; i < nr; i++) {
2117 - node = radix_tree_replace_clear_tags(&mapping->page_tree,
2118 - page->index + i, shadow);
2119 + struct radix_tree_node *node;
2120 + void **slot;
2121 +
2122 + __radix_tree_lookup(&mapping->page_tree, page->index + i,
2123 + &node, &slot);
2124 +
2125 + radix_tree_clear_tags(&mapping->page_tree, node, slot);
2126 +
2127 if (!node) {
2128 VM_BUG_ON_PAGE(nr != 1, page);
2129 - return;
2130 + /*
2131 + * We need a node to properly account shadow
2132 + * entries. Don't plant any without. XXX
2133 + */
2134 + shadow = NULL;
2135 }
2136
2137 + radix_tree_replace_slot(slot, shadow);
2138 +
2139 + if (!node)
2140 + break;
2141 +
2142 workingset_node_pages_dec(node);
2143 if (shadow)
2144 workingset_node_shadows_inc(node);
2145 @@ -219,6 +221,18 @@ static void page_cache_tree_delete(struct address_space *mapping,
2146 &node->private_list);
2147 }
2148 }
2149 +
2150 + if (shadow) {
2151 + mapping->nrexceptional += nr;
2152 + /*
2153 + * Make sure the nrexceptional update is committed before
2154 + * the nrpages update so that final truncate racing
2155 + * with reclaim does not see both counters 0 at the
2156 + * same time and miss a shadow entry.
2157 + */
2158 + smp_wmb();
2159 + }
2160 + mapping->nrpages -= nr;
2161 }
2162
2163 /*
2164 @@ -619,7 +633,6 @@ int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask)
2165 __delete_from_page_cache(old, NULL);
2166 error = page_cache_tree_insert(mapping, new, NULL);
2167 BUG_ON(error);
2168 - mapping->nrpages++;
2169
2170 /*
2171 * hugetlb pages do not participate in page cache accounting.
2172 @@ -1674,6 +1687,10 @@ static ssize_t do_generic_file_read(struct file *filp, loff_t *ppos,
2173 unsigned int prev_offset;
2174 int error = 0;
2175
2176 + if (unlikely(*ppos >= inode->i_sb->s_maxbytes))
2177 + return -EINVAL;
2178 + iov_iter_truncate(iter, inode->i_sb->s_maxbytes);
2179 +
2180 index = *ppos >> PAGE_SHIFT;
2181 prev_index = ra->prev_pos >> PAGE_SHIFT;
2182 prev_offset = ra->prev_pos & (PAGE_SIZE-1);
2183 diff --git a/mm/hugetlb.c b/mm/hugetlb.c
2184 index 87e11d8ad536..603bdd01ec2c 100644
2185 --- a/mm/hugetlb.c
2186 +++ b/mm/hugetlb.c
2187 @@ -1443,13 +1443,14 @@ static void dissolve_free_huge_page(struct page *page)
2188 {
2189 spin_lock(&hugetlb_lock);
2190 if (PageHuge(page) && !page_count(page)) {
2191 - struct hstate *h = page_hstate(page);
2192 - int nid = page_to_nid(page);
2193 - list_del(&page->lru);
2194 + struct page *head = compound_head(page);
2195 + struct hstate *h = page_hstate(head);
2196 + int nid = page_to_nid(head);
2197 + list_del(&head->lru);
2198 h->free_huge_pages--;
2199 h->free_huge_pages_node[nid]--;
2200 h->max_huge_pages--;
2201 - update_and_free_page(h, page);
2202 + update_and_free_page(h, head);
2203 }
2204 spin_unlock(&hugetlb_lock);
2205 }
2206 @@ -1457,7 +1458,8 @@ static void dissolve_free_huge_page(struct page *page)
2207 /*
2208 * Dissolve free hugepages in a given pfn range. Used by memory hotplug to
2209 * make specified memory blocks removable from the system.
2210 - * Note that start_pfn should aligned with (minimum) hugepage size.
2211 + * Note that this will dissolve a free gigantic hugepage completely, if any
2212 + * part of it lies within the given range.
2213 */
2214 void dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn)
2215 {
2216 @@ -1466,7 +1468,6 @@ void dissolve_free_huge_pages(unsigned long start_pfn, unsigned long end_pfn)
2217 if (!hugepages_supported())
2218 return;
2219
2220 - VM_BUG_ON(!IS_ALIGNED(start_pfn, 1 << minimum_order));
2221 for (pfn = start_pfn; pfn < end_pfn; pfn += 1 << minimum_order)
2222 dissolve_free_huge_page(pfn_to_page(pfn));
2223 }
2224 diff --git a/sound/soc/codecs/nau8825.c b/sound/soc/codecs/nau8825.c
2225 index 2e59a85e360b..ff566376da40 100644
2226 --- a/sound/soc/codecs/nau8825.c
2227 +++ b/sound/soc/codecs/nau8825.c
2228 @@ -1907,7 +1907,7 @@ static int nau8825_calc_fll_param(unsigned int fll_in, unsigned int fs,
2229 /* Calculate the FLL 10-bit integer input and the FLL 16-bit fractional
2230 * input based on FDCO, FREF and FLL ratio.
2231 */
2232 - fvco = div_u64(fvco << 16, fref * fll_param->ratio);
2233 + fvco = div_u64(fvco_max << 16, fref * fll_param->ratio);
2234 fll_param->fll_int = (fvco >> 16) & 0x3FF;
2235 fll_param->fll_frac = fvco & 0xFFFF;
2236 return 0;
2237 diff --git a/sound/soc/intel/atom/sst/sst_pvt.c b/sound/soc/intel/atom/sst/sst_pvt.c
2238 index adb32fefd693..b1e6b8f34a6a 100644
2239 --- a/sound/soc/intel/atom/sst/sst_pvt.c
2240 +++ b/sound/soc/intel/atom/sst/sst_pvt.c
2241 @@ -279,17 +279,15 @@ int sst_prepare_and_post_msg(struct intel_sst_drv *sst,
2242
2243 if (response) {
2244 ret = sst_wait_timeout(sst, block);
2245 - if (ret < 0) {
2246 + if (ret < 0)
2247 goto out;
2248 - } else if(block->data) {
2249 - if (!data)
2250 - goto out;
2251 - *data = kzalloc(block->size, GFP_KERNEL);
2252 - if (!(*data)) {
2253 +
2254 + if (data && block->data) {
2255 + *data = kmemdup(block->data, block->size, GFP_KERNEL);
2256 + if (!*data) {
2257 ret = -ENOMEM;
2258 goto out;
2259 - } else
2260 - memcpy(data, (void *) block->data, block->size);
2261 + }
2262 }
2263 }
2264 out: