Magellan Linux

Annotation of /trunk/kernel-alx/patches-3.4/0114-3.4.15-all-fixes.patch

Parent Directory Parent Directory | Revision Log Revision Log


Revision 1946 - (hide annotations) (download)
Wed Nov 14 15:25:09 2012 UTC (11 years, 6 months ago) by niro
File size: 81054 byte(s)
3.4.18-alx-r1
1 niro 1946 diff --git a/arch/arm/Kconfig b/arch/arm/Kconfig
2     index 352322a..e14ae11 100644
3     --- a/arch/arm/Kconfig
4     +++ b/arch/arm/Kconfig
5     @@ -1405,6 +1405,16 @@ config PL310_ERRATA_769419
6     on systems with an outer cache, the store buffer is drained
7     explicitly.
8    
9     +config ARM_ERRATA_775420
10     + bool "ARM errata: A data cache maintenance operation which aborts, might lead to deadlock"
11     + depends on CPU_V7
12     + help
13     + This option enables the workaround for the 775420 Cortex-A9 (r2p2,
14     + r2p6,r2p8,r2p10,r3p0) erratum. In case a date cache maintenance
15     + operation aborts with MMU exception, it might cause the processor
16     + to deadlock. This workaround puts DSB before executing ISB if
17     + an abort may occur on cache maintenance.
18     +
19     endmenu
20    
21     source "arch/arm/common/Kconfig"
22     diff --git a/arch/arm/include/asm/vfpmacros.h b/arch/arm/include/asm/vfpmacros.h
23     index 3d5fc41..bf53047 100644
24     --- a/arch/arm/include/asm/vfpmacros.h
25     +++ b/arch/arm/include/asm/vfpmacros.h
26     @@ -28,7 +28,7 @@
27     ldr \tmp, =elf_hwcap @ may not have MVFR regs
28     ldr \tmp, [\tmp, #0]
29     tst \tmp, #HWCAP_VFPv3D16
30     - ldceq p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31}
31     + ldceql p11, cr0, [\base],#32*4 @ FLDMIAD \base!, {d16-d31}
32     addne \base, \base, #32*4 @ step over unused register space
33     #else
34     VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0
35     @@ -52,7 +52,7 @@
36     ldr \tmp, =elf_hwcap @ may not have MVFR regs
37     ldr \tmp, [\tmp, #0]
38     tst \tmp, #HWCAP_VFPv3D16
39     - stceq p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31}
40     + stceql p11, cr0, [\base],#32*4 @ FSTMIAD \base!, {d16-d31}
41     addne \base, \base, #32*4 @ step over unused register space
42     #else
43     VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0
44     diff --git a/arch/arm/mm/cache-v7.S b/arch/arm/mm/cache-v7.S
45     index a655d3d..82ab2c5 100644
46     --- a/arch/arm/mm/cache-v7.S
47     +++ b/arch/arm/mm/cache-v7.S
48     @@ -211,6 +211,9 @@ ENTRY(v7_coherent_user_range)
49     * isn't mapped, just try the next page.
50     */
51     9001:
52     +#ifdef CONFIG_ARM_ERRATA_775420
53     + dsb
54     +#endif
55     mov r12, r12, lsr #12
56     mov r12, r12, lsl #12
57     add r12, r12, #4096
58     diff --git a/arch/mips/kernel/kgdb.c b/arch/mips/kernel/kgdb.c
59     index f4546e9..23817a6 100644
60     --- a/arch/mips/kernel/kgdb.c
61     +++ b/arch/mips/kernel/kgdb.c
62     @@ -283,6 +283,15 @@ static int kgdb_mips_notify(struct notifier_block *self, unsigned long cmd,
63     struct pt_regs *regs = args->regs;
64     int trap = (regs->cp0_cause & 0x7c) >> 2;
65    
66     +#ifdef CONFIG_KPROBES
67     + /*
68     + * Return immediately if the kprobes fault notifier has set
69     + * DIE_PAGE_FAULT.
70     + */
71     + if (cmd == DIE_PAGE_FAULT)
72     + return NOTIFY_DONE;
73     +#endif /* CONFIG_KPROBES */
74     +
75     /* Userspace events, ignore. */
76     if (user_mode(regs))
77     return NOTIFY_DONE;
78     diff --git a/arch/x86/xen/enlighten.c b/arch/x86/xen/enlighten.c
79     index 40edfc3..5910054 100644
80     --- a/arch/x86/xen/enlighten.c
81     +++ b/arch/x86/xen/enlighten.c
82     @@ -942,7 +942,16 @@ static void xen_write_cr4(unsigned long cr4)
83    
84     native_write_cr4(cr4);
85     }
86     -
87     +#ifdef CONFIG_X86_64
88     +static inline unsigned long xen_read_cr8(void)
89     +{
90     + return 0;
91     +}
92     +static inline void xen_write_cr8(unsigned long val)
93     +{
94     + BUG_ON(val);
95     +}
96     +#endif
97     static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high)
98     {
99     int ret;
100     @@ -1111,6 +1120,11 @@ static const struct pv_cpu_ops xen_cpu_ops __initconst = {
101     .read_cr4_safe = native_read_cr4_safe,
102     .write_cr4 = xen_write_cr4,
103    
104     +#ifdef CONFIG_X86_64
105     + .read_cr8 = xen_read_cr8,
106     + .write_cr8 = xen_write_cr8,
107     +#endif
108     +
109     .wbinvd = native_wbinvd,
110    
111     .read_msr = native_read_msr_safe,
112     @@ -1121,6 +1135,8 @@ static const struct pv_cpu_ops xen_cpu_ops __initconst = {
113     .read_tsc = native_read_tsc,
114     .read_pmc = native_read_pmc,
115    
116     + .read_tscp = native_read_tscp,
117     +
118     .iret = xen_iret,
119     .irq_enable_sysexit = xen_sysexit,
120     #ifdef CONFIG_X86_64
121     diff --git a/block/blk-core.c b/block/blk-core.c
122     index 1f61b74..85fd410 100644
123     --- a/block/blk-core.c
124     +++ b/block/blk-core.c
125     @@ -601,7 +601,7 @@ blk_init_allocated_queue(struct request_queue *q, request_fn_proc *rfn,
126     q->request_fn = rfn;
127     q->prep_rq_fn = NULL;
128     q->unprep_rq_fn = NULL;
129     - q->queue_flags = QUEUE_FLAG_DEFAULT;
130     + q->queue_flags |= QUEUE_FLAG_DEFAULT;
131    
132     /* Override internal queue lock with supplied lock pointer */
133     if (lock)
134     diff --git a/drivers/acpi/ec.c b/drivers/acpi/ec.c
135     index 7edaccc..a51df96 100644
136     --- a/drivers/acpi/ec.c
137     +++ b/drivers/acpi/ec.c
138     @@ -71,9 +71,6 @@ enum ec_command {
139     #define ACPI_EC_UDELAY_GLK 1000 /* Wait 1ms max. to get global lock */
140     #define ACPI_EC_MSI_UDELAY 550 /* Wait 550us for MSI EC */
141    
142     -#define ACPI_EC_STORM_THRESHOLD 8 /* number of false interrupts
143     - per one transaction */
144     -
145     enum {
146     EC_FLAGS_QUERY_PENDING, /* Query is pending */
147     EC_FLAGS_GPE_STORM, /* GPE storm detected */
148     @@ -87,6 +84,15 @@ static unsigned int ec_delay __read_mostly = ACPI_EC_DELAY;
149     module_param(ec_delay, uint, 0644);
150     MODULE_PARM_DESC(ec_delay, "Timeout(ms) waited until an EC command completes");
151    
152     +/*
153     + * If the number of false interrupts per one transaction exceeds
154     + * this threshold, will think there is a GPE storm happened and
155     + * will disable the GPE for normal transaction.
156     + */
157     +static unsigned int ec_storm_threshold __read_mostly = 8;
158     +module_param(ec_storm_threshold, uint, 0644);
159     +MODULE_PARM_DESC(ec_storm_threshold, "Maxim false GPE numbers not considered as GPE storm");
160     +
161     /* If we find an EC via the ECDT, we need to keep a ptr to its context */
162     /* External interfaces use first EC only, so remember */
163     typedef int (*acpi_ec_query_func) (void *data);
164     @@ -319,7 +325,7 @@ static int acpi_ec_transaction(struct acpi_ec *ec, struct transaction *t)
165     msleep(1);
166     /* It is safe to enable the GPE outside of the transaction. */
167     acpi_enable_gpe(NULL, ec->gpe);
168     - } else if (t->irq_count > ACPI_EC_STORM_THRESHOLD) {
169     + } else if (t->irq_count > ec_storm_threshold) {
170     pr_info(PREFIX "GPE storm detected, "
171     "transactions will use polling mode\n");
172     set_bit(EC_FLAGS_GPE_STORM, &ec->flags);
173     @@ -924,6 +930,17 @@ static int ec_flag_msi(const struct dmi_system_id *id)
174     return 0;
175     }
176    
177     +/*
178     + * Clevo M720 notebook actually works ok with IRQ mode, if we lifted
179     + * the GPE storm threshold back to 20
180     + */
181     +static int ec_enlarge_storm_threshold(const struct dmi_system_id *id)
182     +{
183     + pr_debug("Setting the EC GPE storm threshold to 20\n");
184     + ec_storm_threshold = 20;
185     + return 0;
186     +}
187     +
188     static struct dmi_system_id __initdata ec_dmi_table[] = {
189     {
190     ec_skip_dsdt_scan, "Compal JFL92", {
191     @@ -955,10 +972,13 @@ static struct dmi_system_id __initdata ec_dmi_table[] = {
192     {
193     ec_validate_ecdt, "ASUS hardware", {
194     DMI_MATCH(DMI_BOARD_VENDOR, "ASUSTeK Computer Inc.") }, NULL},
195     + {
196     + ec_enlarge_storm_threshold, "CLEVO hardware", {
197     + DMI_MATCH(DMI_SYS_VENDOR, "CLEVO Co."),
198     + DMI_MATCH(DMI_PRODUCT_NAME, "M720T/M730T"),}, NULL},
199     {},
200     };
201    
202     -
203     int __init acpi_ec_ecdt_probe(void)
204     {
205     acpi_status status;
206     diff --git a/drivers/char/tpm/tpm.c b/drivers/char/tpm/tpm.c
207     index 08427ab..27f8ddf 100644
208     --- a/drivers/char/tpm/tpm.c
209     +++ b/drivers/char/tpm/tpm.c
210     @@ -1186,17 +1186,20 @@ ssize_t tpm_write(struct file *file, const char __user *buf,
211     size_t size, loff_t *off)
212     {
213     struct tpm_chip *chip = file->private_data;
214     - size_t in_size = size, out_size;
215     + size_t in_size = size;
216     + ssize_t out_size;
217    
218     /* cannot perform a write until the read has cleared
219     - either via tpm_read or a user_read_timer timeout */
220     - while (atomic_read(&chip->data_pending) != 0)
221     - msleep(TPM_TIMEOUT);
222     -
223     - mutex_lock(&chip->buffer_mutex);
224     + either via tpm_read or a user_read_timer timeout.
225     + This also prevents splitted buffered writes from blocking here.
226     + */
227     + if (atomic_read(&chip->data_pending) != 0)
228     + return -EBUSY;
229    
230     if (in_size > TPM_BUFSIZE)
231     - in_size = TPM_BUFSIZE;
232     + return -E2BIG;
233     +
234     + mutex_lock(&chip->buffer_mutex);
235    
236     if (copy_from_user
237     (chip->data_buffer, (void __user *) buf, in_size)) {
238     @@ -1206,6 +1209,10 @@ ssize_t tpm_write(struct file *file, const char __user *buf,
239    
240     /* atomic tpm command send and result receive */
241     out_size = tpm_transmit(chip, chip->data_buffer, TPM_BUFSIZE);
242     + if (out_size < 0) {
243     + mutex_unlock(&chip->buffer_mutex);
244     + return out_size;
245     + }
246    
247     atomic_set(&chip->data_pending, out_size);
248     mutex_unlock(&chip->buffer_mutex);
249     diff --git a/drivers/firewire/core-cdev.c b/drivers/firewire/core-cdev.c
250     index 2e6b245..b8e4809 100644
251     --- a/drivers/firewire/core-cdev.c
252     +++ b/drivers/firewire/core-cdev.c
253     @@ -471,8 +471,8 @@ static int ioctl_get_info(struct client *client, union ioctl_arg *arg)
254     client->bus_reset_closure = a->bus_reset_closure;
255     if (a->bus_reset != 0) {
256     fill_bus_reset_event(&bus_reset, client);
257     - ret = copy_to_user(u64_to_uptr(a->bus_reset),
258     - &bus_reset, sizeof(bus_reset));
259     + /* unaligned size of bus_reset is 36 bytes */
260     + ret = copy_to_user(u64_to_uptr(a->bus_reset), &bus_reset, 36);
261     }
262     if (ret == 0 && list_empty(&client->link))
263     list_add_tail(&client->link, &client->device->client_list);
264     diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c
265     index 498bcbe..f3b06f0 100644
266     --- a/drivers/gpu/drm/i915/intel_display.c
267     +++ b/drivers/gpu/drm/i915/intel_display.c
268     @@ -5318,7 +5318,7 @@ static int i9xx_crtc_mode_set(struct drm_crtc *crtc,
269     /* default to 8bpc */
270     pipeconf &= ~(PIPECONF_BPP_MASK | PIPECONF_DITHER_EN);
271     if (is_dp) {
272     - if (mode->private_flags & INTEL_MODE_DP_FORCE_6BPC) {
273     + if (adjusted_mode->private_flags & INTEL_MODE_DP_FORCE_6BPC) {
274     pipeconf |= PIPECONF_BPP_6 |
275     PIPECONF_DITHER_EN |
276     PIPECONF_DITHER_TYPE_SP;
277     @@ -5782,7 +5782,7 @@ static int ironlake_crtc_mode_set(struct drm_crtc *crtc,
278     /* determine panel color depth */
279     temp = I915_READ(PIPECONF(pipe));
280     temp &= ~PIPE_BPC_MASK;
281     - dither = intel_choose_pipe_bpp_dither(crtc, &pipe_bpp, mode);
282     + dither = intel_choose_pipe_bpp_dither(crtc, &pipe_bpp, adjusted_mode);
283     switch (pipe_bpp) {
284     case 18:
285     temp |= PIPE_6BPC;
286     diff --git a/drivers/gpu/drm/radeon/radeon_legacy_encoders.c b/drivers/gpu/drm/radeon/radeon_legacy_encoders.c
287     index 42db254..1461e2c 100644
288     --- a/drivers/gpu/drm/radeon/radeon_legacy_encoders.c
289     +++ b/drivers/gpu/drm/radeon/radeon_legacy_encoders.c
290     @@ -973,11 +973,7 @@ static void radeon_legacy_tmds_ext_mode_set(struct drm_encoder *encoder,
291     static void radeon_ext_tmds_enc_destroy(struct drm_encoder *encoder)
292     {
293     struct radeon_encoder *radeon_encoder = to_radeon_encoder(encoder);
294     - struct radeon_encoder_ext_tmds *tmds = radeon_encoder->enc_priv;
295     - if (tmds) {
296     - if (tmds->i2c_bus)
297     - radeon_i2c_destroy(tmds->i2c_bus);
298     - }
299     + /* don't destroy the i2c bus record here, this will be done in radeon_i2c_fini */
300     kfree(radeon_encoder->enc_priv);
301     drm_encoder_cleanup(encoder);
302     kfree(radeon_encoder);
303     diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c
304     index 1f7e8cd..4a038cd 100644
305     --- a/drivers/md/raid10.c
306     +++ b/drivers/md/raid10.c
307     @@ -3019,7 +3019,7 @@ static sector_t sync_request(struct mddev *mddev, sector_t sector_nr,
308     else {
309     bad_sectors -= (sector - first_bad);
310     if (max_sync > bad_sectors)
311     - max_sync = max_sync;
312     + max_sync = bad_sectors;
313     continue;
314     }
315     }
316     diff --git a/drivers/net/ethernet/intel/e1000e/e1000.h b/drivers/net/ethernet/intel/e1000e/e1000.h
317     index b83897f..1ab8067 100644
318     --- a/drivers/net/ethernet/intel/e1000e/e1000.h
319     +++ b/drivers/net/ethernet/intel/e1000e/e1000.h
320     @@ -175,13 +175,13 @@ struct e1000_info;
321     /*
322     * in the case of WTHRESH, it appears at least the 82571/2 hardware
323     * writes back 4 descriptors when WTHRESH=5, and 3 descriptors when
324     - * WTHRESH=4, and since we want 64 bytes at a time written back, set
325     - * it to 5
326     + * WTHRESH=4, so a setting of 5 gives the most efficient bus
327     + * utilization but to avoid possible Tx stalls, set it to 1
328     */
329     #define E1000_TXDCTL_DMA_BURST_ENABLE \
330     (E1000_TXDCTL_GRAN | /* set descriptor granularity */ \
331     E1000_TXDCTL_COUNT_DESC | \
332     - (5 << 16) | /* wthresh must be +1 more than desired */\
333     + (1 << 16) | /* wthresh must be +1 more than desired */\
334     (1 << 8) | /* hthresh */ \
335     0x1f) /* pthresh */
336    
337     diff --git a/drivers/net/ethernet/intel/e1000e/netdev.c b/drivers/net/ethernet/intel/e1000e/netdev.c
338     index 5621d5b..7e88aaf 100644
339     --- a/drivers/net/ethernet/intel/e1000e/netdev.c
340     +++ b/drivers/net/ethernet/intel/e1000e/netdev.c
341     @@ -2806,7 +2806,7 @@ static void e1000_configure_tx(struct e1000_adapter *adapter)
342     * set up some performance related parameters to encourage the
343     * hardware to use the bus more efficiently in bursts, depends
344     * on the tx_int_delay to be enabled,
345     - * wthresh = 5 ==> burst write a cacheline (64 bytes) at a time
346     + * wthresh = 1 ==> burst write is disabled to avoid Tx stalls
347     * hthresh = 1 ==> prefetch when one or more available
348     * pthresh = 0x1f ==> prefetch if internal cache 31 or less
349     * BEWARE: this seems to work but should be considered first if
350     diff --git a/drivers/net/wireless/ath/ath9k/beacon.c b/drivers/net/wireless/ath/ath9k/beacon.c
351     index 6264182..5eb53c9 100644
352     --- a/drivers/net/wireless/ath/ath9k/beacon.c
353     +++ b/drivers/net/wireless/ath/ath9k/beacon.c
354     @@ -121,7 +121,7 @@ static void ath_tx_cabq(struct ieee80211_hw *hw, struct sk_buff *skb)
355    
356     if (ath_tx_start(hw, skb, &txctl) != 0) {
357     ath_dbg(common, XMIT, "CABQ TX failed\n");
358     - dev_kfree_skb_any(skb);
359     + ieee80211_free_txskb(hw, skb);
360     }
361     }
362    
363     diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
364     index d5dabcb..91e2c4f 100644
365     --- a/drivers/net/wireless/ath/ath9k/main.c
366     +++ b/drivers/net/wireless/ath/ath9k/main.c
367     @@ -1147,7 +1147,7 @@ static void ath9k_tx(struct ieee80211_hw *hw, struct sk_buff *skb)
368    
369     return;
370     exit:
371     - dev_kfree_skb_any(skb);
372     + ieee80211_free_txskb(hw, skb);
373     }
374    
375     static void ath9k_stop(struct ieee80211_hw *hw)
376     diff --git a/drivers/net/wireless/ath/ath9k/xmit.c b/drivers/net/wireless/ath/ath9k/xmit.c
377     index 4d57139..b78773b 100644
378     --- a/drivers/net/wireless/ath/ath9k/xmit.c
379     +++ b/drivers/net/wireless/ath/ath9k/xmit.c
380     @@ -64,8 +64,7 @@ static void ath_tx_update_baw(struct ath_softc *sc, struct ath_atx_tid *tid,
381     static struct ath_buf *ath_tx_setup_buffer(struct ath_softc *sc,
382     struct ath_txq *txq,
383     struct ath_atx_tid *tid,
384     - struct sk_buff *skb,
385     - bool dequeue);
386     + struct sk_buff *skb);
387    
388     enum {
389     MCS_HT20,
390     @@ -201,7 +200,15 @@ static void ath_tx_flush_tid(struct ath_softc *sc, struct ath_atx_tid *tid)
391     fi = get_frame_info(skb);
392     bf = fi->bf;
393    
394     - if (bf && fi->retries) {
395     + if (!bf) {
396     + bf = ath_tx_setup_buffer(sc, txq, tid, skb);
397     + if (!bf) {
398     + ieee80211_free_txskb(sc->hw, skb);
399     + continue;
400     + }
401     + }
402     +
403     + if (fi->retries) {
404     list_add_tail(&bf->list, &bf_head);
405     ath_tx_update_baw(sc, tid, bf->bf_state.seqno);
406     ath_tx_complete_buf(sc, bf, txq, &bf_head, &ts, 0);
407     @@ -812,10 +819,13 @@ static enum ATH_AGGR_STATUS ath_tx_form_aggr(struct ath_softc *sc,
408     fi = get_frame_info(skb);
409     bf = fi->bf;
410     if (!fi->bf)
411     - bf = ath_tx_setup_buffer(sc, txq, tid, skb, true);
412     + bf = ath_tx_setup_buffer(sc, txq, tid, skb);
413    
414     - if (!bf)
415     + if (!bf) {
416     + __skb_unlink(skb, &tid->buf_q);
417     + ieee80211_free_txskb(sc->hw, skb);
418     continue;
419     + }
420    
421     bf->bf_state.bf_type = BUF_AMPDU | BUF_AGGR;
422     seqno = bf->bf_state.seqno;
423     @@ -1717,9 +1727,11 @@ static void ath_tx_send_ampdu(struct ath_softc *sc, struct ath_atx_tid *tid,
424     return;
425     }
426    
427     - bf = ath_tx_setup_buffer(sc, txctl->txq, tid, skb, false);
428     - if (!bf)
429     + bf = ath_tx_setup_buffer(sc, txctl->txq, tid, skb);
430     + if (!bf) {
431     + ieee80211_free_txskb(sc->hw, skb);
432     return;
433     + }
434    
435     bf->bf_state.bf_type = BUF_AMPDU;
436     INIT_LIST_HEAD(&bf_head);
437     @@ -1743,11 +1755,6 @@ static void ath_tx_send_normal(struct ath_softc *sc, struct ath_txq *txq,
438     struct ath_buf *bf;
439    
440     bf = fi->bf;
441     - if (!bf)
442     - bf = ath_tx_setup_buffer(sc, txq, tid, skb, false);
443     -
444     - if (!bf)
445     - return;
446    
447     INIT_LIST_HEAD(&bf_head);
448     list_add_tail(&bf->list, &bf_head);
449     @@ -1820,8 +1827,7 @@ u8 ath_txchainmask_reduction(struct ath_softc *sc, u8 chainmask, u32 rate)
450     static struct ath_buf *ath_tx_setup_buffer(struct ath_softc *sc,
451     struct ath_txq *txq,
452     struct ath_atx_tid *tid,
453     - struct sk_buff *skb,
454     - bool dequeue)
455     + struct sk_buff *skb)
456     {
457     struct ath_common *common = ath9k_hw_common(sc->sc_ah);
458     struct ath_frame_info *fi = get_frame_info(skb);
459     @@ -1833,7 +1839,7 @@ static struct ath_buf *ath_tx_setup_buffer(struct ath_softc *sc,
460     bf = ath_tx_get_buffer(sc);
461     if (!bf) {
462     ath_dbg(common, XMIT, "TX buffers are full\n");
463     - goto error;
464     + return NULL;
465     }
466    
467     ATH_TXBUF_RESET(bf);
468     @@ -1862,18 +1868,12 @@ static struct ath_buf *ath_tx_setup_buffer(struct ath_softc *sc,
469     ath_err(ath9k_hw_common(sc->sc_ah),
470     "dma_mapping_error() on TX\n");
471     ath_tx_return_buffer(sc, bf);
472     - goto error;
473     + return NULL;
474     }
475    
476     fi->bf = bf;
477    
478     return bf;
479     -
480     -error:
481     - if (dequeue)
482     - __skb_unlink(skb, &tid->buf_q);
483     - dev_kfree_skb_any(skb);
484     - return NULL;
485     }
486    
487     /* FIXME: tx power */
488     @@ -1902,9 +1902,14 @@ static void ath_tx_start_dma(struct ath_softc *sc, struct sk_buff *skb,
489     */
490     ath_tx_send_ampdu(sc, tid, skb, txctl);
491     } else {
492     - bf = ath_tx_setup_buffer(sc, txctl->txq, tid, skb, false);
493     - if (!bf)
494     + bf = ath_tx_setup_buffer(sc, txctl->txq, tid, skb);
495     + if (!bf) {
496     + if (txctl->paprd)
497     + dev_kfree_skb_any(skb);
498     + else
499     + ieee80211_free_txskb(sc->hw, skb);
500     return;
501     + }
502    
503     bf->bf_state.bfs_paprd = txctl->paprd;
504    
505     diff --git a/drivers/scsi/hpsa.c b/drivers/scsi/hpsa.c
506     index 8a5e25d..b0fefc4 100644
507     --- a/drivers/scsi/hpsa.c
508     +++ b/drivers/scsi/hpsa.c
509     @@ -548,12 +548,42 @@ static void set_performant_mode(struct ctlr_info *h, struct CommandList *c)
510     c->busaddr |= 1 | (h->blockFetchTable[c->Header.SGList] << 1);
511     }
512    
513     +static int is_firmware_flash_cmd(u8 *cdb)
514     +{
515     + return cdb[0] == BMIC_WRITE && cdb[6] == BMIC_FLASH_FIRMWARE;
516     +}
517     +
518     +/*
519     + * During firmware flash, the heartbeat register may not update as frequently
520     + * as it should. So we dial down lockup detection during firmware flash. and
521     + * dial it back up when firmware flash completes.
522     + */
523     +#define HEARTBEAT_SAMPLE_INTERVAL_DURING_FLASH (240 * HZ)
524     +#define HEARTBEAT_SAMPLE_INTERVAL (30 * HZ)
525     +static void dial_down_lockup_detection_during_fw_flash(struct ctlr_info *h,
526     + struct CommandList *c)
527     +{
528     + if (!is_firmware_flash_cmd(c->Request.CDB))
529     + return;
530     + atomic_inc(&h->firmware_flash_in_progress);
531     + h->heartbeat_sample_interval = HEARTBEAT_SAMPLE_INTERVAL_DURING_FLASH;
532     +}
533     +
534     +static void dial_up_lockup_detection_on_fw_flash_complete(struct ctlr_info *h,
535     + struct CommandList *c)
536     +{
537     + if (is_firmware_flash_cmd(c->Request.CDB) &&
538     + atomic_dec_and_test(&h->firmware_flash_in_progress))
539     + h->heartbeat_sample_interval = HEARTBEAT_SAMPLE_INTERVAL;
540     +}
541     +
542     static void enqueue_cmd_and_start_io(struct ctlr_info *h,
543     struct CommandList *c)
544     {
545     unsigned long flags;
546    
547     set_performant_mode(h, c);
548     + dial_down_lockup_detection_during_fw_flash(h, c);
549     spin_lock_irqsave(&h->lock, flags);
550     addQ(&h->reqQ, c);
551     h->Qdepth++;
552     @@ -3049,6 +3079,7 @@ static inline int bad_tag(struct ctlr_info *h, u32 tag_index,
553     static inline void finish_cmd(struct CommandList *c, u32 raw_tag)
554     {
555     removeQ(c);
556     + dial_up_lockup_detection_on_fw_flash_complete(c->h, c);
557     if (likely(c->cmd_type == CMD_SCSI))
558     complete_scsi_command(c);
559     else if (c->cmd_type == CMD_IOCTL_PEND)
560     @@ -4189,9 +4220,6 @@ static void controller_lockup_detected(struct ctlr_info *h)
561     spin_unlock_irqrestore(&h->lock, flags);
562     }
563    
564     -#define HEARTBEAT_SAMPLE_INTERVAL (10 * HZ)
565     -#define HEARTBEAT_CHECK_MINIMUM_INTERVAL (HEARTBEAT_SAMPLE_INTERVAL / 2)
566     -
567     static void detect_controller_lockup(struct ctlr_info *h)
568     {
569     u64 now;
570     @@ -4202,7 +4230,7 @@ static void detect_controller_lockup(struct ctlr_info *h)
571     now = get_jiffies_64();
572     /* If we've received an interrupt recently, we're ok. */
573     if (time_after64(h->last_intr_timestamp +
574     - (HEARTBEAT_CHECK_MINIMUM_INTERVAL), now))
575     + (h->heartbeat_sample_interval), now))
576     return;
577    
578     /*
579     @@ -4211,7 +4239,7 @@ static void detect_controller_lockup(struct ctlr_info *h)
580     * otherwise don't care about signals in this thread.
581     */
582     if (time_after64(h->last_heartbeat_timestamp +
583     - (HEARTBEAT_CHECK_MINIMUM_INTERVAL), now))
584     + (h->heartbeat_sample_interval), now))
585     return;
586    
587     /* If heartbeat has not changed since we last looked, we're not ok. */
588     @@ -4253,6 +4281,7 @@ static void add_ctlr_to_lockup_detector_list(struct ctlr_info *h)
589     {
590     unsigned long flags;
591    
592     + h->heartbeat_sample_interval = HEARTBEAT_SAMPLE_INTERVAL;
593     spin_lock_irqsave(&lockup_detector_lock, flags);
594     list_add_tail(&h->lockup_list, &hpsa_ctlr_list);
595     spin_unlock_irqrestore(&lockup_detector_lock, flags);
596     diff --git a/drivers/scsi/hpsa.h b/drivers/scsi/hpsa.h
597     index 7b28d54..6f30a6f 100644
598     --- a/drivers/scsi/hpsa.h
599     +++ b/drivers/scsi/hpsa.h
600     @@ -123,6 +123,8 @@ struct ctlr_info {
601     u64 last_intr_timestamp;
602     u32 last_heartbeat;
603     u64 last_heartbeat_timestamp;
604     + u32 heartbeat_sample_interval;
605     + atomic_t firmware_flash_in_progress;
606     u32 lockup_detected;
607     struct list_head lockup_list;
608     };
609     diff --git a/drivers/scsi/hpsa_cmd.h b/drivers/scsi/hpsa_cmd.h
610     index 8049815..cdd742e 100644
611     --- a/drivers/scsi/hpsa_cmd.h
612     +++ b/drivers/scsi/hpsa_cmd.h
613     @@ -162,6 +162,7 @@ struct SenseSubsystem_info {
614     #define BMIC_WRITE 0x27
615     #define BMIC_CACHE_FLUSH 0xc2
616     #define HPSA_CACHE_FLUSH 0x01 /* C2 was already being used by HPSA */
617     +#define BMIC_FLASH_FIRMWARE 0xF7
618    
619     /* Command List Structure */
620     union SCSI3Addr {
621     diff --git a/drivers/scsi/scsi_debug.c b/drivers/scsi/scsi_debug.c
622     index 182d5a5..f4cc413 100644
623     --- a/drivers/scsi/scsi_debug.c
624     +++ b/drivers/scsi/scsi_debug.c
625     @@ -2054,7 +2054,7 @@ static void unmap_region(sector_t lba, unsigned int len)
626     block = lba + alignment;
627     rem = do_div(block, granularity);
628    
629     - if (rem == 0 && lba + granularity <= end && block < map_size) {
630     + if (rem == 0 && lba + granularity < end && block < map_size) {
631     clear_bit(block, map_storep);
632     if (scsi_debug_lbprz)
633     memset(fake_storep +
634     diff --git a/drivers/scsi/storvsc_drv.c b/drivers/scsi/storvsc_drv.c
635     index 83a1972..40a4570 100644
636     --- a/drivers/scsi/storvsc_drv.c
637     +++ b/drivers/scsi/storvsc_drv.c
638     @@ -1211,7 +1211,12 @@ static int storvsc_host_reset_handler(struct scsi_cmnd *scmnd)
639     /*
640     * At this point, all outstanding requests in the adapter
641     * should have been flushed out and return to us
642     + * There is a potential race here where the host may be in
643     + * the process of responding when we return from here.
644     + * Just wait for all in-transit packets to be accounted for
645     + * before we return from here.
646     */
647     + storvsc_wait_to_drain(stor_device);
648    
649     return SUCCESS;
650     }
651     diff --git a/drivers/target/iscsi/iscsi_target.c b/drivers/target/iscsi/iscsi_target.c
652     index e326d17..23e122a 100644
653     --- a/drivers/target/iscsi/iscsi_target.c
654     +++ b/drivers/target/iscsi/iscsi_target.c
655     @@ -3196,7 +3196,6 @@ static int iscsit_build_sendtargets_response(struct iscsi_cmd *cmd)
656     len += 1;
657    
658     if ((len + payload_len) > buffer_len) {
659     - spin_unlock(&tiqn->tiqn_tpg_lock);
660     end_of_buf = 1;
661     goto eob;
662     }
663     @@ -3349,6 +3348,7 @@ static int iscsit_send_reject(
664     hdr->opcode = ISCSI_OP_REJECT;
665     hdr->flags |= ISCSI_FLAG_CMD_FINAL;
666     hton24(hdr->dlength, ISCSI_HDR_LEN);
667     + hdr->ffffffff = 0xffffffff;
668     cmd->stat_sn = conn->stat_sn++;
669     hdr->statsn = cpu_to_be32(cmd->stat_sn);
670     hdr->exp_cmdsn = cpu_to_be32(conn->sess->exp_cmd_sn);
671     diff --git a/drivers/target/iscsi/iscsi_target_core.h b/drivers/target/iscsi/iscsi_target_core.h
672     index d1c4bc2..1596aec 100644
673     --- a/drivers/target/iscsi/iscsi_target_core.h
674     +++ b/drivers/target/iscsi/iscsi_target_core.h
675     @@ -25,10 +25,10 @@
676     #define NA_DATAOUT_TIMEOUT_RETRIES 5
677     #define NA_DATAOUT_TIMEOUT_RETRIES_MAX 15
678     #define NA_DATAOUT_TIMEOUT_RETRIES_MIN 1
679     -#define NA_NOPIN_TIMEOUT 5
680     +#define NA_NOPIN_TIMEOUT 15
681     #define NA_NOPIN_TIMEOUT_MAX 60
682     #define NA_NOPIN_TIMEOUT_MIN 3
683     -#define NA_NOPIN_RESPONSE_TIMEOUT 5
684     +#define NA_NOPIN_RESPONSE_TIMEOUT 30
685     #define NA_NOPIN_RESPONSE_TIMEOUT_MAX 60
686     #define NA_NOPIN_RESPONSE_TIMEOUT_MIN 3
687     #define NA_RANDOM_DATAIN_PDU_OFFSETS 0
688     diff --git a/drivers/target/iscsi/iscsi_target_tpg.c b/drivers/target/iscsi/iscsi_target_tpg.c
689     index 879d8d0..c3d7bf54 100644
690     --- a/drivers/target/iscsi/iscsi_target_tpg.c
691     +++ b/drivers/target/iscsi/iscsi_target_tpg.c
692     @@ -672,6 +672,12 @@ int iscsit_ta_generate_node_acls(
693     pr_debug("iSCSI_TPG[%hu] - Generate Initiator Portal Group ACLs: %s\n",
694     tpg->tpgt, (a->generate_node_acls) ? "Enabled" : "Disabled");
695    
696     + if (flag == 1 && a->cache_dynamic_acls == 0) {
697     + pr_debug("Explicitly setting cache_dynamic_acls=1 when "
698     + "generate_node_acls=1\n");
699     + a->cache_dynamic_acls = 1;
700     + }
701     +
702     return 0;
703     }
704    
705     @@ -711,6 +717,12 @@ int iscsit_ta_cache_dynamic_acls(
706     return -EINVAL;
707     }
708    
709     + if (a->generate_node_acls == 1 && flag == 0) {
710     + pr_debug("Skipping cache_dynamic_acls=0 when"
711     + " generate_node_acls=1\n");
712     + return 0;
713     + }
714     +
715     a->cache_dynamic_acls = flag;
716     pr_debug("iSCSI_TPG[%hu] - Cache Dynamic Initiator Portal Group"
717     " ACLs %s\n", tpg->tpgt, (a->cache_dynamic_acls) ?
718     diff --git a/drivers/target/target_core_configfs.c b/drivers/target/target_core_configfs.c
719     index cbb6653..3f90d4b 100644
720     --- a/drivers/target/target_core_configfs.c
721     +++ b/drivers/target/target_core_configfs.c
722     @@ -3115,6 +3115,7 @@ static int __init target_core_init_configfs(void)
723     GFP_KERNEL);
724     if (!target_cg->default_groups) {
725     pr_err("Unable to allocate target_cg->default_groups\n");
726     + ret = -ENOMEM;
727     goto out_global;
728     }
729    
730     @@ -3130,6 +3131,7 @@ static int __init target_core_init_configfs(void)
731     GFP_KERNEL);
732     if (!hba_cg->default_groups) {
733     pr_err("Unable to allocate hba_cg->default_groups\n");
734     + ret = -ENOMEM;
735     goto out_global;
736     }
737     config_group_init_type_name(&alua_group,
738     @@ -3145,6 +3147,7 @@ static int __init target_core_init_configfs(void)
739     GFP_KERNEL);
740     if (!alua_cg->default_groups) {
741     pr_err("Unable to allocate alua_cg->default_groups\n");
742     + ret = -ENOMEM;
743     goto out_global;
744     }
745    
746     @@ -3156,14 +3159,17 @@ static int __init target_core_init_configfs(void)
747     * Add core/alua/lu_gps/default_lu_gp
748     */
749     lu_gp = core_alua_allocate_lu_gp("default_lu_gp", 1);
750     - if (IS_ERR(lu_gp))
751     + if (IS_ERR(lu_gp)) {
752     + ret = -ENOMEM;
753     goto out_global;
754     + }
755    
756     lu_gp_cg = &alua_lu_gps_group;
757     lu_gp_cg->default_groups = kzalloc(sizeof(struct config_group) * 2,
758     GFP_KERNEL);
759     if (!lu_gp_cg->default_groups) {
760     pr_err("Unable to allocate lu_gp_cg->default_groups\n");
761     + ret = -ENOMEM;
762     goto out_global;
763     }
764    
765     diff --git a/drivers/tty/vt/vt.c b/drivers/tty/vt/vt.c
766     index 2156188..18d06be 100644
767     --- a/drivers/tty/vt/vt.c
768     +++ b/drivers/tty/vt/vt.c
769     @@ -3475,6 +3475,19 @@ int con_debug_enter(struct vc_data *vc)
770     kdb_set(2, setargs);
771     }
772     }
773     + if (vc->vc_cols < 999) {
774     + int colcount;
775     + char cols[4];
776     + const char *setargs[3] = {
777     + "set",
778     + "COLUMNS",
779     + cols,
780     + };
781     + if (kdbgetintenv(setargs[0], &colcount)) {
782     + snprintf(cols, 4, "%i", vc->vc_cols);
783     + kdb_set(2, setargs);
784     + }
785     + }
786     #endif /* CONFIG_KGDB_KDB */
787     return ret;
788     }
789     diff --git a/drivers/usb/class/cdc-acm.c b/drivers/usb/class/cdc-acm.c
790     index 1b632cb..7f2fac1 100644
791     --- a/drivers/usb/class/cdc-acm.c
792     +++ b/drivers/usb/class/cdc-acm.c
793     @@ -1551,6 +1551,9 @@ static const struct usb_device_id acm_ids[] = {
794     Maybe we should define a new
795     quirk for this. */
796     },
797     + { USB_DEVICE(0x0572, 0x1340), /* Conexant CX93010-2x UCMxx */
798     + .driver_info = NO_UNION_NORMAL,
799     + },
800     { USB_DEVICE(0x1bbb, 0x0003), /* Alcatel OT-I650 */
801     .driver_info = NO_UNION_NORMAL, /* reports zero length descriptor */
802     },
803     diff --git a/drivers/usb/gadget/at91_udc.c b/drivers/usb/gadget/at91_udc.c
804     index 9d7bcd9..be6952e 100644
805     --- a/drivers/usb/gadget/at91_udc.c
806     +++ b/drivers/usb/gadget/at91_udc.c
807     @@ -1735,7 +1735,7 @@ static int __devinit at91udc_probe(struct platform_device *pdev)
808     int retval;
809     struct resource *res;
810    
811     - if (!dev->platform_data) {
812     + if (!dev->platform_data && !pdev->dev.of_node) {
813     /* small (so we copy it) but critical! */
814     DBG("missing platform_data\n");
815     return -ENODEV;
816     diff --git a/drivers/video/udlfb.c b/drivers/video/udlfb.c
817     index a159b63..85d8110 100644
818     --- a/drivers/video/udlfb.c
819     +++ b/drivers/video/udlfb.c
820     @@ -647,7 +647,7 @@ static ssize_t dlfb_ops_write(struct fb_info *info, const char __user *buf,
821     result = fb_sys_write(info, buf, count, ppos);
822    
823     if (result > 0) {
824     - int start = max((int)(offset / info->fix.line_length) - 1, 0);
825     + int start = max((int)(offset / info->fix.line_length), 0);
826     int lines = min((u32)((result / info->fix.line_length) + 1),
827     (u32)info->var.yres);
828    
829     diff --git a/drivers/video/via/via_clock.c b/drivers/video/via/via_clock.c
830     index af8f26b..db1e392 100644
831     --- a/drivers/video/via/via_clock.c
832     +++ b/drivers/video/via/via_clock.c
833     @@ -25,6 +25,7 @@
834    
835     #include <linux/kernel.h>
836     #include <linux/via-core.h>
837     +#include <asm/olpc.h>
838     #include "via_clock.h"
839     #include "global.h"
840     #include "debug.h"
841     @@ -289,6 +290,10 @@ static void dummy_set_pll(struct via_pll_config config)
842     printk(KERN_INFO "Using undocumented set PLL.\n%s", via_slap);
843     }
844    
845     +static void noop_set_clock_state(u8 state)
846     +{
847     +}
848     +
849     void via_clock_init(struct via_clock *clock, int gfx_chip)
850     {
851     switch (gfx_chip) {
852     @@ -346,4 +351,18 @@ void via_clock_init(struct via_clock *clock, int gfx_chip)
853     break;
854    
855     }
856     +
857     + if (machine_is_olpc()) {
858     + /* The OLPC XO-1.5 cannot suspend/resume reliably if the
859     + * IGA1/IGA2 clocks are set as on or off (memory rot
860     + * occasionally happens during suspend under such
861     + * configurations).
862     + *
863     + * The only known stable scenario is to leave this bits as-is,
864     + * which in their default states are documented to enable the
865     + * clock only when it is needed.
866     + */
867     + clock->set_primary_clock_state = noop_set_clock_state;
868     + clock->set_secondary_clock_state = noop_set_clock_state;
869     + }
870     }
871     diff --git a/fs/autofs4/root.c b/fs/autofs4/root.c
872     index 75e5f1c..8c4292f 100644
873     --- a/fs/autofs4/root.c
874     +++ b/fs/autofs4/root.c
875     @@ -392,10 +392,12 @@ static struct vfsmount *autofs4_d_automount(struct path *path)
876     ino->flags |= AUTOFS_INF_PENDING;
877     spin_unlock(&sbi->fs_lock);
878     status = autofs4_mount_wait(dentry);
879     - if (status)
880     - return ERR_PTR(status);
881     spin_lock(&sbi->fs_lock);
882     ino->flags &= ~AUTOFS_INF_PENDING;
883     + if (status) {
884     + spin_unlock(&sbi->fs_lock);
885     + return ERR_PTR(status);
886     + }
887     }
888     done:
889     if (!(ino->flags & AUTOFS_INF_EXPIRING)) {
890     diff --git a/fs/ceph/export.c b/fs/ceph/export.c
891     index fbb2a64..4098ccf 100644
892     --- a/fs/ceph/export.c
893     +++ b/fs/ceph/export.c
894     @@ -89,7 +89,7 @@ static int ceph_encode_fh(struct dentry *dentry, u32 *rawfh, int *max_len,
895     * FIXME: we should try harder by querying the mds for the ino.
896     */
897     static struct dentry *__fh_to_dentry(struct super_block *sb,
898     - struct ceph_nfs_fh *fh)
899     + struct ceph_nfs_fh *fh, int fh_len)
900     {
901     struct ceph_mds_client *mdsc = ceph_sb_to_client(sb)->mdsc;
902     struct inode *inode;
903     @@ -97,6 +97,9 @@ static struct dentry *__fh_to_dentry(struct super_block *sb,
904     struct ceph_vino vino;
905     int err;
906    
907     + if (fh_len < sizeof(*fh) / 4)
908     + return ERR_PTR(-ESTALE);
909     +
910     dout("__fh_to_dentry %llx\n", fh->ino);
911     vino.ino = fh->ino;
912     vino.snap = CEPH_NOSNAP;
913     @@ -140,7 +143,7 @@ static struct dentry *__fh_to_dentry(struct super_block *sb,
914     * convert connectable fh to dentry
915     */
916     static struct dentry *__cfh_to_dentry(struct super_block *sb,
917     - struct ceph_nfs_confh *cfh)
918     + struct ceph_nfs_confh *cfh, int fh_len)
919     {
920     struct ceph_mds_client *mdsc = ceph_sb_to_client(sb)->mdsc;
921     struct inode *inode;
922     @@ -148,6 +151,9 @@ static struct dentry *__cfh_to_dentry(struct super_block *sb,
923     struct ceph_vino vino;
924     int err;
925    
926     + if (fh_len < sizeof(*cfh) / 4)
927     + return ERR_PTR(-ESTALE);
928     +
929     dout("__cfh_to_dentry %llx (%llx/%x)\n",
930     cfh->ino, cfh->parent_ino, cfh->parent_name_hash);
931    
932     @@ -197,9 +203,11 @@ static struct dentry *ceph_fh_to_dentry(struct super_block *sb, struct fid *fid,
933     int fh_len, int fh_type)
934     {
935     if (fh_type == 1)
936     - return __fh_to_dentry(sb, (struct ceph_nfs_fh *)fid->raw);
937     + return __fh_to_dentry(sb, (struct ceph_nfs_fh *)fid->raw,
938     + fh_len);
939     else
940     - return __cfh_to_dentry(sb, (struct ceph_nfs_confh *)fid->raw);
941     + return __cfh_to_dentry(sb, (struct ceph_nfs_confh *)fid->raw,
942     + fh_len);
943     }
944    
945     /*
946     @@ -220,6 +228,8 @@ static struct dentry *ceph_fh_to_parent(struct super_block *sb,
947    
948     if (fh_type == 1)
949     return ERR_PTR(-ESTALE);
950     + if (fh_len < sizeof(*cfh) / 4)
951     + return ERR_PTR(-ESTALE);
952    
953     pr_debug("fh_to_parent %llx/%d\n", cfh->parent_ino,
954     cfh->parent_name_hash);
955     diff --git a/fs/ecryptfs/ecryptfs_kernel.h b/fs/ecryptfs/ecryptfs_kernel.h
956     index 867b64c..56e3aa5 100644
957     --- a/fs/ecryptfs/ecryptfs_kernel.h
958     +++ b/fs/ecryptfs/ecryptfs_kernel.h
959     @@ -568,6 +568,8 @@ struct ecryptfs_open_req {
960     struct inode *ecryptfs_get_inode(struct inode *lower_inode,
961     struct super_block *sb);
962     void ecryptfs_i_size_init(const char *page_virt, struct inode *inode);
963     +int ecryptfs_initialize_file(struct dentry *ecryptfs_dentry,
964     + struct inode *ecryptfs_inode);
965     int ecryptfs_decode_and_decrypt_filename(char **decrypted_name,
966     size_t *decrypted_name_size,
967     struct dentry *ecryptfs_dentry,
968     diff --git a/fs/ecryptfs/file.c b/fs/ecryptfs/file.c
969     index 2b17f2f..d45ba45 100644
970     --- a/fs/ecryptfs/file.c
971     +++ b/fs/ecryptfs/file.c
972     @@ -138,29 +138,50 @@ out:
973     return rc;
974     }
975    
976     -static void ecryptfs_vma_close(struct vm_area_struct *vma)
977     -{
978     - filemap_write_and_wait(vma->vm_file->f_mapping);
979     -}
980     -
981     -static const struct vm_operations_struct ecryptfs_file_vm_ops = {
982     - .close = ecryptfs_vma_close,
983     - .fault = filemap_fault,
984     -};
985     +struct kmem_cache *ecryptfs_file_info_cache;
986    
987     -static int ecryptfs_file_mmap(struct file *file, struct vm_area_struct *vma)
988     +static int read_or_initialize_metadata(struct dentry *dentry)
989     {
990     + struct inode *inode = dentry->d_inode;
991     + struct ecryptfs_mount_crypt_stat *mount_crypt_stat;
992     + struct ecryptfs_crypt_stat *crypt_stat;
993     int rc;
994    
995     - rc = generic_file_mmap(file, vma);
996     + crypt_stat = &ecryptfs_inode_to_private(inode)->crypt_stat;
997     + mount_crypt_stat = &ecryptfs_superblock_to_private(
998     + inode->i_sb)->mount_crypt_stat;
999     + mutex_lock(&crypt_stat->cs_mutex);
1000     +
1001     + if (crypt_stat->flags & ECRYPTFS_POLICY_APPLIED &&
1002     + crypt_stat->flags & ECRYPTFS_KEY_VALID) {
1003     + rc = 0;
1004     + goto out;
1005     + }
1006     +
1007     + rc = ecryptfs_read_metadata(dentry);
1008     if (!rc)
1009     - vma->vm_ops = &ecryptfs_file_vm_ops;
1010     + goto out;
1011     +
1012     + if (mount_crypt_stat->flags & ECRYPTFS_PLAINTEXT_PASSTHROUGH_ENABLED) {
1013     + crypt_stat->flags &= ~(ECRYPTFS_I_SIZE_INITIALIZED
1014     + | ECRYPTFS_ENCRYPTED);
1015     + rc = 0;
1016     + goto out;
1017     + }
1018     +
1019     + if (!(mount_crypt_stat->flags & ECRYPTFS_XATTR_METADATA_ENABLED) &&
1020     + !i_size_read(ecryptfs_inode_to_lower(inode))) {
1021     + rc = ecryptfs_initialize_file(dentry, inode);
1022     + if (!rc)
1023     + goto out;
1024     + }
1025    
1026     + rc = -EIO;
1027     +out:
1028     + mutex_unlock(&crypt_stat->cs_mutex);
1029     return rc;
1030     }
1031    
1032     -struct kmem_cache *ecryptfs_file_info_cache;
1033     -
1034     /**
1035     * ecryptfs_open
1036     * @inode: inode speciying file to open
1037     @@ -236,32 +257,9 @@ static int ecryptfs_open(struct inode *inode, struct file *file)
1038     rc = 0;
1039     goto out;
1040     }
1041     - mutex_lock(&crypt_stat->cs_mutex);
1042     - if (!(crypt_stat->flags & ECRYPTFS_POLICY_APPLIED)
1043     - || !(crypt_stat->flags & ECRYPTFS_KEY_VALID)) {
1044     - rc = ecryptfs_read_metadata(ecryptfs_dentry);
1045     - if (rc) {
1046     - ecryptfs_printk(KERN_DEBUG,
1047     - "Valid headers not found\n");
1048     - if (!(mount_crypt_stat->flags
1049     - & ECRYPTFS_PLAINTEXT_PASSTHROUGH_ENABLED)) {
1050     - rc = -EIO;
1051     - printk(KERN_WARNING "Either the lower file "
1052     - "is not in a valid eCryptfs format, "
1053     - "or the key could not be retrieved. "
1054     - "Plaintext passthrough mode is not "
1055     - "enabled; returning -EIO\n");
1056     - mutex_unlock(&crypt_stat->cs_mutex);
1057     - goto out_put;
1058     - }
1059     - rc = 0;
1060     - crypt_stat->flags &= ~(ECRYPTFS_I_SIZE_INITIALIZED
1061     - | ECRYPTFS_ENCRYPTED);
1062     - mutex_unlock(&crypt_stat->cs_mutex);
1063     - goto out;
1064     - }
1065     - }
1066     - mutex_unlock(&crypt_stat->cs_mutex);
1067     + rc = read_or_initialize_metadata(ecryptfs_dentry);
1068     + if (rc)
1069     + goto out_put;
1070     ecryptfs_printk(KERN_DEBUG, "inode w/ addr = [0x%p], i_ino = "
1071     "[0x%.16lx] size: [0x%.16llx]\n", inode, inode->i_ino,
1072     (unsigned long long)i_size_read(inode));
1073     @@ -277,8 +275,14 @@ out:
1074    
1075     static int ecryptfs_flush(struct file *file, fl_owner_t td)
1076     {
1077     - return file->f_mode & FMODE_WRITE
1078     - ? filemap_write_and_wait(file->f_mapping) : 0;
1079     + struct file *lower_file = ecryptfs_file_to_lower(file);
1080     +
1081     + if (lower_file->f_op && lower_file->f_op->flush) {
1082     + filemap_write_and_wait(file->f_mapping);
1083     + return lower_file->f_op->flush(lower_file, td);
1084     + }
1085     +
1086     + return 0;
1087     }
1088    
1089     static int ecryptfs_release(struct inode *inode, struct file *file)
1090     @@ -292,15 +296,7 @@ static int ecryptfs_release(struct inode *inode, struct file *file)
1091     static int
1092     ecryptfs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
1093     {
1094     - int rc = 0;
1095     -
1096     - rc = generic_file_fsync(file, start, end, datasync);
1097     - if (rc)
1098     - goto out;
1099     - rc = vfs_fsync_range(ecryptfs_file_to_lower(file), start, end,
1100     - datasync);
1101     -out:
1102     - return rc;
1103     + return vfs_fsync(ecryptfs_file_to_lower(file), datasync);
1104     }
1105    
1106     static int ecryptfs_fasync(int fd, struct file *file, int flag)
1107     @@ -369,7 +365,7 @@ const struct file_operations ecryptfs_main_fops = {
1108     #ifdef CONFIG_COMPAT
1109     .compat_ioctl = ecryptfs_compat_ioctl,
1110     #endif
1111     - .mmap = ecryptfs_file_mmap,
1112     + .mmap = generic_file_mmap,
1113     .open = ecryptfs_open,
1114     .flush = ecryptfs_flush,
1115     .release = ecryptfs_release,
1116     diff --git a/fs/ecryptfs/inode.c b/fs/ecryptfs/inode.c
1117     index 6f5fb1a..11030b2 100644
1118     --- a/fs/ecryptfs/inode.c
1119     +++ b/fs/ecryptfs/inode.c
1120     @@ -143,6 +143,31 @@ static int ecryptfs_interpose(struct dentry *lower_dentry,
1121     return 0;
1122     }
1123    
1124     +static int ecryptfs_do_unlink(struct inode *dir, struct dentry *dentry,
1125     + struct inode *inode)
1126     +{
1127     + struct dentry *lower_dentry = ecryptfs_dentry_to_lower(dentry);
1128     + struct inode *lower_dir_inode = ecryptfs_inode_to_lower(dir);
1129     + struct dentry *lower_dir_dentry;
1130     + int rc;
1131     +
1132     + dget(lower_dentry);
1133     + lower_dir_dentry = lock_parent(lower_dentry);
1134     + rc = vfs_unlink(lower_dir_inode, lower_dentry);
1135     + if (rc) {
1136     + printk(KERN_ERR "Error in vfs_unlink; rc = [%d]\n", rc);
1137     + goto out_unlock;
1138     + }
1139     + fsstack_copy_attr_times(dir, lower_dir_inode);
1140     + set_nlink(inode, ecryptfs_inode_to_lower(inode)->i_nlink);
1141     + inode->i_ctime = dir->i_ctime;
1142     + d_drop(dentry);
1143     +out_unlock:
1144     + unlock_dir(lower_dir_dentry);
1145     + dput(lower_dentry);
1146     + return rc;
1147     +}
1148     +
1149     /**
1150     * ecryptfs_do_create
1151     * @directory_inode: inode of the new file's dentry's parent in ecryptfs
1152     @@ -182,8 +207,10 @@ ecryptfs_do_create(struct inode *directory_inode,
1153     }
1154     inode = __ecryptfs_get_inode(lower_dentry->d_inode,
1155     directory_inode->i_sb);
1156     - if (IS_ERR(inode))
1157     + if (IS_ERR(inode)) {
1158     + vfs_unlink(lower_dir_dentry->d_inode, lower_dentry);
1159     goto out_lock;
1160     + }
1161     fsstack_copy_attr_times(directory_inode, lower_dir_dentry->d_inode);
1162     fsstack_copy_inode_size(directory_inode, lower_dir_dentry->d_inode);
1163     out_lock:
1164     @@ -200,8 +227,8 @@ out:
1165     *
1166     * Returns zero on success
1167     */
1168     -static int ecryptfs_initialize_file(struct dentry *ecryptfs_dentry,
1169     - struct inode *ecryptfs_inode)
1170     +int ecryptfs_initialize_file(struct dentry *ecryptfs_dentry,
1171     + struct inode *ecryptfs_inode)
1172     {
1173     struct ecryptfs_crypt_stat *crypt_stat =
1174     &ecryptfs_inode_to_private(ecryptfs_inode)->crypt_stat;
1175     @@ -265,7 +292,9 @@ ecryptfs_create(struct inode *directory_inode, struct dentry *ecryptfs_dentry,
1176     * that this on disk file is prepared to be an ecryptfs file */
1177     rc = ecryptfs_initialize_file(ecryptfs_dentry, ecryptfs_inode);
1178     if (rc) {
1179     - drop_nlink(ecryptfs_inode);
1180     + ecryptfs_do_unlink(directory_inode, ecryptfs_dentry,
1181     + ecryptfs_inode);
1182     + make_bad_inode(ecryptfs_inode);
1183     unlock_new_inode(ecryptfs_inode);
1184     iput(ecryptfs_inode);
1185     goto out;
1186     @@ -477,27 +506,7 @@ out_lock:
1187    
1188     static int ecryptfs_unlink(struct inode *dir, struct dentry *dentry)
1189     {
1190     - int rc = 0;
1191     - struct dentry *lower_dentry = ecryptfs_dentry_to_lower(dentry);
1192     - struct inode *lower_dir_inode = ecryptfs_inode_to_lower(dir);
1193     - struct dentry *lower_dir_dentry;
1194     -
1195     - dget(lower_dentry);
1196     - lower_dir_dentry = lock_parent(lower_dentry);
1197     - rc = vfs_unlink(lower_dir_inode, lower_dentry);
1198     - if (rc) {
1199     - printk(KERN_ERR "Error in vfs_unlink; rc = [%d]\n", rc);
1200     - goto out_unlock;
1201     - }
1202     - fsstack_copy_attr_times(dir, lower_dir_inode);
1203     - set_nlink(dentry->d_inode,
1204     - ecryptfs_inode_to_lower(dentry->d_inode)->i_nlink);
1205     - dentry->d_inode->i_ctime = dir->i_ctime;
1206     - d_drop(dentry);
1207     -out_unlock:
1208     - unlock_dir(lower_dir_dentry);
1209     - dput(lower_dentry);
1210     - return rc;
1211     + return ecryptfs_do_unlink(dir, dentry, dentry->d_inode);
1212     }
1213    
1214     static int ecryptfs_symlink(struct inode *dir, struct dentry *dentry,
1215     @@ -1007,12 +1016,6 @@ static int ecryptfs_setattr(struct dentry *dentry, struct iattr *ia)
1216     goto out;
1217     }
1218    
1219     - if (S_ISREG(inode->i_mode)) {
1220     - rc = filemap_write_and_wait(inode->i_mapping);
1221     - if (rc)
1222     - goto out;
1223     - fsstack_copy_attr_all(inode, lower_inode);
1224     - }
1225     memcpy(&lower_ia, ia, sizeof(lower_ia));
1226     if (ia->ia_valid & ATTR_FILE)
1227     lower_ia.ia_file = ecryptfs_file_to_lower(ia->ia_file);
1228     diff --git a/fs/ecryptfs/main.c b/fs/ecryptfs/main.c
1229     index 6895493..0f04d2e 100644
1230     --- a/fs/ecryptfs/main.c
1231     +++ b/fs/ecryptfs/main.c
1232     @@ -162,6 +162,7 @@ void ecryptfs_put_lower_file(struct inode *inode)
1233     inode_info = ecryptfs_inode_to_private(inode);
1234     if (atomic_dec_and_mutex_lock(&inode_info->lower_file_count,
1235     &inode_info->lower_file_mutex)) {
1236     + filemap_write_and_wait(inode->i_mapping);
1237     fput(inode_info->lower_file);
1238     inode_info->lower_file = NULL;
1239     mutex_unlock(&inode_info->lower_file_mutex);
1240     diff --git a/fs/ecryptfs/mmap.c b/fs/ecryptfs/mmap.c
1241     index a46b3a8..bd1d57f 100644
1242     --- a/fs/ecryptfs/mmap.c
1243     +++ b/fs/ecryptfs/mmap.c
1244     @@ -66,18 +66,6 @@ static int ecryptfs_writepage(struct page *page, struct writeback_control *wbc)
1245     {
1246     int rc;
1247    
1248     - /*
1249     - * Refuse to write the page out if we are called from reclaim context
1250     - * since our writepage() path may potentially allocate memory when
1251     - * calling into the lower fs vfs_write() which may in turn invoke
1252     - * us again.
1253     - */
1254     - if (current->flags & PF_MEMALLOC) {
1255     - redirty_page_for_writepage(wbc, page);
1256     - rc = 0;
1257     - goto out;
1258     - }
1259     -
1260     rc = ecryptfs_encrypt_page(page);
1261     if (rc) {
1262     ecryptfs_printk(KERN_WARNING, "Error encrypting "
1263     @@ -498,7 +486,6 @@ static int ecryptfs_write_end(struct file *file,
1264     struct ecryptfs_crypt_stat *crypt_stat =
1265     &ecryptfs_inode_to_private(ecryptfs_inode)->crypt_stat;
1266     int rc;
1267     - int need_unlock_page = 1;
1268    
1269     ecryptfs_printk(KERN_DEBUG, "Calling fill_zeros_to_end_of_page"
1270     "(page w/ index = [0x%.16lx], to = [%d])\n", index, to);
1271     @@ -519,26 +506,26 @@ static int ecryptfs_write_end(struct file *file,
1272     "zeros in page with index = [0x%.16lx]\n", index);
1273     goto out;
1274     }
1275     - set_page_dirty(page);
1276     - unlock_page(page);
1277     - need_unlock_page = 0;
1278     + rc = ecryptfs_encrypt_page(page);
1279     + if (rc) {
1280     + ecryptfs_printk(KERN_WARNING, "Error encrypting page (upper "
1281     + "index [0x%.16lx])\n", index);
1282     + goto out;
1283     + }
1284     if (pos + copied > i_size_read(ecryptfs_inode)) {
1285     i_size_write(ecryptfs_inode, pos + copied);
1286     ecryptfs_printk(KERN_DEBUG, "Expanded file size to "
1287     "[0x%.16llx]\n",
1288     (unsigned long long)i_size_read(ecryptfs_inode));
1289     - balance_dirty_pages_ratelimited(mapping);
1290     - rc = ecryptfs_write_inode_size_to_metadata(ecryptfs_inode);
1291     - if (rc) {
1292     - printk(KERN_ERR "Error writing inode size to metadata; "
1293     - "rc = [%d]\n", rc);
1294     - goto out;
1295     - }
1296     }
1297     - rc = copied;
1298     + rc = ecryptfs_write_inode_size_to_metadata(ecryptfs_inode);
1299     + if (rc)
1300     + printk(KERN_ERR "Error writing inode size to metadata; "
1301     + "rc = [%d]\n", rc);
1302     + else
1303     + rc = copied;
1304     out:
1305     - if (need_unlock_page)
1306     - unlock_page(page);
1307     + unlock_page(page);
1308     page_cache_release(page);
1309     return rc;
1310     }
1311     diff --git a/fs/gfs2/export.c b/fs/gfs2/export.c
1312     index 70ba891..fdef7f0 100644
1313     --- a/fs/gfs2/export.c
1314     +++ b/fs/gfs2/export.c
1315     @@ -168,6 +168,8 @@ static struct dentry *gfs2_fh_to_dentry(struct super_block *sb, struct fid *fid,
1316     case GFS2_SMALL_FH_SIZE:
1317     case GFS2_LARGE_FH_SIZE:
1318     case GFS2_OLD_FH_SIZE:
1319     + if (fh_len < GFS2_SMALL_FH_SIZE)
1320     + return NULL;
1321     this.no_formal_ino = ((u64)be32_to_cpu(fh[0])) << 32;
1322     this.no_formal_ino |= be32_to_cpu(fh[1]);
1323     this.no_addr = ((u64)be32_to_cpu(fh[2])) << 32;
1324     @@ -187,6 +189,8 @@ static struct dentry *gfs2_fh_to_parent(struct super_block *sb, struct fid *fid,
1325     switch (fh_type) {
1326     case GFS2_LARGE_FH_SIZE:
1327     case GFS2_OLD_FH_SIZE:
1328     + if (fh_len < GFS2_LARGE_FH_SIZE)
1329     + return NULL;
1330     parent.no_formal_ino = ((u64)be32_to_cpu(fh[4])) << 32;
1331     parent.no_formal_ino |= be32_to_cpu(fh[5]);
1332     parent.no_addr = ((u64)be32_to_cpu(fh[6])) << 32;
1333     diff --git a/fs/isofs/export.c b/fs/isofs/export.c
1334     index dd4687f..516eb21 100644
1335     --- a/fs/isofs/export.c
1336     +++ b/fs/isofs/export.c
1337     @@ -179,7 +179,7 @@ static struct dentry *isofs_fh_to_parent(struct super_block *sb,
1338     {
1339     struct isofs_fid *ifid = (struct isofs_fid *)fid;
1340    
1341     - if (fh_type != 2)
1342     + if (fh_len < 2 || fh_type != 2)
1343     return NULL;
1344    
1345     return isofs_export_iget(sb,
1346     diff --git a/fs/jbd/commit.c b/fs/jbd/commit.c
1347     index f2b9a57..9626bc8 100644
1348     --- a/fs/jbd/commit.c
1349     +++ b/fs/jbd/commit.c
1350     @@ -86,7 +86,12 @@ nope:
1351     static void release_data_buffer(struct buffer_head *bh)
1352     {
1353     if (buffer_freed(bh)) {
1354     + WARN_ON_ONCE(buffer_dirty(bh));
1355     clear_buffer_freed(bh);
1356     + clear_buffer_mapped(bh);
1357     + clear_buffer_new(bh);
1358     + clear_buffer_req(bh);
1359     + bh->b_bdev = NULL;
1360     release_buffer_page(bh);
1361     } else
1362     put_bh(bh);
1363     @@ -853,17 +858,35 @@ restart_loop:
1364     * there's no point in keeping a checkpoint record for
1365     * it. */
1366    
1367     - /* A buffer which has been freed while still being
1368     - * journaled by a previous transaction may end up still
1369     - * being dirty here, but we want to avoid writing back
1370     - * that buffer in the future after the "add to orphan"
1371     - * operation been committed, That's not only a performance
1372     - * gain, it also stops aliasing problems if the buffer is
1373     - * left behind for writeback and gets reallocated for another
1374     - * use in a different page. */
1375     - if (buffer_freed(bh) && !jh->b_next_transaction) {
1376     - clear_buffer_freed(bh);
1377     - clear_buffer_jbddirty(bh);
1378     + /*
1379     + * A buffer which has been freed while still being journaled by
1380     + * a previous transaction.
1381     + */
1382     + if (buffer_freed(bh)) {
1383     + /*
1384     + * If the running transaction is the one containing
1385     + * "add to orphan" operation (b_next_transaction !=
1386     + * NULL), we have to wait for that transaction to
1387     + * commit before we can really get rid of the buffer.
1388     + * So just clear b_modified to not confuse transaction
1389     + * credit accounting and refile the buffer to
1390     + * BJ_Forget of the running transaction. If the just
1391     + * committed transaction contains "add to orphan"
1392     + * operation, we can completely invalidate the buffer
1393     + * now. We are rather throughout in that since the
1394     + * buffer may be still accessible when blocksize <
1395     + * pagesize and it is attached to the last partial
1396     + * page.
1397     + */
1398     + jh->b_modified = 0;
1399     + if (!jh->b_next_transaction) {
1400     + clear_buffer_freed(bh);
1401     + clear_buffer_jbddirty(bh);
1402     + clear_buffer_mapped(bh);
1403     + clear_buffer_new(bh);
1404     + clear_buffer_req(bh);
1405     + bh->b_bdev = NULL;
1406     + }
1407     }
1408    
1409     if (buffer_jbddirty(bh)) {
1410     diff --git a/fs/jbd/transaction.c b/fs/jbd/transaction.c
1411     index b2a7e52..841cabc 100644
1412     --- a/fs/jbd/transaction.c
1413     +++ b/fs/jbd/transaction.c
1414     @@ -1845,15 +1845,16 @@ static int __dispose_buffer(struct journal_head *jh, transaction_t *transaction)
1415     * We're outside-transaction here. Either or both of j_running_transaction
1416     * and j_committing_transaction may be NULL.
1417     */
1418     -static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1419     +static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh,
1420     + int partial_page)
1421     {
1422     transaction_t *transaction;
1423     struct journal_head *jh;
1424     int may_free = 1;
1425     - int ret;
1426    
1427     BUFFER_TRACE(bh, "entry");
1428    
1429     +retry:
1430     /*
1431     * It is safe to proceed here without the j_list_lock because the
1432     * buffers cannot be stolen by try_to_free_buffers as long as we are
1433     @@ -1881,10 +1882,18 @@ static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1434     * clear the buffer dirty bit at latest at the moment when the
1435     * transaction marking the buffer as freed in the filesystem
1436     * structures is committed because from that moment on the
1437     - * buffer can be reallocated and used by a different page.
1438     + * block can be reallocated and used by a different page.
1439     * Since the block hasn't been freed yet but the inode has
1440     * already been added to orphan list, it is safe for us to add
1441     * the buffer to BJ_Forget list of the newest transaction.
1442     + *
1443     + * Also we have to clear buffer_mapped flag of a truncated buffer
1444     + * because the buffer_head may be attached to the page straddling
1445     + * i_size (can happen only when blocksize < pagesize) and thus the
1446     + * buffer_head can be reused when the file is extended again. So we end
1447     + * up keeping around invalidated buffers attached to transactions'
1448     + * BJ_Forget list just to stop checkpointing code from cleaning up
1449     + * the transaction this buffer was modified in.
1450     */
1451     transaction = jh->b_transaction;
1452     if (transaction == NULL) {
1453     @@ -1911,13 +1920,9 @@ static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1454     * committed, the buffer won't be needed any
1455     * longer. */
1456     JBUFFER_TRACE(jh, "checkpointed: add to BJ_Forget");
1457     - ret = __dispose_buffer(jh,
1458     + may_free = __dispose_buffer(jh,
1459     journal->j_running_transaction);
1460     - journal_put_journal_head(jh);
1461     - spin_unlock(&journal->j_list_lock);
1462     - jbd_unlock_bh_state(bh);
1463     - spin_unlock(&journal->j_state_lock);
1464     - return ret;
1465     + goto zap_buffer;
1466     } else {
1467     /* There is no currently-running transaction. So the
1468     * orphan record which we wrote for this file must have
1469     @@ -1925,13 +1930,9 @@ static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1470     * the committing transaction, if it exists. */
1471     if (journal->j_committing_transaction) {
1472     JBUFFER_TRACE(jh, "give to committing trans");
1473     - ret = __dispose_buffer(jh,
1474     + may_free = __dispose_buffer(jh,
1475     journal->j_committing_transaction);
1476     - journal_put_journal_head(jh);
1477     - spin_unlock(&journal->j_list_lock);
1478     - jbd_unlock_bh_state(bh);
1479     - spin_unlock(&journal->j_state_lock);
1480     - return ret;
1481     + goto zap_buffer;
1482     } else {
1483     /* The orphan record's transaction has
1484     * committed. We can cleanse this buffer */
1485     @@ -1952,10 +1953,24 @@ static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1486     }
1487     /*
1488     * The buffer is committing, we simply cannot touch
1489     - * it. So we just set j_next_transaction to the
1490     - * running transaction (if there is one) and mark
1491     - * buffer as freed so that commit code knows it should
1492     - * clear dirty bits when it is done with the buffer.
1493     + * it. If the page is straddling i_size we have to wait
1494     + * for commit and try again.
1495     + */
1496     + if (partial_page) {
1497     + tid_t tid = journal->j_committing_transaction->t_tid;
1498     +
1499     + journal_put_journal_head(jh);
1500     + spin_unlock(&journal->j_list_lock);
1501     + jbd_unlock_bh_state(bh);
1502     + spin_unlock(&journal->j_state_lock);
1503     + log_wait_commit(journal, tid);
1504     + goto retry;
1505     + }
1506     + /*
1507     + * OK, buffer won't be reachable after truncate. We just set
1508     + * j_next_transaction to the running transaction (if there is
1509     + * one) and mark buffer as freed so that commit code knows it
1510     + * should clear dirty bits when it is done with the buffer.
1511     */
1512     set_buffer_freed(bh);
1513     if (journal->j_running_transaction && buffer_jbddirty(bh))
1514     @@ -1978,6 +1993,14 @@ static int journal_unmap_buffer(journal_t *journal, struct buffer_head *bh)
1515     }
1516    
1517     zap_buffer:
1518     + /*
1519     + * This is tricky. Although the buffer is truncated, it may be reused
1520     + * if blocksize < pagesize and it is attached to the page straddling
1521     + * EOF. Since the buffer might have been added to BJ_Forget list of the
1522     + * running transaction, journal_get_write_access() won't clear
1523     + * b_modified and credit accounting gets confused. So clear b_modified
1524     + * here. */
1525     + jh->b_modified = 0;
1526     journal_put_journal_head(jh);
1527     zap_buffer_no_jh:
1528     spin_unlock(&journal->j_list_lock);
1529     @@ -2026,7 +2049,8 @@ void journal_invalidatepage(journal_t *journal,
1530     if (offset <= curr_off) {
1531     /* This block is wholly outside the truncation point */
1532     lock_buffer(bh);
1533     - may_free &= journal_unmap_buffer(journal, bh);
1534     + may_free &= journal_unmap_buffer(journal, bh,
1535     + offset > 0);
1536     unlock_buffer(bh);
1537     }
1538     curr_off = next_off;
1539     diff --git a/fs/lockd/mon.c b/fs/lockd/mon.c
1540     index 7ef14b3..606a8dd 100644
1541     --- a/fs/lockd/mon.c
1542     +++ b/fs/lockd/mon.c
1543     @@ -40,6 +40,7 @@ struct nsm_args {
1544     u32 proc;
1545    
1546     char *mon_name;
1547     + char *nodename;
1548     };
1549    
1550     struct nsm_res {
1551     @@ -94,6 +95,7 @@ static int nsm_mon_unmon(struct nsm_handle *nsm, u32 proc, struct nsm_res *res,
1552     .vers = 3,
1553     .proc = NLMPROC_NSM_NOTIFY,
1554     .mon_name = nsm->sm_mon_name,
1555     + .nodename = utsname()->nodename,
1556     };
1557     struct rpc_message msg = {
1558     .rpc_argp = &args,
1559     @@ -430,7 +432,7 @@ static void encode_my_id(struct xdr_stream *xdr, const struct nsm_args *argp)
1560     {
1561     __be32 *p;
1562    
1563     - encode_nsm_string(xdr, utsname()->nodename);
1564     + encode_nsm_string(xdr, argp->nodename);
1565     p = xdr_reserve_space(xdr, 4 + 4 + 4);
1566     *p++ = cpu_to_be32(argp->prog);
1567     *p++ = cpu_to_be32(argp->vers);
1568     diff --git a/fs/nfsd/nfs4idmap.c b/fs/nfsd/nfs4idmap.c
1569     index 322d11c..01b090d 100644
1570     --- a/fs/nfsd/nfs4idmap.c
1571     +++ b/fs/nfsd/nfs4idmap.c
1572     @@ -581,7 +581,7 @@ numeric_name_to_id(struct svc_rqst *rqstp, int type, const char *name, u32 namel
1573     /* Just to make sure it's null-terminated: */
1574     memcpy(buf, name, namelen);
1575     buf[namelen] = '\0';
1576     - ret = kstrtouint(name, 10, id);
1577     + ret = kstrtouint(buf, 10, id);
1578     return ret == 0;
1579     }
1580    
1581     diff --git a/fs/nfsd/nfs4state.c b/fs/nfsd/nfs4state.c
1582     index e79c24e..a2f99d1 100644
1583     --- a/fs/nfsd/nfs4state.c
1584     +++ b/fs/nfsd/nfs4state.c
1585     @@ -3783,6 +3783,7 @@ nfsd4_close(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
1586     memcpy(&close->cl_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
1587    
1588     nfsd4_close_open_stateid(stp);
1589     + release_last_closed_stateid(oo);
1590     oo->oo_last_closed_stid = stp;
1591    
1592     /* place unused nfs4_stateowners on so_close_lru list to be
1593     diff --git a/fs/reiserfs/inode.c b/fs/reiserfs/inode.c
1594     index 494c315..f99c1b4 100644
1595     --- a/fs/reiserfs/inode.c
1596     +++ b/fs/reiserfs/inode.c
1597     @@ -1573,8 +1573,10 @@ struct dentry *reiserfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
1598     reiserfs_warning(sb, "reiserfs-13077",
1599     "nfsd/reiserfs, fhtype=%d, len=%d - odd",
1600     fh_type, fh_len);
1601     - fh_type = 5;
1602     + fh_type = fh_len;
1603     }
1604     + if (fh_len < 2)
1605     + return NULL;
1606    
1607     return reiserfs_get_dentry(sb, fid->raw[0], fid->raw[1],
1608     (fh_type == 3 || fh_type >= 5) ? fid->raw[2] : 0);
1609     @@ -1583,6 +1585,8 @@ struct dentry *reiserfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
1610     struct dentry *reiserfs_fh_to_parent(struct super_block *sb, struct fid *fid,
1611     int fh_len, int fh_type)
1612     {
1613     + if (fh_type > fh_len)
1614     + fh_type = fh_len;
1615     if (fh_type < 4)
1616     return NULL;
1617    
1618     diff --git a/fs/xfs/xfs_export.c b/fs/xfs/xfs_export.c
1619     index 558910f..5703fb8 100644
1620     --- a/fs/xfs/xfs_export.c
1621     +++ b/fs/xfs/xfs_export.c
1622     @@ -195,6 +195,9 @@ xfs_fs_fh_to_parent(struct super_block *sb, struct fid *fid,
1623     struct xfs_fid64 *fid64 = (struct xfs_fid64 *)fid;
1624     struct inode *inode = NULL;
1625    
1626     + if (fh_len < xfs_fileid_length(fileid_type))
1627     + return NULL;
1628     +
1629     switch (fileid_type) {
1630     case FILEID_INO32_GEN_PARENT:
1631     inode = xfs_nfs_get_inode(sb, fid->i32.parent_ino,
1632     diff --git a/include/linux/netfilter/ipset/ip_set_timeout.h b/include/linux/netfilter/ipset/ip_set_timeout.h
1633     index 4792320..41d9cfa 100644
1634     --- a/include/linux/netfilter/ipset/ip_set_timeout.h
1635     +++ b/include/linux/netfilter/ipset/ip_set_timeout.h
1636     @@ -30,6 +30,10 @@ ip_set_timeout_uget(struct nlattr *tb)
1637     {
1638     unsigned int timeout = ip_set_get_h32(tb);
1639    
1640     + /* Normalize to fit into jiffies */
1641     + if (timeout > UINT_MAX/MSEC_PER_SEC)
1642     + timeout = UINT_MAX/MSEC_PER_SEC;
1643     +
1644     /* Userspace supplied TIMEOUT parameter: adjust crazy size */
1645     return timeout == IPSET_NO_TIMEOUT ? IPSET_NO_TIMEOUT - 1 : timeout;
1646     }
1647     diff --git a/include/net/ip_vs.h b/include/net/ip_vs.h
1648     index 72522f0..2389959 100644
1649     --- a/include/net/ip_vs.h
1650     +++ b/include/net/ip_vs.h
1651     @@ -1356,7 +1356,7 @@ static inline void ip_vs_notrack(struct sk_buff *skb)
1652     struct nf_conn *ct = nf_ct_get(skb, &ctinfo);
1653    
1654     if (!ct || !nf_ct_is_untracked(ct)) {
1655     - nf_reset(skb);
1656     + nf_conntrack_put(skb->nfct);
1657     skb->nfct = &nf_ct_untracked_get()->ct_general;
1658     skb->nfctinfo = IP_CT_NEW;
1659     nf_conntrack_get(skb->nfct);
1660     diff --git a/include/net/netfilter/nf_conntrack_ecache.h b/include/net/netfilter/nf_conntrack_ecache.h
1661     index a88fb69..ea6f8a4 100644
1662     --- a/include/net/netfilter/nf_conntrack_ecache.h
1663     +++ b/include/net/netfilter/nf_conntrack_ecache.h
1664     @@ -18,6 +18,7 @@ struct nf_conntrack_ecache {
1665     u16 ctmask; /* bitmask of ct events to be delivered */
1666     u16 expmask; /* bitmask of expect events to be delivered */
1667     u32 pid; /* netlink pid of destroyer */
1668     + struct timer_list timeout;
1669     };
1670    
1671     static inline struct nf_conntrack_ecache *
1672     diff --git a/kernel/debug/kdb/kdb_io.c b/kernel/debug/kdb/kdb_io.c
1673     index bb9520f..572e604 100644
1674     --- a/kernel/debug/kdb/kdb_io.c
1675     +++ b/kernel/debug/kdb/kdb_io.c
1676     @@ -552,6 +552,7 @@ int vkdb_printf(const char *fmt, va_list ap)
1677     {
1678     int diag;
1679     int linecount;
1680     + int colcount;
1681     int logging, saved_loglevel = 0;
1682     int saved_trap_printk;
1683     int got_printf_lock = 0;
1684     @@ -584,6 +585,10 @@ int vkdb_printf(const char *fmt, va_list ap)
1685     if (diag || linecount <= 1)
1686     linecount = 24;
1687    
1688     + diag = kdbgetintenv("COLUMNS", &colcount);
1689     + if (diag || colcount <= 1)
1690     + colcount = 80;
1691     +
1692     diag = kdbgetintenv("LOGGING", &logging);
1693     if (diag)
1694     logging = 0;
1695     @@ -690,7 +695,7 @@ kdb_printit:
1696     gdbstub_msg_write(kdb_buffer, retlen);
1697     } else {
1698     if (dbg_io_ops && !dbg_io_ops->is_console) {
1699     - len = strlen(kdb_buffer);
1700     + len = retlen;
1701     cp = kdb_buffer;
1702     while (len--) {
1703     dbg_io_ops->write_char(*cp);
1704     @@ -709,11 +714,29 @@ kdb_printit:
1705     printk(KERN_INFO "%s", kdb_buffer);
1706     }
1707    
1708     - if (KDB_STATE(PAGER) && strchr(kdb_buffer, '\n'))
1709     - kdb_nextline++;
1710     + if (KDB_STATE(PAGER)) {
1711     + /*
1712     + * Check printed string to decide how to bump the
1713     + * kdb_nextline to control when the more prompt should
1714     + * show up.
1715     + */
1716     + int got = 0;
1717     + len = retlen;
1718     + while (len--) {
1719     + if (kdb_buffer[len] == '\n') {
1720     + kdb_nextline++;
1721     + got = 0;
1722     + } else if (kdb_buffer[len] == '\r') {
1723     + got = 0;
1724     + } else {
1725     + got++;
1726     + }
1727     + }
1728     + kdb_nextline += got / (colcount + 1);
1729     + }
1730    
1731     /* check for having reached the LINES number of printed lines */
1732     - if (kdb_nextline == linecount) {
1733     + if (kdb_nextline >= linecount) {
1734     char buf1[16] = "";
1735     #if defined(CONFIG_SMP)
1736     char buf2[32];
1737     @@ -776,7 +799,7 @@ kdb_printit:
1738     kdb_grepping_flag = 0;
1739     kdb_printf("\n");
1740     } else if (buf1[0] == ' ') {
1741     - kdb_printf("\n");
1742     + kdb_printf("\r");
1743     suspend_grep = 1; /* for this recursion */
1744     } else if (buf1[0] == '\n') {
1745     kdb_nextline = linecount - 1;
1746     diff --git a/kernel/module.c b/kernel/module.c
1747     index 78ac6ec..61ea75e 100644
1748     --- a/kernel/module.c
1749     +++ b/kernel/module.c
1750     @@ -2729,6 +2729,10 @@ static int check_module_license_and_versions(struct module *mod)
1751     if (strcmp(mod->name, "driverloader") == 0)
1752     add_taint_module(mod, TAINT_PROPRIETARY_MODULE);
1753    
1754     + /* lve claims to be GPL but upstream won't provide source */
1755     + if (strcmp(mod->name, "lve") == 0)
1756     + add_taint_module(mod, TAINT_PROPRIETARY_MODULE);
1757     +
1758     #ifdef CONFIG_MODVERSIONS
1759     if ((mod->num_syms && !mod->crcs)
1760     || (mod->num_gpl_syms && !mod->gpl_crcs)
1761     diff --git a/kernel/timer.c b/kernel/timer.c
1762     index a297ffc..6dfdb72 100644
1763     --- a/kernel/timer.c
1764     +++ b/kernel/timer.c
1765     @@ -63,6 +63,7 @@ EXPORT_SYMBOL(jiffies_64);
1766     #define TVR_SIZE (1 << TVR_BITS)
1767     #define TVN_MASK (TVN_SIZE - 1)
1768     #define TVR_MASK (TVR_SIZE - 1)
1769     +#define MAX_TVAL ((unsigned long)((1ULL << (TVR_BITS + 4*TVN_BITS)) - 1))
1770    
1771     struct tvec {
1772     struct list_head vec[TVN_SIZE];
1773     @@ -356,11 +357,12 @@ static void internal_add_timer(struct tvec_base *base, struct timer_list *timer)
1774     vec = base->tv1.vec + (base->timer_jiffies & TVR_MASK);
1775     } else {
1776     int i;
1777     - /* If the timeout is larger than 0xffffffff on 64-bit
1778     - * architectures then we use the maximum timeout:
1779     + /* If the timeout is larger than MAX_TVAL (on 64-bit
1780     + * architectures or with CONFIG_BASE_SMALL=1) then we
1781     + * use the maximum timeout.
1782     */
1783     - if (idx > 0xffffffffUL) {
1784     - idx = 0xffffffffUL;
1785     + if (idx > MAX_TVAL) {
1786     + idx = MAX_TVAL;
1787     expires = idx + base->timer_jiffies;
1788     }
1789     i = (expires >> (TVR_BITS + 3 * TVN_BITS)) & TVN_MASK;
1790     diff --git a/mm/shmem.c b/mm/shmem.c
1791     index 9d65a02..40383cd 100644
1792     --- a/mm/shmem.c
1793     +++ b/mm/shmem.c
1794     @@ -2018,12 +2018,14 @@ static struct dentry *shmem_fh_to_dentry(struct super_block *sb,
1795     {
1796     struct inode *inode;
1797     struct dentry *dentry = NULL;
1798     - u64 inum = fid->raw[2];
1799     - inum = (inum << 32) | fid->raw[1];
1800     + u64 inum;
1801    
1802     if (fh_len < 3)
1803     return NULL;
1804    
1805     + inum = fid->raw[2];
1806     + inum = (inum << 32) | fid->raw[1];
1807     +
1808     inode = ilookup5(sb, (unsigned long)(inum + fid->raw[0]),
1809     shmem_match, fid->raw);
1810     if (inode) {
1811     diff --git a/net/core/pktgen.c b/net/core/pktgen.c
1812     index b81369b..8dae76f 100644
1813     --- a/net/core/pktgen.c
1814     +++ b/net/core/pktgen.c
1815     @@ -2932,7 +2932,7 @@ static struct sk_buff *fill_packet_ipv6(struct net_device *odev,
1816     sizeof(struct ipv6hdr) - sizeof(struct udphdr) -
1817     pkt_dev->pkt_overhead;
1818    
1819     - if (datalen < sizeof(struct pktgen_hdr)) {
1820     + if (datalen < 0 || datalen < sizeof(struct pktgen_hdr)) {
1821     datalen = sizeof(struct pktgen_hdr);
1822     if (net_ratelimit())
1823     pr_info("increased datalen to %d\n", datalen);
1824     diff --git a/net/ipv4/netfilter/nf_nat_sip.c b/net/ipv4/netfilter/nf_nat_sip.c
1825     index 57932c4..566be2d 100644
1826     --- a/net/ipv4/netfilter/nf_nat_sip.c
1827     +++ b/net/ipv4/netfilter/nf_nat_sip.c
1828     @@ -148,7 +148,7 @@ static unsigned int ip_nat_sip(struct sk_buff *skb, unsigned int dataoff,
1829     if (ct_sip_parse_header_uri(ct, *dptr, NULL, *datalen,
1830     hdr, NULL, &matchoff, &matchlen,
1831     &addr, &port) > 0) {
1832     - unsigned int matchend, poff, plen, buflen, n;
1833     + unsigned int olen, matchend, poff, plen, buflen, n;
1834     char buffer[sizeof("nnn.nnn.nnn.nnn:nnnnn")];
1835    
1836     /* We're only interested in headers related to this
1837     @@ -163,11 +163,12 @@ static unsigned int ip_nat_sip(struct sk_buff *skb, unsigned int dataoff,
1838     goto next;
1839     }
1840    
1841     + olen = *datalen;
1842     if (!map_addr(skb, dataoff, dptr, datalen, matchoff, matchlen,
1843     &addr, port))
1844     return NF_DROP;
1845    
1846     - matchend = matchoff + matchlen;
1847     + matchend = matchoff + matchlen + *datalen - olen;
1848    
1849     /* The maddr= parameter (RFC 2361) specifies where to send
1850     * the reply. */
1851     @@ -501,7 +502,10 @@ static unsigned int ip_nat_sdp_media(struct sk_buff *skb, unsigned int dataoff,
1852     ret = nf_ct_expect_related(rtcp_exp);
1853     if (ret == 0)
1854     break;
1855     - else if (ret != -EBUSY) {
1856     + else if (ret == -EBUSY) {
1857     + nf_ct_unexpect_related(rtp_exp);
1858     + continue;
1859     + } else if (ret < 0) {
1860     nf_ct_unexpect_related(rtp_exp);
1861     port = 0;
1862     break;
1863     diff --git a/net/netfilter/ipvs/ip_vs_ctl.c b/net/netfilter/ipvs/ip_vs_ctl.c
1864     index 97e7380..cbc5bfd 100644
1865     --- a/net/netfilter/ipvs/ip_vs_ctl.c
1866     +++ b/net/netfilter/ipvs/ip_vs_ctl.c
1867     @@ -1521,11 +1521,12 @@ static int ip_vs_dst_event(struct notifier_block *this, unsigned long event,
1868     {
1869     struct net_device *dev = ptr;
1870     struct net *net = dev_net(dev);
1871     + struct netns_ipvs *ipvs = net_ipvs(net);
1872     struct ip_vs_service *svc;
1873     struct ip_vs_dest *dest;
1874     unsigned int idx;
1875    
1876     - if (event != NETDEV_UNREGISTER)
1877     + if (event != NETDEV_UNREGISTER || !ipvs)
1878     return NOTIFY_DONE;
1879     IP_VS_DBG(3, "%s() dev=%s\n", __func__, dev->name);
1880     EnterFunction(2);
1881     @@ -1551,7 +1552,7 @@ static int ip_vs_dst_event(struct notifier_block *this, unsigned long event,
1882     }
1883     }
1884    
1885     - list_for_each_entry(dest, &net_ipvs(net)->dest_trash, n_list) {
1886     + list_for_each_entry(dest, &ipvs->dest_trash, n_list) {
1887     __ip_vs_dev_reset(dest, dev);
1888     }
1889     mutex_unlock(&__ip_vs_mutex);
1890     diff --git a/net/netfilter/nf_conntrack_core.c b/net/netfilter/nf_conntrack_core.c
1891     index 729f157..9a171b2 100644
1892     --- a/net/netfilter/nf_conntrack_core.c
1893     +++ b/net/netfilter/nf_conntrack_core.c
1894     @@ -249,12 +249,15 @@ static void death_by_event(unsigned long ul_conntrack)
1895     {
1896     struct nf_conn *ct = (void *)ul_conntrack;
1897     struct net *net = nf_ct_net(ct);
1898     + struct nf_conntrack_ecache *ecache = nf_ct_ecache_find(ct);
1899     +
1900     + BUG_ON(ecache == NULL);
1901    
1902     if (nf_conntrack_event(IPCT_DESTROY, ct) < 0) {
1903     /* bad luck, let's retry again */
1904     - ct->timeout.expires = jiffies +
1905     + ecache->timeout.expires = jiffies +
1906     (random32() % net->ct.sysctl_events_retry_timeout);
1907     - add_timer(&ct->timeout);
1908     + add_timer(&ecache->timeout);
1909     return;
1910     }
1911     /* we've got the event delivered, now it's dying */
1912     @@ -268,6 +271,9 @@ static void death_by_event(unsigned long ul_conntrack)
1913     void nf_ct_insert_dying_list(struct nf_conn *ct)
1914     {
1915     struct net *net = nf_ct_net(ct);
1916     + struct nf_conntrack_ecache *ecache = nf_ct_ecache_find(ct);
1917     +
1918     + BUG_ON(ecache == NULL);
1919    
1920     /* add this conntrack to the dying list */
1921     spin_lock_bh(&nf_conntrack_lock);
1922     @@ -275,10 +281,10 @@ void nf_ct_insert_dying_list(struct nf_conn *ct)
1923     &net->ct.dying);
1924     spin_unlock_bh(&nf_conntrack_lock);
1925     /* set a new timer to retry event delivery */
1926     - setup_timer(&ct->timeout, death_by_event, (unsigned long)ct);
1927     - ct->timeout.expires = jiffies +
1928     + setup_timer(&ecache->timeout, death_by_event, (unsigned long)ct);
1929     + ecache->timeout.expires = jiffies +
1930     (random32() % net->ct.sysctl_events_retry_timeout);
1931     - add_timer(&ct->timeout);
1932     + add_timer(&ecache->timeout);
1933     }
1934     EXPORT_SYMBOL_GPL(nf_ct_insert_dying_list);
1935    
1936     diff --git a/net/netfilter/nf_conntrack_expect.c b/net/netfilter/nf_conntrack_expect.c
1937     index 4147ba3..e41ec84 100644
1938     --- a/net/netfilter/nf_conntrack_expect.c
1939     +++ b/net/netfilter/nf_conntrack_expect.c
1940     @@ -361,23 +361,6 @@ static void evict_oldest_expect(struct nf_conn *master,
1941     }
1942     }
1943    
1944     -static inline int refresh_timer(struct nf_conntrack_expect *i)
1945     -{
1946     - struct nf_conn_help *master_help = nfct_help(i->master);
1947     - const struct nf_conntrack_expect_policy *p;
1948     -
1949     - if (!del_timer(&i->timeout))
1950     - return 0;
1951     -
1952     - p = &rcu_dereference_protected(
1953     - master_help->helper,
1954     - lockdep_is_held(&nf_conntrack_lock)
1955     - )->expect_policy[i->class];
1956     - i->timeout.expires = jiffies + p->timeout * HZ;
1957     - add_timer(&i->timeout);
1958     - return 1;
1959     -}
1960     -
1961     static inline int __nf_ct_expect_check(struct nf_conntrack_expect *expect)
1962     {
1963     const struct nf_conntrack_expect_policy *p;
1964     @@ -386,7 +369,7 @@ static inline int __nf_ct_expect_check(struct nf_conntrack_expect *expect)
1965     struct nf_conn_help *master_help = nfct_help(master);
1966     struct nf_conntrack_helper *helper;
1967     struct net *net = nf_ct_exp_net(expect);
1968     - struct hlist_node *n;
1969     + struct hlist_node *n, *next;
1970     unsigned int h;
1971     int ret = 1;
1972    
1973     @@ -395,12 +378,12 @@ static inline int __nf_ct_expect_check(struct nf_conntrack_expect *expect)
1974     goto out;
1975     }
1976     h = nf_ct_expect_dst_hash(&expect->tuple);
1977     - hlist_for_each_entry(i, n, &net->ct.expect_hash[h], hnode) {
1978     + hlist_for_each_entry_safe(i, n, next, &net->ct.expect_hash[h], hnode) {
1979     if (expect_matches(i, expect)) {
1980     - /* Refresh timer: if it's dying, ignore.. */
1981     - if (refresh_timer(i)) {
1982     - ret = 0;
1983     - goto out;
1984     + if (del_timer(&i->timeout)) {
1985     + nf_ct_unlink_expect(i);
1986     + nf_ct_expect_put(i);
1987     + break;
1988     }
1989     } else if (expect_clash(i, expect)) {
1990     ret = -EBUSY;
1991     diff --git a/net/netfilter/xt_hashlimit.c b/net/netfilter/xt_hashlimit.c
1992     index d95f9c9..2195eb0 100644
1993     --- a/net/netfilter/xt_hashlimit.c
1994     +++ b/net/netfilter/xt_hashlimit.c
1995     @@ -389,8 +389,7 @@ static void htable_put(struct xt_hashlimit_htable *hinfo)
1996     #define CREDITS_PER_JIFFY POW2_BELOW32(MAX_CPJ)
1997    
1998     /* Precision saver. */
1999     -static inline u_int32_t
2000     -user2credits(u_int32_t user)
2001     +static u32 user2credits(u32 user)
2002     {
2003     /* If multiplying would overflow... */
2004     if (user > 0xFFFFFFFF / (HZ*CREDITS_PER_JIFFY))
2005     @@ -400,7 +399,7 @@ user2credits(u_int32_t user)
2006     return (user * HZ * CREDITS_PER_JIFFY) / XT_HASHLIMIT_SCALE;
2007     }
2008    
2009     -static inline void rateinfo_recalc(struct dsthash_ent *dh, unsigned long now)
2010     +static void rateinfo_recalc(struct dsthash_ent *dh, unsigned long now)
2011     {
2012     dh->rateinfo.credit += (now - dh->rateinfo.prev) * CREDITS_PER_JIFFY;
2013     if (dh->rateinfo.credit > dh->rateinfo.credit_cap)
2014     @@ -535,8 +534,7 @@ hashlimit_mt(const struct sk_buff *skb, struct xt_action_param *par)
2015     dh->rateinfo.prev = jiffies;
2016     dh->rateinfo.credit = user2credits(hinfo->cfg.avg *
2017     hinfo->cfg.burst);
2018     - dh->rateinfo.credit_cap = user2credits(hinfo->cfg.avg *
2019     - hinfo->cfg.burst);
2020     + dh->rateinfo.credit_cap = dh->rateinfo.credit;
2021     dh->rateinfo.cost = user2credits(hinfo->cfg.avg);
2022     } else {
2023     /* update expiration timeout */
2024     diff --git a/net/netfilter/xt_limit.c b/net/netfilter/xt_limit.c
2025     index 32b7a57..a4c1e45 100644
2026     --- a/net/netfilter/xt_limit.c
2027     +++ b/net/netfilter/xt_limit.c
2028     @@ -88,8 +88,7 @@ limit_mt(const struct sk_buff *skb, struct xt_action_param *par)
2029     }
2030    
2031     /* Precision saver. */
2032     -static u_int32_t
2033     -user2credits(u_int32_t user)
2034     +static u32 user2credits(u32 user)
2035     {
2036     /* If multiplying would overflow... */
2037     if (user > 0xFFFFFFFF / (HZ*CREDITS_PER_JIFFY))
2038     @@ -118,12 +117,12 @@ static int limit_mt_check(const struct xt_mtchk_param *par)
2039    
2040     /* For SMP, we only want to use one set of state. */
2041     r->master = priv;
2042     + /* User avg in seconds * XT_LIMIT_SCALE: convert to jiffies *
2043     + 128. */
2044     + priv->prev = jiffies;
2045     + priv->credit = user2credits(r->avg * r->burst); /* Credits full. */
2046     if (r->cost == 0) {
2047     - /* User avg in seconds * XT_LIMIT_SCALE: convert to jiffies *
2048     - 128. */
2049     - priv->prev = jiffies;
2050     - priv->credit = user2credits(r->avg * r->burst); /* Credits full. */
2051     - r->credit_cap = user2credits(r->avg * r->burst); /* Credits full. */
2052     + r->credit_cap = priv->credit; /* Credits full. */
2053     r->cost = user2credits(r->avg);
2054     }
2055     return 0;
2056     diff --git a/net/netfilter/xt_set.c b/net/netfilter/xt_set.c
2057     index 0ec8138..c6f7db7 100644
2058     --- a/net/netfilter/xt_set.c
2059     +++ b/net/netfilter/xt_set.c
2060     @@ -16,6 +16,7 @@
2061    
2062     #include <linux/netfilter/x_tables.h>
2063     #include <linux/netfilter/xt_set.h>
2064     +#include <linux/netfilter/ipset/ip_set_timeout.h>
2065    
2066     MODULE_LICENSE("GPL");
2067     MODULE_AUTHOR("Jozsef Kadlecsik <kadlec@blackhole.kfki.hu>");
2068     @@ -44,6 +45,14 @@ const struct ip_set_adt_opt n = { \
2069     .cmdflags = cfs, \
2070     .timeout = t, \
2071     }
2072     +#define ADT_MOPT(n, f, d, fs, cfs, t) \
2073     +struct ip_set_adt_opt n = { \
2074     + .family = f, \
2075     + .dim = d, \
2076     + .flags = fs, \
2077     + .cmdflags = cfs, \
2078     + .timeout = t, \
2079     +}
2080    
2081     /* Revision 0 interface: backward compatible with netfilter/iptables */
2082    
2083     @@ -296,11 +305,15 @@ static unsigned int
2084     set_target_v2(struct sk_buff *skb, const struct xt_action_param *par)
2085     {
2086     const struct xt_set_info_target_v2 *info = par->targinfo;
2087     - ADT_OPT(add_opt, par->family, info->add_set.dim,
2088     - info->add_set.flags, info->flags, info->timeout);
2089     + ADT_MOPT(add_opt, par->family, info->add_set.dim,
2090     + info->add_set.flags, info->flags, info->timeout);
2091     ADT_OPT(del_opt, par->family, info->del_set.dim,
2092     info->del_set.flags, 0, UINT_MAX);
2093    
2094     + /* Normalize to fit into jiffies */
2095     + if (add_opt.timeout != IPSET_NO_TIMEOUT &&
2096     + add_opt.timeout > UINT_MAX/MSEC_PER_SEC)
2097     + add_opt.timeout = UINT_MAX/MSEC_PER_SEC;
2098     if (info->add_set.index != IPSET_INVALID_ID)
2099     ip_set_add(info->add_set.index, skb, par, &add_opt);
2100     if (info->del_set.index != IPSET_INVALID_ID)
2101     diff --git a/net/sunrpc/xprtsock.c b/net/sunrpc/xprtsock.c
2102     index b88c6bf..00ff343 100644
2103     --- a/net/sunrpc/xprtsock.c
2104     +++ b/net/sunrpc/xprtsock.c
2105     @@ -1028,6 +1028,16 @@ static void xs_udp_data_ready(struct sock *sk, int len)
2106     read_unlock_bh(&sk->sk_callback_lock);
2107     }
2108    
2109     +/*
2110     + * Helper function to force a TCP close if the server is sending
2111     + * junk and/or it has put us in CLOSE_WAIT
2112     + */
2113     +static void xs_tcp_force_close(struct rpc_xprt *xprt)
2114     +{
2115     + set_bit(XPRT_CONNECTION_CLOSE, &xprt->state);
2116     + xprt_force_disconnect(xprt);
2117     +}
2118     +
2119     static inline void xs_tcp_read_fraghdr(struct rpc_xprt *xprt, struct xdr_skb_reader *desc)
2120     {
2121     struct sock_xprt *transport = container_of(xprt, struct sock_xprt, xprt);
2122     @@ -1054,7 +1064,7 @@ static inline void xs_tcp_read_fraghdr(struct rpc_xprt *xprt, struct xdr_skb_rea
2123     /* Sanity check of the record length */
2124     if (unlikely(transport->tcp_reclen < 8)) {
2125     dprintk("RPC: invalid TCP record fragment length\n");
2126     - xprt_force_disconnect(xprt);
2127     + xs_tcp_force_close(xprt);
2128     return;
2129     }
2130     dprintk("RPC: reading TCP record fragment of length %d\n",
2131     @@ -1135,7 +1145,7 @@ static inline void xs_tcp_read_calldir(struct sock_xprt *transport,
2132     break;
2133     default:
2134     dprintk("RPC: invalid request message type\n");
2135     - xprt_force_disconnect(&transport->xprt);
2136     + xs_tcp_force_close(&transport->xprt);
2137     }
2138     xs_tcp_check_fraghdr(transport);
2139     }
2140     @@ -1458,6 +1468,8 @@ static void xs_tcp_cancel_linger_timeout(struct rpc_xprt *xprt)
2141     static void xs_sock_mark_closed(struct rpc_xprt *xprt)
2142     {
2143     smp_mb__before_clear_bit();
2144     + clear_bit(XPRT_CONNECTION_ABORT, &xprt->state);
2145     + clear_bit(XPRT_CONNECTION_CLOSE, &xprt->state);
2146     clear_bit(XPRT_CLOSE_WAIT, &xprt->state);
2147     clear_bit(XPRT_CLOSING, &xprt->state);
2148     smp_mb__after_clear_bit();
2149     @@ -1515,8 +1527,8 @@ static void xs_tcp_state_change(struct sock *sk)
2150     break;
2151     case TCP_CLOSE_WAIT:
2152     /* The server initiated a shutdown of the socket */
2153     - xprt_force_disconnect(xprt);
2154     xprt->connect_cookie++;
2155     + xs_tcp_force_close(xprt);
2156     case TCP_CLOSING:
2157     /*
2158     * If the server closed down the connection, make sure that
2159     @@ -2159,8 +2171,7 @@ static void xs_tcp_setup_socket(struct work_struct *work)
2160     /* We're probably in TIME_WAIT. Get rid of existing socket,
2161     * and retry
2162     */
2163     - set_bit(XPRT_CONNECTION_CLOSE, &xprt->state);
2164     - xprt_force_disconnect(xprt);
2165     + xs_tcp_force_close(xprt);
2166     break;
2167     case -ECONNREFUSED:
2168     case -ECONNRESET:
2169     diff --git a/sound/pci/ac97/ac97_codec.c b/sound/pci/ac97/ac97_codec.c
2170     index 9473fca..8b0f996 100644
2171     --- a/sound/pci/ac97/ac97_codec.c
2172     +++ b/sound/pci/ac97/ac97_codec.c
2173     @@ -1271,6 +1271,8 @@ static int snd_ac97_cvol_new(struct snd_card *card, char *name, int reg, unsigne
2174     tmp.index = ac97->num;
2175     kctl = snd_ctl_new1(&tmp, ac97);
2176     }
2177     + if (!kctl)
2178     + return -ENOMEM;
2179     if (reg >= AC97_PHONE && reg <= AC97_PCM)
2180     set_tlv_db_scale(kctl, db_scale_5bit_12db_max);
2181     else
2182     diff --git a/sound/pci/emu10k1/emu10k1_main.c b/sound/pci/emu10k1/emu10k1_main.c
2183     index 7549240..a78fdf4 100644
2184     --- a/sound/pci/emu10k1/emu10k1_main.c
2185     +++ b/sound/pci/emu10k1/emu10k1_main.c
2186     @@ -1416,6 +1416,15 @@ static struct snd_emu_chip_details emu_chip_details[] = {
2187     .ca0108_chip = 1,
2188     .spk71 = 1,
2189     .emu_model = EMU_MODEL_EMU1010B}, /* EMU 1010 new revision */
2190     + /* Tested by Maxim Kachur <mcdebugger@duganet.ru> 17th Oct 2012. */
2191     + /* This is MAEM8986, 0202 is MAEM8980 */
2192     + {.vendor = 0x1102, .device = 0x0008, .subsystem = 0x40071102,
2193     + .driver = "Audigy2", .name = "E-mu 1010 PCIe [MAEM8986]",
2194     + .id = "EMU1010",
2195     + .emu10k2_chip = 1,
2196     + .ca0108_chip = 1,
2197     + .spk71 = 1,
2198     + .emu_model = EMU_MODEL_EMU1010B}, /* EMU 1010 PCIe */
2199     /* Tested by James@superbug.co.uk 8th July 2005. */
2200     /* This is MAEM8810, 0202 is MAEM8820 */
2201     {.vendor = 0x1102, .device = 0x0004, .subsystem = 0x40011102,
2202     diff --git a/sound/pci/hda/patch_cirrus.c b/sound/pci/hda/patch_cirrus.c
2203     index c83ccdb..2bc6c51 100644
2204     --- a/sound/pci/hda/patch_cirrus.c
2205     +++ b/sound/pci/hda/patch_cirrus.c
2206     @@ -1427,7 +1427,7 @@ static int patch_cs420x(struct hda_codec *codec)
2207     return 0;
2208    
2209     error:
2210     - kfree(codec->spec);
2211     + cs_free(codec);
2212     codec->spec = NULL;
2213     return err;
2214     }
2215     @@ -1984,7 +1984,7 @@ static int patch_cs4210(struct hda_codec *codec)
2216     return 0;
2217    
2218     error:
2219     - kfree(codec->spec);
2220     + cs_free(codec);
2221     codec->spec = NULL;
2222     return err;
2223     }
2224     @@ -2009,7 +2009,7 @@ static int patch_cs4213(struct hda_codec *codec)
2225     return 0;
2226    
2227     error:
2228     - kfree(codec->spec);
2229     + cs_free(codec);
2230     codec->spec = NULL;
2231     return err;
2232     }
2233     diff --git a/sound/pci/hda/patch_realtek.c b/sound/pci/hda/patch_realtek.c
2234     index e7cb4bd..94b765b 100644
2235     --- a/sound/pci/hda/patch_realtek.c
2236     +++ b/sound/pci/hda/patch_realtek.c
2237     @@ -602,6 +602,8 @@ static void alc_line_automute(struct hda_codec *codec)
2238     {
2239     struct alc_spec *spec = codec->spec;
2240    
2241     + if (spec->autocfg.line_out_type == AUTO_PIN_SPEAKER_OUT)
2242     + return;
2243     /* check LO jack only when it's different from HP */
2244     if (spec->autocfg.line_out_pins[0] == spec->autocfg.hp_pins[0])
2245     return;
2246     @@ -2663,8 +2665,10 @@ static const char *alc_get_line_out_pfx(struct alc_spec *spec, int ch,
2247     return "PCM";
2248     break;
2249     }
2250     - if (snd_BUG_ON(ch >= ARRAY_SIZE(channel_name)))
2251     + if (ch >= ARRAY_SIZE(channel_name)) {
2252     + snd_BUG();
2253     return "PCM";
2254     + }
2255    
2256     return channel_name[ch];
2257     }
2258     diff --git a/sound/soc/codecs/wm2200.c b/sound/soc/codecs/wm2200.c
2259     index 32682c1..c8bff6d 100644
2260     --- a/sound/soc/codecs/wm2200.c
2261     +++ b/sound/soc/codecs/wm2200.c
2262     @@ -1028,7 +1028,7 @@ SOC_DOUBLE_R_TLV("OUT2 Digital Volume", WM2200_DAC_DIGITAL_VOLUME_2L,
2263     WM2200_DAC_DIGITAL_VOLUME_2R, WM2200_OUT2L_VOL_SHIFT, 0x9f, 0,
2264     digital_tlv),
2265     SOC_DOUBLE("OUT2 Switch", WM2200_PDM_1, WM2200_SPK1L_MUTE_SHIFT,
2266     - WM2200_SPK1R_MUTE_SHIFT, 1, 0),
2267     + WM2200_SPK1R_MUTE_SHIFT, 1, 1),
2268     };
2269    
2270     WM2200_MIXER_ENUMS(OUT1L, WM2200_OUT1LMIX_INPUT_1_SOURCE);
2271     @@ -2091,6 +2091,7 @@ static __devinit int wm2200_i2c_probe(struct i2c_client *i2c,
2272    
2273     switch (wm2200->rev) {
2274     case 0:
2275     + case 1:
2276     ret = regmap_register_patch(wm2200->regmap, wm2200_reva_patch,
2277     ARRAY_SIZE(wm2200_reva_patch));
2278     if (ret != 0) {
2279     diff --git a/sound/soc/omap/omap-abe-twl6040.c b/sound/soc/omap/omap-abe-twl6040.c
2280     index 93bb8ee..9c2f090 100644
2281     --- a/sound/soc/omap/omap-abe-twl6040.c
2282     +++ b/sound/soc/omap/omap-abe-twl6040.c
2283     @@ -181,7 +181,7 @@ static int omap_abe_twl6040_init(struct snd_soc_pcm_runtime *rtd)
2284     twl6040_disconnect_pin(dapm, pdata->has_hf, "Ext Spk");
2285     twl6040_disconnect_pin(dapm, pdata->has_ep, "Earphone Spk");
2286     twl6040_disconnect_pin(dapm, pdata->has_aux, "Line Out");
2287     - twl6040_disconnect_pin(dapm, pdata->has_vibra, "Vinrator");
2288     + twl6040_disconnect_pin(dapm, pdata->has_vibra, "Vibrator");
2289     twl6040_disconnect_pin(dapm, pdata->has_hsmic, "Headset Mic");
2290     twl6040_disconnect_pin(dapm, pdata->has_mainmic, "Main Handset Mic");
2291     twl6040_disconnect_pin(dapm, pdata->has_submic, "Sub Handset Mic");
2292     diff --git a/sound/soc/sh/fsi.c b/sound/soc/sh/fsi.c
2293     index 74ed2df..91b7287 100644
2294     --- a/sound/soc/sh/fsi.c
2295     +++ b/sound/soc/sh/fsi.c
2296     @@ -20,6 +20,7 @@
2297     #include <linux/sh_dma.h>
2298     #include <linux/slab.h>
2299     #include <linux/module.h>
2300     +#include <linux/workqueue.h>
2301     #include <sound/soc.h>
2302     #include <sound/sh_fsi.h>
2303    
2304     @@ -199,7 +200,7 @@ struct fsi_stream {
2305     */
2306     struct dma_chan *chan;
2307     struct sh_dmae_slave slave; /* see fsi_handler_init() */
2308     - struct tasklet_struct tasklet;
2309     + struct work_struct work;
2310     dma_addr_t dma;
2311     };
2312    
2313     @@ -968,9 +969,9 @@ static dma_addr_t fsi_dma_get_area(struct fsi_stream *io)
2314     return io->dma + samples_to_bytes(runtime, io->buff_sample_pos);
2315     }
2316    
2317     -static void fsi_dma_do_tasklet(unsigned long data)
2318     +static void fsi_dma_do_work(struct work_struct *work)
2319     {
2320     - struct fsi_stream *io = (struct fsi_stream *)data;
2321     + struct fsi_stream *io = container_of(work, struct fsi_stream, work);
2322     struct fsi_priv *fsi = fsi_stream_to_priv(io);
2323     struct dma_chan *chan;
2324     struct snd_soc_dai *dai;
2325     @@ -1023,7 +1024,7 @@ static void fsi_dma_do_tasklet(unsigned long data)
2326     * FIXME
2327     *
2328     * In DMAEngine case, codec and FSI cannot be started simultaneously
2329     - * since FSI is using tasklet.
2330     + * since FSI is using the scheduler work queue.
2331     * Therefore, in capture case, probably FSI FIFO will have got
2332     * overflow error in this point.
2333     * in that case, DMA cannot start transfer until error was cleared.
2334     @@ -1047,7 +1048,7 @@ static bool fsi_dma_filter(struct dma_chan *chan, void *param)
2335    
2336     static int fsi_dma_transfer(struct fsi_priv *fsi, struct fsi_stream *io)
2337     {
2338     - tasklet_schedule(&io->tasklet);
2339     + schedule_work(&io->work);
2340    
2341     return 0;
2342     }
2343     @@ -1087,14 +1088,14 @@ static int fsi_dma_probe(struct fsi_priv *fsi, struct fsi_stream *io)
2344     if (!io->chan)
2345     return -EIO;
2346    
2347     - tasklet_init(&io->tasklet, fsi_dma_do_tasklet, (unsigned long)io);
2348     + INIT_WORK(&io->work, fsi_dma_do_work);
2349    
2350     return 0;
2351     }
2352    
2353     static int fsi_dma_remove(struct fsi_priv *fsi, struct fsi_stream *io)
2354     {
2355     - tasklet_kill(&io->tasklet);
2356     + cancel_work_sync(&io->work);
2357    
2358     fsi_stream_stop(fsi, io);
2359