Contents of /trunk/kernel26-alx/patches-3.10/0164-3.10.65-all-fixes.patch
Parent Directory | Revision Log
Revision 2672 -
(show annotations)
(download)
Tue Jul 21 16:46:35 2015 UTC (9 years, 2 months ago) by niro
File size: 51228 byte(s)
Tue Jul 21 16:46:35 2015 UTC (9 years, 2 months ago) by niro
File size: 51228 byte(s)
-3.10.84-alx-r1
1 | diff --git a/Documentation/ramoops.txt b/Documentation/ramoops.txt |
2 | index 69b3cac4749d..5d8675615e59 100644 |
3 | --- a/Documentation/ramoops.txt |
4 | +++ b/Documentation/ramoops.txt |
5 | @@ -14,11 +14,19 @@ survive after a restart. |
6 | |
7 | 1. Ramoops concepts |
8 | |
9 | -Ramoops uses a predefined memory area to store the dump. The start and size of |
10 | -the memory area are set using two variables: |
11 | +Ramoops uses a predefined memory area to store the dump. The start and size |
12 | +and type of the memory area are set using three variables: |
13 | * "mem_address" for the start |
14 | * "mem_size" for the size. The memory size will be rounded down to a |
15 | power of two. |
16 | + * "mem_type" to specifiy if the memory type (default is pgprot_writecombine). |
17 | + |
18 | +Typically the default value of mem_type=0 should be used as that sets the pstore |
19 | +mapping to pgprot_writecombine. Setting mem_type=1 attempts to use |
20 | +pgprot_noncached, which only works on some platforms. This is because pstore |
21 | +depends on atomic operations. At least on ARM, pgprot_noncached causes the |
22 | +memory to be mapped strongly ordered, and atomic operations on strongly ordered |
23 | +memory are implementation defined, and won't work on many ARMs such as omaps. |
24 | |
25 | The memory area is divided into "record_size" chunks (also rounded down to |
26 | power of two) and each oops/panic writes a "record_size" chunk of |
27 | @@ -55,6 +63,7 @@ Setting the ramoops parameters can be done in 2 different manners: |
28 | static struct ramoops_platform_data ramoops_data = { |
29 | .mem_size = <...>, |
30 | .mem_address = <...>, |
31 | + .mem_type = <...>, |
32 | .record_size = <...>, |
33 | .dump_oops = <...>, |
34 | .ecc = <...>, |
35 | diff --git a/Makefile b/Makefile |
36 | index e5b63fb3d0e1..7889b38766db 100644 |
37 | --- a/Makefile |
38 | +++ b/Makefile |
39 | @@ -1,6 +1,6 @@ |
40 | VERSION = 3 |
41 | PATCHLEVEL = 10 |
42 | -SUBLEVEL = 64 |
43 | +SUBLEVEL = 65 |
44 | EXTRAVERSION = |
45 | NAME = TOSSUG Baby Fish |
46 | |
47 | diff --git a/arch/arm/mach-mvebu/coherency.c b/arch/arm/mach-mvebu/coherency.c |
48 | index 8278960066c3..3ee701f1d38e 100644 |
49 | --- a/arch/arm/mach-mvebu/coherency.c |
50 | +++ b/arch/arm/mach-mvebu/coherency.c |
51 | @@ -141,6 +141,29 @@ int __init coherency_init(void) |
52 | { |
53 | struct device_node *np; |
54 | |
55 | + /* |
56 | + * The coherency fabric is needed: |
57 | + * - For coherency between processors on Armada XP, so only |
58 | + * when SMP is enabled. |
59 | + * - For coherency between the processor and I/O devices, but |
60 | + * this coherency requires many pre-requisites (write |
61 | + * allocate cache policy, shareable pages, SMP bit set) that |
62 | + * are only meant in SMP situations. |
63 | + * |
64 | + * Note that this means that on Armada 370, there is currently |
65 | + * no way to use hardware I/O coherency, because even when |
66 | + * CONFIG_SMP is enabled, is_smp() returns false due to the |
67 | + * Armada 370 being a single-core processor. To lift this |
68 | + * limitation, we would have to find a way to make the cache |
69 | + * policy set to write-allocate (on all Armada SoCs), and to |
70 | + * set the shareable attribute in page tables (on all Armada |
71 | + * SoCs except the Armada 370). Unfortunately, such decisions |
72 | + * are taken very early in the kernel boot process, at a point |
73 | + * where we don't know yet on which SoC we are running. |
74 | + */ |
75 | + if (!is_smp()) |
76 | + return 0; |
77 | + |
78 | np = of_find_matching_node(NULL, of_coherency_table); |
79 | if (np) { |
80 | pr_info("Initializing Coherency fabric\n"); |
81 | diff --git a/arch/x86/include/asm/vsyscall.h b/arch/x86/include/asm/vsyscall.h |
82 | index 2a46ca720afc..2874be9aef0a 100644 |
83 | --- a/arch/x86/include/asm/vsyscall.h |
84 | +++ b/arch/x86/include/asm/vsyscall.h |
85 | @@ -34,7 +34,7 @@ static inline unsigned int __getcpu(void) |
86 | native_read_tscp(&p); |
87 | } else { |
88 | /* Load per CPU data from GDT */ |
89 | - asm("lsl %1,%0" : "=r" (p) : "r" (__PER_CPU_SEG)); |
90 | + asm volatile ("lsl %1,%0" : "=r" (p) : "r" (__PER_CPU_SEG)); |
91 | } |
92 | |
93 | return p; |
94 | diff --git a/arch/x86/kernel/cpu/perf_event_intel_uncore.c b/arch/x86/kernel/cpu/perf_event_intel_uncore.c |
95 | index 8aac56bda7dc..7185af255fb5 100644 |
96 | --- a/arch/x86/kernel/cpu/perf_event_intel_uncore.c |
97 | +++ b/arch/x86/kernel/cpu/perf_event_intel_uncore.c |
98 | @@ -2657,6 +2657,17 @@ static struct intel_uncore_box *uncore_event_to_box(struct perf_event *event) |
99 | return uncore_pmu_to_box(uncore_event_to_pmu(event), smp_processor_id()); |
100 | } |
101 | |
102 | +/* |
103 | + * Using uncore_pmu_event_init pmu event_init callback |
104 | + * as a detection point for uncore events. |
105 | + */ |
106 | +static int uncore_pmu_event_init(struct perf_event *event); |
107 | + |
108 | +static bool is_uncore_event(struct perf_event *event) |
109 | +{ |
110 | + return event->pmu->event_init == uncore_pmu_event_init; |
111 | +} |
112 | + |
113 | static int |
114 | uncore_collect_events(struct intel_uncore_box *box, struct perf_event *leader, bool dogrp) |
115 | { |
116 | @@ -2671,13 +2682,18 @@ uncore_collect_events(struct intel_uncore_box *box, struct perf_event *leader, b |
117 | return -EINVAL; |
118 | |
119 | n = box->n_events; |
120 | - box->event_list[n] = leader; |
121 | - n++; |
122 | + |
123 | + if (is_uncore_event(leader)) { |
124 | + box->event_list[n] = leader; |
125 | + n++; |
126 | + } |
127 | + |
128 | if (!dogrp) |
129 | return n; |
130 | |
131 | list_for_each_entry(event, &leader->sibling_list, group_entry) { |
132 | - if (event->state <= PERF_EVENT_STATE_OFF) |
133 | + if (!is_uncore_event(event) || |
134 | + event->state <= PERF_EVENT_STATE_OFF) |
135 | continue; |
136 | |
137 | if (n >= max_count) |
138 | diff --git a/arch/x86/vdso/vma.c b/arch/x86/vdso/vma.c |
139 | index 431e87544411..ab6ba35a9357 100644 |
140 | --- a/arch/x86/vdso/vma.c |
141 | +++ b/arch/x86/vdso/vma.c |
142 | @@ -117,30 +117,45 @@ subsys_initcall(init_vdso); |
143 | |
144 | struct linux_binprm; |
145 | |
146 | -/* Put the vdso above the (randomized) stack with another randomized offset. |
147 | - This way there is no hole in the middle of address space. |
148 | - To save memory make sure it is still in the same PTE as the stack top. |
149 | - This doesn't give that many random bits */ |
150 | +/* |
151 | + * Put the vdso above the (randomized) stack with another randomized |
152 | + * offset. This way there is no hole in the middle of address space. |
153 | + * To save memory make sure it is still in the same PTE as the stack |
154 | + * top. This doesn't give that many random bits. |
155 | + * |
156 | + * Note that this algorithm is imperfect: the distribution of the vdso |
157 | + * start address within a PMD is biased toward the end. |
158 | + * |
159 | + * Only used for the 64-bit and x32 vdsos. |
160 | + */ |
161 | static unsigned long vdso_addr(unsigned long start, unsigned len) |
162 | { |
163 | unsigned long addr, end; |
164 | unsigned offset; |
165 | - end = (start + PMD_SIZE - 1) & PMD_MASK; |
166 | + |
167 | + /* |
168 | + * Round up the start address. It can start out unaligned as a result |
169 | + * of stack start randomization. |
170 | + */ |
171 | + start = PAGE_ALIGN(start); |
172 | + |
173 | + /* Round the lowest possible end address up to a PMD boundary. */ |
174 | + end = (start + len + PMD_SIZE - 1) & PMD_MASK; |
175 | if (end >= TASK_SIZE_MAX) |
176 | end = TASK_SIZE_MAX; |
177 | end -= len; |
178 | - /* This loses some more bits than a modulo, but is cheaper */ |
179 | - offset = get_random_int() & (PTRS_PER_PTE - 1); |
180 | - addr = start + (offset << PAGE_SHIFT); |
181 | - if (addr >= end) |
182 | - addr = end; |
183 | + |
184 | + if (end > start) { |
185 | + offset = get_random_int() % (((end - start) >> PAGE_SHIFT) + 1); |
186 | + addr = start + (offset << PAGE_SHIFT); |
187 | + } else { |
188 | + addr = start; |
189 | + } |
190 | |
191 | /* |
192 | - * page-align it here so that get_unmapped_area doesn't |
193 | - * align it wrongfully again to the next page. addr can come in 4K |
194 | - * unaligned here as a result of stack start randomization. |
195 | + * Forcibly align the final address in case we have a hardware |
196 | + * issue that requires alignment for performance reasons. |
197 | */ |
198 | - addr = PAGE_ALIGN(addr); |
199 | addr = align_vdso_addr(addr); |
200 | |
201 | return addr; |
202 | diff --git a/block/genhd.c b/block/genhd.c |
203 | index e670148c3773..7694dffe9f0e 100644 |
204 | --- a/block/genhd.c |
205 | +++ b/block/genhd.c |
206 | @@ -1070,9 +1070,16 @@ int disk_expand_part_tbl(struct gendisk *disk, int partno) |
207 | struct disk_part_tbl *old_ptbl = disk->part_tbl; |
208 | struct disk_part_tbl *new_ptbl; |
209 | int len = old_ptbl ? old_ptbl->len : 0; |
210 | - int target = partno + 1; |
211 | + int i, target; |
212 | size_t size; |
213 | - int i; |
214 | + |
215 | + /* |
216 | + * check for int overflow, since we can get here from blkpg_ioctl() |
217 | + * with a user passed 'partno'. |
218 | + */ |
219 | + target = partno + 1; |
220 | + if (target < 0) |
221 | + return -EINVAL; |
222 | |
223 | /* disk_max_parts() is zero during initialization, ignore if so */ |
224 | if (disk_max_parts(disk) && target > disk_max_parts(disk)) |
225 | diff --git a/drivers/base/bus.c b/drivers/base/bus.c |
226 | index d414331b480e..558d562f4901 100644 |
227 | --- a/drivers/base/bus.c |
228 | +++ b/drivers/base/bus.c |
229 | @@ -242,13 +242,15 @@ static ssize_t store_drivers_probe(struct bus_type *bus, |
230 | const char *buf, size_t count) |
231 | { |
232 | struct device *dev; |
233 | + int err = -EINVAL; |
234 | |
235 | dev = bus_find_device_by_name(bus, NULL, buf); |
236 | if (!dev) |
237 | return -ENODEV; |
238 | - if (bus_rescan_devices_helper(dev, NULL) != 0) |
239 | - return -EINVAL; |
240 | - return count; |
241 | + if (bus_rescan_devices_helper(dev, NULL) == 0) |
242 | + err = count; |
243 | + put_device(dev); |
244 | + return err; |
245 | } |
246 | |
247 | static struct device *next_device(struct klist_iter *i) |
248 | diff --git a/drivers/hid/hid-core.c b/drivers/hid/hid-core.c |
249 | index 81d0e6e1f754..2bd798a7d9aa 100644 |
250 | --- a/drivers/hid/hid-core.c |
251 | +++ b/drivers/hid/hid-core.c |
252 | @@ -1687,6 +1687,7 @@ static const struct hid_device_id hid_have_special_driver[] = { |
253 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_ERGO_525V) }, |
254 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_EASYPEN_I405X) }, |
255 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_MOUSEPEN_I608X) }, |
256 | + { HID_USB_DEVICE(USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2) }, |
257 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_EASYPEN_M610X) }, |
258 | { HID_USB_DEVICE(USB_VENDOR_ID_LABTEC, USB_DEVICE_ID_LABTEC_WIRELESS_KEYBOARD) }, |
259 | { HID_USB_DEVICE(USB_VENDOR_ID_LCPOWER, USB_DEVICE_ID_LCPOWER_LC1000 ) }, |
260 | diff --git a/drivers/hid/hid-ids.h b/drivers/hid/hid-ids.h |
261 | index a1e431f830e3..45c593dbf5cd 100644 |
262 | --- a/drivers/hid/hid-ids.h |
263 | +++ b/drivers/hid/hid-ids.h |
264 | @@ -478,6 +478,7 @@ |
265 | #define USB_DEVICE_ID_KYE_GPEN_560 0x5003 |
266 | #define USB_DEVICE_ID_KYE_EASYPEN_I405X 0x5010 |
267 | #define USB_DEVICE_ID_KYE_MOUSEPEN_I608X 0x5011 |
268 | +#define USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2 0x501a |
269 | #define USB_DEVICE_ID_KYE_EASYPEN_M610X 0x5013 |
270 | |
271 | #define USB_VENDOR_ID_LABTEC 0x1020 |
272 | diff --git a/drivers/hid/hid-input.c b/drivers/hid/hid-input.c |
273 | index 012880a2228c..03a6acffed5d 100644 |
274 | --- a/drivers/hid/hid-input.c |
275 | +++ b/drivers/hid/hid-input.c |
276 | @@ -317,6 +317,9 @@ static const struct hid_device_id hid_battery_quirks[] = { |
277 | USB_DEVICE_ID_APPLE_ALU_WIRELESS_2011_ANSI), |
278 | HID_BATTERY_QUIRK_PERCENT | HID_BATTERY_QUIRK_FEATURE }, |
279 | { HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, |
280 | + USB_DEVICE_ID_APPLE_ALU_WIRELESS_2011_ISO), |
281 | + HID_BATTERY_QUIRK_PERCENT | HID_BATTERY_QUIRK_FEATURE }, |
282 | + { HID_BLUETOOTH_DEVICE(USB_VENDOR_ID_APPLE, |
283 | USB_DEVICE_ID_APPLE_ALU_WIRELESS_ANSI), |
284 | HID_BATTERY_QUIRK_PERCENT | HID_BATTERY_QUIRK_FEATURE }, |
285 | {} |
286 | diff --git a/drivers/hid/hid-kye.c b/drivers/hid/hid-kye.c |
287 | index 843f2dd55200..973eed788cc6 100644 |
288 | --- a/drivers/hid/hid-kye.c |
289 | +++ b/drivers/hid/hid-kye.c |
290 | @@ -303,6 +303,7 @@ static __u8 *kye_report_fixup(struct hid_device *hdev, __u8 *rdesc, |
291 | } |
292 | break; |
293 | case USB_DEVICE_ID_KYE_MOUSEPEN_I608X: |
294 | + case USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2: |
295 | if (*rsize == MOUSEPEN_I608X_RDESC_ORIG_SIZE) { |
296 | rdesc = mousepen_i608x_rdesc_fixed; |
297 | *rsize = sizeof(mousepen_i608x_rdesc_fixed); |
298 | @@ -383,6 +384,7 @@ static int kye_probe(struct hid_device *hdev, const struct hid_device_id *id) |
299 | switch (id->product) { |
300 | case USB_DEVICE_ID_KYE_EASYPEN_I405X: |
301 | case USB_DEVICE_ID_KYE_MOUSEPEN_I608X: |
302 | + case USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2: |
303 | case USB_DEVICE_ID_KYE_EASYPEN_M610X: |
304 | ret = kye_tablet_enable(hdev); |
305 | if (ret) { |
306 | @@ -406,6 +408,8 @@ static const struct hid_device_id kye_devices[] = { |
307 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, |
308 | USB_DEVICE_ID_KYE_MOUSEPEN_I608X) }, |
309 | { HID_USB_DEVICE(USB_VENDOR_ID_KYE, |
310 | + USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2) }, |
311 | + { HID_USB_DEVICE(USB_VENDOR_ID_KYE, |
312 | USB_DEVICE_ID_KYE_EASYPEN_M610X) }, |
313 | { } |
314 | }; |
315 | diff --git a/drivers/hid/hid-roccat-pyra.c b/drivers/hid/hid-roccat-pyra.c |
316 | index d4f1e3bee590..264ddc4a0118 100644 |
317 | --- a/drivers/hid/hid-roccat-pyra.c |
318 | +++ b/drivers/hid/hid-roccat-pyra.c |
319 | @@ -35,6 +35,8 @@ static struct class *pyra_class; |
320 | static void profile_activated(struct pyra_device *pyra, |
321 | unsigned int new_profile) |
322 | { |
323 | + if (new_profile >= ARRAY_SIZE(pyra->profile_settings)) |
324 | + return; |
325 | pyra->actual_profile = new_profile; |
326 | pyra->actual_cpi = pyra->profile_settings[pyra->actual_profile].y_cpi; |
327 | } |
328 | @@ -236,9 +238,11 @@ static ssize_t pyra_sysfs_write_settings(struct file *fp, |
329 | if (off != 0 || count != PYRA_SIZE_SETTINGS) |
330 | return -EINVAL; |
331 | |
332 | - mutex_lock(&pyra->pyra_lock); |
333 | - |
334 | settings = (struct pyra_settings const *)buf; |
335 | + if (settings->startup_profile >= ARRAY_SIZE(pyra->profile_settings)) |
336 | + return -EINVAL; |
337 | + |
338 | + mutex_lock(&pyra->pyra_lock); |
339 | |
340 | retval = pyra_set_settings(usb_dev, settings); |
341 | if (retval) { |
342 | diff --git a/drivers/hid/i2c-hid/i2c-hid.c b/drivers/hid/i2c-hid/i2c-hid.c |
343 | index 2b1799a3b212..469daa04dadb 100644 |
344 | --- a/drivers/hid/i2c-hid/i2c-hid.c |
345 | +++ b/drivers/hid/i2c-hid/i2c-hid.c |
346 | @@ -134,6 +134,7 @@ struct i2c_hid { |
347 | * descriptor. */ |
348 | unsigned int bufsize; /* i2c buffer size */ |
349 | char *inbuf; /* Input buffer */ |
350 | + char *rawbuf; /* Raw Input buffer */ |
351 | char *cmdbuf; /* Command buffer */ |
352 | char *argsbuf; /* Command arguments buffer */ |
353 | |
354 | @@ -340,7 +341,7 @@ static int i2c_hid_hwreset(struct i2c_client *client) |
355 | static void i2c_hid_get_input(struct i2c_hid *ihid) |
356 | { |
357 | int ret, ret_size; |
358 | - int size = le16_to_cpu(ihid->hdesc.wMaxInputLength); |
359 | + int size = ihid->bufsize; |
360 | |
361 | ret = i2c_master_recv(ihid->client, ihid->inbuf, size); |
362 | if (ret != size) { |
363 | @@ -471,9 +472,11 @@ static void i2c_hid_find_max_report(struct hid_device *hid, unsigned int type, |
364 | static void i2c_hid_free_buffers(struct i2c_hid *ihid) |
365 | { |
366 | kfree(ihid->inbuf); |
367 | + kfree(ihid->rawbuf); |
368 | kfree(ihid->argsbuf); |
369 | kfree(ihid->cmdbuf); |
370 | ihid->inbuf = NULL; |
371 | + ihid->rawbuf = NULL; |
372 | ihid->cmdbuf = NULL; |
373 | ihid->argsbuf = NULL; |
374 | ihid->bufsize = 0; |
375 | @@ -489,10 +492,11 @@ static int i2c_hid_alloc_buffers(struct i2c_hid *ihid, size_t report_size) |
376 | report_size; /* report */ |
377 | |
378 | ihid->inbuf = kzalloc(report_size, GFP_KERNEL); |
379 | + ihid->rawbuf = kzalloc(report_size, GFP_KERNEL); |
380 | ihid->argsbuf = kzalloc(args_len, GFP_KERNEL); |
381 | ihid->cmdbuf = kzalloc(sizeof(union command) + args_len, GFP_KERNEL); |
382 | |
383 | - if (!ihid->inbuf || !ihid->argsbuf || !ihid->cmdbuf) { |
384 | + if (!ihid->inbuf || !ihid->rawbuf || !ihid->argsbuf || !ihid->cmdbuf) { |
385 | i2c_hid_free_buffers(ihid); |
386 | return -ENOMEM; |
387 | } |
388 | @@ -519,12 +523,12 @@ static int i2c_hid_get_raw_report(struct hid_device *hid, |
389 | |
390 | ret = i2c_hid_get_report(client, |
391 | report_type == HID_FEATURE_REPORT ? 0x03 : 0x01, |
392 | - report_number, ihid->inbuf, ask_count); |
393 | + report_number, ihid->rawbuf, ask_count); |
394 | |
395 | if (ret < 0) |
396 | return ret; |
397 | |
398 | - ret_count = ihid->inbuf[0] | (ihid->inbuf[1] << 8); |
399 | + ret_count = ihid->rawbuf[0] | (ihid->rawbuf[1] << 8); |
400 | |
401 | if (ret_count <= 2) |
402 | return 0; |
403 | @@ -533,7 +537,7 @@ static int i2c_hid_get_raw_report(struct hid_device *hid, |
404 | |
405 | /* The query buffer contains the size, dropping it in the reply */ |
406 | count = min(count, ret_count - 2); |
407 | - memcpy(buf, ihid->inbuf + 2, count); |
408 | + memcpy(buf, ihid->rawbuf + 2, count); |
409 | |
410 | return count; |
411 | } |
412 | diff --git a/drivers/hid/usbhid/hid-quirks.c b/drivers/hid/usbhid/hid-quirks.c |
413 | index 0db9a67278ba..5b46a79dcb1f 100644 |
414 | --- a/drivers/hid/usbhid/hid-quirks.c |
415 | +++ b/drivers/hid/usbhid/hid-quirks.c |
416 | @@ -110,6 +110,7 @@ static const struct hid_blacklist { |
417 | { USB_VENDOR_ID_CHICONY, USB_DEVICE_ID_CHICONY_WIRELESS, HID_QUIRK_MULTI_INPUT }, |
418 | { USB_VENDOR_ID_SIGMA_MICRO, USB_DEVICE_ID_SIGMA_MICRO_KEYBOARD, HID_QUIRK_NO_INIT_REPORTS }, |
419 | { USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_MOUSEPEN_I608X, HID_QUIRK_MULTI_INPUT }, |
420 | + { USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_MOUSEPEN_I608X_2, HID_QUIRK_MULTI_INPUT }, |
421 | { USB_VENDOR_ID_KYE, USB_DEVICE_ID_KYE_EASYPEN_M610X, HID_QUIRK_MULTI_INPUT }, |
422 | { USB_VENDOR_ID_NTRIG, USB_DEVICE_ID_NTRIG_DUOSENSE, HID_QUIRK_NO_INIT_REPORTS }, |
423 | { USB_VENDOR_ID_SYNAPTICS, USB_DEVICE_ID_SYNAPTICS_LTS1, HID_QUIRK_NO_INIT_REPORTS }, |
424 | diff --git a/drivers/iommu/intel-iommu.c b/drivers/iommu/intel-iommu.c |
425 | index 6771e3c94801..db4e10d4c7f5 100644 |
426 | --- a/drivers/iommu/intel-iommu.c |
427 | +++ b/drivers/iommu/intel-iommu.c |
428 | @@ -1796,7 +1796,7 @@ static int __domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn, |
429 | struct dma_pte *first_pte = NULL, *pte = NULL; |
430 | phys_addr_t uninitialized_var(pteval); |
431 | int addr_width = agaw_to_width(domain->agaw) - VTD_PAGE_SHIFT; |
432 | - unsigned long sg_res; |
433 | + unsigned long sg_res = 0; |
434 | unsigned int largepage_lvl = 0; |
435 | unsigned long lvl_pages = 0; |
436 | |
437 | @@ -1807,10 +1807,8 @@ static int __domain_mapping(struct dmar_domain *domain, unsigned long iov_pfn, |
438 | |
439 | prot &= DMA_PTE_READ | DMA_PTE_WRITE | DMA_PTE_SNP; |
440 | |
441 | - if (sg) |
442 | - sg_res = 0; |
443 | - else { |
444 | - sg_res = nr_pages + 1; |
445 | + if (!sg) { |
446 | + sg_res = nr_pages; |
447 | pteval = ((phys_addr_t)phys_pfn << VTD_PAGE_SHIFT) | prot; |
448 | } |
449 | |
450 | diff --git a/drivers/mtd/ubi/upd.c b/drivers/mtd/ubi/upd.c |
451 | index ec2c2dc1c1ca..2a1b6e037e1a 100644 |
452 | --- a/drivers/mtd/ubi/upd.c |
453 | +++ b/drivers/mtd/ubi/upd.c |
454 | @@ -133,6 +133,10 @@ int ubi_start_update(struct ubi_device *ubi, struct ubi_volume *vol, |
455 | ubi_assert(!vol->updating && !vol->changing_leb); |
456 | vol->updating = 1; |
457 | |
458 | + vol->upd_buf = vmalloc(ubi->leb_size); |
459 | + if (!vol->upd_buf) |
460 | + return -ENOMEM; |
461 | + |
462 | err = set_update_marker(ubi, vol); |
463 | if (err) |
464 | return err; |
465 | @@ -152,14 +156,12 @@ int ubi_start_update(struct ubi_device *ubi, struct ubi_volume *vol, |
466 | err = clear_update_marker(ubi, vol, 0); |
467 | if (err) |
468 | return err; |
469 | + |
470 | + vfree(vol->upd_buf); |
471 | vol->updating = 0; |
472 | return 0; |
473 | } |
474 | |
475 | - vol->upd_buf = vmalloc(ubi->leb_size); |
476 | - if (!vol->upd_buf) |
477 | - return -ENOMEM; |
478 | - |
479 | vol->upd_ebs = div_u64(bytes + vol->usable_leb_size - 1, |
480 | vol->usable_leb_size); |
481 | vol->upd_bytes = bytes; |
482 | diff --git a/drivers/mtd/ubi/wl.c b/drivers/mtd/ubi/wl.c |
483 | index c95bfb183c62..49e570abe58b 100644 |
484 | --- a/drivers/mtd/ubi/wl.c |
485 | +++ b/drivers/mtd/ubi/wl.c |
486 | @@ -1209,7 +1209,6 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, |
487 | |
488 | err = do_sync_erase(ubi, e1, vol_id, lnum, 0); |
489 | if (err) { |
490 | - kmem_cache_free(ubi_wl_entry_slab, e1); |
491 | if (e2) |
492 | kmem_cache_free(ubi_wl_entry_slab, e2); |
493 | goto out_ro; |
494 | @@ -1223,10 +1222,8 @@ static int wear_leveling_worker(struct ubi_device *ubi, struct ubi_work *wrk, |
495 | dbg_wl("PEB %d (LEB %d:%d) was put meanwhile, erase", |
496 | e2->pnum, vol_id, lnum); |
497 | err = do_sync_erase(ubi, e2, vol_id, lnum, 0); |
498 | - if (err) { |
499 | - kmem_cache_free(ubi_wl_entry_slab, e2); |
500 | + if (err) |
501 | goto out_ro; |
502 | - } |
503 | } |
504 | |
505 | dbg_wl("done"); |
506 | @@ -1262,10 +1259,9 @@ out_not_moved: |
507 | |
508 | ubi_free_vid_hdr(ubi, vid_hdr); |
509 | err = do_sync_erase(ubi, e2, vol_id, lnum, torture); |
510 | - if (err) { |
511 | - kmem_cache_free(ubi_wl_entry_slab, e2); |
512 | + if (err) |
513 | goto out_ro; |
514 | - } |
515 | + |
516 | mutex_unlock(&ubi->move_mutex); |
517 | return 0; |
518 | |
519 | diff --git a/drivers/net/can/usb/peak_usb/pcan_usb_core.c b/drivers/net/can/usb/peak_usb/pcan_usb_core.c |
520 | index a0f647f92bf5..3a220d2f2ee1 100644 |
521 | --- a/drivers/net/can/usb/peak_usb/pcan_usb_core.c |
522 | +++ b/drivers/net/can/usb/peak_usb/pcan_usb_core.c |
523 | @@ -727,7 +727,7 @@ static int peak_usb_create_dev(struct peak_usb_adapter *peak_usb_adapter, |
524 | dev->cmd_buf = kmalloc(PCAN_USB_MAX_CMD_LEN, GFP_KERNEL); |
525 | if (!dev->cmd_buf) { |
526 | err = -ENOMEM; |
527 | - goto lbl_set_intf_data; |
528 | + goto lbl_free_candev; |
529 | } |
530 | |
531 | dev->udev = usb_dev; |
532 | @@ -766,7 +766,7 @@ static int peak_usb_create_dev(struct peak_usb_adapter *peak_usb_adapter, |
533 | err = register_candev(netdev); |
534 | if (err) { |
535 | dev_err(&intf->dev, "couldn't register CAN device: %d\n", err); |
536 | - goto lbl_free_cmd_buf; |
537 | + goto lbl_restore_intf_data; |
538 | } |
539 | |
540 | if (dev->prev_siblings) |
541 | @@ -779,14 +779,14 @@ static int peak_usb_create_dev(struct peak_usb_adapter *peak_usb_adapter, |
542 | if (dev->adapter->dev_init) { |
543 | err = dev->adapter->dev_init(dev); |
544 | if (err) |
545 | - goto lbl_free_cmd_buf; |
546 | + goto lbl_unregister_candev; |
547 | } |
548 | |
549 | /* set bus off */ |
550 | if (dev->adapter->dev_set_bus) { |
551 | err = dev->adapter->dev_set_bus(dev, 0); |
552 | if (err) |
553 | - goto lbl_free_cmd_buf; |
554 | + goto lbl_unregister_candev; |
555 | } |
556 | |
557 | /* get device number early */ |
558 | @@ -798,11 +798,14 @@ static int peak_usb_create_dev(struct peak_usb_adapter *peak_usb_adapter, |
559 | |
560 | return 0; |
561 | |
562 | -lbl_free_cmd_buf: |
563 | - kfree(dev->cmd_buf); |
564 | +lbl_unregister_candev: |
565 | + unregister_candev(netdev); |
566 | |
567 | -lbl_set_intf_data: |
568 | +lbl_restore_intf_data: |
569 | usb_set_intfdata(intf, dev->prev_siblings); |
570 | + kfree(dev->cmd_buf); |
571 | + |
572 | +lbl_free_candev: |
573 | free_candev(netdev); |
574 | |
575 | return err; |
576 | diff --git a/drivers/net/can/usb/peak_usb/pcan_usb_pro.c b/drivers/net/can/usb/peak_usb/pcan_usb_pro.c |
577 | index 263dd921edc4..f7f796a2c50b 100644 |
578 | --- a/drivers/net/can/usb/peak_usb/pcan_usb_pro.c |
579 | +++ b/drivers/net/can/usb/peak_usb/pcan_usb_pro.c |
580 | @@ -333,8 +333,6 @@ static int pcan_usb_pro_send_req(struct peak_usb_device *dev, int req_id, |
581 | if (!(dev->state & PCAN_USB_STATE_CONNECTED)) |
582 | return 0; |
583 | |
584 | - memset(req_addr, '\0', req_size); |
585 | - |
586 | req_type = USB_TYPE_VENDOR | USB_RECIP_OTHER; |
587 | |
588 | switch (req_id) { |
589 | @@ -345,6 +343,7 @@ static int pcan_usb_pro_send_req(struct peak_usb_device *dev, int req_id, |
590 | default: |
591 | p = usb_rcvctrlpipe(dev->udev, 0); |
592 | req_type |= USB_DIR_IN; |
593 | + memset(req_addr, '\0', req_size); |
594 | break; |
595 | } |
596 | |
597 | diff --git a/drivers/net/wireless/ath/ath5k/qcu.c b/drivers/net/wireless/ath/ath5k/qcu.c |
598 | index 65fe929529a8..3bfd0b88016e 100644 |
599 | --- a/drivers/net/wireless/ath/ath5k/qcu.c |
600 | +++ b/drivers/net/wireless/ath/ath5k/qcu.c |
601 | @@ -225,13 +225,7 @@ ath5k_hw_setup_tx_queue(struct ath5k_hw *ah, enum ath5k_tx_queue queue_type, |
602 | } else { |
603 | switch (queue_type) { |
604 | case AR5K_TX_QUEUE_DATA: |
605 | - for (queue = AR5K_TX_QUEUE_ID_DATA_MIN; |
606 | - ah->ah_txq[queue].tqi_type != |
607 | - AR5K_TX_QUEUE_INACTIVE; queue++) { |
608 | - |
609 | - if (queue > AR5K_TX_QUEUE_ID_DATA_MAX) |
610 | - return -EINVAL; |
611 | - } |
612 | + queue = queue_info->tqi_subtype; |
613 | break; |
614 | case AR5K_TX_QUEUE_UAPSD: |
615 | queue = AR5K_TX_QUEUE_ID_UAPSD; |
616 | diff --git a/drivers/net/wireless/ath/ath9k/hw.h b/drivers/net/wireless/ath/ath9k/hw.h |
617 | index ae3034374bc4..d7d9e311089f 100644 |
618 | --- a/drivers/net/wireless/ath/ath9k/hw.h |
619 | +++ b/drivers/net/wireless/ath/ath9k/hw.h |
620 | @@ -215,8 +215,8 @@ |
621 | #define AH_WOW_BEACON_MISS BIT(3) |
622 | |
623 | enum ath_hw_txq_subtype { |
624 | - ATH_TXQ_AC_BE = 0, |
625 | - ATH_TXQ_AC_BK = 1, |
626 | + ATH_TXQ_AC_BK = 0, |
627 | + ATH_TXQ_AC_BE = 1, |
628 | ATH_TXQ_AC_VI = 2, |
629 | ATH_TXQ_AC_VO = 3, |
630 | }; |
631 | diff --git a/drivers/net/wireless/ath/ath9k/mac.c b/drivers/net/wireless/ath/ath9k/mac.c |
632 | index 566109a40fb3..941b08b71308 100644 |
633 | --- a/drivers/net/wireless/ath/ath9k/mac.c |
634 | +++ b/drivers/net/wireless/ath/ath9k/mac.c |
635 | @@ -311,14 +311,7 @@ int ath9k_hw_setuptxqueue(struct ath_hw *ah, enum ath9k_tx_queue type, |
636 | q = ATH9K_NUM_TX_QUEUES - 3; |
637 | break; |
638 | case ATH9K_TX_QUEUE_DATA: |
639 | - for (q = 0; q < ATH9K_NUM_TX_QUEUES; q++) |
640 | - if (ah->txq[q].tqi_type == |
641 | - ATH9K_TX_QUEUE_INACTIVE) |
642 | - break; |
643 | - if (q == ATH9K_NUM_TX_QUEUES) { |
644 | - ath_err(common, "No available TX queue\n"); |
645 | - return -1; |
646 | - } |
647 | + q = qinfo->tqi_subtype; |
648 | break; |
649 | default: |
650 | ath_err(common, "Invalid TX queue type: %u\n", type); |
651 | diff --git a/drivers/pci/probe.c b/drivers/pci/probe.c |
652 | index ea37072e8bf2..034a4d2964d6 100644 |
653 | --- a/drivers/pci/probe.c |
654 | +++ b/drivers/pci/probe.c |
655 | @@ -210,14 +210,17 @@ int __pci_read_base(struct pci_dev *dev, enum pci_bar_type type, |
656 | res->flags |= IORESOURCE_SIZEALIGN; |
657 | if (res->flags & IORESOURCE_IO) { |
658 | l &= PCI_BASE_ADDRESS_IO_MASK; |
659 | + sz &= PCI_BASE_ADDRESS_IO_MASK; |
660 | mask = PCI_BASE_ADDRESS_IO_MASK & (u32) IO_SPACE_LIMIT; |
661 | } else { |
662 | l &= PCI_BASE_ADDRESS_MEM_MASK; |
663 | + sz &= PCI_BASE_ADDRESS_MEM_MASK; |
664 | mask = (u32)PCI_BASE_ADDRESS_MEM_MASK; |
665 | } |
666 | } else { |
667 | res->flags |= (l & IORESOURCE_ROM_ENABLE); |
668 | l &= PCI_ROM_ADDRESS_MASK; |
669 | + sz &= PCI_ROM_ADDRESS_MASK; |
670 | mask = (u32)PCI_ROM_ADDRESS_MASK; |
671 | } |
672 | |
673 | diff --git a/drivers/tty/serial/samsung.c b/drivers/tty/serial/samsung.c |
674 | index 0c8a9fa2be6c..b8366b154fb9 100644 |
675 | --- a/drivers/tty/serial/samsung.c |
676 | +++ b/drivers/tty/serial/samsung.c |
677 | @@ -534,11 +534,15 @@ static void s3c24xx_serial_pm(struct uart_port *port, unsigned int level, |
678 | unsigned int old) |
679 | { |
680 | struct s3c24xx_uart_port *ourport = to_ourport(port); |
681 | + int timeout = 10000; |
682 | |
683 | ourport->pm_level = level; |
684 | |
685 | switch (level) { |
686 | case 3: |
687 | + while (--timeout && !s3c24xx_serial_txempty_nofifo(port)) |
688 | + udelay(100); |
689 | + |
690 | if (!IS_ERR(ourport->baudclk)) |
691 | clk_disable_unprepare(ourport->baudclk); |
692 | |
693 | diff --git a/drivers/usb/class/cdc-acm.c b/drivers/usb/class/cdc-acm.c |
694 | index 1e71f918eb9f..2800776b2e91 100644 |
695 | --- a/drivers/usb/class/cdc-acm.c |
696 | +++ b/drivers/usb/class/cdc-acm.c |
697 | @@ -1087,10 +1087,11 @@ next_desc: |
698 | } else { |
699 | control_interface = usb_ifnum_to_if(usb_dev, union_header->bMasterInterface0); |
700 | data_interface = usb_ifnum_to_if(usb_dev, (data_interface_num = union_header->bSlaveInterface0)); |
701 | - if (!control_interface || !data_interface) { |
702 | - dev_dbg(&intf->dev, "no interfaces\n"); |
703 | - return -ENODEV; |
704 | - } |
705 | + } |
706 | + |
707 | + if (!control_interface || !data_interface) { |
708 | + dev_dbg(&intf->dev, "no interfaces\n"); |
709 | + return -ENODEV; |
710 | } |
711 | |
712 | if (data_interface_num != call_interface_num) |
713 | @@ -1365,6 +1366,7 @@ alloc_fail8: |
714 | &dev_attr_wCountryCodes); |
715 | device_remove_file(&acm->control->dev, |
716 | &dev_attr_iCountryCodeRelDate); |
717 | + kfree(acm->country_codes); |
718 | } |
719 | device_remove_file(&acm->control->dev, &dev_attr_bmCapabilities); |
720 | alloc_fail7: |
721 | diff --git a/drivers/xen/swiotlb-xen.c b/drivers/xen/swiotlb-xen.c |
722 | index 1d94316f0ea4..301b08496478 100644 |
723 | --- a/drivers/xen/swiotlb-xen.c |
724 | +++ b/drivers/xen/swiotlb-xen.c |
725 | @@ -390,7 +390,7 @@ static void xen_unmap_single(struct device *hwdev, dma_addr_t dev_addr, |
726 | |
727 | /* NOTE: We use dev_addr here, not paddr! */ |
728 | if (is_xen_swiotlb_buffer(dev_addr)) { |
729 | - swiotlb_tbl_unmap_single(hwdev, paddr, size, dir); |
730 | + swiotlb_tbl_unmap_single(hwdev, dev_addr, size, dir); |
731 | return; |
732 | } |
733 | |
734 | diff --git a/fs/btrfs/delayed-inode.c b/fs/btrfs/delayed-inode.c |
735 | index f26f38ccd194..019fc5a68a14 100644 |
736 | --- a/fs/btrfs/delayed-inode.c |
737 | +++ b/fs/btrfs/delayed-inode.c |
738 | @@ -1843,6 +1843,14 @@ int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans, |
739 | struct btrfs_delayed_node *delayed_node; |
740 | int ret = 0; |
741 | |
742 | + /* |
743 | + * we don't do delayed inode updates during log recovery because it |
744 | + * leads to enospc problems. This means we also can't do |
745 | + * delayed inode refs |
746 | + */ |
747 | + if (BTRFS_I(inode)->root->fs_info->log_root_recovering) |
748 | + return -EAGAIN; |
749 | + |
750 | delayed_node = btrfs_get_or_create_delayed_node(inode); |
751 | if (IS_ERR(delayed_node)) |
752 | return PTR_ERR(delayed_node); |
753 | diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c |
754 | index 387213ac2608..b44306378193 100644 |
755 | --- a/fs/fs-writeback.c |
756 | +++ b/fs/fs-writeback.c |
757 | @@ -470,12 +470,28 @@ __writeback_single_inode(struct inode *inode, struct writeback_control *wbc) |
758 | * write_inode() |
759 | */ |
760 | spin_lock(&inode->i_lock); |
761 | - /* Clear I_DIRTY_PAGES if we've written out all dirty pages */ |
762 | - if (!mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) |
763 | - inode->i_state &= ~I_DIRTY_PAGES; |
764 | + |
765 | dirty = inode->i_state & I_DIRTY; |
766 | - inode->i_state &= ~(I_DIRTY_SYNC | I_DIRTY_DATASYNC); |
767 | + inode->i_state &= ~I_DIRTY; |
768 | + |
769 | + /* |
770 | + * Paired with smp_mb() in __mark_inode_dirty(). This allows |
771 | + * __mark_inode_dirty() to test i_state without grabbing i_lock - |
772 | + * either they see the I_DIRTY bits cleared or we see the dirtied |
773 | + * inode. |
774 | + * |
775 | + * I_DIRTY_PAGES is always cleared together above even if @mapping |
776 | + * still has dirty pages. The flag is reinstated after smp_mb() if |
777 | + * necessary. This guarantees that either __mark_inode_dirty() |
778 | + * sees clear I_DIRTY_PAGES or we see PAGECACHE_TAG_DIRTY. |
779 | + */ |
780 | + smp_mb(); |
781 | + |
782 | + if (mapping_tagged(mapping, PAGECACHE_TAG_DIRTY)) |
783 | + inode->i_state |= I_DIRTY_PAGES; |
784 | + |
785 | spin_unlock(&inode->i_lock); |
786 | + |
787 | /* Don't write the inode if only I_DIRTY_PAGES was set */ |
788 | if (dirty & (I_DIRTY_SYNC | I_DIRTY_DATASYNC)) { |
789 | int err = write_inode(inode, wbc); |
790 | @@ -1146,12 +1162,11 @@ void __mark_inode_dirty(struct inode *inode, int flags) |
791 | } |
792 | |
793 | /* |
794 | - * make sure that changes are seen by all cpus before we test i_state |
795 | - * -- mikulas |
796 | + * Paired with smp_mb() in __writeback_single_inode() for the |
797 | + * following lockless i_state test. See there for details. |
798 | */ |
799 | smp_mb(); |
800 | |
801 | - /* avoid the locking if we can */ |
802 | if ((inode->i_state & flags) == flags) |
803 | return; |
804 | |
805 | diff --git a/fs/nfsd/nfs4state.c b/fs/nfsd/nfs4state.c |
806 | index 836307ae1f08..4a58afa99654 100644 |
807 | --- a/fs/nfsd/nfs4state.c |
808 | +++ b/fs/nfsd/nfs4state.c |
809 | @@ -1200,15 +1200,14 @@ static int copy_cred(struct svc_cred *target, struct svc_cred *source) |
810 | return 0; |
811 | } |
812 | |
813 | -static long long |
814 | +static int |
815 | compare_blob(const struct xdr_netobj *o1, const struct xdr_netobj *o2) |
816 | { |
817 | - long long res; |
818 | - |
819 | - res = o1->len - o2->len; |
820 | - if (res) |
821 | - return res; |
822 | - return (long long)memcmp(o1->data, o2->data, o1->len); |
823 | + if (o1->len < o2->len) |
824 | + return -1; |
825 | + if (o1->len > o2->len) |
826 | + return 1; |
827 | + return memcmp(o1->data, o2->data, o1->len); |
828 | } |
829 | |
830 | static int same_name(const char *n1, const char *n2) |
831 | @@ -1365,7 +1364,7 @@ add_clp_to_name_tree(struct nfs4_client *new_clp, struct rb_root *root) |
832 | static struct nfs4_client * |
833 | find_clp_in_name_tree(struct xdr_netobj *name, struct rb_root *root) |
834 | { |
835 | - long long cmp; |
836 | + int cmp; |
837 | struct rb_node *node = root->rb_node; |
838 | struct nfs4_client *clp; |
839 | |
840 | diff --git a/fs/nfsd/nfs4xdr.c b/fs/nfsd/nfs4xdr.c |
841 | index 9b45f0666cfc..acf179d7615f 100644 |
842 | --- a/fs/nfsd/nfs4xdr.c |
843 | +++ b/fs/nfsd/nfs4xdr.c |
844 | @@ -1743,6 +1743,9 @@ static __be32 nfsd4_encode_components_esc(char sep, char *components, |
845 | } |
846 | else |
847 | end++; |
848 | + if (found_esc) |
849 | + end = next; |
850 | + |
851 | str = end; |
852 | } |
853 | *pp = p; |
854 | diff --git a/fs/nilfs2/inode.c b/fs/nilfs2/inode.c |
855 | index 2e1372efbb00..587d699bdc2c 100644 |
856 | --- a/fs/nilfs2/inode.c |
857 | +++ b/fs/nilfs2/inode.c |
858 | @@ -49,6 +49,8 @@ struct nilfs_iget_args { |
859 | int for_gc; |
860 | }; |
861 | |
862 | +static int nilfs_iget_test(struct inode *inode, void *opaque); |
863 | + |
864 | void nilfs_inode_add_blocks(struct inode *inode, int n) |
865 | { |
866 | struct nilfs_root *root = NILFS_I(inode)->i_root; |
867 | @@ -347,6 +349,17 @@ const struct address_space_operations nilfs_aops = { |
868 | .is_partially_uptodate = block_is_partially_uptodate, |
869 | }; |
870 | |
871 | +static int nilfs_insert_inode_locked(struct inode *inode, |
872 | + struct nilfs_root *root, |
873 | + unsigned long ino) |
874 | +{ |
875 | + struct nilfs_iget_args args = { |
876 | + .ino = ino, .root = root, .cno = 0, .for_gc = 0 |
877 | + }; |
878 | + |
879 | + return insert_inode_locked4(inode, ino, nilfs_iget_test, &args); |
880 | +} |
881 | + |
882 | struct inode *nilfs_new_inode(struct inode *dir, umode_t mode) |
883 | { |
884 | struct super_block *sb = dir->i_sb; |
885 | @@ -382,7 +395,7 @@ struct inode *nilfs_new_inode(struct inode *dir, umode_t mode) |
886 | if (S_ISREG(mode) || S_ISDIR(mode) || S_ISLNK(mode)) { |
887 | err = nilfs_bmap_read(ii->i_bmap, NULL); |
888 | if (err < 0) |
889 | - goto failed_bmap; |
890 | + goto failed_after_creation; |
891 | |
892 | set_bit(NILFS_I_BMAP, &ii->i_state); |
893 | /* No lock is needed; iget() ensures it. */ |
894 | @@ -398,21 +411,24 @@ struct inode *nilfs_new_inode(struct inode *dir, umode_t mode) |
895 | spin_lock(&nilfs->ns_next_gen_lock); |
896 | inode->i_generation = nilfs->ns_next_generation++; |
897 | spin_unlock(&nilfs->ns_next_gen_lock); |
898 | - insert_inode_hash(inode); |
899 | + if (nilfs_insert_inode_locked(inode, root, ino) < 0) { |
900 | + err = -EIO; |
901 | + goto failed_after_creation; |
902 | + } |
903 | |
904 | err = nilfs_init_acl(inode, dir); |
905 | if (unlikely(err)) |
906 | - goto failed_acl; /* never occur. When supporting |
907 | + goto failed_after_creation; /* never occur. When supporting |
908 | nilfs_init_acl(), proper cancellation of |
909 | above jobs should be considered */ |
910 | |
911 | return inode; |
912 | |
913 | - failed_acl: |
914 | - failed_bmap: |
915 | + failed_after_creation: |
916 | clear_nlink(inode); |
917 | + unlock_new_inode(inode); |
918 | iput(inode); /* raw_inode will be deleted through |
919 | - generic_delete_inode() */ |
920 | + nilfs_evict_inode() */ |
921 | goto failed; |
922 | |
923 | failed_ifile_create_inode: |
924 | @@ -460,8 +476,8 @@ int nilfs_read_inode_common(struct inode *inode, |
925 | inode->i_atime.tv_nsec = le32_to_cpu(raw_inode->i_mtime_nsec); |
926 | inode->i_ctime.tv_nsec = le32_to_cpu(raw_inode->i_ctime_nsec); |
927 | inode->i_mtime.tv_nsec = le32_to_cpu(raw_inode->i_mtime_nsec); |
928 | - if (inode->i_nlink == 0 && inode->i_mode == 0) |
929 | - return -EINVAL; /* this inode is deleted */ |
930 | + if (inode->i_nlink == 0) |
931 | + return -ESTALE; /* this inode is deleted */ |
932 | |
933 | inode->i_blocks = le64_to_cpu(raw_inode->i_blocks); |
934 | ii->i_flags = le32_to_cpu(raw_inode->i_flags); |
935 | diff --git a/fs/nilfs2/namei.c b/fs/nilfs2/namei.c |
936 | index 9de78f08989e..0f84b257932c 100644 |
937 | --- a/fs/nilfs2/namei.c |
938 | +++ b/fs/nilfs2/namei.c |
939 | @@ -51,9 +51,11 @@ static inline int nilfs_add_nondir(struct dentry *dentry, struct inode *inode) |
940 | int err = nilfs_add_link(dentry, inode); |
941 | if (!err) { |
942 | d_instantiate(dentry, inode); |
943 | + unlock_new_inode(inode); |
944 | return 0; |
945 | } |
946 | inode_dec_link_count(inode); |
947 | + unlock_new_inode(inode); |
948 | iput(inode); |
949 | return err; |
950 | } |
951 | @@ -182,6 +184,7 @@ out: |
952 | out_fail: |
953 | drop_nlink(inode); |
954 | nilfs_mark_inode_dirty(inode); |
955 | + unlock_new_inode(inode); |
956 | iput(inode); |
957 | goto out; |
958 | } |
959 | @@ -201,11 +204,15 @@ static int nilfs_link(struct dentry *old_dentry, struct inode *dir, |
960 | inode_inc_link_count(inode); |
961 | ihold(inode); |
962 | |
963 | - err = nilfs_add_nondir(dentry, inode); |
964 | - if (!err) |
965 | + err = nilfs_add_link(dentry, inode); |
966 | + if (!err) { |
967 | + d_instantiate(dentry, inode); |
968 | err = nilfs_transaction_commit(dir->i_sb); |
969 | - else |
970 | + } else { |
971 | + inode_dec_link_count(inode); |
972 | + iput(inode); |
973 | nilfs_transaction_abort(dir->i_sb); |
974 | + } |
975 | |
976 | return err; |
977 | } |
978 | @@ -243,6 +250,7 @@ static int nilfs_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode) |
979 | |
980 | nilfs_mark_inode_dirty(inode); |
981 | d_instantiate(dentry, inode); |
982 | + unlock_new_inode(inode); |
983 | out: |
984 | if (!err) |
985 | err = nilfs_transaction_commit(dir->i_sb); |
986 | @@ -255,6 +263,7 @@ out_fail: |
987 | drop_nlink(inode); |
988 | drop_nlink(inode); |
989 | nilfs_mark_inode_dirty(inode); |
990 | + unlock_new_inode(inode); |
991 | iput(inode); |
992 | out_dir: |
993 | drop_nlink(dir); |
994 | diff --git a/fs/ocfs2/aops.c b/fs/ocfs2/aops.c |
995 | index 20dfec72e903..f998c6009ad4 100644 |
996 | --- a/fs/ocfs2/aops.c |
997 | +++ b/fs/ocfs2/aops.c |
998 | @@ -917,7 +917,7 @@ void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages) |
999 | } |
1000 | } |
1001 | |
1002 | -static void ocfs2_free_write_ctxt(struct ocfs2_write_ctxt *wc) |
1003 | +static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) |
1004 | { |
1005 | int i; |
1006 | |
1007 | @@ -938,7 +938,11 @@ static void ocfs2_free_write_ctxt(struct ocfs2_write_ctxt *wc) |
1008 | page_cache_release(wc->w_target_page); |
1009 | } |
1010 | ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); |
1011 | +} |
1012 | |
1013 | +static void ocfs2_free_write_ctxt(struct ocfs2_write_ctxt *wc) |
1014 | +{ |
1015 | + ocfs2_unlock_pages(wc); |
1016 | brelse(wc->w_di_bh); |
1017 | kfree(wc); |
1018 | } |
1019 | @@ -2060,11 +2064,19 @@ out_write_size: |
1020 | di->i_mtime_nsec = di->i_ctime_nsec = cpu_to_le32(inode->i_mtime.tv_nsec); |
1021 | ocfs2_journal_dirty(handle, wc->w_di_bh); |
1022 | |
1023 | + /* unlock pages before dealloc since it needs acquiring j_trans_barrier |
1024 | + * lock, or it will cause a deadlock since journal commit threads holds |
1025 | + * this lock and will ask for the page lock when flushing the data. |
1026 | + * put it here to preserve the unlock order. |
1027 | + */ |
1028 | + ocfs2_unlock_pages(wc); |
1029 | + |
1030 | ocfs2_commit_trans(osb, handle); |
1031 | |
1032 | ocfs2_run_deallocs(osb, &wc->w_dealloc); |
1033 | |
1034 | - ocfs2_free_write_ctxt(wc); |
1035 | + brelse(wc->w_di_bh); |
1036 | + kfree(wc); |
1037 | |
1038 | return copied; |
1039 | } |
1040 | diff --git a/fs/pstore/ram.c b/fs/pstore/ram.c |
1041 | index 1376e5a8f0d6..42d5911c7e29 100644 |
1042 | --- a/fs/pstore/ram.c |
1043 | +++ b/fs/pstore/ram.c |
1044 | @@ -61,6 +61,11 @@ module_param(mem_size, ulong, 0400); |
1045 | MODULE_PARM_DESC(mem_size, |
1046 | "size of reserved RAM used to store oops/panic logs"); |
1047 | |
1048 | +static unsigned int mem_type; |
1049 | +module_param(mem_type, uint, 0600); |
1050 | +MODULE_PARM_DESC(mem_type, |
1051 | + "set to 1 to try to use unbuffered memory (default 0)"); |
1052 | + |
1053 | static int dump_oops = 1; |
1054 | module_param(dump_oops, int, 0600); |
1055 | MODULE_PARM_DESC(dump_oops, |
1056 | @@ -79,6 +84,7 @@ struct ramoops_context { |
1057 | struct persistent_ram_zone *fprz; |
1058 | phys_addr_t phys_addr; |
1059 | unsigned long size; |
1060 | + unsigned int memtype; |
1061 | size_t record_size; |
1062 | size_t console_size; |
1063 | size_t ftrace_size; |
1064 | @@ -331,7 +337,8 @@ static int ramoops_init_przs(struct device *dev, struct ramoops_context *cxt, |
1065 | size_t sz = cxt->record_size; |
1066 | |
1067 | cxt->przs[i] = persistent_ram_new(*paddr, sz, 0, |
1068 | - &cxt->ecc_info); |
1069 | + &cxt->ecc_info, |
1070 | + cxt->memtype); |
1071 | if (IS_ERR(cxt->przs[i])) { |
1072 | err = PTR_ERR(cxt->przs[i]); |
1073 | dev_err(dev, "failed to request mem region (0x%zx@0x%llx): %d\n", |
1074 | @@ -361,7 +368,7 @@ static int ramoops_init_prz(struct device *dev, struct ramoops_context *cxt, |
1075 | return -ENOMEM; |
1076 | } |
1077 | |
1078 | - *prz = persistent_ram_new(*paddr, sz, sig, &cxt->ecc_info); |
1079 | + *prz = persistent_ram_new(*paddr, sz, sig, &cxt->ecc_info, cxt->memtype); |
1080 | if (IS_ERR(*prz)) { |
1081 | int err = PTR_ERR(*prz); |
1082 | |
1083 | @@ -411,6 +418,7 @@ static int ramoops_probe(struct platform_device *pdev) |
1084 | cxt->dump_read_cnt = 0; |
1085 | cxt->size = pdata->mem_size; |
1086 | cxt->phys_addr = pdata->mem_address; |
1087 | + cxt->memtype = pdata->mem_type; |
1088 | cxt->record_size = pdata->record_size; |
1089 | cxt->console_size = pdata->console_size; |
1090 | cxt->ftrace_size = pdata->ftrace_size; |
1091 | @@ -541,6 +549,7 @@ static void ramoops_register_dummy(void) |
1092 | |
1093 | dummy_data->mem_size = mem_size; |
1094 | dummy_data->mem_address = mem_address; |
1095 | + dummy_data->mem_type = 0; |
1096 | dummy_data->record_size = record_size; |
1097 | dummy_data->console_size = ramoops_console_size; |
1098 | dummy_data->ftrace_size = ramoops_ftrace_size; |
1099 | diff --git a/fs/pstore/ram_core.c b/fs/pstore/ram_core.c |
1100 | index 59337326e288..6ff97553331b 100644 |
1101 | --- a/fs/pstore/ram_core.c |
1102 | +++ b/fs/pstore/ram_core.c |
1103 | @@ -333,7 +333,8 @@ void persistent_ram_zap(struct persistent_ram_zone *prz) |
1104 | persistent_ram_update_header_ecc(prz); |
1105 | } |
1106 | |
1107 | -static void *persistent_ram_vmap(phys_addr_t start, size_t size) |
1108 | +static void *persistent_ram_vmap(phys_addr_t start, size_t size, |
1109 | + unsigned int memtype) |
1110 | { |
1111 | struct page **pages; |
1112 | phys_addr_t page_start; |
1113 | @@ -345,7 +346,10 @@ static void *persistent_ram_vmap(phys_addr_t start, size_t size) |
1114 | page_start = start - offset_in_page(start); |
1115 | page_count = DIV_ROUND_UP(size + offset_in_page(start), PAGE_SIZE); |
1116 | |
1117 | - prot = pgprot_noncached(PAGE_KERNEL); |
1118 | + if (memtype) |
1119 | + prot = pgprot_noncached(PAGE_KERNEL); |
1120 | + else |
1121 | + prot = pgprot_writecombine(PAGE_KERNEL); |
1122 | |
1123 | pages = kmalloc(sizeof(struct page *) * page_count, GFP_KERNEL); |
1124 | if (!pages) { |
1125 | @@ -364,27 +368,35 @@ static void *persistent_ram_vmap(phys_addr_t start, size_t size) |
1126 | return vaddr; |
1127 | } |
1128 | |
1129 | -static void *persistent_ram_iomap(phys_addr_t start, size_t size) |
1130 | +static void *persistent_ram_iomap(phys_addr_t start, size_t size, |
1131 | + unsigned int memtype) |
1132 | { |
1133 | + void *va; |
1134 | + |
1135 | if (!request_mem_region(start, size, "persistent_ram")) { |
1136 | pr_err("request mem region (0x%llx@0x%llx) failed\n", |
1137 | (unsigned long long)size, (unsigned long long)start); |
1138 | return NULL; |
1139 | } |
1140 | |
1141 | - return ioremap(start, size); |
1142 | + if (memtype) |
1143 | + va = ioremap(start, size); |
1144 | + else |
1145 | + va = ioremap_wc(start, size); |
1146 | + |
1147 | + return va; |
1148 | } |
1149 | |
1150 | static int persistent_ram_buffer_map(phys_addr_t start, phys_addr_t size, |
1151 | - struct persistent_ram_zone *prz) |
1152 | + struct persistent_ram_zone *prz, int memtype) |
1153 | { |
1154 | prz->paddr = start; |
1155 | prz->size = size; |
1156 | |
1157 | if (pfn_valid(start >> PAGE_SHIFT)) |
1158 | - prz->vaddr = persistent_ram_vmap(start, size); |
1159 | + prz->vaddr = persistent_ram_vmap(start, size, memtype); |
1160 | else |
1161 | - prz->vaddr = persistent_ram_iomap(start, size); |
1162 | + prz->vaddr = persistent_ram_iomap(start, size, memtype); |
1163 | |
1164 | if (!prz->vaddr) { |
1165 | pr_err("%s: Failed to map 0x%llx pages at 0x%llx\n", __func__, |
1166 | @@ -452,7 +464,8 @@ void persistent_ram_free(struct persistent_ram_zone *prz) |
1167 | } |
1168 | |
1169 | struct persistent_ram_zone *persistent_ram_new(phys_addr_t start, size_t size, |
1170 | - u32 sig, struct persistent_ram_ecc_info *ecc_info) |
1171 | + u32 sig, struct persistent_ram_ecc_info *ecc_info, |
1172 | + unsigned int memtype) |
1173 | { |
1174 | struct persistent_ram_zone *prz; |
1175 | int ret = -ENOMEM; |
1176 | @@ -463,7 +476,7 @@ struct persistent_ram_zone *persistent_ram_new(phys_addr_t start, size_t size, |
1177 | goto err; |
1178 | } |
1179 | |
1180 | - ret = persistent_ram_buffer_map(start, size, prz); |
1181 | + ret = persistent_ram_buffer_map(start, size, prz, memtype); |
1182 | if (ret) |
1183 | goto err; |
1184 | |
1185 | diff --git a/include/linux/mm.h b/include/linux/mm.h |
1186 | index d4cdac903468..c4085192c2b6 100644 |
1187 | --- a/include/linux/mm.h |
1188 | +++ b/include/linux/mm.h |
1189 | @@ -1630,7 +1630,7 @@ extern int expand_downwards(struct vm_area_struct *vma, |
1190 | #if VM_GROWSUP |
1191 | extern int expand_upwards(struct vm_area_struct *vma, unsigned long address); |
1192 | #else |
1193 | - #define expand_upwards(vma, address) do { } while (0) |
1194 | + #define expand_upwards(vma, address) (0) |
1195 | #endif |
1196 | |
1197 | /* Look up the first VMA which satisfies addr < vm_end, NULL if none. */ |
1198 | diff --git a/include/linux/pstore_ram.h b/include/linux/pstore_ram.h |
1199 | index 9974975d40db..4af3fdc85b01 100644 |
1200 | --- a/include/linux/pstore_ram.h |
1201 | +++ b/include/linux/pstore_ram.h |
1202 | @@ -53,7 +53,8 @@ struct persistent_ram_zone { |
1203 | }; |
1204 | |
1205 | struct persistent_ram_zone *persistent_ram_new(phys_addr_t start, size_t size, |
1206 | - u32 sig, struct persistent_ram_ecc_info *ecc_info); |
1207 | + u32 sig, struct persistent_ram_ecc_info *ecc_info, |
1208 | + unsigned int memtype); |
1209 | void persistent_ram_free(struct persistent_ram_zone *prz); |
1210 | void persistent_ram_zap(struct persistent_ram_zone *prz); |
1211 | |
1212 | @@ -76,6 +77,7 @@ ssize_t persistent_ram_ecc_string(struct persistent_ram_zone *prz, |
1213 | struct ramoops_platform_data { |
1214 | unsigned long mem_size; |
1215 | unsigned long mem_address; |
1216 | + unsigned int mem_type; |
1217 | unsigned long record_size; |
1218 | unsigned long console_size; |
1219 | unsigned long ftrace_size; |
1220 | diff --git a/kernel/events/core.c b/kernel/events/core.c |
1221 | index 3f63ea6464ca..7bf4d519c20f 100644 |
1222 | --- a/kernel/events/core.c |
1223 | +++ b/kernel/events/core.c |
1224 | @@ -6887,11 +6887,11 @@ SYSCALL_DEFINE5(perf_event_open, |
1225 | |
1226 | if (move_group) { |
1227 | synchronize_rcu(); |
1228 | - perf_install_in_context(ctx, group_leader, event->cpu); |
1229 | + perf_install_in_context(ctx, group_leader, group_leader->cpu); |
1230 | get_ctx(ctx); |
1231 | list_for_each_entry(sibling, &group_leader->sibling_list, |
1232 | group_entry) { |
1233 | - perf_install_in_context(ctx, sibling, event->cpu); |
1234 | + perf_install_in_context(ctx, sibling, sibling->cpu); |
1235 | get_ctx(ctx); |
1236 | } |
1237 | } |
1238 | diff --git a/mm/memory.c b/mm/memory.c |
1239 | index 0926ccd04d7a..8b2d75f61b32 100644 |
1240 | --- a/mm/memory.c |
1241 | +++ b/mm/memory.c |
1242 | @@ -3200,7 +3200,7 @@ static inline int check_stack_guard_page(struct vm_area_struct *vma, unsigned lo |
1243 | if (prev && prev->vm_end == address) |
1244 | return prev->vm_flags & VM_GROWSDOWN ? 0 : -ENOMEM; |
1245 | |
1246 | - expand_downwards(vma, address - PAGE_SIZE); |
1247 | + return expand_downwards(vma, address - PAGE_SIZE); |
1248 | } |
1249 | if ((vma->vm_flags & VM_GROWSUP) && address + PAGE_SIZE == vma->vm_end) { |
1250 | struct vm_area_struct *next = vma->vm_next; |
1251 | @@ -3209,7 +3209,7 @@ static inline int check_stack_guard_page(struct vm_area_struct *vma, unsigned lo |
1252 | if (next && next->vm_start == address + PAGE_SIZE) |
1253 | return next->vm_flags & VM_GROWSUP ? 0 : -ENOMEM; |
1254 | |
1255 | - expand_upwards(vma, address + PAGE_SIZE); |
1256 | + return expand_upwards(vma, address + PAGE_SIZE); |
1257 | } |
1258 | return 0; |
1259 | } |
1260 | diff --git a/mm/mmap.c b/mm/mmap.c |
1261 | index 8f87b14c7968..43a7089c6a7c 100644 |
1262 | --- a/mm/mmap.c |
1263 | +++ b/mm/mmap.c |
1264 | @@ -2056,14 +2056,17 @@ static int acct_stack_growth(struct vm_area_struct *vma, unsigned long size, uns |
1265 | { |
1266 | struct mm_struct *mm = vma->vm_mm; |
1267 | struct rlimit *rlim = current->signal->rlim; |
1268 | - unsigned long new_start; |
1269 | + unsigned long new_start, actual_size; |
1270 | |
1271 | /* address space limit tests */ |
1272 | if (!may_expand_vm(mm, grow)) |
1273 | return -ENOMEM; |
1274 | |
1275 | /* Stack limit test */ |
1276 | - if (size > ACCESS_ONCE(rlim[RLIMIT_STACK].rlim_cur)) |
1277 | + actual_size = size; |
1278 | + if (size && (vma->vm_flags & (VM_GROWSUP | VM_GROWSDOWN))) |
1279 | + actual_size -= PAGE_SIZE; |
1280 | + if (actual_size > ACCESS_ONCE(rlim[RLIMIT_STACK].rlim_cur)) |
1281 | return -ENOMEM; |
1282 | |
1283 | /* mlock limit tests */ |
1284 | diff --git a/mm/vmscan.c b/mm/vmscan.c |
1285 | index 4e89500391dc..a2fd7e759cb7 100644 |
1286 | --- a/mm/vmscan.c |
1287 | +++ b/mm/vmscan.c |
1288 | @@ -2631,18 +2631,20 @@ static bool prepare_kswapd_sleep(pg_data_t *pgdat, int order, long remaining, |
1289 | return false; |
1290 | |
1291 | /* |
1292 | - * There is a potential race between when kswapd checks its watermarks |
1293 | - * and a process gets throttled. There is also a potential race if |
1294 | - * processes get throttled, kswapd wakes, a large process exits therby |
1295 | - * balancing the zones that causes kswapd to miss a wakeup. If kswapd |
1296 | - * is going to sleep, no process should be sleeping on pfmemalloc_wait |
1297 | - * so wake them now if necessary. If necessary, processes will wake |
1298 | - * kswapd and get throttled again |
1299 | + * The throttled processes are normally woken up in balance_pgdat() as |
1300 | + * soon as pfmemalloc_watermark_ok() is true. But there is a potential |
1301 | + * race between when kswapd checks the watermarks and a process gets |
1302 | + * throttled. There is also a potential race if processes get |
1303 | + * throttled, kswapd wakes, a large process exits thereby balancing the |
1304 | + * zones, which causes kswapd to exit balance_pgdat() before reaching |
1305 | + * the wake up checks. If kswapd is going to sleep, no process should |
1306 | + * be sleeping on pfmemalloc_wait, so wake them now if necessary. If |
1307 | + * the wake up is premature, processes will wake kswapd and get |
1308 | + * throttled again. The difference from wake ups in balance_pgdat() is |
1309 | + * that here we are under prepare_to_wait(). |
1310 | */ |
1311 | - if (waitqueue_active(&pgdat->pfmemalloc_wait)) { |
1312 | - wake_up(&pgdat->pfmemalloc_wait); |
1313 | - return false; |
1314 | - } |
1315 | + if (waitqueue_active(&pgdat->pfmemalloc_wait)) |
1316 | + wake_up_all(&pgdat->pfmemalloc_wait); |
1317 | |
1318 | return pgdat_balanced(pgdat, order, classzone_idx); |
1319 | } |
1320 | diff --git a/scripts/kernel-doc b/scripts/kernel-doc |
1321 | index 4305b2f2ec5e..8c0e07b7a70b 100755 |
1322 | --- a/scripts/kernel-doc |
1323 | +++ b/scripts/kernel-doc |
1324 | @@ -1750,7 +1750,7 @@ sub dump_struct($$) { |
1325 | # strip kmemcheck_bitfield_{begin,end}.*; |
1326 | $members =~ s/kmemcheck_bitfield_.*?;//gos; |
1327 | # strip attributes |
1328 | - $members =~ s/__aligned\s*\(.+\)//gos; |
1329 | + $members =~ s/__aligned\s*\([^;]*\)//gos; |
1330 | |
1331 | create_parameterlist($members, ';', $file); |
1332 | check_sections($file, $declaration_name, "struct", $sectcheck, $struct_actual, $nested); |
1333 | diff --git a/sound/pci/hda/hda_codec.c b/sound/pci/hda/hda_codec.c |
1334 | index aeefec74a061..83a0f9b4452b 100644 |
1335 | --- a/sound/pci/hda/hda_codec.c |
1336 | +++ b/sound/pci/hda/hda_codec.c |
1337 | @@ -327,8 +327,10 @@ int snd_hda_get_sub_nodes(struct hda_codec *codec, hda_nid_t nid, |
1338 | unsigned int parm; |
1339 | |
1340 | parm = snd_hda_param_read(codec, nid, AC_PAR_NODE_COUNT); |
1341 | - if (parm == -1) |
1342 | + if (parm == -1) { |
1343 | + *start_id = 0; |
1344 | return 0; |
1345 | + } |
1346 | *start_id = (parm >> 16) & 0x7fff; |
1347 | return (int)(parm & 0x7fff); |
1348 | } |
1349 | diff --git a/sound/pci/hda/patch_sigmatel.c b/sound/pci/hda/patch_sigmatel.c |
1350 | index 5dd4c4af9c9f..4ae5767a2cf5 100644 |
1351 | --- a/sound/pci/hda/patch_sigmatel.c |
1352 | +++ b/sound/pci/hda/patch_sigmatel.c |
1353 | @@ -573,9 +573,9 @@ static void stac_store_hints(struct hda_codec *codec) |
1354 | spec->gpio_mask; |
1355 | } |
1356 | if (get_int_hint(codec, "gpio_dir", &spec->gpio_dir)) |
1357 | - spec->gpio_mask &= spec->gpio_mask; |
1358 | - if (get_int_hint(codec, "gpio_data", &spec->gpio_data)) |
1359 | spec->gpio_dir &= spec->gpio_mask; |
1360 | + if (get_int_hint(codec, "gpio_data", &spec->gpio_data)) |
1361 | + spec->gpio_data &= spec->gpio_mask; |
1362 | if (get_int_hint(codec, "eapd_mask", &spec->eapd_mask)) |
1363 | spec->eapd_mask &= spec->gpio_mask; |
1364 | if (get_int_hint(codec, "gpio_mute", &spec->gpio_mute)) |
1365 | diff --git a/sound/soc/codecs/max98090.c b/sound/soc/codecs/max98090.c |
1366 | index 76bfeb3c3e30..be8de7ce1cda 100644 |
1367 | --- a/sound/soc/codecs/max98090.c |
1368 | +++ b/sound/soc/codecs/max98090.c |
1369 | @@ -1364,8 +1364,8 @@ static const struct snd_soc_dapm_route max98090_dapm_routes[] = { |
1370 | {"STENL Mux", "Sidetone Left", "DMICL"}, |
1371 | {"STENR Mux", "Sidetone Right", "ADCR"}, |
1372 | {"STENR Mux", "Sidetone Right", "DMICR"}, |
1373 | - {"DACL", "NULL", "STENL Mux"}, |
1374 | - {"DACR", "NULL", "STENL Mux"}, |
1375 | + {"DACL", NULL, "STENL Mux"}, |
1376 | + {"DACR", NULL, "STENL Mux"}, |
1377 | |
1378 | {"AIFINL", NULL, "SHDN"}, |
1379 | {"AIFINR", NULL, "SHDN"}, |
1380 | diff --git a/sound/soc/codecs/sigmadsp.c b/sound/soc/codecs/sigmadsp.c |
1381 | index 4068f2491232..bb3878c9625f 100644 |
1382 | --- a/sound/soc/codecs/sigmadsp.c |
1383 | +++ b/sound/soc/codecs/sigmadsp.c |
1384 | @@ -176,6 +176,13 @@ static int _process_sigma_firmware(struct device *dev, |
1385 | goto done; |
1386 | } |
1387 | |
1388 | + if (ssfw_head->version != 1) { |
1389 | + dev_err(dev, |
1390 | + "Failed to load firmware: Invalid version %d. Supported firmware versions: 1\n", |
1391 | + ssfw_head->version); |
1392 | + goto done; |
1393 | + } |
1394 | + |
1395 | crc = crc32(0, fw->data + sizeof(*ssfw_head), |
1396 | fw->size - sizeof(*ssfw_head)); |
1397 | pr_debug("%s: crc=%x\n", __func__, crc); |
1398 | diff --git a/sound/soc/dwc/designware_i2s.c b/sound/soc/dwc/designware_i2s.c |
1399 | index 593a3ea12d4c..489a9abf112b 100644 |
1400 | --- a/sound/soc/dwc/designware_i2s.c |
1401 | +++ b/sound/soc/dwc/designware_i2s.c |
1402 | @@ -263,6 +263,19 @@ static void dw_i2s_shutdown(struct snd_pcm_substream *substream, |
1403 | snd_soc_dai_set_dma_data(dai, substream, NULL); |
1404 | } |
1405 | |
1406 | +static int dw_i2s_prepare(struct snd_pcm_substream *substream, |
1407 | + struct snd_soc_dai *dai) |
1408 | +{ |
1409 | + struct dw_i2s_dev *dev = snd_soc_dai_get_drvdata(dai); |
1410 | + |
1411 | + if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) |
1412 | + i2s_write_reg(dev->i2s_base, TXFFR, 1); |
1413 | + else |
1414 | + i2s_write_reg(dev->i2s_base, RXFFR, 1); |
1415 | + |
1416 | + return 0; |
1417 | +} |
1418 | + |
1419 | static int dw_i2s_trigger(struct snd_pcm_substream *substream, |
1420 | int cmd, struct snd_soc_dai *dai) |
1421 | { |
1422 | @@ -294,6 +307,7 @@ static struct snd_soc_dai_ops dw_i2s_dai_ops = { |
1423 | .startup = dw_i2s_startup, |
1424 | .shutdown = dw_i2s_shutdown, |
1425 | .hw_params = dw_i2s_hw_params, |
1426 | + .prepare = dw_i2s_prepare, |
1427 | .trigger = dw_i2s_trigger, |
1428 | }; |
1429 | |
1430 | diff --git a/sound/usb/mixer_maps.c b/sound/usb/mixer_maps.c |
1431 | index 0339d464791a..4df31b0f94a3 100644 |
1432 | --- a/sound/usb/mixer_maps.c |
1433 | +++ b/sound/usb/mixer_maps.c |
1434 | @@ -322,8 +322,11 @@ static struct usbmix_name_map hercules_usb51_map[] = { |
1435 | { 0 } /* terminator */ |
1436 | }; |
1437 | |
1438 | -static const struct usbmix_name_map kef_x300a_map[] = { |
1439 | - { 10, NULL }, /* firmware locks up (?) when we try to access this FU */ |
1440 | +/* some (all?) SCMS USB3318 devices are affected by a firmware lock up |
1441 | + * when anything attempts to access FU 10 (control) |
1442 | + */ |
1443 | +static const struct usbmix_name_map scms_usb3318_map[] = { |
1444 | + { 10, NULL }, |
1445 | { 0 } |
1446 | }; |
1447 | |
1448 | @@ -415,8 +418,14 @@ static struct usbmix_ctl_map usbmix_ctl_maps[] = { |
1449 | .map = ebox44_map, |
1450 | }, |
1451 | { |
1452 | + /* KEF X300A */ |
1453 | .id = USB_ID(0x27ac, 0x1000), |
1454 | - .map = kef_x300a_map, |
1455 | + .map = scms_usb3318_map, |
1456 | + }, |
1457 | + { |
1458 | + /* Arcam rPAC */ |
1459 | + .id = USB_ID(0x25c4, 0x0003), |
1460 | + .map = scms_usb3318_map, |
1461 | }, |
1462 | { 0 } /* terminator */ |
1463 | }; |
1464 | diff --git a/tools/perf/util/hist.h b/tools/perf/util/hist.h |
1465 | index 14c2fe20aa62..20764e01df16 100644 |
1466 | --- a/tools/perf/util/hist.h |
1467 | +++ b/tools/perf/util/hist.h |
1468 | @@ -34,6 +34,7 @@ struct events_stats { |
1469 | u32 nr_invalid_chains; |
1470 | u32 nr_unknown_id; |
1471 | u32 nr_unprocessable_samples; |
1472 | + u32 nr_unordered_events; |
1473 | }; |
1474 | |
1475 | enum hist_column { |
1476 | diff --git a/tools/perf/util/session.c b/tools/perf/util/session.c |
1477 | index e392202b96bc..6f593a704ea5 100644 |
1478 | --- a/tools/perf/util/session.c |
1479 | +++ b/tools/perf/util/session.c |
1480 | @@ -656,8 +656,7 @@ static int perf_session_queue_event(struct perf_session *s, union perf_event *ev |
1481 | return -ETIME; |
1482 | |
1483 | if (timestamp < s->ordered_samples.last_flush) { |
1484 | - printf("Warning: Timestamp below last timeslice flush\n"); |
1485 | - return -EINVAL; |
1486 | + s->stats.nr_unordered_events++; |
1487 | } |
1488 | |
1489 | if (!list_empty(sc)) { |
1490 | @@ -1057,6 +1056,8 @@ static void perf_session__warn_about_errors(const struct perf_session *session, |
1491 | "Do you have a KVM guest running and not using 'perf kvm'?\n", |
1492 | session->stats.nr_unprocessable_samples); |
1493 | } |
1494 | + if (session->stats.nr_unordered_events != 0) |
1495 | + ui__warning("%u out of order events recorded.\n", session->stats.nr_unordered_events); |
1496 | } |
1497 | |
1498 | #define session_done() (*(volatile int *)(&session_done)) |