Contents of /trunk/kernel-alx/patches-3.10/0111-3.10.12-all-fixes.patch
Parent Directory | Revision Log
Revision 2288 -
(show annotations)
(download)
Wed Sep 25 06:47:47 2013 UTC (11 years ago) by niro
File size: 45530 byte(s)
Wed Sep 25 06:47:47 2013 UTC (11 years ago) by niro
File size: 45530 byte(s)
-linux-3.10 patches
1 | diff --git a/Makefile b/Makefile |
2 | index 595076d..afe001e 100644 |
3 | --- a/Makefile |
4 | +++ b/Makefile |
5 | @@ -1,6 +1,6 @@ |
6 | VERSION = 3 |
7 | PATCHLEVEL = 10 |
8 | -SUBLEVEL = 11 |
9 | +SUBLEVEL = 12 |
10 | EXTRAVERSION = |
11 | NAME = TOSSUG Baby Fish |
12 | |
13 | diff --git a/arch/arm/boot/dts/at91sam9260.dtsi b/arch/arm/boot/dts/at91sam9260.dtsi |
14 | index 84c4bef..43a18f7 100644 |
15 | --- a/arch/arm/boot/dts/at91sam9260.dtsi |
16 | +++ b/arch/arm/boot/dts/at91sam9260.dtsi |
17 | @@ -340,6 +340,14 @@ |
18 | }; |
19 | }; |
20 | |
21 | + i2c_gpio0 { |
22 | + pinctrl_i2c_gpio0: i2c_gpio0-0 { |
23 | + atmel,pins = |
24 | + <0 23 0x0 0x3 /* PA23 gpio I2C_SDA pin */ |
25 | + 0 24 0x0 0x3>; /* PA24 gpio I2C_SCL pin */ |
26 | + }; |
27 | + }; |
28 | + |
29 | pioA: gpio@fffff400 { |
30 | compatible = "atmel,at91rm9200-gpio"; |
31 | reg = <0xfffff400 0x200>; |
32 | @@ -592,6 +600,8 @@ |
33 | i2c-gpio,delay-us = <2>; /* ~100 kHz */ |
34 | #address-cells = <1>; |
35 | #size-cells = <0>; |
36 | + pinctrl-names = "default"; |
37 | + pinctrl-0 = <&pinctrl_i2c_gpio0>; |
38 | status = "disabled"; |
39 | }; |
40 | }; |
41 | diff --git a/arch/x86/include/asm/xor_avx.h b/arch/x86/include/asm/xor_avx.h |
42 | index 7ea79c5..492b298 100644 |
43 | --- a/arch/x86/include/asm/xor_avx.h |
44 | +++ b/arch/x86/include/asm/xor_avx.h |
45 | @@ -167,12 +167,12 @@ static struct xor_block_template xor_block_avx = { |
46 | |
47 | #define AVX_XOR_SPEED \ |
48 | do { \ |
49 | - if (cpu_has_avx) \ |
50 | + if (cpu_has_avx && cpu_has_osxsave) \ |
51 | xor_speed(&xor_block_avx); \ |
52 | } while (0) |
53 | |
54 | #define AVX_SELECT(FASTEST) \ |
55 | - (cpu_has_avx ? &xor_block_avx : FASTEST) |
56 | + (cpu_has_avx && cpu_has_osxsave ? &xor_block_avx : FASTEST) |
57 | |
58 | #else |
59 | |
60 | diff --git a/drivers/net/bonding/bond_main.c b/drivers/net/bonding/bond_main.c |
61 | index f975696..666cf3a 100644 |
62 | --- a/drivers/net/bonding/bond_main.c |
63 | +++ b/drivers/net/bonding/bond_main.c |
64 | @@ -3770,11 +3770,17 @@ static int bond_neigh_init(struct neighbour *n) |
65 | * The bonding ndo_neigh_setup is called at init time beofre any |
66 | * slave exists. So we must declare proxy setup function which will |
67 | * be used at run time to resolve the actual slave neigh param setup. |
68 | + * |
69 | + * It's also called by master devices (such as vlans) to setup their |
70 | + * underlying devices. In that case - do nothing, we're already set up from |
71 | + * our init. |
72 | */ |
73 | static int bond_neigh_setup(struct net_device *dev, |
74 | struct neigh_parms *parms) |
75 | { |
76 | - parms->neigh_setup = bond_neigh_init; |
77 | + /* modify only our neigh_parms */ |
78 | + if (parms->dev == dev) |
79 | + parms->neigh_setup = bond_neigh_init; |
80 | |
81 | return 0; |
82 | } |
83 | diff --git a/drivers/net/ethernet/broadcom/tg3.c b/drivers/net/ethernet/broadcom/tg3.c |
84 | index a13463e..0877a05 100644 |
85 | --- a/drivers/net/ethernet/broadcom/tg3.c |
86 | +++ b/drivers/net/ethernet/broadcom/tg3.c |
87 | @@ -3003,6 +3003,19 @@ static bool tg3_phy_power_bug(struct tg3 *tp) |
88 | return false; |
89 | } |
90 | |
91 | +static bool tg3_phy_led_bug(struct tg3 *tp) |
92 | +{ |
93 | + switch (tg3_asic_rev(tp)) { |
94 | + case ASIC_REV_5719: |
95 | + if ((tp->phy_flags & TG3_PHYFLG_MII_SERDES) && |
96 | + !tp->pci_fn) |
97 | + return true; |
98 | + return false; |
99 | + } |
100 | + |
101 | + return false; |
102 | +} |
103 | + |
104 | static void tg3_power_down_phy(struct tg3 *tp, bool do_low_power) |
105 | { |
106 | u32 val; |
107 | @@ -3050,8 +3063,9 @@ static void tg3_power_down_phy(struct tg3 *tp, bool do_low_power) |
108 | } |
109 | return; |
110 | } else if (do_low_power) { |
111 | - tg3_writephy(tp, MII_TG3_EXT_CTRL, |
112 | - MII_TG3_EXT_CTRL_FORCE_LED_OFF); |
113 | + if (!tg3_phy_led_bug(tp)) |
114 | + tg3_writephy(tp, MII_TG3_EXT_CTRL, |
115 | + MII_TG3_EXT_CTRL_FORCE_LED_OFF); |
116 | |
117 | val = MII_TG3_AUXCTL_PCTL_100TX_LPWR | |
118 | MII_TG3_AUXCTL_PCTL_SPR_ISOLATE | |
119 | diff --git a/drivers/net/ethernet/emulex/benet/be_main.c b/drivers/net/ethernet/emulex/benet/be_main.c |
120 | index 6e43426..7371626 100644 |
121 | --- a/drivers/net/ethernet/emulex/benet/be_main.c |
122 | +++ b/drivers/net/ethernet/emulex/benet/be_main.c |
123 | @@ -2561,8 +2561,8 @@ static int be_close(struct net_device *netdev) |
124 | /* Wait for all pending tx completions to arrive so that |
125 | * all tx skbs are freed. |
126 | */ |
127 | - be_tx_compl_clean(adapter); |
128 | netif_tx_disable(netdev); |
129 | + be_tx_compl_clean(adapter); |
130 | |
131 | be_rx_qs_destroy(adapter); |
132 | |
133 | diff --git a/drivers/net/ethernet/marvell/mvneta.c b/drivers/net/ethernet/marvell/mvneta.c |
134 | index c966785..254f255 100644 |
135 | --- a/drivers/net/ethernet/marvell/mvneta.c |
136 | +++ b/drivers/net/ethernet/marvell/mvneta.c |
137 | @@ -136,7 +136,9 @@ |
138 | #define MVNETA_GMAC_FORCE_LINK_PASS BIT(1) |
139 | #define MVNETA_GMAC_CONFIG_MII_SPEED BIT(5) |
140 | #define MVNETA_GMAC_CONFIG_GMII_SPEED BIT(6) |
141 | +#define MVNETA_GMAC_AN_SPEED_EN BIT(7) |
142 | #define MVNETA_GMAC_CONFIG_FULL_DUPLEX BIT(12) |
143 | +#define MVNETA_GMAC_AN_DUPLEX_EN BIT(13) |
144 | #define MVNETA_MIB_COUNTERS_BASE 0x3080 |
145 | #define MVNETA_MIB_LATE_COLLISION 0x7c |
146 | #define MVNETA_DA_FILT_SPEC_MCAST 0x3400 |
147 | @@ -911,6 +913,13 @@ static void mvneta_defaults_set(struct mvneta_port *pp) |
148 | /* Assign port SDMA configuration */ |
149 | mvreg_write(pp, MVNETA_SDMA_CONFIG, val); |
150 | |
151 | + /* Disable PHY polling in hardware, since we're using the |
152 | + * kernel phylib to do this. |
153 | + */ |
154 | + val = mvreg_read(pp, MVNETA_UNIT_CONTROL); |
155 | + val &= ~MVNETA_PHY_POLLING_ENABLE; |
156 | + mvreg_write(pp, MVNETA_UNIT_CONTROL, val); |
157 | + |
158 | mvneta_set_ucast_table(pp, -1); |
159 | mvneta_set_special_mcast_table(pp, -1); |
160 | mvneta_set_other_mcast_table(pp, -1); |
161 | @@ -2288,7 +2297,9 @@ static void mvneta_adjust_link(struct net_device *ndev) |
162 | val = mvreg_read(pp, MVNETA_GMAC_AUTONEG_CONFIG); |
163 | val &= ~(MVNETA_GMAC_CONFIG_MII_SPEED | |
164 | MVNETA_GMAC_CONFIG_GMII_SPEED | |
165 | - MVNETA_GMAC_CONFIG_FULL_DUPLEX); |
166 | + MVNETA_GMAC_CONFIG_FULL_DUPLEX | |
167 | + MVNETA_GMAC_AN_SPEED_EN | |
168 | + MVNETA_GMAC_AN_DUPLEX_EN); |
169 | |
170 | if (phydev->duplex) |
171 | val |= MVNETA_GMAC_CONFIG_FULL_DUPLEX; |
172 | diff --git a/drivers/net/ethernet/realtek/8139cp.c b/drivers/net/ethernet/realtek/8139cp.c |
173 | index 887aebe..9095ff9 100644 |
174 | --- a/drivers/net/ethernet/realtek/8139cp.c |
175 | +++ b/drivers/net/ethernet/realtek/8139cp.c |
176 | @@ -524,6 +524,7 @@ rx_status_loop: |
177 | PCI_DMA_FROMDEVICE); |
178 | if (dma_mapping_error(&cp->pdev->dev, new_mapping)) { |
179 | dev->stats.rx_dropped++; |
180 | + kfree_skb(new_skb); |
181 | goto rx_next; |
182 | } |
183 | |
184 | diff --git a/drivers/net/ethernet/sfc/filter.c b/drivers/net/ethernet/sfc/filter.c |
185 | index 2738b5f..a520465 100644 |
186 | --- a/drivers/net/ethernet/sfc/filter.c |
187 | +++ b/drivers/net/ethernet/sfc/filter.c |
188 | @@ -675,7 +675,7 @@ s32 efx_filter_insert_filter(struct efx_nic *efx, struct efx_filter_spec *spec, |
189 | BUILD_BUG_ON(EFX_FILTER_INDEX_UC_DEF != 0); |
190 | BUILD_BUG_ON(EFX_FILTER_INDEX_MC_DEF != |
191 | EFX_FILTER_MC_DEF - EFX_FILTER_UC_DEF); |
192 | - rep_index = spec->type - EFX_FILTER_INDEX_UC_DEF; |
193 | + rep_index = spec->type - EFX_FILTER_UC_DEF; |
194 | ins_index = rep_index; |
195 | |
196 | spin_lock_bh(&state->lock); |
197 | diff --git a/drivers/net/macvlan.c b/drivers/net/macvlan.c |
198 | index 6e91931..06eba6e 100644 |
199 | --- a/drivers/net/macvlan.c |
200 | +++ b/drivers/net/macvlan.c |
201 | @@ -727,6 +727,10 @@ static int macvlan_validate(struct nlattr *tb[], struct nlattr *data[]) |
202 | return -EADDRNOTAVAIL; |
203 | } |
204 | |
205 | + if (data && data[IFLA_MACVLAN_FLAGS] && |
206 | + nla_get_u16(data[IFLA_MACVLAN_FLAGS]) & ~MACVLAN_FLAG_NOPROMISC) |
207 | + return -EINVAL; |
208 | + |
209 | if (data && data[IFLA_MACVLAN_MODE]) { |
210 | switch (nla_get_u32(data[IFLA_MACVLAN_MODE])) { |
211 | case MACVLAN_MODE_PRIVATE: |
212 | diff --git a/drivers/net/tun.c b/drivers/net/tun.c |
213 | index 2491eb2..7b54f4f 100644 |
214 | --- a/drivers/net/tun.c |
215 | +++ b/drivers/net/tun.c |
216 | @@ -1076,8 +1076,9 @@ static ssize_t tun_get_user(struct tun_struct *tun, struct tun_file *tfile, |
217 | u32 rxhash; |
218 | |
219 | if (!(tun->flags & TUN_NO_PI)) { |
220 | - if ((len -= sizeof(pi)) > total_len) |
221 | + if (len < sizeof(pi)) |
222 | return -EINVAL; |
223 | + len -= sizeof(pi); |
224 | |
225 | if (memcpy_fromiovecend((void *)&pi, iv, 0, sizeof(pi))) |
226 | return -EFAULT; |
227 | @@ -1085,8 +1086,9 @@ static ssize_t tun_get_user(struct tun_struct *tun, struct tun_file *tfile, |
228 | } |
229 | |
230 | if (tun->flags & TUN_VNET_HDR) { |
231 | - if ((len -= tun->vnet_hdr_sz) > total_len) |
232 | + if (len < tun->vnet_hdr_sz) |
233 | return -EINVAL; |
234 | + len -= tun->vnet_hdr_sz; |
235 | |
236 | if (memcpy_fromiovecend((void *)&gso, iv, offset, sizeof(gso))) |
237 | return -EFAULT; |
238 | diff --git a/drivers/net/usb/cdc_mbim.c b/drivers/net/usb/cdc_mbim.c |
239 | index 8728198..25ba7ec 100644 |
240 | --- a/drivers/net/usb/cdc_mbim.c |
241 | +++ b/drivers/net/usb/cdc_mbim.c |
242 | @@ -400,6 +400,10 @@ static const struct usb_device_id mbim_devs[] = { |
243 | { USB_DEVICE_AND_INTERFACE_INFO(0x1199, 0x68a2, USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE), |
244 | .driver_info = (unsigned long)&cdc_mbim_info_zlp, |
245 | }, |
246 | + /* HP hs2434 Mobile Broadband Module needs ZLPs */ |
247 | + { USB_DEVICE_AND_INTERFACE_INFO(0x3f0, 0x4b1d, USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE), |
248 | + .driver_info = (unsigned long)&cdc_mbim_info_zlp, |
249 | + }, |
250 | { USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE), |
251 | .driver_info = (unsigned long)&cdc_mbim_info, |
252 | }, |
253 | diff --git a/drivers/net/vxlan.c b/drivers/net/vxlan.c |
254 | index 57325f3..054489f 100644 |
255 | --- a/drivers/net/vxlan.c |
256 | +++ b/drivers/net/vxlan.c |
257 | @@ -1090,7 +1090,7 @@ static netdev_tx_t vxlan_xmit_one(struct sk_buff *skb, struct net_device *dev, |
258 | iph->daddr = dst; |
259 | iph->saddr = fl4.saddr; |
260 | iph->ttl = ttl ? : ip4_dst_hoplimit(&rt->dst); |
261 | - tunnel_ip_select_ident(skb, old_iph, &rt->dst); |
262 | + __ip_select_ident(iph, &rt->dst, (skb_shinfo(skb)->gso_segs ?: 1) - 1); |
263 | |
264 | nf_reset(skb); |
265 | |
266 | diff --git a/drivers/net/wireless/mwifiex/main.c b/drivers/net/wireless/mwifiex/main.c |
267 | index 2eb88ea..c4a2e77 100644 |
268 | --- a/drivers/net/wireless/mwifiex/main.c |
269 | +++ b/drivers/net/wireless/mwifiex/main.c |
270 | @@ -363,20 +363,6 @@ static void mwifiex_fw_dpc(const struct firmware *firmware, void *context) |
271 | dev_err(adapter->dev, "cannot create default STA interface\n"); |
272 | goto err_add_intf; |
273 | } |
274 | - |
275 | - /* Create AP interface by default */ |
276 | - if (!mwifiex_add_virtual_intf(adapter->wiphy, "uap%d", |
277 | - NL80211_IFTYPE_AP, NULL, NULL)) { |
278 | - dev_err(adapter->dev, "cannot create default AP interface\n"); |
279 | - goto err_add_intf; |
280 | - } |
281 | - |
282 | - /* Create P2P interface by default */ |
283 | - if (!mwifiex_add_virtual_intf(adapter->wiphy, "p2p%d", |
284 | - NL80211_IFTYPE_P2P_CLIENT, NULL, NULL)) { |
285 | - dev_err(adapter->dev, "cannot create default P2P interface\n"); |
286 | - goto err_add_intf; |
287 | - } |
288 | rtnl_unlock(); |
289 | |
290 | mwifiex_drv_get_driver_version(adapter, fmt, sizeof(fmt) - 1); |
291 | diff --git a/drivers/rtc/rtc-max77686.c b/drivers/rtc/rtc-max77686.c |
292 | index 771812d..3bb9401 100644 |
293 | --- a/drivers/rtc/rtc-max77686.c |
294 | +++ b/drivers/rtc/rtc-max77686.c |
295 | @@ -240,9 +240,9 @@ static int max77686_rtc_read_alarm(struct device *dev, struct rtc_wkalrm *alrm) |
296 | } |
297 | |
298 | alrm->pending = 0; |
299 | - ret = regmap_read(info->max77686->regmap, MAX77686_REG_STATUS1, &val); |
300 | + ret = regmap_read(info->max77686->regmap, MAX77686_REG_STATUS2, &val); |
301 | if (ret < 0) { |
302 | - dev_err(info->dev, "%s:%d fail to read status1 reg(%d)\n", |
303 | + dev_err(info->dev, "%s:%d fail to read status2 reg(%d)\n", |
304 | __func__, __LINE__, ret); |
305 | goto out; |
306 | } |
307 | diff --git a/drivers/vhost/net.c b/drivers/vhost/net.c |
308 | index 8ca5ac7..d6a518c 100644 |
309 | --- a/drivers/vhost/net.c |
310 | +++ b/drivers/vhost/net.c |
311 | @@ -307,6 +307,11 @@ static void vhost_zerocopy_callback(struct ubuf_info *ubuf, bool success) |
312 | struct vhost_virtqueue *vq = ubufs->vq; |
313 | int cnt = atomic_read(&ubufs->kref.refcount); |
314 | |
315 | + /* set len to mark this desc buffers done DMA */ |
316 | + vq->heads[ubuf->desc].len = success ? |
317 | + VHOST_DMA_DONE_LEN : VHOST_DMA_FAILED_LEN; |
318 | + vhost_net_ubuf_put(ubufs); |
319 | + |
320 | /* |
321 | * Trigger polling thread if guest stopped submitting new buffers: |
322 | * in this case, the refcount after decrement will eventually reach 1 |
323 | @@ -317,10 +322,6 @@ static void vhost_zerocopy_callback(struct ubuf_info *ubuf, bool success) |
324 | */ |
325 | if (cnt <= 2 || !(cnt % 16)) |
326 | vhost_poll_queue(&vq->poll); |
327 | - /* set len to mark this desc buffers done DMA */ |
328 | - vq->heads[ubuf->desc].len = success ? |
329 | - VHOST_DMA_DONE_LEN : VHOST_DMA_FAILED_LEN; |
330 | - vhost_net_ubuf_put(ubufs); |
331 | } |
332 | |
333 | /* Expects to be always run from workqueue - which acts as |
334 | diff --git a/include/linux/ipv6.h b/include/linux/ipv6.h |
335 | index 850e95b..b8b7dc7 100644 |
336 | --- a/include/linux/ipv6.h |
337 | +++ b/include/linux/ipv6.h |
338 | @@ -101,6 +101,7 @@ struct inet6_skb_parm { |
339 | #define IP6SKB_FORWARDED 2 |
340 | #define IP6SKB_REROUTED 4 |
341 | #define IP6SKB_ROUTERALERT 8 |
342 | +#define IP6SKB_FRAGMENTED 16 |
343 | }; |
344 | |
345 | #define IP6CB(skb) ((struct inet6_skb_parm*)((skb)->cb)) |
346 | diff --git a/include/net/genetlink.h b/include/net/genetlink.h |
347 | index 93024a4..8e0b6c8 100644 |
348 | --- a/include/net/genetlink.h |
349 | +++ b/include/net/genetlink.h |
350 | @@ -61,6 +61,7 @@ struct genl_family { |
351 | struct list_head ops_list; /* private */ |
352 | struct list_head family_list; /* private */ |
353 | struct list_head mcast_groups; /* private */ |
354 | + struct module *module; |
355 | }; |
356 | |
357 | /** |
358 | @@ -121,9 +122,24 @@ struct genl_ops { |
359 | struct list_head ops_list; |
360 | }; |
361 | |
362 | -extern int genl_register_family(struct genl_family *family); |
363 | -extern int genl_register_family_with_ops(struct genl_family *family, |
364 | +extern int __genl_register_family(struct genl_family *family); |
365 | + |
366 | +static inline int genl_register_family(struct genl_family *family) |
367 | +{ |
368 | + family->module = THIS_MODULE; |
369 | + return __genl_register_family(family); |
370 | +} |
371 | + |
372 | +extern int __genl_register_family_with_ops(struct genl_family *family, |
373 | struct genl_ops *ops, size_t n_ops); |
374 | + |
375 | +static inline int genl_register_family_with_ops(struct genl_family *family, |
376 | + struct genl_ops *ops, size_t n_ops) |
377 | +{ |
378 | + family->module = THIS_MODULE; |
379 | + return __genl_register_family_with_ops(family, ops, n_ops); |
380 | +} |
381 | + |
382 | extern int genl_unregister_family(struct genl_family *family); |
383 | extern int genl_register_ops(struct genl_family *, struct genl_ops *ops); |
384 | extern int genl_unregister_ops(struct genl_family *, struct genl_ops *ops); |
385 | diff --git a/include/net/ip_tunnels.h b/include/net/ip_tunnels.h |
386 | index 09b1360..a9942e1 100644 |
387 | --- a/include/net/ip_tunnels.h |
388 | +++ b/include/net/ip_tunnels.h |
389 | @@ -141,20 +141,6 @@ static inline u8 ip_tunnel_ecn_encap(u8 tos, const struct iphdr *iph, |
390 | return INET_ECN_encapsulate(tos, inner); |
391 | } |
392 | |
393 | -static inline void tunnel_ip_select_ident(struct sk_buff *skb, |
394 | - const struct iphdr *old_iph, |
395 | - struct dst_entry *dst) |
396 | -{ |
397 | - struct iphdr *iph = ip_hdr(skb); |
398 | - |
399 | - /* Use inner packet iph-id if possible. */ |
400 | - if (skb->protocol == htons(ETH_P_IP) && old_iph->id) |
401 | - iph->id = old_iph->id; |
402 | - else |
403 | - __ip_select_ident(iph, dst, |
404 | - (skb_shinfo(skb)->gso_segs ?: 1) - 1); |
405 | -} |
406 | - |
407 | static inline void iptunnel_xmit(struct sk_buff *skb, struct net_device *dev) |
408 | { |
409 | int err; |
410 | diff --git a/include/net/sch_generic.h b/include/net/sch_generic.h |
411 | index e7f4e21..63ed1d1 100644 |
412 | --- a/include/net/sch_generic.h |
413 | +++ b/include/net/sch_generic.h |
414 | @@ -682,13 +682,19 @@ struct psched_ratecfg { |
415 | u64 rate_bps; |
416 | u32 mult; |
417 | u16 overhead; |
418 | + u8 linklayer; |
419 | u8 shift; |
420 | }; |
421 | |
422 | static inline u64 psched_l2t_ns(const struct psched_ratecfg *r, |
423 | unsigned int len) |
424 | { |
425 | - return ((u64)(len + r->overhead) * r->mult) >> r->shift; |
426 | + len += r->overhead; |
427 | + |
428 | + if (unlikely(r->linklayer == TC_LINKLAYER_ATM)) |
429 | + return ((u64)(DIV_ROUND_UP(len,48)*53) * r->mult) >> r->shift; |
430 | + |
431 | + return ((u64)len * r->mult) >> r->shift; |
432 | } |
433 | |
434 | extern void psched_ratecfg_precompute(struct psched_ratecfg *r, const struct tc_ratespec *conf); |
435 | @@ -699,6 +705,7 @@ static inline void psched_ratecfg_getrate(struct tc_ratespec *res, |
436 | memset(res, 0, sizeof(*res)); |
437 | res->rate = r->rate_bps >> 3; |
438 | res->overhead = r->overhead; |
439 | + res->linklayer = (r->linklayer & TC_LINKLAYER_MASK); |
440 | } |
441 | |
442 | #endif |
443 | diff --git a/include/uapi/linux/icmpv6.h b/include/uapi/linux/icmpv6.h |
444 | index e0133c7..590beda 100644 |
445 | --- a/include/uapi/linux/icmpv6.h |
446 | +++ b/include/uapi/linux/icmpv6.h |
447 | @@ -115,6 +115,8 @@ struct icmp6hdr { |
448 | #define ICMPV6_NOT_NEIGHBOUR 2 |
449 | #define ICMPV6_ADDR_UNREACH 3 |
450 | #define ICMPV6_PORT_UNREACH 4 |
451 | +#define ICMPV6_POLICY_FAIL 5 |
452 | +#define ICMPV6_REJECT_ROUTE 6 |
453 | |
454 | /* |
455 | * Codes for Time Exceeded |
456 | diff --git a/include/uapi/linux/pkt_sched.h b/include/uapi/linux/pkt_sched.h |
457 | index dbd71b0..09d62b92 100644 |
458 | --- a/include/uapi/linux/pkt_sched.h |
459 | +++ b/include/uapi/linux/pkt_sched.h |
460 | @@ -73,9 +73,17 @@ struct tc_estimator { |
461 | #define TC_H_ROOT (0xFFFFFFFFU) |
462 | #define TC_H_INGRESS (0xFFFFFFF1U) |
463 | |
464 | +/* Need to corrospond to iproute2 tc/tc_core.h "enum link_layer" */ |
465 | +enum tc_link_layer { |
466 | + TC_LINKLAYER_UNAWARE, /* Indicate unaware old iproute2 util */ |
467 | + TC_LINKLAYER_ETHERNET, |
468 | + TC_LINKLAYER_ATM, |
469 | +}; |
470 | +#define TC_LINKLAYER_MASK 0x0F /* limit use to lower 4 bits */ |
471 | + |
472 | struct tc_ratespec { |
473 | unsigned char cell_log; |
474 | - unsigned char __reserved; |
475 | + __u8 linklayer; /* lower 4 bits */ |
476 | unsigned short overhead; |
477 | short cell_align; |
478 | unsigned short mpu; |
479 | diff --git a/net/bridge/br_fdb.c b/net/bridge/br_fdb.c |
480 | index ebfa444..84dd783 100644 |
481 | --- a/net/bridge/br_fdb.c |
482 | +++ b/net/bridge/br_fdb.c |
483 | @@ -161,7 +161,7 @@ void br_fdb_change_mac_address(struct net_bridge *br, const u8 *newaddr) |
484 | if (!pv) |
485 | return; |
486 | |
487 | - for_each_set_bit_from(vid, pv->vlan_bitmap, BR_VLAN_BITMAP_LEN) { |
488 | + for_each_set_bit_from(vid, pv->vlan_bitmap, VLAN_N_VID) { |
489 | f = __br_fdb_get(br, br->dev->dev_addr, vid); |
490 | if (f && f->is_local && !f->dst) |
491 | fdb_delete(br, f); |
492 | @@ -725,7 +725,7 @@ int br_fdb_add(struct ndmsg *ndm, struct nlattr *tb[], |
493 | /* VID was specified, so use it. */ |
494 | err = __br_fdb_add(ndm, p, addr, nlh_flags, vid); |
495 | } else { |
496 | - if (!pv || bitmap_empty(pv->vlan_bitmap, BR_VLAN_BITMAP_LEN)) { |
497 | + if (!pv || bitmap_empty(pv->vlan_bitmap, VLAN_N_VID)) { |
498 | err = __br_fdb_add(ndm, p, addr, nlh_flags, 0); |
499 | goto out; |
500 | } |
501 | @@ -734,7 +734,7 @@ int br_fdb_add(struct ndmsg *ndm, struct nlattr *tb[], |
502 | * specify a VLAN. To be nice, add/update entry for every |
503 | * vlan on this port. |
504 | */ |
505 | - for_each_set_bit(vid, pv->vlan_bitmap, BR_VLAN_BITMAP_LEN) { |
506 | + for_each_set_bit(vid, pv->vlan_bitmap, VLAN_N_VID) { |
507 | err = __br_fdb_add(ndm, p, addr, nlh_flags, vid); |
508 | if (err) |
509 | goto out; |
510 | @@ -812,7 +812,7 @@ int br_fdb_delete(struct ndmsg *ndm, struct nlattr *tb[], |
511 | |
512 | err = __br_fdb_delete(p, addr, vid); |
513 | } else { |
514 | - if (!pv || bitmap_empty(pv->vlan_bitmap, BR_VLAN_BITMAP_LEN)) { |
515 | + if (!pv || bitmap_empty(pv->vlan_bitmap, VLAN_N_VID)) { |
516 | err = __br_fdb_delete(p, addr, 0); |
517 | goto out; |
518 | } |
519 | @@ -822,7 +822,7 @@ int br_fdb_delete(struct ndmsg *ndm, struct nlattr *tb[], |
520 | * vlan on this port. |
521 | */ |
522 | err = -ENOENT; |
523 | - for_each_set_bit(vid, pv->vlan_bitmap, BR_VLAN_BITMAP_LEN) { |
524 | + for_each_set_bit(vid, pv->vlan_bitmap, VLAN_N_VID) { |
525 | err &= __br_fdb_delete(p, addr, vid); |
526 | } |
527 | } |
528 | diff --git a/net/bridge/br_multicast.c b/net/bridge/br_multicast.c |
529 | index d6448e3..d82058f 100644 |
530 | --- a/net/bridge/br_multicast.c |
531 | +++ b/net/bridge/br_multicast.c |
532 | @@ -1185,7 +1185,7 @@ static int br_ip6_multicast_query(struct net_bridge *br, |
533 | max_delay = msecs_to_jiffies(ntohs(mld->mld_maxdelay)); |
534 | if (max_delay) |
535 | group = &mld->mld_mca; |
536 | - } else if (skb->len >= sizeof(*mld2q)) { |
537 | + } else { |
538 | if (!pskb_may_pull(skb, sizeof(*mld2q))) { |
539 | err = -EINVAL; |
540 | goto out; |
541 | @@ -1193,7 +1193,8 @@ static int br_ip6_multicast_query(struct net_bridge *br, |
542 | mld2q = (struct mld2_query *)icmp6_hdr(skb); |
543 | if (!mld2q->mld2q_nsrcs) |
544 | group = &mld2q->mld2q_mca; |
545 | - max_delay = mld2q->mld2q_mrc ? MLDV2_MRC(ntohs(mld2q->mld2q_mrc)) : 1; |
546 | + |
547 | + max_delay = max(msecs_to_jiffies(MLDV2_MRC(ntohs(mld2q->mld2q_mrc))), 1UL); |
548 | } |
549 | |
550 | if (!group) |
551 | diff --git a/net/bridge/br_netlink.c b/net/bridge/br_netlink.c |
552 | index 8e3abf5..f66a034 100644 |
553 | --- a/net/bridge/br_netlink.c |
554 | +++ b/net/bridge/br_netlink.c |
555 | @@ -128,7 +128,7 @@ static int br_fill_ifinfo(struct sk_buff *skb, |
556 | else |
557 | pv = br_get_vlan_info(br); |
558 | |
559 | - if (!pv || bitmap_empty(pv->vlan_bitmap, BR_VLAN_BITMAP_LEN)) |
560 | + if (!pv || bitmap_empty(pv->vlan_bitmap, VLAN_N_VID)) |
561 | goto done; |
562 | |
563 | af = nla_nest_start(skb, IFLA_AF_SPEC); |
564 | @@ -136,7 +136,7 @@ static int br_fill_ifinfo(struct sk_buff *skb, |
565 | goto nla_put_failure; |
566 | |
567 | pvid = br_get_pvid(pv); |
568 | - for_each_set_bit(vid, pv->vlan_bitmap, BR_VLAN_BITMAP_LEN) { |
569 | + for_each_set_bit(vid, pv->vlan_bitmap, VLAN_N_VID) { |
570 | vinfo.vid = vid; |
571 | vinfo.flags = 0; |
572 | if (vid == pvid) |
573 | diff --git a/net/bridge/br_vlan.c b/net/bridge/br_vlan.c |
574 | index bd58b45..9a9ffe7 100644 |
575 | --- a/net/bridge/br_vlan.c |
576 | +++ b/net/bridge/br_vlan.c |
577 | @@ -108,7 +108,7 @@ static int __vlan_del(struct net_port_vlans *v, u16 vid) |
578 | |
579 | clear_bit(vid, v->vlan_bitmap); |
580 | v->num_vlans--; |
581 | - if (bitmap_empty(v->vlan_bitmap, BR_VLAN_BITMAP_LEN)) { |
582 | + if (bitmap_empty(v->vlan_bitmap, VLAN_N_VID)) { |
583 | if (v->port_idx) |
584 | rcu_assign_pointer(v->parent.port->vlan_info, NULL); |
585 | else |
586 | @@ -122,7 +122,7 @@ static void __vlan_flush(struct net_port_vlans *v) |
587 | { |
588 | smp_wmb(); |
589 | v->pvid = 0; |
590 | - bitmap_zero(v->vlan_bitmap, BR_VLAN_BITMAP_LEN); |
591 | + bitmap_zero(v->vlan_bitmap, VLAN_N_VID); |
592 | if (v->port_idx) |
593 | rcu_assign_pointer(v->parent.port->vlan_info, NULL); |
594 | else |
595 | diff --git a/net/core/flow_dissector.c b/net/core/flow_dissector.c |
596 | index 00ee068..c99cc37 100644 |
597 | --- a/net/core/flow_dissector.c |
598 | +++ b/net/core/flow_dissector.c |
599 | @@ -345,14 +345,9 @@ u16 __netdev_pick_tx(struct net_device *dev, struct sk_buff *skb) |
600 | if (new_index < 0) |
601 | new_index = skb_tx_hash(dev, skb); |
602 | |
603 | - if (queue_index != new_index && sk) { |
604 | - struct dst_entry *dst = |
605 | - rcu_dereference_check(sk->sk_dst_cache, 1); |
606 | - |
607 | - if (dst && skb_dst(skb) == dst) |
608 | - sk_tx_queue_set(sk, queue_index); |
609 | - |
610 | - } |
611 | + if (queue_index != new_index && sk && |
612 | + rcu_access_pointer(sk->sk_dst_cache)) |
613 | + sk_tx_queue_set(sk, queue_index); |
614 | |
615 | queue_index = new_index; |
616 | } |
617 | diff --git a/net/core/neighbour.c b/net/core/neighbour.c |
618 | index ce90b02..0034b61 100644 |
619 | --- a/net/core/neighbour.c |
620 | +++ b/net/core/neighbour.c |
621 | @@ -1445,16 +1445,18 @@ struct neigh_parms *neigh_parms_alloc(struct net_device *dev, |
622 | atomic_set(&p->refcnt, 1); |
623 | p->reachable_time = |
624 | neigh_rand_reach_time(p->base_reachable_time); |
625 | + dev_hold(dev); |
626 | + p->dev = dev; |
627 | + write_pnet(&p->net, hold_net(net)); |
628 | + p->sysctl_table = NULL; |
629 | |
630 | if (ops->ndo_neigh_setup && ops->ndo_neigh_setup(dev, p)) { |
631 | + release_net(net); |
632 | + dev_put(dev); |
633 | kfree(p); |
634 | return NULL; |
635 | } |
636 | |
637 | - dev_hold(dev); |
638 | - p->dev = dev; |
639 | - write_pnet(&p->net, hold_net(net)); |
640 | - p->sysctl_table = NULL; |
641 | write_lock_bh(&tbl->lock); |
642 | p->next = tbl->parms.next; |
643 | tbl->parms.next = p; |
644 | diff --git a/net/core/rtnetlink.c b/net/core/rtnetlink.c |
645 | index a08bd2b..fd01eca 100644 |
646 | --- a/net/core/rtnetlink.c |
647 | +++ b/net/core/rtnetlink.c |
648 | @@ -2142,7 +2142,7 @@ int ndo_dflt_fdb_del(struct ndmsg *ndm, |
649 | /* If aging addresses are supported device will need to |
650 | * implement its own handler for this. |
651 | */ |
652 | - if (ndm->ndm_state & NUD_PERMANENT) { |
653 | + if (!(ndm->ndm_state & NUD_PERMANENT)) { |
654 | pr_info("%s: FDB only supports static addresses\n", dev->name); |
655 | return -EINVAL; |
656 | } |
657 | @@ -2374,7 +2374,7 @@ static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb) |
658 | struct nlattr *extfilt; |
659 | u32 filter_mask = 0; |
660 | |
661 | - extfilt = nlmsg_find_attr(cb->nlh, sizeof(struct rtgenmsg), |
662 | + extfilt = nlmsg_find_attr(cb->nlh, sizeof(struct ifinfomsg), |
663 | IFLA_EXT_MASK); |
664 | if (extfilt) |
665 | filter_mask = nla_get_u32(extfilt); |
666 | diff --git a/net/core/sysctl_net_core.c b/net/core/sysctl_net_core.c |
667 | index cfdb46a..2ff093b 100644 |
668 | --- a/net/core/sysctl_net_core.c |
669 | +++ b/net/core/sysctl_net_core.c |
670 | @@ -20,7 +20,9 @@ |
671 | #include <net/sock.h> |
672 | #include <net/net_ratelimit.h> |
673 | |
674 | +static int zero = 0; |
675 | static int one = 1; |
676 | +static int ushort_max = USHRT_MAX; |
677 | |
678 | #ifdef CONFIG_RPS |
679 | static int rps_sock_flow_sysctl(ctl_table *table, int write, |
680 | @@ -204,7 +206,9 @@ static struct ctl_table netns_core_table[] = { |
681 | .data = &init_net.core.sysctl_somaxconn, |
682 | .maxlen = sizeof(int), |
683 | .mode = 0644, |
684 | - .proc_handler = proc_dointvec |
685 | + .extra1 = &zero, |
686 | + .extra2 = &ushort_max, |
687 | + .proc_handler = proc_dointvec_minmax |
688 | }, |
689 | { } |
690 | }; |
691 | diff --git a/net/ipv4/devinet.c b/net/ipv4/devinet.c |
692 | index dfc39d4..9e38217 100644 |
693 | --- a/net/ipv4/devinet.c |
694 | +++ b/net/ipv4/devinet.c |
695 | @@ -771,7 +771,7 @@ static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh, |
696 | ci = nla_data(tb[IFA_CACHEINFO]); |
697 | if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) { |
698 | err = -EINVAL; |
699 | - goto errout; |
700 | + goto errout_free; |
701 | } |
702 | *pvalid_lft = ci->ifa_valid; |
703 | *pprefered_lft = ci->ifa_prefered; |
704 | @@ -779,6 +779,8 @@ static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh, |
705 | |
706 | return ifa; |
707 | |
708 | +errout_free: |
709 | + inet_free_ifa(ifa); |
710 | errout: |
711 | return ERR_PTR(err); |
712 | } |
713 | diff --git a/net/ipv4/fib_trie.c b/net/ipv4/fib_trie.c |
714 | index 49616fe..6e8a13d 100644 |
715 | --- a/net/ipv4/fib_trie.c |
716 | +++ b/net/ipv4/fib_trie.c |
717 | @@ -71,7 +71,6 @@ |
718 | #include <linux/init.h> |
719 | #include <linux/list.h> |
720 | #include <linux/slab.h> |
721 | -#include <linux/prefetch.h> |
722 | #include <linux/export.h> |
723 | #include <net/net_namespace.h> |
724 | #include <net/ip.h> |
725 | @@ -1761,10 +1760,8 @@ static struct leaf *leaf_walk_rcu(struct tnode *p, struct rt_trie_node *c) |
726 | if (!c) |
727 | continue; |
728 | |
729 | - if (IS_LEAF(c)) { |
730 | - prefetch(rcu_dereference_rtnl(p->child[idx])); |
731 | + if (IS_LEAF(c)) |
732 | return (struct leaf *) c; |
733 | - } |
734 | |
735 | /* Rescan start scanning in new node */ |
736 | p = (struct tnode *) c; |
737 | diff --git a/net/ipv4/ip_gre.c b/net/ipv4/ip_gre.c |
738 | index 855004f..c52fee0 100644 |
739 | --- a/net/ipv4/ip_gre.c |
740 | +++ b/net/ipv4/ip_gre.c |
741 | @@ -572,7 +572,7 @@ static int ipgre_header(struct sk_buff *skb, struct net_device *dev, |
742 | if (daddr) |
743 | memcpy(&iph->daddr, daddr, 4); |
744 | if (iph->daddr) |
745 | - return t->hlen; |
746 | + return t->hlen + sizeof(*iph); |
747 | |
748 | return -(t->hlen + sizeof(*iph)); |
749 | } |
750 | diff --git a/net/ipv4/ip_tunnel.c b/net/ipv4/ip_tunnel.c |
751 | index cbfc37f..b7a4c21 100644 |
752 | --- a/net/ipv4/ip_tunnel.c |
753 | +++ b/net/ipv4/ip_tunnel.c |
754 | @@ -686,7 +686,7 @@ void ip_tunnel_xmit(struct sk_buff *skb, struct net_device *dev, |
755 | iph->daddr = fl4.daddr; |
756 | iph->saddr = fl4.saddr; |
757 | iph->ttl = ttl; |
758 | - tunnel_ip_select_ident(skb, inner_iph, &rt->dst); |
759 | + __ip_select_ident(iph, &rt->dst, (skb_shinfo(skb)->gso_segs ?: 1) - 1); |
760 | |
761 | iptunnel_xmit(skb, dev); |
762 | return; |
763 | diff --git a/net/ipv4/raw.c b/net/ipv4/raw.c |
764 | index dd44e0a..61e60d6 100644 |
765 | --- a/net/ipv4/raw.c |
766 | +++ b/net/ipv4/raw.c |
767 | @@ -571,7 +571,8 @@ static int raw_sendmsg(struct kiocb *iocb, struct sock *sk, struct msghdr *msg, |
768 | flowi4_init_output(&fl4, ipc.oif, sk->sk_mark, tos, |
769 | RT_SCOPE_UNIVERSE, |
770 | inet->hdrincl ? IPPROTO_RAW : sk->sk_protocol, |
771 | - inet_sk_flowi_flags(sk) | FLOWI_FLAG_CAN_SLEEP, |
772 | + inet_sk_flowi_flags(sk) | FLOWI_FLAG_CAN_SLEEP | |
773 | + (inet->hdrincl ? FLOWI_FLAG_KNOWN_NH : 0), |
774 | daddr, saddr, 0, 0); |
775 | |
776 | if (!inet->hdrincl) { |
777 | diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c |
778 | index ab450c0..2005561 100644 |
779 | --- a/net/ipv4/tcp.c |
780 | +++ b/net/ipv4/tcp.c |
781 | @@ -1118,6 +1118,13 @@ new_segment: |
782 | goto wait_for_memory; |
783 | |
784 | /* |
785 | + * All packets are restored as if they have |
786 | + * already been sent. |
787 | + */ |
788 | + if (tp->repair) |
789 | + TCP_SKB_CB(skb)->when = tcp_time_stamp; |
790 | + |
791 | + /* |
792 | * Check whether we can use HW checksum. |
793 | */ |
794 | if (sk->sk_route_caps & NETIF_F_ALL_CSUM) |
795 | diff --git a/net/ipv4/tcp_cubic.c b/net/ipv4/tcp_cubic.c |
796 | index a9077f4..b6ae92a 100644 |
797 | --- a/net/ipv4/tcp_cubic.c |
798 | +++ b/net/ipv4/tcp_cubic.c |
799 | @@ -206,8 +206,8 @@ static u32 cubic_root(u64 a) |
800 | */ |
801 | static inline void bictcp_update(struct bictcp *ca, u32 cwnd) |
802 | { |
803 | - u64 offs; |
804 | - u32 delta, t, bic_target, max_cnt; |
805 | + u32 delta, bic_target, max_cnt; |
806 | + u64 offs, t; |
807 | |
808 | ca->ack_cnt++; /* count the number of ACKs */ |
809 | |
810 | @@ -250,9 +250,11 @@ static inline void bictcp_update(struct bictcp *ca, u32 cwnd) |
811 | * if the cwnd < 1 million packets !!! |
812 | */ |
813 | |
814 | + t = (s32)(tcp_time_stamp - ca->epoch_start); |
815 | + t += msecs_to_jiffies(ca->delay_min >> 3); |
816 | /* change the unit from HZ to bictcp_HZ */ |
817 | - t = ((tcp_time_stamp + msecs_to_jiffies(ca->delay_min>>3) |
818 | - - ca->epoch_start) << BICTCP_HZ) / HZ; |
819 | + t <<= BICTCP_HZ; |
820 | + do_div(t, HZ); |
821 | |
822 | if (t < ca->bic_K) /* t - K */ |
823 | offs = ca->bic_K - t; |
824 | @@ -414,7 +416,7 @@ static void bictcp_acked(struct sock *sk, u32 cnt, s32 rtt_us) |
825 | return; |
826 | |
827 | /* Discard delay samples right after fast recovery */ |
828 | - if ((s32)(tcp_time_stamp - ca->epoch_start) < HZ) |
829 | + if (ca->epoch_start && (s32)(tcp_time_stamp - ca->epoch_start) < HZ) |
830 | return; |
831 | |
832 | delay = (rtt_us << 3) / USEC_PER_MSEC; |
833 | diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c |
834 | index 9c62257..4b75aad 100644 |
835 | --- a/net/ipv4/tcp_input.c |
836 | +++ b/net/ipv4/tcp_input.c |
837 | @@ -3598,7 +3598,10 @@ static bool tcp_parse_aligned_timestamp(struct tcp_sock *tp, const struct tcphdr |
838 | ++ptr; |
839 | tp->rx_opt.rcv_tsval = ntohl(*ptr); |
840 | ++ptr; |
841 | - tp->rx_opt.rcv_tsecr = ntohl(*ptr) - tp->tsoffset; |
842 | + if (*ptr) |
843 | + tp->rx_opt.rcv_tsecr = ntohl(*ptr) - tp->tsoffset; |
844 | + else |
845 | + tp->rx_opt.rcv_tsecr = 0; |
846 | return true; |
847 | } |
848 | return false; |
849 | @@ -3623,7 +3626,7 @@ static bool tcp_fast_parse_options(const struct sk_buff *skb, |
850 | } |
851 | |
852 | tcp_parse_options(skb, &tp->rx_opt, 1, NULL); |
853 | - if (tp->rx_opt.saw_tstamp) |
854 | + if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr) |
855 | tp->rx_opt.rcv_tsecr -= tp->tsoffset; |
856 | |
857 | return true; |
858 | @@ -5376,7 +5379,7 @@ static int tcp_rcv_synsent_state_process(struct sock *sk, struct sk_buff *skb, |
859 | int saved_clamp = tp->rx_opt.mss_clamp; |
860 | |
861 | tcp_parse_options(skb, &tp->rx_opt, 0, &foc); |
862 | - if (tp->rx_opt.saw_tstamp) |
863 | + if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr) |
864 | tp->rx_opt.rcv_tsecr -= tp->tsoffset; |
865 | |
866 | if (th->ack) { |
867 | diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c |
868 | index ec335fa..0145ce7 100644 |
869 | --- a/net/ipv4/tcp_output.c |
870 | +++ b/net/ipv4/tcp_output.c |
871 | @@ -2664,7 +2664,7 @@ struct sk_buff *tcp_make_synack(struct sock *sk, struct dst_entry *dst, |
872 | int tcp_header_size; |
873 | int mss; |
874 | |
875 | - skb = alloc_skb(MAX_TCP_HEADER + 15, sk_gfp_atomic(sk, GFP_ATOMIC)); |
876 | + skb = sock_wmalloc(sk, MAX_TCP_HEADER + 15, 1, GFP_ATOMIC); |
877 | if (unlikely(!skb)) { |
878 | dst_release(dst); |
879 | return NULL; |
880 | @@ -2808,6 +2808,8 @@ void tcp_connect_init(struct sock *sk) |
881 | |
882 | if (likely(!tp->repair)) |
883 | tp->rcv_nxt = 0; |
884 | + else |
885 | + tp->rcv_tstamp = tcp_time_stamp; |
886 | tp->rcv_wup = tp->rcv_nxt; |
887 | tp->copied_seq = tp->rcv_nxt; |
888 | |
889 | diff --git a/net/ipv6/addrconf.c b/net/ipv6/addrconf.c |
890 | index fb8c94c..d3057f9 100644 |
891 | --- a/net/ipv6/addrconf.c |
892 | +++ b/net/ipv6/addrconf.c |
893 | @@ -1124,12 +1124,10 @@ retry: |
894 | if (ifp->flags & IFA_F_OPTIMISTIC) |
895 | addr_flags |= IFA_F_OPTIMISTIC; |
896 | |
897 | - ift = !max_addresses || |
898 | - ipv6_count_addresses(idev) < max_addresses ? |
899 | - ipv6_add_addr(idev, &addr, tmp_plen, |
900 | - ipv6_addr_type(&addr)&IPV6_ADDR_SCOPE_MASK, |
901 | - addr_flags) : NULL; |
902 | - if (IS_ERR_OR_NULL(ift)) { |
903 | + ift = ipv6_add_addr(idev, &addr, tmp_plen, |
904 | + ipv6_addr_type(&addr)&IPV6_ADDR_SCOPE_MASK, |
905 | + addr_flags); |
906 | + if (IS_ERR(ift)) { |
907 | in6_ifa_put(ifp); |
908 | in6_dev_put(idev); |
909 | pr_info("%s: retry temporary address regeneration\n", __func__); |
910 | diff --git a/net/ipv6/addrlabel.c b/net/ipv6/addrlabel.c |
911 | index f083a58..b30ad37 100644 |
912 | --- a/net/ipv6/addrlabel.c |
913 | +++ b/net/ipv6/addrlabel.c |
914 | @@ -251,38 +251,36 @@ static struct ip6addrlbl_entry *ip6addrlbl_alloc(struct net *net, |
915 | /* add a label */ |
916 | static int __ip6addrlbl_add(struct ip6addrlbl_entry *newp, int replace) |
917 | { |
918 | + struct hlist_node *n; |
919 | + struct ip6addrlbl_entry *last = NULL, *p = NULL; |
920 | int ret = 0; |
921 | |
922 | - ADDRLABEL(KERN_DEBUG "%s(newp=%p, replace=%d)\n", |
923 | - __func__, |
924 | - newp, replace); |
925 | + ADDRLABEL(KERN_DEBUG "%s(newp=%p, replace=%d)\n", __func__, newp, |
926 | + replace); |
927 | |
928 | - if (hlist_empty(&ip6addrlbl_table.head)) { |
929 | - hlist_add_head_rcu(&newp->list, &ip6addrlbl_table.head); |
930 | - } else { |
931 | - struct hlist_node *n; |
932 | - struct ip6addrlbl_entry *p = NULL; |
933 | - hlist_for_each_entry_safe(p, n, |
934 | - &ip6addrlbl_table.head, list) { |
935 | - if (p->prefixlen == newp->prefixlen && |
936 | - net_eq(ip6addrlbl_net(p), ip6addrlbl_net(newp)) && |
937 | - p->ifindex == newp->ifindex && |
938 | - ipv6_addr_equal(&p->prefix, &newp->prefix)) { |
939 | - if (!replace) { |
940 | - ret = -EEXIST; |
941 | - goto out; |
942 | - } |
943 | - hlist_replace_rcu(&p->list, &newp->list); |
944 | - ip6addrlbl_put(p); |
945 | - goto out; |
946 | - } else if ((p->prefixlen == newp->prefixlen && !p->ifindex) || |
947 | - (p->prefixlen < newp->prefixlen)) { |
948 | - hlist_add_before_rcu(&newp->list, &p->list); |
949 | + hlist_for_each_entry_safe(p, n, &ip6addrlbl_table.head, list) { |
950 | + if (p->prefixlen == newp->prefixlen && |
951 | + net_eq(ip6addrlbl_net(p), ip6addrlbl_net(newp)) && |
952 | + p->ifindex == newp->ifindex && |
953 | + ipv6_addr_equal(&p->prefix, &newp->prefix)) { |
954 | + if (!replace) { |
955 | + ret = -EEXIST; |
956 | goto out; |
957 | } |
958 | + hlist_replace_rcu(&p->list, &newp->list); |
959 | + ip6addrlbl_put(p); |
960 | + goto out; |
961 | + } else if ((p->prefixlen == newp->prefixlen && !p->ifindex) || |
962 | + (p->prefixlen < newp->prefixlen)) { |
963 | + hlist_add_before_rcu(&newp->list, &p->list); |
964 | + goto out; |
965 | } |
966 | - hlist_add_after_rcu(&p->list, &newp->list); |
967 | + last = p; |
968 | } |
969 | + if (last) |
970 | + hlist_add_after_rcu(&last->list, &newp->list); |
971 | + else |
972 | + hlist_add_head_rcu(&newp->list, &ip6addrlbl_table.head); |
973 | out: |
974 | if (!ret) |
975 | ip6addrlbl_table.seq++; |
976 | diff --git a/net/ipv6/icmp.c b/net/ipv6/icmp.c |
977 | index b4ff0a4..70e704d 100644 |
978 | --- a/net/ipv6/icmp.c |
979 | +++ b/net/ipv6/icmp.c |
980 | @@ -931,6 +931,14 @@ static const struct icmp6_err { |
981 | .err = ECONNREFUSED, |
982 | .fatal = 1, |
983 | }, |
984 | + { /* POLICY_FAIL */ |
985 | + .err = EACCES, |
986 | + .fatal = 1, |
987 | + }, |
988 | + { /* REJECT_ROUTE */ |
989 | + .err = EACCES, |
990 | + .fatal = 1, |
991 | + }, |
992 | }; |
993 | |
994 | int icmpv6_err_convert(u8 type, u8 code, int *err) |
995 | @@ -942,7 +950,7 @@ int icmpv6_err_convert(u8 type, u8 code, int *err) |
996 | switch (type) { |
997 | case ICMPV6_DEST_UNREACH: |
998 | fatal = 1; |
999 | - if (code <= ICMPV6_PORT_UNREACH) { |
1000 | + if (code < ARRAY_SIZE(tab_unreach)) { |
1001 | *err = tab_unreach[code].err; |
1002 | fatal = tab_unreach[code].fatal; |
1003 | } |
1004 | diff --git a/net/ipv6/ip6_fib.c b/net/ipv6/ip6_fib.c |
1005 | index 5fc9c7a..2221ff6 100644 |
1006 | --- a/net/ipv6/ip6_fib.c |
1007 | +++ b/net/ipv6/ip6_fib.c |
1008 | @@ -993,14 +993,22 @@ static struct fib6_node * fib6_lookup_1(struct fib6_node *root, |
1009 | |
1010 | if (ipv6_prefix_equal(&key->addr, args->addr, key->plen)) { |
1011 | #ifdef CONFIG_IPV6_SUBTREES |
1012 | - if (fn->subtree) |
1013 | - fn = fib6_lookup_1(fn->subtree, args + 1); |
1014 | + if (fn->subtree) { |
1015 | + struct fib6_node *sfn; |
1016 | + sfn = fib6_lookup_1(fn->subtree, |
1017 | + args + 1); |
1018 | + if (!sfn) |
1019 | + goto backtrack; |
1020 | + fn = sfn; |
1021 | + } |
1022 | #endif |
1023 | - if (!fn || fn->fn_flags & RTN_RTINFO) |
1024 | + if (fn->fn_flags & RTN_RTINFO) |
1025 | return fn; |
1026 | } |
1027 | } |
1028 | - |
1029 | +#ifdef CONFIG_IPV6_SUBTREES |
1030 | +backtrack: |
1031 | +#endif |
1032 | if (fn->fn_flags & RTN_ROOT) |
1033 | break; |
1034 | |
1035 | diff --git a/net/ipv6/ndisc.c b/net/ipv6/ndisc.c |
1036 | index ca4ffcc..060a044 100644 |
1037 | --- a/net/ipv6/ndisc.c |
1038 | +++ b/net/ipv6/ndisc.c |
1039 | @@ -372,14 +372,11 @@ static struct sk_buff *ndisc_alloc_skb(struct net_device *dev, |
1040 | int tlen = dev->needed_tailroom; |
1041 | struct sock *sk = dev_net(dev)->ipv6.ndisc_sk; |
1042 | struct sk_buff *skb; |
1043 | - int err; |
1044 | |
1045 | - skb = sock_alloc_send_skb(sk, |
1046 | - hlen + sizeof(struct ipv6hdr) + len + tlen, |
1047 | - 1, &err); |
1048 | + skb = alloc_skb(hlen + sizeof(struct ipv6hdr) + len + tlen, GFP_ATOMIC); |
1049 | if (!skb) { |
1050 | - ND_PRINTK(0, err, "ndisc: %s failed to allocate an skb, err=%d\n", |
1051 | - __func__, err); |
1052 | + ND_PRINTK(0, err, "ndisc: %s failed to allocate an skb\n", |
1053 | + __func__); |
1054 | return NULL; |
1055 | } |
1056 | |
1057 | @@ -389,6 +386,11 @@ static struct sk_buff *ndisc_alloc_skb(struct net_device *dev, |
1058 | skb_reserve(skb, hlen + sizeof(struct ipv6hdr)); |
1059 | skb_reset_transport_header(skb); |
1060 | |
1061 | + /* Manually assign socket ownership as we avoid calling |
1062 | + * sock_alloc_send_pskb() to bypass wmem buffer limits |
1063 | + */ |
1064 | + skb_set_owner_w(skb, sk); |
1065 | + |
1066 | return skb; |
1067 | } |
1068 | |
1069 | diff --git a/net/ipv6/reassembly.c b/net/ipv6/reassembly.c |
1070 | index 790d9f4..1aeb473 100644 |
1071 | --- a/net/ipv6/reassembly.c |
1072 | +++ b/net/ipv6/reassembly.c |
1073 | @@ -490,6 +490,7 @@ static int ip6_frag_reasm(struct frag_queue *fq, struct sk_buff *prev, |
1074 | ipv6_hdr(head)->payload_len = htons(payload_len); |
1075 | ipv6_change_dsfield(ipv6_hdr(head), 0xff, ecn); |
1076 | IP6CB(head)->nhoff = nhoff; |
1077 | + IP6CB(head)->flags |= IP6SKB_FRAGMENTED; |
1078 | |
1079 | /* Yes, and fold redundant checksum back. 8) */ |
1080 | if (head->ip_summed == CHECKSUM_COMPLETE) |
1081 | @@ -524,6 +525,9 @@ static int ipv6_frag_rcv(struct sk_buff *skb) |
1082 | struct net *net = dev_net(skb_dst(skb)->dev); |
1083 | int evicted; |
1084 | |
1085 | + if (IP6CB(skb)->flags & IP6SKB_FRAGMENTED) |
1086 | + goto fail_hdr; |
1087 | + |
1088 | IP6_INC_STATS_BH(net, ip6_dst_idev(skb_dst(skb)), IPSTATS_MIB_REASMREQDS); |
1089 | |
1090 | /* Jumbo payload inhibits frag. header */ |
1091 | @@ -544,6 +548,7 @@ static int ipv6_frag_rcv(struct sk_buff *skb) |
1092 | ip6_dst_idev(skb_dst(skb)), IPSTATS_MIB_REASMOKS); |
1093 | |
1094 | IP6CB(skb)->nhoff = (u8 *)fhdr - skb_network_header(skb); |
1095 | + IP6CB(skb)->flags |= IP6SKB_FRAGMENTED; |
1096 | return 1; |
1097 | } |
1098 | |
1099 | diff --git a/net/ipv6/tcp_ipv6.c b/net/ipv6/tcp_ipv6.c |
1100 | index 0a17ed9..66c7188 100644 |
1101 | --- a/net/ipv6/tcp_ipv6.c |
1102 | +++ b/net/ipv6/tcp_ipv6.c |
1103 | @@ -1426,7 +1426,7 @@ ipv6_pktoptions: |
1104 | if (np->rxopt.bits.rxhlim || np->rxopt.bits.rxohlim) |
1105 | np->mcast_hops = ipv6_hdr(opt_skb)->hop_limit; |
1106 | if (np->rxopt.bits.rxtclass) |
1107 | - np->rcv_tclass = ipv6_get_dsfield(ipv6_hdr(skb)); |
1108 | + np->rcv_tclass = ipv6_get_dsfield(ipv6_hdr(opt_skb)); |
1109 | if (ipv6_opt_accepted(sk, opt_skb)) { |
1110 | skb_set_owner_r(opt_skb, sk); |
1111 | opt_skb = xchg(&np->pktoptions, opt_skb); |
1112 | diff --git a/net/netlink/genetlink.c b/net/netlink/genetlink.c |
1113 | index 1076fe1..393f17e 100644 |
1114 | --- a/net/netlink/genetlink.c |
1115 | +++ b/net/netlink/genetlink.c |
1116 | @@ -364,7 +364,7 @@ int genl_unregister_ops(struct genl_family *family, struct genl_ops *ops) |
1117 | EXPORT_SYMBOL(genl_unregister_ops); |
1118 | |
1119 | /** |
1120 | - * genl_register_family - register a generic netlink family |
1121 | + * __genl_register_family - register a generic netlink family |
1122 | * @family: generic netlink family |
1123 | * |
1124 | * Registers the specified family after validating it first. Only one |
1125 | @@ -374,7 +374,7 @@ EXPORT_SYMBOL(genl_unregister_ops); |
1126 | * |
1127 | * Return 0 on success or a negative error code. |
1128 | */ |
1129 | -int genl_register_family(struct genl_family *family) |
1130 | +int __genl_register_family(struct genl_family *family) |
1131 | { |
1132 | int err = -EINVAL; |
1133 | |
1134 | @@ -430,10 +430,10 @@ errout_locked: |
1135 | errout: |
1136 | return err; |
1137 | } |
1138 | -EXPORT_SYMBOL(genl_register_family); |
1139 | +EXPORT_SYMBOL(__genl_register_family); |
1140 | |
1141 | /** |
1142 | - * genl_register_family_with_ops - register a generic netlink family |
1143 | + * __genl_register_family_with_ops - register a generic netlink family |
1144 | * @family: generic netlink family |
1145 | * @ops: operations to be registered |
1146 | * @n_ops: number of elements to register |
1147 | @@ -457,12 +457,12 @@ EXPORT_SYMBOL(genl_register_family); |
1148 | * |
1149 | * Return 0 on success or a negative error code. |
1150 | */ |
1151 | -int genl_register_family_with_ops(struct genl_family *family, |
1152 | +int __genl_register_family_with_ops(struct genl_family *family, |
1153 | struct genl_ops *ops, size_t n_ops) |
1154 | { |
1155 | int err, i; |
1156 | |
1157 | - err = genl_register_family(family); |
1158 | + err = __genl_register_family(family); |
1159 | if (err) |
1160 | return err; |
1161 | |
1162 | @@ -476,7 +476,7 @@ err_out: |
1163 | genl_unregister_family(family); |
1164 | return err; |
1165 | } |
1166 | -EXPORT_SYMBOL(genl_register_family_with_ops); |
1167 | +EXPORT_SYMBOL(__genl_register_family_with_ops); |
1168 | |
1169 | /** |
1170 | * genl_unregister_family - unregister generic netlink family |
1171 | @@ -544,6 +544,30 @@ void *genlmsg_put(struct sk_buff *skb, u32 portid, u32 seq, |
1172 | } |
1173 | EXPORT_SYMBOL(genlmsg_put); |
1174 | |
1175 | +static int genl_lock_dumpit(struct sk_buff *skb, struct netlink_callback *cb) |
1176 | +{ |
1177 | + struct genl_ops *ops = cb->data; |
1178 | + int rc; |
1179 | + |
1180 | + genl_lock(); |
1181 | + rc = ops->dumpit(skb, cb); |
1182 | + genl_unlock(); |
1183 | + return rc; |
1184 | +} |
1185 | + |
1186 | +static int genl_lock_done(struct netlink_callback *cb) |
1187 | +{ |
1188 | + struct genl_ops *ops = cb->data; |
1189 | + int rc = 0; |
1190 | + |
1191 | + if (ops->done) { |
1192 | + genl_lock(); |
1193 | + rc = ops->done(cb); |
1194 | + genl_unlock(); |
1195 | + } |
1196 | + return rc; |
1197 | +} |
1198 | + |
1199 | static int genl_family_rcv_msg(struct genl_family *family, |
1200 | struct sk_buff *skb, |
1201 | struct nlmsghdr *nlh) |
1202 | @@ -572,15 +596,34 @@ static int genl_family_rcv_msg(struct genl_family *family, |
1203 | return -EPERM; |
1204 | |
1205 | if (nlh->nlmsg_flags & NLM_F_DUMP) { |
1206 | - struct netlink_dump_control c = { |
1207 | - .dump = ops->dumpit, |
1208 | - .done = ops->done, |
1209 | - }; |
1210 | + int rc; |
1211 | |
1212 | if (ops->dumpit == NULL) |
1213 | return -EOPNOTSUPP; |
1214 | |
1215 | - return netlink_dump_start(net->genl_sock, skb, nlh, &c); |
1216 | + if (!family->parallel_ops) { |
1217 | + struct netlink_dump_control c = { |
1218 | + .module = family->module, |
1219 | + .data = ops, |
1220 | + .dump = genl_lock_dumpit, |
1221 | + .done = genl_lock_done, |
1222 | + }; |
1223 | + |
1224 | + genl_unlock(); |
1225 | + rc = __netlink_dump_start(net->genl_sock, skb, nlh, &c); |
1226 | + genl_lock(); |
1227 | + |
1228 | + } else { |
1229 | + struct netlink_dump_control c = { |
1230 | + .module = family->module, |
1231 | + .dump = ops->dumpit, |
1232 | + .done = ops->done, |
1233 | + }; |
1234 | + |
1235 | + rc = __netlink_dump_start(net->genl_sock, skb, nlh, &c); |
1236 | + } |
1237 | + |
1238 | + return rc; |
1239 | } |
1240 | |
1241 | if (ops->doit == NULL) |
1242 | diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c |
1243 | index 20a1bd0..a6895ab 100644 |
1244 | --- a/net/packet/af_packet.c |
1245 | +++ b/net/packet/af_packet.c |
1246 | @@ -3259,9 +3259,11 @@ static int packet_getsockopt(struct socket *sock, int level, int optname, |
1247 | |
1248 | if (po->tp_version == TPACKET_V3) { |
1249 | lv = sizeof(struct tpacket_stats_v3); |
1250 | + st.stats3.tp_packets += st.stats3.tp_drops; |
1251 | data = &st.stats3; |
1252 | } else { |
1253 | lv = sizeof(struct tpacket_stats); |
1254 | + st.stats1.tp_packets += st.stats1.tp_drops; |
1255 | data = &st.stats1; |
1256 | } |
1257 | |
1258 | diff --git a/net/sched/sch_api.c b/net/sched/sch_api.c |
1259 | index 281c1bd..51b968d 100644 |
1260 | --- a/net/sched/sch_api.c |
1261 | +++ b/net/sched/sch_api.c |
1262 | @@ -285,6 +285,45 @@ static struct Qdisc_ops *qdisc_lookup_ops(struct nlattr *kind) |
1263 | return q; |
1264 | } |
1265 | |
1266 | +/* The linklayer setting were not transferred from iproute2, in older |
1267 | + * versions, and the rate tables lookup systems have been dropped in |
1268 | + * the kernel. To keep backward compatible with older iproute2 tc |
1269 | + * utils, we detect the linklayer setting by detecting if the rate |
1270 | + * table were modified. |
1271 | + * |
1272 | + * For linklayer ATM table entries, the rate table will be aligned to |
1273 | + * 48 bytes, thus some table entries will contain the same value. The |
1274 | + * mpu (min packet unit) is also encoded into the old rate table, thus |
1275 | + * starting from the mpu, we find low and high table entries for |
1276 | + * mapping this cell. If these entries contain the same value, when |
1277 | + * the rate tables have been modified for linklayer ATM. |
1278 | + * |
1279 | + * This is done by rounding mpu to the nearest 48 bytes cell/entry, |
1280 | + * and then roundup to the next cell, calc the table entry one below, |
1281 | + * and compare. |
1282 | + */ |
1283 | +static __u8 __detect_linklayer(struct tc_ratespec *r, __u32 *rtab) |
1284 | +{ |
1285 | + int low = roundup(r->mpu, 48); |
1286 | + int high = roundup(low+1, 48); |
1287 | + int cell_low = low >> r->cell_log; |
1288 | + int cell_high = (high >> r->cell_log) - 1; |
1289 | + |
1290 | + /* rtab is too inaccurate at rates > 100Mbit/s */ |
1291 | + if ((r->rate > (100000000/8)) || (rtab[0] == 0)) { |
1292 | + pr_debug("TC linklayer: Giving up ATM detection\n"); |
1293 | + return TC_LINKLAYER_ETHERNET; |
1294 | + } |
1295 | + |
1296 | + if ((cell_high > cell_low) && (cell_high < 256) |
1297 | + && (rtab[cell_low] == rtab[cell_high])) { |
1298 | + pr_debug("TC linklayer: Detected ATM, low(%d)=high(%d)=%u\n", |
1299 | + cell_low, cell_high, rtab[cell_high]); |
1300 | + return TC_LINKLAYER_ATM; |
1301 | + } |
1302 | + return TC_LINKLAYER_ETHERNET; |
1303 | +} |
1304 | + |
1305 | static struct qdisc_rate_table *qdisc_rtab_list; |
1306 | |
1307 | struct qdisc_rate_table *qdisc_get_rtab(struct tc_ratespec *r, struct nlattr *tab) |
1308 | @@ -308,6 +347,8 @@ struct qdisc_rate_table *qdisc_get_rtab(struct tc_ratespec *r, struct nlattr *ta |
1309 | rtab->rate = *r; |
1310 | rtab->refcnt = 1; |
1311 | memcpy(rtab->data, nla_data(tab), 1024); |
1312 | + if (r->linklayer == TC_LINKLAYER_UNAWARE) |
1313 | + r->linklayer = __detect_linklayer(r, rtab->data); |
1314 | rtab->next = qdisc_rtab_list; |
1315 | qdisc_rtab_list = rtab; |
1316 | } |
1317 | diff --git a/net/sched/sch_generic.c b/net/sched/sch_generic.c |
1318 | index 2022408..a7f838b 100644 |
1319 | --- a/net/sched/sch_generic.c |
1320 | +++ b/net/sched/sch_generic.c |
1321 | @@ -908,6 +908,7 @@ void psched_ratecfg_precompute(struct psched_ratecfg *r, |
1322 | memset(r, 0, sizeof(*r)); |
1323 | r->overhead = conf->overhead; |
1324 | r->rate_bps = (u64)conf->rate << 3; |
1325 | + r->linklayer = (conf->linklayer & TC_LINKLAYER_MASK); |
1326 | r->mult = 1; |
1327 | /* |
1328 | * Calibrate mult, shift so that token counting is accurate |
1329 | diff --git a/net/sched/sch_htb.c b/net/sched/sch_htb.c |
1330 | index adaedd7..910667c 100644 |
1331 | --- a/net/sched/sch_htb.c |
1332 | +++ b/net/sched/sch_htb.c |
1333 | @@ -87,7 +87,7 @@ struct htb_class { |
1334 | unsigned int children; |
1335 | struct htb_class *parent; /* parent class */ |
1336 | |
1337 | - int prio; /* these two are used only by leaves... */ |
1338 | + u32 prio; /* these two are used only by leaves... */ |
1339 | int quantum; /* but stored for parent-to-leaf return */ |
1340 | |
1341 | union { |
1342 | @@ -1312,6 +1312,7 @@ static int htb_change_class(struct Qdisc *sch, u32 classid, |
1343 | struct htb_sched *q = qdisc_priv(sch); |
1344 | struct htb_class *cl = (struct htb_class *)*arg, *parent; |
1345 | struct nlattr *opt = tca[TCA_OPTIONS]; |
1346 | + struct qdisc_rate_table *rtab = NULL, *ctab = NULL; |
1347 | struct nlattr *tb[TCA_HTB_MAX + 1]; |
1348 | struct tc_htb_opt *hopt; |
1349 | |
1350 | @@ -1333,6 +1334,18 @@ static int htb_change_class(struct Qdisc *sch, u32 classid, |
1351 | if (!hopt->rate.rate || !hopt->ceil.rate) |
1352 | goto failure; |
1353 | |
1354 | + /* Keeping backward compatible with rate_table based iproute2 tc */ |
1355 | + if (hopt->rate.linklayer == TC_LINKLAYER_UNAWARE) { |
1356 | + rtab = qdisc_get_rtab(&hopt->rate, tb[TCA_HTB_RTAB]); |
1357 | + if (rtab) |
1358 | + qdisc_put_rtab(rtab); |
1359 | + } |
1360 | + if (hopt->ceil.linklayer == TC_LINKLAYER_UNAWARE) { |
1361 | + ctab = qdisc_get_rtab(&hopt->ceil, tb[TCA_HTB_CTAB]); |
1362 | + if (ctab) |
1363 | + qdisc_put_rtab(ctab); |
1364 | + } |
1365 | + |
1366 | if (!cl) { /* new class */ |
1367 | struct Qdisc *new_q; |
1368 | int prio; |
1369 | diff --git a/net/tipc/socket.c b/net/tipc/socket.c |
1370 | index 515ce38..7e26ad4 100644 |
1371 | --- a/net/tipc/socket.c |
1372 | +++ b/net/tipc/socket.c |
1373 | @@ -1179,7 +1179,7 @@ static u32 filter_connect(struct tipc_sock *tsock, struct sk_buff **buf) |
1374 | /* Accept only ACK or NACK message */ |
1375 | if (unlikely(msg_errcode(msg))) { |
1376 | sock->state = SS_DISCONNECTING; |
1377 | - sk->sk_err = -ECONNREFUSED; |
1378 | + sk->sk_err = ECONNREFUSED; |
1379 | retval = TIPC_OK; |
1380 | break; |
1381 | } |
1382 | @@ -1190,7 +1190,7 @@ static u32 filter_connect(struct tipc_sock *tsock, struct sk_buff **buf) |
1383 | res = auto_connect(sock, msg); |
1384 | if (res) { |
1385 | sock->state = SS_DISCONNECTING; |
1386 | - sk->sk_err = res; |
1387 | + sk->sk_err = -res; |
1388 | retval = TIPC_OK; |
1389 | break; |
1390 | } |