This source file includes following definitions.
- is_supervision_frame
- create_stripped_skb
- frame_get_stripped_skb
- hsr_fill_tag
- create_tagged_skb
- frame_get_tagged_skb
- hsr_deliver_master
- hsr_xmit
- hsr_forward_do
- check_local_dest
- hsr_fill_frame_info
- hsr_forward_skb
1
2
3
4
5
6
7
8 #include "hsr_forward.h"
9 #include <linux/types.h>
10 #include <linux/skbuff.h>
11 #include <linux/etherdevice.h>
12 #include <linux/if_vlan.h>
13 #include "hsr_main.h"
14 #include "hsr_framereg.h"
15
16 struct hsr_node;
17
18 struct hsr_frame_info {
19 struct sk_buff *skb_std;
20 struct sk_buff *skb_hsr;
21 struct hsr_port *port_rcv;
22 struct hsr_node *node_src;
23 u16 sequence_nr;
24 bool is_supervision;
25 bool is_vlan;
26 bool is_local_dest;
27 bool is_local_exclusive;
28 };
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45 static bool is_supervision_frame(struct hsr_priv *hsr, struct sk_buff *skb)
46 {
47 struct ethhdr *eth_hdr;
48 struct hsr_sup_tag *hsr_sup_tag;
49 struct hsrv1_ethhdr_sp *hsr_V1_hdr;
50
51 WARN_ON_ONCE(!skb_mac_header_was_set(skb));
52 eth_hdr = (struct ethhdr *)skb_mac_header(skb);
53
54
55 if (!ether_addr_equal(eth_hdr->h_dest,
56 hsr->sup_multicast_addr))
57 return false;
58
59
60 if (!(eth_hdr->h_proto == htons(ETH_P_PRP) ||
61 eth_hdr->h_proto == htons(ETH_P_HSR)))
62 return false;
63
64
65 if (eth_hdr->h_proto == htons(ETH_P_HSR)) {
66 hsr_V1_hdr = (struct hsrv1_ethhdr_sp *)skb_mac_header(skb);
67 if (hsr_V1_hdr->hsr.encap_proto != htons(ETH_P_PRP))
68 return false;
69
70 hsr_sup_tag = &hsr_V1_hdr->hsr_sup;
71 } else {
72 hsr_sup_tag =
73 &((struct hsrv0_ethhdr_sp *)skb_mac_header(skb))->hsr_sup;
74 }
75
76 if (hsr_sup_tag->HSR_TLV_type != HSR_TLV_ANNOUNCE &&
77 hsr_sup_tag->HSR_TLV_type != HSR_TLV_LIFE_CHECK)
78 return false;
79 if (hsr_sup_tag->HSR_TLV_length != 12 &&
80 hsr_sup_tag->HSR_TLV_length != sizeof(struct hsr_sup_payload))
81 return false;
82
83 return true;
84 }
85
86 static struct sk_buff *create_stripped_skb(struct sk_buff *skb_in,
87 struct hsr_frame_info *frame)
88 {
89 struct sk_buff *skb;
90 int copylen;
91 unsigned char *dst, *src;
92
93 skb_pull(skb_in, HSR_HLEN);
94 skb = __pskb_copy(skb_in, skb_headroom(skb_in) - HSR_HLEN, GFP_ATOMIC);
95 skb_push(skb_in, HSR_HLEN);
96 if (!skb)
97 return NULL;
98
99 skb_reset_mac_header(skb);
100
101 if (skb->ip_summed == CHECKSUM_PARTIAL)
102 skb->csum_start -= HSR_HLEN;
103
104 copylen = 2 * ETH_ALEN;
105 if (frame->is_vlan)
106 copylen += VLAN_HLEN;
107 src = skb_mac_header(skb_in);
108 dst = skb_mac_header(skb);
109 memcpy(dst, src, copylen);
110
111 skb->protocol = eth_hdr(skb)->h_proto;
112 return skb;
113 }
114
115 static struct sk_buff *frame_get_stripped_skb(struct hsr_frame_info *frame,
116 struct hsr_port *port)
117 {
118 if (!frame->skb_std)
119 frame->skb_std = create_stripped_skb(frame->skb_hsr, frame);
120 return skb_clone(frame->skb_std, GFP_ATOMIC);
121 }
122
123 static void hsr_fill_tag(struct sk_buff *skb, struct hsr_frame_info *frame,
124 struct hsr_port *port, u8 proto_version)
125 {
126 struct hsr_ethhdr *hsr_ethhdr;
127 int lane_id;
128 int lsdu_size;
129
130 if (port->type == HSR_PT_SLAVE_A)
131 lane_id = 0;
132 else
133 lane_id = 1;
134
135 lsdu_size = skb->len - 14;
136 if (frame->is_vlan)
137 lsdu_size -= 4;
138
139 hsr_ethhdr = (struct hsr_ethhdr *)skb_mac_header(skb);
140
141 set_hsr_tag_path(&hsr_ethhdr->hsr_tag, lane_id);
142 set_hsr_tag_LSDU_size(&hsr_ethhdr->hsr_tag, lsdu_size);
143 hsr_ethhdr->hsr_tag.sequence_nr = htons(frame->sequence_nr);
144 hsr_ethhdr->hsr_tag.encap_proto = hsr_ethhdr->ethhdr.h_proto;
145 hsr_ethhdr->ethhdr.h_proto = htons(proto_version ?
146 ETH_P_HSR : ETH_P_PRP);
147 }
148
149 static struct sk_buff *create_tagged_skb(struct sk_buff *skb_o,
150 struct hsr_frame_info *frame,
151 struct hsr_port *port)
152 {
153 int movelen;
154 unsigned char *dst, *src;
155 struct sk_buff *skb;
156
157
158 skb = __pskb_copy(skb_o, skb_headroom(skb_o) + HSR_HLEN, GFP_ATOMIC);
159 if (!skb)
160 return NULL;
161 skb_reset_mac_header(skb);
162
163 if (skb->ip_summed == CHECKSUM_PARTIAL)
164 skb->csum_start += HSR_HLEN;
165
166 movelen = ETH_HLEN;
167 if (frame->is_vlan)
168 movelen += VLAN_HLEN;
169
170 src = skb_mac_header(skb);
171 dst = skb_push(skb, HSR_HLEN);
172 memmove(dst, src, movelen);
173 skb_reset_mac_header(skb);
174
175 hsr_fill_tag(skb, frame, port, port->hsr->prot_version);
176
177 return skb;
178 }
179
180
181
182
183 static struct sk_buff *frame_get_tagged_skb(struct hsr_frame_info *frame,
184 struct hsr_port *port)
185 {
186 if (frame->skb_hsr)
187 return skb_clone(frame->skb_hsr, GFP_ATOMIC);
188
189 if (port->type != HSR_PT_SLAVE_A && port->type != HSR_PT_SLAVE_B) {
190 WARN_ONCE(1, "HSR: Bug: trying to create a tagged frame for a non-ring port");
191 return NULL;
192 }
193
194 return create_tagged_skb(frame->skb_std, frame, port);
195 }
196
197 static void hsr_deliver_master(struct sk_buff *skb, struct net_device *dev,
198 struct hsr_node *node_src)
199 {
200 bool was_multicast_frame;
201 int res;
202
203 was_multicast_frame = (skb->pkt_type == PACKET_MULTICAST);
204 hsr_addr_subst_source(node_src, skb);
205 skb_pull(skb, ETH_HLEN);
206 res = netif_rx(skb);
207 if (res == NET_RX_DROP) {
208 dev->stats.rx_dropped++;
209 } else {
210 dev->stats.rx_packets++;
211 dev->stats.rx_bytes += skb->len;
212 if (was_multicast_frame)
213 dev->stats.multicast++;
214 }
215 }
216
217 static int hsr_xmit(struct sk_buff *skb, struct hsr_port *port,
218 struct hsr_frame_info *frame)
219 {
220 if (frame->port_rcv->type == HSR_PT_MASTER) {
221 hsr_addr_subst_dest(frame->node_src, skb, port);
222
223
224
225
226 ether_addr_copy(eth_hdr(skb)->h_source, port->dev->dev_addr);
227 }
228 return dev_queue_xmit(skb);
229 }
230
231
232
233
234
235
236
237
238
239
240
241
242 static void hsr_forward_do(struct hsr_frame_info *frame)
243 {
244 struct hsr_port *port;
245 struct sk_buff *skb;
246
247 hsr_for_each_port(frame->port_rcv->hsr, port) {
248
249 if (port == frame->port_rcv)
250 continue;
251
252
253 if (port->type == HSR_PT_MASTER && !frame->is_local_dest)
254 continue;
255
256
257 if (port->type != HSR_PT_MASTER && frame->is_local_exclusive)
258 continue;
259
260
261 if (hsr_register_frame_out(port, frame->node_src,
262 frame->sequence_nr))
263 continue;
264
265 if (frame->is_supervision && port->type == HSR_PT_MASTER) {
266 hsr_handle_sup_frame(frame->skb_hsr,
267 frame->node_src,
268 frame->port_rcv);
269 continue;
270 }
271
272 if (port->type != HSR_PT_MASTER)
273 skb = frame_get_tagged_skb(frame, port);
274 else
275 skb = frame_get_stripped_skb(frame, port);
276 if (!skb) {
277
278 continue;
279 }
280
281 skb->dev = port->dev;
282 if (port->type == HSR_PT_MASTER)
283 hsr_deliver_master(skb, port->dev, frame->node_src);
284 else
285 hsr_xmit(skb, port, frame);
286 }
287 }
288
289 static void check_local_dest(struct hsr_priv *hsr, struct sk_buff *skb,
290 struct hsr_frame_info *frame)
291 {
292 if (hsr_addr_is_self(hsr, eth_hdr(skb)->h_dest)) {
293 frame->is_local_exclusive = true;
294 skb->pkt_type = PACKET_HOST;
295 } else {
296 frame->is_local_exclusive = false;
297 }
298
299 if (skb->pkt_type == PACKET_HOST ||
300 skb->pkt_type == PACKET_MULTICAST ||
301 skb->pkt_type == PACKET_BROADCAST) {
302 frame->is_local_dest = true;
303 } else {
304 frame->is_local_dest = false;
305 }
306 }
307
308 static int hsr_fill_frame_info(struct hsr_frame_info *frame,
309 struct sk_buff *skb, struct hsr_port *port)
310 {
311 struct ethhdr *ethhdr;
312 unsigned long irqflags;
313
314 frame->is_supervision = is_supervision_frame(port->hsr, skb);
315 frame->node_src = hsr_get_node(port, skb, frame->is_supervision);
316 if (!frame->node_src)
317 return -1;
318
319 ethhdr = (struct ethhdr *)skb_mac_header(skb);
320 frame->is_vlan = false;
321 if (ethhdr->h_proto == htons(ETH_P_8021Q)) {
322 frame->is_vlan = true;
323
324 WARN_ONCE(1, "HSR: VLAN not yet supported");
325 }
326 if (ethhdr->h_proto == htons(ETH_P_PRP) ||
327 ethhdr->h_proto == htons(ETH_P_HSR)) {
328 frame->skb_std = NULL;
329 frame->skb_hsr = skb;
330 frame->sequence_nr = hsr_get_skb_sequence_nr(skb);
331 } else {
332 frame->skb_std = skb;
333 frame->skb_hsr = NULL;
334
335 spin_lock_irqsave(&port->hsr->seqnr_lock, irqflags);
336 frame->sequence_nr = port->hsr->sequence_nr;
337 port->hsr->sequence_nr++;
338 spin_unlock_irqrestore(&port->hsr->seqnr_lock, irqflags);
339 }
340
341 frame->port_rcv = port;
342 check_local_dest(port->hsr, skb, frame);
343
344 return 0;
345 }
346
347
348 void hsr_forward_skb(struct sk_buff *skb, struct hsr_port *port)
349 {
350 struct hsr_frame_info frame;
351
352 if (skb_mac_header(skb) != skb->data) {
353 WARN_ONCE(1, "%s:%d: Malformed frame (port_src %s)\n",
354 __FILE__, __LINE__, port->dev->name);
355 goto out_drop;
356 }
357
358 if (hsr_fill_frame_info(&frame, skb, port) < 0)
359 goto out_drop;
360 hsr_register_frame_in(frame.node_src, port, frame.sequence_nr);
361 hsr_forward_do(&frame);
362
363
364
365 if (port->type == HSR_PT_MASTER) {
366 port->dev->stats.tx_packets++;
367 port->dev->stats.tx_bytes += skb->len;
368 }
369
370 if (frame.skb_hsr)
371 kfree_skb(frame.skb_hsr);
372 if (frame.skb_std)
373 kfree_skb(frame.skb_std);
374 return;
375
376 out_drop:
377 port->dev->stats.tx_dropped++;
378 kfree_skb(skb);
379 }