1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (c) 2010-2011 EIA Electronics,
3 // Pieter Beyens <pieter.beyens@eia.be>
4 // Copyright (c) 2010-2011 EIA Electronics,
5 // Kurt Van Dijck <kurt.van.dijck@eia.be>
6 // Copyright (c) 2018 Protonic,
7 // Robin van der Gracht <robin@protonic.nl>
8 // Copyright (c) 2017-2019 Pengutronix,
9 // Marc Kleine-Budde <kernel@pengutronix.de>
10 // Copyright (c) 2017-2019 Pengutronix,
11 // Oleksij Rempel <kernel@pengutronix.de>
12
13 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14
15 #include <linux/can/can-ml.h>
16 #include <linux/can/core.h>
17 #include <linux/can/skb.h>
18 #include <linux/errqueue.h>
19 #include <linux/if_arp.h>
20
21 #include "j1939-priv.h"
22
23 #define J1939_MIN_NAMELEN CAN_REQUIRED_SIZE(struct sockaddr_can, can_addr.j1939)
24
25 /* conversion function between struct sock::sk_priority from linux and
26 * j1939 priority field
27 */
j1939_prio(u32 sk_priority)28 static inline priority_t j1939_prio(u32 sk_priority)
29 {
30 sk_priority = min(sk_priority, 7U);
31
32 return 7 - sk_priority;
33 }
34
j1939_to_sk_priority(priority_t prio)35 static inline u32 j1939_to_sk_priority(priority_t prio)
36 {
37 return 7 - prio;
38 }
39
40 /* function to see if pgn is to be evaluated */
j1939_pgn_is_valid(pgn_t pgn)41 static inline bool j1939_pgn_is_valid(pgn_t pgn)
42 {
43 return pgn <= J1939_PGN_MAX;
44 }
45
46 /* test function to avoid non-zero DA placeholder for pdu1 pgn's */
j1939_pgn_is_clean_pdu(pgn_t pgn)47 static inline bool j1939_pgn_is_clean_pdu(pgn_t pgn)
48 {
49 if (j1939_pgn_is_pdu1(pgn))
50 return !(pgn & 0xff);
51 else
52 return true;
53 }
54
j1939_sock_pending_add(struct sock * sk)55 static inline void j1939_sock_pending_add(struct sock *sk)
56 {
57 struct j1939_sock *jsk = j1939_sk(sk);
58
59 atomic_inc(&jsk->skb_pending);
60 }
61
j1939_sock_pending_get(struct sock * sk)62 static int j1939_sock_pending_get(struct sock *sk)
63 {
64 struct j1939_sock *jsk = j1939_sk(sk);
65
66 return atomic_read(&jsk->skb_pending);
67 }
68
j1939_sock_pending_del(struct sock * sk)69 void j1939_sock_pending_del(struct sock *sk)
70 {
71 struct j1939_sock *jsk = j1939_sk(sk);
72
73 /* atomic_dec_return returns the new value */
74 if (!atomic_dec_return(&jsk->skb_pending))
75 wake_up(&jsk->waitq); /* no pending SKB's */
76 }
77
j1939_jsk_add(struct j1939_priv * priv,struct j1939_sock * jsk)78 static void j1939_jsk_add(struct j1939_priv *priv, struct j1939_sock *jsk)
79 {
80 jsk->state |= J1939_SOCK_BOUND;
81 j1939_priv_get(priv);
82
83 spin_lock_bh(&priv->j1939_socks_lock);
84 list_add_tail(&jsk->list, &priv->j1939_socks);
85 spin_unlock_bh(&priv->j1939_socks_lock);
86 }
87
j1939_jsk_del(struct j1939_priv * priv,struct j1939_sock * jsk)88 static void j1939_jsk_del(struct j1939_priv *priv, struct j1939_sock *jsk)
89 {
90 spin_lock_bh(&priv->j1939_socks_lock);
91 list_del_init(&jsk->list);
92 spin_unlock_bh(&priv->j1939_socks_lock);
93
94 j1939_priv_put(priv);
95 jsk->state &= ~J1939_SOCK_BOUND;
96 }
97
j1939_sk_queue_session(struct j1939_session * session)98 static bool j1939_sk_queue_session(struct j1939_session *session)
99 {
100 struct j1939_sock *jsk = j1939_sk(session->sk);
101 bool empty;
102
103 spin_lock_bh(&jsk->sk_session_queue_lock);
104 empty = list_empty(&jsk->sk_session_queue);
105 j1939_session_get(session);
106 list_add_tail(&session->sk_session_queue_entry, &jsk->sk_session_queue);
107 spin_unlock_bh(&jsk->sk_session_queue_lock);
108 j1939_sock_pending_add(&jsk->sk);
109
110 return empty;
111 }
112
113 static struct
j1939_sk_get_incomplete_session(struct j1939_sock * jsk)114 j1939_session *j1939_sk_get_incomplete_session(struct j1939_sock *jsk)
115 {
116 struct j1939_session *session = NULL;
117
118 spin_lock_bh(&jsk->sk_session_queue_lock);
119 if (!list_empty(&jsk->sk_session_queue)) {
120 session = list_last_entry(&jsk->sk_session_queue,
121 struct j1939_session,
122 sk_session_queue_entry);
123 if (session->total_queued_size == session->total_message_size)
124 session = NULL;
125 else
126 j1939_session_get(session);
127 }
128 spin_unlock_bh(&jsk->sk_session_queue_lock);
129
130 return session;
131 }
132
j1939_sk_queue_drop_all(struct j1939_priv * priv,struct j1939_sock * jsk,int err)133 static void j1939_sk_queue_drop_all(struct j1939_priv *priv,
134 struct j1939_sock *jsk, int err)
135 {
136 struct j1939_session *session, *tmp;
137
138 netdev_dbg(priv->ndev, "%s: err: %i\n", __func__, err);
139 spin_lock_bh(&jsk->sk_session_queue_lock);
140 list_for_each_entry_safe(session, tmp, &jsk->sk_session_queue,
141 sk_session_queue_entry) {
142 list_del_init(&session->sk_session_queue_entry);
143 session->err = err;
144 j1939_session_put(session);
145 }
146 spin_unlock_bh(&jsk->sk_session_queue_lock);
147 }
148
j1939_sk_queue_activate_next_locked(struct j1939_session * session)149 static void j1939_sk_queue_activate_next_locked(struct j1939_session *session)
150 {
151 struct j1939_sock *jsk;
152 struct j1939_session *first;
153 int err;
154
155 /* RX-Session don't have a socket (yet) */
156 if (!session->sk)
157 return;
158
159 jsk = j1939_sk(session->sk);
160 lockdep_assert_held(&jsk->sk_session_queue_lock);
161
162 err = session->err;
163
164 first = list_first_entry_or_null(&jsk->sk_session_queue,
165 struct j1939_session,
166 sk_session_queue_entry);
167
168 /* Some else has already activated the next session */
169 if (first != session)
170 return;
171
172 activate_next:
173 list_del_init(&first->sk_session_queue_entry);
174 j1939_session_put(first);
175 first = list_first_entry_or_null(&jsk->sk_session_queue,
176 struct j1939_session,
177 sk_session_queue_entry);
178 if (!first)
179 return;
180
181 if (WARN_ON_ONCE(j1939_session_activate(first))) {
182 first->err = -EBUSY;
183 goto activate_next;
184 } else {
185 /* Give receiver some time (arbitrary chosen) to recover */
186 int time_ms = 0;
187
188 if (err)
189 time_ms = 10 + prandom_u32_max(16);
190
191 j1939_tp_schedule_txtimer(first, time_ms);
192 }
193 }
194
j1939_sk_queue_activate_next(struct j1939_session * session)195 void j1939_sk_queue_activate_next(struct j1939_session *session)
196 {
197 struct j1939_sock *jsk;
198
199 if (!session->sk)
200 return;
201
202 jsk = j1939_sk(session->sk);
203
204 spin_lock_bh(&jsk->sk_session_queue_lock);
205 j1939_sk_queue_activate_next_locked(session);
206 spin_unlock_bh(&jsk->sk_session_queue_lock);
207 }
208
j1939_sk_match_dst(struct j1939_sock * jsk,const struct j1939_sk_buff_cb * skcb)209 static bool j1939_sk_match_dst(struct j1939_sock *jsk,
210 const struct j1939_sk_buff_cb *skcb)
211 {
212 if ((jsk->state & J1939_SOCK_PROMISC))
213 return true;
214
215 /* Destination address filter */
216 if (jsk->addr.src_name && skcb->addr.dst_name) {
217 if (jsk->addr.src_name != skcb->addr.dst_name)
218 return false;
219 } else {
220 /* receive (all sockets) if
221 * - all packages that match our bind() address
222 * - all broadcast on a socket if SO_BROADCAST
223 * is set
224 */
225 if (j1939_address_is_unicast(skcb->addr.da)) {
226 if (jsk->addr.sa != skcb->addr.da)
227 return false;
228 } else if (!sock_flag(&jsk->sk, SOCK_BROADCAST)) {
229 /* receiving broadcast without SO_BROADCAST
230 * flag is not allowed
231 */
232 return false;
233 }
234 }
235
236 /* Source address filter */
237 if (jsk->state & J1939_SOCK_CONNECTED) {
238 /* receive (all sockets) if
239 * - all packages that match our connect() name or address
240 */
241 if (jsk->addr.dst_name && skcb->addr.src_name) {
242 if (jsk->addr.dst_name != skcb->addr.src_name)
243 return false;
244 } else {
245 if (jsk->addr.da != skcb->addr.sa)
246 return false;
247 }
248 }
249
250 /* PGN filter */
251 if (j1939_pgn_is_valid(jsk->pgn_rx_filter) &&
252 jsk->pgn_rx_filter != skcb->addr.pgn)
253 return false;
254
255 return true;
256 }
257
258 /* matches skb control buffer (addr) with a j1939 filter */
j1939_sk_match_filter(struct j1939_sock * jsk,const struct j1939_sk_buff_cb * skcb)259 static bool j1939_sk_match_filter(struct j1939_sock *jsk,
260 const struct j1939_sk_buff_cb *skcb)
261 {
262 const struct j1939_filter *f = jsk->filters;
263 int nfilter = jsk->nfilters;
264
265 if (!nfilter)
266 /* receive all when no filters are assigned */
267 return true;
268
269 for (; nfilter; ++f, --nfilter) {
270 if ((skcb->addr.pgn & f->pgn_mask) != f->pgn)
271 continue;
272 if ((skcb->addr.sa & f->addr_mask) != f->addr)
273 continue;
274 if ((skcb->addr.src_name & f->name_mask) != f->name)
275 continue;
276 return true;
277 }
278 return false;
279 }
280
j1939_sk_recv_match_one(struct j1939_sock * jsk,const struct j1939_sk_buff_cb * skcb)281 static bool j1939_sk_recv_match_one(struct j1939_sock *jsk,
282 const struct j1939_sk_buff_cb *skcb)
283 {
284 if (!(jsk->state & J1939_SOCK_BOUND))
285 return false;
286
287 if (!j1939_sk_match_dst(jsk, skcb))
288 return false;
289
290 if (!j1939_sk_match_filter(jsk, skcb))
291 return false;
292
293 return true;
294 }
295
j1939_sk_recv_one(struct j1939_sock * jsk,struct sk_buff * oskb)296 static void j1939_sk_recv_one(struct j1939_sock *jsk, struct sk_buff *oskb)
297 {
298 const struct j1939_sk_buff_cb *oskcb = j1939_skb_to_cb(oskb);
299 struct j1939_sk_buff_cb *skcb;
300 struct sk_buff *skb;
301
302 if (oskb->sk == &jsk->sk)
303 return;
304
305 if (!j1939_sk_recv_match_one(jsk, oskcb))
306 return;
307
308 skb = skb_clone(oskb, GFP_ATOMIC);
309 if (!skb) {
310 pr_warn("skb clone failed\n");
311 return;
312 }
313 can_skb_set_owner(skb, oskb->sk);
314
315 skcb = j1939_skb_to_cb(skb);
316 skcb->msg_flags &= ~(MSG_DONTROUTE);
317 if (skb->sk)
318 skcb->msg_flags |= MSG_DONTROUTE;
319
320 if (sock_queue_rcv_skb(&jsk->sk, skb) < 0)
321 kfree_skb(skb);
322 }
323
j1939_sk_recv_match(struct j1939_priv * priv,struct j1939_sk_buff_cb * skcb)324 bool j1939_sk_recv_match(struct j1939_priv *priv, struct j1939_sk_buff_cb *skcb)
325 {
326 struct j1939_sock *jsk;
327 bool match = false;
328
329 spin_lock_bh(&priv->j1939_socks_lock);
330 list_for_each_entry(jsk, &priv->j1939_socks, list) {
331 match = j1939_sk_recv_match_one(jsk, skcb);
332 if (match)
333 break;
334 }
335 spin_unlock_bh(&priv->j1939_socks_lock);
336
337 return match;
338 }
339
j1939_sk_recv(struct j1939_priv * priv,struct sk_buff * skb)340 void j1939_sk_recv(struct j1939_priv *priv, struct sk_buff *skb)
341 {
342 struct j1939_sock *jsk;
343
344 spin_lock_bh(&priv->j1939_socks_lock);
345 list_for_each_entry(jsk, &priv->j1939_socks, list) {
346 j1939_sk_recv_one(jsk, skb);
347 }
348 spin_unlock_bh(&priv->j1939_socks_lock);
349 }
350
j1939_sk_sock_destruct(struct sock * sk)351 static void j1939_sk_sock_destruct(struct sock *sk)
352 {
353 struct j1939_sock *jsk = j1939_sk(sk);
354
355 /* This function will be called by the generic networking code, when
356 * the socket is ultimately closed (sk->sk_destruct).
357 *
358 * The race between
359 * - processing a received CAN frame
360 * (can_receive -> j1939_can_recv)
361 * and accessing j1939_priv
362 * ... and ...
363 * - closing a socket
364 * (j1939_can_rx_unregister -> can_rx_unregister)
365 * and calling the final j1939_priv_put()
366 *
367 * is avoided by calling the final j1939_priv_put() from this
368 * RCU deferred cleanup call.
369 */
370 if (jsk->priv) {
371 j1939_priv_put(jsk->priv);
372 jsk->priv = NULL;
373 }
374
375 /* call generic CAN sock destruct */
376 can_sock_destruct(sk);
377 }
378
j1939_sk_init(struct sock * sk)379 static int j1939_sk_init(struct sock *sk)
380 {
381 struct j1939_sock *jsk = j1939_sk(sk);
382
383 /* Ensure that "sk" is first member in "struct j1939_sock", so that we
384 * can skip it during memset().
385 */
386 BUILD_BUG_ON(offsetof(struct j1939_sock, sk) != 0);
387 memset((void *)jsk + sizeof(jsk->sk), 0x0,
388 sizeof(*jsk) - sizeof(jsk->sk));
389
390 INIT_LIST_HEAD(&jsk->list);
391 init_waitqueue_head(&jsk->waitq);
392 jsk->sk.sk_priority = j1939_to_sk_priority(6);
393 jsk->sk.sk_reuse = 1; /* per default */
394 jsk->addr.sa = J1939_NO_ADDR;
395 jsk->addr.da = J1939_NO_ADDR;
396 jsk->addr.pgn = J1939_NO_PGN;
397 jsk->pgn_rx_filter = J1939_NO_PGN;
398 atomic_set(&jsk->skb_pending, 0);
399 spin_lock_init(&jsk->sk_session_queue_lock);
400 INIT_LIST_HEAD(&jsk->sk_session_queue);
401
402 /* j1939_sk_sock_destruct() depends on SOCK_RCU_FREE flag */
403 sock_set_flag(sk, SOCK_RCU_FREE);
404 sk->sk_destruct = j1939_sk_sock_destruct;
405 sk->sk_protocol = CAN_J1939;
406
407 return 0;
408 }
409
j1939_sk_sanity_check(struct sockaddr_can * addr,int len)410 static int j1939_sk_sanity_check(struct sockaddr_can *addr, int len)
411 {
412 if (!addr)
413 return -EDESTADDRREQ;
414 if (len < J1939_MIN_NAMELEN)
415 return -EINVAL;
416 if (addr->can_family != AF_CAN)
417 return -EINVAL;
418 if (!addr->can_ifindex)
419 return -ENODEV;
420 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) &&
421 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn))
422 return -EINVAL;
423
424 return 0;
425 }
426
j1939_sk_bind(struct socket * sock,struct sockaddr * uaddr,int len)427 static int j1939_sk_bind(struct socket *sock, struct sockaddr *uaddr, int len)
428 {
429 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
430 struct j1939_sock *jsk = j1939_sk(sock->sk);
431 struct j1939_priv *priv;
432 struct sock *sk;
433 struct net *net;
434 int ret = 0;
435
436 ret = j1939_sk_sanity_check(addr, len);
437 if (ret)
438 return ret;
439
440 lock_sock(sock->sk);
441
442 priv = jsk->priv;
443 sk = sock->sk;
444 net = sock_net(sk);
445
446 /* Already bound to an interface? */
447 if (jsk->state & J1939_SOCK_BOUND) {
448 /* A re-bind() to a different interface is not
449 * supported.
450 */
451 if (jsk->ifindex != addr->can_ifindex) {
452 ret = -EINVAL;
453 goto out_release_sock;
454 }
455
456 /* drop old references */
457 j1939_jsk_del(priv, jsk);
458 j1939_local_ecu_put(priv, jsk->addr.src_name, jsk->addr.sa);
459 } else {
460 struct can_ml_priv *can_ml;
461 struct net_device *ndev;
462
463 ndev = dev_get_by_index(net, addr->can_ifindex);
464 if (!ndev) {
465 ret = -ENODEV;
466 goto out_release_sock;
467 }
468
469 can_ml = can_get_ml_priv(ndev);
470 if (!can_ml) {
471 dev_put(ndev);
472 ret = -ENODEV;
473 goto out_release_sock;
474 }
475
476 if (!(ndev->flags & IFF_UP)) {
477 dev_put(ndev);
478 ret = -ENETDOWN;
479 goto out_release_sock;
480 }
481
482 priv = j1939_netdev_start(ndev);
483 dev_put(ndev);
484 if (IS_ERR(priv)) {
485 ret = PTR_ERR(priv);
486 goto out_release_sock;
487 }
488
489 jsk->ifindex = addr->can_ifindex;
490
491 /* the corresponding j1939_priv_put() is called via
492 * sk->sk_destruct, which points to j1939_sk_sock_destruct()
493 */
494 j1939_priv_get(priv);
495 jsk->priv = priv;
496 }
497
498 /* set default transmit pgn */
499 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
500 jsk->pgn_rx_filter = addr->can_addr.j1939.pgn;
501 jsk->addr.src_name = addr->can_addr.j1939.name;
502 jsk->addr.sa = addr->can_addr.j1939.addr;
503
504 /* get new references */
505 ret = j1939_local_ecu_get(priv, jsk->addr.src_name, jsk->addr.sa);
506 if (ret) {
507 j1939_netdev_stop(priv);
508 goto out_release_sock;
509 }
510
511 j1939_jsk_add(priv, jsk);
512
513 out_release_sock: /* fall through */
514 release_sock(sock->sk);
515
516 return ret;
517 }
518
j1939_sk_connect(struct socket * sock,struct sockaddr * uaddr,int len,int flags)519 static int j1939_sk_connect(struct socket *sock, struct sockaddr *uaddr,
520 int len, int flags)
521 {
522 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
523 struct j1939_sock *jsk = j1939_sk(sock->sk);
524 int ret = 0;
525
526 ret = j1939_sk_sanity_check(addr, len);
527 if (ret)
528 return ret;
529
530 lock_sock(sock->sk);
531
532 /* bind() before connect() is mandatory */
533 if (!(jsk->state & J1939_SOCK_BOUND)) {
534 ret = -EINVAL;
535 goto out_release_sock;
536 }
537
538 /* A connect() to a different interface is not supported. */
539 if (jsk->ifindex != addr->can_ifindex) {
540 ret = -EINVAL;
541 goto out_release_sock;
542 }
543
544 if (!addr->can_addr.j1939.name &&
545 addr->can_addr.j1939.addr == J1939_NO_ADDR &&
546 !sock_flag(&jsk->sk, SOCK_BROADCAST)) {
547 /* broadcast, but SO_BROADCAST not set */
548 ret = -EACCES;
549 goto out_release_sock;
550 }
551
552 jsk->addr.dst_name = addr->can_addr.j1939.name;
553 jsk->addr.da = addr->can_addr.j1939.addr;
554
555 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
556 jsk->addr.pgn = addr->can_addr.j1939.pgn;
557
558 jsk->state |= J1939_SOCK_CONNECTED;
559
560 out_release_sock: /* fall through */
561 release_sock(sock->sk);
562
563 return ret;
564 }
565
j1939_sk_sock2sockaddr_can(struct sockaddr_can * addr,const struct j1939_sock * jsk,int peer)566 static void j1939_sk_sock2sockaddr_can(struct sockaddr_can *addr,
567 const struct j1939_sock *jsk, int peer)
568 {
569 /* There are two holes (2 bytes and 3 bytes) to clear to avoid
570 * leaking kernel information to user space.
571 */
572 memset(addr, 0, J1939_MIN_NAMELEN);
573
574 addr->can_family = AF_CAN;
575 addr->can_ifindex = jsk->ifindex;
576 addr->can_addr.j1939.pgn = jsk->addr.pgn;
577 if (peer) {
578 addr->can_addr.j1939.name = jsk->addr.dst_name;
579 addr->can_addr.j1939.addr = jsk->addr.da;
580 } else {
581 addr->can_addr.j1939.name = jsk->addr.src_name;
582 addr->can_addr.j1939.addr = jsk->addr.sa;
583 }
584 }
585
j1939_sk_getname(struct socket * sock,struct sockaddr * uaddr,int peer)586 static int j1939_sk_getname(struct socket *sock, struct sockaddr *uaddr,
587 int peer)
588 {
589 struct sockaddr_can *addr = (struct sockaddr_can *)uaddr;
590 struct sock *sk = sock->sk;
591 struct j1939_sock *jsk = j1939_sk(sk);
592 int ret = 0;
593
594 lock_sock(sk);
595
596 if (peer && !(jsk->state & J1939_SOCK_CONNECTED)) {
597 ret = -EADDRNOTAVAIL;
598 goto failure;
599 }
600
601 j1939_sk_sock2sockaddr_can(addr, jsk, peer);
602 ret = J1939_MIN_NAMELEN;
603
604 failure:
605 release_sock(sk);
606
607 return ret;
608 }
609
j1939_sk_release(struct socket * sock)610 static int j1939_sk_release(struct socket *sock)
611 {
612 struct sock *sk = sock->sk;
613 struct j1939_sock *jsk;
614
615 if (!sk)
616 return 0;
617
618 lock_sock(sk);
619 jsk = j1939_sk(sk);
620
621 if (jsk->state & J1939_SOCK_BOUND) {
622 struct j1939_priv *priv = jsk->priv;
623
624 if (wait_event_interruptible(jsk->waitq,
625 !j1939_sock_pending_get(&jsk->sk))) {
626 j1939_cancel_active_session(priv, sk);
627 j1939_sk_queue_drop_all(priv, jsk, ESHUTDOWN);
628 }
629
630 j1939_jsk_del(priv, jsk);
631
632 j1939_local_ecu_put(priv, jsk->addr.src_name,
633 jsk->addr.sa);
634
635 j1939_netdev_stop(priv);
636 }
637
638 kfree(jsk->filters);
639 sock_orphan(sk);
640 sock->sk = NULL;
641
642 release_sock(sk);
643 sock_put(sk);
644
645 return 0;
646 }
647
j1939_sk_setsockopt_flag(struct j1939_sock * jsk,sockptr_t optval,unsigned int optlen,int flag)648 static int j1939_sk_setsockopt_flag(struct j1939_sock *jsk, sockptr_t optval,
649 unsigned int optlen, int flag)
650 {
651 int tmp;
652
653 if (optlen != sizeof(tmp))
654 return -EINVAL;
655 if (copy_from_sockptr(&tmp, optval, optlen))
656 return -EFAULT;
657 lock_sock(&jsk->sk);
658 if (tmp)
659 jsk->state |= flag;
660 else
661 jsk->state &= ~flag;
662 release_sock(&jsk->sk);
663 return tmp;
664 }
665
j1939_sk_setsockopt(struct socket * sock,int level,int optname,sockptr_t optval,unsigned int optlen)666 static int j1939_sk_setsockopt(struct socket *sock, int level, int optname,
667 sockptr_t optval, unsigned int optlen)
668 {
669 struct sock *sk = sock->sk;
670 struct j1939_sock *jsk = j1939_sk(sk);
671 int tmp, count = 0, ret = 0;
672 struct j1939_filter *filters = NULL, *ofilters;
673
674 if (level != SOL_CAN_J1939)
675 return -EINVAL;
676
677 switch (optname) {
678 case SO_J1939_FILTER:
679 if (!sockptr_is_null(optval) && optlen != 0) {
680 struct j1939_filter *f;
681 int c;
682
683 if (optlen % sizeof(*filters) != 0)
684 return -EINVAL;
685
686 if (optlen > J1939_FILTER_MAX *
687 sizeof(struct j1939_filter))
688 return -EINVAL;
689
690 count = optlen / sizeof(*filters);
691 filters = memdup_sockptr(optval, optlen);
692 if (IS_ERR(filters))
693 return PTR_ERR(filters);
694
695 for (f = filters, c = count; c; f++, c--) {
696 f->name &= f->name_mask;
697 f->pgn &= f->pgn_mask;
698 f->addr &= f->addr_mask;
699 }
700 }
701
702 lock_sock(&jsk->sk);
703 ofilters = jsk->filters;
704 jsk->filters = filters;
705 jsk->nfilters = count;
706 release_sock(&jsk->sk);
707 kfree(ofilters);
708 return 0;
709 case SO_J1939_PROMISC:
710 return j1939_sk_setsockopt_flag(jsk, optval, optlen,
711 J1939_SOCK_PROMISC);
712 case SO_J1939_ERRQUEUE:
713 ret = j1939_sk_setsockopt_flag(jsk, optval, optlen,
714 J1939_SOCK_ERRQUEUE);
715 if (ret < 0)
716 return ret;
717
718 if (!(jsk->state & J1939_SOCK_ERRQUEUE))
719 skb_queue_purge(&sk->sk_error_queue);
720 return ret;
721 case SO_J1939_SEND_PRIO:
722 if (optlen != sizeof(tmp))
723 return -EINVAL;
724 if (copy_from_sockptr(&tmp, optval, optlen))
725 return -EFAULT;
726 if (tmp < 0 || tmp > 7)
727 return -EDOM;
728 if (tmp < 2 && !capable(CAP_NET_ADMIN))
729 return -EPERM;
730 lock_sock(&jsk->sk);
731 jsk->sk.sk_priority = j1939_to_sk_priority(tmp);
732 release_sock(&jsk->sk);
733 return 0;
734 default:
735 return -ENOPROTOOPT;
736 }
737 }
738
j1939_sk_getsockopt(struct socket * sock,int level,int optname,char __user * optval,int __user * optlen)739 static int j1939_sk_getsockopt(struct socket *sock, int level, int optname,
740 char __user *optval, int __user *optlen)
741 {
742 struct sock *sk = sock->sk;
743 struct j1939_sock *jsk = j1939_sk(sk);
744 int ret, ulen;
745 /* set defaults for using 'int' properties */
746 int tmp = 0;
747 int len = sizeof(tmp);
748 void *val = &tmp;
749
750 if (level != SOL_CAN_J1939)
751 return -EINVAL;
752 if (get_user(ulen, optlen))
753 return -EFAULT;
754 if (ulen < 0)
755 return -EINVAL;
756
757 lock_sock(&jsk->sk);
758 switch (optname) {
759 case SO_J1939_PROMISC:
760 tmp = (jsk->state & J1939_SOCK_PROMISC) ? 1 : 0;
761 break;
762 case SO_J1939_ERRQUEUE:
763 tmp = (jsk->state & J1939_SOCK_ERRQUEUE) ? 1 : 0;
764 break;
765 case SO_J1939_SEND_PRIO:
766 tmp = j1939_prio(jsk->sk.sk_priority);
767 break;
768 default:
769 ret = -ENOPROTOOPT;
770 goto no_copy;
771 }
772
773 /* copy to user, based on 'len' & 'val'
774 * but most sockopt's are 'int' properties, and have 'len' & 'val'
775 * left unchanged, but instead modified 'tmp'
776 */
777 if (len > ulen)
778 ret = -EFAULT;
779 else if (put_user(len, optlen))
780 ret = -EFAULT;
781 else if (copy_to_user(optval, val, len))
782 ret = -EFAULT;
783 else
784 ret = 0;
785 no_copy:
786 release_sock(&jsk->sk);
787 return ret;
788 }
789
j1939_sk_recvmsg(struct socket * sock,struct msghdr * msg,size_t size,int flags)790 static int j1939_sk_recvmsg(struct socket *sock, struct msghdr *msg,
791 size_t size, int flags)
792 {
793 struct sock *sk = sock->sk;
794 struct sk_buff *skb;
795 struct j1939_sk_buff_cb *skcb;
796 int ret = 0;
797
798 if (flags & ~(MSG_DONTWAIT | MSG_ERRQUEUE))
799 return -EINVAL;
800
801 if (flags & MSG_ERRQUEUE)
802 return sock_recv_errqueue(sock->sk, msg, size, SOL_CAN_J1939,
803 SCM_J1939_ERRQUEUE);
804
805 skb = skb_recv_datagram(sk, flags, 0, &ret);
806 if (!skb)
807 return ret;
808
809 if (size < skb->len)
810 msg->msg_flags |= MSG_TRUNC;
811 else
812 size = skb->len;
813
814 ret = memcpy_to_msg(msg, skb->data, size);
815 if (ret < 0) {
816 skb_free_datagram(sk, skb);
817 return ret;
818 }
819
820 skcb = j1939_skb_to_cb(skb);
821 if (j1939_address_is_valid(skcb->addr.da))
822 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_ADDR,
823 sizeof(skcb->addr.da), &skcb->addr.da);
824
825 if (skcb->addr.dst_name)
826 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_DEST_NAME,
827 sizeof(skcb->addr.dst_name), &skcb->addr.dst_name);
828
829 put_cmsg(msg, SOL_CAN_J1939, SCM_J1939_PRIO,
830 sizeof(skcb->priority), &skcb->priority);
831
832 if (msg->msg_name) {
833 struct sockaddr_can *paddr = msg->msg_name;
834
835 msg->msg_namelen = J1939_MIN_NAMELEN;
836 memset(msg->msg_name, 0, msg->msg_namelen);
837 paddr->can_family = AF_CAN;
838 paddr->can_ifindex = skb->skb_iif;
839 paddr->can_addr.j1939.name = skcb->addr.src_name;
840 paddr->can_addr.j1939.addr = skcb->addr.sa;
841 paddr->can_addr.j1939.pgn = skcb->addr.pgn;
842 }
843
844 sock_recv_ts_and_drops(msg, sk, skb);
845 msg->msg_flags |= skcb->msg_flags;
846 skb_free_datagram(sk, skb);
847
848 return size;
849 }
850
j1939_sk_alloc_skb(struct net_device * ndev,struct sock * sk,struct msghdr * msg,size_t size,int * errcode)851 static struct sk_buff *j1939_sk_alloc_skb(struct net_device *ndev,
852 struct sock *sk,
853 struct msghdr *msg, size_t size,
854 int *errcode)
855 {
856 struct j1939_sock *jsk = j1939_sk(sk);
857 struct j1939_sk_buff_cb *skcb;
858 struct sk_buff *skb;
859 int ret;
860
861 skb = sock_alloc_send_skb(sk,
862 size +
863 sizeof(struct can_frame) -
864 sizeof(((struct can_frame *)NULL)->data) +
865 sizeof(struct can_skb_priv),
866 msg->msg_flags & MSG_DONTWAIT, &ret);
867 if (!skb)
868 goto failure;
869
870 can_skb_reserve(skb);
871 can_skb_prv(skb)->ifindex = ndev->ifindex;
872 can_skb_prv(skb)->skbcnt = 0;
873 skb_reserve(skb, offsetof(struct can_frame, data));
874
875 ret = memcpy_from_msg(skb_put(skb, size), msg, size);
876 if (ret < 0)
877 goto free_skb;
878
879 skb->dev = ndev;
880
881 skcb = j1939_skb_to_cb(skb);
882 memset(skcb, 0, sizeof(*skcb));
883 skcb->addr = jsk->addr;
884 skcb->priority = j1939_prio(sk->sk_priority);
885
886 if (msg->msg_name) {
887 struct sockaddr_can *addr = msg->msg_name;
888
889 if (addr->can_addr.j1939.name ||
890 addr->can_addr.j1939.addr != J1939_NO_ADDR) {
891 skcb->addr.dst_name = addr->can_addr.j1939.name;
892 skcb->addr.da = addr->can_addr.j1939.addr;
893 }
894 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn))
895 skcb->addr.pgn = addr->can_addr.j1939.pgn;
896 }
897
898 *errcode = ret;
899 return skb;
900
901 free_skb:
902 kfree_skb(skb);
903 failure:
904 *errcode = ret;
905 return NULL;
906 }
907
j1939_sk_opt_stats_get_size(enum j1939_sk_errqueue_type type)908 static size_t j1939_sk_opt_stats_get_size(enum j1939_sk_errqueue_type type)
909 {
910 switch (type) {
911 case J1939_ERRQUEUE_RX_RTS:
912 return
913 nla_total_size(sizeof(u32)) + /* J1939_NLA_TOTAL_SIZE */
914 nla_total_size(sizeof(u32)) + /* J1939_NLA_PGN */
915 nla_total_size(sizeof(u64)) + /* J1939_NLA_SRC_NAME */
916 nla_total_size(sizeof(u64)) + /* J1939_NLA_DEST_NAME */
917 nla_total_size(sizeof(u8)) + /* J1939_NLA_SRC_ADDR */
918 nla_total_size(sizeof(u8)) + /* J1939_NLA_DEST_ADDR */
919 0;
920 default:
921 return
922 nla_total_size(sizeof(u32)) + /* J1939_NLA_BYTES_ACKED */
923 0;
924 }
925 }
926
927 static struct sk_buff *
j1939_sk_get_timestamping_opt_stats(struct j1939_session * session,enum j1939_sk_errqueue_type type)928 j1939_sk_get_timestamping_opt_stats(struct j1939_session *session,
929 enum j1939_sk_errqueue_type type)
930 {
931 struct sk_buff *stats;
932 u32 size;
933
934 stats = alloc_skb(j1939_sk_opt_stats_get_size(type), GFP_ATOMIC);
935 if (!stats)
936 return NULL;
937
938 if (session->skcb.addr.type == J1939_SIMPLE)
939 size = session->total_message_size;
940 else
941 size = min(session->pkt.tx_acked * 7,
942 session->total_message_size);
943
944 switch (type) {
945 case J1939_ERRQUEUE_RX_RTS:
946 nla_put_u32(stats, J1939_NLA_TOTAL_SIZE,
947 session->total_message_size);
948 nla_put_u32(stats, J1939_NLA_PGN,
949 session->skcb.addr.pgn);
950 nla_put_u64_64bit(stats, J1939_NLA_SRC_NAME,
951 session->skcb.addr.src_name, J1939_NLA_PAD);
952 nla_put_u64_64bit(stats, J1939_NLA_DEST_NAME,
953 session->skcb.addr.dst_name, J1939_NLA_PAD);
954 nla_put_u8(stats, J1939_NLA_SRC_ADDR,
955 session->skcb.addr.sa);
956 nla_put_u8(stats, J1939_NLA_DEST_ADDR,
957 session->skcb.addr.da);
958 break;
959 default:
960 nla_put_u32(stats, J1939_NLA_BYTES_ACKED, size);
961 }
962
963 return stats;
964 }
965
__j1939_sk_errqueue(struct j1939_session * session,struct sock * sk,enum j1939_sk_errqueue_type type)966 static void __j1939_sk_errqueue(struct j1939_session *session, struct sock *sk,
967 enum j1939_sk_errqueue_type type)
968 {
969 struct j1939_priv *priv = session->priv;
970 struct j1939_sock *jsk;
971 struct sock_exterr_skb *serr;
972 struct sk_buff *skb;
973 char *state = "UNK";
974 int err;
975
976 jsk = j1939_sk(sk);
977
978 if (!(jsk->state & J1939_SOCK_ERRQUEUE))
979 return;
980
981 switch (type) {
982 case J1939_ERRQUEUE_TX_ACK:
983 if (!(sk->sk_tsflags & SOF_TIMESTAMPING_TX_ACK))
984 return;
985 break;
986 case J1939_ERRQUEUE_TX_SCHED:
987 if (!(sk->sk_tsflags & SOF_TIMESTAMPING_TX_SCHED))
988 return;
989 break;
990 case J1939_ERRQUEUE_TX_ABORT:
991 break;
992 case J1939_ERRQUEUE_RX_RTS:
993 fallthrough;
994 case J1939_ERRQUEUE_RX_DPO:
995 fallthrough;
996 case J1939_ERRQUEUE_RX_ABORT:
997 if (!(sk->sk_tsflags & SOF_TIMESTAMPING_RX_SOFTWARE))
998 return;
999 break;
1000 default:
1001 netdev_err(priv->ndev, "Unknown errqueue type %i\n", type);
1002 }
1003
1004 skb = j1939_sk_get_timestamping_opt_stats(session, type);
1005 if (!skb)
1006 return;
1007
1008 skb->tstamp = ktime_get_real();
1009
1010 BUILD_BUG_ON(sizeof(struct sock_exterr_skb) > sizeof(skb->cb));
1011
1012 serr = SKB_EXT_ERR(skb);
1013 memset(serr, 0, sizeof(*serr));
1014 switch (type) {
1015 case J1939_ERRQUEUE_TX_ACK:
1016 serr->ee.ee_errno = ENOMSG;
1017 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING;
1018 serr->ee.ee_info = SCM_TSTAMP_ACK;
1019 state = "TX ACK";
1020 break;
1021 case J1939_ERRQUEUE_TX_SCHED:
1022 serr->ee.ee_errno = ENOMSG;
1023 serr->ee.ee_origin = SO_EE_ORIGIN_TIMESTAMPING;
1024 serr->ee.ee_info = SCM_TSTAMP_SCHED;
1025 state = "TX SCH";
1026 break;
1027 case J1939_ERRQUEUE_TX_ABORT:
1028 serr->ee.ee_errno = session->err;
1029 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1030 serr->ee.ee_info = J1939_EE_INFO_TX_ABORT;
1031 state = "TX ABT";
1032 break;
1033 case J1939_ERRQUEUE_RX_RTS:
1034 serr->ee.ee_errno = ENOMSG;
1035 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1036 serr->ee.ee_info = J1939_EE_INFO_RX_RTS;
1037 state = "RX RTS";
1038 break;
1039 case J1939_ERRQUEUE_RX_DPO:
1040 serr->ee.ee_errno = ENOMSG;
1041 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1042 serr->ee.ee_info = J1939_EE_INFO_RX_DPO;
1043 state = "RX DPO";
1044 break;
1045 case J1939_ERRQUEUE_RX_ABORT:
1046 serr->ee.ee_errno = session->err;
1047 serr->ee.ee_origin = SO_EE_ORIGIN_LOCAL;
1048 serr->ee.ee_info = J1939_EE_INFO_RX_ABORT;
1049 state = "RX ABT";
1050 break;
1051 }
1052
1053 serr->opt_stats = true;
1054 if (sk->sk_tsflags & SOF_TIMESTAMPING_OPT_ID)
1055 serr->ee.ee_data = session->tskey;
1056
1057 netdev_dbg(session->priv->ndev, "%s: 0x%p tskey: %i, state: %s\n",
1058 __func__, session, session->tskey, state);
1059 err = sock_queue_err_skb(sk, skb);
1060
1061 if (err)
1062 kfree_skb(skb);
1063 };
1064
j1939_sk_errqueue(struct j1939_session * session,enum j1939_sk_errqueue_type type)1065 void j1939_sk_errqueue(struct j1939_session *session,
1066 enum j1939_sk_errqueue_type type)
1067 {
1068 struct j1939_priv *priv = session->priv;
1069 struct j1939_sock *jsk;
1070
1071 if (session->sk) {
1072 /* send TX notifications to the socket of origin */
1073 __j1939_sk_errqueue(session, session->sk, type);
1074 return;
1075 }
1076
1077 /* spread RX notifications to all sockets subscribed to this session */
1078 spin_lock_bh(&priv->j1939_socks_lock);
1079 list_for_each_entry(jsk, &priv->j1939_socks, list) {
1080 if (j1939_sk_recv_match_one(jsk, &session->skcb))
1081 __j1939_sk_errqueue(session, &jsk->sk, type);
1082 }
1083 spin_unlock_bh(&priv->j1939_socks_lock);
1084 };
1085
j1939_sk_send_loop_abort(struct sock * sk,int err)1086 void j1939_sk_send_loop_abort(struct sock *sk, int err)
1087 {
1088 sk->sk_err = err;
1089
1090 sk_error_report(sk);
1091 }
1092
j1939_sk_send_loop(struct j1939_priv * priv,struct sock * sk,struct msghdr * msg,size_t size)1093 static int j1939_sk_send_loop(struct j1939_priv *priv, struct sock *sk,
1094 struct msghdr *msg, size_t size)
1095
1096 {
1097 struct j1939_sock *jsk = j1939_sk(sk);
1098 struct j1939_session *session = j1939_sk_get_incomplete_session(jsk);
1099 struct sk_buff *skb;
1100 size_t segment_size, todo_size;
1101 int ret = 0;
1102
1103 if (session &&
1104 session->total_message_size != session->total_queued_size + size) {
1105 j1939_session_put(session);
1106 return -EIO;
1107 }
1108
1109 todo_size = size;
1110
1111 while (todo_size) {
1112 struct j1939_sk_buff_cb *skcb;
1113
1114 segment_size = min_t(size_t, J1939_MAX_TP_PACKET_SIZE,
1115 todo_size);
1116
1117 /* Allocate skb for one segment */
1118 skb = j1939_sk_alloc_skb(priv->ndev, sk, msg, segment_size,
1119 &ret);
1120 if (ret)
1121 break;
1122
1123 skcb = j1939_skb_to_cb(skb);
1124
1125 if (!session) {
1126 /* at this point the size should be full size
1127 * of the session
1128 */
1129 skcb->offset = 0;
1130 session = j1939_tp_send(priv, skb, size);
1131 if (IS_ERR(session)) {
1132 ret = PTR_ERR(session);
1133 goto kfree_skb;
1134 }
1135 if (j1939_sk_queue_session(session)) {
1136 /* try to activate session if we a
1137 * fist in the queue
1138 */
1139 if (!j1939_session_activate(session)) {
1140 j1939_tp_schedule_txtimer(session, 0);
1141 } else {
1142 ret = -EBUSY;
1143 session->err = ret;
1144 j1939_sk_queue_drop_all(priv, jsk,
1145 EBUSY);
1146 break;
1147 }
1148 }
1149 } else {
1150 skcb->offset = session->total_queued_size;
1151 j1939_session_skb_queue(session, skb);
1152 }
1153
1154 todo_size -= segment_size;
1155 session->total_queued_size += segment_size;
1156 }
1157
1158 switch (ret) {
1159 case 0: /* OK */
1160 if (todo_size)
1161 netdev_warn(priv->ndev,
1162 "no error found and not completely queued?! %zu\n",
1163 todo_size);
1164 ret = size;
1165 break;
1166 case -ERESTARTSYS:
1167 ret = -EINTR;
1168 fallthrough;
1169 case -EAGAIN: /* OK */
1170 if (todo_size != size)
1171 ret = size - todo_size;
1172 break;
1173 default: /* ERROR */
1174 break;
1175 }
1176
1177 if (session)
1178 j1939_session_put(session);
1179
1180 return ret;
1181
1182 kfree_skb:
1183 kfree_skb(skb);
1184 return ret;
1185 }
1186
j1939_sk_sendmsg(struct socket * sock,struct msghdr * msg,size_t size)1187 static int j1939_sk_sendmsg(struct socket *sock, struct msghdr *msg,
1188 size_t size)
1189 {
1190 struct sock *sk = sock->sk;
1191 struct j1939_sock *jsk = j1939_sk(sk);
1192 struct j1939_priv *priv;
1193 int ifindex;
1194 int ret;
1195
1196 lock_sock(sock->sk);
1197 /* various socket state tests */
1198 if (!(jsk->state & J1939_SOCK_BOUND)) {
1199 ret = -EBADFD;
1200 goto sendmsg_done;
1201 }
1202
1203 priv = jsk->priv;
1204 ifindex = jsk->ifindex;
1205
1206 if (!jsk->addr.src_name && jsk->addr.sa == J1939_NO_ADDR) {
1207 /* no source address assigned yet */
1208 ret = -EBADFD;
1209 goto sendmsg_done;
1210 }
1211
1212 /* deal with provided destination address info */
1213 if (msg->msg_name) {
1214 struct sockaddr_can *addr = msg->msg_name;
1215
1216 if (msg->msg_namelen < J1939_MIN_NAMELEN) {
1217 ret = -EINVAL;
1218 goto sendmsg_done;
1219 }
1220
1221 if (addr->can_family != AF_CAN) {
1222 ret = -EINVAL;
1223 goto sendmsg_done;
1224 }
1225
1226 if (addr->can_ifindex && addr->can_ifindex != ifindex) {
1227 ret = -EBADFD;
1228 goto sendmsg_done;
1229 }
1230
1231 if (j1939_pgn_is_valid(addr->can_addr.j1939.pgn) &&
1232 !j1939_pgn_is_clean_pdu(addr->can_addr.j1939.pgn)) {
1233 ret = -EINVAL;
1234 goto sendmsg_done;
1235 }
1236
1237 if (!addr->can_addr.j1939.name &&
1238 addr->can_addr.j1939.addr == J1939_NO_ADDR &&
1239 !sock_flag(sk, SOCK_BROADCAST)) {
1240 /* broadcast, but SO_BROADCAST not set */
1241 ret = -EACCES;
1242 goto sendmsg_done;
1243 }
1244 } else {
1245 if (!jsk->addr.dst_name && jsk->addr.da == J1939_NO_ADDR &&
1246 !sock_flag(sk, SOCK_BROADCAST)) {
1247 /* broadcast, but SO_BROADCAST not set */
1248 ret = -EACCES;
1249 goto sendmsg_done;
1250 }
1251 }
1252
1253 ret = j1939_sk_send_loop(priv, sk, msg, size);
1254
1255 sendmsg_done:
1256 release_sock(sock->sk);
1257
1258 return ret;
1259 }
1260
j1939_sk_netdev_event_netdown(struct j1939_priv * priv)1261 void j1939_sk_netdev_event_netdown(struct j1939_priv *priv)
1262 {
1263 struct j1939_sock *jsk;
1264 int error_code = ENETDOWN;
1265
1266 spin_lock_bh(&priv->j1939_socks_lock);
1267 list_for_each_entry(jsk, &priv->j1939_socks, list) {
1268 jsk->sk.sk_err = error_code;
1269 if (!sock_flag(&jsk->sk, SOCK_DEAD))
1270 sk_error_report(&jsk->sk);
1271
1272 j1939_sk_queue_drop_all(priv, jsk, error_code);
1273 }
1274 spin_unlock_bh(&priv->j1939_socks_lock);
1275 }
1276
j1939_sk_no_ioctlcmd(struct socket * sock,unsigned int cmd,unsigned long arg)1277 static int j1939_sk_no_ioctlcmd(struct socket *sock, unsigned int cmd,
1278 unsigned long arg)
1279 {
1280 /* no ioctls for socket layer -> hand it down to NIC layer */
1281 return -ENOIOCTLCMD;
1282 }
1283
1284 static const struct proto_ops j1939_ops = {
1285 .family = PF_CAN,
1286 .release = j1939_sk_release,
1287 .bind = j1939_sk_bind,
1288 .connect = j1939_sk_connect,
1289 .socketpair = sock_no_socketpair,
1290 .accept = sock_no_accept,
1291 .getname = j1939_sk_getname,
1292 .poll = datagram_poll,
1293 .ioctl = j1939_sk_no_ioctlcmd,
1294 .listen = sock_no_listen,
1295 .shutdown = sock_no_shutdown,
1296 .setsockopt = j1939_sk_setsockopt,
1297 .getsockopt = j1939_sk_getsockopt,
1298 .sendmsg = j1939_sk_sendmsg,
1299 .recvmsg = j1939_sk_recvmsg,
1300 .mmap = sock_no_mmap,
1301 .sendpage = sock_no_sendpage,
1302 };
1303
1304 static struct proto j1939_proto __read_mostly = {
1305 .name = "CAN_J1939",
1306 .owner = THIS_MODULE,
1307 .obj_size = sizeof(struct j1939_sock),
1308 .init = j1939_sk_init,
1309 };
1310
1311 const struct can_proto j1939_can_proto = {
1312 .type = SOCK_DGRAM,
1313 .protocol = CAN_J1939,
1314 .ops = &j1939_ops,
1315 .prot = &j1939_proto,
1316 };
1317