| // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB |
| /* |
| * Copyright (c) 2020 Mellanox Technologies. All rights reserved. |
| */ |
| |
| #include <rdma/ib_verbs.h> |
| #include <rdma/ib_cache.h> |
| #include <rdma/lag.h> |
| |
| static struct sk_buff *rdma_build_skb(struct net_device *netdev, |
| struct rdma_ah_attr *ah_attr, |
| gfp_t flags) |
| { |
| struct ipv6hdr *ip6h; |
| struct sk_buff *skb; |
| struct ethhdr *eth; |
| struct iphdr *iph; |
| struct udphdr *uh; |
| u8 smac[ETH_ALEN]; |
| bool is_ipv4; |
| int hdr_len; |
| |
| is_ipv4 = ipv6_addr_v4mapped((struct in6_addr *)ah_attr->grh.dgid.raw); |
| hdr_len = ETH_HLEN + sizeof(struct udphdr) + LL_RESERVED_SPACE(netdev); |
| hdr_len += is_ipv4 ? sizeof(struct iphdr) : sizeof(struct ipv6hdr); |
| |
| skb = alloc_skb(hdr_len, flags); |
| if (!skb) |
| return NULL; |
| |
| skb->dev = netdev; |
| skb_reserve(skb, hdr_len); |
| skb_push(skb, sizeof(struct udphdr)); |
| skb_reset_transport_header(skb); |
| uh = udp_hdr(skb); |
| uh->source = |
| htons(rdma_flow_label_to_udp_sport(ah_attr->grh.flow_label)); |
| uh->dest = htons(ROCE_V2_UDP_DPORT); |
| uh->len = htons(sizeof(struct udphdr)); |
| |
| if (is_ipv4) { |
| skb_push(skb, sizeof(struct iphdr)); |
| skb_reset_network_header(skb); |
| iph = ip_hdr(skb); |
| iph->frag_off = 0; |
| iph->version = 4; |
| iph->protocol = IPPROTO_UDP; |
| iph->ihl = 0x5; |
| iph->tot_len = htons(sizeof(struct udphdr) + sizeof(struct |
| iphdr)); |
| memcpy(&iph->saddr, ah_attr->grh.sgid_attr->gid.raw + 12, |
| sizeof(struct in_addr)); |
| memcpy(&iph->daddr, ah_attr->grh.dgid.raw + 12, |
| sizeof(struct in_addr)); |
| } else { |
| skb_push(skb, sizeof(struct ipv6hdr)); |
| skb_reset_network_header(skb); |
| ip6h = ipv6_hdr(skb); |
| ip6h->version = 6; |
| ip6h->nexthdr = IPPROTO_UDP; |
| memcpy(&ip6h->flow_lbl, &ah_attr->grh.flow_label, |
| sizeof(*ip6h->flow_lbl)); |
| memcpy(&ip6h->saddr, ah_attr->grh.sgid_attr->gid.raw, |
| sizeof(struct in6_addr)); |
| memcpy(&ip6h->daddr, ah_attr->grh.dgid.raw, |
| sizeof(struct in6_addr)); |
| } |
| |
| skb_push(skb, sizeof(struct ethhdr)); |
| skb_reset_mac_header(skb); |
| eth = eth_hdr(skb); |
| skb->protocol = eth->h_proto = htons(is_ipv4 ? ETH_P_IP : ETH_P_IPV6); |
| rdma_read_gid_l2_fields(ah_attr->grh.sgid_attr, NULL, smac); |
| memcpy(eth->h_source, smac, ETH_ALEN); |
| memcpy(eth->h_dest, ah_attr->roce.dmac, ETH_ALEN); |
| |
| return skb; |
| } |
| |
| static struct net_device *rdma_get_xmit_slave_udp(struct ib_device *device, |
| struct net_device *master, |
| struct rdma_ah_attr *ah_attr, |
| gfp_t flags) |
| { |
| struct net_device *slave; |
| struct sk_buff *skb; |
| |
| skb = rdma_build_skb(master, ah_attr, flags); |
| if (!skb) |
| return ERR_PTR(-ENOMEM); |
| |
| rcu_read_lock(); |
| slave = netdev_get_xmit_slave(master, skb, |
| !!(device->lag_flags & |
| RDMA_LAG_FLAGS_HASH_ALL_SLAVES)); |
| if (slave) |
| dev_hold(slave); |
| rcu_read_unlock(); |
| kfree_skb(skb); |
| return slave; |
| } |
| |
| void rdma_lag_put_ah_roce_slave(struct net_device *xmit_slave) |
| { |
| if (xmit_slave) |
| dev_put(xmit_slave); |
| } |
| |
| struct net_device *rdma_lag_get_ah_roce_slave(struct ib_device *device, |
| struct rdma_ah_attr *ah_attr, |
| gfp_t flags) |
| { |
| struct net_device *slave = NULL; |
| struct net_device *master; |
| |
| if (!(ah_attr->type == RDMA_AH_ATTR_TYPE_ROCE && |
| ah_attr->grh.sgid_attr->gid_type == IB_GID_TYPE_ROCE_UDP_ENCAP && |
| ah_attr->grh.flow_label)) |
| return NULL; |
| |
| rcu_read_lock(); |
| master = rdma_read_gid_attr_ndev_rcu(ah_attr->grh.sgid_attr); |
| if (IS_ERR(master)) { |
| rcu_read_unlock(); |
| return master; |
| } |
| dev_hold(master); |
| rcu_read_unlock(); |
| |
| if (!netif_is_bond_master(master)) |
| goto put; |
| |
| slave = rdma_get_xmit_slave_udp(device, master, ah_attr, flags); |
| put: |
| dev_put(master); |
| return slave; |
| } |