/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2010-2016 Intel Corporation */ #ifndef __L3FWD_EM_H__ #define __L3FWD_EM_H__ static __rte_always_inline uint16_t l3fwd_em_handle_ipv4(struct rte_mbuf *m, uint16_t portid, struct rte_ether_hdr *eth_hdr, struct lcore_conf *qconf) { struct rte_ipv4_hdr *ipv4_hdr; uint16_t dst_port; /* Handle IPv4 headers.*/ ipv4_hdr = rte_pktmbuf_mtod_offset(m, struct rte_ipv4_hdr *, sizeof(struct rte_ether_hdr)); #ifdef DO_RFC_1812_CHECKS /* Check to make sure the packet is valid (RFC1812) */ if (is_valid_ipv4_pkt(ipv4_hdr, m->pkt_len) < 0) { rte_pktmbuf_free(m); return BAD_PORT; } #endif dst_port = em_get_ipv4_dst_port(ipv4_hdr, portid, qconf->ipv4_lookup_struct); if (dst_port >= RTE_MAX_ETHPORTS || (enabled_port_mask & 1 << dst_port) == 0) dst_port = portid; #ifdef DO_RFC_1812_CHECKS /* Update time to live and header checksum */ --(ipv4_hdr->time_to_live); ++(ipv4_hdr->hdr_checksum); #endif /* dst addr */ *(uint64_t *)ð_hdr->d_addr = dest_eth_addr[dst_port]; /* src addr */ rte_ether_addr_copy(&ports_eth_addr[dst_port], ð_hdr->s_addr); return dst_port; } static __rte_always_inline uint16_t l3fwd_em_handle_ipv6(struct rte_mbuf *m, uint16_t portid, struct rte_ether_hdr *eth_hdr, struct lcore_conf *qconf) { /* Handle IPv6 headers.*/ struct rte_ipv6_hdr *ipv6_hdr; uint16_t dst_port; ipv6_hdr = rte_pktmbuf_mtod_offset(m, struct rte_ipv6_hdr *, sizeof(struct rte_ether_hdr)); dst_port = em_get_ipv6_dst_port(ipv6_hdr, portid, qconf->ipv6_lookup_struct); if (dst_port >= RTE_MAX_ETHPORTS || (enabled_port_mask & 1 << dst_port) == 0) dst_port = portid; /* dst addr */ *(uint64_t *)ð_hdr->d_addr = dest_eth_addr[dst_port]; /* src addr */ rte_ether_addr_copy(&ports_eth_addr[dst_port], ð_hdr->s_addr); return dst_port; } static __rte_always_inline void l3fwd_em_simple_forward(struct rte_mbuf *m, uint16_t portid, struct lcore_conf *qconf) { struct rte_ether_hdr *eth_hdr; uint16_t dst_port; uint32_t tcp_or_udp; uint32_t l3_ptypes; eth_hdr = rte_pktmbuf_mtod(m, struct rte_ether_hdr *); tcp_or_udp = m->packet_type & (RTE_PTYPE_L4_TCP | RTE_PTYPE_L4_UDP); l3_ptypes = m->packet_type & RTE_PTYPE_L3_MASK; if (tcp_or_udp && (l3_ptypes == RTE_PTYPE_L3_IPV4)) { dst_port = l3fwd_em_handle_ipv4(m, portid, eth_hdr, qconf); send_single_packet(qconf, m, dst_port); } else if (tcp_or_udp && (l3_ptypes == RTE_PTYPE_L3_IPV6)) { dst_port = l3fwd_em_handle_ipv6(m, portid, eth_hdr, qconf); send_single_packet(qconf, m, dst_port); } else { /* Free the mbuf that contains non-IPV4/IPV6 packet */ rte_pktmbuf_free(m); } } static __rte_always_inline void l3fwd_em_simple_process(struct rte_mbuf *m, struct lcore_conf *qconf) { struct rte_ether_hdr *eth_hdr; uint32_t tcp_or_udp; uint32_t l3_ptypes; eth_hdr = rte_pktmbuf_mtod(m, struct rte_ether_hdr *); tcp_or_udp = m->packet_type & (RTE_PTYPE_L4_TCP | RTE_PTYPE_L4_UDP); l3_ptypes = m->packet_type & RTE_PTYPE_L3_MASK; if (tcp_or_udp && (l3_ptypes == RTE_PTYPE_L3_IPV4)) m->port = l3fwd_em_handle_ipv4(m, m->port, eth_hdr, qconf); else if (tcp_or_udp && (l3_ptypes == RTE_PTYPE_L3_IPV6)) m->port = l3fwd_em_handle_ipv6(m, m->port, eth_hdr, qconf); else m->port = BAD_PORT; } /* * Buffer non-optimized handling of packets, invoked * from main_loop. */ static inline void l3fwd_em_no_opt_send_packets(int nb_rx, struct rte_mbuf **pkts_burst, uint16_t portid, struct lcore_conf *qconf) { int32_t j; /* Prefetch first packets */ for (j = 0; j < PREFETCH_OFFSET && j < nb_rx; j++) rte_prefetch0(rte_pktmbuf_mtod(pkts_burst[j], void *)); /* * Prefetch and forward already prefetched * packets. */ for (j = 0; j < (nb_rx - PREFETCH_OFFSET); j++) { rte_prefetch0(rte_pktmbuf_mtod(pkts_burst[ j + PREFETCH_OFFSET], void *)); l3fwd_em_simple_forward(pkts_burst[j], portid, qconf); } /* Forward remaining prefetched packets */ for (; j < nb_rx; j++) l3fwd_em_simple_forward(pkts_burst[j], portid, qconf); } /* * Buffer non-optimized handling of events, invoked * from main_loop. */ static inline void l3fwd_em_no_opt_process_events(int nb_rx, struct rte_event **events, struct lcore_conf *qconf) { int32_t j; /* Prefetch first packets */ for (j = 0; j < PREFETCH_OFFSET && j < nb_rx; j++) rte_prefetch0(rte_pktmbuf_mtod(events[j]->mbuf, void *)); /* * Prefetch and forward already prefetched * packets. */ for (j = 0; j < (nb_rx - PREFETCH_OFFSET); j++) { rte_prefetch0(rte_pktmbuf_mtod(events[ j + PREFETCH_OFFSET]->mbuf, void *)); l3fwd_em_simple_process(events[j]->mbuf, qconf); } /* Forward remaining prefetched packets */ for (; j < nb_rx; j++) l3fwd_em_simple_process(events[j]->mbuf, qconf); } #endif /* __L3FWD_EM_H__ */