From 638a9e433ecd61e64761352dbec1fa4f5874c941 Mon Sep 17 00:00:00 2001 From: Daniel Baumann Date: Wed, 7 Aug 2024 15:18:06 +0200 Subject: Merging upstream version 6.10.3. Signed-off-by: Daniel Baumann --- drivers/net/team/Makefile | 1 + drivers/net/team/team.c | 3101 ------------------------------------------ drivers/net/team/team_core.c | 3056 +++++++++++++++++++++++++++++++++++++++++ drivers/net/team/team_nl.c | 59 + drivers/net/team/team_nl.h | 29 + 5 files changed, 3145 insertions(+), 3101 deletions(-) delete mode 100644 drivers/net/team/team.c create mode 100644 drivers/net/team/team_core.c create mode 100644 drivers/net/team/team_nl.c create mode 100644 drivers/net/team/team_nl.h (limited to 'drivers/net/team') diff --git a/drivers/net/team/Makefile b/drivers/net/team/Makefile index f582d81a50..7a5aa20d28 100644 --- a/drivers/net/team/Makefile +++ b/drivers/net/team/Makefile @@ -3,6 +3,7 @@ # Makefile for the network team driver # +team-y:= team_core.o team_nl.o obj-$(CONFIG_NET_TEAM) += team.o obj-$(CONFIG_NET_TEAM_MODE_BROADCAST) += team_mode_broadcast.o obj-$(CONFIG_NET_TEAM_MODE_ROUNDROBIN) += team_mode_roundrobin.o diff --git a/drivers/net/team/team.c b/drivers/net/team/team.c deleted file mode 100644 index 0a44bbdcfb..0000000000 --- a/drivers/net/team/team.c +++ /dev/null @@ -1,3101 +0,0 @@ -// SPDX-License-Identifier: GPL-2.0-or-later -/* - * drivers/net/team/team.c - Network team device driver - * Copyright (c) 2011 Jiri Pirko - */ - -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include -#include - -#define DRV_NAME "team" - - -/********** - * Helpers - **********/ - -static struct team_port *team_port_get_rtnl(const struct net_device *dev) -{ - struct team_port *port = rtnl_dereference(dev->rx_handler_data); - - return netif_is_team_port(dev) ? port : NULL; -} - -/* - * Since the ability to change device address for open port device is tested in - * team_port_add, this function can be called without control of return value - */ -static int __set_port_dev_addr(struct net_device *port_dev, - const unsigned char *dev_addr) -{ - struct sockaddr_storage addr; - - memcpy(addr.__data, dev_addr, port_dev->addr_len); - addr.ss_family = port_dev->type; - return dev_set_mac_address(port_dev, (struct sockaddr *)&addr, NULL); -} - -static int team_port_set_orig_dev_addr(struct team_port *port) -{ - return __set_port_dev_addr(port->dev, port->orig.dev_addr); -} - -static int team_port_set_team_dev_addr(struct team *team, - struct team_port *port) -{ - return __set_port_dev_addr(port->dev, team->dev->dev_addr); -} - -int team_modeop_port_enter(struct team *team, struct team_port *port) -{ - return team_port_set_team_dev_addr(team, port); -} -EXPORT_SYMBOL(team_modeop_port_enter); - -void team_modeop_port_change_dev_addr(struct team *team, - struct team_port *port) -{ - team_port_set_team_dev_addr(team, port); -} -EXPORT_SYMBOL(team_modeop_port_change_dev_addr); - -static void team_lower_state_changed(struct team_port *port) -{ - struct netdev_lag_lower_state_info info; - - info.link_up = port->linkup; - info.tx_enabled = team_port_enabled(port); - netdev_lower_state_changed(port->dev, &info); -} - -static void team_refresh_port_linkup(struct team_port *port) -{ - bool new_linkup = port->user.linkup_enabled ? port->user.linkup : - port->state.linkup; - - if (port->linkup != new_linkup) { - port->linkup = new_linkup; - team_lower_state_changed(port); - } -} - - -/******************* - * Options handling - *******************/ - -struct team_option_inst { /* One for each option instance */ - struct list_head list; - struct list_head tmp_list; - struct team_option *option; - struct team_option_inst_info info; - bool changed; - bool removed; -}; - -static struct team_option *__team_find_option(struct team *team, - const char *opt_name) -{ - struct team_option *option; - - list_for_each_entry(option, &team->option_list, list) { - if (strcmp(option->name, opt_name) == 0) - return option; - } - return NULL; -} - -static void __team_option_inst_del(struct team_option_inst *opt_inst) -{ - list_del(&opt_inst->list); - kfree(opt_inst); -} - -static void __team_option_inst_del_option(struct team *team, - struct team_option *option) -{ - struct team_option_inst *opt_inst, *tmp; - - list_for_each_entry_safe(opt_inst, tmp, &team->option_inst_list, list) { - if (opt_inst->option == option) - __team_option_inst_del(opt_inst); - } -} - -static int __team_option_inst_add(struct team *team, struct team_option *option, - struct team_port *port) -{ - struct team_option_inst *opt_inst; - unsigned int array_size; - unsigned int i; - - array_size = option->array_size; - if (!array_size) - array_size = 1; /* No array but still need one instance */ - - for (i = 0; i < array_size; i++) { - opt_inst = kmalloc(sizeof(*opt_inst), GFP_KERNEL); - if (!opt_inst) - return -ENOMEM; - opt_inst->option = option; - opt_inst->info.port = port; - opt_inst->info.array_index = i; - opt_inst->changed = true; - opt_inst->removed = false; - list_add_tail(&opt_inst->list, &team->option_inst_list); - if (option->init) - option->init(team, &opt_inst->info); - - } - return 0; -} - -static int __team_option_inst_add_option(struct team *team, - struct team_option *option) -{ - int err; - - if (!option->per_port) { - err = __team_option_inst_add(team, option, NULL); - if (err) - goto inst_del_option; - } - return 0; - -inst_del_option: - __team_option_inst_del_option(team, option); - return err; -} - -static void __team_option_inst_mark_removed_option(struct team *team, - struct team_option *option) -{ - struct team_option_inst *opt_inst; - - list_for_each_entry(opt_inst, &team->option_inst_list, list) { - if (opt_inst->option == option) { - opt_inst->changed = true; - opt_inst->removed = true; - } - } -} - -static void __team_option_inst_del_port(struct team *team, - struct team_port *port) -{ - struct team_option_inst *opt_inst, *tmp; - - list_for_each_entry_safe(opt_inst, tmp, &team->option_inst_list, list) { - if (opt_inst->option->per_port && - opt_inst->info.port == port) - __team_option_inst_del(opt_inst); - } -} - -static int __team_option_inst_add_port(struct team *team, - struct team_port *port) -{ - struct team_option *option; - int err; - - list_for_each_entry(option, &team->option_list, list) { - if (!option->per_port) - continue; - err = __team_option_inst_add(team, option, port); - if (err) - goto inst_del_port; - } - return 0; - -inst_del_port: - __team_option_inst_del_port(team, port); - return err; -} - -static void __team_option_inst_mark_removed_port(struct team *team, - struct team_port *port) -{ - struct team_option_inst *opt_inst; - - list_for_each_entry(opt_inst, &team->option_inst_list, list) { - if (opt_inst->info.port == port) { - opt_inst->changed = true; - opt_inst->removed = true; - } - } -} - -static int __team_options_register(struct team *team, - const struct team_option *option, - size_t option_count) -{ - int i; - struct team_option **dst_opts; - int err; - - dst_opts = kcalloc(option_count, sizeof(struct team_option *), - GFP_KERNEL); - if (!dst_opts) - return -ENOMEM; - for (i = 0; i < option_count; i++, option++) { - if (__team_find_option(team, option->name)) { - err = -EEXIST; - goto alloc_rollback; - } - dst_opts[i] = kmemdup(option, sizeof(*option), GFP_KERNEL); - if (!dst_opts[i]) { - err = -ENOMEM; - goto alloc_rollback; - } - } - - for (i = 0; i < option_count; i++) { - err = __team_option_inst_add_option(team, dst_opts[i]); - if (err) - goto inst_rollback; - list_add_tail(&dst_opts[i]->list, &team->option_list); - } - - kfree(dst_opts); - return 0; - -inst_rollback: - for (i--; i >= 0; i--) { - __team_option_inst_del_option(team, dst_opts[i]); - list_del(&dst_opts[i]->list); - } - - i = option_count; -alloc_rollback: - for (i--; i >= 0; i--) - kfree(dst_opts[i]); - - kfree(dst_opts); - return err; -} - -static void __team_options_mark_removed(struct team *team, - const struct team_option *option, - size_t option_count) -{ - int i; - - for (i = 0; i < option_count; i++, option++) { - struct team_option *del_opt; - - del_opt = __team_find_option(team, option->name); - if (del_opt) - __team_option_inst_mark_removed_option(team, del_opt); - } -} - -static void __team_options_unregister(struct team *team, - const struct team_option *option, - size_t option_count) -{ - int i; - - for (i = 0; i < option_count; i++, option++) { - struct team_option *del_opt; - - del_opt = __team_find_option(team, option->name); - if (del_opt) { - __team_option_inst_del_option(team, del_opt); - list_del(&del_opt->list); - kfree(del_opt); - } - } -} - -static void __team_options_change_check(struct team *team); - -int team_options_register(struct team *team, - const struct team_option *option, - size_t option_count) -{ - int err; - - err = __team_options_register(team, option, option_count); - if (err) - return err; - __team_options_change_check(team); - return 0; -} -EXPORT_SYMBOL(team_options_register); - -void team_options_unregister(struct team *team, - const struct team_option *option, - size_t option_count) -{ - __team_options_mark_removed(team, option, option_count); - __team_options_change_check(team); - __team_options_unregister(team, option, option_count); -} -EXPORT_SYMBOL(team_options_unregister); - -static int team_option_get(struct team *team, - struct team_option_inst *opt_inst, - struct team_gsetter_ctx *ctx) -{ - if (!opt_inst->option->getter) - return -EOPNOTSUPP; - - opt_inst->option->getter(team, ctx); - return 0; -} - -static int team_option_set(struct team *team, - struct team_option_inst *opt_inst, - struct team_gsetter_ctx *ctx) -{ - if (!opt_inst->option->setter) - return -EOPNOTSUPP; - return opt_inst->option->setter(team, ctx); -} - -void team_option_inst_set_change(struct team_option_inst_info *opt_inst_info) -{ - struct team_option_inst *opt_inst; - - opt_inst = container_of(opt_inst_info, struct team_option_inst, info); - opt_inst->changed = true; -} -EXPORT_SYMBOL(team_option_inst_set_change); - -void team_options_change_check(struct team *team) -{ - __team_options_change_check(team); -} -EXPORT_SYMBOL(team_options_change_check); - - -/**************** - * Mode handling - ****************/ - -static LIST_HEAD(mode_list); -static DEFINE_SPINLOCK(mode_list_lock); - -struct team_mode_item { - struct list_head list; - const struct team_mode *mode; -}; - -static struct team_mode_item *__find_mode(const char *kind) -{ - struct team_mode_item *mitem; - - list_for_each_entry(mitem, &mode_list, list) { - if (strcmp(mitem->mode->kind, kind) == 0) - return mitem; - } - return NULL; -} - -static bool is_good_mode_name(const char *name) -{ - while (*name != '\0') { - if (!isalpha(*name) && !isdigit(*name) && *name != '_') - return false; - name++; - } - return true; -} - -int team_mode_register(const struct team_mode *mode) -{ - int err = 0; - struct team_mode_item *mitem; - - if (!is_good_mode_name(mode->kind) || - mode->priv_size > TEAM_MODE_PRIV_SIZE) - return -EINVAL; - - mitem = kmalloc(sizeof(*mitem), GFP_KERNEL); - if (!mitem) - return -ENOMEM; - - spin_lock(&mode_list_lock); - if (__find_mode(mode->kind)) { - err = -EEXIST; - kfree(mitem); - goto unlock; - } - mitem->mode = mode; - list_add_tail(&mitem->list, &mode_list); -unlock: - spin_unlock(&mode_list_lock); - return err; -} -EXPORT_SYMBOL(team_mode_register); - -void team_mode_unregister(const struct team_mode *mode) -{ - struct team_mode_item *mitem; - - spin_lock(&mode_list_lock); - mitem = __find_mode(mode->kind); - if (mitem) { - list_del_init(&mitem->list); - kfree(mitem); - } - spin_unlock(&mode_list_lock); -} -EXPORT_SYMBOL(team_mode_unregister); - -static const struct team_mode *team_mode_get(const char *kind) -{ - struct team_mode_item *mitem; - const struct team_mode *mode = NULL; - - if (!try_module_get(THIS_MODULE)) - return NULL; - - spin_lock(&mode_list_lock); - mitem = __find_mode(kind); - if (!mitem) { - spin_unlock(&mode_list_lock); - request_module("team-mode-%s", kind); - spin_lock(&mode_list_lock); - mitem = __find_mode(kind); - } - if (mitem) { - mode = mitem->mode; - if (!try_module_get(mode->owner)) - mode = NULL; - } - - spin_unlock(&mode_list_lock); - module_put(THIS_MODULE); - return mode; -} - -static void team_mode_put(const struct team_mode *mode) -{ - module_put(mode->owner); -} - -static bool team_dummy_transmit(struct team *team, struct sk_buff *skb) -{ - dev_kfree_skb_any(skb); - return false; -} - -static rx_handler_result_t team_dummy_receive(struct team *team, - struct team_port *port, - struct sk_buff *skb) -{ - return RX_HANDLER_ANOTHER; -} - -static const struct team_mode __team_no_mode = { - .kind = "*NOMODE*", -}; - -static bool team_is_mode_set(struct team *team) -{ - return team->mode != &__team_no_mode; -} - -static void team_set_no_mode(struct team *team) -{ - team->user_carrier_enabled = false; - team->mode = &__team_no_mode; -} - -static void team_adjust_ops(struct team *team) -{ - /* - * To avoid checks in rx/tx skb paths, ensure here that non-null and - * correct ops are always set. - */ - - if (!team->en_port_count || !team_is_mode_set(team) || - !team->mode->ops->transmit) - team->ops.transmit = team_dummy_transmit; - else - team->ops.transmit = team->mode->ops->transmit; - - if (!team->en_port_count || !team_is_mode_set(team) || - !team->mode->ops->receive) - team->ops.receive = team_dummy_receive; - else - team->ops.receive = team->mode->ops->receive; -} - -/* - * We can benefit from the fact that it's ensured no port is present - * at the time of mode change. Therefore no packets are in fly so there's no - * need to set mode operations in any special way. - */ -static int __team_change_mode(struct team *team, - const struct team_mode *new_mode) -{ - /* Check if mode was previously set and do cleanup if so */ - if (team_is_mode_set(team)) { - void (*exit_op)(struct team *team) = team->ops.exit; - - /* Clear ops area so no callback is called any longer */ - memset(&team->ops, 0, sizeof(struct team_mode_ops)); - team_adjust_ops(team); - - if (exit_op) - exit_op(team); - team_mode_put(team->mode); - team_set_no_mode(team); - /* zero private data area */ - memset(&team->mode_priv, 0, - sizeof(struct team) - offsetof(struct team, mode_priv)); - } - - if (!new_mode) - return 0; - - if (new_mode->ops->init) { - int err; - - err = new_mode->ops->init(team); - if (err) - return err; - } - - team->mode = new_mode; - memcpy(&team->ops, new_mode->ops, sizeof(struct team_mode_ops)); - team_adjust_ops(team); - - return 0; -} - -static int team_change_mode(struct team *team, const char *kind) -{ - const struct team_mode *new_mode; - struct net_device *dev = team->dev; - int err; - - if (!list_empty(&team->port_list)) { - netdev_err(dev, "No ports can be present during mode change\n"); - return -EBUSY; - } - - if (team_is_mode_set(team) && strcmp(team->mode->kind, kind) == 0) { - netdev_err(dev, "Unable to change to the same mode the team is in\n"); - return -EINVAL; - } - - new_mode = team_mode_get(kind); - if (!new_mode) { - netdev_err(dev, "Mode \"%s\" not found\n", kind); - return -EINVAL; - } - - err = __team_change_mode(team, new_mode); - if (err) { - netdev_err(dev, "Failed to change to mode \"%s\"\n", kind); - team_mode_put(new_mode); - return err; - } - - netdev_info(dev, "Mode changed to \"%s\"\n", kind); - return 0; -} - - -/********************* - * Peers notification - *********************/ - -static void team_notify_peers_work(struct work_struct *work) -{ - struct team *team; - int val; - - team = container_of(work, struct team, notify_peers.dw.work); - - if (!rtnl_trylock()) { - schedule_delayed_work(&team->notify_peers.dw, 0); - return; - } - val = atomic_dec_if_positive(&team->notify_peers.count_pending); - if (val < 0) { - rtnl_unlock(); - return; - } - call_netdevice_notifiers(NETDEV_NOTIFY_PEERS, team->dev); - rtnl_unlock(); - if (val) - schedule_delayed_work(&team->notify_peers.dw, - msecs_to_jiffies(team->notify_peers.interval)); -} - -static void team_notify_peers(struct team *team) -{ - if (!team->notify_peers.count || !netif_running(team->dev)) - return; - atomic_add(team->notify_peers.count, &team->notify_peers.count_pending); - schedule_delayed_work(&team->notify_peers.dw, 0); -} - -static void team_notify_peers_init(struct team *team) -{ - INIT_DELAYED_WORK(&team->notify_peers.dw, team_notify_peers_work); -} - -static void team_notify_peers_fini(struct team *team) -{ - cancel_delayed_work_sync(&team->notify_peers.dw); -} - - -/******************************* - * Send multicast group rejoins - *******************************/ - -static void team_mcast_rejoin_work(struct work_struct *work) -{ - struct team *team; - int val; - - team = container_of(work, struct team, mcast_rejoin.dw.work); - - if (!rtnl_trylock()) { - schedule_delayed_work(&team->mcast_rejoin.dw, 0); - return; - } - val = atomic_dec_if_positive(&team->mcast_rejoin.count_pending); - if (val < 0) { - rtnl_unlock(); - return; - } - call_netdevice_notifiers(NETDEV_RESEND_IGMP, team->dev); - rtnl_unlock(); - if (val) - schedule_delayed_work(&team->mcast_rejoin.dw, - msecs_to_jiffies(team->mcast_rejoin.interval)); -} - -static void team_mcast_rejoin(struct team *team) -{ - if (!team->mcast_rejoin.count || !netif_running(team->dev)) - return; - atomic_add(team->mcast_rejoin.count, &team->mcast_rejoin.count_pending); - schedule_delayed_work(&team->mcast_rejoin.dw, 0); -} - -static void team_mcast_rejoin_init(struct team *team) -{ - INIT_DELAYED_WORK(&team->mcast_rejoin.dw, team_mcast_rejoin_work); -} - -static void team_mcast_rejoin_fini(struct team *team) -{ - cancel_delayed_work_sync(&team->mcast_rejoin.dw); -} - - -/************************ - * Rx path frame handler - ************************/ - -/* note: already called with rcu_read_lock */ -static rx_handler_result_t team_handle_frame(struct sk_buff **pskb) -{ - struct sk_buff *skb = *pskb; - struct team_port *port; - struct team *team; - rx_handler_result_t res; - - skb = skb_share_check(skb, GFP_ATOMIC); - if (!skb) - return RX_HANDLER_CONSUMED; - - *pskb = skb; - - port = team_port_get_rcu(skb->dev); - team = port->team; - if (!team_port_enabled(port)) { - if (is_link_local_ether_addr(eth_hdr(skb)->h_dest)) - /* link-local packets are mostly useful when stack receives them - * with the link they arrive on. - */ - return RX_HANDLER_PASS; - /* allow exact match delivery for disabled ports */ - res = RX_HANDLER_EXACT; - } else { - res = team->ops.receive(team, port, skb); - } - if (res == RX_HANDLER_ANOTHER) { - struct team_pcpu_stats *pcpu_stats; - - pcpu_stats = this_cpu_ptr(team->pcpu_stats); - u64_stats_update_begin(&pcpu_stats->syncp); - u64_stats_inc(&pcpu_stats->rx_packets); - u64_stats_add(&pcpu_stats->rx_bytes, skb->len); - if (skb->pkt_type == PACKET_MULTICAST) - u64_stats_inc(&pcpu_stats->rx_multicast); - u64_stats_update_end(&pcpu_stats->syncp); - - skb->dev = team->dev; - } else if (res == RX_HANDLER_EXACT) { - this_cpu_inc(team->pcpu_stats->rx_nohandler); - } else { - this_cpu_inc(team->pcpu_stats->rx_dropped); - } - - return res; -} - - -/************************************* - * Multiqueue Tx port select override - *************************************/ - -static int team_queue_override_init(struct team *team) -{ - struct list_head *listarr; - unsigned int queue_cnt = team->dev->num_tx_queues - 1; - unsigned int i; - - if (!queue_cnt) - return 0; - listarr = kmalloc_array(queue_cnt, sizeof(struct list_head), - GFP_KERNEL); - if (!listarr) - return -ENOMEM; - team->qom_lists = listarr; - for (i = 0; i < queue_cnt; i++) - INIT_LIST_HEAD(listarr++); - return 0; -} - -static void team_queue_override_fini(struct team *team) -{ - kfree(team->qom_lists); -} - -static struct list_head *__team_get_qom_list(struct team *team, u16 queue_id) -{ - return &team->qom_lists[queue_id - 1]; -} - -/* - * note: already called with rcu_read_lock - */ -static bool team_queue_override_transmit(struct team *team, struct sk_buff *skb) -{ - struct list_head *qom_list; - struct team_port *port; - - if (!team->queue_override_enabled || !skb->queue_mapping) - return false; - qom_list = __team_get_qom_list(team, skb->queue_mapping); - list_for_each_entry_rcu(port, qom_list, qom_list) { - if (!team_dev_queue_xmit(team, port, skb)) - return true; - } - return false; -} - -static void __team_queue_override_port_del(struct team *team, - struct team_port *port) -{ - if (!port->queue_id) - return; - list_del_rcu(&port->qom_list); -} - -static bool team_queue_override_port_has_gt_prio_than(struct team_port *port, - struct team_port *cur) -{ - if (port->priority < cur->priority) - return true; - if (port->priority > cur->priority) - return false; - if (port->index < cur->index) - return true; - return false; -} - -static void __team_queue_override_port_add(struct team *team, - struct team_port *port) -{ - struct team_port *cur; - struct list_head *qom_list; - struct list_head *node; - - if (!port->queue_id) - return; - qom_list = __team_get_qom_list(team, port->queue_id); - node = qom_list; - list_for_each_entry(cur, qom_list, qom_list) { - if (team_queue_override_port_has_gt_prio_than(port, cur)) - break; - node = &cur->qom_list; - } - list_add_tail_rcu(&port->qom_list, node); -} - -static void __team_queue_override_enabled_check(struct team *team) -{ - struct team_port *port; - bool enabled = false; - - list_for_each_entry(port, &team->port_list, list) { - if (port->queue_id) { - enabled = true; - break; - } - } - if (enabled == team->queue_override_enabled) - return; - netdev_dbg(team->dev, "%s queue override\n", - enabled ? "Enabling" : "Disabling"); - team->queue_override_enabled = enabled; -} - -static void team_queue_override_port_prio_changed(struct team *team, - struct team_port *port) -{ - if (!port->queue_id || team_port_enabled(port)) - return; - __team_queue_override_port_del(team, port); - __team_queue_override_port_add(team, port); - __team_queue_override_enabled_check(team); -} - -static void team_queue_override_port_change_queue_id(struct team *team, - struct team_port *port, - u16 new_queue_id) -{ - if (team_port_enabled(port)) { - __team_queue_override_port_del(team, port); - port->queue_id = new_queue_id; - __team_queue_override_port_add(team, port); - __team_queue_override_enabled_check(team); - } else { - port->queue_id = new_queue_id; - } -} - -static void team_queue_override_port_add(struct team *team, - struct team_port *port) -{ - __team_queue_override_port_add(team, port); - __team_queue_override_enabled_check(team); -} - -static void team_queue_override_port_del(struct team *team, - struct team_port *port) -{ - __team_queue_override_port_del(team, port); - __team_queue_override_enabled_check(team); -} - - -/**************** - * Port handling - ****************/ - -static bool team_port_find(const struct team *team, - const struct team_port *port) -{ - struct team_port *cur; - - list_for_each_entry(cur, &team->port_list, list) - if (cur == port) - return true; - return false; -} - -/* - * Enable/disable port by adding to enabled port hashlist and setting - * port->index (Might be racy so reader could see incorrect ifindex when - * processing a flying packet, but that is not a problem). Write guarded - * by team->lock. - */ -static void team_port_enable(struct team *team, - struct team_port *port) -{ - if (team_port_enabled(port)) - return; - port->index = team->en_port_count++; - hlist_add_head_rcu(&port->hlist, - team_port_index_hash(team, port->index)); - team_adjust_ops(team); - team_queue_override_port_add(team, port); - if (team->ops.port_enabled) - team->ops.port_enabled(team, port); - team_notify_peers(team); - team_mcast_rejoin(team); - team_lower_state_changed(port); -} - -static void __reconstruct_port_hlist(struct team *team, int rm_index) -{ - int i; - struct team_port *port; - - for (i = rm_index + 1; i < team->en_port_count; i++) { - port = team_get_port_by_index(team, i); - hlist_del_rcu(&port->hlist); - port->index--; - hlist_add_head_rcu(&port->hlist, - team_port_index_hash(team, port->index)); - } -} - -static void team_port_disable(struct team *team, - struct team_port *port) -{ - if (!team_port_enabled(port)) - return; - if (team->ops.port_disabled) - team->ops.port_disabled(team, port); - hlist_del_rcu(&port->hlist); - __reconstruct_port_hlist(team, port->index); - port->index = -1; - team->en_port_count--; - team_queue_override_port_del(team, port); - team_adjust_ops(team); - team_lower_state_changed(port); -} - -#define TEAM_VLAN_FEATURES (NETIF_F_HW_CSUM | NETIF_F_SG | \ - NETIF_F_FRAGLIST | NETIF_F_GSO_SOFTWARE | \ - NETIF_F_HIGHDMA | NETIF_F_LRO) - -#define TEAM_ENC_FEATURES (NETIF_F_HW_CSUM | NETIF_F_SG | \ - NETIF_F_RXCSUM | NETIF_F_GSO_SOFTWARE) - -static void __team_compute_features(struct team *team) -{ - struct team_port *port; - netdev_features_t vlan_features = TEAM_VLAN_FEATURES & - NETIF_F_ALL_FOR_ALL; - netdev_features_t enc_features = TEAM_ENC_FEATURES; - unsigned short max_hard_header_len = ETH_HLEN; - unsigned int dst_release_flag = IFF_XMIT_DST_RELEASE | - IFF_XMIT_DST_RELEASE_PERM; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &team->port_list, list) { - vlan_features = netdev_increment_features(vlan_features, - port->dev->vlan_features, - TEAM_VLAN_FEATURES); - enc_features = - netdev_increment_features(enc_features, - port->dev->hw_enc_features, - TEAM_ENC_FEATURES); - - - dst_release_flag &= port->dev->priv_flags; - if (port->dev->hard_header_len > max_hard_header_len) - max_hard_header_len = port->dev->hard_header_len; - } - rcu_read_unlock(); - - team->dev->vlan_features = vlan_features; - team->dev->hw_enc_features = enc_features | NETIF_F_GSO_ENCAP_ALL | - NETIF_F_HW_VLAN_CTAG_TX | - NETIF_F_HW_VLAN_STAG_TX; - team->dev->hard_header_len = max_hard_header_len; - - team->dev->priv_flags &= ~IFF_XMIT_DST_RELEASE; - if (dst_release_flag == (IFF_XMIT_DST_RELEASE | IFF_XMIT_DST_RELEASE_PERM)) - team->dev->priv_flags |= IFF_XMIT_DST_RELEASE; -} - -static void team_compute_features(struct team *team) -{ - __team_compute_features(team); - netdev_change_features(team->dev); -} - -static int team_port_enter(struct team *team, struct team_port *port) -{ - int err = 0; - - dev_hold(team->dev); - if (team->ops.port_enter) { - err = team->ops.port_enter(team, port); - if (err) { - netdev_err(team->dev, "Device %s failed to enter team mode\n", - port->dev->name); - goto err_port_enter; - } - } - - return 0; - -err_port_enter: - dev_put(team->dev); - - return err; -} - -static void team_port_leave(struct team *team, struct team_port *port) -{ - if (team->ops.port_leave) - team->ops.port_leave(team, port); - dev_put(team->dev); -} - -#ifdef CONFIG_NET_POLL_CONTROLLER -static int __team_port_enable_netpoll(struct team_port *port) -{ - struct netpoll *np; - int err; - - np = kzalloc(sizeof(*np), GFP_KERNEL); - if (!np) - return -ENOMEM; - - err = __netpoll_setup(np, port->dev); - if (err) { - kfree(np); - return err; - } - port->np = np; - return err; -} - -static int team_port_enable_netpoll(struct team_port *port) -{ - if (!port->team->dev->npinfo) - return 0; - - return __team_port_enable_netpoll(port); -} - -static void team_port_disable_netpoll(struct team_port *port) -{ - struct netpoll *np = port->np; - - if (!np) - return; - port->np = NULL; - - __netpoll_free(np); -} -#else -static int team_port_enable_netpoll(struct team_port *port) -{ - return 0; -} -static void team_port_disable_netpoll(struct team_port *port) -{ -} -#endif - -static int team_upper_dev_link(struct team *team, struct team_port *port, - struct netlink_ext_ack *extack) -{ - struct netdev_lag_upper_info lag_upper_info; - int err; - - lag_upper_info.tx_type = team->mode->lag_tx_type; - lag_upper_info.hash_type = NETDEV_LAG_HASH_UNKNOWN; - err = netdev_master_upper_dev_link(port->dev, team->dev, NULL, - &lag_upper_info, extack); - if (err) - return err; - port->dev->priv_flags |= IFF_TEAM_PORT; - return 0; -} - -static void team_upper_dev_unlink(struct team *team, struct team_port *port) -{ - netdev_upper_dev_unlink(port->dev, team->dev); - port->dev->priv_flags &= ~IFF_TEAM_PORT; -} - -static void __team_port_change_port_added(struct team_port *port, bool linkup); -static int team_dev_type_check_change(struct net_device *dev, - struct net_device *port_dev); - -static int team_port_add(struct team *team, struct net_device *port_dev, - struct netlink_ext_ack *extack) -{ - struct net_device *dev = team->dev; - struct team_port *port; - char *portname = port_dev->name; - int err; - - if (port_dev->flags & IFF_LOOPBACK) { - NL_SET_ERR_MSG(extack, "Loopback device can't be added as a team port"); - netdev_err(dev, "Device %s is loopback device. Loopback devices can't be added as a team port\n", - portname); - return -EINVAL; - } - - if (netif_is_team_port(port_dev)) { - NL_SET_ERR_MSG(extack, "Device is already a port of a team device"); - netdev_err(dev, "Device %s is already a port " - "of a team device\n", portname); - return -EBUSY; - } - - if (dev == port_dev) { - NL_SET_ERR_MSG(extack, "Cannot enslave team device to itself"); - netdev_err(dev, "Cannot enslave team device to itself\n"); - return -EINVAL; - } - - if (netdev_has_upper_dev(dev, port_dev)) { - NL_SET_ERR_MSG(extack, "Device is already an upper device of the team interface"); - netdev_err(dev, "Device %s is already an upper device of the team interface\n", - portname); - return -EBUSY; - } - - if (port_dev->features & NETIF_F_VLAN_CHALLENGED && - vlan_uses_dev(dev)) { - NL_SET_ERR_MSG(extack, "Device is VLAN challenged and team device has VLAN set up"); - netdev_err(dev, "Device %s is VLAN challenged and team device has VLAN set up\n", - portname); - return -EPERM; - } - - err = team_dev_type_check_change(dev, port_dev); - if (err) - return err; - - if (port_dev->flags & IFF_UP) { - NL_SET_ERR_MSG(extack, "Device is up. Set it down before adding it as a team port"); - netdev_err(dev, "Device %s is up. Set it down before adding it as a team port\n", - portname); - return -EBUSY; - } - - port = kzalloc(sizeof(struct team_port) + team->mode->port_priv_size, - GFP_KERNEL); - if (!port) - return -ENOMEM; - - port->dev = port_dev; - port->team = team; - INIT_LIST_HEAD(&port->qom_list); - - port->orig.mtu = port_dev->mtu; - err = dev_set_mtu(port_dev, dev->mtu); - if (err) { - netdev_dbg(dev, "Error %d calling dev_set_mtu\n", err); - goto err_set_mtu; - } - - memcpy(port->orig.dev_addr, port_dev->dev_addr, port_dev->addr_len); - - err = team_port_enter(team, port); - if (err) { - netdev_err(dev, "Device %s failed to enter team mode\n", - portname); - goto err_port_enter; - } - - err = dev_open(port_dev, extack); - if (err) { - netdev_dbg(dev, "Device %s opening failed\n", - portname); - goto err_dev_open; - } - - err = vlan_vids_add_by_dev(port_dev, dev); - if (err) { - netdev_err(dev, "Failed to add vlan ids to device %s\n", - portname); - goto err_vids_add; - } - - err = team_port_enable_netpoll(port); - if (err) { - netdev_err(dev, "Failed to enable netpoll on device %s\n", - portname); - goto err_enable_netpoll; - } - - if (!(dev->features & NETIF_F_LRO)) - dev_disable_lro(port_dev); - - err = netdev_rx_handler_register(port_dev, team_handle_frame, - port); - if (err) { - netdev_err(dev, "Device %s failed to register rx_handler\n", - portname); - goto err_handler_register; - } - - err = team_upper_dev_link(team, port, extack); - if (err) { - netdev_err(dev, "Device %s failed to set upper link\n", - portname); - goto err_set_upper_link; - } - - err = __team_option_inst_add_port(team, port); - if (err) { - netdev_err(dev, "Device %s failed to add per-port options\n", - portname); - goto err_option_port_add; - } - - /* set promiscuity level to new slave */ - if (dev->flags & IFF_PROMISC) { - err = dev_set_promiscuity(port_dev, 1); - if (err) - goto err_set_slave_promisc; - } - - /* set allmulti level to new slave */ - if (dev->flags & IFF_ALLMULTI) { - err = dev_set_allmulti(port_dev, 1); - if (err) { - if (dev->flags & IFF_PROMISC) - dev_set_promiscuity(port_dev, -1); - goto err_set_slave_promisc; - } - } - - if (dev->flags & IFF_UP) { - netif_addr_lock_bh(dev); - dev_uc_sync_multiple(port_dev, dev); - dev_mc_sync_multiple(port_dev, dev); - netif_addr_unlock_bh(dev); - } - - port->index = -1; - list_add_tail_rcu(&port->list, &team->port_list); - team_port_enable(team, port); - __team_compute_features(team); - __team_port_change_port_added(port, !!netif_oper_up(port_dev)); - __team_options_change_check(team); - - netdev_info(dev, "Port device %s added\n", portname); - - return 0; - -err_set_slave_promisc: - __team_option_inst_del_port(team, port); - -err_option_port_add: - team_upper_dev_unlink(team, port); - -err_set_upper_link: - netdev_rx_handler_unregister(port_dev); - -err_handler_register: - team_port_disable_netpoll(port); - -err_enable_netpoll: - vlan_vids_del_by_dev(port_dev, dev); - -err_vids_add: - dev_close(port_dev); - -err_dev_open: - team_port_leave(team, port); - team_port_set_orig_dev_addr(port); - -err_port_enter: - dev_set_mtu(port_dev, port->orig.mtu); - -err_set_mtu: - kfree(port); - - return err; -} - -static void __team_port_change_port_removed(struct team_port *port); - -static int team_port_del(struct team *team, struct net_device *port_dev) -{ - struct net_device *dev = team->dev; - struct team_port *port; - char *portname = port_dev->name; - - port = team_port_get_rtnl(port_dev); - if (!port || !team_port_find(team, port)) { - netdev_err(dev, "Device %s does not act as a port of this team\n", - portname); - return -ENOENT; - } - - team_port_disable(team, port); - list_del_rcu(&port->list); - - if (dev->flags & IFF_PROMISC) - dev_set_promiscuity(port_dev, -1); - if (dev->flags & IFF_ALLMULTI) - dev_set_allmulti(port_dev, -1); - - team_upper_dev_unlink(team, port); - netdev_rx_handler_unregister(port_dev); - team_port_disable_netpoll(port); - vlan_vids_del_by_dev(port_dev, dev); - if (dev->flags & IFF_UP) { - dev_uc_unsync(port_dev, dev); - dev_mc_unsync(port_dev, dev); - } - dev_close(port_dev); - team_port_leave(team, port); - - __team_option_inst_mark_removed_port(team, port); - __team_options_change_check(team); - __team_option_inst_del_port(team, port); - __team_port_change_port_removed(port); - - team_port_set_orig_dev_addr(port); - dev_set_mtu(port_dev, port->orig.mtu); - kfree_rcu(port, rcu); - netdev_info(dev, "Port device %s removed\n", portname); - __team_compute_features(team); - - return 0; -} - - -/***************** - * Net device ops - *****************/ - -static void team_mode_option_get(struct team *team, struct team_gsetter_ctx *ctx) -{ - ctx->data.str_val = team->mode->kind; -} - -static int team_mode_option_set(struct team *team, struct team_gsetter_ctx *ctx) -{ - return team_change_mode(team, ctx->data.str_val); -} - -static void team_notify_peers_count_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - ctx->data.u32_val = team->notify_peers.count; -} - -static int team_notify_peers_count_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - team->notify_peers.count = ctx->data.u32_val; - return 0; -} - -static void team_notify_peers_interval_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - ctx->data.u32_val = team->notify_peers.interval; -} - -static int team_notify_peers_interval_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - team->notify_peers.interval = ctx->data.u32_val; - return 0; -} - -static void team_mcast_rejoin_count_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - ctx->data.u32_val = team->mcast_rejoin.count; -} - -static int team_mcast_rejoin_count_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - team->mcast_rejoin.count = ctx->data.u32_val; - return 0; -} - -static void team_mcast_rejoin_interval_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - ctx->data.u32_val = team->mcast_rejoin.interval; -} - -static int team_mcast_rejoin_interval_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - team->mcast_rejoin.interval = ctx->data.u32_val; - return 0; -} - -static void team_port_en_option_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - ctx->data.bool_val = team_port_enabled(port); -} - -static int team_port_en_option_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - if (ctx->data.bool_val) - team_port_enable(team, port); - else - team_port_disable(team, port); - return 0; -} - -static void team_user_linkup_option_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - ctx->data.bool_val = port->user.linkup; -} - -static void __team_carrier_check(struct team *team); - -static int team_user_linkup_option_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - port->user.linkup = ctx->data.bool_val; - team_refresh_port_linkup(port); - __team_carrier_check(port->team); - return 0; -} - -static void team_user_linkup_en_option_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - ctx->data.bool_val = port->user.linkup_enabled; -} - -static int team_user_linkup_en_option_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - port->user.linkup_enabled = ctx->data.bool_val; - team_refresh_port_linkup(port); - __team_carrier_check(port->team); - return 0; -} - -static void team_priority_option_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - ctx->data.s32_val = port->priority; -} - -static int team_priority_option_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - s32 priority = ctx->data.s32_val; - - if (port->priority == priority) - return 0; - port->priority = priority; - team_queue_override_port_prio_changed(team, port); - return 0; -} - -static void team_queue_id_option_get(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - - ctx->data.u32_val = port->queue_id; -} - -static int team_queue_id_option_set(struct team *team, - struct team_gsetter_ctx *ctx) -{ - struct team_port *port = ctx->info->port; - u16 new_queue_id = ctx->data.u32_val; - - if (port->queue_id == new_queue_id) - return 0; - if (new_queue_id >= team->dev->real_num_tx_queues) - return -EINVAL; - team_queue_override_port_change_queue_id(team, port, new_queue_id); - return 0; -} - -static const struct team_option team_options[] = { - { - .name = "mode", - .type = TEAM_OPTION_TYPE_STRING, - .getter = team_mode_option_get, - .setter = team_mode_option_set, - }, - { - .name = "notify_peers_count", - .type = TEAM_OPTION_TYPE_U32, - .getter = team_notify_peers_count_get, - .setter = team_notify_peers_count_set, - }, - { - .name = "notify_peers_interval", - .type = TEAM_OPTION_TYPE_U32, - .getter = team_notify_peers_interval_get, - .setter = team_notify_peers_interval_set, - }, - { - .name = "mcast_rejoin_count", - .type = TEAM_OPTION_TYPE_U32, - .getter = team_mcast_rejoin_count_get, - .setter = team_mcast_rejoin_count_set, - }, - { - .name = "mcast_rejoin_interval", - .type = TEAM_OPTION_TYPE_U32, - .getter = team_mcast_rejoin_interval_get, - .setter = team_mcast_rejoin_interval_set, - }, - { - .name = "enabled", - .type = TEAM_OPTION_TYPE_BOOL, - .per_port = true, - .getter = team_port_en_option_get, - .setter = team_port_en_option_set, - }, - { - .name = "user_linkup", - .type = TEAM_OPTION_TYPE_BOOL, - .per_port = true, - .getter = team_user_linkup_option_get, - .setter = team_user_linkup_option_set, - }, - { - .name = "user_linkup_enabled", - .type = TEAM_OPTION_TYPE_BOOL, - .per_port = true, - .getter = team_user_linkup_en_option_get, - .setter = team_user_linkup_en_option_set, - }, - { - .name = "priority", - .type = TEAM_OPTION_TYPE_S32, - .per_port = true, - .getter = team_priority_option_get, - .setter = team_priority_option_set, - }, - { - .name = "queue_id", - .type = TEAM_OPTION_TYPE_U32, - .per_port = true, - .getter = team_queue_id_option_get, - .setter = team_queue_id_option_set, - }, -}; - - -static int team_init(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - int i; - int err; - - team->dev = dev; - team_set_no_mode(team); - team->notifier_ctx = false; - - team->pcpu_stats = netdev_alloc_pcpu_stats(struct team_pcpu_stats); - if (!team->pcpu_stats) - return -ENOMEM; - - for (i = 0; i < TEAM_PORT_HASHENTRIES; i++) - INIT_HLIST_HEAD(&team->en_port_hlist[i]); - INIT_LIST_HEAD(&team->port_list); - err = team_queue_override_init(team); - if (err) - goto err_team_queue_override_init; - - team_adjust_ops(team); - - INIT_LIST_HEAD(&team->option_list); - INIT_LIST_HEAD(&team->option_inst_list); - - team_notify_peers_init(team); - team_mcast_rejoin_init(team); - - err = team_options_register(team, team_options, ARRAY_SIZE(team_options)); - if (err) - goto err_options_register; - netif_carrier_off(dev); - - lockdep_register_key(&team->team_lock_key); - __mutex_init(&team->lock, "team->team_lock_key", &team->team_lock_key); - netdev_lockdep_set_classes(dev); - - return 0; - -err_options_register: - team_mcast_rejoin_fini(team); - team_notify_peers_fini(team); - team_queue_override_fini(team); -err_team_queue_override_init: - free_percpu(team->pcpu_stats); - - return err; -} - -static void team_uninit(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - struct team_port *tmp; - - mutex_lock(&team->lock); - list_for_each_entry_safe(port, tmp, &team->port_list, list) - team_port_del(team, port->dev); - - __team_change_mode(team, NULL); /* cleanup */ - __team_options_unregister(team, team_options, ARRAY_SIZE(team_options)); - team_mcast_rejoin_fini(team); - team_notify_peers_fini(team); - team_queue_override_fini(team); - mutex_unlock(&team->lock); - netdev_change_features(dev); - lockdep_unregister_key(&team->team_lock_key); -} - -static void team_destructor(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - - free_percpu(team->pcpu_stats); -} - -static int team_open(struct net_device *dev) -{ - return 0; -} - -static int team_close(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - - list_for_each_entry(port, &team->port_list, list) { - dev_uc_unsync(port->dev, dev); - dev_mc_unsync(port->dev, dev); - } - - return 0; -} - -/* - * note: already called with rcu_read_lock - */ -static netdev_tx_t team_xmit(struct sk_buff *skb, struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - bool tx_success; - unsigned int len = skb->len; - - tx_success = team_queue_override_transmit(team, skb); - if (!tx_success) - tx_success = team->ops.transmit(team, skb); - if (tx_success) { - struct team_pcpu_stats *pcpu_stats; - - pcpu_stats = this_cpu_ptr(team->pcpu_stats); - u64_stats_update_begin(&pcpu_stats->syncp); - u64_stats_inc(&pcpu_stats->tx_packets); - u64_stats_add(&pcpu_stats->tx_bytes, len); - u64_stats_update_end(&pcpu_stats->syncp); - } else { - this_cpu_inc(team->pcpu_stats->tx_dropped); - } - - return NETDEV_TX_OK; -} - -static u16 team_select_queue(struct net_device *dev, struct sk_buff *skb, - struct net_device *sb_dev) -{ - /* - * This helper function exists to help dev_pick_tx get the correct - * destination queue. Using a helper function skips a call to - * skb_tx_hash and will put the skbs in the queue we expect on their - * way down to the team driver. - */ - u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0; - - /* - * Save the original txq to restore before passing to the driver - */ - qdisc_skb_cb(skb)->slave_dev_queue_mapping = skb->queue_mapping; - - if (unlikely(txq >= dev->real_num_tx_queues)) { - do { - txq -= dev->real_num_tx_queues; - } while (txq >= dev->real_num_tx_queues); - } - return txq; -} - -static void team_change_rx_flags(struct net_device *dev, int change) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - int inc; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &team->port_list, list) { - if (change & IFF_PROMISC) { - inc = dev->flags & IFF_PROMISC ? 1 : -1; - dev_set_promiscuity(port->dev, inc); - } - if (change & IFF_ALLMULTI) { - inc = dev->flags & IFF_ALLMULTI ? 1 : -1; - dev_set_allmulti(port->dev, inc); - } - } - rcu_read_unlock(); -} - -static void team_set_rx_mode(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &team->port_list, list) { - dev_uc_sync_multiple(port->dev, dev); - dev_mc_sync_multiple(port->dev, dev); - } - rcu_read_unlock(); -} - -static int team_set_mac_address(struct net_device *dev, void *p) -{ - struct sockaddr *addr = p; - struct team *team = netdev_priv(dev); - struct team_port *port; - - if (dev->type == ARPHRD_ETHER && !is_valid_ether_addr(addr->sa_data)) - return -EADDRNOTAVAIL; - dev_addr_set(dev, addr->sa_data); - mutex_lock(&team->lock); - list_for_each_entry(port, &team->port_list, list) - if (team->ops.port_change_dev_addr) - team->ops.port_change_dev_addr(team, port); - mutex_unlock(&team->lock); - return 0; -} - -static int team_change_mtu(struct net_device *dev, int new_mtu) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - int err; - - /* - * Alhough this is reader, it's guarded by team lock. It's not possible - * to traverse list in reverse under rcu_read_lock - */ - mutex_lock(&team->lock); - team->port_mtu_change_allowed = true; - list_for_each_entry(port, &team->port_list, list) { - err = dev_set_mtu(port->dev, new_mtu); - if (err) { - netdev_err(dev, "Device %s failed to change mtu", - port->dev->name); - goto unwind; - } - } - team->port_mtu_change_allowed = false; - mutex_unlock(&team->lock); - - dev->mtu = new_mtu; - - return 0; - -unwind: - list_for_each_entry_continue_reverse(port, &team->port_list, list) - dev_set_mtu(port->dev, dev->mtu); - team->port_mtu_change_allowed = false; - mutex_unlock(&team->lock); - - return err; -} - -static void -team_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats) -{ - struct team *team = netdev_priv(dev); - struct team_pcpu_stats *p; - u64 rx_packets, rx_bytes, rx_multicast, tx_packets, tx_bytes; - u32 rx_dropped = 0, tx_dropped = 0, rx_nohandler = 0; - unsigned int start; - int i; - - for_each_possible_cpu(i) { - p = per_cpu_ptr(team->pcpu_stats, i); - do { - start = u64_stats_fetch_begin(&p->syncp); - rx_packets = u64_stats_read(&p->rx_packets); - rx_bytes = u64_stats_read(&p->rx_bytes); - rx_multicast = u64_stats_read(&p->rx_multicast); - tx_packets = u64_stats_read(&p->tx_packets); - tx_bytes = u64_stats_read(&p->tx_bytes); - } while (u64_stats_fetch_retry(&p->syncp, start)); - - stats->rx_packets += rx_packets; - stats->rx_bytes += rx_bytes; - stats->multicast += rx_multicast; - stats->tx_packets += tx_packets; - stats->tx_bytes += tx_bytes; - /* - * rx_dropped, tx_dropped & rx_nohandler are u32, - * updated without syncp protection. - */ - rx_dropped += READ_ONCE(p->rx_dropped); - tx_dropped += READ_ONCE(p->tx_dropped); - rx_nohandler += READ_ONCE(p->rx_nohandler); - } - stats->rx_dropped = rx_dropped; - stats->tx_dropped = tx_dropped; - stats->rx_nohandler = rx_nohandler; -} - -static int team_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - int err; - - /* - * Alhough this is reader, it's guarded by team lock. It's not possible - * to traverse list in reverse under rcu_read_lock - */ - mutex_lock(&team->lock); - list_for_each_entry(port, &team->port_list, list) { - err = vlan_vid_add(port->dev, proto, vid); - if (err) - goto unwind; - } - mutex_unlock(&team->lock); - - return 0; - -unwind: - list_for_each_entry_continue_reverse(port, &team->port_list, list) - vlan_vid_del(port->dev, proto, vid); - mutex_unlock(&team->lock); - - return err; -} - -static int team_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - - mutex_lock(&team->lock); - list_for_each_entry(port, &team->port_list, list) - vlan_vid_del(port->dev, proto, vid); - mutex_unlock(&team->lock); - - return 0; -} - -#ifdef CONFIG_NET_POLL_CONTROLLER -static void team_poll_controller(struct net_device *dev) -{ -} - -static void __team_netpoll_cleanup(struct team *team) -{ - struct team_port *port; - - list_for_each_entry(port, &team->port_list, list) - team_port_disable_netpoll(port); -} - -static void team_netpoll_cleanup(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - - mutex_lock(&team->lock); - __team_netpoll_cleanup(team); - mutex_unlock(&team->lock); -} - -static int team_netpoll_setup(struct net_device *dev, - struct netpoll_info *npifo) -{ - struct team *team = netdev_priv(dev); - struct team_port *port; - int err = 0; - - mutex_lock(&team->lock); - list_for_each_entry(port, &team->port_list, list) { - err = __team_port_enable_netpoll(port); - if (err) { - __team_netpoll_cleanup(team); - break; - } - } - mutex_unlock(&team->lock); - return err; -} -#endif - -static int team_add_slave(struct net_device *dev, struct net_device *port_dev, - struct netlink_ext_ack *extack) -{ - struct team *team = netdev_priv(dev); - int err; - - mutex_lock(&team->lock); - err = team_port_add(team, port_dev, extack); - mutex_unlock(&team->lock); - - if (!err) - netdev_change_features(dev); - - return err; -} - -static int team_del_slave(struct net_device *dev, struct net_device *port_dev) -{ - struct team *team = netdev_priv(dev); - int err; - - mutex_lock(&team->lock); - err = team_port_del(team, port_dev); - mutex_unlock(&team->lock); - - if (err) - return err; - - if (netif_is_team_master(port_dev)) { - lockdep_unregister_key(&team->team_lock_key); - lockdep_register_key(&team->team_lock_key); - lockdep_set_class(&team->lock, &team->team_lock_key); - } - netdev_change_features(dev); - - return err; -} - -static netdev_features_t team_fix_features(struct net_device *dev, - netdev_features_t features) -{ - struct team_port *port; - struct team *team = netdev_priv(dev); - netdev_features_t mask; - - mask = features; - features &= ~NETIF_F_ONE_FOR_ALL; - features |= NETIF_F_ALL_FOR_ALL; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &team->port_list, list) { - features = netdev_increment_features(features, - port->dev->features, - mask); - } - rcu_read_unlock(); - - features = netdev_add_tso_features(features, mask); - - return features; -} - -static int team_change_carrier(struct net_device *dev, bool new_carrier) -{ - struct team *team = netdev_priv(dev); - - team->user_carrier_enabled = true; - - if (new_carrier) - netif_carrier_on(dev); - else - netif_carrier_off(dev); - return 0; -} - -static const struct net_device_ops team_netdev_ops = { - .ndo_init = team_init, - .ndo_uninit = team_uninit, - .ndo_open = team_open, - .ndo_stop = team_close, - .ndo_start_xmit = team_xmit, - .ndo_select_queue = team_select_queue, - .ndo_change_rx_flags = team_change_rx_flags, - .ndo_set_rx_mode = team_set_rx_mode, - .ndo_set_mac_address = team_set_mac_address, - .ndo_change_mtu = team_change_mtu, - .ndo_get_stats64 = team_get_stats64, - .ndo_vlan_rx_add_vid = team_vlan_rx_add_vid, - .ndo_vlan_rx_kill_vid = team_vlan_rx_kill_vid, -#ifdef CONFIG_NET_POLL_CONTROLLER - .ndo_poll_controller = team_poll_controller, - .ndo_netpoll_setup = team_netpoll_setup, - .ndo_netpoll_cleanup = team_netpoll_cleanup, -#endif - .ndo_add_slave = team_add_slave, - .ndo_del_slave = team_del_slave, - .ndo_fix_features = team_fix_features, - .ndo_change_carrier = team_change_carrier, - .ndo_features_check = passthru_features_check, -}; - -/*********************** - * ethtool interface - ***********************/ - -static void team_ethtool_get_drvinfo(struct net_device *dev, - struct ethtool_drvinfo *drvinfo) -{ - strscpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver)); -} - -static int team_ethtool_get_link_ksettings(struct net_device *dev, - struct ethtool_link_ksettings *cmd) -{ - struct team *team= netdev_priv(dev); - unsigned long speed = 0; - struct team_port *port; - - cmd->base.duplex = DUPLEX_UNKNOWN; - cmd->base.port = PORT_OTHER; - - rcu_read_lock(); - list_for_each_entry_rcu(port, &team->port_list, list) { - if (team_port_txable(port)) { - if (port->state.speed != SPEED_UNKNOWN) - speed += port->state.speed; - if (cmd->base.duplex == DUPLEX_UNKNOWN && - port->state.duplex != DUPLEX_UNKNOWN) - cmd->base.duplex = port->state.duplex; - } - } - rcu_read_unlock(); - - cmd->base.speed = speed ? : SPEED_UNKNOWN; - - return 0; -} - -static const struct ethtool_ops team_ethtool_ops = { - .get_drvinfo = team_ethtool_get_drvinfo, - .get_link = ethtool_op_get_link, - .get_link_ksettings = team_ethtool_get_link_ksettings, -}; - -/*********************** - * rt netlink interface - ***********************/ - -static void team_setup_by_port(struct net_device *dev, - struct net_device *port_dev) -{ - struct team *team = netdev_priv(dev); - - if (port_dev->type == ARPHRD_ETHER) - dev->header_ops = team->header_ops_cache; - else - dev->header_ops = port_dev->header_ops; - dev->type = port_dev->type; - dev->hard_header_len = port_dev->hard_header_len; - dev->needed_headroom = port_dev->needed_headroom; - dev->addr_len = port_dev->addr_len; - dev->mtu = port_dev->mtu; - memcpy(dev->broadcast, port_dev->broadcast, port_dev->addr_len); - eth_hw_addr_inherit(dev, port_dev); - - if (port_dev->flags & IFF_POINTOPOINT) { - dev->flags &= ~(IFF_BROADCAST | IFF_MULTICAST); - dev->flags |= (IFF_POINTOPOINT | IFF_NOARP); - } else if ((port_dev->flags & (IFF_BROADCAST | IFF_MULTICAST)) == - (IFF_BROADCAST | IFF_MULTICAST)) { - dev->flags |= (IFF_BROADCAST | IFF_MULTICAST); - dev->flags &= ~(IFF_POINTOPOINT | IFF_NOARP); - } -} - -static int team_dev_type_check_change(struct net_device *dev, - struct net_device *port_dev) -{ - struct team *team = netdev_priv(dev); - char *portname = port_dev->name; - int err; - - if (dev->type == port_dev->type) - return 0; - if (!list_empty(&team->port_list)) { - netdev_err(dev, "Device %s is of different type\n", portname); - return -EBUSY; - } - err = call_netdevice_notifiers(NETDEV_PRE_TYPE_CHANGE, dev); - err = notifier_to_errno(err); - if (err) { - netdev_err(dev, "Refused to change device type\n"); - return err; - } - dev_uc_flush(dev); - dev_mc_flush(dev); - team_setup_by_port(dev, port_dev); - call_netdevice_notifiers(NETDEV_POST_TYPE_CHANGE, dev); - return 0; -} - -static void team_setup(struct net_device *dev) -{ - struct team *team = netdev_priv(dev); - - ether_setup(dev); - dev->max_mtu = ETH_MAX_MTU; - team->header_ops_cache = dev->header_ops; - - dev->netdev_ops = &team_netdev_ops; - dev->ethtool_ops = &team_ethtool_ops; - dev->needs_free_netdev = true; - dev->priv_destructor = team_destructor; - dev->priv_flags &= ~(IFF_XMIT_DST_RELEASE | IFF_TX_SKB_SHARING); - dev->priv_flags |= IFF_NO_QUEUE; - dev->priv_flags |= IFF_TEAM; - - /* - * Indicate we support unicast address filtering. That way core won't - * bring us to promisc mode in case a unicast addr is added. - * Let this up to underlay drivers. - */ - dev->priv_flags |= IFF_UNICAST_FLT | IFF_LIVE_ADDR_CHANGE; - - dev->features |= NETIF_F_LLTX; - dev->features |= NETIF_F_GRO; - - /* Don't allow team devices to change network namespaces. */ - dev->features |= NETIF_F_NETNS_LOCAL; - - dev->hw_features = TEAM_VLAN_FEATURES | - NETIF_F_HW_VLAN_CTAG_RX | - NETIF_F_HW_VLAN_CTAG_FILTER | - NETIF_F_HW_VLAN_STAG_RX | - NETIF_F_HW_VLAN_STAG_FILTER; - - dev->hw_features |= NETIF_F_GSO_ENCAP_ALL; - dev->features |= dev->hw_features; - dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX; -} - -static int team_newlink(struct net *src_net, struct net_device *dev, - struct nlattr *tb[], struct nlattr *data[], - struct netlink_ext_ack *extack) -{ - if (tb[IFLA_ADDRESS] == NULL) - eth_hw_addr_random(dev); - - return register_netdevice(dev); -} - -static int team_validate(struct nlattr *tb[], struct nlattr *data[], - struct netlink_ext_ack *extack) -{ - if (tb[IFLA_ADDRESS]) { - if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN) - return -EINVAL; - if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS]))) - return -EADDRNOTAVAIL; - } - return 0; -} - -static unsigned int team_get_num_tx_queues(void) -{ - return TEAM_DEFAULT_NUM_TX_QUEUES; -} - -static unsigned int team_get_num_rx_queues(void) -{ - return TEAM_DEFAULT_NUM_RX_QUEUES; -} - -static struct rtnl_link_ops team_link_ops __read_mostly = { - .kind = DRV_NAME, - .priv_size = sizeof(struct team), - .setup = team_setup, - .newlink = team_newlink, - .validate = team_validate, - .get_num_tx_queues = team_get_num_tx_queues, - .get_num_rx_queues = team_get_num_rx_queues, -}; - - -/*********************************** - * Generic netlink custom interface - ***********************************/ - -static struct genl_family team_nl_family; - -static const struct nla_policy team_nl_policy[TEAM_ATTR_MAX + 1] = { - [TEAM_ATTR_UNSPEC] = { .type = NLA_UNSPEC, }, - [TEAM_ATTR_TEAM_IFINDEX] = { .type = NLA_U32 }, - [TEAM_ATTR_LIST_OPTION] = { .type = NLA_NESTED }, - [TEAM_ATTR_LIST_PORT] = { .type = NLA_NESTED }, -}; - -static const struct nla_policy -team_nl_option_policy[TEAM_ATTR_OPTION_MAX + 1] = { - [TEAM_ATTR_OPTION_UNSPEC] = { .type = NLA_UNSPEC, }, - [TEAM_ATTR_OPTION_NAME] = { - .type = NLA_STRING, - .len = TEAM_STRING_MAX_LEN, - }, - [TEAM_ATTR_OPTION_CHANGED] = { .type = NLA_FLAG }, - [TEAM_ATTR_OPTION_TYPE] = { .type = NLA_U8 }, - [TEAM_ATTR_OPTION_DATA] = { .type = NLA_BINARY }, - [TEAM_ATTR_OPTION_PORT_IFINDEX] = { .type = NLA_U32 }, - [TEAM_ATTR_OPTION_ARRAY_INDEX] = { .type = NLA_U32 }, -}; - -static int team_nl_cmd_noop(struct sk_buff *skb, struct genl_info *info) -{ - struct sk_buff *msg; - void *hdr; - int err; - - msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); - if (!msg) - return -ENOMEM; - - hdr = genlmsg_put(msg, info->snd_portid, info->snd_seq, - &team_nl_family, 0, TEAM_CMD_NOOP); - if (!hdr) { - err = -EMSGSIZE; - goto err_msg_put; - } - - genlmsg_end(msg, hdr); - - return genlmsg_unicast(genl_info_net(info), msg, info->snd_portid); - -err_msg_put: - nlmsg_free(msg); - - return err; -} - -/* - * Netlink cmd functions should be locked by following two functions. - * Since dev gets held here, that ensures dev won't disappear in between. - */ -static struct team *team_nl_team_get(struct genl_info *info) -{ - struct net *net = genl_info_net(info); - int ifindex; - struct net_device *dev; - struct team *team; - - if (!info->attrs[TEAM_ATTR_TEAM_IFINDEX]) - return NULL; - - ifindex = nla_get_u32(info->attrs[TEAM_ATTR_TEAM_IFINDEX]); - dev = dev_get_by_index(net, ifindex); - if (!dev || dev->netdev_ops != &team_netdev_ops) { - dev_put(dev); - return NULL; - } - - team = netdev_priv(dev); - mutex_lock(&team->lock); - return team; -} - -static void team_nl_team_put(struct team *team) -{ - mutex_unlock(&team->lock); - dev_put(team->dev); -} - -typedef int team_nl_send_func_t(struct sk_buff *skb, - struct team *team, u32 portid); - -static int team_nl_send_unicast(struct sk_buff *skb, struct team *team, u32 portid) -{ - return genlmsg_unicast(dev_net(team->dev), skb, portid); -} - -static int team_nl_fill_one_option_get(struct sk_buff *skb, struct team *team, - struct team_option_inst *opt_inst) -{ - struct nlattr *option_item; - struct team_option *option = opt_inst->option; - struct team_option_inst_info *opt_inst_info = &opt_inst->info; - struct team_gsetter_ctx ctx; - int err; - - ctx.info = opt_inst_info; - err = team_option_get(team, opt_inst, &ctx); - if (err) - return err; - - option_item = nla_nest_start_noflag(skb, TEAM_ATTR_ITEM_OPTION); - if (!option_item) - return -EMSGSIZE; - - if (nla_put_string(skb, TEAM_ATTR_OPTION_NAME, option->name)) - goto nest_cancel; - if (opt_inst_info->port && - nla_put_u32(skb, TEAM_ATTR_OPTION_PORT_IFINDEX, - opt_inst_info->port->dev->ifindex)) - goto nest_cancel; - if (opt_inst->option->array_size && - nla_put_u32(skb, TEAM_ATTR_OPTION_ARRAY_INDEX, - opt_inst_info->array_index)) - goto nest_cancel; - - switch (option->type) { - case TEAM_OPTION_TYPE_U32: - if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_U32)) - goto nest_cancel; - if (nla_put_u32(skb, TEAM_ATTR_OPTION_DATA, ctx.data.u32_val)) - goto nest_cancel; - break; - case TEAM_OPTION_TYPE_STRING: - if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_STRING)) - goto nest_cancel; - if (nla_put_string(skb, TEAM_ATTR_OPTION_DATA, - ctx.data.str_val)) - goto nest_cancel; - break; - case TEAM_OPTION_TYPE_BINARY: - if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_BINARY)) - goto nest_cancel; - if (nla_put(skb, TEAM_ATTR_OPTION_DATA, ctx.data.bin_val.len, - ctx.data.bin_val.ptr)) - goto nest_cancel; - break; - case TEAM_OPTION_TYPE_BOOL: - if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_FLAG)) - goto nest_cancel; - if (ctx.data.bool_val && - nla_put_flag(skb, TEAM_ATTR_OPTION_DATA)) - goto nest_cancel; - break; - case TEAM_OPTION_TYPE_S32: - if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_S32)) - goto nest_cancel; - if (nla_put_s32(skb, TEAM_ATTR_OPTION_DATA, ctx.data.s32_val)) - goto nest_cancel; - break; - default: - BUG(); - } - if (opt_inst->removed && nla_put_flag(skb, TEAM_ATTR_OPTION_REMOVED)) - goto nest_cancel; - if (opt_inst->changed) { - if (nla_put_flag(skb, TEAM_ATTR_OPTION_CHANGED)) - goto nest_cancel; - opt_inst->changed = false; - } - nla_nest_end(skb, option_item); - return 0; - -nest_cancel: - nla_nest_cancel(skb, option_item); - return -EMSGSIZE; -} - -static int __send_and_alloc_skb(struct sk_buff **pskb, - struct team *team, u32 portid, - team_nl_send_func_t *send_func) -{ - int err; - - if (*pskb) { - err = send_func(*pskb, team, portid); - if (err) - return err; - } - *pskb = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_KERNEL); - if (!*pskb) - return -ENOMEM; - return 0; -} - -static int team_nl_send_options_get(struct team *team, u32 portid, u32 seq, - int flags, team_nl_send_func_t *send_func, - struct list_head *sel_opt_inst_list) -{ - struct nlattr *option_list; - struct nlmsghdr *nlh; - void *hdr; - struct team_option_inst *opt_inst; - int err; - struct sk_buff *skb = NULL; - bool incomplete; - int i; - - opt_inst = list_first_entry(sel_opt_inst_list, - struct team_option_inst, tmp_list); - -start_again: - err = __send_and_alloc_skb(&skb, team, portid, send_func); - if (err) - return err; - - hdr = genlmsg_put(skb, portid, seq, &team_nl_family, flags | NLM_F_MULTI, - TEAM_CMD_OPTIONS_GET); - if (!hdr) { - nlmsg_free(skb); - return -EMSGSIZE; - } - - if (nla_put_u32(skb, TEAM_ATTR_TEAM_IFINDEX, team->dev->ifindex)) - goto nla_put_failure; - option_list = nla_nest_start_noflag(skb, TEAM_ATTR_LIST_OPTION); - if (!option_list) - goto nla_put_failure; - - i = 0; - incomplete = false; - list_for_each_entry_from(opt_inst, sel_opt_inst_list, tmp_list) { - err = team_nl_fill_one_option_get(skb, team, opt_inst); - if (err) { - if (err == -EMSGSIZE) { - if (!i) - goto errout; - incomplete = true; - break; - } - goto errout; - } - i++; - } - - nla_nest_end(skb, option_list); - genlmsg_end(skb, hdr); - if (incomplete) - goto start_again; - -send_done: - nlh = nlmsg_put(skb, portid, seq, NLMSG_DONE, 0, flags | NLM_F_MULTI); - if (!nlh) { - err = __send_and_alloc_skb(&skb, team, portid, send_func); - if (err) - return err; - goto send_done; - } - - return send_func(skb, team, portid); - -nla_put_failure: - err = -EMSGSIZE; -errout: - nlmsg_free(skb); - return err; -} - -static int team_nl_cmd_options_get(struct sk_buff *skb, struct genl_info *info) -{ - struct team *team; - struct team_option_inst *opt_inst; - int err; - LIST_HEAD(sel_opt_inst_list); - - team = team_nl_team_get(info); - if (!team) - return -EINVAL; - - list_for_each_entry(opt_inst, &team->option_inst_list, list) - list_add_tail(&opt_inst->tmp_list, &sel_opt_inst_list); - err = team_nl_send_options_get(team, info->snd_portid, info->snd_seq, - NLM_F_ACK, team_nl_send_unicast, - &sel_opt_inst_list); - - team_nl_team_put(team); - - return err; -} - -static int team_nl_send_event_options_get(struct team *team, - struct list_head *sel_opt_inst_list); - -static int team_nl_cmd_options_set(struct sk_buff *skb, struct genl_info *info) -{ - struct team *team; - int err = 0; - int i; - struct nlattr *nl_option; - - rtnl_lock(); - - team = team_nl_team_get(info); - if (!team) { - err = -EINVAL; - goto rtnl_unlock; - } - - err = -EINVAL; - if (!info->attrs[TEAM_ATTR_LIST_OPTION]) { - err = -EINVAL; - goto team_put; - } - - nla_for_each_nested(nl_option, info->attrs[TEAM_ATTR_LIST_OPTION], i) { - struct nlattr *opt_attrs[TEAM_ATTR_OPTION_MAX + 1]; - struct nlattr *attr; - struct nlattr *attr_data; - LIST_HEAD(opt_inst_list); - enum team_option_type opt_type; - int opt_port_ifindex = 0; /* != 0 for per-port options */ - u32 opt_array_index = 0; - bool opt_is_array = false; - struct team_option_inst *opt_inst; - char *opt_name; - bool opt_found = false; - - if (nla_type(nl_option) != TEAM_ATTR_ITEM_OPTION) { - err = -EINVAL; - goto team_put; - } - err = nla_parse_nested_deprecated(opt_attrs, - TEAM_ATTR_OPTION_MAX, - nl_option, - team_nl_option_policy, - info->extack); - if (err) - goto team_put; - if (!opt_attrs[TEAM_ATTR_OPTION_NAME] || - !opt_attrs[TEAM_ATTR_OPTION_TYPE]) { - err = -EINVAL; - goto team_put; - } - switch (nla_get_u8(opt_attrs[TEAM_ATTR_OPTION_TYPE])) { - case NLA_U32: - opt_type = TEAM_OPTION_TYPE_U32; - break; - case NLA_STRING: - opt_type = TEAM_OPTION_TYPE_STRING; - break; - case NLA_BINARY: - opt_type = TEAM_OPTION_TYPE_BINARY; - break; - case NLA_FLAG: - opt_type = TEAM_OPTION_TYPE_BOOL; - break; - case NLA_S32: - opt_type = TEAM_OPTION_TYPE_S32; - break; - default: - goto team_put; - } - - attr_data = opt_attrs[TEAM_ATTR_OPTION_DATA]; - if (opt_type != TEAM_OPTION_TYPE_BOOL && !attr_data) { - err = -EINVAL; - goto team_put; - } - - opt_name = nla_data(opt_attrs[TEAM_ATTR_OPTION_NAME]); - attr = opt_attrs[TEAM_ATTR_OPTION_PORT_IFINDEX]; - if (attr) - opt_port_ifindex = nla_get_u32(attr); - - attr = opt_attrs[TEAM_ATTR_OPTION_ARRAY_INDEX]; - if (attr) { - opt_is_array = true; - opt_array_index = nla_get_u32(attr); - } - - list_for_each_entry(opt_inst, &team->option_inst_list, list) { - struct team_option *option = opt_inst->option; - struct team_gsetter_ctx ctx; - struct team_option_inst_info *opt_inst_info; - int tmp_ifindex; - - opt_inst_info = &opt_inst->info; - tmp_ifindex = opt_inst_info->port ? - opt_inst_info->port->dev->ifindex : 0; - if (option->type != opt_type || - strcmp(option->name, opt_name) || - tmp_ifindex != opt_port_ifindex || - (option->array_size && !opt_is_array) || - opt_inst_info->array_index != opt_array_index) - continue; - opt_found = true; - ctx.info = opt_inst_info; - switch (opt_type) { - case TEAM_OPTION_TYPE_U32: - ctx.data.u32_val = nla_get_u32(attr_data); - break; - case TEAM_OPTION_TYPE_STRING: - if (nla_len(attr_data) > TEAM_STRING_MAX_LEN) { - err = -EINVAL; - goto team_put; - } - ctx.data.str_val = nla_data(attr_data); - break; - case TEAM_OPTION_TYPE_BINARY: - ctx.data.bin_val.len = nla_len(attr_data); - ctx.data.bin_val.ptr = nla_data(attr_data); - break; - case TEAM_OPTION_TYPE_BOOL: - ctx.data.bool_val = attr_data ? true : false; - break; - case TEAM_OPTION_TYPE_S32: - ctx.data.s32_val = nla_get_s32(attr_data); - break; - default: - BUG(); - } - err = team_option_set(team, opt_inst, &ctx); - if (err) - goto team_put; - opt_inst->changed = true; - list_add(&opt_inst->tmp_list, &opt_inst_list); - } - if (!opt_found) { - err = -ENOENT; - goto team_put; - } - - err = team_nl_send_event_options_get(team, &opt_inst_list); - if (err) - break; - } - -team_put: - team_nl_team_put(team); -rtnl_unlock: - rtnl_unlock(); - return err; -} - -static int team_nl_fill_one_port_get(struct sk_buff *skb, - struct team_port *port) -{ - struct nlattr *port_item; - - port_item = nla_nest_start_noflag(skb, TEAM_ATTR_ITEM_PORT); - if (!port_item) - goto nest_cancel; - if (nla_put_u32(skb, TEAM_ATTR_PORT_IFINDEX, port->dev->ifindex)) - goto nest_cancel; - if (port->changed) { - if (nla_put_flag(skb, TEAM_ATTR_PORT_CHANGED)) - goto nest_cancel; - port->changed = false; - } - if ((port->removed && - nla_put_flag(skb, TEAM_ATTR_PORT_REMOVED)) || - (port->state.linkup && - nla_put_flag(skb, TEAM_ATTR_PORT_LINKUP)) || - nla_put_u32(skb, TEAM_ATTR_PORT_SPEED, port->state.speed) || - nla_put_u8(skb, TEAM_ATTR_PORT_DUPLEX, port->state.duplex)) - goto nest_cancel; - nla_nest_end(skb, port_item); - return 0; - -nest_cancel: - nla_nest_cancel(skb, port_item); - return -EMSGSIZE; -} - -static int team_nl_send_port_list_get(struct team *team, u32 portid, u32 seq, - int flags, team_nl_send_func_t *send_func, - struct team_port *one_port) -{ - struct nlattr *port_list; - struct nlmsghdr *nlh; - void *hdr; - struct team_port *port; - int err; - struct sk_buff *skb = NULL; - bool incomplete; - int i; - - port = list_first_entry_or_null(&team->port_list, - struct team_port, list); - -start_again: - err = __send_and_alloc_skb(&skb, team, portid, send_func); - if (err) - return err; - - hdr = genlmsg_put(skb, portid, seq, &team_nl_family, flags | NLM_F_MULTI, - TEAM_CMD_PORT_LIST_GET); - if (!hdr) { - nlmsg_free(skb); - return -EMSGSIZE; - } - - if (nla_put_u32(skb, TEAM_ATTR_TEAM_IFINDEX, team->dev->ifindex)) - goto nla_put_failure; - port_list = nla_nest_start_noflag(skb, TEAM_ATTR_LIST_PORT); - if (!port_list) - goto nla_put_failure; - - i = 0; - incomplete = false; - - /* If one port is selected, called wants to send port list containing - * only this port. Otherwise go through all listed ports and send all - */ - if (one_port) { - err = team_nl_fill_one_port_get(skb, one_port); - if (err) - goto errout; - } else if (port) { - list_for_each_entry_from(port, &team->port_list, list) { - err = team_nl_fill_one_port_get(skb, port); - if (err) { - if (err == -EMSGSIZE) { - if (!i) - goto errout; - incomplete = true; - break; - } - goto errout; - } - i++; - } - } - - nla_nest_end(skb, port_list); - genlmsg_end(skb, hdr); - if (incomplete) - goto start_again; - -send_done: - nlh = nlmsg_put(skb, portid, seq, NLMSG_DONE, 0, flags | NLM_F_MULTI); - if (!nlh) { - err = __send_and_alloc_skb(&skb, team, portid, send_func); - if (err) - return err; - goto send_done; - } - - return send_func(skb, team, portid); - -nla_put_failure: - err = -EMSGSIZE; -errout: - nlmsg_free(skb); - return err; -} - -static int team_nl_cmd_port_list_get(struct sk_buff *skb, - struct genl_info *info) -{ - struct team *team; - int err; - - team = team_nl_team_get(info); - if (!team) - return -EINVAL; - - err = team_nl_send_port_list_get(team, info->snd_portid, info->snd_seq, - NLM_F_ACK, team_nl_send_unicast, NULL); - - team_nl_team_put(team); - - return err; -} - -static const struct genl_small_ops team_nl_ops[] = { - { - .cmd = TEAM_CMD_NOOP, - .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, - .doit = team_nl_cmd_noop, - }, - { - .cmd = TEAM_CMD_OPTIONS_SET, - .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, - .doit = team_nl_cmd_options_set, - .flags = GENL_ADMIN_PERM, - }, - { - .cmd = TEAM_CMD_OPTIONS_GET, - .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, - .doit = team_nl_cmd_options_get, - .flags = GENL_ADMIN_PERM, - }, - { - .cmd = TEAM_CMD_PORT_LIST_GET, - .validate = GENL_DONT_VALIDATE_STRICT | GENL_DONT_VALIDATE_DUMP, - .doit = team_nl_cmd_port_list_get, - .flags = GENL_ADMIN_PERM, - }, -}; - -static const struct genl_multicast_group team_nl_mcgrps[] = { - { .name = TEAM_GENL_CHANGE_EVENT_MC_GRP_NAME, }, -}; - -static struct genl_family team_nl_family __ro_after_init = { - .name = TEAM_GENL_NAME, - .version = TEAM_GENL_VERSION, - .maxattr = TEAM_ATTR_MAX, - .policy = team_nl_policy, - .netnsok = true, - .module = THIS_MODULE, - .small_ops = team_nl_ops, - .n_small_ops = ARRAY_SIZE(team_nl_ops), - .resv_start_op = TEAM_CMD_PORT_LIST_GET + 1, - .mcgrps = team_nl_mcgrps, - .n_mcgrps = ARRAY_SIZE(team_nl_mcgrps), -}; - -static int team_nl_send_multicast(struct sk_buff *skb, - struct team *team, u32 portid) -{ - return genlmsg_multicast_netns(&team_nl_family, dev_net(team->dev), - skb, 0, 0, GFP_KERNEL); -} - -static int team_nl_send_event_options_get(struct team *team, - struct list_head *sel_opt_inst_list) -{ - return team_nl_send_options_get(team, 0, 0, 0, team_nl_send_multicast, - sel_opt_inst_list); -} - -static int team_nl_send_event_port_get(struct team *team, - struct team_port *port) -{ - return team_nl_send_port_list_get(team, 0, 0, 0, team_nl_send_multicast, - port); -} - -static int __init team_nl_init(void) -{ - return genl_register_family(&team_nl_family); -} - -static void __exit team_nl_fini(void) -{ - genl_unregister_family(&team_nl_family); -} - - -/****************** - * Change checkers - ******************/ - -static void __team_options_change_check(struct team *team) -{ - int err; - struct team_option_inst *opt_inst; - LIST_HEAD(sel_opt_inst_list); - - list_for_each_entry(opt_inst, &team->option_inst_list, list) { - if (opt_inst->changed) - list_add_tail(&opt_inst->tmp_list, &sel_opt_inst_list); - } - err = team_nl_send_event_options_get(team, &sel_opt_inst_list); - if (err && err != -ESRCH) - netdev_warn(team->dev, "Failed to send options change via netlink (err %d)\n", - err); -} - -/* rtnl lock is held */ - -static void __team_port_change_send(struct team_port *port, bool linkup) -{ - int err; - - port->changed = true; - port->state.linkup = linkup; - team_refresh_port_linkup(port); - if (linkup) { - struct ethtool_link_ksettings ecmd; - - err = __ethtool_get_link_ksettings(port->dev, &ecmd); - if (!err) { - port->state.speed = ecmd.base.speed; - port->state.duplex = ecmd.base.duplex; - goto send_event; - } - } - port->state.speed = 0; - port->state.duplex = 0; - -send_event: - err = team_nl_send_event_port_get(port->team, port); - if (err && err != -ESRCH) - netdev_warn(port->team->dev, "Failed to send port change of device %s via netlink (err %d)\n", - port->dev->name, err); - -} - -static void __team_carrier_check(struct team *team) -{ - struct team_port *port; - bool team_linkup; - - if (team->user_carrier_enabled) - return; - - team_linkup = false; - list_for_each_entry(port, &team->port_list, list) { - if (port->linkup) { - team_linkup = true; - break; - } - } - - if (team_linkup) - netif_carrier_on(team->dev); - else - netif_carrier_off(team->dev); -} - -static void __team_port_change_check(struct team_port *port, bool linkup) -{ - if (port->state.linkup != linkup) - __team_port_change_send(port, linkup); - __team_carrier_check(port->team); -} - -static void __team_port_change_port_added(struct team_port *port, bool linkup) -{ - __team_port_change_send(port, linkup); - __team_carrier_check(port->team); -} - -static void __team_port_change_port_removed(struct team_port *port) -{ - port->removed = true; - __team_port_change_send(port, false); - __team_carrier_check(port->team); -} - -static void team_port_change_check(struct team_port *port, bool linkup) -{ - struct team *team = port->team; - - mutex_lock(&team->lock); - __team_port_change_check(port, linkup); - mutex_unlock(&team->lock); -} - - -/************************************ - * Net device notifier event handler - ************************************/ - -static int team_device_event(struct notifier_block *unused, - unsigned long event, void *ptr) -{ - struct net_device *dev = netdev_notifier_info_to_dev(ptr); - struct team_port *port; - - port = team_port_get_rtnl(dev); - if (!port) - return NOTIFY_DONE; - - switch (event) { - case NETDEV_UP: - if (netif_oper_up(dev)) - team_port_change_check(port, true); - break; - case NETDEV_DOWN: - team_port_change_check(port, false); - break; - case NETDEV_CHANGE: - if (netif_running(port->dev)) - team_port_change_check(port, - !!netif_oper_up(port->dev)); - break; - case NETDEV_UNREGISTER: - team_del_slave(port->team->dev, dev); - break; - case NETDEV_FEAT_CHANGE: - if (!port->team->notifier_ctx) { - port->team->notifier_ctx = true; - team_compute_features(port->team); - port->team->notifier_ctx = false; - } - break; - case NETDEV_PRECHANGEMTU: - /* Forbid to change mtu of underlaying device */ - if (!port->team->port_mtu_change_allowed) - return NOTIFY_BAD; - break; - case NETDEV_PRE_TYPE_CHANGE: - /* Forbid to change type of underlaying device */ - return NOTIFY_BAD; - case NETDEV_RESEND_IGMP: - /* Propagate to master device */ - call_netdevice_notifiers(event, port->team->dev); - break; - } - return NOTIFY_DONE; -} - -static struct notifier_block team_notifier_block __read_mostly = { - .notifier_call = team_device_event, -}; - - -/*********************** - * Module init and exit - ***********************/ - -static int __init team_module_init(void) -{ - int err; - - register_netdevice_notifier(&team_notifier_block); - - err = rtnl_link_register(&team_link_ops); - if (err) - goto err_rtnl_reg; - - err = team_nl_init(); - if (err) - goto err_nl_init; - - return 0; - -err_nl_init: - rtnl_link_unregister(&team_link_ops); - -err_rtnl_reg: - unregister_netdevice_notifier(&team_notifier_block); - - return err; -} - -static void __exit team_module_exit(void) -{ - team_nl_fini(); - rtnl_link_unregister(&team_link_ops); - unregister_netdevice_notifier(&team_notifier_block); -} - -module_init(team_module_init); -module_exit(team_module_exit); - -MODULE_LICENSE("GPL v2"); -MODULE_AUTHOR("Jiri Pirko "); -MODULE_DESCRIPTION("Ethernet team device driver"); -MODULE_ALIAS_RTNL_LINK(DRV_NAME); diff --git a/drivers/net/team/team_core.c b/drivers/net/team/team_core.c new file mode 100644 index 0000000000..ab1935a4aa --- /dev/null +++ b/drivers/net/team/team_core.c @@ -0,0 +1,3056 @@ +// SPDX-License-Identifier: GPL-2.0-or-later +/* + * drivers/net/team/team.c - Network team device driver + * Copyright (c) 2011 Jiri Pirko + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "team_nl.h" + +#define DRV_NAME "team" + + +/********** + * Helpers + **********/ + +static struct team_port *team_port_get_rtnl(const struct net_device *dev) +{ + struct team_port *port = rtnl_dereference(dev->rx_handler_data); + + return netif_is_team_port(dev) ? port : NULL; +} + +/* + * Since the ability to change device address for open port device is tested in + * team_port_add, this function can be called without control of return value + */ +static int __set_port_dev_addr(struct net_device *port_dev, + const unsigned char *dev_addr) +{ + struct sockaddr_storage addr; + + memcpy(addr.__data, dev_addr, port_dev->addr_len); + addr.ss_family = port_dev->type; + return dev_set_mac_address(port_dev, (struct sockaddr *)&addr, NULL); +} + +static int team_port_set_orig_dev_addr(struct team_port *port) +{ + return __set_port_dev_addr(port->dev, port->orig.dev_addr); +} + +static int team_port_set_team_dev_addr(struct team *team, + struct team_port *port) +{ + return __set_port_dev_addr(port->dev, team->dev->dev_addr); +} + +int team_modeop_port_enter(struct team *team, struct team_port *port) +{ + return team_port_set_team_dev_addr(team, port); +} +EXPORT_SYMBOL(team_modeop_port_enter); + +void team_modeop_port_change_dev_addr(struct team *team, + struct team_port *port) +{ + team_port_set_team_dev_addr(team, port); +} +EXPORT_SYMBOL(team_modeop_port_change_dev_addr); + +static void team_lower_state_changed(struct team_port *port) +{ + struct netdev_lag_lower_state_info info; + + info.link_up = port->linkup; + info.tx_enabled = team_port_enabled(port); + netdev_lower_state_changed(port->dev, &info); +} + +static void team_refresh_port_linkup(struct team_port *port) +{ + bool new_linkup = port->user.linkup_enabled ? port->user.linkup : + port->state.linkup; + + if (port->linkup != new_linkup) { + port->linkup = new_linkup; + team_lower_state_changed(port); + } +} + + +/******************* + * Options handling + *******************/ + +struct team_option_inst { /* One for each option instance */ + struct list_head list; + struct list_head tmp_list; + struct team_option *option; + struct team_option_inst_info info; + bool changed; + bool removed; +}; + +static struct team_option *__team_find_option(struct team *team, + const char *opt_name) +{ + struct team_option *option; + + list_for_each_entry(option, &team->option_list, list) { + if (strcmp(option->name, opt_name) == 0) + return option; + } + return NULL; +} + +static void __team_option_inst_del(struct team_option_inst *opt_inst) +{ + list_del(&opt_inst->list); + kfree(opt_inst); +} + +static void __team_option_inst_del_option(struct team *team, + struct team_option *option) +{ + struct team_option_inst *opt_inst, *tmp; + + list_for_each_entry_safe(opt_inst, tmp, &team->option_inst_list, list) { + if (opt_inst->option == option) + __team_option_inst_del(opt_inst); + } +} + +static int __team_option_inst_add(struct team *team, struct team_option *option, + struct team_port *port) +{ + struct team_option_inst *opt_inst; + unsigned int array_size; + unsigned int i; + + array_size = option->array_size; + if (!array_size) + array_size = 1; /* No array but still need one instance */ + + for (i = 0; i < array_size; i++) { + opt_inst = kmalloc(sizeof(*opt_inst), GFP_KERNEL); + if (!opt_inst) + return -ENOMEM; + opt_inst->option = option; + opt_inst->info.port = port; + opt_inst->info.array_index = i; + opt_inst->changed = true; + opt_inst->removed = false; + list_add_tail(&opt_inst->list, &team->option_inst_list); + if (option->init) + option->init(team, &opt_inst->info); + + } + return 0; +} + +static int __team_option_inst_add_option(struct team *team, + struct team_option *option) +{ + int err; + + if (!option->per_port) { + err = __team_option_inst_add(team, option, NULL); + if (err) + goto inst_del_option; + } + return 0; + +inst_del_option: + __team_option_inst_del_option(team, option); + return err; +} + +static void __team_option_inst_mark_removed_option(struct team *team, + struct team_option *option) +{ + struct team_option_inst *opt_inst; + + list_for_each_entry(opt_inst, &team->option_inst_list, list) { + if (opt_inst->option == option) { + opt_inst->changed = true; + opt_inst->removed = true; + } + } +} + +static void __team_option_inst_del_port(struct team *team, + struct team_port *port) +{ + struct team_option_inst *opt_inst, *tmp; + + list_for_each_entry_safe(opt_inst, tmp, &team->option_inst_list, list) { + if (opt_inst->option->per_port && + opt_inst->info.port == port) + __team_option_inst_del(opt_inst); + } +} + +static int __team_option_inst_add_port(struct team *team, + struct team_port *port) +{ + struct team_option *option; + int err; + + list_for_each_entry(option, &team->option_list, list) { + if (!option->per_port) + continue; + err = __team_option_inst_add(team, option, port); + if (err) + goto inst_del_port; + } + return 0; + +inst_del_port: + __team_option_inst_del_port(team, port); + return err; +} + +static void __team_option_inst_mark_removed_port(struct team *team, + struct team_port *port) +{ + struct team_option_inst *opt_inst; + + list_for_each_entry(opt_inst, &team->option_inst_list, list) { + if (opt_inst->info.port == port) { + opt_inst->changed = true; + opt_inst->removed = true; + } + } +} + +static int __team_options_register(struct team *team, + const struct team_option *option, + size_t option_count) +{ + int i; + struct team_option **dst_opts; + int err; + + dst_opts = kcalloc(option_count, sizeof(struct team_option *), + GFP_KERNEL); + if (!dst_opts) + return -ENOMEM; + for (i = 0; i < option_count; i++, option++) { + if (__team_find_option(team, option->name)) { + err = -EEXIST; + goto alloc_rollback; + } + dst_opts[i] = kmemdup(option, sizeof(*option), GFP_KERNEL); + if (!dst_opts[i]) { + err = -ENOMEM; + goto alloc_rollback; + } + } + + for (i = 0; i < option_count; i++) { + err = __team_option_inst_add_option(team, dst_opts[i]); + if (err) + goto inst_rollback; + list_add_tail(&dst_opts[i]->list, &team->option_list); + } + + kfree(dst_opts); + return 0; + +inst_rollback: + for (i--; i >= 0; i--) { + __team_option_inst_del_option(team, dst_opts[i]); + list_del(&dst_opts[i]->list); + } + + i = option_count; +alloc_rollback: + for (i--; i >= 0; i--) + kfree(dst_opts[i]); + + kfree(dst_opts); + return err; +} + +static void __team_options_mark_removed(struct team *team, + const struct team_option *option, + size_t option_count) +{ + int i; + + for (i = 0; i < option_count; i++, option++) { + struct team_option *del_opt; + + del_opt = __team_find_option(team, option->name); + if (del_opt) + __team_option_inst_mark_removed_option(team, del_opt); + } +} + +static void __team_options_unregister(struct team *team, + const struct team_option *option, + size_t option_count) +{ + int i; + + for (i = 0; i < option_count; i++, option++) { + struct team_option *del_opt; + + del_opt = __team_find_option(team, option->name); + if (del_opt) { + __team_option_inst_del_option(team, del_opt); + list_del(&del_opt->list); + kfree(del_opt); + } + } +} + +static void __team_options_change_check(struct team *team); + +int team_options_register(struct team *team, + const struct team_option *option, + size_t option_count) +{ + int err; + + err = __team_options_register(team, option, option_count); + if (err) + return err; + __team_options_change_check(team); + return 0; +} +EXPORT_SYMBOL(team_options_register); + +void team_options_unregister(struct team *team, + const struct team_option *option, + size_t option_count) +{ + __team_options_mark_removed(team, option, option_count); + __team_options_change_check(team); + __team_options_unregister(team, option, option_count); +} +EXPORT_SYMBOL(team_options_unregister); + +static int team_option_get(struct team *team, + struct team_option_inst *opt_inst, + struct team_gsetter_ctx *ctx) +{ + if (!opt_inst->option->getter) + return -EOPNOTSUPP; + + opt_inst->option->getter(team, ctx); + return 0; +} + +static int team_option_set(struct team *team, + struct team_option_inst *opt_inst, + struct team_gsetter_ctx *ctx) +{ + if (!opt_inst->option->setter) + return -EOPNOTSUPP; + return opt_inst->option->setter(team, ctx); +} + +void team_option_inst_set_change(struct team_option_inst_info *opt_inst_info) +{ + struct team_option_inst *opt_inst; + + opt_inst = container_of(opt_inst_info, struct team_option_inst, info); + opt_inst->changed = true; +} +EXPORT_SYMBOL(team_option_inst_set_change); + +void team_options_change_check(struct team *team) +{ + __team_options_change_check(team); +} +EXPORT_SYMBOL(team_options_change_check); + + +/**************** + * Mode handling + ****************/ + +static LIST_HEAD(mode_list); +static DEFINE_SPINLOCK(mode_list_lock); + +struct team_mode_item { + struct list_head list; + const struct team_mode *mode; +}; + +static struct team_mode_item *__find_mode(const char *kind) +{ + struct team_mode_item *mitem; + + list_for_each_entry(mitem, &mode_list, list) { + if (strcmp(mitem->mode->kind, kind) == 0) + return mitem; + } + return NULL; +} + +static bool is_good_mode_name(const char *name) +{ + while (*name != '\0') { + if (!isalpha(*name) && !isdigit(*name) && *name != '_') + return false; + name++; + } + return true; +} + +int team_mode_register(const struct team_mode *mode) +{ + int err = 0; + struct team_mode_item *mitem; + + if (!is_good_mode_name(mode->kind) || + mode->priv_size > TEAM_MODE_PRIV_SIZE) + return -EINVAL; + + mitem = kmalloc(sizeof(*mitem), GFP_KERNEL); + if (!mitem) + return -ENOMEM; + + spin_lock(&mode_list_lock); + if (__find_mode(mode->kind)) { + err = -EEXIST; + kfree(mitem); + goto unlock; + } + mitem->mode = mode; + list_add_tail(&mitem->list, &mode_list); +unlock: + spin_unlock(&mode_list_lock); + return err; +} +EXPORT_SYMBOL(team_mode_register); + +void team_mode_unregister(const struct team_mode *mode) +{ + struct team_mode_item *mitem; + + spin_lock(&mode_list_lock); + mitem = __find_mode(mode->kind); + if (mitem) { + list_del_init(&mitem->list); + kfree(mitem); + } + spin_unlock(&mode_list_lock); +} +EXPORT_SYMBOL(team_mode_unregister); + +static const struct team_mode *team_mode_get(const char *kind) +{ + struct team_mode_item *mitem; + const struct team_mode *mode = NULL; + + if (!try_module_get(THIS_MODULE)) + return NULL; + + spin_lock(&mode_list_lock); + mitem = __find_mode(kind); + if (!mitem) { + spin_unlock(&mode_list_lock); + request_module("team-mode-%s", kind); + spin_lock(&mode_list_lock); + mitem = __find_mode(kind); + } + if (mitem) { + mode = mitem->mode; + if (!try_module_get(mode->owner)) + mode = NULL; + } + + spin_unlock(&mode_list_lock); + module_put(THIS_MODULE); + return mode; +} + +static void team_mode_put(const struct team_mode *mode) +{ + module_put(mode->owner); +} + +static bool team_dummy_transmit(struct team *team, struct sk_buff *skb) +{ + dev_kfree_skb_any(skb); + return false; +} + +static rx_handler_result_t team_dummy_receive(struct team *team, + struct team_port *port, + struct sk_buff *skb) +{ + return RX_HANDLER_ANOTHER; +} + +static const struct team_mode __team_no_mode = { + .kind = "*NOMODE*", +}; + +static bool team_is_mode_set(struct team *team) +{ + return team->mode != &__team_no_mode; +} + +static void team_set_no_mode(struct team *team) +{ + team->user_carrier_enabled = false; + team->mode = &__team_no_mode; +} + +static void team_adjust_ops(struct team *team) +{ + /* + * To avoid checks in rx/tx skb paths, ensure here that non-null and + * correct ops are always set. + */ + + if (!team->en_port_count || !team_is_mode_set(team) || + !team->mode->ops->transmit) + team->ops.transmit = team_dummy_transmit; + else + team->ops.transmit = team->mode->ops->transmit; + + if (!team->en_port_count || !team_is_mode_set(team) || + !team->mode->ops->receive) + team->ops.receive = team_dummy_receive; + else + team->ops.receive = team->mode->ops->receive; +} + +/* + * We can benefit from the fact that it's ensured no port is present + * at the time of mode change. Therefore no packets are in fly so there's no + * need to set mode operations in any special way. + */ +static int __team_change_mode(struct team *team, + const struct team_mode *new_mode) +{ + /* Check if mode was previously set and do cleanup if so */ + if (team_is_mode_set(team)) { + void (*exit_op)(struct team *team) = team->ops.exit; + + /* Clear ops area so no callback is called any longer */ + memset(&team->ops, 0, sizeof(struct team_mode_ops)); + team_adjust_ops(team); + + if (exit_op) + exit_op(team); + team_mode_put(team->mode); + team_set_no_mode(team); + /* zero private data area */ + memset(&team->mode_priv, 0, + sizeof(struct team) - offsetof(struct team, mode_priv)); + } + + if (!new_mode) + return 0; + + if (new_mode->ops->init) { + int err; + + err = new_mode->ops->init(team); + if (err) + return err; + } + + team->mode = new_mode; + memcpy(&team->ops, new_mode->ops, sizeof(struct team_mode_ops)); + team_adjust_ops(team); + + return 0; +} + +static int team_change_mode(struct team *team, const char *kind) +{ + const struct team_mode *new_mode; + struct net_device *dev = team->dev; + int err; + + if (!list_empty(&team->port_list)) { + netdev_err(dev, "No ports can be present during mode change\n"); + return -EBUSY; + } + + if (team_is_mode_set(team) && strcmp(team->mode->kind, kind) == 0) { + netdev_err(dev, "Unable to change to the same mode the team is in\n"); + return -EINVAL; + } + + new_mode = team_mode_get(kind); + if (!new_mode) { + netdev_err(dev, "Mode \"%s\" not found\n", kind); + return -EINVAL; + } + + err = __team_change_mode(team, new_mode); + if (err) { + netdev_err(dev, "Failed to change to mode \"%s\"\n", kind); + team_mode_put(new_mode); + return err; + } + + netdev_info(dev, "Mode changed to \"%s\"\n", kind); + return 0; +} + + +/********************* + * Peers notification + *********************/ + +static void team_notify_peers_work(struct work_struct *work) +{ + struct team *team; + int val; + + team = container_of(work, struct team, notify_peers.dw.work); + + if (!rtnl_trylock()) { + schedule_delayed_work(&team->notify_peers.dw, 0); + return; + } + val = atomic_dec_if_positive(&team->notify_peers.count_pending); + if (val < 0) { + rtnl_unlock(); + return; + } + call_netdevice_notifiers(NETDEV_NOTIFY_PEERS, team->dev); + rtnl_unlock(); + if (val) + schedule_delayed_work(&team->notify_peers.dw, + msecs_to_jiffies(team->notify_peers.interval)); +} + +static void team_notify_peers(struct team *team) +{ + if (!team->notify_peers.count || !netif_running(team->dev)) + return; + atomic_add(team->notify_peers.count, &team->notify_peers.count_pending); + schedule_delayed_work(&team->notify_peers.dw, 0); +} + +static void team_notify_peers_init(struct team *team) +{ + INIT_DELAYED_WORK(&team->notify_peers.dw, team_notify_peers_work); +} + +static void team_notify_peers_fini(struct team *team) +{ + cancel_delayed_work_sync(&team->notify_peers.dw); +} + + +/******************************* + * Send multicast group rejoins + *******************************/ + +static void team_mcast_rejoin_work(struct work_struct *work) +{ + struct team *team; + int val; + + team = container_of(work, struct team, mcast_rejoin.dw.work); + + if (!rtnl_trylock()) { + schedule_delayed_work(&team->mcast_rejoin.dw, 0); + return; + } + val = atomic_dec_if_positive(&team->mcast_rejoin.count_pending); + if (val < 0) { + rtnl_unlock(); + return; + } + call_netdevice_notifiers(NETDEV_RESEND_IGMP, team->dev); + rtnl_unlock(); + if (val) + schedule_delayed_work(&team->mcast_rejoin.dw, + msecs_to_jiffies(team->mcast_rejoin.interval)); +} + +static void team_mcast_rejoin(struct team *team) +{ + if (!team->mcast_rejoin.count || !netif_running(team->dev)) + return; + atomic_add(team->mcast_rejoin.count, &team->mcast_rejoin.count_pending); + schedule_delayed_work(&team->mcast_rejoin.dw, 0); +} + +static void team_mcast_rejoin_init(struct team *team) +{ + INIT_DELAYED_WORK(&team->mcast_rejoin.dw, team_mcast_rejoin_work); +} + +static void team_mcast_rejoin_fini(struct team *team) +{ + cancel_delayed_work_sync(&team->mcast_rejoin.dw); +} + + +/************************ + * Rx path frame handler + ************************/ + +/* note: already called with rcu_read_lock */ +static rx_handler_result_t team_handle_frame(struct sk_buff **pskb) +{ + struct sk_buff *skb = *pskb; + struct team_port *port; + struct team *team; + rx_handler_result_t res; + + skb = skb_share_check(skb, GFP_ATOMIC); + if (!skb) + return RX_HANDLER_CONSUMED; + + *pskb = skb; + + port = team_port_get_rcu(skb->dev); + team = port->team; + if (!team_port_enabled(port)) { + if (is_link_local_ether_addr(eth_hdr(skb)->h_dest)) + /* link-local packets are mostly useful when stack receives them + * with the link they arrive on. + */ + return RX_HANDLER_PASS; + /* allow exact match delivery for disabled ports */ + res = RX_HANDLER_EXACT; + } else { + res = team->ops.receive(team, port, skb); + } + if (res == RX_HANDLER_ANOTHER) { + struct team_pcpu_stats *pcpu_stats; + + pcpu_stats = this_cpu_ptr(team->pcpu_stats); + u64_stats_update_begin(&pcpu_stats->syncp); + u64_stats_inc(&pcpu_stats->rx_packets); + u64_stats_add(&pcpu_stats->rx_bytes, skb->len); + if (skb->pkt_type == PACKET_MULTICAST) + u64_stats_inc(&pcpu_stats->rx_multicast); + u64_stats_update_end(&pcpu_stats->syncp); + + skb->dev = team->dev; + } else if (res == RX_HANDLER_EXACT) { + this_cpu_inc(team->pcpu_stats->rx_nohandler); + } else { + this_cpu_inc(team->pcpu_stats->rx_dropped); + } + + return res; +} + + +/************************************* + * Multiqueue Tx port select override + *************************************/ + +static int team_queue_override_init(struct team *team) +{ + struct list_head *listarr; + unsigned int queue_cnt = team->dev->num_tx_queues - 1; + unsigned int i; + + if (!queue_cnt) + return 0; + listarr = kmalloc_array(queue_cnt, sizeof(struct list_head), + GFP_KERNEL); + if (!listarr) + return -ENOMEM; + team->qom_lists = listarr; + for (i = 0; i < queue_cnt; i++) + INIT_LIST_HEAD(listarr++); + return 0; +} + +static void team_queue_override_fini(struct team *team) +{ + kfree(team->qom_lists); +} + +static struct list_head *__team_get_qom_list(struct team *team, u16 queue_id) +{ + return &team->qom_lists[queue_id - 1]; +} + +/* + * note: already called with rcu_read_lock + */ +static bool team_queue_override_transmit(struct team *team, struct sk_buff *skb) +{ + struct list_head *qom_list; + struct team_port *port; + + if (!team->queue_override_enabled || !skb->queue_mapping) + return false; + qom_list = __team_get_qom_list(team, skb->queue_mapping); + list_for_each_entry_rcu(port, qom_list, qom_list) { + if (!team_dev_queue_xmit(team, port, skb)) + return true; + } + return false; +} + +static void __team_queue_override_port_del(struct team *team, + struct team_port *port) +{ + if (!port->queue_id) + return; + list_del_rcu(&port->qom_list); +} + +static bool team_queue_override_port_has_gt_prio_than(struct team_port *port, + struct team_port *cur) +{ + if (port->priority < cur->priority) + return true; + if (port->priority > cur->priority) + return false; + if (port->index < cur->index) + return true; + return false; +} + +static void __team_queue_override_port_add(struct team *team, + struct team_port *port) +{ + struct team_port *cur; + struct list_head *qom_list; + struct list_head *node; + + if (!port->queue_id) + return; + qom_list = __team_get_qom_list(team, port->queue_id); + node = qom_list; + list_for_each_entry(cur, qom_list, qom_list) { + if (team_queue_override_port_has_gt_prio_than(port, cur)) + break; + node = &cur->qom_list; + } + list_add_tail_rcu(&port->qom_list, node); +} + +static void __team_queue_override_enabled_check(struct team *team) +{ + struct team_port *port; + bool enabled = false; + + list_for_each_entry(port, &team->port_list, list) { + if (port->queue_id) { + enabled = true; + break; + } + } + if (enabled == team->queue_override_enabled) + return; + netdev_dbg(team->dev, "%s queue override\n", + enabled ? "Enabling" : "Disabling"); + team->queue_override_enabled = enabled; +} + +static void team_queue_override_port_prio_changed(struct team *team, + struct team_port *port) +{ + if (!port->queue_id || team_port_enabled(port)) + return; + __team_queue_override_port_del(team, port); + __team_queue_override_port_add(team, port); + __team_queue_override_enabled_check(team); +} + +static void team_queue_override_port_change_queue_id(struct team *team, + struct team_port *port, + u16 new_queue_id) +{ + if (team_port_enabled(port)) { + __team_queue_override_port_del(team, port); + port->queue_id = new_queue_id; + __team_queue_override_port_add(team, port); + __team_queue_override_enabled_check(team); + } else { + port->queue_id = new_queue_id; + } +} + +static void team_queue_override_port_add(struct team *team, + struct team_port *port) +{ + __team_queue_override_port_add(team, port); + __team_queue_override_enabled_check(team); +} + +static void team_queue_override_port_del(struct team *team, + struct team_port *port) +{ + __team_queue_override_port_del(team, port); + __team_queue_override_enabled_check(team); +} + + +/**************** + * Port handling + ****************/ + +static bool team_port_find(const struct team *team, + const struct team_port *port) +{ + struct team_port *cur; + + list_for_each_entry(cur, &team->port_list, list) + if (cur == port) + return true; + return false; +} + +/* + * Enable/disable port by adding to enabled port hashlist and setting + * port->index (Might be racy so reader could see incorrect ifindex when + * processing a flying packet, but that is not a problem). Write guarded + * by team->lock. + */ +static void team_port_enable(struct team *team, + struct team_port *port) +{ + if (team_port_enabled(port)) + return; + port->index = team->en_port_count++; + hlist_add_head_rcu(&port->hlist, + team_port_index_hash(team, port->index)); + team_adjust_ops(team); + team_queue_override_port_add(team, port); + if (team->ops.port_enabled) + team->ops.port_enabled(team, port); + team_notify_peers(team); + team_mcast_rejoin(team); + team_lower_state_changed(port); +} + +static void __reconstruct_port_hlist(struct team *team, int rm_index) +{ + int i; + struct team_port *port; + + for (i = rm_index + 1; i < team->en_port_count; i++) { + port = team_get_port_by_index(team, i); + hlist_del_rcu(&port->hlist); + port->index--; + hlist_add_head_rcu(&port->hlist, + team_port_index_hash(team, port->index)); + } +} + +static void team_port_disable(struct team *team, + struct team_port *port) +{ + if (!team_port_enabled(port)) + return; + if (team->ops.port_disabled) + team->ops.port_disabled(team, port); + hlist_del_rcu(&port->hlist); + __reconstruct_port_hlist(team, port->index); + port->index = -1; + team->en_port_count--; + team_queue_override_port_del(team, port); + team_adjust_ops(team); + team_lower_state_changed(port); +} + +#define TEAM_VLAN_FEATURES (NETIF_F_HW_CSUM | NETIF_F_SG | \ + NETIF_F_FRAGLIST | NETIF_F_GSO_SOFTWARE | \ + NETIF_F_HIGHDMA | NETIF_F_LRO) + +#define TEAM_ENC_FEATURES (NETIF_F_HW_CSUM | NETIF_F_SG | \ + NETIF_F_RXCSUM | NETIF_F_GSO_SOFTWARE) + +static void __team_compute_features(struct team *team) +{ + struct team_port *port; + netdev_features_t vlan_features = TEAM_VLAN_FEATURES & + NETIF_F_ALL_FOR_ALL; + netdev_features_t enc_features = TEAM_ENC_FEATURES; + unsigned short max_hard_header_len = ETH_HLEN; + unsigned int dst_release_flag = IFF_XMIT_DST_RELEASE | + IFF_XMIT_DST_RELEASE_PERM; + + rcu_read_lock(); + list_for_each_entry_rcu(port, &team->port_list, list) { + vlan_features = netdev_increment_features(vlan_features, + port->dev->vlan_features, + TEAM_VLAN_FEATURES); + enc_features = + netdev_increment_features(enc_features, + port->dev->hw_enc_features, + TEAM_ENC_FEATURES); + + + dst_release_flag &= port->dev->priv_flags; + if (port->dev->hard_header_len > max_hard_header_len) + max_hard_header_len = port->dev->hard_header_len; + } + rcu_read_unlock(); + + team->dev->vlan_features = vlan_features; + team->dev->hw_enc_features = enc_features | NETIF_F_GSO_ENCAP_ALL | + NETIF_F_HW_VLAN_CTAG_TX | + NETIF_F_HW_VLAN_STAG_TX; + team->dev->hard_header_len = max_hard_header_len; + + team->dev->priv_flags &= ~IFF_XMIT_DST_RELEASE; + if (dst_release_flag == (IFF_XMIT_DST_RELEASE | IFF_XMIT_DST_RELEASE_PERM)) + team->dev->priv_flags |= IFF_XMIT_DST_RELEASE; +} + +static void team_compute_features(struct team *team) +{ + __team_compute_features(team); + netdev_change_features(team->dev); +} + +static int team_port_enter(struct team *team, struct team_port *port) +{ + int err = 0; + + dev_hold(team->dev); + if (team->ops.port_enter) { + err = team->ops.port_enter(team, port); + if (err) { + netdev_err(team->dev, "Device %s failed to enter team mode\n", + port->dev->name); + goto err_port_enter; + } + } + + return 0; + +err_port_enter: + dev_put(team->dev); + + return err; +} + +static void team_port_leave(struct team *team, struct team_port *port) +{ + if (team->ops.port_leave) + team->ops.port_leave(team, port); + dev_put(team->dev); +} + +#ifdef CONFIG_NET_POLL_CONTROLLER +static int __team_port_enable_netpoll(struct team_port *port) +{ + struct netpoll *np; + int err; + + np = kzalloc(sizeof(*np), GFP_KERNEL); + if (!np) + return -ENOMEM; + + err = __netpoll_setup(np, port->dev); + if (err) { + kfree(np); + return err; + } + port->np = np; + return err; +} + +static int team_port_enable_netpoll(struct team_port *port) +{ + if (!port->team->dev->npinfo) + return 0; + + return __team_port_enable_netpoll(port); +} + +static void team_port_disable_netpoll(struct team_port *port) +{ + struct netpoll *np = port->np; + + if (!np) + return; + port->np = NULL; + + __netpoll_free(np); +} +#else +static int team_port_enable_netpoll(struct team_port *port) +{ + return 0; +} +static void team_port_disable_netpoll(struct team_port *port) +{ +} +#endif + +static int team_upper_dev_link(struct team *team, struct team_port *port, + struct netlink_ext_ack *extack) +{ + struct netdev_lag_upper_info lag_upper_info; + int err; + + lag_upper_info.tx_type = team->mode->lag_tx_type; + lag_upper_info.hash_type = NETDEV_LAG_HASH_UNKNOWN; + err = netdev_master_upper_dev_link(port->dev, team->dev, NULL, + &lag_upper_info, extack); + if (err) + return err; + port->dev->priv_flags |= IFF_TEAM_PORT; + return 0; +} + +static void team_upper_dev_unlink(struct team *team, struct team_port *port) +{ + netdev_upper_dev_unlink(port->dev, team->dev); + port->dev->priv_flags &= ~IFF_TEAM_PORT; +} + +static void __team_port_change_port_added(struct team_port *port, bool linkup); +static int team_dev_type_check_change(struct net_device *dev, + struct net_device *port_dev); + +static int team_port_add(struct team *team, struct net_device *port_dev, + struct netlink_ext_ack *extack) +{ + struct net_device *dev = team->dev; + struct team_port *port; + char *portname = port_dev->name; + int err; + + if (port_dev->flags & IFF_LOOPBACK) { + NL_SET_ERR_MSG(extack, "Loopback device can't be added as a team port"); + netdev_err(dev, "Device %s is loopback device. Loopback devices can't be added as a team port\n", + portname); + return -EINVAL; + } + + if (netif_is_team_port(port_dev)) { + NL_SET_ERR_MSG(extack, "Device is already a port of a team device"); + netdev_err(dev, "Device %s is already a port " + "of a team device\n", portname); + return -EBUSY; + } + + if (dev == port_dev) { + NL_SET_ERR_MSG(extack, "Cannot enslave team device to itself"); + netdev_err(dev, "Cannot enslave team device to itself\n"); + return -EINVAL; + } + + if (netdev_has_upper_dev(dev, port_dev)) { + NL_SET_ERR_MSG(extack, "Device is already an upper device of the team interface"); + netdev_err(dev, "Device %s is already an upper device of the team interface\n", + portname); + return -EBUSY; + } + + if (port_dev->features & NETIF_F_VLAN_CHALLENGED && + vlan_uses_dev(dev)) { + NL_SET_ERR_MSG(extack, "Device is VLAN challenged and team device has VLAN set up"); + netdev_err(dev, "Device %s is VLAN challenged and team device has VLAN set up\n", + portname); + return -EPERM; + } + + err = team_dev_type_check_change(dev, port_dev); + if (err) + return err; + + if (port_dev->flags & IFF_UP) { + NL_SET_ERR_MSG(extack, "Device is up. Set it down before adding it as a team port"); + netdev_err(dev, "Device %s is up. Set it down before adding it as a team port\n", + portname); + return -EBUSY; + } + + port = kzalloc(sizeof(struct team_port) + team->mode->port_priv_size, + GFP_KERNEL); + if (!port) + return -ENOMEM; + + port->dev = port_dev; + port->team = team; + INIT_LIST_HEAD(&port->qom_list); + + port->orig.mtu = port_dev->mtu; + err = dev_set_mtu(port_dev, dev->mtu); + if (err) { + netdev_dbg(dev, "Error %d calling dev_set_mtu\n", err); + goto err_set_mtu; + } + + memcpy(port->orig.dev_addr, port_dev->dev_addr, port_dev->addr_len); + + err = team_port_enter(team, port); + if (err) { + netdev_err(dev, "Device %s failed to enter team mode\n", + portname); + goto err_port_enter; + } + + err = dev_open(port_dev, extack); + if (err) { + netdev_dbg(dev, "Device %s opening failed\n", + portname); + goto err_dev_open; + } + + err = vlan_vids_add_by_dev(port_dev, dev); + if (err) { + netdev_err(dev, "Failed to add vlan ids to device %s\n", + portname); + goto err_vids_add; + } + + err = team_port_enable_netpoll(port); + if (err) { + netdev_err(dev, "Failed to enable netpoll on device %s\n", + portname); + goto err_enable_netpoll; + } + + if (!(dev->features & NETIF_F_LRO)) + dev_disable_lro(port_dev); + + err = netdev_rx_handler_register(port_dev, team_handle_frame, + port); + if (err) { + netdev_err(dev, "Device %s failed to register rx_handler\n", + portname); + goto err_handler_register; + } + + err = team_upper_dev_link(team, port, extack); + if (err) { + netdev_err(dev, "Device %s failed to set upper link\n", + portname); + goto err_set_upper_link; + } + + err = __team_option_inst_add_port(team, port); + if (err) { + netdev_err(dev, "Device %s failed to add per-port options\n", + portname); + goto err_option_port_add; + } + + /* set promiscuity level to new slave */ + if (dev->flags & IFF_PROMISC) { + err = dev_set_promiscuity(port_dev, 1); + if (err) + goto err_set_slave_promisc; + } + + /* set allmulti level to new slave */ + if (dev->flags & IFF_ALLMULTI) { + err = dev_set_allmulti(port_dev, 1); + if (err) { + if (dev->flags & IFF_PROMISC) + dev_set_promiscuity(port_dev, -1); + goto err_set_slave_promisc; + } + } + + if (dev->flags & IFF_UP) { + netif_addr_lock_bh(dev); + dev_uc_sync_multiple(port_dev, dev); + dev_mc_sync_multiple(port_dev, dev); + netif_addr_unlock_bh(dev); + } + + port->index = -1; + list_add_tail_rcu(&port->list, &team->port_list); + team_port_enable(team, port); + __team_compute_features(team); + __team_port_change_port_added(port, !!netif_oper_up(port_dev)); + __team_options_change_check(team); + + netdev_info(dev, "Port device %s added\n", portname); + + return 0; + +err_set_slave_promisc: + __team_option_inst_del_port(team, port); + +err_option_port_add: + team_upper_dev_unlink(team, port); + +err_set_upper_link: + netdev_rx_handler_unregister(port_dev); + +err_handler_register: + team_port_disable_netpoll(port); + +err_enable_netpoll: + vlan_vids_del_by_dev(port_dev, dev); + +err_vids_add: + dev_close(port_dev); + +err_dev_open: + team_port_leave(team, port); + team_port_set_orig_dev_addr(port); + +err_port_enter: + dev_set_mtu(port_dev, port->orig.mtu); + +err_set_mtu: + kfree(port); + + return err; +} + +static void __team_port_change_port_removed(struct team_port *port); + +static int team_port_del(struct team *team, struct net_device *port_dev) +{ + struct net_device *dev = team->dev; + struct team_port *port; + char *portname = port_dev->name; + + port = team_port_get_rtnl(port_dev); + if (!port || !team_port_find(team, port)) { + netdev_err(dev, "Device %s does not act as a port of this team\n", + portname); + return -ENOENT; + } + + team_port_disable(team, port); + list_del_rcu(&port->list); + + if (dev->flags & IFF_PROMISC) + dev_set_promiscuity(port_dev, -1); + if (dev->flags & IFF_ALLMULTI) + dev_set_allmulti(port_dev, -1); + + team_upper_dev_unlink(team, port); + netdev_rx_handler_unregister(port_dev); + team_port_disable_netpoll(port); + vlan_vids_del_by_dev(port_dev, dev); + if (dev->flags & IFF_UP) { + dev_uc_unsync(port_dev, dev); + dev_mc_unsync(port_dev, dev); + } + dev_close(port_dev); + team_port_leave(team, port); + + __team_option_inst_mark_removed_port(team, port); + __team_options_change_check(team); + __team_option_inst_del_port(team, port); + __team_port_change_port_removed(port); + + team_port_set_orig_dev_addr(port); + dev_set_mtu(port_dev, port->orig.mtu); + kfree_rcu(port, rcu); + netdev_info(dev, "Port device %s removed\n", portname); + __team_compute_features(team); + + return 0; +} + + +/***************** + * Net device ops + *****************/ + +static void team_mode_option_get(struct team *team, struct team_gsetter_ctx *ctx) +{ + ctx->data.str_val = team->mode->kind; +} + +static int team_mode_option_set(struct team *team, struct team_gsetter_ctx *ctx) +{ + return team_change_mode(team, ctx->data.str_val); +} + +static void team_notify_peers_count_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + ctx->data.u32_val = team->notify_peers.count; +} + +static int team_notify_peers_count_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + team->notify_peers.count = ctx->data.u32_val; + return 0; +} + +static void team_notify_peers_interval_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + ctx->data.u32_val = team->notify_peers.interval; +} + +static int team_notify_peers_interval_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + team->notify_peers.interval = ctx->data.u32_val; + return 0; +} + +static void team_mcast_rejoin_count_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + ctx->data.u32_val = team->mcast_rejoin.count; +} + +static int team_mcast_rejoin_count_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + team->mcast_rejoin.count = ctx->data.u32_val; + return 0; +} + +static void team_mcast_rejoin_interval_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + ctx->data.u32_val = team->mcast_rejoin.interval; +} + +static int team_mcast_rejoin_interval_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + team->mcast_rejoin.interval = ctx->data.u32_val; + return 0; +} + +static void team_port_en_option_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + ctx->data.bool_val = team_port_enabled(port); +} + +static int team_port_en_option_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + if (ctx->data.bool_val) + team_port_enable(team, port); + else + team_port_disable(team, port); + return 0; +} + +static void team_user_linkup_option_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + ctx->data.bool_val = port->user.linkup; +} + +static void __team_carrier_check(struct team *team); + +static int team_user_linkup_option_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + port->user.linkup = ctx->data.bool_val; + team_refresh_port_linkup(port); + __team_carrier_check(port->team); + return 0; +} + +static void team_user_linkup_en_option_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + ctx->data.bool_val = port->user.linkup_enabled; +} + +static int team_user_linkup_en_option_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + port->user.linkup_enabled = ctx->data.bool_val; + team_refresh_port_linkup(port); + __team_carrier_check(port->team); + return 0; +} + +static void team_priority_option_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + ctx->data.s32_val = port->priority; +} + +static int team_priority_option_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + s32 priority = ctx->data.s32_val; + + if (port->priority == priority) + return 0; + port->priority = priority; + team_queue_override_port_prio_changed(team, port); + return 0; +} + +static void team_queue_id_option_get(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + + ctx->data.u32_val = port->queue_id; +} + +static int team_queue_id_option_set(struct team *team, + struct team_gsetter_ctx *ctx) +{ + struct team_port *port = ctx->info->port; + u16 new_queue_id = ctx->data.u32_val; + + if (port->queue_id == new_queue_id) + return 0; + if (new_queue_id >= team->dev->real_num_tx_queues) + return -EINVAL; + team_queue_override_port_change_queue_id(team, port, new_queue_id); + return 0; +} + +static const struct team_option team_options[] = { + { + .name = "mode", + .type = TEAM_OPTION_TYPE_STRING, + .getter = team_mode_option_get, + .setter = team_mode_option_set, + }, + { + .name = "notify_peers_count", + .type = TEAM_OPTION_TYPE_U32, + .getter = team_notify_peers_count_get, + .setter = team_notify_peers_count_set, + }, + { + .name = "notify_peers_interval", + .type = TEAM_OPTION_TYPE_U32, + .getter = team_notify_peers_interval_get, + .setter = team_notify_peers_interval_set, + }, + { + .name = "mcast_rejoin_count", + .type = TEAM_OPTION_TYPE_U32, + .getter = team_mcast_rejoin_count_get, + .setter = team_mcast_rejoin_count_set, + }, + { + .name = "mcast_rejoin_interval", + .type = TEAM_OPTION_TYPE_U32, + .getter = team_mcast_rejoin_interval_get, + .setter = team_mcast_rejoin_interval_set, + }, + { + .name = "enabled", + .type = TEAM_OPTION_TYPE_BOOL, + .per_port = true, + .getter = team_port_en_option_get, + .setter = team_port_en_option_set, + }, + { + .name = "user_linkup", + .type = TEAM_OPTION_TYPE_BOOL, + .per_port = true, + .getter = team_user_linkup_option_get, + .setter = team_user_linkup_option_set, + }, + { + .name = "user_linkup_enabled", + .type = TEAM_OPTION_TYPE_BOOL, + .per_port = true, + .getter = team_user_linkup_en_option_get, + .setter = team_user_linkup_en_option_set, + }, + { + .name = "priority", + .type = TEAM_OPTION_TYPE_S32, + .per_port = true, + .getter = team_priority_option_get, + .setter = team_priority_option_set, + }, + { + .name = "queue_id", + .type = TEAM_OPTION_TYPE_U32, + .per_port = true, + .getter = team_queue_id_option_get, + .setter = team_queue_id_option_set, + }, +}; + + +static int team_init(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + int i; + int err; + + team->dev = dev; + team_set_no_mode(team); + team->notifier_ctx = false; + + team->pcpu_stats = netdev_alloc_pcpu_stats(struct team_pcpu_stats); + if (!team->pcpu_stats) + return -ENOMEM; + + for (i = 0; i < TEAM_PORT_HASHENTRIES; i++) + INIT_HLIST_HEAD(&team->en_port_hlist[i]); + INIT_LIST_HEAD(&team->port_list); + err = team_queue_override_init(team); + if (err) + goto err_team_queue_override_init; + + team_adjust_ops(team); + + INIT_LIST_HEAD(&team->option_list); + INIT_LIST_HEAD(&team->option_inst_list); + + team_notify_peers_init(team); + team_mcast_rejoin_init(team); + + err = team_options_register(team, team_options, ARRAY_SIZE(team_options)); + if (err) + goto err_options_register; + netif_carrier_off(dev); + + lockdep_register_key(&team->team_lock_key); + __mutex_init(&team->lock, "team->team_lock_key", &team->team_lock_key); + netdev_lockdep_set_classes(dev); + + return 0; + +err_options_register: + team_mcast_rejoin_fini(team); + team_notify_peers_fini(team); + team_queue_override_fini(team); +err_team_queue_override_init: + free_percpu(team->pcpu_stats); + + return err; +} + +static void team_uninit(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + struct team_port *tmp; + + mutex_lock(&team->lock); + list_for_each_entry_safe(port, tmp, &team->port_list, list) + team_port_del(team, port->dev); + + __team_change_mode(team, NULL); /* cleanup */ + __team_options_unregister(team, team_options, ARRAY_SIZE(team_options)); + team_mcast_rejoin_fini(team); + team_notify_peers_fini(team); + team_queue_override_fini(team); + mutex_unlock(&team->lock); + netdev_change_features(dev); + lockdep_unregister_key(&team->team_lock_key); +} + +static void team_destructor(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + + free_percpu(team->pcpu_stats); +} + +static int team_open(struct net_device *dev) +{ + return 0; +} + +static int team_close(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + + list_for_each_entry(port, &team->port_list, list) { + dev_uc_unsync(port->dev, dev); + dev_mc_unsync(port->dev, dev); + } + + return 0; +} + +/* + * note: already called with rcu_read_lock + */ +static netdev_tx_t team_xmit(struct sk_buff *skb, struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + bool tx_success; + unsigned int len = skb->len; + + tx_success = team_queue_override_transmit(team, skb); + if (!tx_success) + tx_success = team->ops.transmit(team, skb); + if (tx_success) { + struct team_pcpu_stats *pcpu_stats; + + pcpu_stats = this_cpu_ptr(team->pcpu_stats); + u64_stats_update_begin(&pcpu_stats->syncp); + u64_stats_inc(&pcpu_stats->tx_packets); + u64_stats_add(&pcpu_stats->tx_bytes, len); + u64_stats_update_end(&pcpu_stats->syncp); + } else { + this_cpu_inc(team->pcpu_stats->tx_dropped); + } + + return NETDEV_TX_OK; +} + +static u16 team_select_queue(struct net_device *dev, struct sk_buff *skb, + struct net_device *sb_dev) +{ + /* + * This helper function exists to help dev_pick_tx get the correct + * destination queue. Using a helper function skips a call to + * skb_tx_hash and will put the skbs in the queue we expect on their + * way down to the team driver. + */ + u16 txq = skb_rx_queue_recorded(skb) ? skb_get_rx_queue(skb) : 0; + + /* + * Save the original txq to restore before passing to the driver + */ + qdisc_skb_cb(skb)->slave_dev_queue_mapping = skb->queue_mapping; + + if (unlikely(txq >= dev->real_num_tx_queues)) { + do { + txq -= dev->real_num_tx_queues; + } while (txq >= dev->real_num_tx_queues); + } + return txq; +} + +static void team_change_rx_flags(struct net_device *dev, int change) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + int inc; + + rcu_read_lock(); + list_for_each_entry_rcu(port, &team->port_list, list) { + if (change & IFF_PROMISC) { + inc = dev->flags & IFF_PROMISC ? 1 : -1; + dev_set_promiscuity(port->dev, inc); + } + if (change & IFF_ALLMULTI) { + inc = dev->flags & IFF_ALLMULTI ? 1 : -1; + dev_set_allmulti(port->dev, inc); + } + } + rcu_read_unlock(); +} + +static void team_set_rx_mode(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + + rcu_read_lock(); + list_for_each_entry_rcu(port, &team->port_list, list) { + dev_uc_sync_multiple(port->dev, dev); + dev_mc_sync_multiple(port->dev, dev); + } + rcu_read_unlock(); +} + +static int team_set_mac_address(struct net_device *dev, void *p) +{ + struct sockaddr *addr = p; + struct team *team = netdev_priv(dev); + struct team_port *port; + + if (dev->type == ARPHRD_ETHER && !is_valid_ether_addr(addr->sa_data)) + return -EADDRNOTAVAIL; + dev_addr_set(dev, addr->sa_data); + mutex_lock(&team->lock); + list_for_each_entry(port, &team->port_list, list) + if (team->ops.port_change_dev_addr) + team->ops.port_change_dev_addr(team, port); + mutex_unlock(&team->lock); + return 0; +} + +static int team_change_mtu(struct net_device *dev, int new_mtu) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + int err; + + /* + * Alhough this is reader, it's guarded by team lock. It's not possible + * to traverse list in reverse under rcu_read_lock + */ + mutex_lock(&team->lock); + team->port_mtu_change_allowed = true; + list_for_each_entry(port, &team->port_list, list) { + err = dev_set_mtu(port->dev, new_mtu); + if (err) { + netdev_err(dev, "Device %s failed to change mtu", + port->dev->name); + goto unwind; + } + } + team->port_mtu_change_allowed = false; + mutex_unlock(&team->lock); + + WRITE_ONCE(dev->mtu, new_mtu); + + return 0; + +unwind: + list_for_each_entry_continue_reverse(port, &team->port_list, list) + dev_set_mtu(port->dev, dev->mtu); + team->port_mtu_change_allowed = false; + mutex_unlock(&team->lock); + + return err; +} + +static void +team_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats) +{ + struct team *team = netdev_priv(dev); + struct team_pcpu_stats *p; + u64 rx_packets, rx_bytes, rx_multicast, tx_packets, tx_bytes; + u32 rx_dropped = 0, tx_dropped = 0, rx_nohandler = 0; + unsigned int start; + int i; + + for_each_possible_cpu(i) { + p = per_cpu_ptr(team->pcpu_stats, i); + do { + start = u64_stats_fetch_begin(&p->syncp); + rx_packets = u64_stats_read(&p->rx_packets); + rx_bytes = u64_stats_read(&p->rx_bytes); + rx_multicast = u64_stats_read(&p->rx_multicast); + tx_packets = u64_stats_read(&p->tx_packets); + tx_bytes = u64_stats_read(&p->tx_bytes); + } while (u64_stats_fetch_retry(&p->syncp, start)); + + stats->rx_packets += rx_packets; + stats->rx_bytes += rx_bytes; + stats->multicast += rx_multicast; + stats->tx_packets += tx_packets; + stats->tx_bytes += tx_bytes; + /* + * rx_dropped, tx_dropped & rx_nohandler are u32, + * updated without syncp protection. + */ + rx_dropped += READ_ONCE(p->rx_dropped); + tx_dropped += READ_ONCE(p->tx_dropped); + rx_nohandler += READ_ONCE(p->rx_nohandler); + } + stats->rx_dropped = rx_dropped; + stats->tx_dropped = tx_dropped; + stats->rx_nohandler = rx_nohandler; +} + +static int team_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + int err; + + /* + * Alhough this is reader, it's guarded by team lock. It's not possible + * to traverse list in reverse under rcu_read_lock + */ + mutex_lock(&team->lock); + list_for_each_entry(port, &team->port_list, list) { + err = vlan_vid_add(port->dev, proto, vid); + if (err) + goto unwind; + } + mutex_unlock(&team->lock); + + return 0; + +unwind: + list_for_each_entry_continue_reverse(port, &team->port_list, list) + vlan_vid_del(port->dev, proto, vid); + mutex_unlock(&team->lock); + + return err; +} + +static int team_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + + mutex_lock(&team->lock); + list_for_each_entry(port, &team->port_list, list) + vlan_vid_del(port->dev, proto, vid); + mutex_unlock(&team->lock); + + return 0; +} + +#ifdef CONFIG_NET_POLL_CONTROLLER +static void team_poll_controller(struct net_device *dev) +{ +} + +static void __team_netpoll_cleanup(struct team *team) +{ + struct team_port *port; + + list_for_each_entry(port, &team->port_list, list) + team_port_disable_netpoll(port); +} + +static void team_netpoll_cleanup(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + + mutex_lock(&team->lock); + __team_netpoll_cleanup(team); + mutex_unlock(&team->lock); +} + +static int team_netpoll_setup(struct net_device *dev, + struct netpoll_info *npifo) +{ + struct team *team = netdev_priv(dev); + struct team_port *port; + int err = 0; + + mutex_lock(&team->lock); + list_for_each_entry(port, &team->port_list, list) { + err = __team_port_enable_netpoll(port); + if (err) { + __team_netpoll_cleanup(team); + break; + } + } + mutex_unlock(&team->lock); + return err; +} +#endif + +static int team_add_slave(struct net_device *dev, struct net_device *port_dev, + struct netlink_ext_ack *extack) +{ + struct team *team = netdev_priv(dev); + int err; + + mutex_lock(&team->lock); + err = team_port_add(team, port_dev, extack); + mutex_unlock(&team->lock); + + if (!err) + netdev_change_features(dev); + + return err; +} + +static int team_del_slave(struct net_device *dev, struct net_device *port_dev) +{ + struct team *team = netdev_priv(dev); + int err; + + mutex_lock(&team->lock); + err = team_port_del(team, port_dev); + mutex_unlock(&team->lock); + + if (err) + return err; + + if (netif_is_team_master(port_dev)) { + lockdep_unregister_key(&team->team_lock_key); + lockdep_register_key(&team->team_lock_key); + lockdep_set_class(&team->lock, &team->team_lock_key); + } + netdev_change_features(dev); + + return err; +} + +static netdev_features_t team_fix_features(struct net_device *dev, + netdev_features_t features) +{ + struct team_port *port; + struct team *team = netdev_priv(dev); + netdev_features_t mask; + + mask = features; + features &= ~NETIF_F_ONE_FOR_ALL; + features |= NETIF_F_ALL_FOR_ALL; + + rcu_read_lock(); + list_for_each_entry_rcu(port, &team->port_list, list) { + features = netdev_increment_features(features, + port->dev->features, + mask); + } + rcu_read_unlock(); + + features = netdev_add_tso_features(features, mask); + + return features; +} + +static int team_change_carrier(struct net_device *dev, bool new_carrier) +{ + struct team *team = netdev_priv(dev); + + team->user_carrier_enabled = true; + + if (new_carrier) + netif_carrier_on(dev); + else + netif_carrier_off(dev); + return 0; +} + +static const struct net_device_ops team_netdev_ops = { + .ndo_init = team_init, + .ndo_uninit = team_uninit, + .ndo_open = team_open, + .ndo_stop = team_close, + .ndo_start_xmit = team_xmit, + .ndo_select_queue = team_select_queue, + .ndo_change_rx_flags = team_change_rx_flags, + .ndo_set_rx_mode = team_set_rx_mode, + .ndo_set_mac_address = team_set_mac_address, + .ndo_change_mtu = team_change_mtu, + .ndo_get_stats64 = team_get_stats64, + .ndo_vlan_rx_add_vid = team_vlan_rx_add_vid, + .ndo_vlan_rx_kill_vid = team_vlan_rx_kill_vid, +#ifdef CONFIG_NET_POLL_CONTROLLER + .ndo_poll_controller = team_poll_controller, + .ndo_netpoll_setup = team_netpoll_setup, + .ndo_netpoll_cleanup = team_netpoll_cleanup, +#endif + .ndo_add_slave = team_add_slave, + .ndo_del_slave = team_del_slave, + .ndo_fix_features = team_fix_features, + .ndo_change_carrier = team_change_carrier, + .ndo_features_check = passthru_features_check, +}; + +/*********************** + * ethtool interface + ***********************/ + +static void team_ethtool_get_drvinfo(struct net_device *dev, + struct ethtool_drvinfo *drvinfo) +{ + strscpy(drvinfo->driver, DRV_NAME, sizeof(drvinfo->driver)); +} + +static int team_ethtool_get_link_ksettings(struct net_device *dev, + struct ethtool_link_ksettings *cmd) +{ + struct team *team= netdev_priv(dev); + unsigned long speed = 0; + struct team_port *port; + + cmd->base.duplex = DUPLEX_UNKNOWN; + cmd->base.port = PORT_OTHER; + + rcu_read_lock(); + list_for_each_entry_rcu(port, &team->port_list, list) { + if (team_port_txable(port)) { + if (port->state.speed != SPEED_UNKNOWN) + speed += port->state.speed; + if (cmd->base.duplex == DUPLEX_UNKNOWN && + port->state.duplex != DUPLEX_UNKNOWN) + cmd->base.duplex = port->state.duplex; + } + } + rcu_read_unlock(); + + cmd->base.speed = speed ? : SPEED_UNKNOWN; + + return 0; +} + +static const struct ethtool_ops team_ethtool_ops = { + .get_drvinfo = team_ethtool_get_drvinfo, + .get_link = ethtool_op_get_link, + .get_link_ksettings = team_ethtool_get_link_ksettings, +}; + +/*********************** + * rt netlink interface + ***********************/ + +static void team_setup_by_port(struct net_device *dev, + struct net_device *port_dev) +{ + struct team *team = netdev_priv(dev); + + if (port_dev->type == ARPHRD_ETHER) + dev->header_ops = team->header_ops_cache; + else + dev->header_ops = port_dev->header_ops; + dev->type = port_dev->type; + dev->hard_header_len = port_dev->hard_header_len; + dev->needed_headroom = port_dev->needed_headroom; + dev->addr_len = port_dev->addr_len; + dev->mtu = port_dev->mtu; + memcpy(dev->broadcast, port_dev->broadcast, port_dev->addr_len); + eth_hw_addr_inherit(dev, port_dev); + + if (port_dev->flags & IFF_POINTOPOINT) { + dev->flags &= ~(IFF_BROADCAST | IFF_MULTICAST); + dev->flags |= (IFF_POINTOPOINT | IFF_NOARP); + } else if ((port_dev->flags & (IFF_BROADCAST | IFF_MULTICAST)) == + (IFF_BROADCAST | IFF_MULTICAST)) { + dev->flags |= (IFF_BROADCAST | IFF_MULTICAST); + dev->flags &= ~(IFF_POINTOPOINT | IFF_NOARP); + } +} + +static int team_dev_type_check_change(struct net_device *dev, + struct net_device *port_dev) +{ + struct team *team = netdev_priv(dev); + char *portname = port_dev->name; + int err; + + if (dev->type == port_dev->type) + return 0; + if (!list_empty(&team->port_list)) { + netdev_err(dev, "Device %s is of different type\n", portname); + return -EBUSY; + } + err = call_netdevice_notifiers(NETDEV_PRE_TYPE_CHANGE, dev); + err = notifier_to_errno(err); + if (err) { + netdev_err(dev, "Refused to change device type\n"); + return err; + } + dev_uc_flush(dev); + dev_mc_flush(dev); + team_setup_by_port(dev, port_dev); + call_netdevice_notifiers(NETDEV_POST_TYPE_CHANGE, dev); + return 0; +} + +static void team_setup(struct net_device *dev) +{ + struct team *team = netdev_priv(dev); + + ether_setup(dev); + dev->max_mtu = ETH_MAX_MTU; + team->header_ops_cache = dev->header_ops; + + dev->netdev_ops = &team_netdev_ops; + dev->ethtool_ops = &team_ethtool_ops; + dev->needs_free_netdev = true; + dev->priv_destructor = team_destructor; + dev->priv_flags &= ~(IFF_XMIT_DST_RELEASE | IFF_TX_SKB_SHARING); + dev->priv_flags |= IFF_NO_QUEUE; + dev->priv_flags |= IFF_TEAM; + + /* + * Indicate we support unicast address filtering. That way core won't + * bring us to promisc mode in case a unicast addr is added. + * Let this up to underlay drivers. + */ + dev->priv_flags |= IFF_UNICAST_FLT | IFF_LIVE_ADDR_CHANGE; + + dev->features |= NETIF_F_LLTX; + dev->features |= NETIF_F_GRO; + + /* Don't allow team devices to change network namespaces. */ + dev->features |= NETIF_F_NETNS_LOCAL; + + dev->hw_features = TEAM_VLAN_FEATURES | + NETIF_F_HW_VLAN_CTAG_RX | + NETIF_F_HW_VLAN_CTAG_FILTER | + NETIF_F_HW_VLAN_STAG_RX | + NETIF_F_HW_VLAN_STAG_FILTER; + + dev->hw_features |= NETIF_F_GSO_ENCAP_ALL; + dev->features |= dev->hw_features; + dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_STAG_TX; +} + +static int team_newlink(struct net *src_net, struct net_device *dev, + struct nlattr *tb[], struct nlattr *data[], + struct netlink_ext_ack *extack) +{ + if (tb[IFLA_ADDRESS] == NULL) + eth_hw_addr_random(dev); + + return register_netdevice(dev); +} + +static int team_validate(struct nlattr *tb[], struct nlattr *data[], + struct netlink_ext_ack *extack) +{ + if (tb[IFLA_ADDRESS]) { + if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN) + return -EINVAL; + if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS]))) + return -EADDRNOTAVAIL; + } + return 0; +} + +static unsigned int team_get_num_tx_queues(void) +{ + return TEAM_DEFAULT_NUM_TX_QUEUES; +} + +static unsigned int team_get_num_rx_queues(void) +{ + return TEAM_DEFAULT_NUM_RX_QUEUES; +} + +static struct rtnl_link_ops team_link_ops __read_mostly = { + .kind = DRV_NAME, + .priv_size = sizeof(struct team), + .setup = team_setup, + .newlink = team_newlink, + .validate = team_validate, + .get_num_tx_queues = team_get_num_tx_queues, + .get_num_rx_queues = team_get_num_rx_queues, +}; + + +/*********************************** + * Generic netlink custom interface + ***********************************/ + +static struct genl_family team_nl_family; + +int team_nl_noop_doit(struct sk_buff *skb, struct genl_info *info) +{ + struct sk_buff *msg; + void *hdr; + int err; + + msg = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); + if (!msg) + return -ENOMEM; + + hdr = genlmsg_put(msg, info->snd_portid, info->snd_seq, + &team_nl_family, 0, TEAM_CMD_NOOP); + if (!hdr) { + err = -EMSGSIZE; + goto err_msg_put; + } + + genlmsg_end(msg, hdr); + + return genlmsg_unicast(genl_info_net(info), msg, info->snd_portid); + +err_msg_put: + nlmsg_free(msg); + + return err; +} + +/* + * Netlink cmd functions should be locked by following two functions. + * Since dev gets held here, that ensures dev won't disappear in between. + */ +static struct team *team_nl_team_get(struct genl_info *info) +{ + struct net *net = genl_info_net(info); + int ifindex; + struct net_device *dev; + struct team *team; + + if (!info->attrs[TEAM_ATTR_TEAM_IFINDEX]) + return NULL; + + ifindex = nla_get_u32(info->attrs[TEAM_ATTR_TEAM_IFINDEX]); + dev = dev_get_by_index(net, ifindex); + if (!dev || dev->netdev_ops != &team_netdev_ops) { + dev_put(dev); + return NULL; + } + + team = netdev_priv(dev); + mutex_lock(&team->lock); + return team; +} + +static void team_nl_team_put(struct team *team) +{ + mutex_unlock(&team->lock); + dev_put(team->dev); +} + +typedef int team_nl_send_func_t(struct sk_buff *skb, + struct team *team, u32 portid); + +static int team_nl_send_unicast(struct sk_buff *skb, struct team *team, u32 portid) +{ + return genlmsg_unicast(dev_net(team->dev), skb, portid); +} + +static int team_nl_fill_one_option_get(struct sk_buff *skb, struct team *team, + struct team_option_inst *opt_inst) +{ + struct nlattr *option_item; + struct team_option *option = opt_inst->option; + struct team_option_inst_info *opt_inst_info = &opt_inst->info; + struct team_gsetter_ctx ctx; + int err; + + ctx.info = opt_inst_info; + err = team_option_get(team, opt_inst, &ctx); + if (err) + return err; + + option_item = nla_nest_start_noflag(skb, TEAM_ATTR_ITEM_OPTION); + if (!option_item) + return -EMSGSIZE; + + if (nla_put_string(skb, TEAM_ATTR_OPTION_NAME, option->name)) + goto nest_cancel; + if (opt_inst_info->port && + nla_put_u32(skb, TEAM_ATTR_OPTION_PORT_IFINDEX, + opt_inst_info->port->dev->ifindex)) + goto nest_cancel; + if (opt_inst->option->array_size && + nla_put_u32(skb, TEAM_ATTR_OPTION_ARRAY_INDEX, + opt_inst_info->array_index)) + goto nest_cancel; + + switch (option->type) { + case TEAM_OPTION_TYPE_U32: + if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_U32)) + goto nest_cancel; + if (nla_put_u32(skb, TEAM_ATTR_OPTION_DATA, ctx.data.u32_val)) + goto nest_cancel; + break; + case TEAM_OPTION_TYPE_STRING: + if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_STRING)) + goto nest_cancel; + if (nla_put_string(skb, TEAM_ATTR_OPTION_DATA, + ctx.data.str_val)) + goto nest_cancel; + break; + case TEAM_OPTION_TYPE_BINARY: + if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_BINARY)) + goto nest_cancel; + if (nla_put(skb, TEAM_ATTR_OPTION_DATA, ctx.data.bin_val.len, + ctx.data.bin_val.ptr)) + goto nest_cancel; + break; + case TEAM_OPTION_TYPE_BOOL: + if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_FLAG)) + goto nest_cancel; + if (ctx.data.bool_val && + nla_put_flag(skb, TEAM_ATTR_OPTION_DATA)) + goto nest_cancel; + break; + case TEAM_OPTION_TYPE_S32: + if (nla_put_u8(skb, TEAM_ATTR_OPTION_TYPE, NLA_S32)) + goto nest_cancel; + if (nla_put_s32(skb, TEAM_ATTR_OPTION_DATA, ctx.data.s32_val)) + goto nest_cancel; + break; + default: + BUG(); + } + if (opt_inst->removed && nla_put_flag(skb, TEAM_ATTR_OPTION_REMOVED)) + goto nest_cancel; + if (opt_inst->changed) { + if (nla_put_flag(skb, TEAM_ATTR_OPTION_CHANGED)) + goto nest_cancel; + opt_inst->changed = false; + } + nla_nest_end(skb, option_item); + return 0; + +nest_cancel: + nla_nest_cancel(skb, option_item); + return -EMSGSIZE; +} + +static int __send_and_alloc_skb(struct sk_buff **pskb, + struct team *team, u32 portid, + team_nl_send_func_t *send_func) +{ + int err; + + if (*pskb) { + err = send_func(*pskb, team, portid); + if (err) + return err; + } + *pskb = genlmsg_new(GENLMSG_DEFAULT_SIZE, GFP_KERNEL); + if (!*pskb) + return -ENOMEM; + return 0; +} + +static int team_nl_send_options_get(struct team *team, u32 portid, u32 seq, + int flags, team_nl_send_func_t *send_func, + struct list_head *sel_opt_inst_list) +{ + struct nlattr *option_list; + struct nlmsghdr *nlh; + void *hdr; + struct team_option_inst *opt_inst; + int err; + struct sk_buff *skb = NULL; + bool incomplete; + int i; + + opt_inst = list_first_entry(sel_opt_inst_list, + struct team_option_inst, tmp_list); + +start_again: + err = __send_and_alloc_skb(&skb, team, portid, send_func); + if (err) + return err; + + hdr = genlmsg_put(skb, portid, seq, &team_nl_family, flags | NLM_F_MULTI, + TEAM_CMD_OPTIONS_GET); + if (!hdr) { + nlmsg_free(skb); + return -EMSGSIZE; + } + + if (nla_put_u32(skb, TEAM_ATTR_TEAM_IFINDEX, team->dev->ifindex)) + goto nla_put_failure; + option_list = nla_nest_start_noflag(skb, TEAM_ATTR_LIST_OPTION); + if (!option_list) + goto nla_put_failure; + + i = 0; + incomplete = false; + list_for_each_entry_from(opt_inst, sel_opt_inst_list, tmp_list) { + err = team_nl_fill_one_option_get(skb, team, opt_inst); + if (err) { + if (err == -EMSGSIZE) { + if (!i) + goto errout; + incomplete = true; + break; + } + goto errout; + } + i++; + } + + nla_nest_end(skb, option_list); + genlmsg_end(skb, hdr); + if (incomplete) + goto start_again; + +send_done: + nlh = nlmsg_put(skb, portid, seq, NLMSG_DONE, 0, flags | NLM_F_MULTI); + if (!nlh) { + err = __send_and_alloc_skb(&skb, team, portid, send_func); + if (err) + return err; + goto send_done; + } + + return send_func(skb, team, portid); + +nla_put_failure: + err = -EMSGSIZE; +errout: + nlmsg_free(skb); + return err; +} + +int team_nl_options_get_doit(struct sk_buff *skb, struct genl_info *info) +{ + struct team *team; + struct team_option_inst *opt_inst; + int err; + LIST_HEAD(sel_opt_inst_list); + + team = team_nl_team_get(info); + if (!team) + return -EINVAL; + + list_for_each_entry(opt_inst, &team->option_inst_list, list) + list_add_tail(&opt_inst->tmp_list, &sel_opt_inst_list); + err = team_nl_send_options_get(team, info->snd_portid, info->snd_seq, + NLM_F_ACK, team_nl_send_unicast, + &sel_opt_inst_list); + + team_nl_team_put(team); + + return err; +} + +static int team_nl_send_event_options_get(struct team *team, + struct list_head *sel_opt_inst_list); + +int team_nl_options_set_doit(struct sk_buff *skb, struct genl_info *info) +{ + struct team *team; + int err = 0; + int i; + struct nlattr *nl_option; + + rtnl_lock(); + + team = team_nl_team_get(info); + if (!team) { + err = -EINVAL; + goto rtnl_unlock; + } + + err = -EINVAL; + if (!info->attrs[TEAM_ATTR_LIST_OPTION]) { + err = -EINVAL; + goto team_put; + } + + nla_for_each_nested(nl_option, info->attrs[TEAM_ATTR_LIST_OPTION], i) { + struct nlattr *opt_attrs[TEAM_ATTR_OPTION_MAX + 1]; + struct nlattr *attr; + struct nlattr *attr_data; + LIST_HEAD(opt_inst_list); + enum team_option_type opt_type; + int opt_port_ifindex = 0; /* != 0 for per-port options */ + u32 opt_array_index = 0; + bool opt_is_array = false; + struct team_option_inst *opt_inst; + char *opt_name; + bool opt_found = false; + + if (nla_type(nl_option) != TEAM_ATTR_ITEM_OPTION) { + err = -EINVAL; + goto team_put; + } + err = nla_parse_nested_deprecated(opt_attrs, + TEAM_ATTR_OPTION_MAX, + nl_option, + team_attr_option_nl_policy, + info->extack); + if (err) + goto team_put; + if (!opt_attrs[TEAM_ATTR_OPTION_NAME] || + !opt_attrs[TEAM_ATTR_OPTION_TYPE]) { + err = -EINVAL; + goto team_put; + } + switch (nla_get_u8(opt_attrs[TEAM_ATTR_OPTION_TYPE])) { + case NLA_U32: + opt_type = TEAM_OPTION_TYPE_U32; + break; + case NLA_STRING: + opt_type = TEAM_OPTION_TYPE_STRING; + break; + case NLA_BINARY: + opt_type = TEAM_OPTION_TYPE_BINARY; + break; + case NLA_FLAG: + opt_type = TEAM_OPTION_TYPE_BOOL; + break; + case NLA_S32: + opt_type = TEAM_OPTION_TYPE_S32; + break; + default: + goto team_put; + } + + attr_data = opt_attrs[TEAM_ATTR_OPTION_DATA]; + if (opt_type != TEAM_OPTION_TYPE_BOOL && !attr_data) { + err = -EINVAL; + goto team_put; + } + + opt_name = nla_data(opt_attrs[TEAM_ATTR_OPTION_NAME]); + attr = opt_attrs[TEAM_ATTR_OPTION_PORT_IFINDEX]; + if (attr) + opt_port_ifindex = nla_get_u32(attr); + + attr = opt_attrs[TEAM_ATTR_OPTION_ARRAY_INDEX]; + if (attr) { + opt_is_array = true; + opt_array_index = nla_get_u32(attr); + } + + list_for_each_entry(opt_inst, &team->option_inst_list, list) { + struct team_option *option = opt_inst->option; + struct team_gsetter_ctx ctx; + struct team_option_inst_info *opt_inst_info; + int tmp_ifindex; + + opt_inst_info = &opt_inst->info; + tmp_ifindex = opt_inst_info->port ? + opt_inst_info->port->dev->ifindex : 0; + if (option->type != opt_type || + strcmp(option->name, opt_name) || + tmp_ifindex != opt_port_ifindex || + (option->array_size && !opt_is_array) || + opt_inst_info->array_index != opt_array_index) + continue; + opt_found = true; + ctx.info = opt_inst_info; + switch (opt_type) { + case TEAM_OPTION_TYPE_U32: + ctx.data.u32_val = nla_get_u32(attr_data); + break; + case TEAM_OPTION_TYPE_STRING: + if (nla_len(attr_data) > TEAM_STRING_MAX_LEN) { + err = -EINVAL; + goto team_put; + } + ctx.data.str_val = nla_data(attr_data); + break; + case TEAM_OPTION_TYPE_BINARY: + ctx.data.bin_val.len = nla_len(attr_data); + ctx.data.bin_val.ptr = nla_data(attr_data); + break; + case TEAM_OPTION_TYPE_BOOL: + ctx.data.bool_val = attr_data ? true : false; + break; + case TEAM_OPTION_TYPE_S32: + ctx.data.s32_val = nla_get_s32(attr_data); + break; + default: + BUG(); + } + err = team_option_set(team, opt_inst, &ctx); + if (err) + goto team_put; + opt_inst->changed = true; + list_add(&opt_inst->tmp_list, &opt_inst_list); + } + if (!opt_found) { + err = -ENOENT; + goto team_put; + } + + err = team_nl_send_event_options_get(team, &opt_inst_list); + if (err) + break; + } + +team_put: + team_nl_team_put(team); +rtnl_unlock: + rtnl_unlock(); + return err; +} + +static int team_nl_fill_one_port_get(struct sk_buff *skb, + struct team_port *port) +{ + struct nlattr *port_item; + + port_item = nla_nest_start_noflag(skb, TEAM_ATTR_ITEM_PORT); + if (!port_item) + goto nest_cancel; + if (nla_put_u32(skb, TEAM_ATTR_PORT_IFINDEX, port->dev->ifindex)) + goto nest_cancel; + if (port->changed) { + if (nla_put_flag(skb, TEAM_ATTR_PORT_CHANGED)) + goto nest_cancel; + port->changed = false; + } + if ((port->removed && + nla_put_flag(skb, TEAM_ATTR_PORT_REMOVED)) || + (port->state.linkup && + nla_put_flag(skb, TEAM_ATTR_PORT_LINKUP)) || + nla_put_u32(skb, TEAM_ATTR_PORT_SPEED, port->state.speed) || + nla_put_u8(skb, TEAM_ATTR_PORT_DUPLEX, port->state.duplex)) + goto nest_cancel; + nla_nest_end(skb, port_item); + return 0; + +nest_cancel: + nla_nest_cancel(skb, port_item); + return -EMSGSIZE; +} + +static int team_nl_send_port_list_get(struct team *team, u32 portid, u32 seq, + int flags, team_nl_send_func_t *send_func, + struct team_port *one_port) +{ + struct nlattr *port_list; + struct nlmsghdr *nlh; + void *hdr; + struct team_port *port; + int err; + struct sk_buff *skb = NULL; + bool incomplete; + int i; + + port = list_first_entry_or_null(&team->port_list, + struct team_port, list); + +start_again: + err = __send_and_alloc_skb(&skb, team, portid, send_func); + if (err) + return err; + + hdr = genlmsg_put(skb, portid, seq, &team_nl_family, flags | NLM_F_MULTI, + TEAM_CMD_PORT_LIST_GET); + if (!hdr) { + nlmsg_free(skb); + return -EMSGSIZE; + } + + if (nla_put_u32(skb, TEAM_ATTR_TEAM_IFINDEX, team->dev->ifindex)) + goto nla_put_failure; + port_list = nla_nest_start_noflag(skb, TEAM_ATTR_LIST_PORT); + if (!port_list) + goto nla_put_failure; + + i = 0; + incomplete = false; + + /* If one port is selected, called wants to send port list containing + * only this port. Otherwise go through all listed ports and send all + */ + if (one_port) { + err = team_nl_fill_one_port_get(skb, one_port); + if (err) + goto errout; + } else if (port) { + list_for_each_entry_from(port, &team->port_list, list) { + err = team_nl_fill_one_port_get(skb, port); + if (err) { + if (err == -EMSGSIZE) { + if (!i) + goto errout; + incomplete = true; + break; + } + goto errout; + } + i++; + } + } + + nla_nest_end(skb, port_list); + genlmsg_end(skb, hdr); + if (incomplete) + goto start_again; + +send_done: + nlh = nlmsg_put(skb, portid, seq, NLMSG_DONE, 0, flags | NLM_F_MULTI); + if (!nlh) { + err = __send_and_alloc_skb(&skb, team, portid, send_func); + if (err) + return err; + goto send_done; + } + + return send_func(skb, team, portid); + +nla_put_failure: + err = -EMSGSIZE; +errout: + nlmsg_free(skb); + return err; +} + +int team_nl_port_list_get_doit(struct sk_buff *skb, + struct genl_info *info) +{ + struct team *team; + int err; + + team = team_nl_team_get(info); + if (!team) + return -EINVAL; + + err = team_nl_send_port_list_get(team, info->snd_portid, info->snd_seq, + NLM_F_ACK, team_nl_send_unicast, NULL); + + team_nl_team_put(team); + + return err; +} + +static const struct genl_multicast_group team_nl_mcgrps[] = { + { .name = TEAM_GENL_CHANGE_EVENT_MC_GRP_NAME, }, +}; + +static struct genl_family team_nl_family __ro_after_init = { + .name = TEAM_GENL_NAME, + .version = TEAM_GENL_VERSION, + .maxattr = ARRAY_SIZE(team_nl_policy) - 1, + .policy = team_nl_policy, + .netnsok = true, + .module = THIS_MODULE, + .small_ops = team_nl_ops, + .n_small_ops = ARRAY_SIZE(team_nl_ops), + .resv_start_op = TEAM_CMD_PORT_LIST_GET + 1, + .mcgrps = team_nl_mcgrps, + .n_mcgrps = ARRAY_SIZE(team_nl_mcgrps), +}; + +static int team_nl_send_multicast(struct sk_buff *skb, + struct team *team, u32 portid) +{ + return genlmsg_multicast_netns(&team_nl_family, dev_net(team->dev), + skb, 0, 0, GFP_KERNEL); +} + +static int team_nl_send_event_options_get(struct team *team, + struct list_head *sel_opt_inst_list) +{ + return team_nl_send_options_get(team, 0, 0, 0, team_nl_send_multicast, + sel_opt_inst_list); +} + +static int team_nl_send_event_port_get(struct team *team, + struct team_port *port) +{ + return team_nl_send_port_list_get(team, 0, 0, 0, team_nl_send_multicast, + port); +} + +static int __init team_nl_init(void) +{ + return genl_register_family(&team_nl_family); +} + +static void __exit team_nl_fini(void) +{ + genl_unregister_family(&team_nl_family); +} + + +/****************** + * Change checkers + ******************/ + +static void __team_options_change_check(struct team *team) +{ + int err; + struct team_option_inst *opt_inst; + LIST_HEAD(sel_opt_inst_list); + + list_for_each_entry(opt_inst, &team->option_inst_list, list) { + if (opt_inst->changed) + list_add_tail(&opt_inst->tmp_list, &sel_opt_inst_list); + } + err = team_nl_send_event_options_get(team, &sel_opt_inst_list); + if (err && err != -ESRCH) + netdev_warn(team->dev, "Failed to send options change via netlink (err %d)\n", + err); +} + +/* rtnl lock is held */ + +static void __team_port_change_send(struct team_port *port, bool linkup) +{ + int err; + + port->changed = true; + port->state.linkup = linkup; + team_refresh_port_linkup(port); + if (linkup) { + struct ethtool_link_ksettings ecmd; + + err = __ethtool_get_link_ksettings(port->dev, &ecmd); + if (!err) { + port->state.speed = ecmd.base.speed; + port->state.duplex = ecmd.base.duplex; + goto send_event; + } + } + port->state.speed = 0; + port->state.duplex = 0; + +send_event: + err = team_nl_send_event_port_get(port->team, port); + if (err && err != -ESRCH) + netdev_warn(port->team->dev, "Failed to send port change of device %s via netlink (err %d)\n", + port->dev->name, err); + +} + +static void __team_carrier_check(struct team *team) +{ + struct team_port *port; + bool team_linkup; + + if (team->user_carrier_enabled) + return; + + team_linkup = false; + list_for_each_entry(port, &team->port_list, list) { + if (port->linkup) { + team_linkup = true; + break; + } + } + + if (team_linkup) + netif_carrier_on(team->dev); + else + netif_carrier_off(team->dev); +} + +static void __team_port_change_check(struct team_port *port, bool linkup) +{ + if (port->state.linkup != linkup) + __team_port_change_send(port, linkup); + __team_carrier_check(port->team); +} + +static void __team_port_change_port_added(struct team_port *port, bool linkup) +{ + __team_port_change_send(port, linkup); + __team_carrier_check(port->team); +} + +static void __team_port_change_port_removed(struct team_port *port) +{ + port->removed = true; + __team_port_change_send(port, false); + __team_carrier_check(port->team); +} + +static void team_port_change_check(struct team_port *port, bool linkup) +{ + struct team *team = port->team; + + mutex_lock(&team->lock); + __team_port_change_check(port, linkup); + mutex_unlock(&team->lock); +} + + +/************************************ + * Net device notifier event handler + ************************************/ + +static int team_device_event(struct notifier_block *unused, + unsigned long event, void *ptr) +{ + struct net_device *dev = netdev_notifier_info_to_dev(ptr); + struct team_port *port; + + port = team_port_get_rtnl(dev); + if (!port) + return NOTIFY_DONE; + + switch (event) { + case NETDEV_UP: + if (netif_oper_up(dev)) + team_port_change_check(port, true); + break; + case NETDEV_DOWN: + team_port_change_check(port, false); + break; + case NETDEV_CHANGE: + if (netif_running(port->dev)) + team_port_change_check(port, + !!netif_oper_up(port->dev)); + break; + case NETDEV_UNREGISTER: + team_del_slave(port->team->dev, dev); + break; + case NETDEV_FEAT_CHANGE: + if (!port->team->notifier_ctx) { + port->team->notifier_ctx = true; + team_compute_features(port->team); + port->team->notifier_ctx = false; + } + break; + case NETDEV_PRECHANGEMTU: + /* Forbid to change mtu of underlaying device */ + if (!port->team->port_mtu_change_allowed) + return NOTIFY_BAD; + break; + case NETDEV_PRE_TYPE_CHANGE: + /* Forbid to change type of underlaying device */ + return NOTIFY_BAD; + case NETDEV_RESEND_IGMP: + /* Propagate to master device */ + call_netdevice_notifiers(event, port->team->dev); + break; + } + return NOTIFY_DONE; +} + +static struct notifier_block team_notifier_block __read_mostly = { + .notifier_call = team_device_event, +}; + + +/*********************** + * Module init and exit + ***********************/ + +static int __init team_module_init(void) +{ + int err; + + register_netdevice_notifier(&team_notifier_block); + + err = rtnl_link_register(&team_link_ops); + if (err) + goto err_rtnl_reg; + + err = team_nl_init(); + if (err) + goto err_nl_init; + + return 0; + +err_nl_init: + rtnl_link_unregister(&team_link_ops); + +err_rtnl_reg: + unregister_netdevice_notifier(&team_notifier_block); + + return err; +} + +static void __exit team_module_exit(void) +{ + team_nl_fini(); + rtnl_link_unregister(&team_link_ops); + unregister_netdevice_notifier(&team_notifier_block); +} + +module_init(team_module_init); +module_exit(team_module_exit); + +MODULE_LICENSE("GPL v2"); +MODULE_AUTHOR("Jiri Pirko "); +MODULE_DESCRIPTION("Ethernet team device driver"); +MODULE_ALIAS_RTNL_LINK(DRV_NAME); diff --git a/drivers/net/team/team_nl.c b/drivers/net/team/team_nl.c new file mode 100644 index 0000000000..208424ab78 --- /dev/null +++ b/drivers/net/team/team_nl.c @@ -0,0 +1,59 @@ +// SPDX-License-Identifier: ((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) +/* Do not edit directly, auto-generated from: */ +/* Documentation/netlink/specs/team.yaml */ +/* YNL-GEN kernel source */ + +#include +#include + +#include "team_nl.h" + +#include + +/* Common nested types */ +const struct nla_policy team_attr_option_nl_policy[TEAM_ATTR_OPTION_ARRAY_INDEX + 1] = { + [TEAM_ATTR_OPTION_NAME] = { .type = NLA_STRING, .len = TEAM_STRING_MAX_LEN, }, + [TEAM_ATTR_OPTION_CHANGED] = { .type = NLA_FLAG, }, + [TEAM_ATTR_OPTION_TYPE] = { .type = NLA_U8, }, + [TEAM_ATTR_OPTION_DATA] = { .type = NLA_BINARY, }, + [TEAM_ATTR_OPTION_REMOVED] = { .type = NLA_FLAG, }, + [TEAM_ATTR_OPTION_PORT_IFINDEX] = { .type = NLA_U32, }, + [TEAM_ATTR_OPTION_ARRAY_INDEX] = { .type = NLA_U32, }, +}; + +const struct nla_policy team_item_option_nl_policy[TEAM_ATTR_ITEM_OPTION + 1] = { + [TEAM_ATTR_ITEM_OPTION] = NLA_POLICY_NESTED(team_attr_option_nl_policy), +}; + +/* Global operation policy for team */ +const struct nla_policy team_nl_policy[TEAM_ATTR_LIST_OPTION + 1] = { + [TEAM_ATTR_TEAM_IFINDEX] = { .type = NLA_U32, }, + [TEAM_ATTR_LIST_OPTION] = NLA_POLICY_NESTED(team_item_option_nl_policy), +}; + +/* Ops table for team */ +const struct genl_small_ops team_nl_ops[4] = { + { + .cmd = TEAM_CMD_NOOP, + .validate = GENL_DONT_VALIDATE_STRICT, + .doit = team_nl_noop_doit, + }, + { + .cmd = TEAM_CMD_OPTIONS_SET, + .validate = GENL_DONT_VALIDATE_STRICT, + .doit = team_nl_options_set_doit, + .flags = GENL_ADMIN_PERM, + }, + { + .cmd = TEAM_CMD_OPTIONS_GET, + .validate = GENL_DONT_VALIDATE_STRICT, + .doit = team_nl_options_get_doit, + .flags = GENL_ADMIN_PERM, + }, + { + .cmd = TEAM_CMD_PORT_LIST_GET, + .validate = GENL_DONT_VALIDATE_STRICT, + .doit = team_nl_port_list_get_doit, + .flags = GENL_ADMIN_PERM, + }, +}; diff --git a/drivers/net/team/team_nl.h b/drivers/net/team/team_nl.h new file mode 100644 index 0000000000..c9ec1b22ac --- /dev/null +++ b/drivers/net/team/team_nl.h @@ -0,0 +1,29 @@ +/* SPDX-License-Identifier: ((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) */ +/* Do not edit directly, auto-generated from: */ +/* Documentation/netlink/specs/team.yaml */ +/* YNL-GEN kernel header */ + +#ifndef _LINUX_TEAM_GEN_H +#define _LINUX_TEAM_GEN_H + +#include +#include + +#include + +/* Common nested types */ +extern const struct nla_policy team_attr_option_nl_policy[TEAM_ATTR_OPTION_ARRAY_INDEX + 1]; +extern const struct nla_policy team_item_option_nl_policy[TEAM_ATTR_ITEM_OPTION + 1]; + +/* Global operation policy for team */ +extern const struct nla_policy team_nl_policy[TEAM_ATTR_LIST_OPTION + 1]; + +/* Ops table for team */ +extern const struct genl_small_ops team_nl_ops[4]; + +int team_nl_noop_doit(struct sk_buff *skb, struct genl_info *info); +int team_nl_options_set_doit(struct sk_buff *skb, struct genl_info *info); +int team_nl_options_get_doit(struct sk_buff *skb, struct genl_info *info); +int team_nl_port_list_get_doit(struct sk_buff *skb, struct genl_info *info); + +#endif /* _LINUX_TEAM_GEN_H */ -- cgit v1.2.3