12874c5fdSThomas Gleixner // SPDX-License-Identifier: GPL-2.0-or-later 2007f790cSJiri Pirko /* 3007f790cSJiri Pirko * net/switchdev/switchdev.c - Switch device API 47ea6eb3fSJiri Pirko * Copyright (c) 2014-2015 Jiri Pirko <[email protected]> 5f8f21471SScott Feldman * Copyright (c) 2014-2015 Scott Feldman <[email protected]> 6007f790cSJiri Pirko */ 7007f790cSJiri Pirko 8007f790cSJiri Pirko #include <linux/kernel.h> 9007f790cSJiri Pirko #include <linux/types.h> 10007f790cSJiri Pirko #include <linux/init.h> 1103bf0c28SJiri Pirko #include <linux/mutex.h> 1203bf0c28SJiri Pirko #include <linux/notifier.h> 13007f790cSJiri Pirko #include <linux/netdevice.h> 14850d0cbcSJiri Pirko #include <linux/etherdevice.h> 1547f8328bSScott Feldman #include <linux/if_bridge.h> 167ea6eb3fSJiri Pirko #include <linux/list.h> 17793f4014SJiri Pirko #include <linux/workqueue.h> 1887aaf2caSNikolay Aleksandrov #include <linux/if_vlan.h> 194f2c6ae5SIdo Schimmel #include <linux/rtnetlink.h> 20007f790cSJiri Pirko #include <net/switchdev.h> 21007f790cSJiri Pirko 22793f4014SJiri Pirko static LIST_HEAD(deferred); 23793f4014SJiri Pirko static DEFINE_SPINLOCK(deferred_lock); 24793f4014SJiri Pirko 25793f4014SJiri Pirko typedef void switchdev_deferred_func_t(struct net_device *dev, 26793f4014SJiri Pirko const void *data); 27793f4014SJiri Pirko 28793f4014SJiri Pirko struct switchdev_deferred_item { 29793f4014SJiri Pirko struct list_head list; 30793f4014SJiri Pirko struct net_device *dev; 31793f4014SJiri Pirko switchdev_deferred_func_t *func; 32fbfc8502SGustavo A. R. Silva unsigned long data[]; 33793f4014SJiri Pirko }; 34793f4014SJiri Pirko 35793f4014SJiri Pirko static struct switchdev_deferred_item *switchdev_deferred_dequeue(void) 36793f4014SJiri Pirko { 37793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 38793f4014SJiri Pirko 39793f4014SJiri Pirko spin_lock_bh(&deferred_lock); 40793f4014SJiri Pirko if (list_empty(&deferred)) { 41793f4014SJiri Pirko dfitem = NULL; 42793f4014SJiri Pirko goto unlock; 43793f4014SJiri Pirko } 44793f4014SJiri Pirko dfitem = list_first_entry(&deferred, 45793f4014SJiri Pirko struct switchdev_deferred_item, list); 46793f4014SJiri Pirko list_del(&dfitem->list); 47793f4014SJiri Pirko unlock: 48793f4014SJiri Pirko spin_unlock_bh(&deferred_lock); 49793f4014SJiri Pirko return dfitem; 50793f4014SJiri Pirko } 51793f4014SJiri Pirko 52793f4014SJiri Pirko /** 53793f4014SJiri Pirko * switchdev_deferred_process - Process ops in deferred queue 54793f4014SJiri Pirko * 55793f4014SJiri Pirko * Called to flush the ops currently queued in deferred ops queue. 56793f4014SJiri Pirko * rtnl_lock must be held. 57793f4014SJiri Pirko */ 58793f4014SJiri Pirko void switchdev_deferred_process(void) 59793f4014SJiri Pirko { 60793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 61793f4014SJiri Pirko 62793f4014SJiri Pirko ASSERT_RTNL(); 63793f4014SJiri Pirko 64793f4014SJiri Pirko while ((dfitem = switchdev_deferred_dequeue())) { 65793f4014SJiri Pirko dfitem->func(dfitem->dev, dfitem->data); 66793f4014SJiri Pirko dev_put(dfitem->dev); 67793f4014SJiri Pirko kfree(dfitem); 68793f4014SJiri Pirko } 69793f4014SJiri Pirko } 70793f4014SJiri Pirko EXPORT_SYMBOL_GPL(switchdev_deferred_process); 71793f4014SJiri Pirko 72793f4014SJiri Pirko static void switchdev_deferred_process_work(struct work_struct *work) 73793f4014SJiri Pirko { 74793f4014SJiri Pirko rtnl_lock(); 75793f4014SJiri Pirko switchdev_deferred_process(); 76793f4014SJiri Pirko rtnl_unlock(); 77793f4014SJiri Pirko } 78793f4014SJiri Pirko 79793f4014SJiri Pirko static DECLARE_WORK(deferred_process_work, switchdev_deferred_process_work); 80793f4014SJiri Pirko 81793f4014SJiri Pirko static int switchdev_deferred_enqueue(struct net_device *dev, 82793f4014SJiri Pirko const void *data, size_t data_len, 83793f4014SJiri Pirko switchdev_deferred_func_t *func) 84793f4014SJiri Pirko { 85793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 86793f4014SJiri Pirko 87793f4014SJiri Pirko dfitem = kmalloc(sizeof(*dfitem) + data_len, GFP_ATOMIC); 88793f4014SJiri Pirko if (!dfitem) 89793f4014SJiri Pirko return -ENOMEM; 90793f4014SJiri Pirko dfitem->dev = dev; 91793f4014SJiri Pirko dfitem->func = func; 92793f4014SJiri Pirko memcpy(dfitem->data, data, data_len); 93793f4014SJiri Pirko dev_hold(dev); 94793f4014SJiri Pirko spin_lock_bh(&deferred_lock); 95793f4014SJiri Pirko list_add_tail(&dfitem->list, &deferred); 96793f4014SJiri Pirko spin_unlock_bh(&deferred_lock); 97793f4014SJiri Pirko schedule_work(&deferred_process_work); 98793f4014SJiri Pirko return 0; 99793f4014SJiri Pirko } 100793f4014SJiri Pirko 101d45224d6SFlorian Fainelli static int switchdev_port_attr_notify(enum switchdev_notifier_type nt, 102d45224d6SFlorian Fainelli struct net_device *dev, 103bae33f2bSVladimir Oltean const struct switchdev_attr *attr) 1043094333dSScott Feldman { 105d45224d6SFlorian Fainelli int err; 106d45224d6SFlorian Fainelli int rc; 1073094333dSScott Feldman 108d45224d6SFlorian Fainelli struct switchdev_notifier_port_attr_info attr_info = { 109d45224d6SFlorian Fainelli .attr = attr, 110d45224d6SFlorian Fainelli .handled = false, 111d45224d6SFlorian Fainelli }; 1123094333dSScott Feldman 113d45224d6SFlorian Fainelli rc = call_switchdev_blocking_notifiers(nt, dev, 114d45224d6SFlorian Fainelli &attr_info.info, NULL); 115d45224d6SFlorian Fainelli err = notifier_to_errno(rc); 116d45224d6SFlorian Fainelli if (err) { 117d45224d6SFlorian Fainelli WARN_ON(!attr_info.handled); 1183094333dSScott Feldman return err; 1193094333dSScott Feldman } 1203094333dSScott Feldman 121d45224d6SFlorian Fainelli if (!attr_info.handled) 122d45224d6SFlorian Fainelli return -EOPNOTSUPP; 123d45224d6SFlorian Fainelli 124d45224d6SFlorian Fainelli return 0; 125d45224d6SFlorian Fainelli } 126d45224d6SFlorian Fainelli 1270bc05d58SJiri Pirko static int switchdev_port_attr_set_now(struct net_device *dev, 128f7fadf30SJiri Pirko const struct switchdev_attr *attr) 1293094333dSScott Feldman { 130bae33f2bSVladimir Oltean return switchdev_port_attr_notify(SWITCHDEV_PORT_ATTR_SET, dev, attr); 1313094333dSScott Feldman } 1320bc05d58SJiri Pirko 1330bc05d58SJiri Pirko static void switchdev_port_attr_set_deferred(struct net_device *dev, 1340bc05d58SJiri Pirko const void *data) 1350bc05d58SJiri Pirko { 1360bc05d58SJiri Pirko const struct switchdev_attr *attr = data; 1370bc05d58SJiri Pirko int err; 1380bc05d58SJiri Pirko 1390bc05d58SJiri Pirko err = switchdev_port_attr_set_now(dev, attr); 1400bc05d58SJiri Pirko if (err && err != -EOPNOTSUPP) 1410bc05d58SJiri Pirko netdev_err(dev, "failed (err=%d) to set attribute (id=%d)\n", 1420bc05d58SJiri Pirko err, attr->id); 1437ceb2afbSElad Raz if (attr->complete) 1447ceb2afbSElad Raz attr->complete(dev, err, attr->complete_priv); 1450bc05d58SJiri Pirko } 1460bc05d58SJiri Pirko 1470bc05d58SJiri Pirko static int switchdev_port_attr_set_defer(struct net_device *dev, 1480bc05d58SJiri Pirko const struct switchdev_attr *attr) 1490bc05d58SJiri Pirko { 1500bc05d58SJiri Pirko return switchdev_deferred_enqueue(dev, attr, sizeof(*attr), 1510bc05d58SJiri Pirko switchdev_port_attr_set_deferred); 1520bc05d58SJiri Pirko } 1530bc05d58SJiri Pirko 1540bc05d58SJiri Pirko /** 1550bc05d58SJiri Pirko * switchdev_port_attr_set - Set port attribute 1560bc05d58SJiri Pirko * 1570bc05d58SJiri Pirko * @dev: port device 1580bc05d58SJiri Pirko * @attr: attribute to set 1590bc05d58SJiri Pirko * 1600bc05d58SJiri Pirko * rtnl_lock must be held and must not be in atomic section, 1610bc05d58SJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 1620bc05d58SJiri Pirko */ 1630bc05d58SJiri Pirko int switchdev_port_attr_set(struct net_device *dev, 1640bc05d58SJiri Pirko const struct switchdev_attr *attr) 1650bc05d58SJiri Pirko { 1660bc05d58SJiri Pirko if (attr->flags & SWITCHDEV_F_DEFER) 1670bc05d58SJiri Pirko return switchdev_port_attr_set_defer(dev, attr); 1680bc05d58SJiri Pirko ASSERT_RTNL(); 1690bc05d58SJiri Pirko return switchdev_port_attr_set_now(dev, attr); 1700bc05d58SJiri Pirko } 1713094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_set); 1723094333dSScott Feldman 173e258d919SScott Feldman static size_t switchdev_obj_size(const struct switchdev_obj *obj) 174e258d919SScott Feldman { 175e258d919SScott Feldman switch (obj->id) { 176e258d919SScott Feldman case SWITCHDEV_OBJ_ID_PORT_VLAN: 177e258d919SScott Feldman return sizeof(struct switchdev_obj_port_vlan); 1784d41e125SElad Raz case SWITCHDEV_OBJ_ID_PORT_MDB: 1794d41e125SElad Raz return sizeof(struct switchdev_obj_port_mdb); 18047d5b6dbSAndrew Lunn case SWITCHDEV_OBJ_ID_HOST_MDB: 18147d5b6dbSAndrew Lunn return sizeof(struct switchdev_obj_port_mdb); 182e258d919SScott Feldman default: 183e258d919SScott Feldman BUG(); 184e258d919SScott Feldman } 185e258d919SScott Feldman return 0; 186e258d919SScott Feldman } 187e258d919SScott Feldman 188d17d9f5eSPetr Machata static int switchdev_port_obj_notify(enum switchdev_notifier_type nt, 189d17d9f5eSPetr Machata struct net_device *dev, 190648b4a99SJiri Pirko const struct switchdev_obj *obj, 19169b7320eSPetr Machata struct netlink_ext_ack *extack) 192491d0f15SScott Feldman { 193d17d9f5eSPetr Machata int rc; 194d17d9f5eSPetr Machata int err; 195491d0f15SScott Feldman 196d17d9f5eSPetr Machata struct switchdev_notifier_port_obj_info obj_info = { 197d17d9f5eSPetr Machata .obj = obj, 198d17d9f5eSPetr Machata .handled = false, 199d17d9f5eSPetr Machata }; 200491d0f15SScott Feldman 201479c86dcSPetr Machata rc = call_switchdev_blocking_notifiers(nt, dev, &obj_info.info, extack); 202d17d9f5eSPetr Machata err = notifier_to_errno(rc); 203d17d9f5eSPetr Machata if (err) { 204d17d9f5eSPetr Machata WARN_ON(!obj_info.handled); 205491d0f15SScott Feldman return err; 206491d0f15SScott Feldman } 207d17d9f5eSPetr Machata if (!obj_info.handled) 208d17d9f5eSPetr Machata return -EOPNOTSUPP; 209d17d9f5eSPetr Machata return 0; 210d17d9f5eSPetr Machata } 211491d0f15SScott Feldman 2124d429c5dSJiri Pirko static void switchdev_port_obj_add_deferred(struct net_device *dev, 2134d429c5dSJiri Pirko const void *data) 2144d429c5dSJiri Pirko { 2154d429c5dSJiri Pirko const struct switchdev_obj *obj = data; 2164d429c5dSJiri Pirko int err; 2174d429c5dSJiri Pirko 218cf6def51SVladimir Oltean ASSERT_RTNL(); 219cf6def51SVladimir Oltean err = switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD, 220cf6def51SVladimir Oltean dev, obj, NULL); 2214d429c5dSJiri Pirko if (err && err != -EOPNOTSUPP) 2224d429c5dSJiri Pirko netdev_err(dev, "failed (err=%d) to add object (id=%d)\n", 2234d429c5dSJiri Pirko err, obj->id); 2247ceb2afbSElad Raz if (obj->complete) 2257ceb2afbSElad Raz obj->complete(dev, err, obj->complete_priv); 2264d429c5dSJiri Pirko } 2274d429c5dSJiri Pirko 2284d429c5dSJiri Pirko static int switchdev_port_obj_add_defer(struct net_device *dev, 2294d429c5dSJiri Pirko const struct switchdev_obj *obj) 2304d429c5dSJiri Pirko { 231e258d919SScott Feldman return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj), 2324d429c5dSJiri Pirko switchdev_port_obj_add_deferred); 2334d429c5dSJiri Pirko } 234491d0f15SScott Feldman 235491d0f15SScott Feldman /** 2364d429c5dSJiri Pirko * switchdev_port_obj_add - Add port object 237491d0f15SScott Feldman * 238491d0f15SScott Feldman * @dev: port device 2394d429c5dSJiri Pirko * @obj: object to add 240c8af73f0SAndrew Lunn * @extack: netlink extended ack 2414d429c5dSJiri Pirko * 2424d429c5dSJiri Pirko * rtnl_lock must be held and must not be in atomic section, 2434d429c5dSJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 244491d0f15SScott Feldman */ 2454d429c5dSJiri Pirko int switchdev_port_obj_add(struct net_device *dev, 24669b7320eSPetr Machata const struct switchdev_obj *obj, 24769b7320eSPetr Machata struct netlink_ext_ack *extack) 2484d429c5dSJiri Pirko { 2494d429c5dSJiri Pirko if (obj->flags & SWITCHDEV_F_DEFER) 2504d429c5dSJiri Pirko return switchdev_port_obj_add_defer(dev, obj); 2514d429c5dSJiri Pirko ASSERT_RTNL(); 252cf6def51SVladimir Oltean return switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD, 253cf6def51SVladimir Oltean dev, obj, extack); 2544d429c5dSJiri Pirko } 2554d429c5dSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_port_obj_add); 2564d429c5dSJiri Pirko 2574d429c5dSJiri Pirko static int switchdev_port_obj_del_now(struct net_device *dev, 258648b4a99SJiri Pirko const struct switchdev_obj *obj) 259491d0f15SScott Feldman { 260d17d9f5eSPetr Machata return switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_DEL, 261ffb68fc5SVladimir Oltean dev, obj, NULL); 262491d0f15SScott Feldman } 2634d429c5dSJiri Pirko 2644d429c5dSJiri Pirko static void switchdev_port_obj_del_deferred(struct net_device *dev, 2654d429c5dSJiri Pirko const void *data) 2664d429c5dSJiri Pirko { 2674d429c5dSJiri Pirko const struct switchdev_obj *obj = data; 2684d429c5dSJiri Pirko int err; 2694d429c5dSJiri Pirko 2704d429c5dSJiri Pirko err = switchdev_port_obj_del_now(dev, obj); 2714d429c5dSJiri Pirko if (err && err != -EOPNOTSUPP) 2724d429c5dSJiri Pirko netdev_err(dev, "failed (err=%d) to del object (id=%d)\n", 2734d429c5dSJiri Pirko err, obj->id); 2747ceb2afbSElad Raz if (obj->complete) 2757ceb2afbSElad Raz obj->complete(dev, err, obj->complete_priv); 2764d429c5dSJiri Pirko } 2774d429c5dSJiri Pirko 2784d429c5dSJiri Pirko static int switchdev_port_obj_del_defer(struct net_device *dev, 2794d429c5dSJiri Pirko const struct switchdev_obj *obj) 2804d429c5dSJiri Pirko { 281e258d919SScott Feldman return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj), 2824d429c5dSJiri Pirko switchdev_port_obj_del_deferred); 2834d429c5dSJiri Pirko } 2844d429c5dSJiri Pirko 2854d429c5dSJiri Pirko /** 2864d429c5dSJiri Pirko * switchdev_port_obj_del - Delete port object 2874d429c5dSJiri Pirko * 2884d429c5dSJiri Pirko * @dev: port device 2894d429c5dSJiri Pirko * @obj: object to delete 2904d429c5dSJiri Pirko * 2914d429c5dSJiri Pirko * rtnl_lock must be held and must not be in atomic section, 2924d429c5dSJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 2934d429c5dSJiri Pirko */ 2944d429c5dSJiri Pirko int switchdev_port_obj_del(struct net_device *dev, 2954d429c5dSJiri Pirko const struct switchdev_obj *obj) 2964d429c5dSJiri Pirko { 2974d429c5dSJiri Pirko if (obj->flags & SWITCHDEV_F_DEFER) 2984d429c5dSJiri Pirko return switchdev_port_obj_del_defer(dev, obj); 2994d429c5dSJiri Pirko ASSERT_RTNL(); 3004d429c5dSJiri Pirko return switchdev_port_obj_del_now(dev, obj); 3014d429c5dSJiri Pirko } 302491d0f15SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_obj_del); 303491d0f15SScott Feldman 304ff5cf100SArkadi Sharshevsky static ATOMIC_NOTIFIER_HEAD(switchdev_notif_chain); 305a93e3b17SPetr Machata static BLOCKING_NOTIFIER_HEAD(switchdev_blocking_notif_chain); 30603bf0c28SJiri Pirko 30703bf0c28SJiri Pirko /** 308ebb9a03aSJiri Pirko * register_switchdev_notifier - Register notifier 30903bf0c28SJiri Pirko * @nb: notifier_block 31003bf0c28SJiri Pirko * 311ff5cf100SArkadi Sharshevsky * Register switch device notifier. 31203bf0c28SJiri Pirko */ 313ebb9a03aSJiri Pirko int register_switchdev_notifier(struct notifier_block *nb) 31403bf0c28SJiri Pirko { 315ff5cf100SArkadi Sharshevsky return atomic_notifier_chain_register(&switchdev_notif_chain, nb); 31603bf0c28SJiri Pirko } 317ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(register_switchdev_notifier); 31803bf0c28SJiri Pirko 31903bf0c28SJiri Pirko /** 320ebb9a03aSJiri Pirko * unregister_switchdev_notifier - Unregister notifier 32103bf0c28SJiri Pirko * @nb: notifier_block 32203bf0c28SJiri Pirko * 32303bf0c28SJiri Pirko * Unregister switch device notifier. 32403bf0c28SJiri Pirko */ 325ebb9a03aSJiri Pirko int unregister_switchdev_notifier(struct notifier_block *nb) 32603bf0c28SJiri Pirko { 327ff5cf100SArkadi Sharshevsky return atomic_notifier_chain_unregister(&switchdev_notif_chain, nb); 32803bf0c28SJiri Pirko } 329ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(unregister_switchdev_notifier); 33003bf0c28SJiri Pirko 33103bf0c28SJiri Pirko /** 332ebb9a03aSJiri Pirko * call_switchdev_notifiers - Call notifiers 33303bf0c28SJiri Pirko * @val: value passed unmodified to notifier function 33403bf0c28SJiri Pirko * @dev: port device 33503bf0c28SJiri Pirko * @info: notifier information data 336ea6754aeSTian Tao * @extack: netlink extended ack 337ff5cf100SArkadi Sharshevsky * Call all network notifier blocks. 33803bf0c28SJiri Pirko */ 339ebb9a03aSJiri Pirko int call_switchdev_notifiers(unsigned long val, struct net_device *dev, 3406685987cSPetr Machata struct switchdev_notifier_info *info, 3416685987cSPetr Machata struct netlink_ext_ack *extack) 34203bf0c28SJiri Pirko { 34303bf0c28SJiri Pirko info->dev = dev; 3446685987cSPetr Machata info->extack = extack; 345ff5cf100SArkadi Sharshevsky return atomic_notifier_call_chain(&switchdev_notif_chain, val, info); 34603bf0c28SJiri Pirko } 347ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(call_switchdev_notifiers); 3488a44dbb2SRoopa Prabhu 349a93e3b17SPetr Machata int register_switchdev_blocking_notifier(struct notifier_block *nb) 350a93e3b17SPetr Machata { 351a93e3b17SPetr Machata struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain; 352a93e3b17SPetr Machata 353a93e3b17SPetr Machata return blocking_notifier_chain_register(chain, nb); 354a93e3b17SPetr Machata } 355a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(register_switchdev_blocking_notifier); 356a93e3b17SPetr Machata 357a93e3b17SPetr Machata int unregister_switchdev_blocking_notifier(struct notifier_block *nb) 358a93e3b17SPetr Machata { 359a93e3b17SPetr Machata struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain; 360a93e3b17SPetr Machata 361a93e3b17SPetr Machata return blocking_notifier_chain_unregister(chain, nb); 362a93e3b17SPetr Machata } 363a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(unregister_switchdev_blocking_notifier); 364a93e3b17SPetr Machata 365a93e3b17SPetr Machata int call_switchdev_blocking_notifiers(unsigned long val, struct net_device *dev, 366479c86dcSPetr Machata struct switchdev_notifier_info *info, 367479c86dcSPetr Machata struct netlink_ext_ack *extack) 368a93e3b17SPetr Machata { 369a93e3b17SPetr Machata info->dev = dev; 370479c86dcSPetr Machata info->extack = extack; 371a93e3b17SPetr Machata return blocking_notifier_call_chain(&switchdev_blocking_notif_chain, 372a93e3b17SPetr Machata val, info); 373a93e3b17SPetr Machata } 374a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(call_switchdev_blocking_notifiers); 375a93e3b17SPetr Machata 376f30f0601SPetr Machata static int __switchdev_handle_port_obj_add(struct net_device *dev, 377f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 378f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 379f30f0601SPetr Machata int (*add_cb)(struct net_device *dev, 380f30f0601SPetr Machata const struct switchdev_obj *obj, 38169213513SPetr Machata struct netlink_ext_ack *extack)) 382f30f0601SPetr Machata { 38369213513SPetr Machata struct netlink_ext_ack *extack; 384f30f0601SPetr Machata struct net_device *lower_dev; 385f30f0601SPetr Machata struct list_head *iter; 386f30f0601SPetr Machata int err = -EOPNOTSUPP; 387f30f0601SPetr Machata 38869213513SPetr Machata extack = switchdev_notifier_info_to_extack(&port_obj_info->info); 38969213513SPetr Machata 390f30f0601SPetr Machata if (check_cb(dev)) { 391c358f952SJakub Kicinski err = add_cb(dev, port_obj_info->obj, extack); 392*20776b46SRasmus Villemoes if (err != -EOPNOTSUPP) 393f30f0601SPetr Machata port_obj_info->handled = true; 394*20776b46SRasmus Villemoes return err; 395f30f0601SPetr Machata } 396f30f0601SPetr Machata 397f30f0601SPetr Machata /* Switch ports might be stacked under e.g. a LAG. Ignore the 398f30f0601SPetr Machata * unsupported devices, another driver might be able to handle them. But 399f30f0601SPetr Machata * propagate to the callers any hard errors. 400f30f0601SPetr Machata * 401f30f0601SPetr Machata * If the driver does its own bookkeeping of stacked ports, it's not 402f30f0601SPetr Machata * necessary to go through this helper. 403f30f0601SPetr Machata */ 404f30f0601SPetr Machata netdev_for_each_lower_dev(dev, lower_dev, iter) { 40507c6f980SRussell King if (netif_is_bridge_master(lower_dev)) 40607c6f980SRussell King continue; 40707c6f980SRussell King 408f30f0601SPetr Machata err = __switchdev_handle_port_obj_add(lower_dev, port_obj_info, 409f30f0601SPetr Machata check_cb, add_cb); 410f30f0601SPetr Machata if (err && err != -EOPNOTSUPP) 411f30f0601SPetr Machata return err; 412f30f0601SPetr Machata } 413f30f0601SPetr Machata 414f30f0601SPetr Machata return err; 415f30f0601SPetr Machata } 416f30f0601SPetr Machata 417f30f0601SPetr Machata int switchdev_handle_port_obj_add(struct net_device *dev, 418f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 419f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 420f30f0601SPetr Machata int (*add_cb)(struct net_device *dev, 421f30f0601SPetr Machata const struct switchdev_obj *obj, 42269213513SPetr Machata struct netlink_ext_ack *extack)) 423f30f0601SPetr Machata { 424f30f0601SPetr Machata int err; 425f30f0601SPetr Machata 426f30f0601SPetr Machata err = __switchdev_handle_port_obj_add(dev, port_obj_info, check_cb, 427f30f0601SPetr Machata add_cb); 428f30f0601SPetr Machata if (err == -EOPNOTSUPP) 429f30f0601SPetr Machata err = 0; 430f30f0601SPetr Machata return err; 431f30f0601SPetr Machata } 432f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_add); 433f30f0601SPetr Machata 434f30f0601SPetr Machata static int __switchdev_handle_port_obj_del(struct net_device *dev, 435f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 436f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 437f30f0601SPetr Machata int (*del_cb)(struct net_device *dev, 438f30f0601SPetr Machata const struct switchdev_obj *obj)) 439f30f0601SPetr Machata { 440f30f0601SPetr Machata struct net_device *lower_dev; 441f30f0601SPetr Machata struct list_head *iter; 442f30f0601SPetr Machata int err = -EOPNOTSUPP; 443f30f0601SPetr Machata 444f30f0601SPetr Machata if (check_cb(dev)) { 445*20776b46SRasmus Villemoes err = del_cb(dev, port_obj_info->obj); 446*20776b46SRasmus Villemoes if (err != -EOPNOTSUPP) 447f30f0601SPetr Machata port_obj_info->handled = true; 448*20776b46SRasmus Villemoes return err; 449f30f0601SPetr Machata } 450f30f0601SPetr Machata 451f30f0601SPetr Machata /* Switch ports might be stacked under e.g. a LAG. Ignore the 452f30f0601SPetr Machata * unsupported devices, another driver might be able to handle them. But 453f30f0601SPetr Machata * propagate to the callers any hard errors. 454f30f0601SPetr Machata * 455f30f0601SPetr Machata * If the driver does its own bookkeeping of stacked ports, it's not 456f30f0601SPetr Machata * necessary to go through this helper. 457f30f0601SPetr Machata */ 458f30f0601SPetr Machata netdev_for_each_lower_dev(dev, lower_dev, iter) { 45907c6f980SRussell King if (netif_is_bridge_master(lower_dev)) 46007c6f980SRussell King continue; 46107c6f980SRussell King 462f30f0601SPetr Machata err = __switchdev_handle_port_obj_del(lower_dev, port_obj_info, 463f30f0601SPetr Machata check_cb, del_cb); 464f30f0601SPetr Machata if (err && err != -EOPNOTSUPP) 465f30f0601SPetr Machata return err; 466f30f0601SPetr Machata } 467f30f0601SPetr Machata 468f30f0601SPetr Machata return err; 469f30f0601SPetr Machata } 470f30f0601SPetr Machata 471f30f0601SPetr Machata int switchdev_handle_port_obj_del(struct net_device *dev, 472f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 473f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 474f30f0601SPetr Machata int (*del_cb)(struct net_device *dev, 475f30f0601SPetr Machata const struct switchdev_obj *obj)) 476f30f0601SPetr Machata { 477f30f0601SPetr Machata int err; 478f30f0601SPetr Machata 479f30f0601SPetr Machata err = __switchdev_handle_port_obj_del(dev, port_obj_info, check_cb, 480f30f0601SPetr Machata del_cb); 481f30f0601SPetr Machata if (err == -EOPNOTSUPP) 482f30f0601SPetr Machata err = 0; 483f30f0601SPetr Machata return err; 484f30f0601SPetr Machata } 485f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_del); 4861cb33af1SFlorian Fainelli 4871cb33af1SFlorian Fainelli static int __switchdev_handle_port_attr_set(struct net_device *dev, 4881cb33af1SFlorian Fainelli struct switchdev_notifier_port_attr_info *port_attr_info, 4891cb33af1SFlorian Fainelli bool (*check_cb)(const struct net_device *dev), 4901cb33af1SFlorian Fainelli int (*set_cb)(struct net_device *dev, 491bae33f2bSVladimir Oltean const struct switchdev_attr *attr)) 4921cb33af1SFlorian Fainelli { 4931cb33af1SFlorian Fainelli struct net_device *lower_dev; 4941cb33af1SFlorian Fainelli struct list_head *iter; 4951cb33af1SFlorian Fainelli int err = -EOPNOTSUPP; 4961cb33af1SFlorian Fainelli 4971cb33af1SFlorian Fainelli if (check_cb(dev)) { 498c358f952SJakub Kicinski err = set_cb(dev, port_attr_info->attr); 499*20776b46SRasmus Villemoes if (err != -EOPNOTSUPP) 5001cb33af1SFlorian Fainelli port_attr_info->handled = true; 501*20776b46SRasmus Villemoes return err; 5021cb33af1SFlorian Fainelli } 5031cb33af1SFlorian Fainelli 5041cb33af1SFlorian Fainelli /* Switch ports might be stacked under e.g. a LAG. Ignore the 5051cb33af1SFlorian Fainelli * unsupported devices, another driver might be able to handle them. But 5061cb33af1SFlorian Fainelli * propagate to the callers any hard errors. 5071cb33af1SFlorian Fainelli * 5081cb33af1SFlorian Fainelli * If the driver does its own bookkeeping of stacked ports, it's not 5091cb33af1SFlorian Fainelli * necessary to go through this helper. 5101cb33af1SFlorian Fainelli */ 5111cb33af1SFlorian Fainelli netdev_for_each_lower_dev(dev, lower_dev, iter) { 51207c6f980SRussell King if (netif_is_bridge_master(lower_dev)) 51307c6f980SRussell King continue; 51407c6f980SRussell King 5151cb33af1SFlorian Fainelli err = __switchdev_handle_port_attr_set(lower_dev, port_attr_info, 5161cb33af1SFlorian Fainelli check_cb, set_cb); 5171cb33af1SFlorian Fainelli if (err && err != -EOPNOTSUPP) 5181cb33af1SFlorian Fainelli return err; 5191cb33af1SFlorian Fainelli } 5201cb33af1SFlorian Fainelli 5211cb33af1SFlorian Fainelli return err; 5221cb33af1SFlorian Fainelli } 5231cb33af1SFlorian Fainelli 5241cb33af1SFlorian Fainelli int switchdev_handle_port_attr_set(struct net_device *dev, 5251cb33af1SFlorian Fainelli struct switchdev_notifier_port_attr_info *port_attr_info, 5261cb33af1SFlorian Fainelli bool (*check_cb)(const struct net_device *dev), 5271cb33af1SFlorian Fainelli int (*set_cb)(struct net_device *dev, 528bae33f2bSVladimir Oltean const struct switchdev_attr *attr)) 5291cb33af1SFlorian Fainelli { 5301cb33af1SFlorian Fainelli int err; 5311cb33af1SFlorian Fainelli 5321cb33af1SFlorian Fainelli err = __switchdev_handle_port_attr_set(dev, port_attr_info, check_cb, 5331cb33af1SFlorian Fainelli set_cb); 5341cb33af1SFlorian Fainelli if (err == -EOPNOTSUPP) 5351cb33af1SFlorian Fainelli err = 0; 5361cb33af1SFlorian Fainelli return err; 5371cb33af1SFlorian Fainelli } 5381cb33af1SFlorian Fainelli EXPORT_SYMBOL_GPL(switchdev_handle_port_attr_set); 539