1007f790cSJiri Pirko /* 2007f790cSJiri Pirko * net/switchdev/switchdev.c - Switch device API 3007f790cSJiri Pirko * Copyright (c) 2014 Jiri Pirko <[email protected]> 4f8f21471SScott Feldman * Copyright (c) 2014-2015 Scott Feldman <[email protected]> 5007f790cSJiri Pirko * 6007f790cSJiri Pirko * This program is free software; you can redistribute it and/or modify 7007f790cSJiri Pirko * it under the terms of the GNU General Public License as published by 8007f790cSJiri Pirko * the Free Software Foundation; either version 2 of the License, or 9007f790cSJiri Pirko * (at your option) any later version. 10007f790cSJiri Pirko */ 11007f790cSJiri Pirko 12007f790cSJiri Pirko #include <linux/kernel.h> 13007f790cSJiri Pirko #include <linux/types.h> 14007f790cSJiri Pirko #include <linux/init.h> 1503bf0c28SJiri Pirko #include <linux/mutex.h> 1603bf0c28SJiri Pirko #include <linux/notifier.h> 17007f790cSJiri Pirko #include <linux/netdevice.h> 1847f8328bSScott Feldman #include <linux/if_bridge.h> 195e8d9049SScott Feldman #include <net/ip_fib.h> 20007f790cSJiri Pirko #include <net/switchdev.h> 21007f790cSJiri Pirko 22007f790cSJiri Pirko /** 233094333dSScott Feldman * switchdev_port_attr_get - Get port attribute 243094333dSScott Feldman * 253094333dSScott Feldman * @dev: port device 263094333dSScott Feldman * @attr: attribute to get 273094333dSScott Feldman */ 283094333dSScott Feldman int switchdev_port_attr_get(struct net_device *dev, struct switchdev_attr *attr) 293094333dSScott Feldman { 303094333dSScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 313094333dSScott Feldman struct net_device *lower_dev; 323094333dSScott Feldman struct list_head *iter; 333094333dSScott Feldman struct switchdev_attr first = { 343094333dSScott Feldman .id = SWITCHDEV_ATTR_UNDEFINED 353094333dSScott Feldman }; 363094333dSScott Feldman int err = -EOPNOTSUPP; 373094333dSScott Feldman 383094333dSScott Feldman if (ops && ops->switchdev_port_attr_get) 393094333dSScott Feldman return ops->switchdev_port_attr_get(dev, attr); 403094333dSScott Feldman 413094333dSScott Feldman if (attr->flags & SWITCHDEV_F_NO_RECURSE) 423094333dSScott Feldman return err; 433094333dSScott Feldman 443094333dSScott Feldman /* Switch device port(s) may be stacked under 453094333dSScott Feldman * bond/team/vlan dev, so recurse down to get attr on 463094333dSScott Feldman * each port. Return -ENODATA if attr values don't 473094333dSScott Feldman * compare across ports. 483094333dSScott Feldman */ 493094333dSScott Feldman 503094333dSScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 513094333dSScott Feldman err = switchdev_port_attr_get(lower_dev, attr); 523094333dSScott Feldman if (err) 533094333dSScott Feldman break; 543094333dSScott Feldman if (first.id == SWITCHDEV_ATTR_UNDEFINED) 553094333dSScott Feldman first = *attr; 563094333dSScott Feldman else if (memcmp(&first, attr, sizeof(*attr))) 573094333dSScott Feldman return -ENODATA; 583094333dSScott Feldman } 593094333dSScott Feldman 603094333dSScott Feldman return err; 613094333dSScott Feldman } 623094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_get); 633094333dSScott Feldman 643094333dSScott Feldman static int __switchdev_port_attr_set(struct net_device *dev, 653094333dSScott Feldman struct switchdev_attr *attr) 663094333dSScott Feldman { 673094333dSScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 683094333dSScott Feldman struct net_device *lower_dev; 693094333dSScott Feldman struct list_head *iter; 703094333dSScott Feldman int err = -EOPNOTSUPP; 713094333dSScott Feldman 723094333dSScott Feldman if (ops && ops->switchdev_port_attr_set) 733094333dSScott Feldman return ops->switchdev_port_attr_set(dev, attr); 743094333dSScott Feldman 753094333dSScott Feldman if (attr->flags & SWITCHDEV_F_NO_RECURSE) 763094333dSScott Feldman return err; 773094333dSScott Feldman 783094333dSScott Feldman /* Switch device port(s) may be stacked under 793094333dSScott Feldman * bond/team/vlan dev, so recurse down to set attr on 803094333dSScott Feldman * each port. 813094333dSScott Feldman */ 823094333dSScott Feldman 833094333dSScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 843094333dSScott Feldman err = __switchdev_port_attr_set(lower_dev, attr); 853094333dSScott Feldman if (err) 863094333dSScott Feldman break; 873094333dSScott Feldman } 883094333dSScott Feldman 893094333dSScott Feldman return err; 903094333dSScott Feldman } 913094333dSScott Feldman 923094333dSScott Feldman struct switchdev_attr_set_work { 933094333dSScott Feldman struct work_struct work; 943094333dSScott Feldman struct net_device *dev; 953094333dSScott Feldman struct switchdev_attr attr; 963094333dSScott Feldman }; 973094333dSScott Feldman 983094333dSScott Feldman static void switchdev_port_attr_set_work(struct work_struct *work) 993094333dSScott Feldman { 1003094333dSScott Feldman struct switchdev_attr_set_work *asw = 1013094333dSScott Feldman container_of(work, struct switchdev_attr_set_work, work); 1023094333dSScott Feldman int err; 1033094333dSScott Feldman 1043094333dSScott Feldman rtnl_lock(); 1053094333dSScott Feldman err = switchdev_port_attr_set(asw->dev, &asw->attr); 1063094333dSScott Feldman BUG_ON(err); 1073094333dSScott Feldman rtnl_unlock(); 1083094333dSScott Feldman 1093094333dSScott Feldman dev_put(asw->dev); 1103094333dSScott Feldman kfree(work); 1113094333dSScott Feldman } 1123094333dSScott Feldman 1133094333dSScott Feldman static int switchdev_port_attr_set_defer(struct net_device *dev, 1143094333dSScott Feldman struct switchdev_attr *attr) 1153094333dSScott Feldman { 1163094333dSScott Feldman struct switchdev_attr_set_work *asw; 1173094333dSScott Feldman 1183094333dSScott Feldman asw = kmalloc(sizeof(*asw), GFP_ATOMIC); 1193094333dSScott Feldman if (!asw) 1203094333dSScott Feldman return -ENOMEM; 1213094333dSScott Feldman 1223094333dSScott Feldman INIT_WORK(&asw->work, switchdev_port_attr_set_work); 1233094333dSScott Feldman 1243094333dSScott Feldman dev_hold(dev); 1253094333dSScott Feldman asw->dev = dev; 1263094333dSScott Feldman memcpy(&asw->attr, attr, sizeof(asw->attr)); 1273094333dSScott Feldman 1283094333dSScott Feldman schedule_work(&asw->work); 1293094333dSScott Feldman 1303094333dSScott Feldman return 0; 1313094333dSScott Feldman } 1323094333dSScott Feldman 1333094333dSScott Feldman /** 1343094333dSScott Feldman * switchdev_port_attr_set - Set port attribute 1353094333dSScott Feldman * 1363094333dSScott Feldman * @dev: port device 1373094333dSScott Feldman * @attr: attribute to set 1383094333dSScott Feldman * 1393094333dSScott Feldman * Use a 2-phase prepare-commit transaction model to ensure 1403094333dSScott Feldman * system is not left in a partially updated state due to 1413094333dSScott Feldman * failure from driver/device. 1423094333dSScott Feldman */ 1433094333dSScott Feldman int switchdev_port_attr_set(struct net_device *dev, struct switchdev_attr *attr) 1443094333dSScott Feldman { 1453094333dSScott Feldman int err; 1463094333dSScott Feldman 1473094333dSScott Feldman if (!rtnl_is_locked()) { 1483094333dSScott Feldman /* Running prepare-commit transaction across stacked 1493094333dSScott Feldman * devices requires nothing moves, so if rtnl_lock is 1503094333dSScott Feldman * not held, schedule a worker thread to hold rtnl_lock 1513094333dSScott Feldman * while setting attr. 1523094333dSScott Feldman */ 1533094333dSScott Feldman 1543094333dSScott Feldman return switchdev_port_attr_set_defer(dev, attr); 1553094333dSScott Feldman } 1563094333dSScott Feldman 1573094333dSScott Feldman /* Phase I: prepare for attr set. Driver/device should fail 1583094333dSScott Feldman * here if there are going to be issues in the commit phase, 1593094333dSScott Feldman * such as lack of resources or support. The driver/device 1603094333dSScott Feldman * should reserve resources needed for the commit phase here, 1613094333dSScott Feldman * but should not commit the attr. 1623094333dSScott Feldman */ 1633094333dSScott Feldman 1643094333dSScott Feldman attr->trans = SWITCHDEV_TRANS_PREPARE; 1653094333dSScott Feldman err = __switchdev_port_attr_set(dev, attr); 1663094333dSScott Feldman if (err) { 1673094333dSScott Feldman /* Prepare phase failed: abort the transaction. Any 1683094333dSScott Feldman * resources reserved in the prepare phase are 1693094333dSScott Feldman * released. 1703094333dSScott Feldman */ 1713094333dSScott Feldman 1723094333dSScott Feldman attr->trans = SWITCHDEV_TRANS_ABORT; 1733094333dSScott Feldman __switchdev_port_attr_set(dev, attr); 1743094333dSScott Feldman 1753094333dSScott Feldman return err; 1763094333dSScott Feldman } 1773094333dSScott Feldman 1783094333dSScott Feldman /* Phase II: commit attr set. This cannot fail as a fault 1793094333dSScott Feldman * of driver/device. If it does, it's a bug in the driver/device 1803094333dSScott Feldman * because the driver said everythings was OK in phase I. 1813094333dSScott Feldman */ 1823094333dSScott Feldman 1833094333dSScott Feldman attr->trans = SWITCHDEV_TRANS_COMMIT; 1843094333dSScott Feldman err = __switchdev_port_attr_set(dev, attr); 1853094333dSScott Feldman BUG_ON(err); 1863094333dSScott Feldman 1873094333dSScott Feldman return err; 1883094333dSScott Feldman } 1893094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_set); 1903094333dSScott Feldman 19122c1f67eSScott Feldman static int __switchdev_port_obj_add(struct net_device *dev, 19222c1f67eSScott Feldman struct switchdev_obj *obj) 193491d0f15SScott Feldman { 194491d0f15SScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 195491d0f15SScott Feldman struct net_device *lower_dev; 196491d0f15SScott Feldman struct list_head *iter; 197491d0f15SScott Feldman int err = -EOPNOTSUPP; 198491d0f15SScott Feldman 199491d0f15SScott Feldman if (ops && ops->switchdev_port_obj_add) 200491d0f15SScott Feldman return ops->switchdev_port_obj_add(dev, obj); 201491d0f15SScott Feldman 202491d0f15SScott Feldman /* Switch device port(s) may be stacked under 203491d0f15SScott Feldman * bond/team/vlan dev, so recurse down to add object on 204491d0f15SScott Feldman * each port. 205491d0f15SScott Feldman */ 206491d0f15SScott Feldman 207491d0f15SScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 208491d0f15SScott Feldman err = __switchdev_port_obj_add(lower_dev, obj); 209491d0f15SScott Feldman if (err) 210491d0f15SScott Feldman break; 211491d0f15SScott Feldman } 212491d0f15SScott Feldman 213491d0f15SScott Feldman return err; 214491d0f15SScott Feldman } 215491d0f15SScott Feldman 216491d0f15SScott Feldman /** 217491d0f15SScott Feldman * switchdev_port_obj_add - Add port object 218491d0f15SScott Feldman * 219491d0f15SScott Feldman * @dev: port device 220491d0f15SScott Feldman * @obj: object to add 221491d0f15SScott Feldman * 222491d0f15SScott Feldman * Use a 2-phase prepare-commit transaction model to ensure 223491d0f15SScott Feldman * system is not left in a partially updated state due to 224491d0f15SScott Feldman * failure from driver/device. 225491d0f15SScott Feldman * 226491d0f15SScott Feldman * rtnl_lock must be held. 227491d0f15SScott Feldman */ 228491d0f15SScott Feldman int switchdev_port_obj_add(struct net_device *dev, struct switchdev_obj *obj) 229491d0f15SScott Feldman { 230491d0f15SScott Feldman int err; 231491d0f15SScott Feldman 232491d0f15SScott Feldman ASSERT_RTNL(); 233491d0f15SScott Feldman 234491d0f15SScott Feldman /* Phase I: prepare for obj add. Driver/device should fail 235491d0f15SScott Feldman * here if there are going to be issues in the commit phase, 236491d0f15SScott Feldman * such as lack of resources or support. The driver/device 237491d0f15SScott Feldman * should reserve resources needed for the commit phase here, 238491d0f15SScott Feldman * but should not commit the obj. 239491d0f15SScott Feldman */ 240491d0f15SScott Feldman 241491d0f15SScott Feldman obj->trans = SWITCHDEV_TRANS_PREPARE; 242491d0f15SScott Feldman err = __switchdev_port_obj_add(dev, obj); 243491d0f15SScott Feldman if (err) { 244491d0f15SScott Feldman /* Prepare phase failed: abort the transaction. Any 245491d0f15SScott Feldman * resources reserved in the prepare phase are 246491d0f15SScott Feldman * released. 247491d0f15SScott Feldman */ 248491d0f15SScott Feldman 249491d0f15SScott Feldman obj->trans = SWITCHDEV_TRANS_ABORT; 250491d0f15SScott Feldman __switchdev_port_obj_add(dev, obj); 251491d0f15SScott Feldman 252491d0f15SScott Feldman return err; 253491d0f15SScott Feldman } 254491d0f15SScott Feldman 255491d0f15SScott Feldman /* Phase II: commit obj add. This cannot fail as a fault 256491d0f15SScott Feldman * of driver/device. If it does, it's a bug in the driver/device 257491d0f15SScott Feldman * because the driver said everythings was OK in phase I. 258491d0f15SScott Feldman */ 259491d0f15SScott Feldman 260491d0f15SScott Feldman obj->trans = SWITCHDEV_TRANS_COMMIT; 261491d0f15SScott Feldman err = __switchdev_port_obj_add(dev, obj); 262491d0f15SScott Feldman WARN(err, "%s: Commit of object (id=%d) failed.\n", dev->name, obj->id); 263491d0f15SScott Feldman 264491d0f15SScott Feldman return err; 265491d0f15SScott Feldman } 266491d0f15SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_obj_add); 267491d0f15SScott Feldman 268491d0f15SScott Feldman /** 269491d0f15SScott Feldman * switchdev_port_obj_del - Delete port object 270491d0f15SScott Feldman * 271491d0f15SScott Feldman * @dev: port device 272491d0f15SScott Feldman * @obj: object to delete 273491d0f15SScott Feldman */ 274491d0f15SScott Feldman int switchdev_port_obj_del(struct net_device *dev, struct switchdev_obj *obj) 275491d0f15SScott Feldman { 276491d0f15SScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 277491d0f15SScott Feldman struct net_device *lower_dev; 278491d0f15SScott Feldman struct list_head *iter; 279491d0f15SScott Feldman int err = -EOPNOTSUPP; 280491d0f15SScott Feldman 281491d0f15SScott Feldman if (ops && ops->switchdev_port_obj_del) 282491d0f15SScott Feldman return ops->switchdev_port_obj_del(dev, obj); 283491d0f15SScott Feldman 284491d0f15SScott Feldman /* Switch device port(s) may be stacked under 285491d0f15SScott Feldman * bond/team/vlan dev, so recurse down to delete object on 286491d0f15SScott Feldman * each port. 287491d0f15SScott Feldman */ 288491d0f15SScott Feldman 289491d0f15SScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 290491d0f15SScott Feldman err = switchdev_port_obj_del(lower_dev, obj); 291491d0f15SScott Feldman if (err) 292491d0f15SScott Feldman break; 293491d0f15SScott Feldman } 294491d0f15SScott Feldman 295491d0f15SScott Feldman return err; 296491d0f15SScott Feldman } 297491d0f15SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_obj_del); 298491d0f15SScott Feldman 299ebb9a03aSJiri Pirko static DEFINE_MUTEX(switchdev_mutex); 300ebb9a03aSJiri Pirko static RAW_NOTIFIER_HEAD(switchdev_notif_chain); 30103bf0c28SJiri Pirko 30203bf0c28SJiri Pirko /** 303ebb9a03aSJiri Pirko * register_switchdev_notifier - Register notifier 30403bf0c28SJiri Pirko * @nb: notifier_block 30503bf0c28SJiri Pirko * 30603bf0c28SJiri Pirko * Register switch device notifier. This should be used by code 30703bf0c28SJiri Pirko * which needs to monitor events happening in particular device. 30803bf0c28SJiri Pirko * Return values are same as for atomic_notifier_chain_register(). 30903bf0c28SJiri Pirko */ 310ebb9a03aSJiri Pirko int register_switchdev_notifier(struct notifier_block *nb) 31103bf0c28SJiri Pirko { 31203bf0c28SJiri Pirko int err; 31303bf0c28SJiri Pirko 314ebb9a03aSJiri Pirko mutex_lock(&switchdev_mutex); 315ebb9a03aSJiri Pirko err = raw_notifier_chain_register(&switchdev_notif_chain, nb); 316ebb9a03aSJiri Pirko mutex_unlock(&switchdev_mutex); 31703bf0c28SJiri Pirko return err; 31803bf0c28SJiri Pirko } 319ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(register_switchdev_notifier); 32003bf0c28SJiri Pirko 32103bf0c28SJiri Pirko /** 322ebb9a03aSJiri Pirko * unregister_switchdev_notifier - Unregister notifier 32303bf0c28SJiri Pirko * @nb: notifier_block 32403bf0c28SJiri Pirko * 32503bf0c28SJiri Pirko * Unregister switch device notifier. 32603bf0c28SJiri Pirko * Return values are same as for atomic_notifier_chain_unregister(). 32703bf0c28SJiri Pirko */ 328ebb9a03aSJiri Pirko int unregister_switchdev_notifier(struct notifier_block *nb) 32903bf0c28SJiri Pirko { 33003bf0c28SJiri Pirko int err; 33103bf0c28SJiri Pirko 332ebb9a03aSJiri Pirko mutex_lock(&switchdev_mutex); 333ebb9a03aSJiri Pirko err = raw_notifier_chain_unregister(&switchdev_notif_chain, nb); 334ebb9a03aSJiri Pirko mutex_unlock(&switchdev_mutex); 33503bf0c28SJiri Pirko return err; 33603bf0c28SJiri Pirko } 337ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(unregister_switchdev_notifier); 33803bf0c28SJiri Pirko 33903bf0c28SJiri Pirko /** 340ebb9a03aSJiri Pirko * call_switchdev_notifiers - Call notifiers 34103bf0c28SJiri Pirko * @val: value passed unmodified to notifier function 34203bf0c28SJiri Pirko * @dev: port device 34303bf0c28SJiri Pirko * @info: notifier information data 34403bf0c28SJiri Pirko * 34503bf0c28SJiri Pirko * Call all network notifier blocks. This should be called by driver 34603bf0c28SJiri Pirko * when it needs to propagate hardware event. 34703bf0c28SJiri Pirko * Return values are same as for atomic_notifier_call_chain(). 34803bf0c28SJiri Pirko */ 349ebb9a03aSJiri Pirko int call_switchdev_notifiers(unsigned long val, struct net_device *dev, 350ebb9a03aSJiri Pirko struct switchdev_notifier_info *info) 35103bf0c28SJiri Pirko { 35203bf0c28SJiri Pirko int err; 35303bf0c28SJiri Pirko 35403bf0c28SJiri Pirko info->dev = dev; 355ebb9a03aSJiri Pirko mutex_lock(&switchdev_mutex); 356ebb9a03aSJiri Pirko err = raw_notifier_call_chain(&switchdev_notif_chain, val, info); 357ebb9a03aSJiri Pirko mutex_unlock(&switchdev_mutex); 35803bf0c28SJiri Pirko return err; 35903bf0c28SJiri Pirko } 360ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(call_switchdev_notifiers); 3618a44dbb2SRoopa Prabhu 3628793d0a6SScott Feldman /** 3638793d0a6SScott Feldman * switchdev_port_bridge_getlink - Get bridge port attributes 3648793d0a6SScott Feldman * 3658793d0a6SScott Feldman * @dev: port device 3668793d0a6SScott Feldman * 3678793d0a6SScott Feldman * Called for SELF on rtnl_bridge_getlink to get bridge port 3688793d0a6SScott Feldman * attributes. 3698793d0a6SScott Feldman */ 3708793d0a6SScott Feldman int switchdev_port_bridge_getlink(struct sk_buff *skb, u32 pid, u32 seq, 3718793d0a6SScott Feldman struct net_device *dev, u32 filter_mask, 3728793d0a6SScott Feldman int nlflags) 3738793d0a6SScott Feldman { 3748793d0a6SScott Feldman struct switchdev_attr attr = { 3758793d0a6SScott Feldman .id = SWITCHDEV_ATTR_PORT_BRIDGE_FLAGS, 3768793d0a6SScott Feldman }; 3778793d0a6SScott Feldman u16 mode = BRIDGE_MODE_UNDEF; 3788793d0a6SScott Feldman u32 mask = BR_LEARNING | BR_LEARNING_SYNC; 3798793d0a6SScott Feldman int err; 3808793d0a6SScott Feldman 3818793d0a6SScott Feldman err = switchdev_port_attr_get(dev, &attr); 3828793d0a6SScott Feldman if (err) 3838793d0a6SScott Feldman return err; 3848793d0a6SScott Feldman 3858793d0a6SScott Feldman return ndo_dflt_bridge_getlink(skb, pid, seq, dev, mode, 386*42275bd8SScott Feldman attr.u.brport_flags, mask, nlflags); 3878793d0a6SScott Feldman } 3888793d0a6SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_bridge_getlink); 3898793d0a6SScott Feldman 39047f8328bSScott Feldman static int switchdev_port_br_setflag(struct net_device *dev, 39147f8328bSScott Feldman struct nlattr *nlattr, 39247f8328bSScott Feldman unsigned long brport_flag) 39347f8328bSScott Feldman { 39447f8328bSScott Feldman struct switchdev_attr attr = { 39547f8328bSScott Feldman .id = SWITCHDEV_ATTR_PORT_BRIDGE_FLAGS, 39647f8328bSScott Feldman }; 39747f8328bSScott Feldman u8 flag = nla_get_u8(nlattr); 39847f8328bSScott Feldman int err; 39947f8328bSScott Feldman 40047f8328bSScott Feldman err = switchdev_port_attr_get(dev, &attr); 40147f8328bSScott Feldman if (err) 40247f8328bSScott Feldman return err; 40347f8328bSScott Feldman 40447f8328bSScott Feldman if (flag) 405*42275bd8SScott Feldman attr.u.brport_flags |= brport_flag; 40647f8328bSScott Feldman else 407*42275bd8SScott Feldman attr.u.brport_flags &= ~brport_flag; 40847f8328bSScott Feldman 40947f8328bSScott Feldman return switchdev_port_attr_set(dev, &attr); 41047f8328bSScott Feldman } 41147f8328bSScott Feldman 41247f8328bSScott Feldman static const struct nla_policy 41347f8328bSScott Feldman switchdev_port_bridge_policy[IFLA_BRPORT_MAX + 1] = { 41447f8328bSScott Feldman [IFLA_BRPORT_STATE] = { .type = NLA_U8 }, 41547f8328bSScott Feldman [IFLA_BRPORT_COST] = { .type = NLA_U32 }, 41647f8328bSScott Feldman [IFLA_BRPORT_PRIORITY] = { .type = NLA_U16 }, 41747f8328bSScott Feldman [IFLA_BRPORT_MODE] = { .type = NLA_U8 }, 41847f8328bSScott Feldman [IFLA_BRPORT_GUARD] = { .type = NLA_U8 }, 41947f8328bSScott Feldman [IFLA_BRPORT_PROTECT] = { .type = NLA_U8 }, 42047f8328bSScott Feldman [IFLA_BRPORT_FAST_LEAVE] = { .type = NLA_U8 }, 42147f8328bSScott Feldman [IFLA_BRPORT_LEARNING] = { .type = NLA_U8 }, 42247f8328bSScott Feldman [IFLA_BRPORT_LEARNING_SYNC] = { .type = NLA_U8 }, 42347f8328bSScott Feldman [IFLA_BRPORT_UNICAST_FLOOD] = { .type = NLA_U8 }, 42447f8328bSScott Feldman }; 42547f8328bSScott Feldman 42647f8328bSScott Feldman static int switchdev_port_br_setlink_protinfo(struct net_device *dev, 42747f8328bSScott Feldman struct nlattr *protinfo) 42847f8328bSScott Feldman { 42947f8328bSScott Feldman struct nlattr *attr; 43047f8328bSScott Feldman int rem; 43147f8328bSScott Feldman int err; 43247f8328bSScott Feldman 43347f8328bSScott Feldman err = nla_validate_nested(protinfo, IFLA_BRPORT_MAX, 43447f8328bSScott Feldman switchdev_port_bridge_policy); 43547f8328bSScott Feldman if (err) 43647f8328bSScott Feldman return err; 43747f8328bSScott Feldman 43847f8328bSScott Feldman nla_for_each_nested(attr, protinfo, rem) { 43947f8328bSScott Feldman switch (nla_type(attr)) { 44047f8328bSScott Feldman case IFLA_BRPORT_LEARNING: 44147f8328bSScott Feldman err = switchdev_port_br_setflag(dev, attr, 44247f8328bSScott Feldman BR_LEARNING); 44347f8328bSScott Feldman break; 44447f8328bSScott Feldman case IFLA_BRPORT_LEARNING_SYNC: 44547f8328bSScott Feldman err = switchdev_port_br_setflag(dev, attr, 44647f8328bSScott Feldman BR_LEARNING_SYNC); 44747f8328bSScott Feldman break; 44847f8328bSScott Feldman default: 44947f8328bSScott Feldman err = -EOPNOTSUPP; 45047f8328bSScott Feldman break; 45147f8328bSScott Feldman } 45247f8328bSScott Feldman if (err) 45347f8328bSScott Feldman return err; 45447f8328bSScott Feldman } 45547f8328bSScott Feldman 45647f8328bSScott Feldman return 0; 45747f8328bSScott Feldman } 45847f8328bSScott Feldman 45947f8328bSScott Feldman static int switchdev_port_br_afspec(struct net_device *dev, 46047f8328bSScott Feldman struct nlattr *afspec, 46147f8328bSScott Feldman int (*f)(struct net_device *dev, 46247f8328bSScott Feldman struct switchdev_obj *obj)) 46347f8328bSScott Feldman { 46447f8328bSScott Feldman struct nlattr *attr; 46547f8328bSScott Feldman struct bridge_vlan_info *vinfo; 46647f8328bSScott Feldman struct switchdev_obj obj = { 46747f8328bSScott Feldman .id = SWITCHDEV_OBJ_PORT_VLAN, 46847f8328bSScott Feldman }; 469*42275bd8SScott Feldman struct switchdev_obj_vlan *vlan = &obj.u.vlan; 47047f8328bSScott Feldman int rem; 47147f8328bSScott Feldman int err; 47247f8328bSScott Feldman 47347f8328bSScott Feldman nla_for_each_nested(attr, afspec, rem) { 47447f8328bSScott Feldman if (nla_type(attr) != IFLA_BRIDGE_VLAN_INFO) 47547f8328bSScott Feldman continue; 47647f8328bSScott Feldman if (nla_len(attr) != sizeof(struct bridge_vlan_info)) 47747f8328bSScott Feldman return -EINVAL; 47847f8328bSScott Feldman vinfo = nla_data(attr); 479*42275bd8SScott Feldman vlan->flags = vinfo->flags; 48047f8328bSScott Feldman if (vinfo->flags & BRIDGE_VLAN_INFO_RANGE_BEGIN) { 481*42275bd8SScott Feldman if (vlan->vid_start) 48247f8328bSScott Feldman return -EINVAL; 483*42275bd8SScott Feldman vlan->vid_start = vinfo->vid; 48447f8328bSScott Feldman } else if (vinfo->flags & BRIDGE_VLAN_INFO_RANGE_END) { 485*42275bd8SScott Feldman if (!vlan->vid_start) 48647f8328bSScott Feldman return -EINVAL; 487*42275bd8SScott Feldman vlan->vid_end = vinfo->vid; 488*42275bd8SScott Feldman if (vlan->vid_end <= vlan->vid_start) 48947f8328bSScott Feldman return -EINVAL; 49047f8328bSScott Feldman err = f(dev, &obj); 49147f8328bSScott Feldman if (err) 49247f8328bSScott Feldman return err; 493*42275bd8SScott Feldman memset(vlan, 0, sizeof(*vlan)); 49447f8328bSScott Feldman } else { 495*42275bd8SScott Feldman if (vlan->vid_start) 49647f8328bSScott Feldman return -EINVAL; 497*42275bd8SScott Feldman vlan->vid_start = vinfo->vid; 498*42275bd8SScott Feldman vlan->vid_end = vinfo->vid; 49947f8328bSScott Feldman err = f(dev, &obj); 50047f8328bSScott Feldman if (err) 50147f8328bSScott Feldman return err; 502*42275bd8SScott Feldman memset(vlan, 0, sizeof(*vlan)); 50347f8328bSScott Feldman } 50447f8328bSScott Feldman } 50547f8328bSScott Feldman 50647f8328bSScott Feldman return 0; 50747f8328bSScott Feldman } 50847f8328bSScott Feldman 5098a44dbb2SRoopa Prabhu /** 51047f8328bSScott Feldman * switchdev_port_bridge_setlink - Set bridge port attributes 5118a44dbb2SRoopa Prabhu * 5128a44dbb2SRoopa Prabhu * @dev: port device 51347f8328bSScott Feldman * @nlh: netlink header 51447f8328bSScott Feldman * @flags: netlink flags 5158a44dbb2SRoopa Prabhu * 51647f8328bSScott Feldman * Called for SELF on rtnl_bridge_setlink to set bridge port 51747f8328bSScott Feldman * attributes. 5188a44dbb2SRoopa Prabhu */ 519ebb9a03aSJiri Pirko int switchdev_port_bridge_setlink(struct net_device *dev, 5208a44dbb2SRoopa Prabhu struct nlmsghdr *nlh, u16 flags) 5218a44dbb2SRoopa Prabhu { 52247f8328bSScott Feldman struct nlattr *protinfo; 52347f8328bSScott Feldman struct nlattr *afspec; 52447f8328bSScott Feldman int err = 0; 5258a44dbb2SRoopa Prabhu 52647f8328bSScott Feldman protinfo = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), 52747f8328bSScott Feldman IFLA_PROTINFO); 52847f8328bSScott Feldman if (protinfo) { 52947f8328bSScott Feldman err = switchdev_port_br_setlink_protinfo(dev, protinfo); 53047f8328bSScott Feldman if (err) 53147f8328bSScott Feldman return err; 53247f8328bSScott Feldman } 5338a44dbb2SRoopa Prabhu 53447f8328bSScott Feldman afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), 53547f8328bSScott Feldman IFLA_AF_SPEC); 53647f8328bSScott Feldman if (afspec) 53747f8328bSScott Feldman err = switchdev_port_br_afspec(dev, afspec, 53847f8328bSScott Feldman switchdev_port_obj_add); 5398a44dbb2SRoopa Prabhu 54047f8328bSScott Feldman return err; 5418a44dbb2SRoopa Prabhu } 542ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_port_bridge_setlink); 5438a44dbb2SRoopa Prabhu 5448a44dbb2SRoopa Prabhu /** 5455c34e022SScott Feldman * switchdev_port_bridge_dellink - Set bridge port attributes 5468a44dbb2SRoopa Prabhu * 5478a44dbb2SRoopa Prabhu * @dev: port device 5485c34e022SScott Feldman * @nlh: netlink header 5495c34e022SScott Feldman * @flags: netlink flags 5508a44dbb2SRoopa Prabhu * 5515c34e022SScott Feldman * Called for SELF on rtnl_bridge_dellink to set bridge port 5525c34e022SScott Feldman * attributes. 5538a44dbb2SRoopa Prabhu */ 554ebb9a03aSJiri Pirko int switchdev_port_bridge_dellink(struct net_device *dev, 5558a44dbb2SRoopa Prabhu struct nlmsghdr *nlh, u16 flags) 5568a44dbb2SRoopa Prabhu { 5575c34e022SScott Feldman struct nlattr *afspec; 5588a44dbb2SRoopa Prabhu 5595c34e022SScott Feldman afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), 5605c34e022SScott Feldman IFLA_AF_SPEC); 5615c34e022SScott Feldman if (afspec) 5625c34e022SScott Feldman return switchdev_port_br_afspec(dev, afspec, 5635c34e022SScott Feldman switchdev_port_obj_del); 5645c34e022SScott Feldman 5658a44dbb2SRoopa Prabhu return 0; 5668a44dbb2SRoopa Prabhu } 567ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_port_bridge_dellink); 5688a44dbb2SRoopa Prabhu 569ebb9a03aSJiri Pirko static struct net_device *switchdev_get_lowest_dev(struct net_device *dev) 570b5d6fbdeSScott Feldman { 5719d47c0a2SJiri Pirko const struct switchdev_ops *ops = dev->switchdev_ops; 572b5d6fbdeSScott Feldman struct net_device *lower_dev; 573b5d6fbdeSScott Feldman struct net_device *port_dev; 574b5d6fbdeSScott Feldman struct list_head *iter; 575b5d6fbdeSScott Feldman 576b5d6fbdeSScott Feldman /* Recusively search down until we find a sw port dev. 577f8e20a9fSScott Feldman * (A sw port dev supports switchdev_port_attr_get). 578b5d6fbdeSScott Feldman */ 579b5d6fbdeSScott Feldman 580f8e20a9fSScott Feldman if (ops && ops->switchdev_port_attr_get) 581b5d6fbdeSScott Feldman return dev; 582b5d6fbdeSScott Feldman 583b5d6fbdeSScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 584ebb9a03aSJiri Pirko port_dev = switchdev_get_lowest_dev(lower_dev); 585b5d6fbdeSScott Feldman if (port_dev) 586b5d6fbdeSScott Feldman return port_dev; 587b5d6fbdeSScott Feldman } 588b5d6fbdeSScott Feldman 589b5d6fbdeSScott Feldman return NULL; 590b5d6fbdeSScott Feldman } 591b5d6fbdeSScott Feldman 592ebb9a03aSJiri Pirko static struct net_device *switchdev_get_dev_by_nhs(struct fib_info *fi) 593b5d6fbdeSScott Feldman { 594f8e20a9fSScott Feldman struct switchdev_attr attr = { 595f8e20a9fSScott Feldman .id = SWITCHDEV_ATTR_PORT_PARENT_ID, 596f8e20a9fSScott Feldman }; 597f8e20a9fSScott Feldman struct switchdev_attr prev_attr; 598b5d6fbdeSScott Feldman struct net_device *dev = NULL; 599b5d6fbdeSScott Feldman int nhsel; 600b5d6fbdeSScott Feldman 601b5d6fbdeSScott Feldman /* For this route, all nexthop devs must be on the same switch. */ 602b5d6fbdeSScott Feldman 603b5d6fbdeSScott Feldman for (nhsel = 0; nhsel < fi->fib_nhs; nhsel++) { 604b5d6fbdeSScott Feldman const struct fib_nh *nh = &fi->fib_nh[nhsel]; 605b5d6fbdeSScott Feldman 606b5d6fbdeSScott Feldman if (!nh->nh_dev) 607b5d6fbdeSScott Feldman return NULL; 608b5d6fbdeSScott Feldman 609ebb9a03aSJiri Pirko dev = switchdev_get_lowest_dev(nh->nh_dev); 610b5d6fbdeSScott Feldman if (!dev) 611b5d6fbdeSScott Feldman return NULL; 612b5d6fbdeSScott Feldman 613f8e20a9fSScott Feldman if (switchdev_port_attr_get(dev, &attr)) 614b5d6fbdeSScott Feldman return NULL; 615b5d6fbdeSScott Feldman 616b5d6fbdeSScott Feldman if (nhsel > 0) { 617*42275bd8SScott Feldman if (prev_attr.u.ppid.id_len != attr.u.ppid.id_len) 618b5d6fbdeSScott Feldman return NULL; 619*42275bd8SScott Feldman if (memcmp(prev_attr.u.ppid.id, attr.u.ppid.id, 620*42275bd8SScott Feldman attr.u.ppid.id_len)) 621b5d6fbdeSScott Feldman return NULL; 622b5d6fbdeSScott Feldman } 623b5d6fbdeSScott Feldman 624f8e20a9fSScott Feldman prev_attr = attr; 625b5d6fbdeSScott Feldman } 626b5d6fbdeSScott Feldman 627b5d6fbdeSScott Feldman return dev; 628b5d6fbdeSScott Feldman } 629b5d6fbdeSScott Feldman 6305e8d9049SScott Feldman /** 631ebb9a03aSJiri Pirko * switchdev_fib_ipv4_add - Add IPv4 route entry to switch 6325e8d9049SScott Feldman * 6335e8d9049SScott Feldman * @dst: route's IPv4 destination address 6345e8d9049SScott Feldman * @dst_len: destination address length (prefix length) 6355e8d9049SScott Feldman * @fi: route FIB info structure 6365e8d9049SScott Feldman * @tos: route TOS 6375e8d9049SScott Feldman * @type: route type 638f8f21471SScott Feldman * @nlflags: netlink flags passed in (NLM_F_*) 6395e8d9049SScott Feldman * @tb_id: route table ID 6405e8d9049SScott Feldman * 6415e8d9049SScott Feldman * Add IPv4 route entry to switch device. 6425e8d9049SScott Feldman */ 643ebb9a03aSJiri Pirko int switchdev_fib_ipv4_add(u32 dst, int dst_len, struct fib_info *fi, 644f8f21471SScott Feldman u8 tos, u8 type, u32 nlflags, u32 tb_id) 6455e8d9049SScott Feldman { 64658c2cb16SScott Feldman struct switchdev_obj fib_obj = { 64758c2cb16SScott Feldman .id = SWITCHDEV_OBJ_IPV4_FIB, 648*42275bd8SScott Feldman .u.ipv4_fib = { 6497a7ee531SScott Feldman .dst = dst, 65058c2cb16SScott Feldman .dst_len = dst_len, 65158c2cb16SScott Feldman .fi = fi, 65258c2cb16SScott Feldman .tos = tos, 65358c2cb16SScott Feldman .type = type, 65458c2cb16SScott Feldman .nlflags = nlflags, 65558c2cb16SScott Feldman .tb_id = tb_id, 65658c2cb16SScott Feldman }, 65758c2cb16SScott Feldman }; 658b5d6fbdeSScott Feldman struct net_device *dev; 659b5d6fbdeSScott Feldman int err = 0; 660b5d6fbdeSScott Feldman 6618e05fd71SScott Feldman /* Don't offload route if using custom ip rules or if 6628e05fd71SScott Feldman * IPv4 FIB offloading has been disabled completely. 6638e05fd71SScott Feldman */ 6648e05fd71SScott Feldman 665e1315db1SScott Feldman #ifdef CONFIG_IP_MULTIPLE_TABLES 666e1315db1SScott Feldman if (fi->fib_net->ipv4.fib_has_custom_rules) 667e1315db1SScott Feldman return 0; 668e1315db1SScott Feldman #endif 669e1315db1SScott Feldman 670e1315db1SScott Feldman if (fi->fib_net->ipv4.fib_offload_disabled) 671104616e7SScott Feldman return 0; 672104616e7SScott Feldman 673ebb9a03aSJiri Pirko dev = switchdev_get_dev_by_nhs(fi); 674b5d6fbdeSScott Feldman if (!dev) 6755e8d9049SScott Feldman return 0; 676b5d6fbdeSScott Feldman 67758c2cb16SScott Feldman err = switchdev_port_obj_add(dev, &fib_obj); 678b5d6fbdeSScott Feldman if (!err) 679b5d6fbdeSScott Feldman fi->fib_flags |= RTNH_F_EXTERNAL; 680b5d6fbdeSScott Feldman 681b5d6fbdeSScott Feldman return err; 6825e8d9049SScott Feldman } 683ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_fib_ipv4_add); 6845e8d9049SScott Feldman 6855e8d9049SScott Feldman /** 686ebb9a03aSJiri Pirko * switchdev_fib_ipv4_del - Delete IPv4 route entry from switch 6875e8d9049SScott Feldman * 6885e8d9049SScott Feldman * @dst: route's IPv4 destination address 6895e8d9049SScott Feldman * @dst_len: destination address length (prefix length) 6905e8d9049SScott Feldman * @fi: route FIB info structure 6915e8d9049SScott Feldman * @tos: route TOS 6925e8d9049SScott Feldman * @type: route type 6935e8d9049SScott Feldman * @tb_id: route table ID 6945e8d9049SScott Feldman * 6955e8d9049SScott Feldman * Delete IPv4 route entry from switch device. 6965e8d9049SScott Feldman */ 697ebb9a03aSJiri Pirko int switchdev_fib_ipv4_del(u32 dst, int dst_len, struct fib_info *fi, 6985e8d9049SScott Feldman u8 tos, u8 type, u32 tb_id) 6995e8d9049SScott Feldman { 70058c2cb16SScott Feldman struct switchdev_obj fib_obj = { 70158c2cb16SScott Feldman .id = SWITCHDEV_OBJ_IPV4_FIB, 702*42275bd8SScott Feldman .u.ipv4_fib = { 7037a7ee531SScott Feldman .dst = dst, 70458c2cb16SScott Feldman .dst_len = dst_len, 70558c2cb16SScott Feldman .fi = fi, 70658c2cb16SScott Feldman .tos = tos, 70758c2cb16SScott Feldman .type = type, 70858c2cb16SScott Feldman .nlflags = 0, 70958c2cb16SScott Feldman .tb_id = tb_id, 71058c2cb16SScott Feldman }, 71158c2cb16SScott Feldman }; 712b5d6fbdeSScott Feldman struct net_device *dev; 713b5d6fbdeSScott Feldman int err = 0; 714b5d6fbdeSScott Feldman 715b5d6fbdeSScott Feldman if (!(fi->fib_flags & RTNH_F_EXTERNAL)) 7165e8d9049SScott Feldman return 0; 717b5d6fbdeSScott Feldman 718ebb9a03aSJiri Pirko dev = switchdev_get_dev_by_nhs(fi); 719b5d6fbdeSScott Feldman if (!dev) 720b5d6fbdeSScott Feldman return 0; 721b5d6fbdeSScott Feldman 72258c2cb16SScott Feldman err = switchdev_port_obj_del(dev, &fib_obj); 723b5d6fbdeSScott Feldman if (!err) 724b5d6fbdeSScott Feldman fi->fib_flags &= ~RTNH_F_EXTERNAL; 725b5d6fbdeSScott Feldman 726b5d6fbdeSScott Feldman return err; 7275e8d9049SScott Feldman } 728ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_fib_ipv4_del); 7298e05fd71SScott Feldman 7308e05fd71SScott Feldman /** 731ebb9a03aSJiri Pirko * switchdev_fib_ipv4_abort - Abort an IPv4 FIB operation 7328e05fd71SScott Feldman * 7338e05fd71SScott Feldman * @fi: route FIB info structure 7348e05fd71SScott Feldman */ 735ebb9a03aSJiri Pirko void switchdev_fib_ipv4_abort(struct fib_info *fi) 7368e05fd71SScott Feldman { 7378e05fd71SScott Feldman /* There was a problem installing this route to the offload 7388e05fd71SScott Feldman * device. For now, until we come up with more refined 7398e05fd71SScott Feldman * policy handling, abruptly end IPv4 fib offloading for 7408e05fd71SScott Feldman * for entire net by flushing offload device(s) of all 7418e05fd71SScott Feldman * IPv4 routes, and mark IPv4 fib offloading broken from 7428e05fd71SScott Feldman * this point forward. 7438e05fd71SScott Feldman */ 7448e05fd71SScott Feldman 7458e05fd71SScott Feldman fib_flush_external(fi->fib_net); 7468e05fd71SScott Feldman fi->fib_net->ipv4.fib_offload_disabled = true; 7478e05fd71SScott Feldman } 748ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_fib_ipv4_abort); 749