1007f790cSJiri Pirko /* 2007f790cSJiri Pirko * net/switchdev/switchdev.c - Switch device API 37ea6eb3fSJiri Pirko * Copyright (c) 2014-2015 Jiri Pirko <[email protected]> 4f8f21471SScott Feldman * Copyright (c) 2014-2015 Scott Feldman <[email protected]> 5007f790cSJiri Pirko * 6007f790cSJiri Pirko * This program is free software; you can redistribute it and/or modify 7007f790cSJiri Pirko * it under the terms of the GNU General Public License as published by 8007f790cSJiri Pirko * the Free Software Foundation; either version 2 of the License, or 9007f790cSJiri Pirko * (at your option) any later version. 10007f790cSJiri Pirko */ 11007f790cSJiri Pirko 12007f790cSJiri Pirko #include <linux/kernel.h> 13007f790cSJiri Pirko #include <linux/types.h> 14007f790cSJiri Pirko #include <linux/init.h> 1503bf0c28SJiri Pirko #include <linux/mutex.h> 1603bf0c28SJiri Pirko #include <linux/notifier.h> 17007f790cSJiri Pirko #include <linux/netdevice.h> 18850d0cbcSJiri Pirko #include <linux/etherdevice.h> 1947f8328bSScott Feldman #include <linux/if_bridge.h> 207ea6eb3fSJiri Pirko #include <linux/list.h> 21793f4014SJiri Pirko #include <linux/workqueue.h> 2287aaf2caSNikolay Aleksandrov #include <linux/if_vlan.h> 234f2c6ae5SIdo Schimmel #include <linux/rtnetlink.h> 24007f790cSJiri Pirko #include <net/switchdev.h> 25007f790cSJiri Pirko 26007f790cSJiri Pirko /** 277ea6eb3fSJiri Pirko * switchdev_trans_item_enqueue - Enqueue data item to transaction queue 287ea6eb3fSJiri Pirko * 297ea6eb3fSJiri Pirko * @trans: transaction 307ea6eb3fSJiri Pirko * @data: pointer to data being queued 317ea6eb3fSJiri Pirko * @destructor: data destructor 327ea6eb3fSJiri Pirko * @tritem: transaction item being queued 337ea6eb3fSJiri Pirko * 347ea6eb3fSJiri Pirko * Enqeueue data item to transaction queue. tritem is typically placed in 357ea6eb3fSJiri Pirko * cointainter pointed at by data pointer. Destructor is called on 367ea6eb3fSJiri Pirko * transaction abort and after successful commit phase in case 377ea6eb3fSJiri Pirko * the caller did not dequeue the item before. 387ea6eb3fSJiri Pirko */ 397ea6eb3fSJiri Pirko void switchdev_trans_item_enqueue(struct switchdev_trans *trans, 407ea6eb3fSJiri Pirko void *data, void (*destructor)(void const *), 417ea6eb3fSJiri Pirko struct switchdev_trans_item *tritem) 427ea6eb3fSJiri Pirko { 437ea6eb3fSJiri Pirko tritem->data = data; 447ea6eb3fSJiri Pirko tritem->destructor = destructor; 457ea6eb3fSJiri Pirko list_add_tail(&tritem->list, &trans->item_list); 467ea6eb3fSJiri Pirko } 477ea6eb3fSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_trans_item_enqueue); 487ea6eb3fSJiri Pirko 497ea6eb3fSJiri Pirko static struct switchdev_trans_item * 507ea6eb3fSJiri Pirko __switchdev_trans_item_dequeue(struct switchdev_trans *trans) 517ea6eb3fSJiri Pirko { 527ea6eb3fSJiri Pirko struct switchdev_trans_item *tritem; 537ea6eb3fSJiri Pirko 547ea6eb3fSJiri Pirko if (list_empty(&trans->item_list)) 557ea6eb3fSJiri Pirko return NULL; 567ea6eb3fSJiri Pirko tritem = list_first_entry(&trans->item_list, 577ea6eb3fSJiri Pirko struct switchdev_trans_item, list); 587ea6eb3fSJiri Pirko list_del(&tritem->list); 597ea6eb3fSJiri Pirko return tritem; 607ea6eb3fSJiri Pirko } 617ea6eb3fSJiri Pirko 627ea6eb3fSJiri Pirko /** 637ea6eb3fSJiri Pirko * switchdev_trans_item_dequeue - Dequeue data item from transaction queue 647ea6eb3fSJiri Pirko * 657ea6eb3fSJiri Pirko * @trans: transaction 667ea6eb3fSJiri Pirko */ 677ea6eb3fSJiri Pirko void *switchdev_trans_item_dequeue(struct switchdev_trans *trans) 687ea6eb3fSJiri Pirko { 697ea6eb3fSJiri Pirko struct switchdev_trans_item *tritem; 707ea6eb3fSJiri Pirko 717ea6eb3fSJiri Pirko tritem = __switchdev_trans_item_dequeue(trans); 727ea6eb3fSJiri Pirko BUG_ON(!tritem); 737ea6eb3fSJiri Pirko return tritem->data; 747ea6eb3fSJiri Pirko } 757ea6eb3fSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_trans_item_dequeue); 767ea6eb3fSJiri Pirko 777ea6eb3fSJiri Pirko static void switchdev_trans_init(struct switchdev_trans *trans) 787ea6eb3fSJiri Pirko { 797ea6eb3fSJiri Pirko INIT_LIST_HEAD(&trans->item_list); 807ea6eb3fSJiri Pirko } 817ea6eb3fSJiri Pirko 827ea6eb3fSJiri Pirko static void switchdev_trans_items_destroy(struct switchdev_trans *trans) 837ea6eb3fSJiri Pirko { 847ea6eb3fSJiri Pirko struct switchdev_trans_item *tritem; 857ea6eb3fSJiri Pirko 867ea6eb3fSJiri Pirko while ((tritem = __switchdev_trans_item_dequeue(trans))) 877ea6eb3fSJiri Pirko tritem->destructor(tritem->data); 887ea6eb3fSJiri Pirko } 897ea6eb3fSJiri Pirko 907ea6eb3fSJiri Pirko static void switchdev_trans_items_warn_destroy(struct net_device *dev, 917ea6eb3fSJiri Pirko struct switchdev_trans *trans) 927ea6eb3fSJiri Pirko { 937ea6eb3fSJiri Pirko WARN(!list_empty(&trans->item_list), "%s: transaction item queue is not empty.\n", 947ea6eb3fSJiri Pirko dev->name); 957ea6eb3fSJiri Pirko switchdev_trans_items_destroy(trans); 967ea6eb3fSJiri Pirko } 977ea6eb3fSJiri Pirko 98793f4014SJiri Pirko static LIST_HEAD(deferred); 99793f4014SJiri Pirko static DEFINE_SPINLOCK(deferred_lock); 100793f4014SJiri Pirko 101793f4014SJiri Pirko typedef void switchdev_deferred_func_t(struct net_device *dev, 102793f4014SJiri Pirko const void *data); 103793f4014SJiri Pirko 104793f4014SJiri Pirko struct switchdev_deferred_item { 105793f4014SJiri Pirko struct list_head list; 106793f4014SJiri Pirko struct net_device *dev; 107793f4014SJiri Pirko switchdev_deferred_func_t *func; 108793f4014SJiri Pirko unsigned long data[0]; 109793f4014SJiri Pirko }; 110793f4014SJiri Pirko 111793f4014SJiri Pirko static struct switchdev_deferred_item *switchdev_deferred_dequeue(void) 112793f4014SJiri Pirko { 113793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 114793f4014SJiri Pirko 115793f4014SJiri Pirko spin_lock_bh(&deferred_lock); 116793f4014SJiri Pirko if (list_empty(&deferred)) { 117793f4014SJiri Pirko dfitem = NULL; 118793f4014SJiri Pirko goto unlock; 119793f4014SJiri Pirko } 120793f4014SJiri Pirko dfitem = list_first_entry(&deferred, 121793f4014SJiri Pirko struct switchdev_deferred_item, list); 122793f4014SJiri Pirko list_del(&dfitem->list); 123793f4014SJiri Pirko unlock: 124793f4014SJiri Pirko spin_unlock_bh(&deferred_lock); 125793f4014SJiri Pirko return dfitem; 126793f4014SJiri Pirko } 127793f4014SJiri Pirko 128793f4014SJiri Pirko /** 129793f4014SJiri Pirko * switchdev_deferred_process - Process ops in deferred queue 130793f4014SJiri Pirko * 131793f4014SJiri Pirko * Called to flush the ops currently queued in deferred ops queue. 132793f4014SJiri Pirko * rtnl_lock must be held. 133793f4014SJiri Pirko */ 134793f4014SJiri Pirko void switchdev_deferred_process(void) 135793f4014SJiri Pirko { 136793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 137793f4014SJiri Pirko 138793f4014SJiri Pirko ASSERT_RTNL(); 139793f4014SJiri Pirko 140793f4014SJiri Pirko while ((dfitem = switchdev_deferred_dequeue())) { 141793f4014SJiri Pirko dfitem->func(dfitem->dev, dfitem->data); 142793f4014SJiri Pirko dev_put(dfitem->dev); 143793f4014SJiri Pirko kfree(dfitem); 144793f4014SJiri Pirko } 145793f4014SJiri Pirko } 146793f4014SJiri Pirko EXPORT_SYMBOL_GPL(switchdev_deferred_process); 147793f4014SJiri Pirko 148793f4014SJiri Pirko static void switchdev_deferred_process_work(struct work_struct *work) 149793f4014SJiri Pirko { 150793f4014SJiri Pirko rtnl_lock(); 151793f4014SJiri Pirko switchdev_deferred_process(); 152793f4014SJiri Pirko rtnl_unlock(); 153793f4014SJiri Pirko } 154793f4014SJiri Pirko 155793f4014SJiri Pirko static DECLARE_WORK(deferred_process_work, switchdev_deferred_process_work); 156793f4014SJiri Pirko 157793f4014SJiri Pirko static int switchdev_deferred_enqueue(struct net_device *dev, 158793f4014SJiri Pirko const void *data, size_t data_len, 159793f4014SJiri Pirko switchdev_deferred_func_t *func) 160793f4014SJiri Pirko { 161793f4014SJiri Pirko struct switchdev_deferred_item *dfitem; 162793f4014SJiri Pirko 163793f4014SJiri Pirko dfitem = kmalloc(sizeof(*dfitem) + data_len, GFP_ATOMIC); 164793f4014SJiri Pirko if (!dfitem) 165793f4014SJiri Pirko return -ENOMEM; 166793f4014SJiri Pirko dfitem->dev = dev; 167793f4014SJiri Pirko dfitem->func = func; 168793f4014SJiri Pirko memcpy(dfitem->data, data, data_len); 169793f4014SJiri Pirko dev_hold(dev); 170793f4014SJiri Pirko spin_lock_bh(&deferred_lock); 171793f4014SJiri Pirko list_add_tail(&dfitem->list, &deferred); 172793f4014SJiri Pirko spin_unlock_bh(&deferred_lock); 173793f4014SJiri Pirko schedule_work(&deferred_process_work); 174793f4014SJiri Pirko return 0; 175793f4014SJiri Pirko } 176793f4014SJiri Pirko 1777ea6eb3fSJiri Pirko /** 1783094333dSScott Feldman * switchdev_port_attr_get - Get port attribute 1793094333dSScott Feldman * 1803094333dSScott Feldman * @dev: port device 1813094333dSScott Feldman * @attr: attribute to get 1823094333dSScott Feldman */ 1833094333dSScott Feldman int switchdev_port_attr_get(struct net_device *dev, struct switchdev_attr *attr) 1843094333dSScott Feldman { 1853094333dSScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 1863094333dSScott Feldman struct net_device *lower_dev; 1873094333dSScott Feldman struct list_head *iter; 1883094333dSScott Feldman struct switchdev_attr first = { 1891f868398SJiri Pirko .id = SWITCHDEV_ATTR_ID_UNDEFINED 1903094333dSScott Feldman }; 1913094333dSScott Feldman int err = -EOPNOTSUPP; 1923094333dSScott Feldman 1933094333dSScott Feldman if (ops && ops->switchdev_port_attr_get) 1943094333dSScott Feldman return ops->switchdev_port_attr_get(dev, attr); 1953094333dSScott Feldman 1963094333dSScott Feldman if (attr->flags & SWITCHDEV_F_NO_RECURSE) 1973094333dSScott Feldman return err; 1983094333dSScott Feldman 1993094333dSScott Feldman /* Switch device port(s) may be stacked under 2003094333dSScott Feldman * bond/team/vlan dev, so recurse down to get attr on 2013094333dSScott Feldman * each port. Return -ENODATA if attr values don't 2023094333dSScott Feldman * compare across ports. 2033094333dSScott Feldman */ 2043094333dSScott Feldman 2053094333dSScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 2063094333dSScott Feldman err = switchdev_port_attr_get(lower_dev, attr); 2073094333dSScott Feldman if (err) 2083094333dSScott Feldman break; 2091f868398SJiri Pirko if (first.id == SWITCHDEV_ATTR_ID_UNDEFINED) 2103094333dSScott Feldman first = *attr; 2113094333dSScott Feldman else if (memcmp(&first, attr, sizeof(*attr))) 2123094333dSScott Feldman return -ENODATA; 2133094333dSScott Feldman } 2143094333dSScott Feldman 2153094333dSScott Feldman return err; 2163094333dSScott Feldman } 2173094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_get); 2183094333dSScott Feldman 2193094333dSScott Feldman static int __switchdev_port_attr_set(struct net_device *dev, 220f7fadf30SJiri Pirko const struct switchdev_attr *attr, 2217ea6eb3fSJiri Pirko struct switchdev_trans *trans) 2223094333dSScott Feldman { 2233094333dSScott Feldman const struct switchdev_ops *ops = dev->switchdev_ops; 2243094333dSScott Feldman struct net_device *lower_dev; 2253094333dSScott Feldman struct list_head *iter; 2263094333dSScott Feldman int err = -EOPNOTSUPP; 2273094333dSScott Feldman 2280c63d80cSJiri Pirko if (ops && ops->switchdev_port_attr_set) { 2290c63d80cSJiri Pirko err = ops->switchdev_port_attr_set(dev, attr, trans); 2300c63d80cSJiri Pirko goto done; 2310c63d80cSJiri Pirko } 2323094333dSScott Feldman 2333094333dSScott Feldman if (attr->flags & SWITCHDEV_F_NO_RECURSE) 234464314eaSScott Feldman goto done; 2353094333dSScott Feldman 2363094333dSScott Feldman /* Switch device port(s) may be stacked under 2373094333dSScott Feldman * bond/team/vlan dev, so recurse down to set attr on 2383094333dSScott Feldman * each port. 2393094333dSScott Feldman */ 2403094333dSScott Feldman 2413094333dSScott Feldman netdev_for_each_lower_dev(dev, lower_dev, iter) { 2427ea6eb3fSJiri Pirko err = __switchdev_port_attr_set(lower_dev, attr, trans); 2433094333dSScott Feldman if (err) 2443094333dSScott Feldman break; 2453094333dSScott Feldman } 2463094333dSScott Feldman 247464314eaSScott Feldman done: 248464314eaSScott Feldman if (err == -EOPNOTSUPP && attr->flags & SWITCHDEV_F_SKIP_EOPNOTSUPP) 249464314eaSScott Feldman err = 0; 250464314eaSScott Feldman 2513094333dSScott Feldman return err; 2523094333dSScott Feldman } 2533094333dSScott Feldman 2540bc05d58SJiri Pirko static int switchdev_port_attr_set_now(struct net_device *dev, 255f7fadf30SJiri Pirko const struct switchdev_attr *attr) 2563094333dSScott Feldman { 2577ea6eb3fSJiri Pirko struct switchdev_trans trans; 2583094333dSScott Feldman int err; 2593094333dSScott Feldman 2607ea6eb3fSJiri Pirko switchdev_trans_init(&trans); 2617ea6eb3fSJiri Pirko 2623094333dSScott Feldman /* Phase I: prepare for attr set. Driver/device should fail 2633094333dSScott Feldman * here if there are going to be issues in the commit phase, 2643094333dSScott Feldman * such as lack of resources or support. The driver/device 2653094333dSScott Feldman * should reserve resources needed for the commit phase here, 2663094333dSScott Feldman * but should not commit the attr. 2673094333dSScott Feldman */ 2683094333dSScott Feldman 269f623ab7fSJiri Pirko trans.ph_prepare = true; 2707ea6eb3fSJiri Pirko err = __switchdev_port_attr_set(dev, attr, &trans); 2713094333dSScott Feldman if (err) { 2723094333dSScott Feldman /* Prepare phase failed: abort the transaction. Any 2733094333dSScott Feldman * resources reserved in the prepare phase are 2743094333dSScott Feldman * released. 2753094333dSScott Feldman */ 2763094333dSScott Feldman 2779f6467cfSJiri Pirko if (err != -EOPNOTSUPP) 2787ea6eb3fSJiri Pirko switchdev_trans_items_destroy(&trans); 2793094333dSScott Feldman 2803094333dSScott Feldman return err; 2813094333dSScott Feldman } 2823094333dSScott Feldman 2833094333dSScott Feldman /* Phase II: commit attr set. This cannot fail as a fault 2843094333dSScott Feldman * of driver/device. If it does, it's a bug in the driver/device 2853094333dSScott Feldman * because the driver said everythings was OK in phase I. 2863094333dSScott Feldman */ 2873094333dSScott Feldman 288f623ab7fSJiri Pirko trans.ph_prepare = false; 2897ea6eb3fSJiri Pirko err = __switchdev_port_attr_set(dev, attr, &trans); 290e9fdaec0SScott Feldman WARN(err, "%s: Commit of attribute (id=%d) failed.\n", 291e9fdaec0SScott Feldman dev->name, attr->id); 2927ea6eb3fSJiri Pirko switchdev_trans_items_warn_destroy(dev, &trans); 2933094333dSScott Feldman 2943094333dSScott Feldman return err; 2953094333dSScott Feldman } 2960bc05d58SJiri Pirko 2970bc05d58SJiri Pirko static void switchdev_port_attr_set_deferred(struct net_device *dev, 2980bc05d58SJiri Pirko const void *data) 2990bc05d58SJiri Pirko { 3000bc05d58SJiri Pirko const struct switchdev_attr *attr = data; 3010bc05d58SJiri Pirko int err; 3020bc05d58SJiri Pirko 3030bc05d58SJiri Pirko err = switchdev_port_attr_set_now(dev, attr); 3040bc05d58SJiri Pirko if (err && err != -EOPNOTSUPP) 3050bc05d58SJiri Pirko netdev_err(dev, "failed (err=%d) to set attribute (id=%d)\n", 3060bc05d58SJiri Pirko err, attr->id); 3077ceb2afbSElad Raz if (attr->complete) 3087ceb2afbSElad Raz attr->complete(dev, err, attr->complete_priv); 3090bc05d58SJiri Pirko } 3100bc05d58SJiri Pirko 3110bc05d58SJiri Pirko static int switchdev_port_attr_set_defer(struct net_device *dev, 3120bc05d58SJiri Pirko const struct switchdev_attr *attr) 3130bc05d58SJiri Pirko { 3140bc05d58SJiri Pirko return switchdev_deferred_enqueue(dev, attr, sizeof(*attr), 3150bc05d58SJiri Pirko switchdev_port_attr_set_deferred); 3160bc05d58SJiri Pirko } 3170bc05d58SJiri Pirko 3180bc05d58SJiri Pirko /** 3190bc05d58SJiri Pirko * switchdev_port_attr_set - Set port attribute 3200bc05d58SJiri Pirko * 3210bc05d58SJiri Pirko * @dev: port device 3220bc05d58SJiri Pirko * @attr: attribute to set 3230bc05d58SJiri Pirko * 3240bc05d58SJiri Pirko * Use a 2-phase prepare-commit transaction model to ensure 3250bc05d58SJiri Pirko * system is not left in a partially updated state due to 3260bc05d58SJiri Pirko * failure from driver/device. 3270bc05d58SJiri Pirko * 3280bc05d58SJiri Pirko * rtnl_lock must be held and must not be in atomic section, 3290bc05d58SJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 3300bc05d58SJiri Pirko */ 3310bc05d58SJiri Pirko int switchdev_port_attr_set(struct net_device *dev, 3320bc05d58SJiri Pirko const struct switchdev_attr *attr) 3330bc05d58SJiri Pirko { 3340bc05d58SJiri Pirko if (attr->flags & SWITCHDEV_F_DEFER) 3350bc05d58SJiri Pirko return switchdev_port_attr_set_defer(dev, attr); 3360bc05d58SJiri Pirko ASSERT_RTNL(); 3370bc05d58SJiri Pirko return switchdev_port_attr_set_now(dev, attr); 3380bc05d58SJiri Pirko } 3393094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_set); 3403094333dSScott Feldman 341e258d919SScott Feldman static size_t switchdev_obj_size(const struct switchdev_obj *obj) 342e258d919SScott Feldman { 343e258d919SScott Feldman switch (obj->id) { 344e258d919SScott Feldman case SWITCHDEV_OBJ_ID_PORT_VLAN: 345e258d919SScott Feldman return sizeof(struct switchdev_obj_port_vlan); 3464d41e125SElad Raz case SWITCHDEV_OBJ_ID_PORT_MDB: 3474d41e125SElad Raz return sizeof(struct switchdev_obj_port_mdb); 34847d5b6dbSAndrew Lunn case SWITCHDEV_OBJ_ID_HOST_MDB: 34947d5b6dbSAndrew Lunn return sizeof(struct switchdev_obj_port_mdb); 350e258d919SScott Feldman default: 351e258d919SScott Feldman BUG(); 352e258d919SScott Feldman } 353e258d919SScott Feldman return 0; 354e258d919SScott Feldman } 355e258d919SScott Feldman 356d17d9f5eSPetr Machata static int switchdev_port_obj_notify(enum switchdev_notifier_type nt, 357d17d9f5eSPetr Machata struct net_device *dev, 358648b4a99SJiri Pirko const struct switchdev_obj *obj, 35969b7320eSPetr Machata struct switchdev_trans *trans, 36069b7320eSPetr Machata struct netlink_ext_ack *extack) 361491d0f15SScott Feldman { 362d17d9f5eSPetr Machata int rc; 363d17d9f5eSPetr Machata int err; 364491d0f15SScott Feldman 365d17d9f5eSPetr Machata struct switchdev_notifier_port_obj_info obj_info = { 366d17d9f5eSPetr Machata .obj = obj, 367d17d9f5eSPetr Machata .trans = trans, 368d17d9f5eSPetr Machata .handled = false, 369d17d9f5eSPetr Machata }; 370491d0f15SScott Feldman 371*479c86dcSPetr Machata rc = call_switchdev_blocking_notifiers(nt, dev, &obj_info.info, extack); 372d17d9f5eSPetr Machata err = notifier_to_errno(rc); 373d17d9f5eSPetr Machata if (err) { 374d17d9f5eSPetr Machata WARN_ON(!obj_info.handled); 375491d0f15SScott Feldman return err; 376491d0f15SScott Feldman } 377d17d9f5eSPetr Machata if (!obj_info.handled) 378d17d9f5eSPetr Machata return -EOPNOTSUPP; 379d17d9f5eSPetr Machata return 0; 380d17d9f5eSPetr Machata } 381491d0f15SScott Feldman 3824d429c5dSJiri Pirko static int switchdev_port_obj_add_now(struct net_device *dev, 38369b7320eSPetr Machata const struct switchdev_obj *obj, 38469b7320eSPetr Machata struct netlink_ext_ack *extack) 385491d0f15SScott Feldman { 3867ea6eb3fSJiri Pirko struct switchdev_trans trans; 387491d0f15SScott Feldman int err; 388491d0f15SScott Feldman 389491d0f15SScott Feldman ASSERT_RTNL(); 390491d0f15SScott Feldman 3917ea6eb3fSJiri Pirko switchdev_trans_init(&trans); 3927ea6eb3fSJiri Pirko 393491d0f15SScott Feldman /* Phase I: prepare for obj add. Driver/device should fail 394491d0f15SScott Feldman * here if there are going to be issues in the commit phase, 395491d0f15SScott Feldman * such as lack of resources or support. The driver/device 396491d0f15SScott Feldman * should reserve resources needed for the commit phase here, 397491d0f15SScott Feldman * but should not commit the obj. 398491d0f15SScott Feldman */ 399491d0f15SScott Feldman 400f623ab7fSJiri Pirko trans.ph_prepare = true; 401d17d9f5eSPetr Machata err = switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD, 40269b7320eSPetr Machata dev, obj, &trans, extack); 403491d0f15SScott Feldman if (err) { 404491d0f15SScott Feldman /* Prepare phase failed: abort the transaction. Any 405491d0f15SScott Feldman * resources reserved in the prepare phase are 406491d0f15SScott Feldman * released. 407491d0f15SScott Feldman */ 408491d0f15SScott Feldman 4099f6467cfSJiri Pirko if (err != -EOPNOTSUPP) 4107ea6eb3fSJiri Pirko switchdev_trans_items_destroy(&trans); 411491d0f15SScott Feldman 412491d0f15SScott Feldman return err; 413491d0f15SScott Feldman } 414491d0f15SScott Feldman 415491d0f15SScott Feldman /* Phase II: commit obj add. This cannot fail as a fault 416491d0f15SScott Feldman * of driver/device. If it does, it's a bug in the driver/device 417491d0f15SScott Feldman * because the driver said everythings was OK in phase I. 418491d0f15SScott Feldman */ 419491d0f15SScott Feldman 420f623ab7fSJiri Pirko trans.ph_prepare = false; 421d17d9f5eSPetr Machata err = switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD, 42269b7320eSPetr Machata dev, obj, &trans, extack); 4239e8f4a54SJiri Pirko WARN(err, "%s: Commit of object (id=%d) failed.\n", dev->name, obj->id); 4247ea6eb3fSJiri Pirko switchdev_trans_items_warn_destroy(dev, &trans); 425491d0f15SScott Feldman 426491d0f15SScott Feldman return err; 427491d0f15SScott Feldman } 4284d429c5dSJiri Pirko 4294d429c5dSJiri Pirko static void switchdev_port_obj_add_deferred(struct net_device *dev, 4304d429c5dSJiri Pirko const void *data) 4314d429c5dSJiri Pirko { 4324d429c5dSJiri Pirko const struct switchdev_obj *obj = data; 4334d429c5dSJiri Pirko int err; 4344d429c5dSJiri Pirko 43569b7320eSPetr Machata err = switchdev_port_obj_add_now(dev, obj, NULL); 4364d429c5dSJiri Pirko if (err && err != -EOPNOTSUPP) 4374d429c5dSJiri Pirko netdev_err(dev, "failed (err=%d) to add object (id=%d)\n", 4384d429c5dSJiri Pirko err, obj->id); 4397ceb2afbSElad Raz if (obj->complete) 4407ceb2afbSElad Raz obj->complete(dev, err, obj->complete_priv); 4414d429c5dSJiri Pirko } 4424d429c5dSJiri Pirko 4434d429c5dSJiri Pirko static int switchdev_port_obj_add_defer(struct net_device *dev, 4444d429c5dSJiri Pirko const struct switchdev_obj *obj) 4454d429c5dSJiri Pirko { 446e258d919SScott Feldman return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj), 4474d429c5dSJiri Pirko switchdev_port_obj_add_deferred); 4484d429c5dSJiri Pirko } 449491d0f15SScott Feldman 450491d0f15SScott Feldman /** 4514d429c5dSJiri Pirko * switchdev_port_obj_add - Add port object 452491d0f15SScott Feldman * 453491d0f15SScott Feldman * @dev: port device 454ab069002SVivien Didelot * @id: object ID 4554d429c5dSJiri Pirko * @obj: object to add 4564d429c5dSJiri Pirko * 4574d429c5dSJiri Pirko * Use a 2-phase prepare-commit transaction model to ensure 4584d429c5dSJiri Pirko * system is not left in a partially updated state due to 4594d429c5dSJiri Pirko * failure from driver/device. 4604d429c5dSJiri Pirko * 4614d429c5dSJiri Pirko * rtnl_lock must be held and must not be in atomic section, 4624d429c5dSJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 463491d0f15SScott Feldman */ 4644d429c5dSJiri Pirko int switchdev_port_obj_add(struct net_device *dev, 46569b7320eSPetr Machata const struct switchdev_obj *obj, 46669b7320eSPetr Machata struct netlink_ext_ack *extack) 4674d429c5dSJiri Pirko { 4684d429c5dSJiri Pirko if (obj->flags & SWITCHDEV_F_DEFER) 4694d429c5dSJiri Pirko return switchdev_port_obj_add_defer(dev, obj); 4704d429c5dSJiri Pirko ASSERT_RTNL(); 47169b7320eSPetr Machata return switchdev_port_obj_add_now(dev, obj, extack); 4724d429c5dSJiri Pirko } 4734d429c5dSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_port_obj_add); 4744d429c5dSJiri Pirko 4754d429c5dSJiri Pirko static int switchdev_port_obj_del_now(struct net_device *dev, 476648b4a99SJiri Pirko const struct switchdev_obj *obj) 477491d0f15SScott Feldman { 478d17d9f5eSPetr Machata return switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_DEL, 47969b7320eSPetr Machata dev, obj, NULL, NULL); 480491d0f15SScott Feldman } 4814d429c5dSJiri Pirko 4824d429c5dSJiri Pirko static void switchdev_port_obj_del_deferred(struct net_device *dev, 4834d429c5dSJiri Pirko const void *data) 4844d429c5dSJiri Pirko { 4854d429c5dSJiri Pirko const struct switchdev_obj *obj = data; 4864d429c5dSJiri Pirko int err; 4874d429c5dSJiri Pirko 4884d429c5dSJiri Pirko err = switchdev_port_obj_del_now(dev, obj); 4894d429c5dSJiri Pirko if (err && err != -EOPNOTSUPP) 4904d429c5dSJiri Pirko netdev_err(dev, "failed (err=%d) to del object (id=%d)\n", 4914d429c5dSJiri Pirko err, obj->id); 4927ceb2afbSElad Raz if (obj->complete) 4937ceb2afbSElad Raz obj->complete(dev, err, obj->complete_priv); 4944d429c5dSJiri Pirko } 4954d429c5dSJiri Pirko 4964d429c5dSJiri Pirko static int switchdev_port_obj_del_defer(struct net_device *dev, 4974d429c5dSJiri Pirko const struct switchdev_obj *obj) 4984d429c5dSJiri Pirko { 499e258d919SScott Feldman return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj), 5004d429c5dSJiri Pirko switchdev_port_obj_del_deferred); 5014d429c5dSJiri Pirko } 5024d429c5dSJiri Pirko 5034d429c5dSJiri Pirko /** 5044d429c5dSJiri Pirko * switchdev_port_obj_del - Delete port object 5054d429c5dSJiri Pirko * 5064d429c5dSJiri Pirko * @dev: port device 5074d429c5dSJiri Pirko * @id: object ID 5084d429c5dSJiri Pirko * @obj: object to delete 5094d429c5dSJiri Pirko * 5104d429c5dSJiri Pirko * rtnl_lock must be held and must not be in atomic section, 5114d429c5dSJiri Pirko * in case SWITCHDEV_F_DEFER flag is not set. 5124d429c5dSJiri Pirko */ 5134d429c5dSJiri Pirko int switchdev_port_obj_del(struct net_device *dev, 5144d429c5dSJiri Pirko const struct switchdev_obj *obj) 5154d429c5dSJiri Pirko { 5164d429c5dSJiri Pirko if (obj->flags & SWITCHDEV_F_DEFER) 5174d429c5dSJiri Pirko return switchdev_port_obj_del_defer(dev, obj); 5184d429c5dSJiri Pirko ASSERT_RTNL(); 5194d429c5dSJiri Pirko return switchdev_port_obj_del_now(dev, obj); 5204d429c5dSJiri Pirko } 521491d0f15SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_obj_del); 522491d0f15SScott Feldman 523ff5cf100SArkadi Sharshevsky static ATOMIC_NOTIFIER_HEAD(switchdev_notif_chain); 524a93e3b17SPetr Machata static BLOCKING_NOTIFIER_HEAD(switchdev_blocking_notif_chain); 52503bf0c28SJiri Pirko 52603bf0c28SJiri Pirko /** 527ebb9a03aSJiri Pirko * register_switchdev_notifier - Register notifier 52803bf0c28SJiri Pirko * @nb: notifier_block 52903bf0c28SJiri Pirko * 530ff5cf100SArkadi Sharshevsky * Register switch device notifier. 53103bf0c28SJiri Pirko */ 532ebb9a03aSJiri Pirko int register_switchdev_notifier(struct notifier_block *nb) 53303bf0c28SJiri Pirko { 534ff5cf100SArkadi Sharshevsky return atomic_notifier_chain_register(&switchdev_notif_chain, nb); 53503bf0c28SJiri Pirko } 536ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(register_switchdev_notifier); 53703bf0c28SJiri Pirko 53803bf0c28SJiri Pirko /** 539ebb9a03aSJiri Pirko * unregister_switchdev_notifier - Unregister notifier 54003bf0c28SJiri Pirko * @nb: notifier_block 54103bf0c28SJiri Pirko * 54203bf0c28SJiri Pirko * Unregister switch device notifier. 54303bf0c28SJiri Pirko */ 544ebb9a03aSJiri Pirko int unregister_switchdev_notifier(struct notifier_block *nb) 54503bf0c28SJiri Pirko { 546ff5cf100SArkadi Sharshevsky return atomic_notifier_chain_unregister(&switchdev_notif_chain, nb); 54703bf0c28SJiri Pirko } 548ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(unregister_switchdev_notifier); 54903bf0c28SJiri Pirko 55003bf0c28SJiri Pirko /** 551ebb9a03aSJiri Pirko * call_switchdev_notifiers - Call notifiers 55203bf0c28SJiri Pirko * @val: value passed unmodified to notifier function 55303bf0c28SJiri Pirko * @dev: port device 55403bf0c28SJiri Pirko * @info: notifier information data 55503bf0c28SJiri Pirko * 556ff5cf100SArkadi Sharshevsky * Call all network notifier blocks. 55703bf0c28SJiri Pirko */ 558ebb9a03aSJiri Pirko int call_switchdev_notifiers(unsigned long val, struct net_device *dev, 559ebb9a03aSJiri Pirko struct switchdev_notifier_info *info) 56003bf0c28SJiri Pirko { 56103bf0c28SJiri Pirko info->dev = dev; 562*479c86dcSPetr Machata info->extack = NULL; 563ff5cf100SArkadi Sharshevsky return atomic_notifier_call_chain(&switchdev_notif_chain, val, info); 56403bf0c28SJiri Pirko } 565ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(call_switchdev_notifiers); 5668a44dbb2SRoopa Prabhu 567a93e3b17SPetr Machata int register_switchdev_blocking_notifier(struct notifier_block *nb) 568a93e3b17SPetr Machata { 569a93e3b17SPetr Machata struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain; 570a93e3b17SPetr Machata 571a93e3b17SPetr Machata return blocking_notifier_chain_register(chain, nb); 572a93e3b17SPetr Machata } 573a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(register_switchdev_blocking_notifier); 574a93e3b17SPetr Machata 575a93e3b17SPetr Machata int unregister_switchdev_blocking_notifier(struct notifier_block *nb) 576a93e3b17SPetr Machata { 577a93e3b17SPetr Machata struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain; 578a93e3b17SPetr Machata 579a93e3b17SPetr Machata return blocking_notifier_chain_unregister(chain, nb); 580a93e3b17SPetr Machata } 581a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(unregister_switchdev_blocking_notifier); 582a93e3b17SPetr Machata 583a93e3b17SPetr Machata int call_switchdev_blocking_notifiers(unsigned long val, struct net_device *dev, 584*479c86dcSPetr Machata struct switchdev_notifier_info *info, 585*479c86dcSPetr Machata struct netlink_ext_ack *extack) 586a93e3b17SPetr Machata { 587a93e3b17SPetr Machata info->dev = dev; 588*479c86dcSPetr Machata info->extack = extack; 589a93e3b17SPetr Machata return blocking_notifier_call_chain(&switchdev_blocking_notif_chain, 590a93e3b17SPetr Machata val, info); 591a93e3b17SPetr Machata } 592a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(call_switchdev_blocking_notifiers); 593a93e3b17SPetr Machata 5948438884dSOr Gerlitz bool switchdev_port_same_parent_id(struct net_device *a, 5951a3b2ec9SScott Feldman struct net_device *b) 5961a3b2ec9SScott Feldman { 5971a3b2ec9SScott Feldman struct switchdev_attr a_attr = { 5986ff64f6fSIdo Schimmel .orig_dev = a, 5991f868398SJiri Pirko .id = SWITCHDEV_ATTR_ID_PORT_PARENT_ID, 6001a3b2ec9SScott Feldman }; 6011a3b2ec9SScott Feldman struct switchdev_attr b_attr = { 6026ff64f6fSIdo Schimmel .orig_dev = b, 6031f868398SJiri Pirko .id = SWITCHDEV_ATTR_ID_PORT_PARENT_ID, 6041a3b2ec9SScott Feldman }; 6051a3b2ec9SScott Feldman 6061a3b2ec9SScott Feldman if (switchdev_port_attr_get(a, &a_attr) || 6071a3b2ec9SScott Feldman switchdev_port_attr_get(b, &b_attr)) 6081a3b2ec9SScott Feldman return false; 6091a3b2ec9SScott Feldman 6101a3b2ec9SScott Feldman return netdev_phys_item_id_same(&a_attr.u.ppid, &b_attr.u.ppid); 6111a3b2ec9SScott Feldman } 6122eb03e6cSOr Gerlitz EXPORT_SYMBOL_GPL(switchdev_port_same_parent_id); 613f30f0601SPetr Machata 614f30f0601SPetr Machata static int __switchdev_handle_port_obj_add(struct net_device *dev, 615f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 616f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 617f30f0601SPetr Machata int (*add_cb)(struct net_device *dev, 618f30f0601SPetr Machata const struct switchdev_obj *obj, 619f30f0601SPetr Machata struct switchdev_trans *trans)) 620f30f0601SPetr Machata { 621f30f0601SPetr Machata struct net_device *lower_dev; 622f30f0601SPetr Machata struct list_head *iter; 623f30f0601SPetr Machata int err = -EOPNOTSUPP; 624f30f0601SPetr Machata 625f30f0601SPetr Machata if (check_cb(dev)) { 626f30f0601SPetr Machata /* This flag is only checked if the return value is success. */ 627f30f0601SPetr Machata port_obj_info->handled = true; 628f30f0601SPetr Machata return add_cb(dev, port_obj_info->obj, port_obj_info->trans); 629f30f0601SPetr Machata } 630f30f0601SPetr Machata 631f30f0601SPetr Machata /* Switch ports might be stacked under e.g. a LAG. Ignore the 632f30f0601SPetr Machata * unsupported devices, another driver might be able to handle them. But 633f30f0601SPetr Machata * propagate to the callers any hard errors. 634f30f0601SPetr Machata * 635f30f0601SPetr Machata * If the driver does its own bookkeeping of stacked ports, it's not 636f30f0601SPetr Machata * necessary to go through this helper. 637f30f0601SPetr Machata */ 638f30f0601SPetr Machata netdev_for_each_lower_dev(dev, lower_dev, iter) { 639f30f0601SPetr Machata err = __switchdev_handle_port_obj_add(lower_dev, port_obj_info, 640f30f0601SPetr Machata check_cb, add_cb); 641f30f0601SPetr Machata if (err && err != -EOPNOTSUPP) 642f30f0601SPetr Machata return err; 643f30f0601SPetr Machata } 644f30f0601SPetr Machata 645f30f0601SPetr Machata return err; 646f30f0601SPetr Machata } 647f30f0601SPetr Machata 648f30f0601SPetr Machata int switchdev_handle_port_obj_add(struct net_device *dev, 649f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 650f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 651f30f0601SPetr Machata int (*add_cb)(struct net_device *dev, 652f30f0601SPetr Machata const struct switchdev_obj *obj, 653f30f0601SPetr Machata struct switchdev_trans *trans)) 654f30f0601SPetr Machata { 655f30f0601SPetr Machata int err; 656f30f0601SPetr Machata 657f30f0601SPetr Machata err = __switchdev_handle_port_obj_add(dev, port_obj_info, check_cb, 658f30f0601SPetr Machata add_cb); 659f30f0601SPetr Machata if (err == -EOPNOTSUPP) 660f30f0601SPetr Machata err = 0; 661f30f0601SPetr Machata return err; 662f30f0601SPetr Machata } 663f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_add); 664f30f0601SPetr Machata 665f30f0601SPetr Machata static int __switchdev_handle_port_obj_del(struct net_device *dev, 666f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 667f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 668f30f0601SPetr Machata int (*del_cb)(struct net_device *dev, 669f30f0601SPetr Machata const struct switchdev_obj *obj)) 670f30f0601SPetr Machata { 671f30f0601SPetr Machata struct net_device *lower_dev; 672f30f0601SPetr Machata struct list_head *iter; 673f30f0601SPetr Machata int err = -EOPNOTSUPP; 674f30f0601SPetr Machata 675f30f0601SPetr Machata if (check_cb(dev)) { 676f30f0601SPetr Machata /* This flag is only checked if the return value is success. */ 677f30f0601SPetr Machata port_obj_info->handled = true; 678f30f0601SPetr Machata return del_cb(dev, port_obj_info->obj); 679f30f0601SPetr Machata } 680f30f0601SPetr Machata 681f30f0601SPetr Machata /* Switch ports might be stacked under e.g. a LAG. Ignore the 682f30f0601SPetr Machata * unsupported devices, another driver might be able to handle them. But 683f30f0601SPetr Machata * propagate to the callers any hard errors. 684f30f0601SPetr Machata * 685f30f0601SPetr Machata * If the driver does its own bookkeeping of stacked ports, it's not 686f30f0601SPetr Machata * necessary to go through this helper. 687f30f0601SPetr Machata */ 688f30f0601SPetr Machata netdev_for_each_lower_dev(dev, lower_dev, iter) { 689f30f0601SPetr Machata err = __switchdev_handle_port_obj_del(lower_dev, port_obj_info, 690f30f0601SPetr Machata check_cb, del_cb); 691f30f0601SPetr Machata if (err && err != -EOPNOTSUPP) 692f30f0601SPetr Machata return err; 693f30f0601SPetr Machata } 694f30f0601SPetr Machata 695f30f0601SPetr Machata return err; 696f30f0601SPetr Machata } 697f30f0601SPetr Machata 698f30f0601SPetr Machata int switchdev_handle_port_obj_del(struct net_device *dev, 699f30f0601SPetr Machata struct switchdev_notifier_port_obj_info *port_obj_info, 700f30f0601SPetr Machata bool (*check_cb)(const struct net_device *dev), 701f30f0601SPetr Machata int (*del_cb)(struct net_device *dev, 702f30f0601SPetr Machata const struct switchdev_obj *obj)) 703f30f0601SPetr Machata { 704f30f0601SPetr Machata int err; 705f30f0601SPetr Machata 706f30f0601SPetr Machata err = __switchdev_handle_port_obj_del(dev, port_obj_info, check_cb, 707f30f0601SPetr Machata del_cb); 708f30f0601SPetr Machata if (err == -EOPNOTSUPP) 709f30f0601SPetr Machata err = 0; 710f30f0601SPetr Machata return err; 711f30f0601SPetr Machata } 712f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_del); 713