xref: /linux-6.15/net/switchdev/switchdev.c (revision 479c86dc)
1007f790cSJiri Pirko /*
2007f790cSJiri Pirko  * net/switchdev/switchdev.c - Switch device API
37ea6eb3fSJiri Pirko  * Copyright (c) 2014-2015 Jiri Pirko <[email protected]>
4f8f21471SScott Feldman  * Copyright (c) 2014-2015 Scott Feldman <[email protected]>
5007f790cSJiri Pirko  *
6007f790cSJiri Pirko  * This program is free software; you can redistribute it and/or modify
7007f790cSJiri Pirko  * it under the terms of the GNU General Public License as published by
8007f790cSJiri Pirko  * the Free Software Foundation; either version 2 of the License, or
9007f790cSJiri Pirko  * (at your option) any later version.
10007f790cSJiri Pirko  */
11007f790cSJiri Pirko 
12007f790cSJiri Pirko #include <linux/kernel.h>
13007f790cSJiri Pirko #include <linux/types.h>
14007f790cSJiri Pirko #include <linux/init.h>
1503bf0c28SJiri Pirko #include <linux/mutex.h>
1603bf0c28SJiri Pirko #include <linux/notifier.h>
17007f790cSJiri Pirko #include <linux/netdevice.h>
18850d0cbcSJiri Pirko #include <linux/etherdevice.h>
1947f8328bSScott Feldman #include <linux/if_bridge.h>
207ea6eb3fSJiri Pirko #include <linux/list.h>
21793f4014SJiri Pirko #include <linux/workqueue.h>
2287aaf2caSNikolay Aleksandrov #include <linux/if_vlan.h>
234f2c6ae5SIdo Schimmel #include <linux/rtnetlink.h>
24007f790cSJiri Pirko #include <net/switchdev.h>
25007f790cSJiri Pirko 
26007f790cSJiri Pirko /**
277ea6eb3fSJiri Pirko  *	switchdev_trans_item_enqueue - Enqueue data item to transaction queue
287ea6eb3fSJiri Pirko  *
297ea6eb3fSJiri Pirko  *	@trans: transaction
307ea6eb3fSJiri Pirko  *	@data: pointer to data being queued
317ea6eb3fSJiri Pirko  *	@destructor: data destructor
327ea6eb3fSJiri Pirko  *	@tritem: transaction item being queued
337ea6eb3fSJiri Pirko  *
347ea6eb3fSJiri Pirko  *	Enqeueue data item to transaction queue. tritem is typically placed in
357ea6eb3fSJiri Pirko  *	cointainter pointed at by data pointer. Destructor is called on
367ea6eb3fSJiri Pirko  *	transaction abort and after successful commit phase in case
377ea6eb3fSJiri Pirko  *	the caller did not dequeue the item before.
387ea6eb3fSJiri Pirko  */
397ea6eb3fSJiri Pirko void switchdev_trans_item_enqueue(struct switchdev_trans *trans,
407ea6eb3fSJiri Pirko 				  void *data, void (*destructor)(void const *),
417ea6eb3fSJiri Pirko 				  struct switchdev_trans_item *tritem)
427ea6eb3fSJiri Pirko {
437ea6eb3fSJiri Pirko 	tritem->data = data;
447ea6eb3fSJiri Pirko 	tritem->destructor = destructor;
457ea6eb3fSJiri Pirko 	list_add_tail(&tritem->list, &trans->item_list);
467ea6eb3fSJiri Pirko }
477ea6eb3fSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_trans_item_enqueue);
487ea6eb3fSJiri Pirko 
497ea6eb3fSJiri Pirko static struct switchdev_trans_item *
507ea6eb3fSJiri Pirko __switchdev_trans_item_dequeue(struct switchdev_trans *trans)
517ea6eb3fSJiri Pirko {
527ea6eb3fSJiri Pirko 	struct switchdev_trans_item *tritem;
537ea6eb3fSJiri Pirko 
547ea6eb3fSJiri Pirko 	if (list_empty(&trans->item_list))
557ea6eb3fSJiri Pirko 		return NULL;
567ea6eb3fSJiri Pirko 	tritem = list_first_entry(&trans->item_list,
577ea6eb3fSJiri Pirko 				  struct switchdev_trans_item, list);
587ea6eb3fSJiri Pirko 	list_del(&tritem->list);
597ea6eb3fSJiri Pirko 	return tritem;
607ea6eb3fSJiri Pirko }
617ea6eb3fSJiri Pirko 
627ea6eb3fSJiri Pirko /**
637ea6eb3fSJiri Pirko  *	switchdev_trans_item_dequeue - Dequeue data item from transaction queue
647ea6eb3fSJiri Pirko  *
657ea6eb3fSJiri Pirko  *	@trans: transaction
667ea6eb3fSJiri Pirko  */
677ea6eb3fSJiri Pirko void *switchdev_trans_item_dequeue(struct switchdev_trans *trans)
687ea6eb3fSJiri Pirko {
697ea6eb3fSJiri Pirko 	struct switchdev_trans_item *tritem;
707ea6eb3fSJiri Pirko 
717ea6eb3fSJiri Pirko 	tritem = __switchdev_trans_item_dequeue(trans);
727ea6eb3fSJiri Pirko 	BUG_ON(!tritem);
737ea6eb3fSJiri Pirko 	return tritem->data;
747ea6eb3fSJiri Pirko }
757ea6eb3fSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_trans_item_dequeue);
767ea6eb3fSJiri Pirko 
777ea6eb3fSJiri Pirko static void switchdev_trans_init(struct switchdev_trans *trans)
787ea6eb3fSJiri Pirko {
797ea6eb3fSJiri Pirko 	INIT_LIST_HEAD(&trans->item_list);
807ea6eb3fSJiri Pirko }
817ea6eb3fSJiri Pirko 
827ea6eb3fSJiri Pirko static void switchdev_trans_items_destroy(struct switchdev_trans *trans)
837ea6eb3fSJiri Pirko {
847ea6eb3fSJiri Pirko 	struct switchdev_trans_item *tritem;
857ea6eb3fSJiri Pirko 
867ea6eb3fSJiri Pirko 	while ((tritem = __switchdev_trans_item_dequeue(trans)))
877ea6eb3fSJiri Pirko 		tritem->destructor(tritem->data);
887ea6eb3fSJiri Pirko }
897ea6eb3fSJiri Pirko 
907ea6eb3fSJiri Pirko static void switchdev_trans_items_warn_destroy(struct net_device *dev,
917ea6eb3fSJiri Pirko 					       struct switchdev_trans *trans)
927ea6eb3fSJiri Pirko {
937ea6eb3fSJiri Pirko 	WARN(!list_empty(&trans->item_list), "%s: transaction item queue is not empty.\n",
947ea6eb3fSJiri Pirko 	     dev->name);
957ea6eb3fSJiri Pirko 	switchdev_trans_items_destroy(trans);
967ea6eb3fSJiri Pirko }
977ea6eb3fSJiri Pirko 
98793f4014SJiri Pirko static LIST_HEAD(deferred);
99793f4014SJiri Pirko static DEFINE_SPINLOCK(deferred_lock);
100793f4014SJiri Pirko 
101793f4014SJiri Pirko typedef void switchdev_deferred_func_t(struct net_device *dev,
102793f4014SJiri Pirko 				       const void *data);
103793f4014SJiri Pirko 
104793f4014SJiri Pirko struct switchdev_deferred_item {
105793f4014SJiri Pirko 	struct list_head list;
106793f4014SJiri Pirko 	struct net_device *dev;
107793f4014SJiri Pirko 	switchdev_deferred_func_t *func;
108793f4014SJiri Pirko 	unsigned long data[0];
109793f4014SJiri Pirko };
110793f4014SJiri Pirko 
111793f4014SJiri Pirko static struct switchdev_deferred_item *switchdev_deferred_dequeue(void)
112793f4014SJiri Pirko {
113793f4014SJiri Pirko 	struct switchdev_deferred_item *dfitem;
114793f4014SJiri Pirko 
115793f4014SJiri Pirko 	spin_lock_bh(&deferred_lock);
116793f4014SJiri Pirko 	if (list_empty(&deferred)) {
117793f4014SJiri Pirko 		dfitem = NULL;
118793f4014SJiri Pirko 		goto unlock;
119793f4014SJiri Pirko 	}
120793f4014SJiri Pirko 	dfitem = list_first_entry(&deferred,
121793f4014SJiri Pirko 				  struct switchdev_deferred_item, list);
122793f4014SJiri Pirko 	list_del(&dfitem->list);
123793f4014SJiri Pirko unlock:
124793f4014SJiri Pirko 	spin_unlock_bh(&deferred_lock);
125793f4014SJiri Pirko 	return dfitem;
126793f4014SJiri Pirko }
127793f4014SJiri Pirko 
128793f4014SJiri Pirko /**
129793f4014SJiri Pirko  *	switchdev_deferred_process - Process ops in deferred queue
130793f4014SJiri Pirko  *
131793f4014SJiri Pirko  *	Called to flush the ops currently queued in deferred ops queue.
132793f4014SJiri Pirko  *	rtnl_lock must be held.
133793f4014SJiri Pirko  */
134793f4014SJiri Pirko void switchdev_deferred_process(void)
135793f4014SJiri Pirko {
136793f4014SJiri Pirko 	struct switchdev_deferred_item *dfitem;
137793f4014SJiri Pirko 
138793f4014SJiri Pirko 	ASSERT_RTNL();
139793f4014SJiri Pirko 
140793f4014SJiri Pirko 	while ((dfitem = switchdev_deferred_dequeue())) {
141793f4014SJiri Pirko 		dfitem->func(dfitem->dev, dfitem->data);
142793f4014SJiri Pirko 		dev_put(dfitem->dev);
143793f4014SJiri Pirko 		kfree(dfitem);
144793f4014SJiri Pirko 	}
145793f4014SJiri Pirko }
146793f4014SJiri Pirko EXPORT_SYMBOL_GPL(switchdev_deferred_process);
147793f4014SJiri Pirko 
148793f4014SJiri Pirko static void switchdev_deferred_process_work(struct work_struct *work)
149793f4014SJiri Pirko {
150793f4014SJiri Pirko 	rtnl_lock();
151793f4014SJiri Pirko 	switchdev_deferred_process();
152793f4014SJiri Pirko 	rtnl_unlock();
153793f4014SJiri Pirko }
154793f4014SJiri Pirko 
155793f4014SJiri Pirko static DECLARE_WORK(deferred_process_work, switchdev_deferred_process_work);
156793f4014SJiri Pirko 
157793f4014SJiri Pirko static int switchdev_deferred_enqueue(struct net_device *dev,
158793f4014SJiri Pirko 				      const void *data, size_t data_len,
159793f4014SJiri Pirko 				      switchdev_deferred_func_t *func)
160793f4014SJiri Pirko {
161793f4014SJiri Pirko 	struct switchdev_deferred_item *dfitem;
162793f4014SJiri Pirko 
163793f4014SJiri Pirko 	dfitem = kmalloc(sizeof(*dfitem) + data_len, GFP_ATOMIC);
164793f4014SJiri Pirko 	if (!dfitem)
165793f4014SJiri Pirko 		return -ENOMEM;
166793f4014SJiri Pirko 	dfitem->dev = dev;
167793f4014SJiri Pirko 	dfitem->func = func;
168793f4014SJiri Pirko 	memcpy(dfitem->data, data, data_len);
169793f4014SJiri Pirko 	dev_hold(dev);
170793f4014SJiri Pirko 	spin_lock_bh(&deferred_lock);
171793f4014SJiri Pirko 	list_add_tail(&dfitem->list, &deferred);
172793f4014SJiri Pirko 	spin_unlock_bh(&deferred_lock);
173793f4014SJiri Pirko 	schedule_work(&deferred_process_work);
174793f4014SJiri Pirko 	return 0;
175793f4014SJiri Pirko }
176793f4014SJiri Pirko 
1777ea6eb3fSJiri Pirko /**
1783094333dSScott Feldman  *	switchdev_port_attr_get - Get port attribute
1793094333dSScott Feldman  *
1803094333dSScott Feldman  *	@dev: port device
1813094333dSScott Feldman  *	@attr: attribute to get
1823094333dSScott Feldman  */
1833094333dSScott Feldman int switchdev_port_attr_get(struct net_device *dev, struct switchdev_attr *attr)
1843094333dSScott Feldman {
1853094333dSScott Feldman 	const struct switchdev_ops *ops = dev->switchdev_ops;
1863094333dSScott Feldman 	struct net_device *lower_dev;
1873094333dSScott Feldman 	struct list_head *iter;
1883094333dSScott Feldman 	struct switchdev_attr first = {
1891f868398SJiri Pirko 		.id = SWITCHDEV_ATTR_ID_UNDEFINED
1903094333dSScott Feldman 	};
1913094333dSScott Feldman 	int err = -EOPNOTSUPP;
1923094333dSScott Feldman 
1933094333dSScott Feldman 	if (ops && ops->switchdev_port_attr_get)
1943094333dSScott Feldman 		return ops->switchdev_port_attr_get(dev, attr);
1953094333dSScott Feldman 
1963094333dSScott Feldman 	if (attr->flags & SWITCHDEV_F_NO_RECURSE)
1973094333dSScott Feldman 		return err;
1983094333dSScott Feldman 
1993094333dSScott Feldman 	/* Switch device port(s) may be stacked under
2003094333dSScott Feldman 	 * bond/team/vlan dev, so recurse down to get attr on
2013094333dSScott Feldman 	 * each port.  Return -ENODATA if attr values don't
2023094333dSScott Feldman 	 * compare across ports.
2033094333dSScott Feldman 	 */
2043094333dSScott Feldman 
2053094333dSScott Feldman 	netdev_for_each_lower_dev(dev, lower_dev, iter) {
2063094333dSScott Feldman 		err = switchdev_port_attr_get(lower_dev, attr);
2073094333dSScott Feldman 		if (err)
2083094333dSScott Feldman 			break;
2091f868398SJiri Pirko 		if (first.id == SWITCHDEV_ATTR_ID_UNDEFINED)
2103094333dSScott Feldman 			first = *attr;
2113094333dSScott Feldman 		else if (memcmp(&first, attr, sizeof(*attr)))
2123094333dSScott Feldman 			return -ENODATA;
2133094333dSScott Feldman 	}
2143094333dSScott Feldman 
2153094333dSScott Feldman 	return err;
2163094333dSScott Feldman }
2173094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_get);
2183094333dSScott Feldman 
2193094333dSScott Feldman static int __switchdev_port_attr_set(struct net_device *dev,
220f7fadf30SJiri Pirko 				     const struct switchdev_attr *attr,
2217ea6eb3fSJiri Pirko 				     struct switchdev_trans *trans)
2223094333dSScott Feldman {
2233094333dSScott Feldman 	const struct switchdev_ops *ops = dev->switchdev_ops;
2243094333dSScott Feldman 	struct net_device *lower_dev;
2253094333dSScott Feldman 	struct list_head *iter;
2263094333dSScott Feldman 	int err = -EOPNOTSUPP;
2273094333dSScott Feldman 
2280c63d80cSJiri Pirko 	if (ops && ops->switchdev_port_attr_set) {
2290c63d80cSJiri Pirko 		err = ops->switchdev_port_attr_set(dev, attr, trans);
2300c63d80cSJiri Pirko 		goto done;
2310c63d80cSJiri Pirko 	}
2323094333dSScott Feldman 
2333094333dSScott Feldman 	if (attr->flags & SWITCHDEV_F_NO_RECURSE)
234464314eaSScott Feldman 		goto done;
2353094333dSScott Feldman 
2363094333dSScott Feldman 	/* Switch device port(s) may be stacked under
2373094333dSScott Feldman 	 * bond/team/vlan dev, so recurse down to set attr on
2383094333dSScott Feldman 	 * each port.
2393094333dSScott Feldman 	 */
2403094333dSScott Feldman 
2413094333dSScott Feldman 	netdev_for_each_lower_dev(dev, lower_dev, iter) {
2427ea6eb3fSJiri Pirko 		err = __switchdev_port_attr_set(lower_dev, attr, trans);
2433094333dSScott Feldman 		if (err)
2443094333dSScott Feldman 			break;
2453094333dSScott Feldman 	}
2463094333dSScott Feldman 
247464314eaSScott Feldman done:
248464314eaSScott Feldman 	if (err == -EOPNOTSUPP && attr->flags & SWITCHDEV_F_SKIP_EOPNOTSUPP)
249464314eaSScott Feldman 		err = 0;
250464314eaSScott Feldman 
2513094333dSScott Feldman 	return err;
2523094333dSScott Feldman }
2533094333dSScott Feldman 
2540bc05d58SJiri Pirko static int switchdev_port_attr_set_now(struct net_device *dev,
255f7fadf30SJiri Pirko 				       const struct switchdev_attr *attr)
2563094333dSScott Feldman {
2577ea6eb3fSJiri Pirko 	struct switchdev_trans trans;
2583094333dSScott Feldman 	int err;
2593094333dSScott Feldman 
2607ea6eb3fSJiri Pirko 	switchdev_trans_init(&trans);
2617ea6eb3fSJiri Pirko 
2623094333dSScott Feldman 	/* Phase I: prepare for attr set. Driver/device should fail
2633094333dSScott Feldman 	 * here if there are going to be issues in the commit phase,
2643094333dSScott Feldman 	 * such as lack of resources or support.  The driver/device
2653094333dSScott Feldman 	 * should reserve resources needed for the commit phase here,
2663094333dSScott Feldman 	 * but should not commit the attr.
2673094333dSScott Feldman 	 */
2683094333dSScott Feldman 
269f623ab7fSJiri Pirko 	trans.ph_prepare = true;
2707ea6eb3fSJiri Pirko 	err = __switchdev_port_attr_set(dev, attr, &trans);
2713094333dSScott Feldman 	if (err) {
2723094333dSScott Feldman 		/* Prepare phase failed: abort the transaction.  Any
2733094333dSScott Feldman 		 * resources reserved in the prepare phase are
2743094333dSScott Feldman 		 * released.
2753094333dSScott Feldman 		 */
2763094333dSScott Feldman 
2779f6467cfSJiri Pirko 		if (err != -EOPNOTSUPP)
2787ea6eb3fSJiri Pirko 			switchdev_trans_items_destroy(&trans);
2793094333dSScott Feldman 
2803094333dSScott Feldman 		return err;
2813094333dSScott Feldman 	}
2823094333dSScott Feldman 
2833094333dSScott Feldman 	/* Phase II: commit attr set.  This cannot fail as a fault
2843094333dSScott Feldman 	 * of driver/device.  If it does, it's a bug in the driver/device
2853094333dSScott Feldman 	 * because the driver said everythings was OK in phase I.
2863094333dSScott Feldman 	 */
2873094333dSScott Feldman 
288f623ab7fSJiri Pirko 	trans.ph_prepare = false;
2897ea6eb3fSJiri Pirko 	err = __switchdev_port_attr_set(dev, attr, &trans);
290e9fdaec0SScott Feldman 	WARN(err, "%s: Commit of attribute (id=%d) failed.\n",
291e9fdaec0SScott Feldman 	     dev->name, attr->id);
2927ea6eb3fSJiri Pirko 	switchdev_trans_items_warn_destroy(dev, &trans);
2933094333dSScott Feldman 
2943094333dSScott Feldman 	return err;
2953094333dSScott Feldman }
2960bc05d58SJiri Pirko 
2970bc05d58SJiri Pirko static void switchdev_port_attr_set_deferred(struct net_device *dev,
2980bc05d58SJiri Pirko 					     const void *data)
2990bc05d58SJiri Pirko {
3000bc05d58SJiri Pirko 	const struct switchdev_attr *attr = data;
3010bc05d58SJiri Pirko 	int err;
3020bc05d58SJiri Pirko 
3030bc05d58SJiri Pirko 	err = switchdev_port_attr_set_now(dev, attr);
3040bc05d58SJiri Pirko 	if (err && err != -EOPNOTSUPP)
3050bc05d58SJiri Pirko 		netdev_err(dev, "failed (err=%d) to set attribute (id=%d)\n",
3060bc05d58SJiri Pirko 			   err, attr->id);
3077ceb2afbSElad Raz 	if (attr->complete)
3087ceb2afbSElad Raz 		attr->complete(dev, err, attr->complete_priv);
3090bc05d58SJiri Pirko }
3100bc05d58SJiri Pirko 
3110bc05d58SJiri Pirko static int switchdev_port_attr_set_defer(struct net_device *dev,
3120bc05d58SJiri Pirko 					 const struct switchdev_attr *attr)
3130bc05d58SJiri Pirko {
3140bc05d58SJiri Pirko 	return switchdev_deferred_enqueue(dev, attr, sizeof(*attr),
3150bc05d58SJiri Pirko 					  switchdev_port_attr_set_deferred);
3160bc05d58SJiri Pirko }
3170bc05d58SJiri Pirko 
3180bc05d58SJiri Pirko /**
3190bc05d58SJiri Pirko  *	switchdev_port_attr_set - Set port attribute
3200bc05d58SJiri Pirko  *
3210bc05d58SJiri Pirko  *	@dev: port device
3220bc05d58SJiri Pirko  *	@attr: attribute to set
3230bc05d58SJiri Pirko  *
3240bc05d58SJiri Pirko  *	Use a 2-phase prepare-commit transaction model to ensure
3250bc05d58SJiri Pirko  *	system is not left in a partially updated state due to
3260bc05d58SJiri Pirko  *	failure from driver/device.
3270bc05d58SJiri Pirko  *
3280bc05d58SJiri Pirko  *	rtnl_lock must be held and must not be in atomic section,
3290bc05d58SJiri Pirko  *	in case SWITCHDEV_F_DEFER flag is not set.
3300bc05d58SJiri Pirko  */
3310bc05d58SJiri Pirko int switchdev_port_attr_set(struct net_device *dev,
3320bc05d58SJiri Pirko 			    const struct switchdev_attr *attr)
3330bc05d58SJiri Pirko {
3340bc05d58SJiri Pirko 	if (attr->flags & SWITCHDEV_F_DEFER)
3350bc05d58SJiri Pirko 		return switchdev_port_attr_set_defer(dev, attr);
3360bc05d58SJiri Pirko 	ASSERT_RTNL();
3370bc05d58SJiri Pirko 	return switchdev_port_attr_set_now(dev, attr);
3380bc05d58SJiri Pirko }
3393094333dSScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_attr_set);
3403094333dSScott Feldman 
341e258d919SScott Feldman static size_t switchdev_obj_size(const struct switchdev_obj *obj)
342e258d919SScott Feldman {
343e258d919SScott Feldman 	switch (obj->id) {
344e258d919SScott Feldman 	case SWITCHDEV_OBJ_ID_PORT_VLAN:
345e258d919SScott Feldman 		return sizeof(struct switchdev_obj_port_vlan);
3464d41e125SElad Raz 	case SWITCHDEV_OBJ_ID_PORT_MDB:
3474d41e125SElad Raz 		return sizeof(struct switchdev_obj_port_mdb);
34847d5b6dbSAndrew Lunn 	case SWITCHDEV_OBJ_ID_HOST_MDB:
34947d5b6dbSAndrew Lunn 		return sizeof(struct switchdev_obj_port_mdb);
350e258d919SScott Feldman 	default:
351e258d919SScott Feldman 		BUG();
352e258d919SScott Feldman 	}
353e258d919SScott Feldman 	return 0;
354e258d919SScott Feldman }
355e258d919SScott Feldman 
356d17d9f5eSPetr Machata static int switchdev_port_obj_notify(enum switchdev_notifier_type nt,
357d17d9f5eSPetr Machata 				     struct net_device *dev,
358648b4a99SJiri Pirko 				     const struct switchdev_obj *obj,
35969b7320eSPetr Machata 				     struct switchdev_trans *trans,
36069b7320eSPetr Machata 				     struct netlink_ext_ack *extack)
361491d0f15SScott Feldman {
362d17d9f5eSPetr Machata 	int rc;
363d17d9f5eSPetr Machata 	int err;
364491d0f15SScott Feldman 
365d17d9f5eSPetr Machata 	struct switchdev_notifier_port_obj_info obj_info = {
366d17d9f5eSPetr Machata 		.obj = obj,
367d17d9f5eSPetr Machata 		.trans = trans,
368d17d9f5eSPetr Machata 		.handled = false,
369d17d9f5eSPetr Machata 	};
370491d0f15SScott Feldman 
371*479c86dcSPetr Machata 	rc = call_switchdev_blocking_notifiers(nt, dev, &obj_info.info, extack);
372d17d9f5eSPetr Machata 	err = notifier_to_errno(rc);
373d17d9f5eSPetr Machata 	if (err) {
374d17d9f5eSPetr Machata 		WARN_ON(!obj_info.handled);
375491d0f15SScott Feldman 		return err;
376491d0f15SScott Feldman 	}
377d17d9f5eSPetr Machata 	if (!obj_info.handled)
378d17d9f5eSPetr Machata 		return -EOPNOTSUPP;
379d17d9f5eSPetr Machata 	return 0;
380d17d9f5eSPetr Machata }
381491d0f15SScott Feldman 
3824d429c5dSJiri Pirko static int switchdev_port_obj_add_now(struct net_device *dev,
38369b7320eSPetr Machata 				      const struct switchdev_obj *obj,
38469b7320eSPetr Machata 				      struct netlink_ext_ack *extack)
385491d0f15SScott Feldman {
3867ea6eb3fSJiri Pirko 	struct switchdev_trans trans;
387491d0f15SScott Feldman 	int err;
388491d0f15SScott Feldman 
389491d0f15SScott Feldman 	ASSERT_RTNL();
390491d0f15SScott Feldman 
3917ea6eb3fSJiri Pirko 	switchdev_trans_init(&trans);
3927ea6eb3fSJiri Pirko 
393491d0f15SScott Feldman 	/* Phase I: prepare for obj add. Driver/device should fail
394491d0f15SScott Feldman 	 * here if there are going to be issues in the commit phase,
395491d0f15SScott Feldman 	 * such as lack of resources or support.  The driver/device
396491d0f15SScott Feldman 	 * should reserve resources needed for the commit phase here,
397491d0f15SScott Feldman 	 * but should not commit the obj.
398491d0f15SScott Feldman 	 */
399491d0f15SScott Feldman 
400f623ab7fSJiri Pirko 	trans.ph_prepare = true;
401d17d9f5eSPetr Machata 	err = switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD,
40269b7320eSPetr Machata 					dev, obj, &trans, extack);
403491d0f15SScott Feldman 	if (err) {
404491d0f15SScott Feldman 		/* Prepare phase failed: abort the transaction.  Any
405491d0f15SScott Feldman 		 * resources reserved in the prepare phase are
406491d0f15SScott Feldman 		 * released.
407491d0f15SScott Feldman 		 */
408491d0f15SScott Feldman 
4099f6467cfSJiri Pirko 		if (err != -EOPNOTSUPP)
4107ea6eb3fSJiri Pirko 			switchdev_trans_items_destroy(&trans);
411491d0f15SScott Feldman 
412491d0f15SScott Feldman 		return err;
413491d0f15SScott Feldman 	}
414491d0f15SScott Feldman 
415491d0f15SScott Feldman 	/* Phase II: commit obj add.  This cannot fail as a fault
416491d0f15SScott Feldman 	 * of driver/device.  If it does, it's a bug in the driver/device
417491d0f15SScott Feldman 	 * because the driver said everythings was OK in phase I.
418491d0f15SScott Feldman 	 */
419491d0f15SScott Feldman 
420f623ab7fSJiri Pirko 	trans.ph_prepare = false;
421d17d9f5eSPetr Machata 	err = switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_ADD,
42269b7320eSPetr Machata 					dev, obj, &trans, extack);
4239e8f4a54SJiri Pirko 	WARN(err, "%s: Commit of object (id=%d) failed.\n", dev->name, obj->id);
4247ea6eb3fSJiri Pirko 	switchdev_trans_items_warn_destroy(dev, &trans);
425491d0f15SScott Feldman 
426491d0f15SScott Feldman 	return err;
427491d0f15SScott Feldman }
4284d429c5dSJiri Pirko 
4294d429c5dSJiri Pirko static void switchdev_port_obj_add_deferred(struct net_device *dev,
4304d429c5dSJiri Pirko 					    const void *data)
4314d429c5dSJiri Pirko {
4324d429c5dSJiri Pirko 	const struct switchdev_obj *obj = data;
4334d429c5dSJiri Pirko 	int err;
4344d429c5dSJiri Pirko 
43569b7320eSPetr Machata 	err = switchdev_port_obj_add_now(dev, obj, NULL);
4364d429c5dSJiri Pirko 	if (err && err != -EOPNOTSUPP)
4374d429c5dSJiri Pirko 		netdev_err(dev, "failed (err=%d) to add object (id=%d)\n",
4384d429c5dSJiri Pirko 			   err, obj->id);
4397ceb2afbSElad Raz 	if (obj->complete)
4407ceb2afbSElad Raz 		obj->complete(dev, err, obj->complete_priv);
4414d429c5dSJiri Pirko }
4424d429c5dSJiri Pirko 
4434d429c5dSJiri Pirko static int switchdev_port_obj_add_defer(struct net_device *dev,
4444d429c5dSJiri Pirko 					const struct switchdev_obj *obj)
4454d429c5dSJiri Pirko {
446e258d919SScott Feldman 	return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj),
4474d429c5dSJiri Pirko 					  switchdev_port_obj_add_deferred);
4484d429c5dSJiri Pirko }
449491d0f15SScott Feldman 
450491d0f15SScott Feldman /**
4514d429c5dSJiri Pirko  *	switchdev_port_obj_add - Add port object
452491d0f15SScott Feldman  *
453491d0f15SScott Feldman  *	@dev: port device
454ab069002SVivien Didelot  *	@id: object ID
4554d429c5dSJiri Pirko  *	@obj: object to add
4564d429c5dSJiri Pirko  *
4574d429c5dSJiri Pirko  *	Use a 2-phase prepare-commit transaction model to ensure
4584d429c5dSJiri Pirko  *	system is not left in a partially updated state due to
4594d429c5dSJiri Pirko  *	failure from driver/device.
4604d429c5dSJiri Pirko  *
4614d429c5dSJiri Pirko  *	rtnl_lock must be held and must not be in atomic section,
4624d429c5dSJiri Pirko  *	in case SWITCHDEV_F_DEFER flag is not set.
463491d0f15SScott Feldman  */
4644d429c5dSJiri Pirko int switchdev_port_obj_add(struct net_device *dev,
46569b7320eSPetr Machata 			   const struct switchdev_obj *obj,
46669b7320eSPetr Machata 			   struct netlink_ext_ack *extack)
4674d429c5dSJiri Pirko {
4684d429c5dSJiri Pirko 	if (obj->flags & SWITCHDEV_F_DEFER)
4694d429c5dSJiri Pirko 		return switchdev_port_obj_add_defer(dev, obj);
4704d429c5dSJiri Pirko 	ASSERT_RTNL();
47169b7320eSPetr Machata 	return switchdev_port_obj_add_now(dev, obj, extack);
4724d429c5dSJiri Pirko }
4734d429c5dSJiri Pirko EXPORT_SYMBOL_GPL(switchdev_port_obj_add);
4744d429c5dSJiri Pirko 
4754d429c5dSJiri Pirko static int switchdev_port_obj_del_now(struct net_device *dev,
476648b4a99SJiri Pirko 				      const struct switchdev_obj *obj)
477491d0f15SScott Feldman {
478d17d9f5eSPetr Machata 	return switchdev_port_obj_notify(SWITCHDEV_PORT_OBJ_DEL,
47969b7320eSPetr Machata 					 dev, obj, NULL, NULL);
480491d0f15SScott Feldman }
4814d429c5dSJiri Pirko 
4824d429c5dSJiri Pirko static void switchdev_port_obj_del_deferred(struct net_device *dev,
4834d429c5dSJiri Pirko 					    const void *data)
4844d429c5dSJiri Pirko {
4854d429c5dSJiri Pirko 	const struct switchdev_obj *obj = data;
4864d429c5dSJiri Pirko 	int err;
4874d429c5dSJiri Pirko 
4884d429c5dSJiri Pirko 	err = switchdev_port_obj_del_now(dev, obj);
4894d429c5dSJiri Pirko 	if (err && err != -EOPNOTSUPP)
4904d429c5dSJiri Pirko 		netdev_err(dev, "failed (err=%d) to del object (id=%d)\n",
4914d429c5dSJiri Pirko 			   err, obj->id);
4927ceb2afbSElad Raz 	if (obj->complete)
4937ceb2afbSElad Raz 		obj->complete(dev, err, obj->complete_priv);
4944d429c5dSJiri Pirko }
4954d429c5dSJiri Pirko 
4964d429c5dSJiri Pirko static int switchdev_port_obj_del_defer(struct net_device *dev,
4974d429c5dSJiri Pirko 					const struct switchdev_obj *obj)
4984d429c5dSJiri Pirko {
499e258d919SScott Feldman 	return switchdev_deferred_enqueue(dev, obj, switchdev_obj_size(obj),
5004d429c5dSJiri Pirko 					  switchdev_port_obj_del_deferred);
5014d429c5dSJiri Pirko }
5024d429c5dSJiri Pirko 
5034d429c5dSJiri Pirko /**
5044d429c5dSJiri Pirko  *	switchdev_port_obj_del - Delete port object
5054d429c5dSJiri Pirko  *
5064d429c5dSJiri Pirko  *	@dev: port device
5074d429c5dSJiri Pirko  *	@id: object ID
5084d429c5dSJiri Pirko  *	@obj: object to delete
5094d429c5dSJiri Pirko  *
5104d429c5dSJiri Pirko  *	rtnl_lock must be held and must not be in atomic section,
5114d429c5dSJiri Pirko  *	in case SWITCHDEV_F_DEFER flag is not set.
5124d429c5dSJiri Pirko  */
5134d429c5dSJiri Pirko int switchdev_port_obj_del(struct net_device *dev,
5144d429c5dSJiri Pirko 			   const struct switchdev_obj *obj)
5154d429c5dSJiri Pirko {
5164d429c5dSJiri Pirko 	if (obj->flags & SWITCHDEV_F_DEFER)
5174d429c5dSJiri Pirko 		return switchdev_port_obj_del_defer(dev, obj);
5184d429c5dSJiri Pirko 	ASSERT_RTNL();
5194d429c5dSJiri Pirko 	return switchdev_port_obj_del_now(dev, obj);
5204d429c5dSJiri Pirko }
521491d0f15SScott Feldman EXPORT_SYMBOL_GPL(switchdev_port_obj_del);
522491d0f15SScott Feldman 
523ff5cf100SArkadi Sharshevsky static ATOMIC_NOTIFIER_HEAD(switchdev_notif_chain);
524a93e3b17SPetr Machata static BLOCKING_NOTIFIER_HEAD(switchdev_blocking_notif_chain);
52503bf0c28SJiri Pirko 
52603bf0c28SJiri Pirko /**
527ebb9a03aSJiri Pirko  *	register_switchdev_notifier - Register notifier
52803bf0c28SJiri Pirko  *	@nb: notifier_block
52903bf0c28SJiri Pirko  *
530ff5cf100SArkadi Sharshevsky  *	Register switch device notifier.
53103bf0c28SJiri Pirko  */
532ebb9a03aSJiri Pirko int register_switchdev_notifier(struct notifier_block *nb)
53303bf0c28SJiri Pirko {
534ff5cf100SArkadi Sharshevsky 	return atomic_notifier_chain_register(&switchdev_notif_chain, nb);
53503bf0c28SJiri Pirko }
536ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(register_switchdev_notifier);
53703bf0c28SJiri Pirko 
53803bf0c28SJiri Pirko /**
539ebb9a03aSJiri Pirko  *	unregister_switchdev_notifier - Unregister notifier
54003bf0c28SJiri Pirko  *	@nb: notifier_block
54103bf0c28SJiri Pirko  *
54203bf0c28SJiri Pirko  *	Unregister switch device notifier.
54303bf0c28SJiri Pirko  */
544ebb9a03aSJiri Pirko int unregister_switchdev_notifier(struct notifier_block *nb)
54503bf0c28SJiri Pirko {
546ff5cf100SArkadi Sharshevsky 	return atomic_notifier_chain_unregister(&switchdev_notif_chain, nb);
54703bf0c28SJiri Pirko }
548ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(unregister_switchdev_notifier);
54903bf0c28SJiri Pirko 
55003bf0c28SJiri Pirko /**
551ebb9a03aSJiri Pirko  *	call_switchdev_notifiers - Call notifiers
55203bf0c28SJiri Pirko  *	@val: value passed unmodified to notifier function
55303bf0c28SJiri Pirko  *	@dev: port device
55403bf0c28SJiri Pirko  *	@info: notifier information data
55503bf0c28SJiri Pirko  *
556ff5cf100SArkadi Sharshevsky  *	Call all network notifier blocks.
55703bf0c28SJiri Pirko  */
558ebb9a03aSJiri Pirko int call_switchdev_notifiers(unsigned long val, struct net_device *dev,
559ebb9a03aSJiri Pirko 			     struct switchdev_notifier_info *info)
56003bf0c28SJiri Pirko {
56103bf0c28SJiri Pirko 	info->dev = dev;
562*479c86dcSPetr Machata 	info->extack = NULL;
563ff5cf100SArkadi Sharshevsky 	return atomic_notifier_call_chain(&switchdev_notif_chain, val, info);
56403bf0c28SJiri Pirko }
565ebb9a03aSJiri Pirko EXPORT_SYMBOL_GPL(call_switchdev_notifiers);
5668a44dbb2SRoopa Prabhu 
567a93e3b17SPetr Machata int register_switchdev_blocking_notifier(struct notifier_block *nb)
568a93e3b17SPetr Machata {
569a93e3b17SPetr Machata 	struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain;
570a93e3b17SPetr Machata 
571a93e3b17SPetr Machata 	return blocking_notifier_chain_register(chain, nb);
572a93e3b17SPetr Machata }
573a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(register_switchdev_blocking_notifier);
574a93e3b17SPetr Machata 
575a93e3b17SPetr Machata int unregister_switchdev_blocking_notifier(struct notifier_block *nb)
576a93e3b17SPetr Machata {
577a93e3b17SPetr Machata 	struct blocking_notifier_head *chain = &switchdev_blocking_notif_chain;
578a93e3b17SPetr Machata 
579a93e3b17SPetr Machata 	return blocking_notifier_chain_unregister(chain, nb);
580a93e3b17SPetr Machata }
581a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(unregister_switchdev_blocking_notifier);
582a93e3b17SPetr Machata 
583a93e3b17SPetr Machata int call_switchdev_blocking_notifiers(unsigned long val, struct net_device *dev,
584*479c86dcSPetr Machata 				      struct switchdev_notifier_info *info,
585*479c86dcSPetr Machata 				      struct netlink_ext_ack *extack)
586a93e3b17SPetr Machata {
587a93e3b17SPetr Machata 	info->dev = dev;
588*479c86dcSPetr Machata 	info->extack = extack;
589a93e3b17SPetr Machata 	return blocking_notifier_call_chain(&switchdev_blocking_notif_chain,
590a93e3b17SPetr Machata 					    val, info);
591a93e3b17SPetr Machata }
592a93e3b17SPetr Machata EXPORT_SYMBOL_GPL(call_switchdev_blocking_notifiers);
593a93e3b17SPetr Machata 
5948438884dSOr Gerlitz bool switchdev_port_same_parent_id(struct net_device *a,
5951a3b2ec9SScott Feldman 				   struct net_device *b)
5961a3b2ec9SScott Feldman {
5971a3b2ec9SScott Feldman 	struct switchdev_attr a_attr = {
5986ff64f6fSIdo Schimmel 		.orig_dev = a,
5991f868398SJiri Pirko 		.id = SWITCHDEV_ATTR_ID_PORT_PARENT_ID,
6001a3b2ec9SScott Feldman 	};
6011a3b2ec9SScott Feldman 	struct switchdev_attr b_attr = {
6026ff64f6fSIdo Schimmel 		.orig_dev = b,
6031f868398SJiri Pirko 		.id = SWITCHDEV_ATTR_ID_PORT_PARENT_ID,
6041a3b2ec9SScott Feldman 	};
6051a3b2ec9SScott Feldman 
6061a3b2ec9SScott Feldman 	if (switchdev_port_attr_get(a, &a_attr) ||
6071a3b2ec9SScott Feldman 	    switchdev_port_attr_get(b, &b_attr))
6081a3b2ec9SScott Feldman 		return false;
6091a3b2ec9SScott Feldman 
6101a3b2ec9SScott Feldman 	return netdev_phys_item_id_same(&a_attr.u.ppid, &b_attr.u.ppid);
6111a3b2ec9SScott Feldman }
6122eb03e6cSOr Gerlitz EXPORT_SYMBOL_GPL(switchdev_port_same_parent_id);
613f30f0601SPetr Machata 
614f30f0601SPetr Machata static int __switchdev_handle_port_obj_add(struct net_device *dev,
615f30f0601SPetr Machata 			struct switchdev_notifier_port_obj_info *port_obj_info,
616f30f0601SPetr Machata 			bool (*check_cb)(const struct net_device *dev),
617f30f0601SPetr Machata 			int (*add_cb)(struct net_device *dev,
618f30f0601SPetr Machata 				      const struct switchdev_obj *obj,
619f30f0601SPetr Machata 				      struct switchdev_trans *trans))
620f30f0601SPetr Machata {
621f30f0601SPetr Machata 	struct net_device *lower_dev;
622f30f0601SPetr Machata 	struct list_head *iter;
623f30f0601SPetr Machata 	int err = -EOPNOTSUPP;
624f30f0601SPetr Machata 
625f30f0601SPetr Machata 	if (check_cb(dev)) {
626f30f0601SPetr Machata 		/* This flag is only checked if the return value is success. */
627f30f0601SPetr Machata 		port_obj_info->handled = true;
628f30f0601SPetr Machata 		return add_cb(dev, port_obj_info->obj, port_obj_info->trans);
629f30f0601SPetr Machata 	}
630f30f0601SPetr Machata 
631f30f0601SPetr Machata 	/* Switch ports might be stacked under e.g. a LAG. Ignore the
632f30f0601SPetr Machata 	 * unsupported devices, another driver might be able to handle them. But
633f30f0601SPetr Machata 	 * propagate to the callers any hard errors.
634f30f0601SPetr Machata 	 *
635f30f0601SPetr Machata 	 * If the driver does its own bookkeeping of stacked ports, it's not
636f30f0601SPetr Machata 	 * necessary to go through this helper.
637f30f0601SPetr Machata 	 */
638f30f0601SPetr Machata 	netdev_for_each_lower_dev(dev, lower_dev, iter) {
639f30f0601SPetr Machata 		err = __switchdev_handle_port_obj_add(lower_dev, port_obj_info,
640f30f0601SPetr Machata 						      check_cb, add_cb);
641f30f0601SPetr Machata 		if (err && err != -EOPNOTSUPP)
642f30f0601SPetr Machata 			return err;
643f30f0601SPetr Machata 	}
644f30f0601SPetr Machata 
645f30f0601SPetr Machata 	return err;
646f30f0601SPetr Machata }
647f30f0601SPetr Machata 
648f30f0601SPetr Machata int switchdev_handle_port_obj_add(struct net_device *dev,
649f30f0601SPetr Machata 			struct switchdev_notifier_port_obj_info *port_obj_info,
650f30f0601SPetr Machata 			bool (*check_cb)(const struct net_device *dev),
651f30f0601SPetr Machata 			int (*add_cb)(struct net_device *dev,
652f30f0601SPetr Machata 				      const struct switchdev_obj *obj,
653f30f0601SPetr Machata 				      struct switchdev_trans *trans))
654f30f0601SPetr Machata {
655f30f0601SPetr Machata 	int err;
656f30f0601SPetr Machata 
657f30f0601SPetr Machata 	err = __switchdev_handle_port_obj_add(dev, port_obj_info, check_cb,
658f30f0601SPetr Machata 					      add_cb);
659f30f0601SPetr Machata 	if (err == -EOPNOTSUPP)
660f30f0601SPetr Machata 		err = 0;
661f30f0601SPetr Machata 	return err;
662f30f0601SPetr Machata }
663f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_add);
664f30f0601SPetr Machata 
665f30f0601SPetr Machata static int __switchdev_handle_port_obj_del(struct net_device *dev,
666f30f0601SPetr Machata 			struct switchdev_notifier_port_obj_info *port_obj_info,
667f30f0601SPetr Machata 			bool (*check_cb)(const struct net_device *dev),
668f30f0601SPetr Machata 			int (*del_cb)(struct net_device *dev,
669f30f0601SPetr Machata 				      const struct switchdev_obj *obj))
670f30f0601SPetr Machata {
671f30f0601SPetr Machata 	struct net_device *lower_dev;
672f30f0601SPetr Machata 	struct list_head *iter;
673f30f0601SPetr Machata 	int err = -EOPNOTSUPP;
674f30f0601SPetr Machata 
675f30f0601SPetr Machata 	if (check_cb(dev)) {
676f30f0601SPetr Machata 		/* This flag is only checked if the return value is success. */
677f30f0601SPetr Machata 		port_obj_info->handled = true;
678f30f0601SPetr Machata 		return del_cb(dev, port_obj_info->obj);
679f30f0601SPetr Machata 	}
680f30f0601SPetr Machata 
681f30f0601SPetr Machata 	/* Switch ports might be stacked under e.g. a LAG. Ignore the
682f30f0601SPetr Machata 	 * unsupported devices, another driver might be able to handle them. But
683f30f0601SPetr Machata 	 * propagate to the callers any hard errors.
684f30f0601SPetr Machata 	 *
685f30f0601SPetr Machata 	 * If the driver does its own bookkeeping of stacked ports, it's not
686f30f0601SPetr Machata 	 * necessary to go through this helper.
687f30f0601SPetr Machata 	 */
688f30f0601SPetr Machata 	netdev_for_each_lower_dev(dev, lower_dev, iter) {
689f30f0601SPetr Machata 		err = __switchdev_handle_port_obj_del(lower_dev, port_obj_info,
690f30f0601SPetr Machata 						      check_cb, del_cb);
691f30f0601SPetr Machata 		if (err && err != -EOPNOTSUPP)
692f30f0601SPetr Machata 			return err;
693f30f0601SPetr Machata 	}
694f30f0601SPetr Machata 
695f30f0601SPetr Machata 	return err;
696f30f0601SPetr Machata }
697f30f0601SPetr Machata 
698f30f0601SPetr Machata int switchdev_handle_port_obj_del(struct net_device *dev,
699f30f0601SPetr Machata 			struct switchdev_notifier_port_obj_info *port_obj_info,
700f30f0601SPetr Machata 			bool (*check_cb)(const struct net_device *dev),
701f30f0601SPetr Machata 			int (*del_cb)(struct net_device *dev,
702f30f0601SPetr Machata 				      const struct switchdev_obj *obj))
703f30f0601SPetr Machata {
704f30f0601SPetr Machata 	int err;
705f30f0601SPetr Machata 
706f30f0601SPetr Machata 	err = __switchdev_handle_port_obj_del(dev, port_obj_info, check_cb,
707f30f0601SPetr Machata 					      del_cb);
708f30f0601SPetr Machata 	if (err == -EOPNOTSUPP)
709f30f0601SPetr Machata 		err = 0;
710f30f0601SPetr Machata 	return err;
711f30f0601SPetr Machata }
712f30f0601SPetr Machata EXPORT_SYMBOL_GPL(switchdev_handle_port_obj_del);
713