xref: /f-stack/dpdk/kernel/linux/kni/kni_misc.c (revision 4418919f)
1d30ea906Sjfb8856606 // SPDX-License-Identifier: GPL-2.0
2d30ea906Sjfb8856606 /*
3d30ea906Sjfb8856606  * Copyright(c) 2010-2014 Intel Corporation.
4d30ea906Sjfb8856606  */
5d30ea906Sjfb8856606 
6d30ea906Sjfb8856606 #include <linux/version.h>
7d30ea906Sjfb8856606 #include <linux/module.h>
8d30ea906Sjfb8856606 #include <linux/miscdevice.h>
9d30ea906Sjfb8856606 #include <linux/netdevice.h>
10d30ea906Sjfb8856606 #include <linux/etherdevice.h>
11d30ea906Sjfb8856606 #include <linux/pci.h>
12d30ea906Sjfb8856606 #include <linux/kthread.h>
13d30ea906Sjfb8856606 #include <linux/rwsem.h>
14d30ea906Sjfb8856606 #include <linux/mutex.h>
15d30ea906Sjfb8856606 #include <linux/nsproxy.h>
16d30ea906Sjfb8856606 #include <net/net_namespace.h>
17d30ea906Sjfb8856606 #include <net/netns/generic.h>
18d30ea906Sjfb8856606 
19*4418919fSjohnjiang #include <rte_kni_common.h>
20d30ea906Sjfb8856606 
21d30ea906Sjfb8856606 #include "compat.h"
22d30ea906Sjfb8856606 #include "kni_dev.h"
23d30ea906Sjfb8856606 
24*4418919fSjohnjiang MODULE_VERSION(KNI_VERSION);
25d30ea906Sjfb8856606 MODULE_LICENSE("Dual BSD/GPL");
26d30ea906Sjfb8856606 MODULE_AUTHOR("Intel Corporation");
27d30ea906Sjfb8856606 MODULE_DESCRIPTION("Kernel Module for managing kni devices");
28d30ea906Sjfb8856606 
29d30ea906Sjfb8856606 #define KNI_RX_LOOP_NUM 1000
30d30ea906Sjfb8856606 
31d30ea906Sjfb8856606 #define KNI_MAX_DEVICES 32
32d30ea906Sjfb8856606 
33d30ea906Sjfb8856606 /* loopback mode */
34d30ea906Sjfb8856606 static char *lo_mode;
35d30ea906Sjfb8856606 
36d30ea906Sjfb8856606 /* Kernel thread mode */
37d30ea906Sjfb8856606 static char *kthread_mode;
38d30ea906Sjfb8856606 static uint32_t multiple_kthread_on;
39d30ea906Sjfb8856606 
40d30ea906Sjfb8856606 /* Default carrier state for created KNI network interfaces */
41d30ea906Sjfb8856606 static char *carrier;
42*4418919fSjohnjiang uint32_t kni_dflt_carrier;
43d30ea906Sjfb8856606 
44d30ea906Sjfb8856606 #define KNI_DEV_IN_USE_BIT_NUM 0 /* Bit number for device in use */
45d30ea906Sjfb8856606 
46d30ea906Sjfb8856606 static int kni_net_id;
47d30ea906Sjfb8856606 
48d30ea906Sjfb8856606 struct kni_net {
49d30ea906Sjfb8856606 	unsigned long device_in_use; /* device in use flag */
50d30ea906Sjfb8856606 	struct mutex kni_kthread_lock;
51d30ea906Sjfb8856606 	struct task_struct *kni_kthread;
52d30ea906Sjfb8856606 	struct rw_semaphore kni_list_lock;
53d30ea906Sjfb8856606 	struct list_head kni_list_head;
54d30ea906Sjfb8856606 };
55d30ea906Sjfb8856606 
56d30ea906Sjfb8856606 static int __net_init
kni_init_net(struct net * net)57d30ea906Sjfb8856606 kni_init_net(struct net *net)
58d30ea906Sjfb8856606 {
59d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
60d30ea906Sjfb8856606 	struct kni_net *knet = net_generic(net, kni_net_id);
61d30ea906Sjfb8856606 
62d30ea906Sjfb8856606 	memset(knet, 0, sizeof(*knet));
63d30ea906Sjfb8856606 #else
64d30ea906Sjfb8856606 	struct kni_net *knet;
65d30ea906Sjfb8856606 	int ret;
66d30ea906Sjfb8856606 
67d30ea906Sjfb8856606 	knet = kzalloc(sizeof(struct kni_net), GFP_KERNEL);
68d30ea906Sjfb8856606 	if (!knet) {
69d30ea906Sjfb8856606 		ret = -ENOMEM;
70d30ea906Sjfb8856606 		return ret;
71d30ea906Sjfb8856606 	}
72d30ea906Sjfb8856606 #endif
73d30ea906Sjfb8856606 
74d30ea906Sjfb8856606 	/* Clear the bit of device in use */
75d30ea906Sjfb8856606 	clear_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use);
76d30ea906Sjfb8856606 
77d30ea906Sjfb8856606 	mutex_init(&knet->kni_kthread_lock);
78d30ea906Sjfb8856606 
79d30ea906Sjfb8856606 	init_rwsem(&knet->kni_list_lock);
80d30ea906Sjfb8856606 	INIT_LIST_HEAD(&knet->kni_list_head);
81d30ea906Sjfb8856606 
82d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
83d30ea906Sjfb8856606 	return 0;
84d30ea906Sjfb8856606 #else
85d30ea906Sjfb8856606 	ret = net_assign_generic(net, kni_net_id, knet);
86d30ea906Sjfb8856606 	if (ret < 0)
87d30ea906Sjfb8856606 		kfree(knet);
88d30ea906Sjfb8856606 
89d30ea906Sjfb8856606 	return ret;
90d30ea906Sjfb8856606 #endif
91d30ea906Sjfb8856606 }
92d30ea906Sjfb8856606 
93d30ea906Sjfb8856606 static void __net_exit
kni_exit_net(struct net * net)94d30ea906Sjfb8856606 kni_exit_net(struct net *net)
95d30ea906Sjfb8856606 {
96d30ea906Sjfb8856606 	struct kni_net *knet __maybe_unused;
97d30ea906Sjfb8856606 
98d30ea906Sjfb8856606 	knet = net_generic(net, kni_net_id);
99d30ea906Sjfb8856606 	mutex_destroy(&knet->kni_kthread_lock);
100d30ea906Sjfb8856606 
101d30ea906Sjfb8856606 #ifndef HAVE_SIMPLIFIED_PERNET_OPERATIONS
102d30ea906Sjfb8856606 	kfree(knet);
103d30ea906Sjfb8856606 #endif
104d30ea906Sjfb8856606 }
105d30ea906Sjfb8856606 
106d30ea906Sjfb8856606 static struct pernet_operations kni_net_ops = {
107d30ea906Sjfb8856606 	.init = kni_init_net,
108d30ea906Sjfb8856606 	.exit = kni_exit_net,
109d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
110d30ea906Sjfb8856606 	.id   = &kni_net_id,
111d30ea906Sjfb8856606 	.size = sizeof(struct kni_net),
112d30ea906Sjfb8856606 #endif
113d30ea906Sjfb8856606 };
114d30ea906Sjfb8856606 
115d30ea906Sjfb8856606 static int
kni_thread_single(void * data)116d30ea906Sjfb8856606 kni_thread_single(void *data)
117d30ea906Sjfb8856606 {
118d30ea906Sjfb8856606 	struct kni_net *knet = data;
119d30ea906Sjfb8856606 	int j;
120d30ea906Sjfb8856606 	struct kni_dev *dev;
121d30ea906Sjfb8856606 
122d30ea906Sjfb8856606 	while (!kthread_should_stop()) {
123d30ea906Sjfb8856606 		down_read(&knet->kni_list_lock);
124d30ea906Sjfb8856606 		for (j = 0; j < KNI_RX_LOOP_NUM; j++) {
125d30ea906Sjfb8856606 			list_for_each_entry(dev, &knet->kni_list_head, list) {
126d30ea906Sjfb8856606 				kni_net_rx(dev);
127d30ea906Sjfb8856606 				kni_net_poll_resp(dev);
128d30ea906Sjfb8856606 			}
129d30ea906Sjfb8856606 		}
130d30ea906Sjfb8856606 		up_read(&knet->kni_list_lock);
131d30ea906Sjfb8856606 #ifdef RTE_KNI_PREEMPT_DEFAULT
132d30ea906Sjfb8856606 		/* reschedule out for a while */
133d30ea906Sjfb8856606 		schedule_timeout_interruptible(
134d30ea906Sjfb8856606 			usecs_to_jiffies(KNI_KTHREAD_RESCHEDULE_INTERVAL));
135d30ea906Sjfb8856606 #endif
136d30ea906Sjfb8856606 	}
137d30ea906Sjfb8856606 
138d30ea906Sjfb8856606 	return 0;
139d30ea906Sjfb8856606 }
140d30ea906Sjfb8856606 
141d30ea906Sjfb8856606 static int
kni_thread_multiple(void * param)142d30ea906Sjfb8856606 kni_thread_multiple(void *param)
143d30ea906Sjfb8856606 {
144d30ea906Sjfb8856606 	int j;
145d30ea906Sjfb8856606 	struct kni_dev *dev = param;
146d30ea906Sjfb8856606 
147d30ea906Sjfb8856606 	while (!kthread_should_stop()) {
148d30ea906Sjfb8856606 		for (j = 0; j < KNI_RX_LOOP_NUM; j++) {
149d30ea906Sjfb8856606 			kni_net_rx(dev);
150d30ea906Sjfb8856606 			kni_net_poll_resp(dev);
151d30ea906Sjfb8856606 		}
152d30ea906Sjfb8856606 #ifdef RTE_KNI_PREEMPT_DEFAULT
153d30ea906Sjfb8856606 		schedule_timeout_interruptible(
154d30ea906Sjfb8856606 			usecs_to_jiffies(KNI_KTHREAD_RESCHEDULE_INTERVAL));
155d30ea906Sjfb8856606 #endif
156d30ea906Sjfb8856606 	}
157d30ea906Sjfb8856606 
158d30ea906Sjfb8856606 	return 0;
159d30ea906Sjfb8856606 }
160d30ea906Sjfb8856606 
161d30ea906Sjfb8856606 static int
kni_open(struct inode * inode,struct file * file)162d30ea906Sjfb8856606 kni_open(struct inode *inode, struct file *file)
163d30ea906Sjfb8856606 {
164d30ea906Sjfb8856606 	struct net *net = current->nsproxy->net_ns;
165d30ea906Sjfb8856606 	struct kni_net *knet = net_generic(net, kni_net_id);
166d30ea906Sjfb8856606 
167d30ea906Sjfb8856606 	/* kni device can be opened by one user only per netns */
168d30ea906Sjfb8856606 	if (test_and_set_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use))
169d30ea906Sjfb8856606 		return -EBUSY;
170d30ea906Sjfb8856606 
171d30ea906Sjfb8856606 	file->private_data = get_net(net);
172d30ea906Sjfb8856606 	pr_debug("/dev/kni opened\n");
173d30ea906Sjfb8856606 
174d30ea906Sjfb8856606 	return 0;
175d30ea906Sjfb8856606 }
176d30ea906Sjfb8856606 
177d30ea906Sjfb8856606 static int
kni_dev_remove(struct kni_dev * dev)178d30ea906Sjfb8856606 kni_dev_remove(struct kni_dev *dev)
179d30ea906Sjfb8856606 {
180d30ea906Sjfb8856606 	if (!dev)
181d30ea906Sjfb8856606 		return -ENODEV;
182d30ea906Sjfb8856606 
183d30ea906Sjfb8856606 	if (dev->net_dev) {
184d30ea906Sjfb8856606 		unregister_netdev(dev->net_dev);
185d30ea906Sjfb8856606 		free_netdev(dev->net_dev);
186d30ea906Sjfb8856606 	}
187d30ea906Sjfb8856606 
188d30ea906Sjfb8856606 	kni_net_release_fifo_phy(dev);
189d30ea906Sjfb8856606 
190d30ea906Sjfb8856606 	return 0;
191d30ea906Sjfb8856606 }
192d30ea906Sjfb8856606 
193d30ea906Sjfb8856606 static int
kni_release(struct inode * inode,struct file * file)194d30ea906Sjfb8856606 kni_release(struct inode *inode, struct file *file)
195d30ea906Sjfb8856606 {
196d30ea906Sjfb8856606 	struct net *net = file->private_data;
197d30ea906Sjfb8856606 	struct kni_net *knet = net_generic(net, kni_net_id);
198d30ea906Sjfb8856606 	struct kni_dev *dev, *n;
199d30ea906Sjfb8856606 
200d30ea906Sjfb8856606 	/* Stop kernel thread for single mode */
201d30ea906Sjfb8856606 	if (multiple_kthread_on == 0) {
202d30ea906Sjfb8856606 		mutex_lock(&knet->kni_kthread_lock);
203d30ea906Sjfb8856606 		/* Stop kernel thread */
204d30ea906Sjfb8856606 		if (knet->kni_kthread != NULL) {
205d30ea906Sjfb8856606 			kthread_stop(knet->kni_kthread);
206d30ea906Sjfb8856606 			knet->kni_kthread = NULL;
207d30ea906Sjfb8856606 		}
208d30ea906Sjfb8856606 		mutex_unlock(&knet->kni_kthread_lock);
209d30ea906Sjfb8856606 	}
210d30ea906Sjfb8856606 
211d30ea906Sjfb8856606 	down_write(&knet->kni_list_lock);
212d30ea906Sjfb8856606 	list_for_each_entry_safe(dev, n, &knet->kni_list_head, list) {
213d30ea906Sjfb8856606 		/* Stop kernel thread for multiple mode */
214d30ea906Sjfb8856606 		if (multiple_kthread_on && dev->pthread != NULL) {
215d30ea906Sjfb8856606 			kthread_stop(dev->pthread);
216d30ea906Sjfb8856606 			dev->pthread = NULL;
217d30ea906Sjfb8856606 		}
218d30ea906Sjfb8856606 
219d30ea906Sjfb8856606 		kni_dev_remove(dev);
220d30ea906Sjfb8856606 		list_del(&dev->list);
221d30ea906Sjfb8856606 	}
222d30ea906Sjfb8856606 	up_write(&knet->kni_list_lock);
223d30ea906Sjfb8856606 
224d30ea906Sjfb8856606 	/* Clear the bit of device in use */
225d30ea906Sjfb8856606 	clear_bit(KNI_DEV_IN_USE_BIT_NUM, &knet->device_in_use);
226d30ea906Sjfb8856606 
227d30ea906Sjfb8856606 	put_net(net);
228d30ea906Sjfb8856606 	pr_debug("/dev/kni closed\n");
229d30ea906Sjfb8856606 
230d30ea906Sjfb8856606 	return 0;
231d30ea906Sjfb8856606 }
232d30ea906Sjfb8856606 
233d30ea906Sjfb8856606 static int
kni_check_param(struct kni_dev * kni,struct rte_kni_device_info * dev)234d30ea906Sjfb8856606 kni_check_param(struct kni_dev *kni, struct rte_kni_device_info *dev)
235d30ea906Sjfb8856606 {
236d30ea906Sjfb8856606 	if (!kni || !dev)
237d30ea906Sjfb8856606 		return -1;
238d30ea906Sjfb8856606 
239d30ea906Sjfb8856606 	/* Check if network name has been used */
240d30ea906Sjfb8856606 	if (!strncmp(kni->name, dev->name, RTE_KNI_NAMESIZE)) {
241d30ea906Sjfb8856606 		pr_err("KNI name %s duplicated\n", dev->name);
242d30ea906Sjfb8856606 		return -1;
243d30ea906Sjfb8856606 	}
244d30ea906Sjfb8856606 
245d30ea906Sjfb8856606 	return 0;
246d30ea906Sjfb8856606 }
247d30ea906Sjfb8856606 
248d30ea906Sjfb8856606 static int
kni_run_thread(struct kni_net * knet,struct kni_dev * kni,uint8_t force_bind)249d30ea906Sjfb8856606 kni_run_thread(struct kni_net *knet, struct kni_dev *kni, uint8_t force_bind)
250d30ea906Sjfb8856606 {
251d30ea906Sjfb8856606 	/**
252d30ea906Sjfb8856606 	 * Create a new kernel thread for multiple mode, set its core affinity,
253d30ea906Sjfb8856606 	 * and finally wake it up.
254d30ea906Sjfb8856606 	 */
255d30ea906Sjfb8856606 	if (multiple_kthread_on) {
256d30ea906Sjfb8856606 		kni->pthread = kthread_create(kni_thread_multiple,
257d30ea906Sjfb8856606 			(void *)kni, "kni_%s", kni->name);
258d30ea906Sjfb8856606 		if (IS_ERR(kni->pthread)) {
259d30ea906Sjfb8856606 			kni_dev_remove(kni);
260d30ea906Sjfb8856606 			return -ECANCELED;
261d30ea906Sjfb8856606 		}
262d30ea906Sjfb8856606 
263d30ea906Sjfb8856606 		if (force_bind)
264d30ea906Sjfb8856606 			kthread_bind(kni->pthread, kni->core_id);
265d30ea906Sjfb8856606 		wake_up_process(kni->pthread);
266d30ea906Sjfb8856606 	} else {
267d30ea906Sjfb8856606 		mutex_lock(&knet->kni_kthread_lock);
268d30ea906Sjfb8856606 
269d30ea906Sjfb8856606 		if (knet->kni_kthread == NULL) {
270d30ea906Sjfb8856606 			knet->kni_kthread = kthread_create(kni_thread_single,
271d30ea906Sjfb8856606 				(void *)knet, "kni_single");
272d30ea906Sjfb8856606 			if (IS_ERR(knet->kni_kthread)) {
273d30ea906Sjfb8856606 				mutex_unlock(&knet->kni_kthread_lock);
274d30ea906Sjfb8856606 				kni_dev_remove(kni);
275d30ea906Sjfb8856606 				return -ECANCELED;
276d30ea906Sjfb8856606 			}
277d30ea906Sjfb8856606 
278d30ea906Sjfb8856606 			if (force_bind)
279d30ea906Sjfb8856606 				kthread_bind(knet->kni_kthread, kni->core_id);
280d30ea906Sjfb8856606 			wake_up_process(knet->kni_kthread);
281d30ea906Sjfb8856606 		}
282d30ea906Sjfb8856606 
283d30ea906Sjfb8856606 		mutex_unlock(&knet->kni_kthread_lock);
284d30ea906Sjfb8856606 	}
285d30ea906Sjfb8856606 
286d30ea906Sjfb8856606 	return 0;
287d30ea906Sjfb8856606 }
288d30ea906Sjfb8856606 
289d30ea906Sjfb8856606 static int
kni_ioctl_create(struct net * net,uint32_t ioctl_num,unsigned long ioctl_param)290d30ea906Sjfb8856606 kni_ioctl_create(struct net *net, uint32_t ioctl_num,
291d30ea906Sjfb8856606 		unsigned long ioctl_param)
292d30ea906Sjfb8856606 {
293d30ea906Sjfb8856606 	struct kni_net *knet = net_generic(net, kni_net_id);
294d30ea906Sjfb8856606 	int ret;
295d30ea906Sjfb8856606 	struct rte_kni_device_info dev_info;
296d30ea906Sjfb8856606 	struct net_device *net_dev = NULL;
297d30ea906Sjfb8856606 	struct kni_dev *kni, *dev, *n;
298d30ea906Sjfb8856606 
299d30ea906Sjfb8856606 	pr_info("Creating kni...\n");
300d30ea906Sjfb8856606 	/* Check the buffer size, to avoid warning */
301d30ea906Sjfb8856606 	if (_IOC_SIZE(ioctl_num) > sizeof(dev_info))
302d30ea906Sjfb8856606 		return -EINVAL;
303d30ea906Sjfb8856606 
304d30ea906Sjfb8856606 	/* Copy kni info from user space */
3054b05018fSfengbojiang 	if (copy_from_user(&dev_info, (void *)ioctl_param, sizeof(dev_info)))
3064b05018fSfengbojiang 		return -EFAULT;
307d30ea906Sjfb8856606 
308d30ea906Sjfb8856606 	/* Check if name is zero-ended */
309d30ea906Sjfb8856606 	if (strnlen(dev_info.name, sizeof(dev_info.name)) == sizeof(dev_info.name)) {
310d30ea906Sjfb8856606 		pr_err("kni.name not zero-terminated");
311d30ea906Sjfb8856606 		return -EINVAL;
312d30ea906Sjfb8856606 	}
313d30ea906Sjfb8856606 
314d30ea906Sjfb8856606 	/**
315d30ea906Sjfb8856606 	 * Check if the cpu core id is valid for binding.
316d30ea906Sjfb8856606 	 */
317d30ea906Sjfb8856606 	if (dev_info.force_bind && !cpu_online(dev_info.core_id)) {
318d30ea906Sjfb8856606 		pr_err("cpu %u is not online\n", dev_info.core_id);
319d30ea906Sjfb8856606 		return -EINVAL;
320d30ea906Sjfb8856606 	}
321d30ea906Sjfb8856606 
322d30ea906Sjfb8856606 	/* Check if it has been created */
323d30ea906Sjfb8856606 	down_read(&knet->kni_list_lock);
324d30ea906Sjfb8856606 	list_for_each_entry_safe(dev, n, &knet->kni_list_head, list) {
325d30ea906Sjfb8856606 		if (kni_check_param(dev, &dev_info) < 0) {
326d30ea906Sjfb8856606 			up_read(&knet->kni_list_lock);
327d30ea906Sjfb8856606 			return -EINVAL;
328d30ea906Sjfb8856606 		}
329d30ea906Sjfb8856606 	}
330d30ea906Sjfb8856606 	up_read(&knet->kni_list_lock);
331d30ea906Sjfb8856606 
332d30ea906Sjfb8856606 	net_dev = alloc_netdev(sizeof(struct kni_dev), dev_info.name,
333d30ea906Sjfb8856606 #ifdef NET_NAME_USER
334d30ea906Sjfb8856606 							NET_NAME_USER,
335d30ea906Sjfb8856606 #endif
336d30ea906Sjfb8856606 							kni_net_init);
337d30ea906Sjfb8856606 	if (net_dev == NULL) {
338d30ea906Sjfb8856606 		pr_err("error allocating device \"%s\"\n", dev_info.name);
339d30ea906Sjfb8856606 		return -EBUSY;
340d30ea906Sjfb8856606 	}
341d30ea906Sjfb8856606 
342d30ea906Sjfb8856606 	dev_net_set(net_dev, net);
343d30ea906Sjfb8856606 
344d30ea906Sjfb8856606 	kni = netdev_priv(net_dev);
345d30ea906Sjfb8856606 
346d30ea906Sjfb8856606 	kni->net_dev = net_dev;
347d30ea906Sjfb8856606 	kni->core_id = dev_info.core_id;
348d30ea906Sjfb8856606 	strncpy(kni->name, dev_info.name, RTE_KNI_NAMESIZE);
349d30ea906Sjfb8856606 
350d30ea906Sjfb8856606 	/* Translate user space info into kernel space info */
351*4418919fSjohnjiang 	if (dev_info.iova_mode) {
352*4418919fSjohnjiang #ifdef HAVE_IOVA_TO_KVA_MAPPING_SUPPORT
353*4418919fSjohnjiang 		kni->tx_q = iova_to_kva(current, dev_info.tx_phys);
354*4418919fSjohnjiang 		kni->rx_q = iova_to_kva(current, dev_info.rx_phys);
355*4418919fSjohnjiang 		kni->alloc_q = iova_to_kva(current, dev_info.alloc_phys);
356*4418919fSjohnjiang 		kni->free_q = iova_to_kva(current, dev_info.free_phys);
357*4418919fSjohnjiang 
358*4418919fSjohnjiang 		kni->req_q = iova_to_kva(current, dev_info.req_phys);
359*4418919fSjohnjiang 		kni->resp_q = iova_to_kva(current, dev_info.resp_phys);
360*4418919fSjohnjiang 		kni->sync_va = dev_info.sync_va;
361*4418919fSjohnjiang 		kni->sync_kva = iova_to_kva(current, dev_info.sync_phys);
362*4418919fSjohnjiang 		kni->usr_tsk = current;
363*4418919fSjohnjiang 		kni->iova_mode = 1;
364*4418919fSjohnjiang #else
365*4418919fSjohnjiang 		pr_err("KNI module does not support IOVA to VA translation\n");
366*4418919fSjohnjiang 		return -EINVAL;
367*4418919fSjohnjiang #endif
368*4418919fSjohnjiang 	} else {
369*4418919fSjohnjiang 
370d30ea906Sjfb8856606 		kni->tx_q = phys_to_virt(dev_info.tx_phys);
371d30ea906Sjfb8856606 		kni->rx_q = phys_to_virt(dev_info.rx_phys);
372d30ea906Sjfb8856606 		kni->alloc_q = phys_to_virt(dev_info.alloc_phys);
373d30ea906Sjfb8856606 		kni->free_q = phys_to_virt(dev_info.free_phys);
374d30ea906Sjfb8856606 
375d30ea906Sjfb8856606 		kni->req_q = phys_to_virt(dev_info.req_phys);
376d30ea906Sjfb8856606 		kni->resp_q = phys_to_virt(dev_info.resp_phys);
377d30ea906Sjfb8856606 		kni->sync_va = dev_info.sync_va;
378d30ea906Sjfb8856606 		kni->sync_kva = phys_to_virt(dev_info.sync_phys);
379*4418919fSjohnjiang 		kni->iova_mode = 0;
380*4418919fSjohnjiang 	}
381d30ea906Sjfb8856606 
382d30ea906Sjfb8856606 	kni->mbuf_size = dev_info.mbuf_size;
383d30ea906Sjfb8856606 
384d30ea906Sjfb8856606 	pr_debug("tx_phys:      0x%016llx, tx_q addr:      0x%p\n",
385d30ea906Sjfb8856606 		(unsigned long long) dev_info.tx_phys, kni->tx_q);
386d30ea906Sjfb8856606 	pr_debug("rx_phys:      0x%016llx, rx_q addr:      0x%p\n",
387d30ea906Sjfb8856606 		(unsigned long long) dev_info.rx_phys, kni->rx_q);
388d30ea906Sjfb8856606 	pr_debug("alloc_phys:   0x%016llx, alloc_q addr:   0x%p\n",
389d30ea906Sjfb8856606 		(unsigned long long) dev_info.alloc_phys, kni->alloc_q);
390d30ea906Sjfb8856606 	pr_debug("free_phys:    0x%016llx, free_q addr:    0x%p\n",
391d30ea906Sjfb8856606 		(unsigned long long) dev_info.free_phys, kni->free_q);
392d30ea906Sjfb8856606 	pr_debug("req_phys:     0x%016llx, req_q addr:     0x%p\n",
393d30ea906Sjfb8856606 		(unsigned long long) dev_info.req_phys, kni->req_q);
394d30ea906Sjfb8856606 	pr_debug("resp_phys:    0x%016llx, resp_q addr:    0x%p\n",
395d30ea906Sjfb8856606 		(unsigned long long) dev_info.resp_phys, kni->resp_q);
396d30ea906Sjfb8856606 	pr_debug("mbuf_size:    %u\n", kni->mbuf_size);
397d30ea906Sjfb8856606 
398d30ea906Sjfb8856606 	/* if user has provided a valid mac address */
3991646932aSjfb8856606 	if (is_valid_ether_addr(dev_info.mac_addr))
400d30ea906Sjfb8856606 		memcpy(net_dev->dev_addr, dev_info.mac_addr, ETH_ALEN);
401d30ea906Sjfb8856606 	else
402d30ea906Sjfb8856606 		/*
403d30ea906Sjfb8856606 		 * Generate random mac address. eth_random_addr() is the
404d30ea906Sjfb8856606 		 * newer version of generating mac address in kernel.
405d30ea906Sjfb8856606 		 */
406d30ea906Sjfb8856606 		random_ether_addr(net_dev->dev_addr);
407d30ea906Sjfb8856606 
408d30ea906Sjfb8856606 	if (dev_info.mtu)
409d30ea906Sjfb8856606 		net_dev->mtu = dev_info.mtu;
410*4418919fSjohnjiang #ifdef HAVE_MAX_MTU_PARAM
411*4418919fSjohnjiang 	net_dev->max_mtu = net_dev->mtu;
412*4418919fSjohnjiang 
413*4418919fSjohnjiang 	if (dev_info.min_mtu)
414*4418919fSjohnjiang 		net_dev->min_mtu = dev_info.min_mtu;
415*4418919fSjohnjiang 
416*4418919fSjohnjiang 	if (dev_info.max_mtu)
417*4418919fSjohnjiang 		net_dev->max_mtu = dev_info.max_mtu;
418*4418919fSjohnjiang #endif
419d30ea906Sjfb8856606 
420d30ea906Sjfb8856606 	ret = register_netdev(net_dev);
421d30ea906Sjfb8856606 	if (ret) {
422d30ea906Sjfb8856606 		pr_err("error %i registering device \"%s\"\n",
423d30ea906Sjfb8856606 					ret, dev_info.name);
424d30ea906Sjfb8856606 		kni->net_dev = NULL;
425d30ea906Sjfb8856606 		kni_dev_remove(kni);
426d30ea906Sjfb8856606 		free_netdev(net_dev);
427d30ea906Sjfb8856606 		return -ENODEV;
428d30ea906Sjfb8856606 	}
429d30ea906Sjfb8856606 
430d30ea906Sjfb8856606 	netif_carrier_off(net_dev);
431d30ea906Sjfb8856606 
432d30ea906Sjfb8856606 	ret = kni_run_thread(knet, kni, dev_info.force_bind);
433d30ea906Sjfb8856606 	if (ret != 0)
434d30ea906Sjfb8856606 		return ret;
435d30ea906Sjfb8856606 
436d30ea906Sjfb8856606 	down_write(&knet->kni_list_lock);
437d30ea906Sjfb8856606 	list_add(&kni->list, &knet->kni_list_head);
438d30ea906Sjfb8856606 	up_write(&knet->kni_list_lock);
439d30ea906Sjfb8856606 
440d30ea906Sjfb8856606 	return 0;
441d30ea906Sjfb8856606 }
442d30ea906Sjfb8856606 
443d30ea906Sjfb8856606 static int
kni_ioctl_release(struct net * net,uint32_t ioctl_num,unsigned long ioctl_param)444d30ea906Sjfb8856606 kni_ioctl_release(struct net *net, uint32_t ioctl_num,
445d30ea906Sjfb8856606 		unsigned long ioctl_param)
446d30ea906Sjfb8856606 {
447d30ea906Sjfb8856606 	struct kni_net *knet = net_generic(net, kni_net_id);
448d30ea906Sjfb8856606 	int ret = -EINVAL;
449d30ea906Sjfb8856606 	struct kni_dev *dev, *n;
450d30ea906Sjfb8856606 	struct rte_kni_device_info dev_info;
451d30ea906Sjfb8856606 
452d30ea906Sjfb8856606 	if (_IOC_SIZE(ioctl_num) > sizeof(dev_info))
453d30ea906Sjfb8856606 		return -EINVAL;
454d30ea906Sjfb8856606 
4554b05018fSfengbojiang 	if (copy_from_user(&dev_info, (void *)ioctl_param, sizeof(dev_info)))
4564b05018fSfengbojiang 		return -EFAULT;
457d30ea906Sjfb8856606 
458d30ea906Sjfb8856606 	/* Release the network device according to its name */
459d30ea906Sjfb8856606 	if (strlen(dev_info.name) == 0)
4604b05018fSfengbojiang 		return -EINVAL;
461d30ea906Sjfb8856606 
462d30ea906Sjfb8856606 	down_write(&knet->kni_list_lock);
463d30ea906Sjfb8856606 	list_for_each_entry_safe(dev, n, &knet->kni_list_head, list) {
464d30ea906Sjfb8856606 		if (strncmp(dev->name, dev_info.name, RTE_KNI_NAMESIZE) != 0)
465d30ea906Sjfb8856606 			continue;
466d30ea906Sjfb8856606 
467d30ea906Sjfb8856606 		if (multiple_kthread_on && dev->pthread != NULL) {
468d30ea906Sjfb8856606 			kthread_stop(dev->pthread);
469d30ea906Sjfb8856606 			dev->pthread = NULL;
470d30ea906Sjfb8856606 		}
471d30ea906Sjfb8856606 
472d30ea906Sjfb8856606 		kni_dev_remove(dev);
473d30ea906Sjfb8856606 		list_del(&dev->list);
474d30ea906Sjfb8856606 		ret = 0;
475d30ea906Sjfb8856606 		break;
476d30ea906Sjfb8856606 	}
477d30ea906Sjfb8856606 	up_write(&knet->kni_list_lock);
478d30ea906Sjfb8856606 	pr_info("%s release kni named %s\n",
479d30ea906Sjfb8856606 		(ret == 0 ? "Successfully" : "Unsuccessfully"), dev_info.name);
480d30ea906Sjfb8856606 
481d30ea906Sjfb8856606 	return ret;
482d30ea906Sjfb8856606 }
483d30ea906Sjfb8856606 
484d30ea906Sjfb8856606 static int
kni_ioctl(struct inode * inode,uint32_t ioctl_num,unsigned long ioctl_param)485d30ea906Sjfb8856606 kni_ioctl(struct inode *inode, uint32_t ioctl_num, unsigned long ioctl_param)
486d30ea906Sjfb8856606 {
487d30ea906Sjfb8856606 	int ret = -EINVAL;
488d30ea906Sjfb8856606 	struct net *net = current->nsproxy->net_ns;
489d30ea906Sjfb8856606 
490d30ea906Sjfb8856606 	pr_debug("IOCTL num=0x%0x param=0x%0lx\n", ioctl_num, ioctl_param);
491d30ea906Sjfb8856606 
492d30ea906Sjfb8856606 	/*
493d30ea906Sjfb8856606 	 * Switch according to the ioctl called
494d30ea906Sjfb8856606 	 */
495d30ea906Sjfb8856606 	switch (_IOC_NR(ioctl_num)) {
496d30ea906Sjfb8856606 	case _IOC_NR(RTE_KNI_IOCTL_TEST):
497d30ea906Sjfb8856606 		/* For test only, not used */
498d30ea906Sjfb8856606 		break;
499d30ea906Sjfb8856606 	case _IOC_NR(RTE_KNI_IOCTL_CREATE):
500d30ea906Sjfb8856606 		ret = kni_ioctl_create(net, ioctl_num, ioctl_param);
501d30ea906Sjfb8856606 		break;
502d30ea906Sjfb8856606 	case _IOC_NR(RTE_KNI_IOCTL_RELEASE):
503d30ea906Sjfb8856606 		ret = kni_ioctl_release(net, ioctl_num, ioctl_param);
504d30ea906Sjfb8856606 		break;
505d30ea906Sjfb8856606 	default:
506d30ea906Sjfb8856606 		pr_debug("IOCTL default\n");
507d30ea906Sjfb8856606 		break;
508d30ea906Sjfb8856606 	}
509d30ea906Sjfb8856606 
510d30ea906Sjfb8856606 	return ret;
511d30ea906Sjfb8856606 }
512d30ea906Sjfb8856606 
513d30ea906Sjfb8856606 static int
kni_compat_ioctl(struct inode * inode,uint32_t ioctl_num,unsigned long ioctl_param)514d30ea906Sjfb8856606 kni_compat_ioctl(struct inode *inode, uint32_t ioctl_num,
515d30ea906Sjfb8856606 		unsigned long ioctl_param)
516d30ea906Sjfb8856606 {
517d30ea906Sjfb8856606 	/* 32 bits app on 64 bits OS to be supported later */
518d30ea906Sjfb8856606 	pr_debug("Not implemented.\n");
519d30ea906Sjfb8856606 
520d30ea906Sjfb8856606 	return -EINVAL;
521d30ea906Sjfb8856606 }
522d30ea906Sjfb8856606 
523d30ea906Sjfb8856606 static const struct file_operations kni_fops = {
524d30ea906Sjfb8856606 	.owner = THIS_MODULE,
525d30ea906Sjfb8856606 	.open = kni_open,
526d30ea906Sjfb8856606 	.release = kni_release,
527d30ea906Sjfb8856606 	.unlocked_ioctl = (void *)kni_ioctl,
528d30ea906Sjfb8856606 	.compat_ioctl = (void *)kni_compat_ioctl,
529d30ea906Sjfb8856606 };
530d30ea906Sjfb8856606 
531d30ea906Sjfb8856606 static struct miscdevice kni_misc = {
532d30ea906Sjfb8856606 	.minor = MISC_DYNAMIC_MINOR,
533d30ea906Sjfb8856606 	.name = KNI_DEVICE,
534d30ea906Sjfb8856606 	.fops = &kni_fops,
535d30ea906Sjfb8856606 };
536d30ea906Sjfb8856606 
537d30ea906Sjfb8856606 static int __init
kni_parse_kthread_mode(void)538d30ea906Sjfb8856606 kni_parse_kthread_mode(void)
539d30ea906Sjfb8856606 {
540d30ea906Sjfb8856606 	if (!kthread_mode)
541d30ea906Sjfb8856606 		return 0;
542d30ea906Sjfb8856606 
543d30ea906Sjfb8856606 	if (strcmp(kthread_mode, "single") == 0)
544d30ea906Sjfb8856606 		return 0;
545d30ea906Sjfb8856606 	else if (strcmp(kthread_mode, "multiple") == 0)
546d30ea906Sjfb8856606 		multiple_kthread_on = 1;
547d30ea906Sjfb8856606 	else
548d30ea906Sjfb8856606 		return -1;
549d30ea906Sjfb8856606 
550d30ea906Sjfb8856606 	return 0;
551d30ea906Sjfb8856606 }
552d30ea906Sjfb8856606 
553d30ea906Sjfb8856606 static int __init
kni_parse_carrier_state(void)554d30ea906Sjfb8856606 kni_parse_carrier_state(void)
555d30ea906Sjfb8856606 {
556d30ea906Sjfb8856606 	if (!carrier) {
557*4418919fSjohnjiang 		kni_dflt_carrier = 0;
558d30ea906Sjfb8856606 		return 0;
559d30ea906Sjfb8856606 	}
560d30ea906Sjfb8856606 
561d30ea906Sjfb8856606 	if (strcmp(carrier, "off") == 0)
562*4418919fSjohnjiang 		kni_dflt_carrier = 0;
563d30ea906Sjfb8856606 	else if (strcmp(carrier, "on") == 0)
564*4418919fSjohnjiang 		kni_dflt_carrier = 1;
565d30ea906Sjfb8856606 	else
566d30ea906Sjfb8856606 		return -1;
567d30ea906Sjfb8856606 
568d30ea906Sjfb8856606 	return 0;
569d30ea906Sjfb8856606 }
570d30ea906Sjfb8856606 
571d30ea906Sjfb8856606 static int __init
kni_init(void)572d30ea906Sjfb8856606 kni_init(void)
573d30ea906Sjfb8856606 {
574d30ea906Sjfb8856606 	int rc;
575d30ea906Sjfb8856606 
576d30ea906Sjfb8856606 	if (kni_parse_kthread_mode() < 0) {
577d30ea906Sjfb8856606 		pr_err("Invalid parameter for kthread_mode\n");
578d30ea906Sjfb8856606 		return -EINVAL;
579d30ea906Sjfb8856606 	}
580d30ea906Sjfb8856606 
581d30ea906Sjfb8856606 	if (multiple_kthread_on == 0)
582d30ea906Sjfb8856606 		pr_debug("Single kernel thread for all KNI devices\n");
583d30ea906Sjfb8856606 	else
584d30ea906Sjfb8856606 		pr_debug("Multiple kernel thread mode enabled\n");
585d30ea906Sjfb8856606 
586d30ea906Sjfb8856606 	if (kni_parse_carrier_state() < 0) {
587d30ea906Sjfb8856606 		pr_err("Invalid parameter for carrier\n");
588d30ea906Sjfb8856606 		return -EINVAL;
589d30ea906Sjfb8856606 	}
590d30ea906Sjfb8856606 
591*4418919fSjohnjiang 	if (kni_dflt_carrier == 0)
592d30ea906Sjfb8856606 		pr_debug("Default carrier state set to off.\n");
593d30ea906Sjfb8856606 	else
594d30ea906Sjfb8856606 		pr_debug("Default carrier state set to on.\n");
595d30ea906Sjfb8856606 
596d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
597d30ea906Sjfb8856606 	rc = register_pernet_subsys(&kni_net_ops);
598d30ea906Sjfb8856606 #else
599d30ea906Sjfb8856606 	rc = register_pernet_gen_subsys(&kni_net_id, &kni_net_ops);
600d30ea906Sjfb8856606 #endif
601d30ea906Sjfb8856606 	if (rc)
602d30ea906Sjfb8856606 		return -EPERM;
603d30ea906Sjfb8856606 
604d30ea906Sjfb8856606 	rc = misc_register(&kni_misc);
605d30ea906Sjfb8856606 	if (rc != 0) {
606d30ea906Sjfb8856606 		pr_err("Misc registration failed\n");
607d30ea906Sjfb8856606 		goto out;
608d30ea906Sjfb8856606 	}
609d30ea906Sjfb8856606 
610d30ea906Sjfb8856606 	/* Configure the lo mode according to the input parameter */
611d30ea906Sjfb8856606 	kni_net_config_lo_mode(lo_mode);
612d30ea906Sjfb8856606 
613d30ea906Sjfb8856606 	return 0;
614d30ea906Sjfb8856606 
615d30ea906Sjfb8856606 out:
616d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
617d30ea906Sjfb8856606 	unregister_pernet_subsys(&kni_net_ops);
618d30ea906Sjfb8856606 #else
619d30ea906Sjfb8856606 	unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
620d30ea906Sjfb8856606 #endif
621d30ea906Sjfb8856606 	return rc;
622d30ea906Sjfb8856606 }
623d30ea906Sjfb8856606 
624d30ea906Sjfb8856606 static void __exit
kni_exit(void)625d30ea906Sjfb8856606 kni_exit(void)
626d30ea906Sjfb8856606 {
627d30ea906Sjfb8856606 	misc_deregister(&kni_misc);
628d30ea906Sjfb8856606 #ifdef HAVE_SIMPLIFIED_PERNET_OPERATIONS
629d30ea906Sjfb8856606 	unregister_pernet_subsys(&kni_net_ops);
630d30ea906Sjfb8856606 #else
631d30ea906Sjfb8856606 	unregister_pernet_gen_subsys(kni_net_id, &kni_net_ops);
632d30ea906Sjfb8856606 #endif
633d30ea906Sjfb8856606 }
634d30ea906Sjfb8856606 
635d30ea906Sjfb8856606 module_init(kni_init);
636d30ea906Sjfb8856606 module_exit(kni_exit);
637d30ea906Sjfb8856606 
638d30ea906Sjfb8856606 module_param(lo_mode, charp, 0644);
639d30ea906Sjfb8856606 MODULE_PARM_DESC(lo_mode,
640d30ea906Sjfb8856606 "KNI loopback mode (default=lo_mode_none):\n"
641d30ea906Sjfb8856606 "\t\tlo_mode_none        Kernel loopback disabled\n"
642d30ea906Sjfb8856606 "\t\tlo_mode_fifo        Enable kernel loopback with fifo\n"
643d30ea906Sjfb8856606 "\t\tlo_mode_fifo_skb    Enable kernel loopback with fifo and skb buffer\n"
644d30ea906Sjfb8856606 "\t\t"
645d30ea906Sjfb8856606 );
646d30ea906Sjfb8856606 
647d30ea906Sjfb8856606 module_param(kthread_mode, charp, 0644);
648d30ea906Sjfb8856606 MODULE_PARM_DESC(kthread_mode,
649d30ea906Sjfb8856606 "Kernel thread mode (default=single):\n"
650d30ea906Sjfb8856606 "\t\tsingle    Single kernel thread mode enabled.\n"
651d30ea906Sjfb8856606 "\t\tmultiple  Multiple kernel thread mode enabled.\n"
652d30ea906Sjfb8856606 "\t\t"
653d30ea906Sjfb8856606 );
654d30ea906Sjfb8856606 
655d30ea906Sjfb8856606 module_param(carrier, charp, 0644);
656d30ea906Sjfb8856606 MODULE_PARM_DESC(carrier,
657d30ea906Sjfb8856606 "Default carrier state for KNI interface (default=off):\n"
658d30ea906Sjfb8856606 "\t\toff   Interfaces will be created with carrier state set to off.\n"
659d30ea906Sjfb8856606 "\t\ton    Interfaces will be created with carrier state set to on.\n"
660d30ea906Sjfb8856606 "\t\t"
661d30ea906Sjfb8856606 );
662