xref: /linux-6.15/lib/tests/usercopy_kunit.c (revision db6fe4d6)
1*db6fe4d6SKees Cook // SPDX-License-Identifier: GPL-2.0-only
2*db6fe4d6SKees Cook /*
3*db6fe4d6SKees Cook  * Kernel module for testing copy_to/from_user infrastructure.
4*db6fe4d6SKees Cook  *
5*db6fe4d6SKees Cook  * Copyright 2013 Google Inc. All Rights Reserved
6*db6fe4d6SKees Cook  *
7*db6fe4d6SKees Cook  * Authors:
8*db6fe4d6SKees Cook  *      Kees Cook       <[email protected]>
9*db6fe4d6SKees Cook  */
10*db6fe4d6SKees Cook 
11*db6fe4d6SKees Cook #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
12*db6fe4d6SKees Cook 
13*db6fe4d6SKees Cook #include <linux/mman.h>
14*db6fe4d6SKees Cook #include <linux/module.h>
15*db6fe4d6SKees Cook #include <linux/sched.h>
16*db6fe4d6SKees Cook #include <linux/slab.h>
17*db6fe4d6SKees Cook #include <linux/uaccess.h>
18*db6fe4d6SKees Cook #include <kunit/test.h>
19*db6fe4d6SKees Cook 
20*db6fe4d6SKees Cook /*
21*db6fe4d6SKees Cook  * Several 32-bit architectures support 64-bit {get,put}_user() calls.
22*db6fe4d6SKees Cook  * As there doesn't appear to be anything that can safely determine
23*db6fe4d6SKees Cook  * their capability at compile-time, we just have to opt-out certain archs.
24*db6fe4d6SKees Cook  */
25*db6fe4d6SKees Cook #if BITS_PER_LONG == 64 || (!(defined(CONFIG_ARM) && !defined(MMU)) && \
26*db6fe4d6SKees Cook 			    !defined(CONFIG_M68K) &&		\
27*db6fe4d6SKees Cook 			    !defined(CONFIG_MICROBLAZE) &&	\
28*db6fe4d6SKees Cook 			    !defined(CONFIG_NIOS2) &&		\
29*db6fe4d6SKees Cook 			    !defined(CONFIG_PPC32) &&		\
30*db6fe4d6SKees Cook 			    !defined(CONFIG_SUPERH))
31*db6fe4d6SKees Cook # define TEST_U64
32*db6fe4d6SKees Cook #endif
33*db6fe4d6SKees Cook 
34*db6fe4d6SKees Cook struct usercopy_test_priv {
35*db6fe4d6SKees Cook 	char *kmem;
36*db6fe4d6SKees Cook 	char __user *umem;
37*db6fe4d6SKees Cook 	size_t size;
38*db6fe4d6SKees Cook };
39*db6fe4d6SKees Cook 
is_zeroed(void * from,size_t size)40*db6fe4d6SKees Cook static bool is_zeroed(void *from, size_t size)
41*db6fe4d6SKees Cook {
42*db6fe4d6SKees Cook 	return memchr_inv(from, 0x0, size) == NULL;
43*db6fe4d6SKees Cook }
44*db6fe4d6SKees Cook 
45*db6fe4d6SKees Cook /* Test usage of check_nonzero_user(). */
usercopy_test_check_nonzero_user(struct kunit * test)46*db6fe4d6SKees Cook static void usercopy_test_check_nonzero_user(struct kunit *test)
47*db6fe4d6SKees Cook {
48*db6fe4d6SKees Cook 	size_t start, end, i, zero_start, zero_end;
49*db6fe4d6SKees Cook 	struct usercopy_test_priv *priv = test->priv;
50*db6fe4d6SKees Cook 	char __user *umem = priv->umem;
51*db6fe4d6SKees Cook 	char *kmem = priv->kmem;
52*db6fe4d6SKees Cook 	size_t size = priv->size;
53*db6fe4d6SKees Cook 
54*db6fe4d6SKees Cook 	KUNIT_ASSERT_GE_MSG(test, size, 2 * PAGE_SIZE, "buffer too small");
55*db6fe4d6SKees Cook 
56*db6fe4d6SKees Cook 	/*
57*db6fe4d6SKees Cook 	 * We want to cross a page boundary to exercise the code more
58*db6fe4d6SKees Cook 	 * effectively. We also don't want to make the size we scan too large,
59*db6fe4d6SKees Cook 	 * otherwise the test can take a long time and cause soft lockups. So
60*db6fe4d6SKees Cook 	 * scan a 1024 byte region across the page boundary.
61*db6fe4d6SKees Cook 	 */
62*db6fe4d6SKees Cook 	size = 1024;
63*db6fe4d6SKees Cook 	start = PAGE_SIZE - (size / 2);
64*db6fe4d6SKees Cook 
65*db6fe4d6SKees Cook 	kmem += start;
66*db6fe4d6SKees Cook 	umem += start;
67*db6fe4d6SKees Cook 
68*db6fe4d6SKees Cook 	zero_start = size / 4;
69*db6fe4d6SKees Cook 	zero_end = size - zero_start;
70*db6fe4d6SKees Cook 
71*db6fe4d6SKees Cook 	/*
72*db6fe4d6SKees Cook 	 * We conduct a series of check_nonzero_user() tests on a block of
73*db6fe4d6SKees Cook 	 * memory with the following byte-pattern (trying every possible
74*db6fe4d6SKees Cook 	 * [start,end] pair):
75*db6fe4d6SKees Cook 	 *
76*db6fe4d6SKees Cook 	 *   [ 00 ff 00 ff ... 00 00 00 00 ... ff 00 ff 00 ]
77*db6fe4d6SKees Cook 	 *
78*db6fe4d6SKees Cook 	 * And we verify that check_nonzero_user() acts identically to
79*db6fe4d6SKees Cook 	 * memchr_inv().
80*db6fe4d6SKees Cook 	 */
81*db6fe4d6SKees Cook 
82*db6fe4d6SKees Cook 	memset(kmem, 0x0, size);
83*db6fe4d6SKees Cook 	for (i = 1; i < zero_start; i += 2)
84*db6fe4d6SKees Cook 		kmem[i] = 0xff;
85*db6fe4d6SKees Cook 	for (i = zero_end; i < size; i += 2)
86*db6fe4d6SKees Cook 		kmem[i] = 0xff;
87*db6fe4d6SKees Cook 
88*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, copy_to_user(umem, kmem, size), 0,
89*db6fe4d6SKees Cook 		"legitimate copy_to_user failed");
90*db6fe4d6SKees Cook 
91*db6fe4d6SKees Cook 	for (start = 0; start <= size; start++) {
92*db6fe4d6SKees Cook 		for (end = start; end <= size; end++) {
93*db6fe4d6SKees Cook 			size_t len = end - start;
94*db6fe4d6SKees Cook 			int retval = check_zeroed_user(umem + start, len);
95*db6fe4d6SKees Cook 			int expected = is_zeroed(kmem + start, len);
96*db6fe4d6SKees Cook 
97*db6fe4d6SKees Cook 			KUNIT_ASSERT_EQ_MSG(test, retval, expected,
98*db6fe4d6SKees Cook 				"check_nonzero_user(=%d) != memchr_inv(=%d) mismatch (start=%zu, end=%zu)",
99*db6fe4d6SKees Cook 				retval, expected, start, end);
100*db6fe4d6SKees Cook 		}
101*db6fe4d6SKees Cook 	}
102*db6fe4d6SKees Cook }
103*db6fe4d6SKees Cook 
104*db6fe4d6SKees Cook /* Test usage of copy_struct_from_user(). */
usercopy_test_copy_struct_from_user(struct kunit * test)105*db6fe4d6SKees Cook static void usercopy_test_copy_struct_from_user(struct kunit *test)
106*db6fe4d6SKees Cook {
107*db6fe4d6SKees Cook 	char *umem_src = NULL, *expected = NULL;
108*db6fe4d6SKees Cook 	struct usercopy_test_priv *priv = test->priv;
109*db6fe4d6SKees Cook 	char __user *umem = priv->umem;
110*db6fe4d6SKees Cook 	char *kmem = priv->kmem;
111*db6fe4d6SKees Cook 	size_t size = priv->size;
112*db6fe4d6SKees Cook 	size_t ksize, usize;
113*db6fe4d6SKees Cook 
114*db6fe4d6SKees Cook 	umem_src = kunit_kmalloc(test, size, GFP_KERNEL);
115*db6fe4d6SKees Cook 	KUNIT_ASSERT_NOT_ERR_OR_NULL(test, umem_src);
116*db6fe4d6SKees Cook 
117*db6fe4d6SKees Cook 	expected = kunit_kmalloc(test, size, GFP_KERNEL);
118*db6fe4d6SKees Cook 	KUNIT_ASSERT_NOT_ERR_OR_NULL(test, expected);
119*db6fe4d6SKees Cook 
120*db6fe4d6SKees Cook 	/* Fill umem with a fixed byte pattern. */
121*db6fe4d6SKees Cook 	memset(umem_src, 0x3e, size);
122*db6fe4d6SKees Cook 	KUNIT_ASSERT_EQ_MSG(test, copy_to_user(umem, umem_src, size), 0,
123*db6fe4d6SKees Cook 		    "legitimate copy_to_user failed");
124*db6fe4d6SKees Cook 
125*db6fe4d6SKees Cook 	/* Check basic case -- (usize == ksize). */
126*db6fe4d6SKees Cook 	ksize = size;
127*db6fe4d6SKees Cook 	usize = size;
128*db6fe4d6SKees Cook 
129*db6fe4d6SKees Cook 	memcpy(expected, umem_src, ksize);
130*db6fe4d6SKees Cook 
131*db6fe4d6SKees Cook 	memset(kmem, 0x0, size);
132*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
133*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize == ksize) failed");
134*db6fe4d6SKees Cook 	KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
135*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize == ksize) gives unexpected copy");
136*db6fe4d6SKees Cook 
137*db6fe4d6SKees Cook 	/* Old userspace case -- (usize < ksize). */
138*db6fe4d6SKees Cook 	ksize = size;
139*db6fe4d6SKees Cook 	usize = size / 2;
140*db6fe4d6SKees Cook 
141*db6fe4d6SKees Cook 	memcpy(expected, umem_src, usize);
142*db6fe4d6SKees Cook 	memset(expected + usize, 0x0, ksize - usize);
143*db6fe4d6SKees Cook 
144*db6fe4d6SKees Cook 	memset(kmem, 0x0, size);
145*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
146*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize < ksize) failed");
147*db6fe4d6SKees Cook 	KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
148*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize < ksize) gives unexpected copy");
149*db6fe4d6SKees Cook 
150*db6fe4d6SKees Cook 	/* New userspace (-E2BIG) case -- (usize > ksize). */
151*db6fe4d6SKees Cook 	ksize = size / 2;
152*db6fe4d6SKees Cook 	usize = size;
153*db6fe4d6SKees Cook 
154*db6fe4d6SKees Cook 	memset(kmem, 0x0, size);
155*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), -E2BIG,
156*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize > ksize) didn't give E2BIG");
157*db6fe4d6SKees Cook 
158*db6fe4d6SKees Cook 	/* New userspace (success) case -- (usize > ksize). */
159*db6fe4d6SKees Cook 	ksize = size / 2;
160*db6fe4d6SKees Cook 	usize = size;
161*db6fe4d6SKees Cook 
162*db6fe4d6SKees Cook 	memcpy(expected, umem_src, ksize);
163*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, clear_user(umem + ksize, usize - ksize), 0,
164*db6fe4d6SKees Cook 		    "legitimate clear_user failed");
165*db6fe4d6SKees Cook 
166*db6fe4d6SKees Cook 	memset(kmem, 0x0, size);
167*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, copy_struct_from_user(kmem, ksize, umem, usize), 0,
168*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize > ksize) failed");
169*db6fe4d6SKees Cook 	KUNIT_EXPECT_MEMEQ_MSG(test, kmem, expected, ksize,
170*db6fe4d6SKees Cook 		    "copy_struct_from_user(usize > ksize) gives unexpected copy");
171*db6fe4d6SKees Cook }
172*db6fe4d6SKees Cook 
173*db6fe4d6SKees Cook /*
174*db6fe4d6SKees Cook  * Legitimate usage: none of these copies should fail.
175*db6fe4d6SKees Cook  */
usercopy_test_valid(struct kunit * test)176*db6fe4d6SKees Cook static void usercopy_test_valid(struct kunit *test)
177*db6fe4d6SKees Cook {
178*db6fe4d6SKees Cook 	struct usercopy_test_priv *priv = test->priv;
179*db6fe4d6SKees Cook 	char __user *usermem = priv->umem;
180*db6fe4d6SKees Cook 	char *kmem = priv->kmem;
181*db6fe4d6SKees Cook 
182*db6fe4d6SKees Cook 	memset(kmem, 0x3a, PAGE_SIZE * 2);
183*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, 0, copy_to_user(usermem, kmem, PAGE_SIZE),
184*db6fe4d6SKees Cook 	     "legitimate copy_to_user failed");
185*db6fe4d6SKees Cook 	memset(kmem, 0x0, PAGE_SIZE);
186*db6fe4d6SKees Cook 	KUNIT_EXPECT_EQ_MSG(test, 0, copy_from_user(kmem, usermem, PAGE_SIZE),
187*db6fe4d6SKees Cook 	     "legitimate copy_from_user failed");
188*db6fe4d6SKees Cook 	KUNIT_EXPECT_MEMEQ_MSG(test, kmem, kmem + PAGE_SIZE, PAGE_SIZE,
189*db6fe4d6SKees Cook 	     "legitimate usercopy failed to copy data");
190*db6fe4d6SKees Cook 
191*db6fe4d6SKees Cook #define test_legit(size, check)						\
192*db6fe4d6SKees Cook 	do {								\
193*db6fe4d6SKees Cook 		size val_##size = (check);				\
194*db6fe4d6SKees Cook 		KUNIT_EXPECT_EQ_MSG(test, 0,				\
195*db6fe4d6SKees Cook 			put_user(val_##size, (size __user *)usermem),	\
196*db6fe4d6SKees Cook 			"legitimate put_user (" #size ") failed");	\
197*db6fe4d6SKees Cook 		val_##size = 0;						\
198*db6fe4d6SKees Cook 		KUNIT_EXPECT_EQ_MSG(test, 0,				\
199*db6fe4d6SKees Cook 			get_user(val_##size, (size __user *)usermem),	\
200*db6fe4d6SKees Cook 			"legitimate get_user (" #size ") failed");	\
201*db6fe4d6SKees Cook 		KUNIT_EXPECT_EQ_MSG(test, val_##size, check,		\
202*db6fe4d6SKees Cook 			"legitimate get_user (" #size ") failed to do copy"); \
203*db6fe4d6SKees Cook 	} while (0)
204*db6fe4d6SKees Cook 
205*db6fe4d6SKees Cook 	test_legit(u8,  0x5a);
206*db6fe4d6SKees Cook 	test_legit(u16, 0x5a5b);
207*db6fe4d6SKees Cook 	test_legit(u32, 0x5a5b5c5d);
208*db6fe4d6SKees Cook #ifdef TEST_U64
209*db6fe4d6SKees Cook 	test_legit(u64, 0x5a5b5c5d6a6b6c6d);
210*db6fe4d6SKees Cook #endif
211*db6fe4d6SKees Cook #undef test_legit
212*db6fe4d6SKees Cook }
213*db6fe4d6SKees Cook 
214*db6fe4d6SKees Cook /*
215*db6fe4d6SKees Cook  * Invalid usage: none of these copies should succeed.
216*db6fe4d6SKees Cook  */
usercopy_test_invalid(struct kunit * test)217*db6fe4d6SKees Cook static void usercopy_test_invalid(struct kunit *test)
218*db6fe4d6SKees Cook {
219*db6fe4d6SKees Cook 	struct usercopy_test_priv *priv = test->priv;
220*db6fe4d6SKees Cook 	char __user *usermem = priv->umem;
221*db6fe4d6SKees Cook 	char *bad_usermem = (char *)usermem;
222*db6fe4d6SKees Cook 	char *kmem = priv->kmem;
223*db6fe4d6SKees Cook 	u64 *kmem_u64 = (u64 *)kmem;
224*db6fe4d6SKees Cook 
225*db6fe4d6SKees Cook 	if (IS_ENABLED(CONFIG_ALTERNATE_USER_ADDRESS_SPACE) ||
226*db6fe4d6SKees Cook 	    !IS_ENABLED(CONFIG_MMU)) {
227*db6fe4d6SKees Cook 		kunit_skip(test, "Testing for kernel/userspace address confusion is only sensible on architectures with a shared address space");
228*db6fe4d6SKees Cook 		return;
229*db6fe4d6SKees Cook 	}
230*db6fe4d6SKees Cook 
231*db6fe4d6SKees Cook 	/* Prepare kernel memory with check values. */
232*db6fe4d6SKees Cook 	memset(kmem, 0x5a, PAGE_SIZE);
233*db6fe4d6SKees Cook 	memset(kmem + PAGE_SIZE, 0, PAGE_SIZE);
234*db6fe4d6SKees Cook 
235*db6fe4d6SKees Cook 	/* Reject kernel-to-kernel copies through copy_from_user(). */
236*db6fe4d6SKees Cook 	KUNIT_EXPECT_NE_MSG(test, copy_from_user(kmem, (char __user *)(kmem + PAGE_SIZE),
237*db6fe4d6SKees Cook 						 PAGE_SIZE), 0,
238*db6fe4d6SKees Cook 		    "illegal all-kernel copy_from_user passed");
239*db6fe4d6SKees Cook 
240*db6fe4d6SKees Cook 	/* Destination half of buffer should have been zeroed. */
241*db6fe4d6SKees Cook 	KUNIT_EXPECT_MEMEQ_MSG(test, kmem + PAGE_SIZE, kmem, PAGE_SIZE,
242*db6fe4d6SKees Cook 		    "zeroing failure for illegal all-kernel copy_from_user");
243*db6fe4d6SKees Cook 
244*db6fe4d6SKees Cook #if 0
245*db6fe4d6SKees Cook 	/*
246*db6fe4d6SKees Cook 	 * When running with SMAP/PAN/etc, this will Oops the kernel
247*db6fe4d6SKees Cook 	 * due to the zeroing of userspace memory on failure. This needs
248*db6fe4d6SKees Cook 	 * to be tested in LKDTM instead, since this test module does not
249*db6fe4d6SKees Cook 	 * expect to explode.
250*db6fe4d6SKees Cook 	 */
251*db6fe4d6SKees Cook 	KUNIT_EXPECT_NE_MSG(test, copy_from_user(bad_usermem, (char __user *)kmem,
252*db6fe4d6SKees Cook 						 PAGE_SIZE), 0,
253*db6fe4d6SKees Cook 		    "illegal reversed copy_from_user passed");
254*db6fe4d6SKees Cook #endif
255*db6fe4d6SKees Cook 	KUNIT_EXPECT_NE_MSG(test, copy_to_user((char __user *)kmem, kmem + PAGE_SIZE,
256*db6fe4d6SKees Cook 					       PAGE_SIZE), 0,
257*db6fe4d6SKees Cook 		    "illegal all-kernel copy_to_user passed");
258*db6fe4d6SKees Cook 
259*db6fe4d6SKees Cook 	KUNIT_EXPECT_NE_MSG(test, copy_to_user((char __user *)kmem, bad_usermem,
260*db6fe4d6SKees Cook 					       PAGE_SIZE), 0,
261*db6fe4d6SKees Cook 		    "illegal reversed copy_to_user passed");
262*db6fe4d6SKees Cook 
263*db6fe4d6SKees Cook #define test_illegal(size, check)							\
264*db6fe4d6SKees Cook 	do {										\
265*db6fe4d6SKees Cook 		size val_##size = (check);						\
266*db6fe4d6SKees Cook 		/* get_user() */							\
267*db6fe4d6SKees Cook 		KUNIT_EXPECT_NE_MSG(test, get_user(val_##size, (size __user *)kmem), 0,	\
268*db6fe4d6SKees Cook 		    "illegal get_user (" #size ") passed");				\
269*db6fe4d6SKees Cook 		KUNIT_EXPECT_EQ_MSG(test, val_##size, 0,				\
270*db6fe4d6SKees Cook 		    "zeroing failure for illegal get_user (" #size ")");		\
271*db6fe4d6SKees Cook 		/* put_user() */							\
272*db6fe4d6SKees Cook 		*kmem_u64 = 0xF09FA4AFF09FA4AF;						\
273*db6fe4d6SKees Cook 		KUNIT_EXPECT_NE_MSG(test, put_user(val_##size, (size __user *)kmem), 0,	\
274*db6fe4d6SKees Cook 		    "illegal put_user (" #size ") passed");				\
275*db6fe4d6SKees Cook 		KUNIT_EXPECT_EQ_MSG(test, *kmem_u64, 0xF09FA4AFF09FA4AF,		\
276*db6fe4d6SKees Cook 		    "illegal put_user (" #size ") wrote to kernel memory!");		\
277*db6fe4d6SKees Cook 	} while (0)
278*db6fe4d6SKees Cook 
279*db6fe4d6SKees Cook 	test_illegal(u8,  0x5a);
280*db6fe4d6SKees Cook 	test_illegal(u16, 0x5a5b);
281*db6fe4d6SKees Cook 	test_illegal(u32, 0x5a5b5c5d);
282*db6fe4d6SKees Cook #ifdef TEST_U64
283*db6fe4d6SKees Cook 	test_illegal(u64, 0x5a5b5c5d6a6b6c6d);
284*db6fe4d6SKees Cook #endif
285*db6fe4d6SKees Cook #undef test_illegal
286*db6fe4d6SKees Cook }
287*db6fe4d6SKees Cook 
usercopy_test_init(struct kunit * test)288*db6fe4d6SKees Cook static int usercopy_test_init(struct kunit *test)
289*db6fe4d6SKees Cook {
290*db6fe4d6SKees Cook 	struct usercopy_test_priv *priv;
291*db6fe4d6SKees Cook 	unsigned long user_addr;
292*db6fe4d6SKees Cook 
293*db6fe4d6SKees Cook 	if (!IS_ENABLED(CONFIG_MMU)) {
294*db6fe4d6SKees Cook 		kunit_skip(test, "Userspace allocation testing not available on non-MMU systems");
295*db6fe4d6SKees Cook 		return 0;
296*db6fe4d6SKees Cook 	}
297*db6fe4d6SKees Cook 
298*db6fe4d6SKees Cook 	priv = kunit_kzalloc(test, sizeof(*priv), GFP_KERNEL);
299*db6fe4d6SKees Cook 	KUNIT_ASSERT_NOT_ERR_OR_NULL(test, priv);
300*db6fe4d6SKees Cook 	test->priv = priv;
301*db6fe4d6SKees Cook 	priv->size = PAGE_SIZE * 2;
302*db6fe4d6SKees Cook 
303*db6fe4d6SKees Cook 	priv->kmem = kunit_kmalloc(test, priv->size, GFP_KERNEL);
304*db6fe4d6SKees Cook 	KUNIT_ASSERT_NOT_ERR_OR_NULL(test, priv->kmem);
305*db6fe4d6SKees Cook 
306*db6fe4d6SKees Cook 	user_addr = kunit_vm_mmap(test, NULL, 0, priv->size,
307*db6fe4d6SKees Cook 			    PROT_READ | PROT_WRITE | PROT_EXEC,
308*db6fe4d6SKees Cook 			    MAP_ANONYMOUS | MAP_PRIVATE, 0);
309*db6fe4d6SKees Cook 	KUNIT_ASSERT_NE_MSG(test, user_addr, 0,
310*db6fe4d6SKees Cook 		"Could not create userspace mm");
311*db6fe4d6SKees Cook 	KUNIT_ASSERT_LT_MSG(test, user_addr, (unsigned long)TASK_SIZE,
312*db6fe4d6SKees Cook 		"Failed to allocate user memory");
313*db6fe4d6SKees Cook 	priv->umem = (char __user *)user_addr;
314*db6fe4d6SKees Cook 
315*db6fe4d6SKees Cook 	return 0;
316*db6fe4d6SKees Cook }
317*db6fe4d6SKees Cook 
318*db6fe4d6SKees Cook static struct kunit_case usercopy_test_cases[] = {
319*db6fe4d6SKees Cook 	KUNIT_CASE(usercopy_test_valid),
320*db6fe4d6SKees Cook 	KUNIT_CASE(usercopy_test_invalid),
321*db6fe4d6SKees Cook 	KUNIT_CASE(usercopy_test_check_nonzero_user),
322*db6fe4d6SKees Cook 	KUNIT_CASE(usercopy_test_copy_struct_from_user),
323*db6fe4d6SKees Cook 	{}
324*db6fe4d6SKees Cook };
325*db6fe4d6SKees Cook 
326*db6fe4d6SKees Cook static struct kunit_suite usercopy_test_suite = {
327*db6fe4d6SKees Cook 	.name = "usercopy",
328*db6fe4d6SKees Cook 	.init = usercopy_test_init,
329*db6fe4d6SKees Cook 	.test_cases = usercopy_test_cases,
330*db6fe4d6SKees Cook };
331*db6fe4d6SKees Cook 
332*db6fe4d6SKees Cook kunit_test_suites(&usercopy_test_suite);
333*db6fe4d6SKees Cook MODULE_AUTHOR("Kees Cook <[email protected]>");
334*db6fe4d6SKees Cook MODULE_DESCRIPTION("Kernel module for testing copy_to/from_user infrastructure");
335*db6fe4d6SKees Cook MODULE_LICENSE("GPL");
336