Home
last modified time | relevance | path

Searched refs:bulk_len (Results 1 – 12 of 12) sorted by relevance

/linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/
H A Dfs_pool.c8 int bulk_len) in mlx5_fs_bulk_init() argument
12 fs_bulk->bitmask = kvcalloc(BITS_TO_LONGS(bulk_len), sizeof(unsigned long), in mlx5_fs_bulk_init()
17 fs_bulk->bulk_len = bulk_len; in mlx5_fs_bulk_init()
18 for (i = 0; i < bulk_len; i++) in mlx5_fs_bulk_init()
31 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount()
37 int free_index = find_first_bit(fs_bulk->bitmask, fs_bulk->bulk_len); in mlx5_fs_bulk_acquire_index()
40 if (free_index >= fs_bulk->bulk_len) in mlx5_fs_bulk_acquire_index()
97 fs_pool->available_units += new_bulk->bulk_len; in mlx5_fs_pool_alloc_new_bulk()
107 fs_pool->available_units -= bulk->bulk_len; in mlx5_fs_pool_free_bulk()
184 if (bulk_free_amount == bulk->bulk_len) { in mlx5_fs_pool_release_index()
H A Dfs_counters.c114 u32 bulk_len = fc_stats->bulk_query_len; in mlx5_fc_stats_query_all_counters() local
131 last_bulk_id = bulk_base_id + bulk_len; in mlx5_fc_stats_query_all_counters()
137 err = mlx5_cmd_fc_bulk_query(dev, bulk_base_id, bulk_len, data); in mlx5_fc_stats_query_all_counters()
434 int bulk_len; in mlx5_fc_bulk_create() local
439 bulk_len = alloc_bitmask > 0 ? MLX5_FC_BULK_NUM_FCS(alloc_bitmask) : 1; in mlx5_fc_bulk_create()
441 fc_bulk = kvzalloc(struct_size(fc_bulk, fcs, bulk_len), GFP_KERNEL); in mlx5_fc_bulk_create()
445 if (mlx5_fs_bulk_init(dev, &fc_bulk->fs_bulk, bulk_len)) in mlx5_fc_bulk_create()
451 for (i = 0; i < bulk_len; i++) in mlx5_fc_bulk_create()
472 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fc_bulk_destroy()
564 fc_bulk->fs_bulk.bulk_len = bulk_size; in mlx5_fc_local_create()
H A Dfs_pool.h11 int bulk_len; member
43 int bulk_len);
H A Dfs_cmd.h118 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len);
119 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len,
H A Dfs_cmd.c853 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len) in mlx5_cmd_fc_get_bulk_query_out_len() argument
856 MLX5_ST_SZ_BYTES(traffic_counter) * bulk_len; in mlx5_cmd_fc_get_bulk_query_out_len()
859 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len, in mlx5_cmd_fc_bulk_query() argument
862 int outlen = mlx5_cmd_fc_get_bulk_query_out_len(bulk_len); in mlx5_cmd_fc_bulk_query()
868 MLX5_SET(query_flow_counter_in, in, num_of_counters, bulk_len); in mlx5_cmd_fc_bulk_query()
/linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/
H A Dfs_hws_pools.c113 int bulk_len; in mlx5_fs_hws_pr_bulk_create() local
119 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_pr_bulk_create()
120 pr_bulk = kvzalloc(struct_size(pr_bulk, prs_data, bulk_len), GFP_KERNEL); in mlx5_fs_hws_pr_bulk_create()
124 if (mlx5_fs_bulk_init(dev, &pr_bulk->fs_bulk, bulk_len)) in mlx5_fs_hws_pr_bulk_create()
127 for (i = 0; i < bulk_len; i++) { in mlx5_fs_hws_pr_bulk_create()
151 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fs_hws_pr_bulk_destroy()
259 int bulk_len; in mlx5_fs_hws_mh_bulk_create() local
273 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_mh_bulk_create()
278 if (mlx5_fs_bulk_init(dev, &mh_bulk->fs_bulk, bulk_len)) in mlx5_fs_hws_mh_bulk_create()
281 for (int i = 0; i < bulk_len; i++) { in mlx5_fs_hws_mh_bulk_create()
[all …]
/linux-6.15/crypto/
H A Dadiantum.c224 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_header() local
229 .message_bits = cpu_to_le64((u64)bulk_len * 8) in adiantum_hash_header()
263 for (i = 0; i < bulk_len; i += n) { in adiantum_hash_message()
265 n = min_t(unsigned int, miter.length, bulk_len - i); in adiantum_hash_message()
306 err = crypto_shash_digest(&rctx->u.hash_desc, virt, bulk_len, in adiantum_finish()
313 memcpy(virt + bulk_len, &rctx->rbuf.bignum, sizeof(le128)); in adiantum_finish()
323 bulk_len, sizeof(le128), 1); in adiantum_finish()
366 err = crypto_shash_digest(&rctx->u.hash_desc, virt, bulk_len, in adiantum_crypt()
368 memcpy(&rctx->rbuf.bignum, virt + bulk_len, sizeof(le128)); in adiantum_crypt()
374 bulk_len, sizeof(le128), 0); in adiantum_crypt()
[all …]
H A Dhctr2.c199 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message() local
201 unsigned int remainder = bulk_len % BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message()
208 for (i = 0; i < bulk_len; i += n) { in hctr2_hash_message()
210 n = min_t(unsigned int, miter.length, bulk_len - i); in hctr2_hash_message()
271 int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_crypt() local
315 rctx->bulk_part_dst, bulk_len, in hctr2_crypt()
/linux-6.15/net/core/
H A Dpage_pool.c852 u32 bulk_len) in page_pool_recycle_ring_bulk() argument
860 for (i = 0; i < bulk_len; i++) { in page_pool_recycle_ring_bulk()
872 if (likely(i == bulk_len)) in page_pool_recycle_ring_bulk()
879 for (; i < bulk_len; i++) in page_pool_recycle_ring_bulk()
899 u32 bulk_len = 0; in page_pool_put_netmem_bulk() local
905 data[bulk_len++] = netmem; in page_pool_put_netmem_bulk()
908 count = bulk_len; in page_pool_put_netmem_bulk()
915 bulk_len = 0; in page_pool_put_netmem_bulk()
938 bulk[bulk_len++] = netmem; in page_pool_put_netmem_bulk()
941 if (bulk_len) in page_pool_put_netmem_bulk()
[all …]
/linux-6.15/drivers/gpu/drm/gud/
H A Dgud_drv.c333 string_get_size(gdrm->bulk_len, 1, STRING_UNITS_2, buf, sizeof(buf)); in gud_stats_debugfs()
395 gdrm->bulk_buf = vmalloc_32(gdrm->bulk_len); in gud_alloc_bulk_buffer()
399 num_pages = DIV_ROUND_UP(gdrm->bulk_len, PAGE_SIZE); in gud_alloc_bulk_buffer()
408 0, gdrm->bulk_len, GFP_KERNEL); in gud_alloc_bulk_buffer()
569 gdrm->bulk_len = max_buffer_size; in gud_probe()
580 gdrm->compress_buf = vmalloc(gdrm->bulk_len); in gud_probe()
H A Dgud_pipe.c166 if (len > gdrm->bulk_len) in gud_prep_flush()
338 if (gdrm->bulk_len < lines * pitch) in gud_flush_damage()
339 lines = gdrm->bulk_len / pitch; in gud_flush_damage()
H A Dgud_internal.h29 size_t bulk_len; member