| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | fs_pool.c | 8 int bulk_len) in mlx5_fs_bulk_init() argument 12 fs_bulk->bitmask = kvcalloc(BITS_TO_LONGS(bulk_len), sizeof(unsigned long), in mlx5_fs_bulk_init() 17 fs_bulk->bulk_len = bulk_len; in mlx5_fs_bulk_init() 18 for (i = 0; i < bulk_len; i++) in mlx5_fs_bulk_init() 31 return bitmap_weight(bulk->bitmask, bulk->bulk_len); in mlx5_fs_bulk_get_free_amount() 37 int free_index = find_first_bit(fs_bulk->bitmask, fs_bulk->bulk_len); in mlx5_fs_bulk_acquire_index() 40 if (free_index >= fs_bulk->bulk_len) in mlx5_fs_bulk_acquire_index() 97 fs_pool->available_units += new_bulk->bulk_len; in mlx5_fs_pool_alloc_new_bulk() 107 fs_pool->available_units -= bulk->bulk_len; in mlx5_fs_pool_free_bulk() 184 if (bulk_free_amount == bulk->bulk_len) { in mlx5_fs_pool_release_index()
|
| H A D | fs_counters.c | 114 u32 bulk_len = fc_stats->bulk_query_len; in mlx5_fc_stats_query_all_counters() local 131 last_bulk_id = bulk_base_id + bulk_len; in mlx5_fc_stats_query_all_counters() 137 err = mlx5_cmd_fc_bulk_query(dev, bulk_base_id, bulk_len, data); in mlx5_fc_stats_query_all_counters() 434 int bulk_len; in mlx5_fc_bulk_create() local 439 bulk_len = alloc_bitmask > 0 ? MLX5_FC_BULK_NUM_FCS(alloc_bitmask) : 1; in mlx5_fc_bulk_create() 441 fc_bulk = kvzalloc(struct_size(fc_bulk, fcs, bulk_len), GFP_KERNEL); in mlx5_fc_bulk_create() 445 if (mlx5_fs_bulk_init(dev, &fc_bulk->fs_bulk, bulk_len)) in mlx5_fc_bulk_create() 451 for (i = 0; i < bulk_len; i++) in mlx5_fc_bulk_create() 472 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fc_bulk_destroy() 564 fc_bulk->fs_bulk.bulk_len = bulk_size; in mlx5_fc_local_create()
|
| H A D | fs_pool.h | 11 int bulk_len; member 43 int bulk_len);
|
| H A D | fs_cmd.h | 118 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len); 119 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len,
|
| H A D | fs_cmd.c | 853 int mlx5_cmd_fc_get_bulk_query_out_len(int bulk_len) in mlx5_cmd_fc_get_bulk_query_out_len() argument 856 MLX5_ST_SZ_BYTES(traffic_counter) * bulk_len; in mlx5_cmd_fc_get_bulk_query_out_len() 859 int mlx5_cmd_fc_bulk_query(struct mlx5_core_dev *dev, u32 base_id, int bulk_len, in mlx5_cmd_fc_bulk_query() argument 862 int outlen = mlx5_cmd_fc_get_bulk_query_out_len(bulk_len); in mlx5_cmd_fc_bulk_query() 868 MLX5_SET(query_flow_counter_in, in, num_of_counters, bulk_len); in mlx5_cmd_fc_bulk_query()
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | fs_hws_pools.c | 113 int bulk_len; in mlx5_fs_hws_pr_bulk_create() local 119 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_pr_bulk_create() 120 pr_bulk = kvzalloc(struct_size(pr_bulk, prs_data, bulk_len), GFP_KERNEL); in mlx5_fs_hws_pr_bulk_create() 124 if (mlx5_fs_bulk_init(dev, &pr_bulk->fs_bulk, bulk_len)) in mlx5_fs_hws_pr_bulk_create() 127 for (i = 0; i < bulk_len; i++) { in mlx5_fs_hws_pr_bulk_create() 151 if (mlx5_fs_bulk_get_free_amount(fs_bulk) < fs_bulk->bulk_len) { in mlx5_fs_hws_pr_bulk_destroy() 259 int bulk_len; in mlx5_fs_hws_mh_bulk_create() local 273 bulk_len = MLX5_FS_HWS_DEFAULT_BULK_LEN; in mlx5_fs_hws_mh_bulk_create() 278 if (mlx5_fs_bulk_init(dev, &mh_bulk->fs_bulk, bulk_len)) in mlx5_fs_hws_mh_bulk_create() 281 for (int i = 0; i < bulk_len; i++) { in mlx5_fs_hws_mh_bulk_create() [all …]
|
| /linux-6.15/crypto/ |
| H A D | adiantum.c | 224 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in adiantum_hash_header() local 229 .message_bits = cpu_to_le64((u64)bulk_len * 8) in adiantum_hash_header() 263 for (i = 0; i < bulk_len; i += n) { in adiantum_hash_message() 265 n = min_t(unsigned int, miter.length, bulk_len - i); in adiantum_hash_message() 306 err = crypto_shash_digest(&rctx->u.hash_desc, virt, bulk_len, in adiantum_finish() 313 memcpy(virt + bulk_len, &rctx->rbuf.bignum, sizeof(le128)); in adiantum_finish() 323 bulk_len, sizeof(le128), 1); in adiantum_finish() 366 err = crypto_shash_digest(&rctx->u.hash_desc, virt, bulk_len, in adiantum_crypt() 368 memcpy(&rctx->rbuf.bignum, virt + bulk_len, sizeof(le128)); in adiantum_crypt() 374 bulk_len, sizeof(le128), 0); in adiantum_crypt() [all …]
|
| H A D | hctr2.c | 199 const unsigned int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message() local 201 unsigned int remainder = bulk_len % BLOCKCIPHER_BLOCK_SIZE; in hctr2_hash_message() 208 for (i = 0; i < bulk_len; i += n) { in hctr2_hash_message() 210 n = min_t(unsigned int, miter.length, bulk_len - i); in hctr2_hash_message() 271 int bulk_len = req->cryptlen - BLOCKCIPHER_BLOCK_SIZE; in hctr2_crypt() local 315 rctx->bulk_part_dst, bulk_len, in hctr2_crypt()
|
| /linux-6.15/net/core/ |
| H A D | page_pool.c | 852 u32 bulk_len) in page_pool_recycle_ring_bulk() argument 860 for (i = 0; i < bulk_len; i++) { in page_pool_recycle_ring_bulk() 872 if (likely(i == bulk_len)) in page_pool_recycle_ring_bulk() 879 for (; i < bulk_len; i++) in page_pool_recycle_ring_bulk() 899 u32 bulk_len = 0; in page_pool_put_netmem_bulk() local 905 data[bulk_len++] = netmem; in page_pool_put_netmem_bulk() 908 count = bulk_len; in page_pool_put_netmem_bulk() 915 bulk_len = 0; in page_pool_put_netmem_bulk() 938 bulk[bulk_len++] = netmem; in page_pool_put_netmem_bulk() 941 if (bulk_len) in page_pool_put_netmem_bulk() [all …]
|
| /linux-6.15/drivers/gpu/drm/gud/ |
| H A D | gud_drv.c | 333 string_get_size(gdrm->bulk_len, 1, STRING_UNITS_2, buf, sizeof(buf)); in gud_stats_debugfs() 395 gdrm->bulk_buf = vmalloc_32(gdrm->bulk_len); in gud_alloc_bulk_buffer() 399 num_pages = DIV_ROUND_UP(gdrm->bulk_len, PAGE_SIZE); in gud_alloc_bulk_buffer() 408 0, gdrm->bulk_len, GFP_KERNEL); in gud_alloc_bulk_buffer() 569 gdrm->bulk_len = max_buffer_size; in gud_probe() 580 gdrm->compress_buf = vmalloc(gdrm->bulk_len); in gud_probe()
|
| H A D | gud_pipe.c | 166 if (len > gdrm->bulk_len) in gud_prep_flush() 338 if (gdrm->bulk_len < lines * pitch) in gud_flush_damage() 339 lines = gdrm->bulk_len / pitch; in gud_flush_damage()
|
| H A D | gud_internal.h | 29 size_t bulk_len; member
|