Home
last modified time | relevance | path

Searched refs:queues (Results 1 – 25 of 481) sorted by relevance

12345678910>>...20

/linux-6.15/tools/testing/selftests/drivers/net/
H A Dqueues.py22 if queues:
41 if not queues:
44 for q in queues:
66 if not queues:
70 ksft_eq(queues, expected)
74 queues = nl_get_queues(cfg, nl)
75 if not queues:
90 queues = nl_get_queues(cfg, nl)
91 ksft_eq(queues, expected)
95 queues = nl_get_queues(cfg, nl)
[all …]
/linux-6.15/drivers/gpu/drm/imagination/
H A Dpvr_context.c204 if (IS_ERR(ctx->queues.geometry)) { in pvr_context_create_queues()
205 err = PTR_ERR(ctx->queues.geometry); in pvr_context_create_queues()
206 ctx->queues.geometry = NULL; in pvr_context_create_queues()
212 if (IS_ERR(ctx->queues.fragment)) { in pvr_context_create_queues()
214 ctx->queues.fragment = NULL; in pvr_context_create_queues()
222 if (IS_ERR(ctx->queues.compute)) { in pvr_context_create_queues()
223 err = PTR_ERR(ctx->queues.compute); in pvr_context_create_queues()
224 ctx->queues.compute = NULL; in pvr_context_create_queues()
232 if (IS_ERR(ctx->queues.transfer)) { in pvr_context_create_queues()
234 ctx->queues.transfer = NULL; in pvr_context_create_queues()
[all …]
H A Dpvr_queue.c533 job->ctx->queues.fragment); in pvr_queue_prepare_job()
606 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_update_active_state()
608 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_update_active_state()
829 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_timedout_job()
831 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_timedout_job()
845 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_timedout_job()
853 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_timedout_job()
1320 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_create()
1322 mutex_unlock(&pvr_dev->queues.lock); in pvr_queue_create()
1349 mutex_lock(&pvr_dev->queues.lock); in pvr_queue_device_pre_reset()
[all …]
H A Dpvr_context.h87 } queues; member
98 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL; in pvr_context_get_queue_for_job()
100 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.fragment : NULL; in pvr_context_get_queue_for_job()
102 return ctx->type == DRM_PVR_CTX_TYPE_COMPUTE ? ctx->queues.compute : NULL; in pvr_context_get_queue_for_job()
104 return ctx->type == DRM_PVR_CTX_TYPE_TRANSFER_FRAG ? ctx->queues.transfer : NULL; in pvr_context_get_queue_for_job()
/linux-6.15/net/sched/
H A Dsch_multiq.c25 struct Qdisc **queues; member
54 return q->queues[0]; in multiq_classify()
56 return q->queues[band]; in multiq_classify()
166 qdisc_put(q->queues[band]); in multiq_destroy()
168 kfree(q->queues); in multiq_destroy()
220 old = q->queues[i]; in multiq_tune()
221 q->queues[i] = child; in multiq_tune()
241 q->queues = NULL; in multiq_init()
253 if (!q->queues) in multiq_init()
299 return q->queues[band]; in multiq_leaf()
[all …]
H A Dsch_prio.c26 struct Qdisc *queues[TCQ_PRIO_BANDS]; member
63 return q->queues[q->prio2band[0]]; in prio_classify()
65 return q->queues[band]; in prio_classify()
137 qdisc_reset(q->queues[prio]); in prio_reset()
173 qdisc_put(q->queues[prio]); in prio_destroy()
201 if (!queues[i]) { in prio_tune()
203 qdisc_put(queues[--i]); in prio_tune()
217 q->queues[i] = queues[i]; in prio_tune()
225 qdisc_put(q->queues[i]); in prio_tune()
322 return q->queues[band]; in prio_leaf()
[all …]
/linux-6.15/Documentation/networking/
H A Dmulti-pf-netdev.rst142 - /sys/class/net/eth2/queues/tx-0/xps_cpus:000001
143 - /sys/class/net/eth2/queues/tx-1/xps_cpus:001000
144 - /sys/class/net/eth2/queues/tx-2/xps_cpus:000002
145 - /sys/class/net/eth2/queues/tx-3/xps_cpus:002000
146 - /sys/class/net/eth2/queues/tx-4/xps_cpus:000004
147 - /sys/class/net/eth2/queues/tx-5/xps_cpus:004000
148 - /sys/class/net/eth2/queues/tx-6/xps_cpus:000008
149 - /sys/class/net/eth2/queues/tx-7/xps_cpus:008000
150 - /sys/class/net/eth2/queues/tx-8/xps_cpus:000010
151 - /sys/class/net/eth2/queues/tx-9/xps_cpus:010000
[all …]
H A Dtc-queue-filters.rst7 TC can be used for directing traffic to either a set of queues or
12 1) TC filter directing traffic to a set of queues is achieved
14 the priority maps to a traffic class (set of queues) when
23 queues and/or a single queue are supported as below:
25 1) TC flower filter directs incoming traffic to a set of queues using
/linux-6.15/drivers/net/wireless/silabs/wfx/
H A Dqueue.c244 queues[num_queues] = &wvif->tx_queue[i]; in wfx_tx_queues_get_skb()
248 swap(queues[j - 1], queues[j]); in wfx_tx_queues_get_skb()
256 skb = skb_dequeue(&queues[i]->offchan); in wfx_tx_queues_get_skb()
264 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
265 trace_queues_stats(wdev, queues[i]); in wfx_tx_queues_get_skb()
278 skb = skb_dequeue(&queues[i]->cab); in wfx_tx_queues_get_skb()
287 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
288 trace_queues_stats(wdev, queues[i]); in wfx_tx_queues_get_skb()
297 skb = skb_dequeue(&queues[i]->normal); in wfx_tx_queues_get_skb()
299 atomic_inc(&queues[i]->pending_frames); in wfx_tx_queues_get_skb()
[all …]
/linux-6.15/drivers/scsi/aacraid/
H A Dcomminit.c373 struct aac_entry * queues; in aac_comm_init() local
399 queues += HOST_NORM_CMD_ENTRIES; in aac_comm_init()
406 queues += HOST_HIGH_CMD_ENTRIES; in aac_comm_init()
413 queues += ADAP_NORM_CMD_ENTRIES; in aac_comm_init()
420 queues += ADAP_HIGH_CMD_ENTRIES; in aac_comm_init()
426 queues += HOST_NORM_RESP_ENTRIES; in aac_comm_init()
637 if (dev->queues == NULL) { in aac_init_adapter()
643 kfree(dev->queues); in aac_init_adapter()
644 dev->queues = NULL; in aac_init_adapter()
651 kfree(dev->queues); in aac_init_adapter()
[all …]
/linux-6.15/Documentation/ABI/testing/
H A Dsysfs-class-net-queues1 What: /sys/class/net/<iface>/queues/rx-<queue>/rps_cpus
11 What: /sys/class/net/<iface>/queues/rx-<queue>/rps_flow_cnt
19 What: /sys/class/net/<iface>/queues/tx-<queue>/tx_timeout
27 What: /sys/class/net/<iface>/queues/tx-<queue>/tx_maxrate
35 What: /sys/class/net/<iface>/queues/tx-<queue>/xps_cpus
45 What: /sys/class/net/<iface>/queues/tx-<queue>/xps_rxqs
56 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/hold_time
65 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/inflight
73 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/limit
82 What: /sys/class/net/<iface>/queues/tx-<queue>/byte_queue_limits/limit_max
[all …]
/linux-6.15/drivers/nvme/target/
H A Dloop.c30 struct nvme_loop_queue *queues; member
71 return queue - queue->ctrl->queues; in nvme_loop_queue_idx()
198 iod->queue = &ctrl->queues[queue_idx]; in nvme_loop_init_iod()
275 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_destroy_admin_queue()
292 kfree(ctrl->queues); in nvme_loop_free_ctrl()
329 ctrl->queues[i].ctrl = ctrl; in nvme_loop_init_io_queues()
362 ctrl->queues[0].ctrl = ctrl; in nvme_loop_configure_admin_queue()
403 nvmet_sq_destroy(&ctrl->queues[0].nvme_sq); in nvme_loop_configure_admin_queue()
583 ctrl->queues = kcalloc(opts->nr_io_queues + 1, sizeof(*ctrl->queues), in nvme_loop_create_ctrl()
585 if (!ctrl->queues) in nvme_loop_create_ctrl()
[all …]
/linux-6.15/drivers/media/platform/nxp/imx8-isi/
H A Dimx8-isi-m2m.c58 } queues; member
85 return &ctx->queues.out; in mxc_isi_m2m_ctx_qdata()
87 return &ctx->queues.cap; in mxc_isi_m2m_ctx_qdata()
152 ctx->queues.out.info, in mxc_isi_m2m_device_run()
153 &ctx->queues.out.format); in mxc_isi_m2m_device_run()
155 ctx->queues.cap.info, in mxc_isi_m2m_device_run()
156 &ctx->queues.cap.format); in mxc_isi_m2m_device_run()
469 ctx->queues.out.format = *pix; in mxc_isi_m2m_s_fmt_vid()
470 ctx->queues.out.info = info; in mxc_isi_m2m_s_fmt_vid()
477 ctx->queues.cap.format = *pix; in mxc_isi_m2m_s_fmt_vid()
[all …]
/linux-6.15/Documentation/devicetree/bindings/soc/ti/
H A Dkeystone-navigator-qmss.txt32 -- managed-queues : the actual queues managed by each queue manager
33 instance, specified as <"base queue #" "# of queues">.
51 - qpend : pool of qpend(interruptible) queues
53 as free descriptor queues or the
54 transmit DMA queues.
58 <"base queue #" "# of queues">.
60 for interruptible queues. The driver additionally sets
67 queues looking for descriptors that have been pushed
90 monitor up to 32 queues starting at the base queue #.
132 managed-queues = <0 0x2000>;
[all …]
/linux-6.15/sound/virtio/
H A Dvirtio_card.h64 struct virtio_snd_queue queues[VIRTIO_SND_VQ_MAX]; member
86 return &snd->queues[VIRTIO_SND_VQ_CONTROL]; in virtsnd_control_queue()
92 return &snd->queues[VIRTIO_SND_VQ_EVENT]; in virtsnd_event_queue()
98 return &snd->queues[VIRTIO_SND_VQ_TX]; in virtsnd_tx_queue()
104 return &snd->queues[VIRTIO_SND_VQ_RX]; in virtsnd_rx_queue()
/linux-6.15/Documentation/networking/device_drivers/ethernet/ti/
H A Dcpsw.rst28 - CBS shapers should be used with rated queues
156 // Check if num of queues is set correctly:
175 // Leave last 2 tx queues not rated.
176 $ echo 40 > /sys/class/net/eth0/queues/tx-0/tx_maxrate
181 // Check maximum rate of tx (cpdma) queues:
182 $ cat /sys/class/net/eth0/queues/tx-*/tx_maxrate
206 map 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 queues 1@4 hw 0
387 // Check if num of queues is set correctly:
407 // Leave last 4 tx queues as not rated
415 // Check maximum rate of tx (cpdma) queues:
[all …]
/linux-6.15/Documentation/devicetree/bindings/misc/
H A Dintel,ixp4xx-ahb-queue-manager.yaml14 The IXP4xx AHB Queue Manager maintains queues as circular buffers in
17 IXP4xx for accelerating queues, especially for networking. Clients pick
18 queues from the queue manager with foo-queue = <&qmgr N> where the
33 - description: Interrupt for queues 0-31
34 - description: Interrupt for queues 32-63
/linux-6.15/drivers/target/
H A Dtarget_core_tmr.c118 flush_work(&dev->queues[i].sq.work); in core_tmr_abort_task()
120 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_abort_task()
121 list_for_each_entry_safe(se_cmd, next, &dev->queues[i].state_list, in core_tmr_abort_task()
148 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task()
163 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_abort_task()
301 flush_work(&dev->queues[i].sq.work); in core_tmr_drain_state_list()
303 spin_lock_irqsave(&dev->queues[i].lock, flags); in core_tmr_drain_state_list()
304 list_for_each_entry_safe(cmd, next, &dev->queues[i].state_list, in core_tmr_drain_state_list()
333 spin_unlock_irqrestore(&dev->queues[i].lock, flags); in core_tmr_drain_state_list()
/linux-6.15/tools/perf/util/
H A Dintel-bts.c46 struct auxtrace_queues queues; member
222 if (bts->queues.new_data) { in intel_bts_update_queues()
223 bts->queues.new_data = false; in intel_bts_update_queues()
541 struct auxtrace_queues *queues = &bts->queues; in intel_bts_process_tid_exit() local
544 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_process_tid_exit()
712 struct auxtrace_queues *queues = &bts->queues; in intel_bts_free_events() local
715 for (i = 0; i < queues->nr_queues; i++) { in intel_bts_free_events()
717 queues->queue_array[i].priv = NULL; in intel_bts_free_events()
719 auxtrace_queues__free(queues); in intel_bts_free_events()
901 if (bts->queues.populated) in intel_bts_process_auxtrace_info()
[all …]
H A Dauxtrace.c223 queues->nr_queues = nr_queues; in auxtrace_queues__init_nr()
224 queues->queue_array = auxtrace_alloc_queue_array(queues->nr_queues); in auxtrace_queues__init_nr()
225 if (!queues->queue_array) in auxtrace_queues__init_nr()
264 queues->nr_queues = nr_queues; in auxtrace_queues__grow()
265 queues->queue_array = queue_array; in auxtrace_queues__grow()
299 if (idx >= queues->nr_queues) { in auxtrace_queues__queue_buffer()
317 queues->new_data = true; in auxtrace_queues__queue_buffer()
318 queues->populated = true; in auxtrace_queues__queue_buffer()
471 zfree(&queues->queue_array); in auxtrace_queues__free()
472 queues->nr_queues = 0; in auxtrace_queues__free()
[all …]
H A Ds390-cpumsf.c170 struct auxtrace_queues queues; member
206 q = &sf->queues.queue_array[sample->cpu]; in s390_cpumcf_dumpctr()
838 if (!sf->queues.new_data) in s390_cpumsf_update_queues()
841 sf->queues.new_data = false; in s390_cpumsf_update_queues()
1013 struct auxtrace_queues *queues = &sf->queues; in s390_cpumsf_free_queues() local
1016 for (i = 0; i < queues->nr_queues; i++) { in s390_cpumsf_free_queues()
1018 queues->queue_array[i].priv; in s390_cpumsf_free_queues()
1030 zfree(&queues->queue_array[i].priv); in s390_cpumsf_free_queues()
1032 auxtrace_queues__free(queues); in s390_cpumsf_free_queues()
1162 if (sf->queues.populated) in s390_cpumsf_process_auxtrace_info()
[all …]
/linux-6.15/Documentation/arch/arm/keystone/
H A Dknav-qmss.rst15 management of the packet queues. Packets are queued/de-queued by writing or
24 knav qmss driver provides a set of APIs to drivers to open/close qmss queues,
25 allocate descriptor pools, map the descriptors, push/pop to queues etc. For
31 Accumulator QMSS queues using PDSP firmware
34 queue or multiple contiguous queues. drivers/soc/ti/knav_qmss_acc.c is the
37 1 or 32 queues per channel. More description on the firmware is available in
56 Use of accumulated queues requires the firmware image to be present in the
57 file system. The driver doesn't acc queues to the supported queue range if
/linux-6.15/include/linux/
H A Dptr_ring.h626 void ***queues; in ptr_ring_resize_multiple_bh_noprof() local
629 queues = kmalloc_array_noprof(nrings, sizeof(*queues), gfp); in ptr_ring_resize_multiple_bh_noprof()
630 if (!queues) in ptr_ring_resize_multiple_bh_noprof()
634 queues[i] = __ptr_ring_init_queue_alloc_noprof(size, gfp); in ptr_ring_resize_multiple_bh_noprof()
635 if (!queues[i]) in ptr_ring_resize_multiple_bh_noprof()
642 queues[i] = __ptr_ring_swap_queue(rings[i], queues[i], in ptr_ring_resize_multiple_bh_noprof()
649 kvfree(queues[i]); in ptr_ring_resize_multiple_bh_noprof()
651 kfree(queues); in ptr_ring_resize_multiple_bh_noprof()
657 kvfree(queues[i]); in ptr_ring_resize_multiple_bh_noprof()
659 kfree(queues); in ptr_ring_resize_multiple_bh_noprof()
/linux-6.15/Documentation/block/
H A Dblk-mq.rst37 spawns multiple queues with individual entry points local to the CPU, removing
49 blk-mq has two group of queues: software staging queues and hardware dispatch
50 queues. When the request arrives at the block layer, it will try the shortest
56 Then, after the requests are processed by software queues, they will be placed
62 Software staging queues
65 The block IO subsystem adds requests in the software staging queues
71 the number of queues is defined by a per-CPU or per-node basis.
99 queue (a.k.a. run the hardware queue), the software queues mapped to that
102 Hardware dispatch queues
116 requests that were ready to be sent first. The number of hardware queues
[all …]
/linux-6.15/drivers/vdpa/alibaba/
H A Deni_vdpa.c45 int queues; member
118 for (i = 0; i < eni_vdpa->queues; i++) { in eni_vdpa_free_irq()
164 int queues = eni_vdpa->queues; in eni_vdpa_request_irq() local
165 int vectors = queues + 1; in eni_vdpa_request_irq()
177 for (i = 0; i < queues; i++) { in eni_vdpa_request_irq()
195 irq = pci_irq_vector(pdev, queues); in eni_vdpa_request_irq()
202 vp_legacy_config_vector(ldev, queues); in eni_vdpa_request_irq()
500 eni_vdpa->queues = eni_vdpa_get_num_queues(eni_vdpa); in eni_vdpa_probe()
502 eni_vdpa->vring = devm_kcalloc(&pdev->dev, eni_vdpa->queues, in eni_vdpa_probe()
511 for (i = 0; i < eni_vdpa->queues; i++) { in eni_vdpa_probe()
[all …]

12345678910>>...20