| /linux-6.15/tools/perf/util/ |
| H A D | intel-pt.c | 353 struct intel_pt *pt = ptq->pt; in intel_pt_findnew_vmcs_info() local 1320 ptq->pt = pt; in intel_pt_alloc_queue() 1509 struct intel_pt *pt = ptq->pt; in intel_pt_sample_flags() local 1765 struct intel_pt *pt = ptq->pt; in intel_pt_synth_branch_sample() local 1840 struct intel_pt *pt = ptq->pt; in intel_pt_synth_instruction_sample() local 1876 struct intel_pt *pt = ptq->pt; in intel_pt_synth_cycle_sample() local 1907 struct intel_pt *pt = ptq->pt; in intel_pt_synth_transaction_sample() local 1944 struct intel_pt *pt = ptq->pt; in intel_pt_synth_ptwrite_sample() local 1970 struct intel_pt *pt = ptq->pt; in intel_pt_synth_cbr_sample() local 2004 struct intel_pt *pt = ptq->pt; in intel_pt_synth_psb_sample() local [all …]
|
| /linux-6.15/drivers/dma/amd/ptdma/ |
| H A D | ptdma-pci.c | 38 struct pt_device *pt; in pt_alloc_struct() local 40 pt = devm_kzalloc(dev, sizeof(*pt), GFP_KERNEL); in pt_alloc_struct() 42 if (!pt) in pt_alloc_struct() 44 pt->dev = dev; in pt_alloc_struct() 48 return pt; in pt_alloc_struct() 118 pt->pt_irq = 0; in pt_free_irqs() 131 if (!pt) in pt_pci_probe() 166 pt->io_regs = iomap_table[pt->dev_vdata->bar]; in pt_pci_probe() 191 if (pt->dev_vdata) in pt_pci_probe() 210 if (!pt) in pt_pci_remove() [all …]
|
| H A D | ptdma-dmaengine.c | 92 struct pt_device *pt = desc->pt; in pt_do_cleanup() local 171 pt = pt_cmd->pt; in pt_dma_start_desc() 204 pt = chan->pt; in pt_handle_active_desc() 276 pt = chan->pt; in pt_cmd_callback() 328 desc->pt = chan->pt; in pt_alloc_dma_desc() 388 struct pt_device *pt = chan->pt; in pt_create_desc() local 399 pt_cmd->pt = pt; in pt_create_desc() 456 pt = chan->pt; in pt_issue_pending() 489 struct pt_device *pt = chan->pt; in pt_tx_status() local 639 chan->pt = pt; in pt_dmaengine_register() [all …]
|
| H A D | ptdma-dev.c | 166 struct pt_device *pt = data; in pt_core_irq_handler() local 170 pt->total_interrupts++; in pt_core_irq_handler() 181 struct device *dev = pt->dev; in pt_core_init() 201 cmd_q->pt = pt; in pt_core_init() 234 ret = request_irq(pt->pt_irq, pt_core_irq_handler, 0, dev_name(pt->dev), pt); in pt_core_init() 261 ptdma_debugfs_setup(pt); in pt_core_init() 266 free_irq(pt->pt_irq, pt); in pt_core_init() 279 struct device *dev = pt->dev; in pt_core_destroy() 284 pt_dmaengine_unregister(pt); in pt_core_destroy() 297 free_irq(pt->pt_irq, pt); in pt_core_destroy() [all …]
|
| H A D | ptdma-debugfs.c | 26 struct pt_device *pt = s->private; in pt_debugfs_info_show() local 32 if (pt->ver == AE4_DMA_VERSION) { in pt_debugfs_info_show() 33 ae4 = container_of(pt, struct ae4_device, pt); in pt_debugfs_info_show() 57 struct pt_device *pt = s->private; in pt_debugfs_stats_show() local 67 struct pt_device *pt; in pt_debugfs_queue_show() local 75 pt = cmd_q->pt; in pt_debugfs_queue_show() 76 if (pt->ver == AE4_DMA_VERSION) { in pt_debugfs_queue_show() 113 debugfs_create_file("info", 0400, pt->dma_dev.dbg_dev_root, pt, in ptdma_debugfs_setup() 116 debugfs_create_file("stats", 0400, pt->dma_dev.dbg_dev_root, pt, in ptdma_debugfs_setup() 121 ae4 = container_of(pt, struct ae4_device, pt); in ptdma_debugfs_setup() [all …]
|
| /linux-6.15/sound/i2c/other/ |
| H A D | pt2258.c | 39 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 45 pt->mute = 1; in snd_pt2258_reset() 47 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 54 pt->volume[i] = 0; in snd_pt2258_reset() 57 snd_i2c_lock(pt->i2c_bus); in snd_pt2258_reset() 104 if (val0 == pt->volume[base] && val1 == pt->volume[base + 1]) in pt2258_stereo_volume_put() 107 pt->volume[base] = val0; in pt2258_stereo_volume_put() 150 if (pt->mute == val) in pt2258_switch_put() 153 pt->mute = val; in pt2258_switch_put() 193 err = snd_ctl_add(pt->card, snd_ctl_new1(&knew, pt)); in snd_pt2258_build_controls() [all …]
|
| /linux-6.15/drivers/gpu/drm/i915/gt/ |
| H A D | intel_ppgtt.c | 20 pt = kmalloc(sizeof(*pt), I915_GFP_ALLOW_FAIL); in alloc_pt() 26 kfree(pt); in alloc_pt() 32 return pt; in alloc_pt() 77 container_of(pt, typeof(*pd), pt); in free_px() 252 pt->stash = stash->pt[0]; in i915_vm_alloc_pt_stash() 253 stash->pt[0] = pt; in i915_vm_alloc_pt_stash() 268 pd->pt.stash = stash->pt[1]; in i915_vm_alloc_pt_stash() 269 stash->pt[1] = &pd->pt; in i915_vm_alloc_pt_stash() 283 for (pt = stash->pt[n]; pt; pt = pt->stash) { in i915_vm_map_pt_stash() 300 while ((pt = stash->pt[n])) { in i915_vm_free_pt_stash() [all …]
|
| H A D | gen6_ppgtt.c | 23 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde() 103 vaddr = px_vaddr(pt); in gen6_ppgtt_clear_range() 153 struct i915_page_table *pt; in gen6_flush_pd() local 187 if (!pt) { in gen6_alloc_va_range() 190 pt = stash->pt[0]; in gen6_alloc_va_range() 197 stash->pt[0] = pt->stash; in gen6_alloc_va_range() 199 pd->entry[pde] = pt; in gen6_alloc_va_range() 201 pt = pd->entry[pde]; in gen6_alloc_va_range() 263 if (pt) in gen6_ppgtt_free_pd() 309 if (!pt || atomic_read(&pt->used)) in pd_vma_unbind() [all …]
|
| /linux-6.15/tools/testing/selftests/bpf/progs/ |
| H A D | kfunc_call_fail.c | 41 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_rdonly() local 47 if (pt) { in kfunc_call_test_get_mem_fail_rdonly() 54 bpf_kfunc_call_test_release(pt); in kfunc_call_test_get_mem_fail_rdonly() 62 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_use_after_free() local 68 if (pt) { in kfunc_call_test_get_mem_fail_use_after_free() 88 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_oob() local 94 if (pt) { in kfunc_call_test_get_mem_fail_oob() 111 struct prog_test_ref_kfunc *pt; in kfunc_call_test_get_mem_fail_not_const() local 117 if (pt) { in kfunc_call_test_get_mem_fail_not_const() 132 struct prog_test_ref_kfunc *pt; in kfunc_call_test_mem_acquire_fail() local [all …]
|
| /linux-6.15/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
| H A D | base.c | 44 const int slot = pt->base >> pt->ptp->shift; in nvkm_mmu_ptp_put() 61 kfree(pt); in nvkm_mmu_ptp_put() 71 if (!(pt = kzalloc(sizeof(*pt), GFP_KERNEL))) in nvkm_mmu_ptp_get() 106 pt->memory = pt->ptp->pt->memory; in nvkm_mmu_ptp_get() 108 pt->addr = pt->ptp->pt->addr + pt->base; in nvkm_mmu_ptp_get() 109 return pt; in nvkm_mmu_ptp_get() 144 if (pt) { in nvkm_mmu_ptc_put() 156 list_add_tail(&pt->head, &pt->ptc->item); in nvkm_mmu_ptc_put() 191 if (pt) { in nvkm_mmu_ptc_get() 202 if (!(pt = kmalloc(sizeof(*pt), GFP_KERNEL))) in nvkm_mmu_ptc_get() [all …]
|
| H A D | vmmnv44.c | 33 tmp[0] = nvkm_ro32(pt->memory, pteo + 0x0); in nv44_vmm_pgt_fill() 34 tmp[1] = nvkm_ro32(pt->memory, pteo + 0x4); in nv44_vmm_pgt_fill() 35 tmp[2] = nvkm_ro32(pt->memory, pteo + 0x8); in nv44_vmm_pgt_fill() 36 tmp[3] = nvkm_ro32(pt->memory, pteo + 0xc); in nv44_vmm_pgt_fill() 66 VMM_WO032(pt, vmm, pteo + 0x0, tmp[0]); in nv44_vmm_pgt_fill() 67 VMM_WO032(pt, vmm, pteo + 0x4, tmp[1]); in nv44_vmm_pgt_fill() 68 VMM_WO032(pt, vmm, pteo + 0x8, tmp[2]); in nv44_vmm_pgt_fill() 116 nvkm_kmap(pt->memory); in nv44_vmm_pgt_dma() 140 nvkm_done(pt->memory); in nv44_vmm_pgt_dma() 150 nvkm_kmap(pt->memory); in nv44_vmm_pgt_unmap() [all …]
|
| H A D | vmmgp100.c | 40 nvkm_kmap(pt->memory); in gp100_vmm_pfn_unmap() 42 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_unmap() 43 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_unmap() 61 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 0); in gp100_vmm_pfn_clear() 62 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 8 + 4); in gp100_vmm_pfn_clear() 242 if (pgt->pt[0] && !gp100_vmm_pde(pgt->pt[0], &data[0])) in gp100_vmm_pd0_pde() 244 if (pgt->pt[1] && !gp100_vmm_pde(pgt->pt[1], &data[1])) in gp100_vmm_pd0_pde() 276 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0); in gp100_vmm_pd0_pfn_unmap() 277 u32 datahi = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 4); in gp100_vmm_pd0_pfn_unmap() 297 u32 datalo = nvkm_ro32(pt->memory, pt->base + ptei * 16 + 0); in gp100_vmm_pd0_pfn_clear() [all …]
|
| /linux-6.15/arch/sparc/lib/ |
| H A D | Memcpy_utils.S | 20 ba,pt %xcc, __restore_asi 24 ba,pt %xcc, __restore_asi 28 ba,pt %xcc, __restore_asi 32 ba,pt %xcc, __restore_asi 36 ba,pt %xcc, __restore_asi 40 ba,pt %xcc, __restore_asi 44 ba,pt %xcc, __restore_asi 48 ba,pt %xcc, __restore_asi 52 ba,pt %xcc, __restore_asi 57 ba,pt %xcc, __restore_asi [all …]
|
| /linux-6.15/drivers/gpu/drm/i915/selftests/ |
| H A D | scatterlist.c | 50 pfn = pt->start; in expect_pfn_sg() 51 for_each_sg(pt->st.sgl, sg, pt->st.nents, n) { in expect_pfn_sg() 72 if (pfn != pt->end) { in expect_pfn_sg() 88 pfn = pt->start; in expect_pfn_sg_page_iter() 89 for_each_sg_page(pt->st.sgl, &sgiter, pt->st.nents, 0) { in expect_pfn_sg_page_iter() 120 pfn = pt->start; in expect_pfn_sgtiter() 239 pfn = pt->start; in alloc_table() 240 sg = pt->st.sgl; in alloc_table() 263 pt->st.nents = n; in alloc_table() 264 pt->end = pfn; in alloc_table() [all …]
|
| /linux-6.15/drivers/hv/ |
| H A D | mshv_eventfd.c | 228 synchronize_srcu(&pt->pt_irq_srcu); in mshv_irqfd_resampler_shutdown() 360 mshv_irqfd_update(pt, irqfd); in mshv_irqfd_routing_update() 388 irqfd->irqfd_partn = pt; in mshv_irqfd_assign() 437 rp->rsmplr_partn = pt; in mshv_irqfd_assign() 485 mshv_irqfd_update(pt, irqfd); in mshv_irqfd_assign() 642 __must_hold(&pt->mutex) in ioeventfd_check_collision() 658 __must_hold(&pt->mutex) in mshv_assign_ioeventfd() 750 __must_hold(&pt->mutex) in mshv_deassign_ioeventfd() 779 ioeventfd_release(p, pt->pt_id); in mshv_deassign_ioeventfd() 791 __must_hold(&pt->mutex) in mshv_set_unset_ioeventfd() [all …]
|
| /linux-6.15/drivers/gpu/drm/xe/ |
| H A D | xe_lmtt.c | 63 pt = kzalloc(struct_size(pt, entries, num_entries), GFP_KERNEL); in lmtt_pt_alloc() 64 if (!pt) { in lmtt_pt_alloc() 83 pt->bo = bo; in lmtt_pt_alloc() 84 return pt; in lmtt_pt_alloc() 87 kfree(pt); in lmtt_pt_alloc() 95 kfree(pt); in lmtt_pt_free() 248 if (!pt) in lmtt_destroy_pt() 264 if (!pt) in lmtt_drop_pages() 333 if (IS_ERR(pt)) in lmtt_alloc_range() 369 pt = pt->entries[lmtt->ops->lmtt_pte_index(addr, pt->level)]; in lmtt_leaf_pt() [all …]
|
| /linux-6.15/drivers/dma-buf/ |
| H A D | sw_sync.c | 160 list_del(&pt->link); in timeline_fence_release() 268 struct sync_pt *pt; in sync_pt_create() local 270 pt = kzalloc(sizeof(*pt), GFP_KERNEL); in sync_pt_create() 271 if (!pt) in sync_pt_create() 298 kfree(pt); in sync_pt_create() 299 pt = other; in sync_pt_create() 315 return pt; in sync_pt_create() 364 struct sync_pt *pt; in sw_sync_ioctl_create_fence() local 377 if (!pt) { in sw_sync_ioctl_create_fence() 427 struct sync_pt *pt; in sw_sync_ioctl_get_deadline() local [all …]
|
| /linux-6.15/drivers/gpu/drm/gma500/ |
| H A D | mmu.c | 227 kfree(pt); in psb_mmu_free_pt() 249 if (pt) in psb_mmu_free_pagedir() 272 if (!pt) in psb_mmu_alloc_pt() 304 return pt; in psb_mmu_alloc_pt() 320 if (!pt) in psb_mmu_pt_alloc_map_lock() 343 pt->v = kmap_atomic(pt->p); in psb_mmu_pt_alloc_map_lock() 344 return pt; in psb_mmu_pt_alloc_map_lock() 356 if (!pt) { in psb_mmu_pt_map_lock() 360 pt->v = kmap_atomic(pt->p); in psb_mmu_pt_map_lock() 361 return pt; in psb_mmu_pt_map_lock() [all …]
|
| /linux-6.15/lib/ |
| H A D | nlattr.c | 123 (pt->min < 0 || pt->max < 0)); in nla_get_range_unsigned() 127 switch (pt->type) { in nla_get_range_unsigned() 413 if (nla_attr_len[pt->type] && attrlen != nla_attr_len[pt->type]) { in validate_nla() 424 if ((pt->type == NLA_NESTED || pt->type == NLA_NESTED_ARRAY) && in validate_nla() 430 if (pt->type != NLA_NESTED && pt->type != NLA_NESTED_ARRAY && in validate_nla() 472 if (pt->len) in validate_nla() 487 if (pt->len) { in validate_nla() 499 if (pt->len && attrlen > pt->len) in validate_nla() 513 pt->len, pt->nested_policy, in validate_nla() 537 pt->len, pt->nested_policy, in validate_nla() [all …]
|
| /linux-6.15/net/netlink/ |
| H A D | policy.c | 215 switch (pt->type) { in netlink_policy_dump_attr_size_estimate() 267 switch (pt->type) { in __netlink_policy_dump_write_attr() 280 if (state && pt->nested_policy && pt->len && in __netlink_policy_dump_write_attr() 284 pt->len)) || in __netlink_policy_dump_write_attr() 286 pt->len))) in __netlink_policy_dump_write_attr() 297 if (pt->type == NLA_U8) in __netlink_policy_dump_write_attr() 310 pt->mask, in __netlink_policy_dump_write_attr() 332 if (pt->type == NLA_S8) in __netlink_policy_dump_write_attr() 355 pt->bitfield32_valid)) in __netlink_policy_dump_write_attr() 383 } else if (pt->len && in __netlink_policy_dump_write_attr() [all …]
|
| /linux-6.15/arch/alpha/kernel/ |
| H A D | process.c | 291 dest[ 0] = pt->r0; in dump_elf_thread() 292 dest[ 1] = pt->r1; in dump_elf_thread() 293 dest[ 2] = pt->r2; in dump_elf_thread() 294 dest[ 3] = pt->r3; in dump_elf_thread() 295 dest[ 4] = pt->r4; in dump_elf_thread() 296 dest[ 5] = pt->r5; in dump_elf_thread() 297 dest[ 6] = pt->r6; in dump_elf_thread() 298 dest[ 7] = pt->r7; in dump_elf_thread() 299 dest[ 8] = pt->r8; in dump_elf_thread() 320 dest[29] = pt->gp; in dump_elf_thread() [all …]
|
| /linux-6.15/arch/x86/events/intel/ |
| H A D | pt.c | 419 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_start() local 465 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_filters() local 505 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config() local 563 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_stop() local 646 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_config_buffer() local 984 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_read_offset() local 1505 struct pt *pt = this_cpu_ptr(&pt_ctx); in intel_pt_interrupt() local 1560 struct pt *pt = this_cpu_ptr(&pt_ctx); in intel_pt_handle_vmx() local 1601 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_start() local 1649 struct pt *pt = this_cpu_ptr(&pt_ctx); in pt_event_stop() local [all …]
|
| /linux-6.15/net/core/ |
| H A D | net-procfs.c | 183 return pt; in ptype_get_idx() 190 return pt; in ptype_get_idx() 196 return pt; in ptype_get_idx() 203 return pt; in ptype_get_idx() 221 struct packet_type *pt; in ptype_seq_next() local 229 pt = v; in ptype_seq_next() 230 nxt = pt->list.next; in ptype_seq_next() 231 if (pt->dev) { in ptype_seq_next() 235 dev = pt->dev; in ptype_seq_next() 285 (!pt->dev || net_eq(dev_net(pt->dev), seq_file_net(seq)))) { in ptype_seq_show() [all …]
|
| /linux-6.15/arch/x86/kernel/ |
| H A D | vm86_32.c | 156 regs->pt.ax = retval; in save_v86_state() 256 vm86regs.pt.bx = v.regs.ebx; in do_sys_vm86() 257 vm86regs.pt.cx = v.regs.ecx; in do_sys_vm86() 258 vm86regs.pt.dx = v.regs.edx; in do_sys_vm86() 259 vm86regs.pt.si = v.regs.esi; in do_sys_vm86() 260 vm86regs.pt.di = v.regs.edi; in do_sys_vm86() 261 vm86regs.pt.bp = v.regs.ebp; in do_sys_vm86() 264 vm86regs.pt.cs = v.regs.cs; in do_sys_vm86() 267 vm86regs.pt.ss = v.regs.ss; in do_sys_vm86() 501 if (regs->pt.cs == BIOSSEG) in do_int() [all …]
|
| /linux-6.15/drivers/dma/amd/ae4dma/ |
| H A D | ae4dma-dev.c | 53 struct pt_device *pt; in ae4_core_irq_handler() local 57 pt = cmd_q->pt; in ae4_core_irq_handler() 59 pt->total_interrupts++; in ae4_core_irq_handler() 91 struct pt_device *pt = &ae4->pt; in ae4_core_init() local 93 struct device *dev = pt->dev; in ae4_core_init() 97 writel(max_hw_q, pt->io_regs); in ae4_core_init() 105 cmd_q->pt = pt; in ae4_core_init() 107 cmd_q->reg_control = pt->io_regs + ((i + 1) * AE4_Q_SZ); in ae4_core_init() 110 dev_name(pt->dev), ae4cmd_q); in ae4_core_init() 150 ret = pt_dmaengine_register(pt); in ae4_core_init() [all …]
|