Home
last modified time | relevance | path

Searched refs:work (Results 1 – 25 of 3985) sorted by relevance

12345678910>>...160

/linux-6.15/fs/smb/server/
H A Dksmbd_work.c23 if (work) { in ksmbd_alloc_work_struct()
31 work->iov = kcalloc(work->iov_alloc_cnt, sizeof(struct kvec), in ksmbd_alloc_work_struct()
35 work = NULL; in ksmbd_alloc_work_struct()
38 return work; in ksmbd_alloc_work_struct()
60 ksmbd_release_id(&work->conn->async_ida, work->async_id); in ksmbd_free_work_struct()
95 return queue_work(ksmbd_wq, &work->work); in ksmbd_queue_work()
101 work->iov[++work->iov_idx].iov_base = ib; in __ksmbd_iov_pin()
102 work->iov[work->iov_idx].iov_len = ib_len; in __ksmbd_iov_pin()
119 if (work->iov_alloc_cnt < work->iov_cnt + need_iov_cnt) { in __ksmbd_iov_pin_rsp()
136 work->iov[work->iov_idx].iov_base = work->response_buf; in __ksmbd_iov_pin_rsp()
[all …]
H A Dserver.c139 if (work->sess && conn->ops->is_sign_req(work, command)) { in __process_request()
147 ret = cmds->proc(work); in __process_request()
234 if (work->sess && in __handle_ksmbd_work()
235 (work->sess->sign || smb3_11_final_sess_setup_resp(work) || in __handle_ksmbd_work()
241 if (work->tcon) in __handle_ksmbd_work()
244 if (work->sess && work->sess->enc && work->encrypted && in __handle_ksmbd_work()
250 if (work->sess) in __handle_ksmbd_work()
264 struct ksmbd_work *work = container_of(wk, struct ksmbd_work, work); in handle_ksmbd_work() local
293 if (!work) { in queue_ksmbd_work()
298 work->conn = conn; in queue_ksmbd_work()
[all …]
H A Dsmb2pdu.h438 bool is_smb2_rsp(struct ksmbd_work *work);
481 int smb2_open(struct ksmbd_work *work);
484 int smb2_close(struct ksmbd_work *work);
485 int smb2_echo(struct ksmbd_work *work);
486 int smb2_set_info(struct ksmbd_work *work);
487 int smb2_read(struct ksmbd_work *work);
488 int smb2_write(struct ksmbd_work *work);
489 int smb2_flush(struct ksmbd_work *work);
490 int smb2_cancel(struct ksmbd_work *work);
491 int smb2_lock(struct ksmbd_work *work);
[all …]
H A Dksmbd_work.h86 struct work_struct work; member
98 static inline void *ksmbd_resp_buf_next(struct ksmbd_work *work) in ksmbd_resp_buf_next() argument
100 return work->response_buf + work->next_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_next()
107 static inline void *ksmbd_resp_buf_curr(struct ksmbd_work *work) in ksmbd_resp_buf_curr() argument
109 return work->response_buf + work->curr_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_curr()
116 static inline void *ksmbd_req_buf_next(struct ksmbd_work *work) in ksmbd_req_buf_next() argument
118 return work->request_buf + work->next_smb2_rcv_hdr_off + 4; in ksmbd_req_buf_next()
122 void ksmbd_free_work_struct(struct ksmbd_work *work);
129 bool ksmbd_queue_work(struct ksmbd_work *work);
132 int ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len);
[all …]
H A Dsmb2pdu.c128 work->tcon = ksmbd_tree_conn_lookup(work->sess, tree_id); in smb2_get_ksmbd_tcon()
402 work->iov[work->iov_idx].iov_len += (new_len - len); in init_chained_smb2_rsp()
407 work->curr_smb2_rsp_hdr_off = work->next_smb2_rsp_hdr_off; in init_chained_smb2_rsp()
483 work->iov[work->iov_idx].iov_len += len; in is_chained_smb2_message()
486 work->curr_smb2_rsp_hdr_off = work->next_smb2_rsp_hdr_off; in is_chained_smb2_message()
4442 query_dir_private.work = work; in smb2_query_dir()
8836 iov = &work->iov[work->iov_idx - 1]; in smb2_set_sign_rsp()
8839 iov = &work->iov[work->iov_idx]; in smb2_set_sign_rsp()
8928 chann = lookup_chann_list(work->sess, work->conn); in smb3_set_sign_rsp()
8942 iov = &work->iov[work->iov_idx - 1]; in smb3_set_sign_rsp()
[all …]
H A Dsmb_common.c143 hdr = work->request_buf; in ksmbd_verify_smb_message()
146 work->conn->outstanding_credits++; in ksmbd_verify_smb_message()
344 unsigned int cmd = work->conn->ops->get_cmd_val(work); in smb1_check_user_session()
364 if (!work->response_buf) { in smb1_allocate_rsp_buf()
380 work->send_no_response = 1; in set_smb1_rsp_status()
600 ret = smb2_handle_negotiate(work); in ksmbd_smb_negotiate_common()
607 init_smb2_neg_rsp(work); in ksmbd_smb_negotiate_common()
782 WARN_ON(work->saved_cred); in __ksmbd_override_fsids()
789 return __ksmbd_override_fsids(work, work->tcon->share_conf); in ksmbd_override_fsids()
795 WARN_ON(!work->saved_cred); in ksmbd_revert_fsids()
[all …]
/linux-6.15/virt/kvm/
H A Dasync_pf.c48 container_of(work, struct kvm_async_pf, work); in async_pf_execute()
114 WARN_ON_ONCE(work->work.func); in kvm_flush_and_free_async_pf_work()
116 flush_work(&work->work); in kvm_flush_and_free_async_pf_work()
130 flush_work(&work->work); in kvm_clear_async_pf_completion_queue()
132 if (cancel_work_sync(&work->work)) in kvm_clear_async_pf_completion_queue()
160 work = list_first_entry(&vcpu->async_pf.done, typeof(*work), in kvm_check_async_pf_completion()
196 if (!work) in kvm_setup_async_pf()
202 work->addr = hva; in kvm_setup_async_pf()
205 INIT_WORK(&work->work, async_pf_execute); in kvm_setup_async_pf()
211 schedule_work(&work->work); in kvm_setup_async_pf()
[all …]
/linux-6.15/drivers/gpu/drm/
H A Ddrm_flip_work.c73 work->func(work, val); in drm_flip_work_queue()
94 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit()
95 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit()
97 queue_work(wq, &work->worker); in drm_flip_work_commit()
120 work->func(work, task->data); in flip_worker()
137 work->name = name; in drm_flip_work_init()
138 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init()
139 INIT_LIST_HEAD(&work->commited); in drm_flip_work_init()
140 spin_lock_init(&work->lock); in drm_flip_work_init()
141 work->func = func; in drm_flip_work_init()
[all …]
H A Ddrm_vblank_work.c58 list_del_init(&work->node); in drm_handle_vblank_works()
80 list_del_init(&work->node); in drm_vblank_cancel_pending_works()
120 if (work->cancelling) in drm_vblank_work_schedule()
129 if (list_empty(&work->node)) { in drm_vblank_work_schedule()
140 work->count = count; in drm_vblank_work_schedule()
153 list_del_init(&work->node); in drm_vblank_work_schedule()
192 if (!list_empty(&work->node)) { in drm_vblank_work_cancel_sync()
193 list_del_init(&work->node); in drm_vblank_work_cancel_sync()
198 work->cancelling++; in drm_vblank_work_cancel_sync()
207 work->cancelling--; in drm_vblank_work_cancel_sync()
[all …]
/linux-6.15/include/trace/events/
H A Dworkqueue.h31 __field( void *, work )
39 __entry->work = work;
63 TP_ARGS(work),
66 __field( void *, work )
71 __entry->work = work;
88 TP_ARGS(work),
91 __field( void *, work )
96 __entry->work = work;
114 TP_ARGS(work, function),
117 __field( void *, work )
[all …]
/linux-6.15/kernel/
H A Dirq_work.c112 irq_work_raise(work); in __irq_work_queue_local()
119 if (!irq_work_claim(work)) in irq_work_queue()
124 __irq_work_queue_local(work); in irq_work_queue()
140 return irq_work_queue(work); in irq_work_queue_on()
147 if (!irq_work_claim(work)) in irq_work_queue_on()
169 if (!irq_work_claim(work)) in irq_work_queue_on()
203 struct irq_work *work = arg; in irq_work_single() local
221 work->func(work); in irq_work_single()
237 struct irq_work *work, *tmp; in irq_work_run_list() local
252 irq_work_single(work); in irq_work_run_list()
[all …]
H A Dtask_work.c72 work->next = head; in task_work_add()
128 while (work) { in task_work_cancel_match()
130 pprev = &work->next; in task_work_cancel_match()
132 } else if (try_cmpxchg(pprev, &work, work->next)) in task_work_cancel_match()
137 return work; in task_work_cancel_match()
207 if (!work) { in task_work_run()
215 if (!work) in task_work_run()
226 next = work->next; in task_work_run()
227 work->func(work); in task_work_run()
228 work = next; in task_work_run()
[all …]
H A Dkthread.c996 work = NULL; in kthread_worker_fn()
1006 if (work) { in kthread_worker_fn()
1010 work->func(work); in kthread_worker_fn()
1145 return !list_empty(&work->node) || work->canceling; in queuing_blocked()
1154 WARN_ON_ONCE(work->worker && work->worker != worker); in kthread_insert_work_sanity_check()
1211 struct kthread_work *work = &dwork->work; in kthread_delayed_work_timer_fn() local
1241 struct kthread_work *work = &dwork->work; in __kthread_queue_delayed_work() local
1284 struct kthread_work *work = &dwork->work; in kthread_queue_delayed_work() local
1308 container_of(work, struct kthread_flush_work, work); in kthread_flush_work_fn()
1361 container_of(work, struct kthread_delayed_work, work); in kthread_cancel_delayed_work_timer()
[all …]
/linux-6.15/include/linux/
H A Dcompletion.h35 #define COMPLETION_INITIALIZER(work) \ argument
39 (*({ init_completion_map(&(work), &(map)); &(work); }))
41 #define COMPLETION_INITIALIZER_ONSTACK(work) \ argument
42 (*({ init_completion(&work); &work; }))
52 #define DECLARE_COMPLETION(work) \ argument
53 struct completion work = COMPLETION_INITIALIZER(work)
68 # define DECLARE_COMPLETION_ONSTACK(work) \ argument
69 struct completion work = COMPLETION_INITIALIZER_ONSTACK(work)
71 struct completion work = COMPLETION_INITIALIZER_ONSTACK_MAP(work, map)
73 # define DECLARE_COMPLETION_ONSTACK(work) DECLARE_COMPLETION(work) argument
[all …]
H A Dworkqueue.h23 #define work_data_bits(work) ((unsigned long *)(&(work)->data)) argument
114 struct work_struct work; member
123 struct work_struct work; member
213 return container_of(work, struct delayed_work, work); in to_delayed_work()
218 return container_of(work, struct rcu_work, work); in to_rcu_work()
222 struct work_struct work; member
245 .work = __WORK_INITIALIZER((n).work, (f)), \
354 #define work_pending(work) \ argument
363 work_pending(&(w)->work)
745 if (enable_work(work)) { in enable_and_queue_work()
[all …]
H A Djump_label_ratelimit.h12 struct delayed_work work; member
18 struct delayed_work work; member
24 struct delayed_work work; member
28 __static_key_slow_dec_deferred(&(x)->key, &(x)->work, (x)->timeout)
30 __static_key_slow_dec_deferred(&(x)->key.key, &(x)->work, (x)->timeout)
33 __static_key_deferred_flush((x), &(x)->work)
37 struct delayed_work *work,
39 extern void __static_key_deferred_flush(void *key, struct delayed_work *work);
43 extern void jump_label_update_timeout(struct work_struct *work);
49 .work = __DELAYED_WORK_INITIALIZER((name).work, \
[all …]
H A Dirq_work.h37 void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) in init_irq_work() argument
39 *work = IRQ_WORK_INIT(func); in init_irq_work()
42 static inline bool irq_work_is_pending(struct irq_work *work) in irq_work_is_pending() argument
44 return atomic_read(&work->node.a_flags) & IRQ_WORK_PENDING; in irq_work_is_pending()
47 static inline bool irq_work_is_busy(struct irq_work *work) in irq_work_is_busy() argument
49 return atomic_read(&work->node.a_flags) & IRQ_WORK_BUSY; in irq_work_is_busy()
52 static inline bool irq_work_is_hard(struct irq_work *work) in irq_work_is_hard() argument
54 return atomic_read(&work->node.a_flags) & IRQ_WORK_HARD_IRQ; in irq_work_is_hard()
57 bool irq_work_queue(struct irq_work *work);
58 bool irq_work_queue_on(struct irq_work *work, int cpu);
[all …]
/linux-6.15/fs/btrfs/
H A Dasync-thread.c57 return work->wq->fs_info; in btrfs_work_owner()
213 struct btrfs_work *work; in run_ordered_work() local
244 work->ordered_func(work, false); in run_ordered_work()
251 if (work == self) { in run_ordered_work()
279 work->ordered_func(work, true); in run_ordered_work()
308 if (work->ordered_func) in btrfs_work_helper()
313 work->func(work); in btrfs_work_helper()
333 work->func = func; in btrfs_init_work()
337 work->flags = 0; in btrfs_init_work()
344 work->wq = wq; in btrfs_queue_work()
[all …]
/linux-6.15/drivers/staging/octeon/
H A Dethernet-rx.c72 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64)) in cvm_oct_check_rcv_error()
112 work->word1.len -= i + 5; in cvm_oct_check_rcv_error()
119 work->word1.len -= i + 4; in cvm_oct_check_rcv_error()
131 cvm_oct_free_work(work); in cvm_oct_check_rcv_error()
138 cvm_oct_free_work(work); in cvm_oct_check_rcv_error()
146 int len = work->word1.len; in copy_segments_to_skb()
222 struct cvmx_wqe *work; in cvm_oct_poll() local
230 prefetch(work); in cvm_oct_poll()
232 if (!work) { in cvm_oct_poll()
299 cvm_oct_free_work(work); in cvm_oct_poll()
[all …]
H A Dethernet-tx.c516 if (unlikely(!work)) { in cvm_oct_xmit_pow()
560 work->word1.len = skb->len; in cvm_oct_xmit_pow()
567 work->word2.u64 = 0; in cvm_oct_xmit_pow()
568 work->word2.s.bufs = 1; in cvm_oct_xmit_pow()
569 work->packet_ptr.u64 = 0; in cvm_oct_xmit_pow()
590 work->word2.s.is_v6 = 0; in cvm_oct_xmit_pow()
601 work->word2.s.IP_exc = 0; in cvm_oct_xmit_pow()
607 work->word2.s.not_IP = 0; in cvm_oct_xmit_pow()
641 memcpy(work->packet_data, skb->data, sizeof(work->packet_data)); in cvm_oct_xmit_pow()
645 cvmx_pow_work_submit(work, work->word1.tag, work->word1.tag_type, in cvm_oct_xmit_pow()
[all …]
/linux-6.15/tools/perf/
H A Dbuiltin-kwork.c362 struct kwork_work *work = zalloc(sizeof(*work)); in work_new() local
389 if (work) in work_findnew()
769 if (work->class && work->class->work_name) { in timehist_print_event()
770 work->class->work_name(work, kwork_name, in timehist_print_event()
921 if (!work) in top_exit_event()
928 work->id, work->cpu); in top_exit_event()
950 if (!work) in top_sched_switch_event()
1360 if (work->class && work->class->work_name) { in report_print_work()
1361 work->class->work_name(work, kwork_name, in report_print_work()
2058 work->id, work->cpu); in top_subtract_irq_runtime()
[all …]
/linux-6.15/net/wireless/
H A Ddebugfs.c115 struct wiphy_work work; member
133 struct debugfs_read_work *w = container_of(work, typeof(*w), work); in wiphy_locked_debugfs_read_work()
171 .cancel_data = &work, in wiphy_locked_debugfs_read()
177 wiphy_work_init(&work.work, wiphy_locked_debugfs_read_work); in wiphy_locked_debugfs_read()
178 wiphy_work_queue(wiphy, &work.work); in wiphy_locked_debugfs_read()
184 if (work.ret < 0) in wiphy_locked_debugfs_read()
185 return work.ret; in wiphy_locked_debugfs_read()
213 struct debugfs_write_work *w = container_of(work, typeof(*w), work); in wiphy_locked_debugfs_write_work()
262 wiphy_work_init(&work.work, wiphy_locked_debugfs_write_work); in wiphy_locked_debugfs_write()
263 wiphy_work_queue(wiphy, &work.work); in wiphy_locked_debugfs_write()
[all …]
/linux-6.15/drivers/accessibility/speakup/
H A Dselection.c20 struct work_struct work; member
25 static void __speakup_set_selection(struct work_struct *work) in __speakup_set_selection() argument
28 container_of(work, struct speakup_selection_work, work); in __speakup_set_selection()
58 .work = __WORK_INITIALIZER(speakup_sel_work.work,
87 schedule_work_on(WORK_CPU_UNBOUND, &speakup_sel_work.work); in speakup_set_selection()
96 cancel_work_sync(&speakup_sel_work.work); in speakup_cancel_selection()
106 static void __speakup_paste_selection(struct work_struct *work) in __speakup_paste_selection() argument
109 container_of(work, struct speakup_selection_work, work); in __speakup_paste_selection()
117 .work = __WORK_INITIALIZER(speakup_paste_work.work,
129 schedule_work_on(WORK_CPU_UNBOUND, &speakup_paste_work.work); in speakup_paste_selection()
[all …]
/linux-6.15/drivers/net/wireless/st/cw1200/
H A Dsta.h59 void cw1200_event_handler(struct work_struct *work);
60 void cw1200_bss_loss_work(struct work_struct *work);
61 void cw1200_bss_params_work(struct work_struct *work);
62 void cw1200_keep_alive_work(struct work_struct *work);
63 void cw1200_tx_failure_work(struct work_struct *work);
79 void cw1200_join_timeout(struct work_struct *work);
80 void cw1200_unjoin_work(struct work_struct *work);
82 void cw1200_wep_key_work(struct work_struct *work);
91 void cw1200_ba_work(struct work_struct *work);
114 void cw1200_set_tim_work(struct work_struct *work);
[all …]
/linux-6.15/drivers/infiniband/core/
H A Dcm.c2018 cm_issue_rej(work->port, work->mad_recv_wc, in cm_match_req()
2035 cm_issue_rej(work->port, work->mad_recv_wc, in cm_match_req()
2147 memset(&work->path[0], 0, sizeof(work->path[0])); in cm_req_handler()
2167 &work->path[1], work->mad_recv_wc->wc); in cm_req_handler()
2535 cm_issue_rej(work->port, work->mad_recv_wc, in cm_rep_handler()
2838 cm_issue_drep(work->port, work->mad_recv_wc); in cm_dreq_handler()
3336 work->port->port_num, &work->path[0], in cm_lap_handler()
3865 struct cm_work *work = container_of(_work, struct cm_work, work.work); in cm_work_handler() local
3929 work = kmalloc(sizeof *work, GFP_ATOMIC); in cm_establish()
3970 queue_delayed_work(cm.wq, &work->work, 0); in cm_establish()
[all …]

12345678910>>...160