/* * Copyright (c) 2020 Nordic Semiconductor ASA * * SPDX-License-Identifier: Apache-2.0 */ /** * @file * * Second generation work queue implementation */ #include #include #include #include #include #include #include static inline void flag_clear(uint32_t *flagp, uint32_t bit) { *flagp &= ~BIT(bit); } static inline void flag_set(uint32_t *flagp, uint32_t bit) { *flagp |= BIT(bit); } static inline bool flag_test(const uint32_t *flagp, uint32_t bit) { return (*flagp & BIT(bit)) != 0U; } static inline bool flag_test_and_clear(uint32_t *flagp, int bit) { bool ret = flag_test(flagp, bit); flag_clear(flagp, bit); return ret; } static inline void flags_set(uint32_t *flagp, uint32_t flags) { *flagp = flags; } static inline uint32_t flags_get(const uint32_t *flagp) { return *flagp; } /* Lock to protect the internal state of all work items, work queues, * and pending_cancels. */ static struct k_spinlock lock; /* Invoked by work thread */ static void handle_flush(struct k_work *work) { } static inline void init_flusher(struct z_work_flusher *flusher) { struct k_work *work = &flusher->work; k_sem_init(&flusher->sem, 0, 1); k_work_init(&flusher->work, handle_flush); flag_set(&work->flags, K_WORK_FLUSHING_BIT); } /* List of pending cancellations. */ static sys_slist_t pending_cancels; /* Initialize a canceler record and add it to the list of pending * cancels. * * Invoked with work lock held. * * @param canceler the structure used to notify a waiting process. * @param work the work structure that is to be canceled */ static inline void init_work_cancel(struct z_work_canceller *canceler, struct k_work *work) { k_sem_init(&canceler->sem, 0, 1); canceler->work = work; sys_slist_append(&pending_cancels, &canceler->node); } /* Complete flushing of a work item. * * Invoked with work lock held. * * Invoked from a work queue thread. * * Reschedules. * * @param work the work structure that has completed flushing. */ static void finalize_flush_locked(struct k_work *work) { struct z_work_flusher *flusher = CONTAINER_OF(work, struct z_work_flusher, work); flag_clear(&work->flags, K_WORK_FLUSHING_BIT); k_sem_give(&flusher->sem); }; /* Complete cancellation of a work item and unlock held lock. * * Invoked with work lock held. * * Invoked from a work queue thread. * * Reschedules. * * @param work the work structure that has completed cancellation */ static void finalize_cancel_locked(struct k_work *work) { struct z_work_canceller *wc, *tmp; sys_snode_t *prev = NULL; /* Clear this first, so released high-priority threads don't * see it when doing things. */ flag_clear(&work->flags, K_WORK_CANCELING_BIT); /* Search for and remove the matching container, and release * what's waiting for the completion. The same work item can * appear multiple times in the list if multiple threads * attempt to cancel it. */ SYS_SLIST_FOR_EACH_CONTAINER_SAFE(&pending_cancels, wc, tmp, node) { if (wc->work == work) { sys_slist_remove(&pending_cancels, prev, &wc->node); k_sem_give(&wc->sem); break; } prev = &wc->node; } } void k_work_init(struct k_work *work, k_work_handler_t handler) { __ASSERT_NO_MSG(work != NULL); __ASSERT_NO_MSG(handler != NULL); *work = (struct k_work)Z_WORK_INITIALIZER(handler); SYS_PORT_TRACING_OBJ_INIT(k_work, work); } static inline int work_busy_get_locked(const struct k_work *work) { return flags_get(&work->flags) & K_WORK_MASK; } int k_work_busy_get(const struct k_work *work) { k_spinlock_key_t key = k_spin_lock(&lock); int ret = work_busy_get_locked(work); k_spin_unlock(&lock, key); return ret; } /* Add a flusher work item to the queue. * * Invoked with work lock held. * * Caller must notify queue of pending work. * * @param queue queue on which a work item may appear. * @param work the work item that is either queued or running on @p * queue * @param flusher an uninitialized/unused flusher object */ static void queue_flusher_locked(struct k_work_q *queue, struct k_work *work, struct z_work_flusher *flusher) { bool in_list = false; struct k_work *wn; /* Determine whether the work item is still queued. */ SYS_SLIST_FOR_EACH_CONTAINER(&queue->pending, wn, node) { if (wn == work) { in_list = true; break; } } init_flusher(flusher); if (in_list) { sys_slist_insert(&queue->pending, &work->node, &flusher->work.node); } else { sys_slist_prepend(&queue->pending, &flusher->work.node); } } /* Try to remove a work item from the given queue. * * Invoked with work lock held. * * @param queue the queue from which the work should be removed * @param work work that may be on the queue */ static inline void queue_remove_locked(struct k_work_q *queue, struct k_work *work) { if (flag_test_and_clear(&work->flags, K_WORK_QUEUED_BIT)) { (void)sys_slist_find_and_remove(&queue->pending, &work->node); } } /* Potentially notify a queue that it needs to look for pending work. * * This may make the work queue thread ready, but as the lock is held it * will not be a reschedule point. Callers should yield after the lock is * released where appropriate (generally if this returns true). * * @param queue to be notified. If this is null no notification is required. * * @return true if and only if the queue was notified and woken, i.e. a * reschedule is pending. */ static inline bool notify_queue_locked(struct k_work_q *queue) { bool rv = false; if (queue != NULL) { rv = z_sched_wake(&queue->notifyq, 0, NULL); } return rv; } /* Submit an work item to a queue if queue state allows new work. * * Submission is rejected if no queue is provided, or if the queue is * draining and the work isn't being submitted from the queue's * thread (chained submission). * * Invoked with work lock held. * Conditionally notifies queue. * * @param queue the queue to which work should be submitted. This may * be null, in which case the submission will fail. * * @param work to be submitted * * @retval 1 if successfully queued * @retval -EINVAL if no queue is provided * @retval -ENODEV if the queue is not started * @retval -EBUSY if the submission was rejected (draining, plugged) */ static inline int queue_submit_locked(struct k_work_q *queue, struct k_work *work) { if (queue == NULL) { return -EINVAL; } int ret; bool chained = (_current == &queue->thread) && !k_is_in_isr(); bool draining = flag_test(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); bool plugged = flag_test(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); /* Test for acceptability, in priority order: * * * -ENODEV if the queue isn't running. * * -EBUSY if draining and not chained * * -EBUSY if plugged and not draining * * otherwise OK */ if (!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)) { ret = -ENODEV; } else if (draining && !chained) { ret = -EBUSY; } else if (plugged && !draining) { ret = -EBUSY; } else { sys_slist_append(&queue->pending, &work->node); ret = 1; (void)notify_queue_locked(queue); } return ret; } /* Attempt to submit work to a queue. * * The submission can fail if: * * the work is cancelling, * * no candidate queue can be identified; * * the candidate queue rejects the submission. * * Invoked with work lock held. * Conditionally notifies queue. * * @param work the work structure to be submitted * @param queuep pointer to a queue reference. On input this should * dereference to the proposed queue (which may be null); after completion it * will be null if the work was not submitted or if submitted will reference * the queue it was submitted to. That may or may not be the queue provided * on input. * * @retval 0 if work was already submitted to a queue * @retval 1 if work was not submitted and has been queued to @p queue * @retval 2 if work was running and has been queued to the queue that was * running it * @retval -EBUSY if canceling or submission was rejected by queue * @retval -EINVAL if no queue is provided * @retval -ENODEV if the queue is not started */ static int submit_to_queue_locked(struct k_work *work, struct k_work_q **queuep) { int ret = 0; if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { /* Disallowed */ ret = -EBUSY; } else if (!flag_test(&work->flags, K_WORK_QUEUED_BIT)) { /* Not currently queued */ ret = 1; /* If no queue specified resubmit to last queue. */ if (*queuep == NULL) { *queuep = work->queue; } /* If the work is currently running we have to use the * queue it's running on to prevent handler * re-entrancy. */ if (flag_test(&work->flags, K_WORK_RUNNING_BIT)) { __ASSERT_NO_MSG(work->queue != NULL); *queuep = work->queue; ret = 2; } int rc = queue_submit_locked(*queuep, work); if (rc < 0) { ret = rc; } else { flag_set(&work->flags, K_WORK_QUEUED_BIT); work->queue = *queuep; } } else { /* Already queued, do nothing. */ } if (ret <= 0) { *queuep = NULL; } return ret; } /* Submit work to a queue but do not yield the current thread. * * Intended for internal use. * * See also submit_to_queue_locked(). * * @param queuep pointer to a queue reference. * @param work the work structure to be submitted * * @retval see submit_to_queue_locked() */ int z_work_submit_to_queue(struct k_work_q *queue, struct k_work *work) { __ASSERT_NO_MSG(work != NULL); __ASSERT_NO_MSG(work->handler != NULL); k_spinlock_key_t key = k_spin_lock(&lock); int ret = submit_to_queue_locked(work, &queue); k_spin_unlock(&lock, key); return ret; } int k_work_submit_to_queue(struct k_work_q *queue, struct k_work *work) { SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, submit_to_queue, queue, work); int ret = z_work_submit_to_queue(queue, work); /* submit_to_queue_locked() won't reschedule on its own * (really it should, otherwise this process will result in * spurious calls to z_swap() due to the race), so do it here * if the queue state changed. */ if (ret > 0) { z_reschedule_unlocked(); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, submit_to_queue, queue, work, ret); return ret; } int k_work_submit(struct k_work *work) { SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, submit, work); int ret = k_work_submit_to_queue(&k_sys_work_q, work); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, submit, work, ret); return ret; } /* Flush the work item if necessary. * * Flushing is necessary only if the work is either queued or running. * * Invoked with work lock held by key. * Sleeps. * * @param work the work item that is to be flushed * @param flusher state used to synchronize the flush * * @retval true if work is queued or running. If this happens the * caller must take the flusher semaphore after releasing the lock. * * @retval false otherwise. No wait required. */ static bool work_flush_locked(struct k_work *work, struct z_work_flusher *flusher) { bool need_flush = (flags_get(&work->flags) & (K_WORK_QUEUED | K_WORK_RUNNING)) != 0U; if (need_flush) { struct k_work_q *queue = work->queue; __ASSERT_NO_MSG(queue != NULL); queue_flusher_locked(queue, work, flusher); notify_queue_locked(queue); } return need_flush; } bool k_work_flush(struct k_work *work, struct k_work_sync *sync) { __ASSERT_NO_MSG(work != NULL); __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); __ASSERT_NO_MSG(!k_is_in_isr()); __ASSERT_NO_MSG(sync != NULL); #ifdef CONFIG_KERNEL_COHERENCE __ASSERT_NO_MSG(arch_mem_coherent(sync)); #endif /* CONFIG_KERNEL_COHERENCE */ SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, flush, work); struct z_work_flusher *flusher = &sync->flusher; k_spinlock_key_t key = k_spin_lock(&lock); bool need_flush = work_flush_locked(work, flusher); k_spin_unlock(&lock, key); /* If necessary wait until the flusher item completes */ if (need_flush) { SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_work, flush, work, K_FOREVER); k_sem_take(&flusher->sem, K_FOREVER); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush, work, need_flush); return need_flush; } /* Execute the non-waiting steps necessary to cancel a work item. * * Invoked with work lock held. * * @param work the work item to be canceled. * * @retval true if we need to wait for the work item to finish canceling * @retval false if the work item is idle * * @return k_busy_wait() captured under lock */ static int cancel_async_locked(struct k_work *work) { /* If we haven't already started canceling, do it now. */ if (!flag_test(&work->flags, K_WORK_CANCELING_BIT)) { /* Remove it from the queue, if it's queued. */ queue_remove_locked(work->queue, work); } /* If it's still busy after it's been dequeued, then flag it * as canceling. */ int ret = work_busy_get_locked(work); if (ret != 0) { flag_set(&work->flags, K_WORK_CANCELING_BIT); ret = work_busy_get_locked(work); } return ret; } /* Complete cancellation necessary, release work lock, and wait if * necessary. * * Invoked with work lock held by key. * Sleeps. * * @param work work that is being canceled * @param canceller state used to synchronize the cancellation * @param key used by work lock * * @retval true if and only if the work was still active on entry. The caller * must wait on the canceller semaphore after releasing the lock. * * @retval false if work was idle on entry. The caller need not wait. */ static bool cancel_sync_locked(struct k_work *work, struct z_work_canceller *canceller) { bool ret = flag_test(&work->flags, K_WORK_CANCELING_BIT); /* If something's still running then we have to wait for * completion, which is indicated when finish_cancel() gets * invoked. */ if (ret) { init_work_cancel(canceller, work); } return ret; } int k_work_cancel(struct k_work *work) { __ASSERT_NO_MSG(work != NULL); __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel, work); k_spinlock_key_t key = k_spin_lock(&lock); int ret = cancel_async_locked(work); k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel, work, ret); return ret; } bool k_work_cancel_sync(struct k_work *work, struct k_work_sync *sync) { __ASSERT_NO_MSG(work != NULL); __ASSERT_NO_MSG(sync != NULL); __ASSERT_NO_MSG(!flag_test(&work->flags, K_WORK_DELAYABLE_BIT)); __ASSERT_NO_MSG(!k_is_in_isr()); #ifdef CONFIG_KERNEL_COHERENCE __ASSERT_NO_MSG(arch_mem_coherent(sync)); #endif /* CONFIG_KERNEL_COHERENCE */ SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_sync, work, sync); struct z_work_canceller *canceller = &sync->canceller; k_spinlock_key_t key = k_spin_lock(&lock); bool pending = (work_busy_get_locked(work) != 0U); bool need_wait = false; if (pending) { (void)cancel_async_locked(work); need_wait = cancel_sync_locked(work, canceller); } k_spin_unlock(&lock, key); if (need_wait) { SYS_PORT_TRACING_OBJ_FUNC_BLOCKING(k_work, cancel_sync, work, sync); k_sem_take(&canceller->sem, K_FOREVER); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_sync, work, sync, pending); return pending; } /* Loop executed by a work queue thread. * * @param workq_ptr pointer to the work queue structure */ static void work_queue_main(void *workq_ptr, void *p2, void *p3) { ARG_UNUSED(p2); ARG_UNUSED(p3); struct k_work_q *queue = (struct k_work_q *)workq_ptr; while (true) { sys_snode_t *node; struct k_work *work = NULL; k_work_handler_t handler = NULL; k_spinlock_key_t key = k_spin_lock(&lock); bool yield; /* Check for and prepare any new work. */ node = sys_slist_get(&queue->pending); if (node != NULL) { /* Mark that there's some work active that's * not on the pending list. */ flag_set(&queue->flags, K_WORK_QUEUE_BUSY_BIT); work = CONTAINER_OF(node, struct k_work, node); flag_set(&work->flags, K_WORK_RUNNING_BIT); flag_clear(&work->flags, K_WORK_QUEUED_BIT); /* Static code analysis tool can raise a false-positive violation * in the line below that 'work' is checked for null after being * dereferenced. * * The work is figured out by CONTAINER_OF, as a container * of type struct k_work that contains the node. * The only way for it to be NULL is if node would be a member * of struct k_work object that has been placed at address NULL, * which should never happen, even line 'if (work != NULL)' * ensures that. * This means that if node is not NULL, then work will not be NULL. */ handler = work->handler; } else if (flag_test_and_clear(&queue->flags, K_WORK_QUEUE_DRAIN_BIT)) { /* Not busy and draining: move threads waiting for * drain to ready state. The held spinlock inhibits * immediate reschedule; released threads get their * chance when this invokes z_sched_wait() below. * * We don't touch K_WORK_QUEUE_PLUGGABLE, so getting * here doesn't mean that the queue will allow new * submissions. */ (void)z_sched_wake_all(&queue->drainq, 1, NULL); } else { /* No work is available and no queue state requires * special handling. */ ; } if (work == NULL) { /* Nothing's had a chance to add work since we took * the lock, and we didn't find work nor got asked to * stop. Just go to sleep: when something happens the * work thread will be woken and we can check again. */ (void)z_sched_wait(&lock, key, &queue->notifyq, K_FOREVER, NULL); continue; } k_spin_unlock(&lock, key); __ASSERT_NO_MSG(handler != NULL); handler(work); /* Mark the work item as no longer running and deal * with any cancellation and flushing issued while it * was running. Clear the BUSY flag and optionally * yield to prevent starving other threads. */ key = k_spin_lock(&lock); flag_clear(&work->flags, K_WORK_RUNNING_BIT); if (flag_test(&work->flags, K_WORK_FLUSHING_BIT)) { finalize_flush_locked(work); } if (flag_test(&work->flags, K_WORK_CANCELING_BIT)) { finalize_cancel_locked(work); } flag_clear(&queue->flags, K_WORK_QUEUE_BUSY_BIT); yield = !flag_test(&queue->flags, K_WORK_QUEUE_NO_YIELD_BIT); k_spin_unlock(&lock, key); /* Optionally yield to prevent the work queue from * starving other threads. */ if (yield) { k_yield(); } } } void k_work_queue_init(struct k_work_q *queue) { __ASSERT_NO_MSG(queue != NULL); *queue = (struct k_work_q) { .flags = 0, }; SYS_PORT_TRACING_OBJ_INIT(k_work_queue, queue); } void k_work_queue_start(struct k_work_q *queue, k_thread_stack_t *stack, size_t stack_size, int prio, const struct k_work_queue_config *cfg) { __ASSERT_NO_MSG(queue); __ASSERT_NO_MSG(stack); __ASSERT_NO_MSG(!flag_test(&queue->flags, K_WORK_QUEUE_STARTED_BIT)); uint32_t flags = K_WORK_QUEUE_STARTED; SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, start, queue); sys_slist_init(&queue->pending); z_waitq_init(&queue->notifyq); z_waitq_init(&queue->drainq); if ((cfg != NULL) && cfg->no_yield) { flags |= K_WORK_QUEUE_NO_YIELD; } /* It hasn't actually been started yet, but all the state is in place * so we can submit things and once the thread gets control it's ready * to roll. */ flags_set(&queue->flags, flags); (void)k_thread_create(&queue->thread, stack, stack_size, work_queue_main, queue, NULL, NULL, prio, 0, K_FOREVER); if ((cfg != NULL) && (cfg->name != NULL)) { k_thread_name_set(&queue->thread, cfg->name); } if ((cfg != NULL) && (cfg->essential)) { queue->thread.base.user_options |= K_ESSENTIAL; } k_thread_start(&queue->thread); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, start, queue); } int k_work_queue_drain(struct k_work_q *queue, bool plug) { __ASSERT_NO_MSG(queue); __ASSERT_NO_MSG(!k_is_in_isr()); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, drain, queue); int ret = 0; k_spinlock_key_t key = k_spin_lock(&lock); if (((flags_get(&queue->flags) & (K_WORK_QUEUE_BUSY | K_WORK_QUEUE_DRAIN)) != 0U) || plug || !sys_slist_is_empty(&queue->pending)) { flag_set(&queue->flags, K_WORK_QUEUE_DRAIN_BIT); if (plug) { flag_set(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT); } notify_queue_locked(queue); ret = z_sched_wait(&lock, key, &queue->drainq, K_FOREVER, NULL); } else { k_spin_unlock(&lock, key); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, drain, queue, ret); return ret; } int k_work_queue_unplug(struct k_work_q *queue) { __ASSERT_NO_MSG(queue); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work_queue, unplug, queue); int ret = -EALREADY; k_spinlock_key_t key = k_spin_lock(&lock); if (flag_test_and_clear(&queue->flags, K_WORK_QUEUE_PLUGGED_BIT)) { ret = 0; } k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work_queue, unplug, queue, ret); return ret; } #ifdef CONFIG_SYS_CLOCK_EXISTS /* Timeout handler for delayable work. * * Invoked by timeout infrastructure. * Takes and releases work lock. * Conditionally reschedules. */ static void work_timeout(struct _timeout *to) { struct k_work_delayable *dw = CONTAINER_OF(to, struct k_work_delayable, timeout); struct k_work *wp = &dw->work; k_spinlock_key_t key = k_spin_lock(&lock); struct k_work_q *queue = NULL; /* If the work is still marked delayed (should be) then clear that * state and submit it to the queue. If successful the queue will be * notified of new work at the next reschedule point. * * If not successful there is no notification that the work has been * abandoned. Sorry. */ if (flag_test_and_clear(&wp->flags, K_WORK_DELAYED_BIT)) { queue = dw->queue; (void)submit_to_queue_locked(wp, &queue); } k_spin_unlock(&lock, key); } void k_work_init_delayable(struct k_work_delayable *dwork, k_work_handler_t handler) { __ASSERT_NO_MSG(dwork != NULL); __ASSERT_NO_MSG(handler != NULL); *dwork = (struct k_work_delayable){ .work = { .handler = handler, .flags = K_WORK_DELAYABLE, }, }; z_init_timeout(&dwork->timeout); SYS_PORT_TRACING_OBJ_INIT(k_work_delayable, dwork); } static inline int work_delayable_busy_get_locked(const struct k_work_delayable *dwork) { return flags_get(&dwork->work.flags) & K_WORK_MASK; } int k_work_delayable_busy_get(const struct k_work_delayable *dwork) { k_spinlock_key_t key = k_spin_lock(&lock); int ret = work_delayable_busy_get_locked(dwork); k_spin_unlock(&lock, key); return ret; } /* Attempt to schedule a work item for future (maybe immediate) * submission. * * Invoked with work lock held. * * See also submit_to_queue_locked(), which implements this for a no-wait * delay. * * Invoked with work lock held. * * @param queuep pointer to a pointer to a queue. On input this * should dereference to the proposed queue (which may be null); after * completion it will be null if the work was not submitted or if * submitted will reference the queue it was submitted to. That may * or may not be the queue provided on input. * * @param dwork the delayed work structure * * @param delay the delay to use before scheduling. * * @retval from submit_to_queue_locked() if delay is K_NO_WAIT; otherwise * @retval 1 to indicate successfully scheduled. */ static int schedule_for_queue_locked(struct k_work_q **queuep, struct k_work_delayable *dwork, k_timeout_t delay) { int ret = 1; struct k_work *work = &dwork->work; if (K_TIMEOUT_EQ(delay, K_NO_WAIT)) { return submit_to_queue_locked(work, queuep); } flag_set(&work->flags, K_WORK_DELAYED_BIT); dwork->queue = *queuep; /* Add timeout */ z_add_timeout(&dwork->timeout, work_timeout, delay); return ret; } /* Unschedule delayable work. * * If the work is delayed, cancel the timeout and clear the delayed * flag. * * Invoked with work lock held. * * @param dwork pointer to delayable work structure. * * @return true if and only if work had been delayed so the timeout * was cancelled. */ static inline bool unschedule_locked(struct k_work_delayable *dwork) { bool ret = false; struct k_work *work = &dwork->work; /* If scheduled, try to cancel. If it fails, that means the * callback has been dequeued and will inevitably run (or has * already run), so treat that as "undelayed" and return * false. */ if (flag_test_and_clear(&work->flags, K_WORK_DELAYED_BIT)) { ret = z_abort_timeout(&dwork->timeout) == 0; } return ret; } /* Full cancellation of a delayable work item. * * Unschedules the delayed part then delegates to standard work * cancellation. * * Invoked with work lock held. * * @param dwork delayable work item * * @return k_work_busy_get() flags */ static int cancel_delayable_async_locked(struct k_work_delayable *dwork) { (void)unschedule_locked(dwork); return cancel_async_locked(&dwork->work); } int k_work_schedule_for_queue(struct k_work_q *queue, struct k_work_delayable *dwork, k_timeout_t delay) { __ASSERT_NO_MSG(dwork != NULL); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, schedule_for_queue, queue, dwork, delay); struct k_work *work = &dwork->work; int ret = 0; k_spinlock_key_t key = k_spin_lock(&lock); /* Schedule the work item if it's idle or running. */ if ((work_busy_get_locked(work) & ~K_WORK_RUNNING) == 0U) { ret = schedule_for_queue_locked(&queue, dwork, delay); } k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, schedule_for_queue, queue, dwork, delay, ret); return ret; } int k_work_schedule(struct k_work_delayable *dwork, k_timeout_t delay) { SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, schedule, dwork, delay); int ret = k_work_schedule_for_queue(&k_sys_work_q, dwork, delay); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, schedule, dwork, delay, ret); return ret; } int k_work_reschedule_for_queue(struct k_work_q *queue, struct k_work_delayable *dwork, k_timeout_t delay) { __ASSERT_NO_MSG(dwork != NULL); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, reschedule_for_queue, queue, dwork, delay); int ret; k_spinlock_key_t key = k_spin_lock(&lock); /* Remove any active scheduling. */ (void)unschedule_locked(dwork); /* Schedule the work item with the new parameters. */ ret = schedule_for_queue_locked(&queue, dwork, delay); k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, reschedule_for_queue, queue, dwork, delay, ret); return ret; } int k_work_reschedule(struct k_work_delayable *dwork, k_timeout_t delay) { SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, reschedule, dwork, delay); int ret = k_work_reschedule_for_queue(&k_sys_work_q, dwork, delay); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, reschedule, dwork, delay, ret); return ret; } int k_work_cancel_delayable(struct k_work_delayable *dwork) { __ASSERT_NO_MSG(dwork != NULL); SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_delayable, dwork); k_spinlock_key_t key = k_spin_lock(&lock); int ret = cancel_delayable_async_locked(dwork); k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_delayable, dwork, ret); return ret; } bool k_work_cancel_delayable_sync(struct k_work_delayable *dwork, struct k_work_sync *sync) { __ASSERT_NO_MSG(dwork != NULL); __ASSERT_NO_MSG(sync != NULL); __ASSERT_NO_MSG(!k_is_in_isr()); #ifdef CONFIG_KERNEL_COHERENCE __ASSERT_NO_MSG(arch_mem_coherent(sync)); #endif /* CONFIG_KERNEL_COHERENCE */ SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, cancel_delayable_sync, dwork, sync); struct z_work_canceller *canceller = &sync->canceller; k_spinlock_key_t key = k_spin_lock(&lock); bool pending = (work_delayable_busy_get_locked(dwork) != 0U); bool need_wait = false; if (pending) { (void)cancel_delayable_async_locked(dwork); need_wait = cancel_sync_locked(&dwork->work, canceller); } k_spin_unlock(&lock, key); if (need_wait) { k_sem_take(&canceller->sem, K_FOREVER); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, cancel_delayable_sync, dwork, sync, pending); return pending; } bool k_work_flush_delayable(struct k_work_delayable *dwork, struct k_work_sync *sync) { __ASSERT_NO_MSG(dwork != NULL); __ASSERT_NO_MSG(sync != NULL); __ASSERT_NO_MSG(!k_is_in_isr()); #ifdef CONFIG_KERNEL_COHERENCE __ASSERT_NO_MSG(arch_mem_coherent(sync)); #endif /* CONFIG_KERNEL_COHERENCE */ SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_work, flush_delayable, dwork, sync); struct k_work *work = &dwork->work; struct z_work_flusher *flusher = &sync->flusher; k_spinlock_key_t key = k_spin_lock(&lock); /* If it's idle release the lock and return immediately. */ if (work_busy_get_locked(work) == 0U) { k_spin_unlock(&lock, key); SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush_delayable, dwork, sync, false); return false; } /* If unscheduling did something then submit it. Ignore a * failed submission (e.g. when cancelling). */ if (unschedule_locked(dwork)) { struct k_work_q *queue = dwork->queue; (void)submit_to_queue_locked(work, &queue); } /* Wait for it to finish */ bool need_flush = work_flush_locked(work, flusher); k_spin_unlock(&lock, key); /* If necessary wait until the flusher item completes */ if (need_flush) { k_sem_take(&flusher->sem, K_FOREVER); } SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_work, flush_delayable, dwork, sync, need_flush); return need_flush; } #endif /* CONFIG_SYS_CLOCK_EXISTS */