mirror of
https://github.com/zephyrproject-rtos/zephyr
synced 2025-08-10 05:25:23 +00:00
Add k_delayed_work_pending similar to k_work_pending to check if the delayed work item has been submitted but not yet completed. This would compliment the API since using k_work_pending or k_delayed_work_remaining_get is not enough to check this condition. This is because the timeout could have run out, but the timeout handler not yet processed and put the work into the workqueue. Signed-off-by: Joakim Andersson <joakim.andersson@nordicsemi.no>
153 lines
3.2 KiB
C
153 lines
3.2 KiB
C
/*
|
|
* Copyright (c) 2016 Intel Corporation
|
|
* Copyright (c) 2016 Wind River Systems, Inc.
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
/**
|
|
* @file
|
|
*
|
|
* Workqueue support functions
|
|
*/
|
|
|
|
#include <kernel_structs.h>
|
|
#include <wait_q.h>
|
|
#include <spinlock.h>
|
|
#include <errno.h>
|
|
#include <stdbool.h>
|
|
#include <sys/check.h>
|
|
|
|
#define WORKQUEUE_THREAD_NAME "workqueue"
|
|
|
|
#ifdef CONFIG_SYS_CLOCK_EXISTS
|
|
static struct k_spinlock lock;
|
|
#endif
|
|
|
|
extern void z_work_q_main(void *work_q_ptr, void *p2, void *p3);
|
|
|
|
void k_work_q_start(struct k_work_q *work_q, k_thread_stack_t *stack,
|
|
size_t stack_size, int prio)
|
|
{
|
|
k_queue_init(&work_q->queue);
|
|
(void)k_thread_create(&work_q->thread, stack, stack_size, z_work_q_main,
|
|
work_q, NULL, NULL, prio, 0, K_NO_WAIT);
|
|
|
|
k_thread_name_set(&work_q->thread, WORKQUEUE_THREAD_NAME);
|
|
}
|
|
|
|
#ifdef CONFIG_SYS_CLOCK_EXISTS
|
|
static void work_timeout(struct _timeout *t)
|
|
{
|
|
struct k_delayed_work *w = CONTAINER_OF(t, struct k_delayed_work,
|
|
timeout);
|
|
|
|
/* submit work to workqueue */
|
|
k_work_submit_to_queue(w->work_q, &w->work);
|
|
}
|
|
|
|
void k_delayed_work_init(struct k_delayed_work *work, k_work_handler_t handler)
|
|
{
|
|
k_work_init(&work->work, handler);
|
|
z_init_timeout(&work->timeout);
|
|
work->work_q = NULL;
|
|
}
|
|
|
|
static int work_cancel(struct k_delayed_work *work)
|
|
{
|
|
CHECKIF(work->work_q == NULL) {
|
|
return -EALREADY;
|
|
}
|
|
|
|
if (k_work_pending(&work->work)) {
|
|
/* Remove from the queue if already submitted */
|
|
if (!k_queue_remove(&work->work_q->queue, &work->work)) {
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
int err = z_abort_timeout(&work->timeout);
|
|
|
|
if (err) {
|
|
return -EALREADY;
|
|
}
|
|
}
|
|
|
|
/* Detach from workqueue */
|
|
work->work_q = NULL;
|
|
|
|
atomic_clear_bit(work->work.flags, K_WORK_STATE_PENDING);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int k_delayed_work_submit_to_queue(struct k_work_q *work_q,
|
|
struct k_delayed_work *work,
|
|
k_timeout_t delay)
|
|
{
|
|
k_spinlock_key_t key = k_spin_lock(&lock);
|
|
int err = 0;
|
|
|
|
/* Work cannot be active in multiple queues */
|
|
if (work->work_q != NULL && work->work_q != work_q) {
|
|
err = -EADDRINUSE;
|
|
goto done;
|
|
}
|
|
|
|
/* Cancel if work has been submitted */
|
|
if (work->work_q == work_q) {
|
|
err = work_cancel(work);
|
|
/* -EALREADY indicates the work has already completed so this
|
|
* is likely a recurring work.
|
|
*/
|
|
if (err == -EALREADY) {
|
|
err = 0;
|
|
} else if (err < 0) {
|
|
goto done;
|
|
}
|
|
}
|
|
|
|
/* Attach workqueue so the timeout callback can submit it */
|
|
work->work_q = work_q;
|
|
|
|
/* Submit work directly if no delay. Note that this is a
|
|
* blocking operation, so release the lock first.
|
|
*/
|
|
if (K_TIMEOUT_EQ(delay, K_NO_WAIT)) {
|
|
k_spin_unlock(&lock, key);
|
|
k_work_submit_to_queue(work_q, &work->work);
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_LEGACY_TIMEOUT_API
|
|
delay = _TICK_ALIGN + k_ms_to_ticks_ceil32(delay);
|
|
#endif
|
|
|
|
/* Add timeout */
|
|
z_add_timeout(&work->timeout, work_timeout, delay);
|
|
|
|
done:
|
|
k_spin_unlock(&lock, key);
|
|
return err;
|
|
}
|
|
|
|
int k_delayed_work_cancel(struct k_delayed_work *work)
|
|
{
|
|
if (!work->work_q) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
k_spinlock_key_t key = k_spin_lock(&lock);
|
|
int ret = work_cancel(work);
|
|
|
|
k_spin_unlock(&lock, key);
|
|
return ret;
|
|
}
|
|
|
|
bool k_delayed_work_pending(struct k_delayed_work *work)
|
|
{
|
|
return !z_is_inactive_timeout(&work->timeout) ||
|
|
k_work_pending(&work->work);
|
|
}
|
|
|
|
#endif /* CONFIG_SYS_CLOCK_EXISTS */
|