zephyr/samples/kernel/condition_variables/simple/src/main.c
Gerard Marull-Paretas 79e6b0e0f6 includes: prefer <zephyr/kernel.h> over <zephyr/zephyr.h>
As of today <zephyr/zephyr.h> is 100% equivalent to <zephyr/kernel.h>.
This patch proposes to then include <zephyr/kernel.h> instead of
<zephyr/zephyr.h> since it is more clear that you are including the
Kernel APIs and (probably) nothing else. <zephyr/zephyr.h> sounds like a
catch-all header that may be confusing. Most applications need to
include a bunch of other things to compile, e.g. driver headers or
subsystem headers like BT, logging, etc.

The idea of a catch-all header in Zephyr is probably not feasible
anyway. Reason is that Zephyr is not a library, like it could be for
example `libpython`. Zephyr provides many utilities nowadays: a kernel,
drivers, subsystems, etc and things will likely grow. A catch-all header
would be massive, difficult to keep up-to-date. It is also likely that
an application will only build a small subset. Note that subsystem-level
headers may use a catch-all approach to make things easier, though.

NOTE: This patch is **NOT** removing the header, just removing its usage
in-tree. I'd advocate for its deprecation (add a #warning on it), but I
understand many people will have concerns.

Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
2022-09-05 16:31:47 +02:00

87 lines
1.9 KiB
C

/*
* Copyright (c) 2020 Intel Corporation
*
* SPDX-License-Identifier: Apache-2.0
*/
#include <zephyr/kernel.h>
#include <zephyr/arch/cpu.h>
#include <zephyr/sys/arch_interface.h>
#define NUM_THREADS 20
#define STACK_SIZE (1024)
K_THREAD_STACK_ARRAY_DEFINE(tstacks, NUM_THREADS, STACK_SIZE);
static struct k_thread t[NUM_THREADS];
K_MUTEX_DEFINE(mutex);
K_CONDVAR_DEFINE(condvar);
static int done;
void worker_thread(void *p1, void *p2, void *p3)
{
const int myid = (long)p1;
const int workloops = 5;
for (int i = 0; i < workloops; i++) {
printk("[thread %d] working (%d/%d)\n", myid, i, workloops);
k_sleep(K_MSEC(500));
}
/*
* we're going to manipulate done and use the cond, so we need the mutex
*/
k_mutex_lock(&mutex, K_FOREVER);
/*
* increase the count of threads that have finished their work.
*/
done++;
printk("[thread %d] done is now %d. Signalling cond.\n", myid, done);
k_condvar_signal(&condvar);
k_mutex_unlock(&mutex);
}
void main(void)
{
k_tid_t tid[NUM_THREADS];
done = 0;
for (int i = 0; i < NUM_THREADS; i++) {
tid[i] =
k_thread_create(&t[i], tstacks[i], STACK_SIZE,
worker_thread, INT_TO_POINTER(i), NULL,
NULL, K_PRIO_PREEMPT(10), 0, K_NO_WAIT);
}
k_sleep(K_MSEC(1000));
k_mutex_lock(&mutex, K_FOREVER);
/*
* are the other threads still busy?
*/
while (done < NUM_THREADS) {
printk("[thread %s] done is %d which is < %d so waiting on cond\n",
__func__, done, (int)NUM_THREADS);
/* block this thread until another thread signals cond. While
* blocked, the mutex is released, then re-acquired before this
* thread is woken up and the call returns.
*/
k_condvar_wait(&condvar, &mutex, K_FOREVER);
printk("[thread %s] wake - cond was signalled.\n", __func__);
/* we go around the loop with the lock held */
}
printk("[thread %s] done == %d so everyone is done\n",
__func__, (int)NUM_THREADS);
k_mutex_unlock(&mutex);
}