As of today <zephyr/zephyr.h> is 100% equivalent to <zephyr/kernel.h>. This patch proposes to then include <zephyr/kernel.h> instead of <zephyr/zephyr.h> since it is more clear that you are including the Kernel APIs and (probably) nothing else. <zephyr/zephyr.h> sounds like a catch-all header that may be confusing. Most applications need to include a bunch of other things to compile, e.g. driver headers or subsystem headers like BT, logging, etc. The idea of a catch-all header in Zephyr is probably not feasible anyway. Reason is that Zephyr is not a library, like it could be for example `libpython`. Zephyr provides many utilities nowadays: a kernel, drivers, subsystems, etc and things will likely grow. A catch-all header would be massive, difficult to keep up-to-date. It is also likely that an application will only build a small subset. Note that subsystem-level headers may use a catch-all approach to make things easier, though. NOTE: This patch is **NOT** removing the header, just removing its usage in-tree. I'd advocate for its deprecation (add a #warning on it), but I understand many people will have concerns. Signed-off-by: Gerard Marull-Paretas <gerard.marull@nordicsemi.no>
87 lines
1.9 KiB
C
87 lines
1.9 KiB
C
/*
|
|
* Copyright (c) 2020 Intel Corporation
|
|
*
|
|
* SPDX-License-Identifier: Apache-2.0
|
|
*/
|
|
|
|
#include <zephyr/kernel.h>
|
|
#include <zephyr/arch/cpu.h>
|
|
#include <zephyr/sys/arch_interface.h>
|
|
|
|
#define NUM_THREADS 20
|
|
#define STACK_SIZE (1024)
|
|
|
|
K_THREAD_STACK_ARRAY_DEFINE(tstacks, NUM_THREADS, STACK_SIZE);
|
|
|
|
static struct k_thread t[NUM_THREADS];
|
|
|
|
K_MUTEX_DEFINE(mutex);
|
|
K_CONDVAR_DEFINE(condvar);
|
|
|
|
static int done;
|
|
|
|
void worker_thread(void *p1, void *p2, void *p3)
|
|
{
|
|
const int myid = (long)p1;
|
|
const int workloops = 5;
|
|
|
|
for (int i = 0; i < workloops; i++) {
|
|
printk("[thread %d] working (%d/%d)\n", myid, i, workloops);
|
|
k_sleep(K_MSEC(500));
|
|
}
|
|
|
|
/*
|
|
* we're going to manipulate done and use the cond, so we need the mutex
|
|
*/
|
|
k_mutex_lock(&mutex, K_FOREVER);
|
|
|
|
/*
|
|
* increase the count of threads that have finished their work.
|
|
*/
|
|
done++;
|
|
printk("[thread %d] done is now %d. Signalling cond.\n", myid, done);
|
|
|
|
k_condvar_signal(&condvar);
|
|
k_mutex_unlock(&mutex);
|
|
}
|
|
|
|
void main(void)
|
|
{
|
|
k_tid_t tid[NUM_THREADS];
|
|
|
|
done = 0;
|
|
|
|
for (int i = 0; i < NUM_THREADS; i++) {
|
|
tid[i] =
|
|
k_thread_create(&t[i], tstacks[i], STACK_SIZE,
|
|
worker_thread, INT_TO_POINTER(i), NULL,
|
|
NULL, K_PRIO_PREEMPT(10), 0, K_NO_WAIT);
|
|
}
|
|
k_sleep(K_MSEC(1000));
|
|
|
|
k_mutex_lock(&mutex, K_FOREVER);
|
|
|
|
/*
|
|
* are the other threads still busy?
|
|
*/
|
|
while (done < NUM_THREADS) {
|
|
printk("[thread %s] done is %d which is < %d so waiting on cond\n",
|
|
__func__, done, (int)NUM_THREADS);
|
|
|
|
/* block this thread until another thread signals cond. While
|
|
* blocked, the mutex is released, then re-acquired before this
|
|
* thread is woken up and the call returns.
|
|
*/
|
|
k_condvar_wait(&condvar, &mutex, K_FOREVER);
|
|
|
|
printk("[thread %s] wake - cond was signalled.\n", __func__);
|
|
|
|
/* we go around the loop with the lock held */
|
|
}
|
|
|
|
printk("[thread %s] done == %d so everyone is done\n",
|
|
__func__, (int)NUM_THREADS);
|
|
|
|
k_mutex_unlock(&mutex);
|
|
}
|