kernel: rename Z_KERNEL_STACK_BUFFER to K_KERNEL_STACK_BUFFER
Simple rename to align the kernel naming scheme. This is being used throughout the tree, especially in the architecture code. As this is not a private API internal to kernel, prefix it appropriately with K_. Signed-off-by: Daniel Leung <daniel.leung@intel.com>
This commit is contained in:
parent
f05111dea0
commit
b69d2486fe
22 changed files with 30 additions and 30 deletions
|
@ -44,11 +44,11 @@ K_KERNEL_STACK_DEFINE(_firq_interrupt_stack, CONFIG_ARC_FIRQ_STACK_SIZE);
|
||||||
void z_arc_firq_stack_set(void)
|
void z_arc_firq_stack_set(void)
|
||||||
{
|
{
|
||||||
#ifdef CONFIG_SMP
|
#ifdef CONFIG_SMP
|
||||||
char *firq_sp = Z_KERNEL_STACK_BUFFER(
|
char *firq_sp = K_KERNEL_STACK_BUFFER(
|
||||||
_firq_interrupt_stack[z_arc_v2_core_id()]) +
|
_firq_interrupt_stack[z_arc_v2_core_id()]) +
|
||||||
CONFIG_ARC_FIRQ_STACK_SIZE;
|
CONFIG_ARC_FIRQ_STACK_SIZE;
|
||||||
#else
|
#else
|
||||||
char *firq_sp = Z_KERNEL_STACK_BUFFER(_firq_interrupt_stack) +
|
char *firq_sp = K_KERNEL_STACK_BUFFER(_firq_interrupt_stack) +
|
||||||
CONFIG_ARC_FIRQ_STACK_SIZE;
|
CONFIG_ARC_FIRQ_STACK_SIZE;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
|
|
@ -50,7 +50,7 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
* arc_cpu_wake_flag will protect arc_cpu_sp that
|
* arc_cpu_wake_flag will protect arc_cpu_sp that
|
||||||
* only one slave cpu can read it per time
|
* only one slave cpu can read it per time
|
||||||
*/
|
*/
|
||||||
arc_cpu_sp = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
arc_cpu_sp = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
|
|
||||||
arc_cpu_wake_flag = cpu_num;
|
arc_cpu_wake_flag = cpu_num;
|
||||||
|
|
||||||
|
|
|
@ -283,7 +283,7 @@ FUNC_NORETURN void z_arc_switch_to_main_no_multithreading(k_thread_entry_t main_
|
||||||
void *p1, void *p2, void *p3)
|
void *p1, void *p2, void *p3)
|
||||||
{
|
{
|
||||||
_kernel.cpus[0].id = 0;
|
_kernel.cpus[0].id = 0;
|
||||||
_kernel.cpus[0].irq_stack = (Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]) +
|
_kernel.cpus[0].irq_stack = (K_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]) +
|
||||||
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
||||||
|
|
||||||
void *main_stack = (Z_THREAD_STACK_BUFFER(z_main_stack) +
|
void *main_stack = (Z_THREAD_STACK_BUFFER(z_main_stack) +
|
||||||
|
|
|
@ -120,16 +120,16 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz, arch_cpustart_
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Pass stack address to secondary core */
|
/* Pass stack address to secondary core */
|
||||||
arm_cpu_boot_params.irq_sp = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
arm_cpu_boot_params.irq_sp = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
arm_cpu_boot_params.fiq_sp = Z_KERNEL_STACK_BUFFER(z_arm_fiq_stack[cpu_num])
|
arm_cpu_boot_params.fiq_sp = K_KERNEL_STACK_BUFFER(z_arm_fiq_stack[cpu_num])
|
||||||
+ CONFIG_ARMV7_FIQ_STACK_SIZE;
|
+ CONFIG_ARMV7_FIQ_STACK_SIZE;
|
||||||
arm_cpu_boot_params.abt_sp = Z_KERNEL_STACK_BUFFER(z_arm_abort_stack[cpu_num])
|
arm_cpu_boot_params.abt_sp = K_KERNEL_STACK_BUFFER(z_arm_abort_stack[cpu_num])
|
||||||
+ CONFIG_ARMV7_EXCEPTION_STACK_SIZE;
|
+ CONFIG_ARMV7_EXCEPTION_STACK_SIZE;
|
||||||
arm_cpu_boot_params.udf_sp = Z_KERNEL_STACK_BUFFER(z_arm_undef_stack[cpu_num])
|
arm_cpu_boot_params.udf_sp = K_KERNEL_STACK_BUFFER(z_arm_undef_stack[cpu_num])
|
||||||
+ CONFIG_ARMV7_EXCEPTION_STACK_SIZE;
|
+ CONFIG_ARMV7_EXCEPTION_STACK_SIZE;
|
||||||
arm_cpu_boot_params.svc_sp = Z_KERNEL_STACK_BUFFER(z_arm_svc_stack[cpu_num])
|
arm_cpu_boot_params.svc_sp = K_KERNEL_STACK_BUFFER(z_arm_svc_stack[cpu_num])
|
||||||
+ CONFIG_ARMV7_SVC_STACK_SIZE;
|
+ CONFIG_ARMV7_SVC_STACK_SIZE;
|
||||||
arm_cpu_boot_params.sys_sp = Z_KERNEL_STACK_BUFFER(z_arm_sys_stack[cpu_num])
|
arm_cpu_boot_params.sys_sp = K_KERNEL_STACK_BUFFER(z_arm_sys_stack[cpu_num])
|
||||||
+ CONFIG_ARMV7_SYS_STACK_SIZE;
|
+ CONFIG_ARMV7_SYS_STACK_SIZE;
|
||||||
|
|
||||||
arm_cpu_boot_params.fn = fn;
|
arm_cpu_boot_params.fn = fn;
|
||||||
|
|
|
@ -28,7 +28,7 @@ void z_arm_init_stacks(void)
|
||||||
memset(z_arm_svc_stack, 0xAA, CONFIG_ARMV7_SVC_STACK_SIZE);
|
memset(z_arm_svc_stack, 0xAA, CONFIG_ARMV7_SVC_STACK_SIZE);
|
||||||
memset(z_arm_abort_stack, 0xAA, CONFIG_ARMV7_EXCEPTION_STACK_SIZE);
|
memset(z_arm_abort_stack, 0xAA, CONFIG_ARMV7_EXCEPTION_STACK_SIZE);
|
||||||
memset(z_arm_undef_stack, 0xAA, CONFIG_ARMV7_EXCEPTION_STACK_SIZE);
|
memset(z_arm_undef_stack, 0xAA, CONFIG_ARMV7_EXCEPTION_STACK_SIZE);
|
||||||
memset(Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]), 0xAA,
|
memset(K_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]), 0xAA,
|
||||||
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
|
|
|
@ -40,7 +40,7 @@ K_KERNEL_STACK_ARRAY_DECLARE(z_interrupt_stacks, CONFIG_MP_MAX_NUM_CPUS,
|
||||||
static ALWAYS_INLINE void z_arm_interrupt_stack_setup(void)
|
static ALWAYS_INLINE void z_arm_interrupt_stack_setup(void)
|
||||||
{
|
{
|
||||||
uint32_t msp =
|
uint32_t msp =
|
||||||
(uint32_t)(Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[0])) +
|
(uint32_t)(K_KERNEL_STACK_BUFFER(z_interrupt_stacks[0])) +
|
||||||
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]);
|
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]);
|
||||||
|
|
||||||
__set_MSP(msp);
|
__set_MSP(msp);
|
||||||
|
|
|
@ -33,7 +33,7 @@ void z_arm64_safe_exception_stack_init(void)
|
||||||
char *safe_exc_sp;
|
char *safe_exc_sp;
|
||||||
|
|
||||||
cpu_id = arch_curr_cpu()->id;
|
cpu_id = arch_curr_cpu()->id;
|
||||||
safe_exc_sp = Z_KERNEL_STACK_BUFFER(z_arm64_safe_exception_stacks[cpu_id]) +
|
safe_exc_sp = K_KERNEL_STACK_BUFFER(z_arm64_safe_exception_stacks[cpu_id]) +
|
||||||
CONFIG_ARM64_SAFE_EXCEPTION_STACK_SIZE;
|
CONFIG_ARM64_SAFE_EXCEPTION_STACK_SIZE;
|
||||||
arch_curr_cpu()->arch.safe_exception_stack = (uint64_t)safe_exc_sp;
|
arch_curr_cpu()->arch.safe_exception_stack = (uint64_t)safe_exc_sp;
|
||||||
write_sp_el0((uint64_t)safe_exc_sp);
|
write_sp_el0((uint64_t)safe_exc_sp);
|
||||||
|
|
|
@ -84,7 +84,7 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
"The count of CPU Cores nodes in dts is not equal to CONFIG_MP_MAX_NUM_CPUS\n");
|
"The count of CPU Cores nodes in dts is not equal to CONFIG_MP_MAX_NUM_CPUS\n");
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
arm64_cpu_boot_params.sp = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
arm64_cpu_boot_params.sp = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
arm64_cpu_boot_params.fn = fn;
|
arm64_cpu_boot_params.fn = fn;
|
||||||
arm64_cpu_boot_params.arg = arg;
|
arm64_cpu_boot_params.arg = arg;
|
||||||
arm64_cpu_boot_params.cpu_num = cpu_num;
|
arm64_cpu_boot_params.cpu_num = cpu_num;
|
||||||
|
|
|
@ -33,7 +33,7 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
riscv_cpu_init[cpu_num].fn = fn;
|
riscv_cpu_init[cpu_num].fn = fn;
|
||||||
riscv_cpu_init[cpu_num].arg = arg;
|
riscv_cpu_init[cpu_num].arg = arg;
|
||||||
|
|
||||||
riscv_cpu_sp = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
riscv_cpu_sp = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
riscv_cpu_boot_flag = 0U;
|
riscv_cpu_boot_flag = 0U;
|
||||||
|
|
||||||
#ifdef CONFIG_PM_CPU_OPS
|
#ifdef CONFIG_PM_CPU_OPS
|
||||||
|
|
|
@ -206,7 +206,7 @@ FUNC_NORETURN void z_riscv_switch_to_main_no_multithreading(k_thread_entry_t mai
|
||||||
ARG_UNUSED(p3);
|
ARG_UNUSED(p3);
|
||||||
|
|
||||||
_kernel.cpus[0].id = 0;
|
_kernel.cpus[0].id = 0;
|
||||||
_kernel.cpus[0].irq_stack = (Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]) +
|
_kernel.cpus[0].irq_stack = (K_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]) +
|
||||||
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]));
|
||||||
|
|
||||||
main_stack = (Z_THREAD_STACK_BUFFER(z_main_stack) +
|
main_stack = (Z_THREAD_STACK_BUFFER(z_main_stack) +
|
||||||
|
|
|
@ -59,7 +59,7 @@ bool z_x86_check_stack_bounds(uintptr_t addr, size_t size, uint16_t cs)
|
||||||
#else
|
#else
|
||||||
cpu_id = 0;
|
cpu_id = 0;
|
||||||
#endif
|
#endif
|
||||||
start = (uintptr_t)Z_KERNEL_STACK_BUFFER(
|
start = (uintptr_t)K_KERNEL_STACK_BUFFER(
|
||||||
z_interrupt_stacks[cpu_id]);
|
z_interrupt_stacks[cpu_id]);
|
||||||
end = start + CONFIG_ISR_STACK_SIZE;
|
end = start + CONFIG_ISR_STACK_SIZE;
|
||||||
#ifdef CONFIG_USERSPACE
|
#ifdef CONFIG_USERSPACE
|
||||||
|
|
|
@ -206,7 +206,7 @@ static FUNC_NORETURN __used void df_handler_top(void)
|
||||||
_df_esf.eflags = _main_tss.eflags;
|
_df_esf.eflags = _main_tss.eflags;
|
||||||
|
|
||||||
/* Restore the main IA task to a runnable state */
|
/* Restore the main IA task to a runnable state */
|
||||||
_main_tss.esp = (uint32_t)(Z_KERNEL_STACK_BUFFER(
|
_main_tss.esp = (uint32_t)(K_KERNEL_STACK_BUFFER(
|
||||||
z_interrupt_stacks[0]) + CONFIG_ISR_STACK_SIZE);
|
z_interrupt_stacks[0]) + CONFIG_ISR_STACK_SIZE);
|
||||||
_main_tss.cs = CODE_SEG;
|
_main_tss.cs = CODE_SEG;
|
||||||
_main_tss.ds = DATA_SEG;
|
_main_tss.ds = DATA_SEG;
|
||||||
|
|
|
@ -156,7 +156,7 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
|
|
||||||
apic_id = x86_cpu_loapics[cpu_num];
|
apic_id = x86_cpu_loapics[cpu_num];
|
||||||
|
|
||||||
x86_cpuboot[cpu_num].sp = (uint64_t) Z_KERNEL_STACK_BUFFER(stack) + sz;
|
x86_cpuboot[cpu_num].sp = (uint64_t) K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
x86_cpuboot[cpu_num].stack_size = sz;
|
x86_cpuboot[cpu_num].stack_size = sz;
|
||||||
x86_cpuboot[cpu_num].fn = fn;
|
x86_cpuboot[cpu_num].fn = fn;
|
||||||
x86_cpuboot[cpu_num].arg = arg;
|
x86_cpuboot[cpu_num].arg = arg;
|
||||||
|
|
|
@ -51,7 +51,7 @@ static ALWAYS_INLINE void arch_kernel_init(void)
|
||||||
XTENSA_WSR(ZSR_CPU_STR, cpu0);
|
XTENSA_WSR(ZSR_CPU_STR, cpu0);
|
||||||
|
|
||||||
#ifdef CONFIG_INIT_STACKS
|
#ifdef CONFIG_INIT_STACKS
|
||||||
char *stack_start = Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]);
|
char *stack_start = K_KERNEL_STACK_BUFFER(z_interrupt_stacks[0]);
|
||||||
size_t stack_sz = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]);
|
size_t stack_sz = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[0]);
|
||||||
char *stack_end = stack_start + stack_sz;
|
char *stack_end = stack_start + stack_sz;
|
||||||
|
|
||||||
|
|
|
@ -283,7 +283,7 @@ static inline char *z_stack_ptr_align(char *ptr)
|
||||||
|
|
||||||
/** @} */
|
/** @} */
|
||||||
|
|
||||||
static inline char *Z_KERNEL_STACK_BUFFER(k_thread_stack_t *sym)
|
static inline char *K_KERNEL_STACK_BUFFER(k_thread_stack_t *sym)
|
||||||
{
|
{
|
||||||
return (char *)sym + K_KERNEL_STACK_RESERVED;
|
return (char *)sym + K_KERNEL_STACK_RESERVED;
|
||||||
}
|
}
|
||||||
|
@ -294,7 +294,7 @@ static inline char *Z_KERNEL_STACK_BUFFER(k_thread_stack_t *sym)
|
||||||
#define K_THREAD_STACK_DEFINE K_KERNEL_STACK_DEFINE
|
#define K_THREAD_STACK_DEFINE K_KERNEL_STACK_DEFINE
|
||||||
#define K_THREAD_STACK_ARRAY_DEFINE K_KERNEL_STACK_ARRAY_DEFINE
|
#define K_THREAD_STACK_ARRAY_DEFINE K_KERNEL_STACK_ARRAY_DEFINE
|
||||||
#define K_THREAD_STACK_MEMBER K_KERNEL_STACK_MEMBER
|
#define K_THREAD_STACK_MEMBER K_KERNEL_STACK_MEMBER
|
||||||
#define Z_THREAD_STACK_BUFFER Z_KERNEL_STACK_BUFFER
|
#define Z_THREAD_STACK_BUFFER K_KERNEL_STACK_BUFFER
|
||||||
#define K_THREAD_STACK_DECLARE K_KERNEL_STACK_DECLARE
|
#define K_THREAD_STACK_DECLARE K_KERNEL_STACK_DECLARE
|
||||||
#define K_THREAD_STACK_ARRAY_DECLARE K_KERNEL_STACK_ARRAY_DECLARE
|
#define K_THREAD_STACK_ARRAY_DECLARE K_KERNEL_STACK_ARRAY_DECLARE
|
||||||
#define K_THREAD_PINNED_STACK_DEFINE K_KERNEL_PINNED_STACK_DEFINE
|
#define K_THREAD_PINNED_STACK_DEFINE K_KERNEL_PINNED_STACK_DEFINE
|
||||||
|
|
|
@ -469,7 +469,7 @@ void z_init_cpu(int id)
|
||||||
_kernel.cpus[id].idle_thread = &z_idle_threads[id];
|
_kernel.cpus[id].idle_thread = &z_idle_threads[id];
|
||||||
_kernel.cpus[id].id = id;
|
_kernel.cpus[id].id = id;
|
||||||
_kernel.cpus[id].irq_stack =
|
_kernel.cpus[id].irq_stack =
|
||||||
(Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[id]) +
|
(K_KERNEL_STACK_BUFFER(z_interrupt_stacks[id]) +
|
||||||
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[id]));
|
K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[id]));
|
||||||
#ifdef CONFIG_SCHED_THREAD_USAGE_ALL
|
#ifdef CONFIG_SCHED_THREAD_USAGE_ALL
|
||||||
_kernel.cpus[id].usage = &_kernel.usage[id];
|
_kernel.cpus[id].usage = &_kernel.usage[id];
|
||||||
|
|
|
@ -392,7 +392,7 @@ static char *setup_thread_stack(struct k_thread *new_thread,
|
||||||
{
|
{
|
||||||
/* Object cannot host a user mode thread */
|
/* Object cannot host a user mode thread */
|
||||||
stack_obj_size = Z_KERNEL_STACK_SIZE_ADJUST(stack_size);
|
stack_obj_size = Z_KERNEL_STACK_SIZE_ADJUST(stack_size);
|
||||||
stack_buf_start = Z_KERNEL_STACK_BUFFER(stack);
|
stack_buf_start = K_KERNEL_STACK_BUFFER(stack);
|
||||||
stack_buf_size = stack_obj_size - K_KERNEL_STACK_RESERVED;
|
stack_buf_size = stack_obj_size - K_KERNEL_STACK_RESERVED;
|
||||||
|
|
||||||
/* Zephyr treats stack overflow as an app bug. But
|
/* Zephyr treats stack overflow as an app bug. But
|
||||||
|
|
|
@ -258,12 +258,12 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
|
|
||||||
sr.cpu = cpu_num;
|
sr.cpu = cpu_num;
|
||||||
sr.fn = fn;
|
sr.fn = fn;
|
||||||
sr.stack_top = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
sr.stack_top = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
sr.arg = arg;
|
sr.arg = arg;
|
||||||
sr.vecbase = vb;
|
sr.vecbase = vb;
|
||||||
sr.alive = &alive_flag;
|
sr.alive = &alive_flag;
|
||||||
|
|
||||||
appcpu_top = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
appcpu_top = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
|
|
||||||
start_rec = &sr;
|
start_rec = &sr;
|
||||||
|
|
||||||
|
|
|
@ -129,7 +129,7 @@ void arch_cpu_start(int cpu_num, k_thread_stack_t *stack, int sz,
|
||||||
start_rec.fn = fn;
|
start_rec.fn = fn;
|
||||||
start_rec.arg = arg;
|
start_rec.arg = arg;
|
||||||
|
|
||||||
z_mp_stack_top = Z_KERNEL_STACK_BUFFER(stack) + sz;
|
z_mp_stack_top = K_KERNEL_STACK_BUFFER(stack) + sz;
|
||||||
|
|
||||||
soc_start_core(cpu_num);
|
soc_start_core(cpu_num);
|
||||||
}
|
}
|
||||||
|
|
|
@ -134,7 +134,7 @@ static void isr_stacks(void)
|
||||||
unsigned int num_cpus = arch_num_cpus();
|
unsigned int num_cpus = arch_num_cpus();
|
||||||
|
|
||||||
for (int i = 0; i < num_cpus; i++) {
|
for (int i = 0; i < num_cpus; i++) {
|
||||||
const uint8_t *buf = Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[i]);
|
const uint8_t *buf = K_KERNEL_STACK_BUFFER(z_interrupt_stacks[i]);
|
||||||
size_t size = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[i]);
|
size_t size = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[i]);
|
||||||
size_t unused;
|
size_t unused;
|
||||||
int err;
|
int err;
|
||||||
|
|
|
@ -257,7 +257,7 @@ static int cmd_kernel_stacks(const struct shell *sh,
|
||||||
|
|
||||||
for (int i = 0; i < num_cpus; i++) {
|
for (int i = 0; i < num_cpus; i++) {
|
||||||
size_t unused;
|
size_t unused;
|
||||||
const uint8_t *buf = Z_KERNEL_STACK_BUFFER(z_interrupt_stacks[i]);
|
const uint8_t *buf = K_KERNEL_STACK_BUFFER(z_interrupt_stacks[i]);
|
||||||
size_t size = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[i]);
|
size_t size = K_KERNEL_STACK_SIZEOF(z_interrupt_stacks[i]);
|
||||||
int err = z_stack_space_get(buf, size, &unused);
|
int err = z_stack_space_get(buf, size, &unused);
|
||||||
|
|
||||||
|
|
|
@ -121,7 +121,7 @@ void stack_buffer_scenarios(void)
|
||||||
#endif
|
#endif
|
||||||
{
|
{
|
||||||
reserved = K_KERNEL_STACK_RESERVED;
|
reserved = K_KERNEL_STACK_RESERVED;
|
||||||
stack_buf = Z_KERNEL_STACK_BUFFER(stack_obj);
|
stack_buf = K_KERNEL_STACK_BUFFER(stack_obj);
|
||||||
alignment = Z_KERNEL_STACK_OBJ_ALIGN;
|
alignment = Z_KERNEL_STACK_OBJ_ALIGN;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue