From 15e2117a0b90a50fcf040faa23ca86e4b61d1ec9 Mon Sep 17 00:00:00 2001 From: Andy Ross Date: Wed, 13 May 2020 15:26:08 +0000 Subject: [PATCH] soc/xtensa/intel_adsp: Add support for L1-cache-aware linkage Xtensa CPUs have incoherent L1 caches, which is deeply inconvenient for SMP systems. But as a treatment for this, the ADSP memory map contains the RAM twice, in separate 512MB regions that can be managed separately by the Xtensa TLB/cacheattr mechanism. The low mapping is set to bypass the cache where the high mapping is cached. Set up linkage to use both as appropriate, then reassemble the final sections to a contiguous region. Read-only areas (.text, .rodata) are cached. Data sections are uncached by default, except for a special ".cache" section that may be used by higher level code to flag static areas (e.g. stacks) which don't store multiprocessor-shared content. Signed-off-by: Andy Ross --- soc/xtensa/intel_adsp/cavs_v15/linker.ld | 71 +++++++++++++++---- soc/xtensa/intel_adsp/common/bootloader.cmake | 5 ++ 2 files changed, 61 insertions(+), 15 deletions(-) diff --git a/soc/xtensa/intel_adsp/cavs_v15/linker.ld b/soc/xtensa/intel_adsp/cavs_v15/linker.ld index baf3293a4dc..7d7ddda5195 100644 --- a/soc/xtensa/intel_adsp/cavs_v15/linker.ld +++ b/soc/xtensa/intel_adsp/cavs_v15/linker.ld @@ -25,10 +25,27 @@ OUTPUT_ARCH(xtensa) PROVIDE(__memctl_default = 0x00000000); PROVIDE(_MemErrorHandler = 0x00000000); -#define RAMABLE_REGION ram :ram_phdr -#define ROMABLE_REGION ram :ram_phdr #define LPRAM_REGION lpram +/* DSP RAM regions (all of them) are mapped twice on the DSP: once in + * a 512MB region from 0x80000000-0x9fffffff and again from + * 0xa0000000-0xbfffffff. The first mapping is set up to bypass the + * L1 cache, so it must be used when multiprocessor coherence is + * desired, where the latter mapping is best used for processor-local + * data (e.g. stacks) or shared data that is managed with explicit + * cache flush/invalidate operations. + * + * These macros will set up a segment start address correctly, + * including alignment to a cache line. Be sure to also emit the + * section to ">ram :ram_phdr" or ">ucram :ucram_phdr" as + * appropriate. (Forgetting the correct PHDR will actually work, as + * the output tooling ignores it, but it will cause the linker to emit + * 512MB of unused data into the output file!) + * + */ +#define SEGSTART_CACHED (ALIGN(64) | 0x20000000) +#define SEGSTART_UNCACHED (ALIGN(64) & ~0x20000000) + MEMORY { vector_memory_lit : @@ -97,6 +114,9 @@ MEMORY ram : org = RAM_BASE, len = RAM_SIZE + ucram : + org = RAM_BASE - 0x20000000, + len = RAM_SIZE #ifdef CONFIG_GEN_ISR_TABLES IDT_LIST : org = IDT_BASE, @@ -138,7 +158,7 @@ PHDRS vector_double_lit_phdr PT_LOAD; vector_double_text_phdr PT_LOAD; ram_phdr PT_LOAD; - + ucram_phdr PT_LOAD; static_uuid_entries_phdr PT_NOTE; static_log_entries_phdr PT_NOTE; } @@ -390,6 +410,8 @@ SECTIONS _module_init_end = ABSOLUTE(.); } >ram :ram_phdr +#define RAMABLE_REGION ram :ram_phdr +#define ROMABLE_REGION ram :ram_phdr #include .fw_ready : ALIGN(4) @@ -404,7 +426,16 @@ SECTIONS *(.noinit.*) } >ram :ram_phdr - .data : ALIGN(4) + .lit4 SEGSTART_CACHED : ALIGN(4) + { + _lit4_start = ABSOLUTE(.); + *(*.lit4) + *(.lit4.*) + *(.gnu.linkonce.lit4.*) + _lit4_end = ABSOLUTE(.); + } >ram :ram_phdr + + .data SEGSTART_UNCACHED : ALIGN(4) { _data_start = ABSOLUTE(.); *(.data) @@ -423,18 +454,25 @@ SECTIONS *(.gna_model) _data_end = ABSOLUTE(.); . = ALIGN(4096); - } >ram :ram_phdr - .lit4 : ALIGN(4) - { - _lit4_start = ABSOLUTE(.); - *(*.lit4) - *(.lit4.*) - *(.gnu.linkonce.lit4.*) - _lit4_end = ABSOLUTE(.); - } >ram :ram_phdr + } >ucram :ucram_phdr + +/* These values need to change in our scheme, where the common-ram + * sections need to be linked in safe/uncached memory but common-rom + * wants to use the cache + */ +#undef RAMABLE_REGION +#undef ROMABLE_REGION +#define RAMABLE_REGION ucram :ucram_phdr +#define ROMABLE_REGION ucram :ucram_phdr #include - .bss (NOLOAD) : ALIGN(4096) + /* ANDY TEST */ + .ucram SEGSTART_UNCACHED : + { + *(.ucram) + } >ucram :ucram_phdr + + .bss SEGSTART_UNCACHED (NOLOAD) : ALIGN(4096) { . = ALIGN(4096); _bss_start = ABSOLUTE(.); @@ -453,7 +491,10 @@ SECTIONS *(COMMON) . = ALIGN(8); _bss_end = ABSOLUTE(.); - } >ram :ram_phdr + } >ucram :ucram_phdr + + /* Re-adjust to the upper mapping for the final symbols below */ + . = SEGSTART_CACHED; /* stack */ _end = ALIGN(8); diff --git a/soc/xtensa/intel_adsp/common/bootloader.cmake b/soc/xtensa/intel_adsp/common/bootloader.cmake index ba26b2e4740..7213e6101b7 100644 --- a/soc/xtensa/intel_adsp/common/bootloader.cmake +++ b/soc/xtensa/intel_adsp/common/bootloader.cmake @@ -14,12 +14,17 @@ if(EXISTS ${CMAKE_CURRENT_LIST_DIR}/bootloader/CMakeLists.txt) add_subdirectory(${CMAKE_CURRENT_LIST_DIR}/bootloader ${build_dir}) endif() +set(ELF_FIX ${SOC_DIR}/${ARCH}/${SOC_FAMILY}/common/fix_elf_addrs.py) + add_custom_target( process_elf ALL DEPENDS base_module DEPENDS ${ZEPHYR_FINAL_EXECUTABLE} COMMAND ${CMAKE_OBJCOPY} --dump-section .data=mod-apl.bin $ COMMAND ${CMAKE_OBJCOPY} --add-section .module=mod-apl.bin --set-section-flags .module=load,readonly ${CMAKE_BINARY_DIR}/zephyr/${KERNEL_ELF_NAME} ${CMAKE_BINARY_DIR}/zephyr/${KERNEL_ELF_NAME}.mod + + # Adjust final section addresses so they all appear in the cached region. + COMMAND ${ELF_FIX} ${CMAKE_OBJCOPY} ${CMAKE_BINARY_DIR}/zephyr/zephyr.elf.mod ) add_custom_target(