2018-12-12 12:29:47 -05:00
|
|
|
// Copyright 2015-2016 Espressif Systems (Shanghai) PTE LTD
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
#include <string.h>
|
|
|
|
#include <sdkconfig.h>
|
|
|
|
#include "soc/soc_memory_layout.h"
|
|
|
|
#include "esp_attr.h"
|
2022-07-21 07:24:42 -04:00
|
|
|
#include "esp_cpu.h"
|
2022-10-13 00:01:27 -04:00
|
|
|
#include "esp_macros.h"
|
2018-12-12 12:29:47 -05:00
|
|
|
|
|
|
|
/* Encode the CPU ID in the LSB of the ccount value */
|
|
|
|
inline static uint32_t get_ccount(void)
|
|
|
|
{
|
2022-07-21 07:24:42 -04:00
|
|
|
uint32_t ccount = esp_cpu_get_cycle_count() & ~3;
|
2023-10-30 02:23:23 -04:00
|
|
|
#ifndef CONFIG_ESP_SYSTEM_SINGLE_CORE_MODE
|
2018-12-12 12:29:47 -05:00
|
|
|
ccount |= xPortGetCoreID();
|
|
|
|
#endif
|
|
|
|
return ccount;
|
|
|
|
}
|
|
|
|
|
2020-01-02 12:42:19 -05:00
|
|
|
/* Architecture-specific return value of __builtin_return_address which
|
|
|
|
* should be interpreted as an invalid address.
|
|
|
|
*/
|
|
|
|
#ifdef __XTENSA__
|
|
|
|
#define HEAP_ARCH_INVALID_PC 0x40000000
|
|
|
|
#else
|
|
|
|
#define HEAP_ARCH_INVALID_PC 0x00000000
|
|
|
|
#endif
|
|
|
|
|
2018-12-12 12:29:47 -05:00
|
|
|
// Caller is 2 stack frames deeper than we care about
|
|
|
|
#define STACK_OFFSET 2
|
|
|
|
|
|
|
|
#define TEST_STACK(N) do { \
|
|
|
|
if (STACK_DEPTH == N) { \
|
|
|
|
return; \
|
|
|
|
} \
|
2020-01-02 12:42:19 -05:00
|
|
|
callers[N] = __builtin_return_address(N+STACK_OFFSET); \
|
|
|
|
if (!esp_ptr_executable(callers[N]) \
|
|
|
|
|| callers[N] == (void*) HEAP_ARCH_INVALID_PC) { \
|
2019-11-20 06:47:07 -05:00
|
|
|
callers[N] = 0; \
|
2018-12-12 12:29:47 -05:00
|
|
|
return; \
|
|
|
|
} \
|
|
|
|
} while(0)
|
|
|
|
|
|
|
|
/* Static function to read the call stack for a traced heap call.
|
|
|
|
|
|
|
|
Calls to __builtin_return_address are "unrolled" via TEST_STACK macro as gcc requires the
|
|
|
|
argument to be a compile-time constant.
|
|
|
|
*/
|
2023-04-03 09:16:55 -04:00
|
|
|
static HEAP_IRAM_ATTR __attribute__((noinline)) void get_call_stack(void **callers)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
|
|
|
bzero(callers, sizeof(void *) * STACK_DEPTH);
|
|
|
|
TEST_STACK(0);
|
|
|
|
TEST_STACK(1);
|
|
|
|
TEST_STACK(2);
|
|
|
|
TEST_STACK(3);
|
|
|
|
TEST_STACK(4);
|
|
|
|
TEST_STACK(5);
|
|
|
|
TEST_STACK(6);
|
|
|
|
TEST_STACK(7);
|
|
|
|
TEST_STACK(8);
|
|
|
|
TEST_STACK(9);
|
2023-05-25 23:29:13 -04:00
|
|
|
TEST_STACK(10);
|
|
|
|
TEST_STACK(11);
|
|
|
|
TEST_STACK(12);
|
|
|
|
TEST_STACK(13);
|
|
|
|
TEST_STACK(14);
|
|
|
|
TEST_STACK(15);
|
|
|
|
TEST_STACK(16);
|
|
|
|
TEST_STACK(17);
|
|
|
|
TEST_STACK(18);
|
|
|
|
TEST_STACK(19);
|
|
|
|
TEST_STACK(20);
|
|
|
|
TEST_STACK(21);
|
|
|
|
TEST_STACK(22);
|
|
|
|
TEST_STACK(23);
|
|
|
|
TEST_STACK(24);
|
|
|
|
TEST_STACK(25);
|
|
|
|
TEST_STACK(26);
|
|
|
|
TEST_STACK(27);
|
|
|
|
TEST_STACK(28);
|
|
|
|
TEST_STACK(29);
|
|
|
|
TEST_STACK(30);
|
|
|
|
TEST_STACK(31);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
2023-06-05 23:08:50 -04:00
|
|
|
ESP_STATIC_ASSERT(STACK_DEPTH >= 0 && STACK_DEPTH <= 32, "CONFIG_HEAP_TRACING_STACK_DEPTH must be in range 0-32");
|
2018-12-12 12:29:47 -05:00
|
|
|
|
|
|
|
typedef enum {
|
2024-03-08 07:58:13 -05:00
|
|
|
TRACE_MALLOC_ALIGNED,
|
2018-12-12 12:29:47 -05:00
|
|
|
TRACE_MALLOC_DEFAULT
|
|
|
|
} trace_malloc_mode_t;
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
void *__real_heap_caps_malloc_base( size_t size, uint32_t caps);
|
|
|
|
void *__real_heap_caps_realloc_base( void *ptr, size_t size, uint32_t caps);
|
|
|
|
void *__real_heap_caps_aligned_alloc_base(size_t alignment, size_t size, uint32_t caps);
|
|
|
|
void __real_heap_caps_free(void *p);
|
2018-12-12 12:29:47 -05:00
|
|
|
|
|
|
|
/* trace any 'malloc' event */
|
2024-03-08 07:58:13 -05:00
|
|
|
static HEAP_IRAM_ATTR __attribute__((noinline)) void *trace_malloc(size_t alignment, size_t size, uint32_t caps, trace_malloc_mode_t mode)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
|
|
|
uint32_t ccount = get_ccount();
|
2024-03-08 07:58:13 -05:00
|
|
|
void *p = NULL;
|
2018-12-12 12:29:47 -05:00
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
if (mode == TRACE_MALLOC_DEFAULT) {
|
|
|
|
p = __real_heap_caps_malloc_base(size, caps);
|
|
|
|
} else {
|
|
|
|
p = __real_heap_caps_aligned_alloc_base(alignment, size, caps);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
|
|
|
heap_trace_record_t rec = {
|
|
|
|
.address = p,
|
|
|
|
.ccount = ccount,
|
|
|
|
.size = size,
|
|
|
|
};
|
|
|
|
get_call_stack(rec.alloced_by);
|
|
|
|
record_allocation(&rec);
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* trace any 'realloc' event */
|
2024-03-08 07:58:13 -05:00
|
|
|
static HEAP_IRAM_ATTR __attribute__((noinline)) void *trace_realloc(void *p, size_t size, uint32_t caps)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
|
|
|
void *callers[STACK_DEPTH];
|
|
|
|
uint32_t ccount = get_ccount();
|
|
|
|
void *r;
|
|
|
|
|
|
|
|
/* trace realloc as free-then-alloc */
|
|
|
|
get_call_stack(callers);
|
|
|
|
record_free(p, callers);
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
r = __real_heap_caps_realloc_base(p, size, caps);
|
|
|
|
|
2018-12-12 12:29:47 -05:00
|
|
|
/* realloc with zero size is a free */
|
|
|
|
if (size != 0) {
|
|
|
|
heap_trace_record_t rec = {
|
|
|
|
.address = r,
|
|
|
|
.ccount = ccount,
|
|
|
|
.size = size,
|
|
|
|
};
|
|
|
|
memcpy(rec.alloced_by, callers, sizeof(void *) * STACK_DEPTH);
|
|
|
|
record_allocation(&rec);
|
|
|
|
}
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
/* trace any 'free' event */
|
|
|
|
static HEAP_IRAM_ATTR __attribute__((noinline)) void trace_free(void *p)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
2024-03-08 07:58:13 -05:00
|
|
|
void *callers[STACK_DEPTH];
|
|
|
|
get_call_stack(callers);
|
|
|
|
record_free(p, callers);
|
2018-12-12 12:29:47 -05:00
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
__real_heap_caps_free(p);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
HEAP_IRAM_ATTR void __wrap_heap_caps_free(void *p) {
|
|
|
|
trace_free(p);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
HEAP_IRAM_ATTR void *__wrap_heap_caps_realloc_base(void *ptr, size_t size, uint32_t caps)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
2024-03-08 07:58:13 -05:00
|
|
|
return trace_realloc(ptr, size, caps);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
HEAP_IRAM_ATTR void *__wrap_heap_caps_malloc_base(size_t size, uint32_t caps)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
2024-03-08 07:58:13 -05:00
|
|
|
return trace_malloc(0, size, caps, TRACE_MALLOC_DEFAULT);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|
|
|
|
|
2024-03-08 07:58:13 -05:00
|
|
|
HEAP_IRAM_ATTR void *__wrap_heap_caps_aligned_alloc_base(size_t alignment, size_t size, uint32_t caps)
|
2018-12-12 12:29:47 -05:00
|
|
|
{
|
2024-03-08 07:58:13 -05:00
|
|
|
(void)alignment;
|
|
|
|
return trace_malloc(alignment, size, caps, TRACE_MALLOC_ALIGNED);
|
2018-12-12 12:29:47 -05:00
|
|
|
}
|