Loading...
/* * Copyright (c) 2016 Intel Corporation * * SPDX-License-Identifier: Apache-2.0 */ #include <ztest.h> #include <sys/atomic.h> #define LOOP 10 #define STACK_SIZE (512 + CONFIG_TEST_EXTRA_STACKSIZE) #define THREAD_NUM 4 #define SLAB_NUM 2 #define TIMEOUT 200 #define BLK_NUM 3 #define BLK_ALIGN 8 #define BLK_SIZE1 16 #define BLK_SIZE2 8 /* Blocks per slab. Note this number carefully, because if it is * smaller than this the test can deadlock. There are 4 threads * allocating 3 blocks each, so it's possible in exhaustion situations * for all four threads to have already allocated 2 blocks, with three * of them having gone to sleep waiting on a fourth. There must * remain at least one free block for that final allocation, otherwise * all four threads will end up sleeping forever (or until the * allocation timeout hits and the test fails). */ #define SLAB_BLOCKS (THREAD_NUM * (BLK_NUM - 1) + 1) K_MEM_SLAB_DEFINE(mslab1, BLK_SIZE1, SLAB_BLOCKS, BLK_ALIGN); static struct k_mem_slab mslab2, *slabs[SLAB_NUM] = { &mslab1, &mslab2 }; static K_THREAD_STACK_ARRAY_DEFINE(tstack, THREAD_NUM, STACK_SIZE); static struct k_thread tdata[THREAD_NUM]; static char __aligned(BLK_ALIGN) tslab[BLK_SIZE2 * SLAB_BLOCKS]; static struct k_sem sync_sema; static atomic_t slab_id; /* thread entry simply invoke the APIs*/ static void tmslab_api(void *p1, void *p2, void *p3) { void *block[BLK_NUM]; struct k_mem_slab *slab = slabs[atomic_inc(&slab_id) % SLAB_NUM]; int i = LOOP, ret; while (i--) { (void)memset(block, 0, sizeof(block)); for (int i = 0; i < BLK_NUM; i++) { ret = k_mem_slab_alloc(slab, &block[i], TIMEOUT); zassert_false(ret, "memory is not allocated"); } for (int i = 0; i < BLK_NUM; i++) { if (block[i]) { k_mem_slab_free(slab, &block[i]); block[i] = NULL; } } } k_sem_give(&sync_sema); } /* test cases*/ /** * @brief Verify alloc and free from multiple equal priority threads * * @details Test creates 4 preemptive threads of equal priority. Then * validates the synchronization of threads by allocating and * freeing up the memory blocks in memory slab. * * @ingroup kernel_memory_slab_tests */ void test_mslab_threadsafe(void) { k_tid_t tid[THREAD_NUM]; k_mem_slab_init(&mslab2, tslab, BLK_SIZE2, SLAB_BLOCKS); k_sem_init(&sync_sema, 0, THREAD_NUM); /* create multiple threads to invoke same memory slab APIs*/ for (int i = 0; i < THREAD_NUM; i++) { tid[i] = k_thread_create(&tdata[i], tstack[i], STACK_SIZE, tmslab_api, NULL, NULL, NULL, K_PRIO_PREEMPT(1), 0, K_NO_WAIT); } /* TESTPOINT: all threads complete and exit the entry function*/ for (int i = 0; i < THREAD_NUM; i++) { k_sem_take(&sync_sema, K_FOREVER); } /* test case tear down*/ for (int i = 0; i < THREAD_NUM; i++) { k_thread_abort(tid[i]); } } |