Changes in / [380553c:2616a75b] in mainline
- Files:
-
- 33 edited
Legend:
- Unmodified
- Added
- Removed
-
kernel/arch/ia64/src/drivers/ski.c
r380553c r2616a75b 150 150 151 151 if (instance) { 152 instance->thread = thread_create(kskipoll, instance, TASK, 153 THREAD_FLAG_UNCOUNTED, "kskipoll");152 instance->thread = thread_create(kskipoll, instance, TASK, 0, 153 "kskipoll", true); 154 154 155 155 if (!instance->thread) { -
kernel/arch/sparc64/src/drivers/niagara.c
r380553c r2616a75b 184 184 185 185 instance = malloc(sizeof(niagara_instance_t), FRAME_ATOMIC); 186 instance->thread = thread_create(kniagarapoll, NULL, TASK, 187 THREAD_FLAG_UNCOUNTED, "kniagarapoll");186 instance->thread = thread_create(kniagarapoll, NULL, TASK, 0, 187 "kniagarapoll", true); 188 188 189 189 if (!instance->thread) { -
kernel/arch/sparc64/src/proc/sun4u/scheduler.c
r380553c r2616a75b 51 51 void before_thread_runs_arch(void) 52 52 { 53 if ( THREAD->uspace) {53 if ((THREAD->flags & THREAD_FLAG_USPACE)) { 54 54 /* 55 55 * Write kernel stack address to %g6 of the alternate and … … 74 74 void after_thread_ran_arch(void) 75 75 { 76 if ( THREAD->uspace) {77 /* sample the state of the userspace window buffer */ 76 if ((THREAD->flags & THREAD_FLAG_USPACE)) { 77 /* sample the state of the userspace window buffer */ 78 78 THREAD->arch.uspace_window_buffer = (uint8_t *) read_from_ag_g7(); 79 79 } -
kernel/arch/sparc64/src/proc/sun4v/scheduler.c
r380553c r2616a75b 54 54 void before_thread_runs_arch(void) 55 55 { 56 if ( THREAD->uspace) {56 if ((THREAD->flags & THREAD_FLAG_USPACE)) { 57 57 uint64_t sp = (uintptr_t) THREAD->kstack + STACK_SIZE - 58 58 (STACK_BIAS + ALIGN_UP(STACK_ITEM_SIZE, STACK_ALIGNMENT)); … … 66 66 void after_thread_ran_arch(void) 67 67 { 68 if ( THREAD->uspace) {69 /* sample the state of the userspace window buffer */ 68 if ((THREAD->flags & THREAD_FLAG_USPACE)) { 69 /* sample the state of the userspace window buffer */ 70 70 THREAD->arch.uspace_window_buffer = 71 71 (uint8_t *) asi_u64_read(ASI_SCRATCHPAD, SCRATCHPAD_WBUF); -
kernel/arch/sparc64/src/proc/thread.c
r380553c r2616a75b 61 61 void thread_create_arch(thread_t *t) 62 62 { 63 if ((t-> uspace) && (!t->arch.uspace_window_buffer))63 if ((t->flags & THREAD_FLAG_USPACE) && (!t->arch.uspace_window_buffer)) 64 64 { 65 65 /* -
kernel/genarch/src/kbrd/kbrd.c
r380553c r2616a75b 158 158 = malloc(sizeof(kbrd_instance_t), FRAME_ATOMIC); 159 159 if (instance) { 160 instance->thread = thread_create(kkbrd, (void *) instance,161 TASK, THREAD_FLAG_NONE, "kkbrd");160 instance->thread 161 = thread_create(kkbrd, (void *) instance, TASK, 0, "kkbrd", false); 162 162 163 163 if (!instance->thread) { -
kernel/genarch/src/mm/as_ht.c
r380553c r2616a75b 78 78 hash_table_create(&page_ht, PAGE_HT_ENTRIES, 2, &ht_operations); 79 79 mutex_initialize(&page_ht_lock, MUTEX_PASSIVE); 80 pte_cache = slab_cache_create("pte_ t", sizeof(pte_t), 0,81 NULL, NULL,SLAB_CACHE_MAGDEFERRED);80 pte_cache = slab_cache_create("pte_cache", sizeof(pte_t), 0, NULL, NULL, 81 SLAB_CACHE_MAGDEFERRED); 82 82 } 83 83 -
kernel/genarch/src/srln/srln.c
r380553c r2616a75b 130 130 = malloc(sizeof(srln_instance_t), FRAME_ATOMIC); 131 131 if (instance) { 132 instance->thread = thread_create(ksrln, (void *) instance,133 TASK, THREAD_FLAG_NONE, "ksrln");132 instance->thread 133 = thread_create(ksrln, (void *) instance, TASK, 0, "ksrln", false); 134 134 135 135 if (!instance->thread) { -
kernel/generic/include/proc/thread.h
r380553c r2616a75b 54 54 55 55 /* Thread flags */ 56 typedef enum { 57 THREAD_FLAG_NONE = 0, 58 /** Thread executes in user space. */ 59 THREAD_FLAG_USPACE = (1 << 0), 60 /** Thread will be attached by the caller. */ 61 THREAD_FLAG_NOATTACH = (1 << 1), 62 /** Thread accounting doesn't affect accumulated task accounting. */ 63 THREAD_FLAG_UNCOUNTED = (1 << 2) 64 } thread_flags_t; 56 57 /** Thread cannot be migrated to another CPU. 58 * 59 * When using this flag, the caller must set cpu in the thread_t 60 * structure manually before calling thread_ready (even on uniprocessor). 61 * 62 */ 63 #define THREAD_FLAG_WIRED (1 << 0) 64 65 /** Thread was migrated to another CPU and has not run yet. */ 66 #define THREAD_FLAG_STOLEN (1 << 1) 67 68 /** Thread executes in userspace. */ 69 #define THREAD_FLAG_USPACE (1 << 2) 70 71 /** Thread will be attached by the caller. */ 72 #define THREAD_FLAG_NOATTACH (1 << 3) 65 73 66 74 /** Thread structure. There is one per thread. */ … … 139 147 140 148 fpu_context_t *saved_fpu_context; 141 boolfpu_context_exists;149 int fpu_context_exists; 142 150 143 151 /* … … 146 154 * thread. This disables migration. 147 155 */ 148 boolfpu_context_engaged;156 int fpu_context_engaged; 149 157 150 158 /* The thread will not be migrated if nomigrate is non-zero. */ 151 unsignedint nomigrate;152 153 /** Thread state. */159 int nomigrate; 160 161 /** Thread's state. */ 154 162 state_t state; 155 156 /** Thread CPU. */ 163 /** Thread's flags. */ 164 unsigned int flags; 165 166 /** Thread's CPU. */ 157 167 cpu_t *cpu; 158 168 /** Containing task. */ 159 169 task_t *task; 160 /** Thread is wired to CPU. */161 bool wired;162 /** Thread was migrated to another CPU and has not run yet. */163 bool stolen;164 /** Thread is executed in user space. */165 bool uspace;166 170 167 171 /** Ticks before preemption. */ … … 212 216 extern void thread_init(void); 213 217 extern thread_t *thread_create(void (*)(void *), void *, task_t *, 214 thread_flags_t, const char *); 215 extern void thread_wire(thread_t *, cpu_t *); 218 unsigned int, const char *, bool); 216 219 extern void thread_attach(thread_t *, task_t *); 217 220 extern void thread_ready(thread_t *); -
kernel/generic/src/adt/btree.c
r380553c r2616a75b 71 71 void btree_init(void) 72 72 { 73 btree_node_slab = slab_cache_create("btree_node_ t",73 btree_node_slab = slab_cache_create("btree_node_slab", 74 74 sizeof(btree_node_t), 0, NULL, NULL, SLAB_CACHE_MAGDEFERRED); 75 75 } -
kernel/generic/src/console/cmd.c
r380553c r2616a75b 724 724 thread_t *thread; 725 725 if ((thread = thread_create((void (*)(void *)) cmd_call0, 726 (void *) argv, TASK, THREAD_FLAG_NONE, "call0"))) { 726 (void *) argv, TASK, THREAD_FLAG_WIRED, "call0", false))) { 727 irq_spinlock_lock(&thread->lock, true); 728 thread->cpu = &cpus[i]; 729 irq_spinlock_unlock(&thread->lock, true); 730 727 731 printf("cpu%u: ", i); 728 thread_wire(thread, &cpus[i]);732 729 733 thread_ready(thread); 730 734 thread_join(thread); -
kernel/generic/src/ipc/ipc.c
r380553c r2616a75b 670 670 void ipc_init(void) 671 671 { 672 ipc_call_slab = slab_cache_create(" call_t", sizeof(call_t), 0, NULL,672 ipc_call_slab = slab_cache_create("ipc_call", sizeof(call_t), 0, NULL, 673 673 NULL, 0); 674 ipc_answerbox_slab = slab_cache_create(" answerbox_t",674 ipc_answerbox_slab = slab_cache_create("ipc_answerbox", 675 675 sizeof(answerbox_t), 0, NULL, NULL, 0); 676 676 } -
kernel/generic/src/ipc/kbox.c
r380553c r2616a75b 244 244 245 245 /* Create a kbox thread */ 246 thread_t *kb_thread = thread_create(kbox_thread_proc, NULL, task, 247 THREAD_FLAG_NONE, "kbox");246 thread_t *kb_thread = thread_create(kbox_thread_proc, NULL, task, 0, 247 "kbox", false); 248 248 if (!kb_thread) { 249 249 mutex_unlock(&task->kb.cleanup_lock); -
kernel/generic/src/lib/ra.c
r380553c r2616a75b 424 424 void ra_init(void) 425 425 { 426 ra_segment_cache = slab_cache_create(" ra_segment_t",426 ra_segment_cache = slab_cache_create("segment_cache", 427 427 sizeof(ra_segment_t), 0, NULL, NULL, SLAB_CACHE_MAGDEFERRED); 428 428 } -
kernel/generic/src/main/kinit.c
r380553c r2616a75b 116 116 * Just a beautification. 117 117 */ 118 thread = thread_create(kmp, NULL, TASK, 119 THREAD_FLAG_UNCOUNTED, "kmp"); 118 thread = thread_create(kmp, NULL, TASK, THREAD_FLAG_WIRED, "kmp", true); 120 119 if (thread != NULL) { 121 thread_wire(thread, &cpus[0]); 120 irq_spinlock_lock(&thread->lock, false); 121 thread->cpu = &cpus[0]; 122 irq_spinlock_unlock(&thread->lock, false); 122 123 thread_ready(thread); 123 124 } else … … 133 134 134 135 for (i = 0; i < config.cpu_count; i++) { 135 thread = thread_create(kcpulb, NULL, TASK, 136 THREAD_FLAG_UNCOUNTED, "kcpulb"); 136 thread = thread_create(kcpulb, NULL, TASK, THREAD_FLAG_WIRED, "kcpulb", true); 137 137 if (thread != NULL) { 138 thread_wire(thread, &cpus[i]); 138 irq_spinlock_lock(&thread->lock, false); 139 thread->cpu = &cpus[i]; 140 irq_spinlock_unlock(&thread->lock, false); 139 141 thread_ready(thread); 140 142 } else … … 150 152 151 153 /* Start thread computing system load */ 152 thread = thread_create(kload, NULL, TASK, THREAD_FLAG_NONE, 153 "kload"); 154 thread = thread_create(kload, NULL, TASK, 0, "kload", false); 154 155 if (thread != NULL) 155 156 thread_ready(thread); … … 162 163 * Create kernel console. 163 164 */ 164 thread = thread_create(kconsole_thread, NULL, TASK, 165 THREAD_FLAG_NONE, "kconsole"); 165 thread = thread_create(kconsole_thread, NULL, TASK, 0, "kconsole", false); 166 166 if (thread != NULL) 167 167 thread_ready(thread); -
kernel/generic/src/main/main.c
r380553c r2616a75b 276 276 * Create the first thread. 277 277 */ 278 thread_t *kinit_thread = thread_create(kinit, NULL, kernel,279 THREAD_FLAG_UNCOUNTED, "kinit");278 thread_t *kinit_thread = 279 thread_create(kinit, NULL, kernel, 0, "kinit", true); 280 280 if (!kinit_thread) 281 281 panic("Cannot create kinit thread."); -
kernel/generic/src/mm/as.c
r380553c r2616a75b 130 130 as_arch_init(); 131 131 132 as_slab = slab_cache_create("as_ t", sizeof(as_t), 0,132 as_slab = slab_cache_create("as_slab", sizeof(as_t), 0, 133 133 as_constructor, as_destructor, SLAB_CACHE_MAGDEFERRED); 134 134 -
kernel/generic/src/mm/slab.c
r380553c r2616a75b 891 891 { 892 892 /* Initialize magazine cache */ 893 _slab_cache_create(&mag_cache, "slab_magazine _t",893 _slab_cache_create(&mag_cache, "slab_magazine", 894 894 sizeof(slab_magazine_t) + SLAB_MAG_SIZE * sizeof(void*), 895 895 sizeof(uintptr_t), NULL, NULL, SLAB_CACHE_NOMAGAZINE | … … 897 897 898 898 /* Initialize slab_cache cache */ 899 _slab_cache_create(&slab_cache_cache, "slab_cache _cache",899 _slab_cache_create(&slab_cache_cache, "slab_cache", 900 900 sizeof(slab_cache_cache), sizeof(uintptr_t), NULL, NULL, 901 901 SLAB_CACHE_NOMAGAZINE | SLAB_CACHE_SLINSIDE); 902 902 903 903 /* Initialize external slab cache */ 904 slab_extern_cache = slab_cache_create("slab_ t", sizeof(slab_t), 0,904 slab_extern_cache = slab_cache_create("slab_extern", sizeof(slab_t), 0, 905 905 NULL, NULL, SLAB_CACHE_SLINSIDE | SLAB_CACHE_MAGDEFERRED); 906 906 -
kernel/generic/src/proc/program.c
r380553c r2616a75b 102 102 */ 103 103 prg->main_thread = thread_create(uinit, kernel_uarg, prg->task, 104 THREAD_FLAG_USPACE, "uinit" );104 THREAD_FLAG_USPACE, "uinit", false); 105 105 if (!prg->main_thread) { 106 106 free(kernel_uarg); -
kernel/generic/src/proc/scheduler.c
r380553c r2616a75b 98 98 else { 99 99 fpu_init(); 100 THREAD->fpu_context_exists = true;100 THREAD->fpu_context_exists = 1; 101 101 } 102 102 #endif … … 142 142 143 143 /* Don't prevent migration */ 144 CPU->fpu_owner->fpu_context_engaged = false;144 CPU->fpu_owner->fpu_context_engaged = 0; 145 145 irq_spinlock_unlock(&CPU->fpu_owner->lock, false); 146 146 CPU->fpu_owner = NULL; … … 163 163 } 164 164 fpu_init(); 165 THREAD->fpu_context_exists = true;165 THREAD->fpu_context_exists = 1; 166 166 } 167 167 168 168 CPU->fpu_owner = THREAD; 169 THREAD->fpu_context_engaged = true;169 THREAD->fpu_context_engaged = 1; 170 170 irq_spinlock_unlock(&THREAD->lock, false); 171 171 … … 248 248 249 249 /* 250 * Clear the stolen flag so that it can be migrated250 * Clear the THREAD_FLAG_STOLEN flag so that t can be migrated 251 251 * when load balancing needs emerge. 252 252 */ 253 thread-> stolen = false;253 thread->flags &= ~THREAD_FLAG_STOLEN; 254 254 irq_spinlock_unlock(&thread->lock, false); 255 255 … … 630 630 irq_spinlock_lock(&thread->lock, false); 631 631 632 if ((!thread->wired) && (!thread->stolen) && 633 (!thread->nomigrate) && 634 (!thread->fpu_context_engaged)) { 632 if (!(thread->flags & THREAD_FLAG_WIRED) && 633 !(thread->flags & THREAD_FLAG_STOLEN) && 634 !thread->nomigrate && 635 !thread->fpu_context_engaged) { 635 636 /* 636 637 * Remove thread from ready queue. … … 669 670 #endif 670 671 671 thread-> stolen = true;672 thread->flags |= THREAD_FLAG_STOLEN; 672 673 thread->state = Entering; 673 674 -
kernel/generic/src/proc/task.c
r380553c r2616a75b 90 90 TASK = NULL; 91 91 avltree_create(&tasks_tree); 92 task_slab = slab_cache_create("task_ t", sizeof(task_t), 0,92 task_slab = slab_cache_create("task_slab", sizeof(task_t), 0, 93 93 tsk_constructor, NULL, 0); 94 94 } -
kernel/generic/src/proc/thread.c
r380553c r2616a75b 191 191 kmflags |= FRAME_LOWMEM; 192 192 kmflags &= ~FRAME_HIGHMEM; 193 193 194 194 thread->kstack = (uint8_t *) frame_alloc(STACK_FRAMES, FRAME_KA | kmflags); 195 195 if (!thread->kstack) { … … 236 236 237 237 atomic_set(&nrdy, 0); 238 thread_slab = slab_cache_create("thread_ t", sizeof(thread_t), 0,238 thread_slab = slab_cache_create("thread_slab", sizeof(thread_t), 0, 239 239 thr_constructor, thr_destructor, 0); 240 240 241 241 #ifdef CONFIG_FPU 242 fpu_context_slab = slab_cache_create("fpu_ context_t",243 sizeof(fpu_context_t),FPU_CONTEXT_ALIGN, NULL, NULL, 0);242 fpu_context_slab = slab_cache_create("fpu_slab", sizeof(fpu_context_t), 243 FPU_CONTEXT_ALIGN, NULL, NULL, 0); 244 244 #endif 245 245 … … 247 247 } 248 248 249 /** Wire thread to the given CPU 250 * 251 * @param cpu CPU to wire the thread to. 252 * 253 */ 254 void thread_wire(thread_t *thread, cpu_t *cpu) 249 /** Make thread ready 250 * 251 * Switch thread to the ready state. 252 * 253 * @param thread Thread to make ready. 254 * 255 */ 256 void thread_ready(thread_t *thread) 255 257 { 256 258 irq_spinlock_lock(&thread->lock, true); 257 thread->cpu = cpu;258 thread->wired = true;259 irq_spinlock_unlock(&thread->lock, true);260 }261 262 /** Make thread ready263 *264 * Switch thread to the ready state.265 *266 * @param thread Thread to make ready.267 *268 */269 void thread_ready(thread_t *thread)270 {271 irq_spinlock_lock(&thread->lock, true);272 259 273 260 ASSERT(thread->state != Ready); 274 261 275 int i = (thread->priority < RQ_COUNT - 1) ?276 ++thread->priority : thread->priority;277 278 cpu_t *cpu ;279 if (thread-> wired || thread->nomigrate || thread->fpu_context_engaged) {262 int i = (thread->priority < RQ_COUNT - 1) 263 ? ++thread->priority : thread->priority; 264 265 cpu_t *cpu = CPU; 266 if (thread->flags & THREAD_FLAG_WIRED) { 280 267 ASSERT(thread->cpu != NULL); 281 268 cpu = thread->cpu; 282 } else 283 cpu = CPU; 284 269 } 285 270 thread->state = Ready; 286 271 … … 313 298 * @param flags Thread flags. 314 299 * @param name Symbolic name (a copy is made). 300 * @param uncounted Thread's accounting doesn't affect accumulated task 301 * accounting. 315 302 * 316 303 * @return New thread's structure on success, NULL on failure. … … 318 305 */ 319 306 thread_t *thread_create(void (* func)(void *), void *arg, task_t *task, 320 thread_flags_t flags, const char *name)307 unsigned int flags, const char *name, bool uncounted) 321 308 { 322 309 thread_t *thread = (thread_t *) slab_alloc(thread_slab, 0); … … 348 335 thread->ucycles = 0; 349 336 thread->kcycles = 0; 350 thread->uncounted = 351 ((flags & THREAD_FLAG_UNCOUNTED) == THREAD_FLAG_UNCOUNTED); 337 thread->uncounted = uncounted; 352 338 thread->priority = -1; /* Start in rq[0] */ 353 339 thread->cpu = NULL; 354 thread->wired = false; 355 thread->stolen = false; 356 thread->uspace = 357 ((flags & THREAD_FLAG_USPACE) == THREAD_FLAG_USPACE); 358 340 thread->flags = flags; 359 341 thread->nomigrate = 0; 360 342 thread->state = Entering; … … 374 356 thread->task = task; 375 357 376 thread->fpu_context_exists = false;377 thread->fpu_context_engaged = false;358 thread->fpu_context_exists = 0; 359 thread->fpu_context_engaged = 0; 378 360 379 361 avltree_node_initialize(&thread->threads_tree_node); … … 389 371 thread_create_arch(thread); 390 372 391 if ( (flags & THREAD_FLAG_NOATTACH) != THREAD_FLAG_NOATTACH)373 if (!(flags & THREAD_FLAG_NOATTACH)) 392 374 thread_attach(thread, task); 393 375 … … 455 437 456 438 /* Must not count kbox thread into lifecount */ 457 if (thread-> uspace)439 if (thread->flags & THREAD_FLAG_USPACE) 458 440 atomic_inc(&task->lifecount); 459 441 … … 477 459 void thread_exit(void) 478 460 { 479 if (THREAD-> uspace) {461 if (THREAD->flags & THREAD_FLAG_USPACE) { 480 462 #ifdef CONFIG_UDEBUG 481 463 /* Generate udebug THREAD_E event */ 482 464 udebug_thread_e_event(); 483 465 484 466 /* 485 467 * This thread will not execute any code or system calls from … … 524 506 { 525 507 ASSERT(THREAD); 526 508 527 509 THREAD->nomigrate++; 528 510 } … … 533 515 ASSERT(THREAD); 534 516 ASSERT(THREAD->nomigrate > 0); 535 536 if (THREAD->nomigrate > 0) 537 THREAD->nomigrate--; 517 518 THREAD->nomigrate--; 538 519 } 539 520 … … 884 865 885 866 thread_t *thread = thread_create(uinit, kernel_uarg, TASK, 886 THREAD_FLAG_USPACE | THREAD_FLAG_NOATTACH, namebuf );867 THREAD_FLAG_USPACE | THREAD_FLAG_NOATTACH, namebuf, false); 887 868 if (thread) { 888 869 if (uspace_thread_id != NULL) { -
kernel/generic/src/sysinfo/sysinfo.c
r380553c r2616a75b 97 97 void sysinfo_init(void) 98 98 { 99 sysinfo_item_slab = slab_cache_create("sysinfo_item_ t",99 sysinfo_item_slab = slab_cache_create("sysinfo_item_slab", 100 100 sizeof(sysinfo_item_t), 0, sysinfo_item_constructor, 101 101 sysinfo_item_destructor, SLAB_CACHE_MAGDEFERRED); -
kernel/generic/src/udebug/udebug.c
r380553c r2616a75b 410 410 411 411 mutex_lock(&thread->udebug.lock); 412 unsigned int flags = thread->flags; 412 413 413 414 /* Only process userspace threads. */ 414 if ( thread->uspace) {415 if ((flags & THREAD_FLAG_USPACE) != 0) { 415 416 /* Prevent any further debug activity in thread. */ 416 417 thread->udebug.active = false; -
kernel/generic/src/udebug/udebug_ops.c
r380553c r2616a75b 95 95 96 96 /* Verify that 'thread' is a userspace thread. */ 97 if ( !thread->uspace) {97 if ((thread->flags & THREAD_FLAG_USPACE) == 0) { 98 98 /* It's not, deny its existence */ 99 99 irq_spinlock_unlock(&thread->lock, true); … … 200 200 201 201 mutex_lock(&thread->udebug.lock); 202 if ( thread->uspace) {202 if ((thread->flags & THREAD_FLAG_USPACE) != 0) { 203 203 thread->udebug.active = true; 204 204 mutex_unlock(&thread->udebug.lock); … … 393 393 394 394 irq_spinlock_lock(&thread->lock, false); 395 bool uspace = thread->uspace;395 int flags = thread->flags; 396 396 irq_spinlock_unlock(&thread->lock, false); 397 397 398 398 /* Not interested in kernel threads. */ 399 if ( !uspace)399 if ((flags & THREAD_FLAG_USPACE) == 0) 400 400 continue; 401 401 -
kernel/test/mm/falloc1.c
r380553c r2616a75b 37 37 #include <align.h> 38 38 39 #define MAX_FRAMES 1024 U39 #define MAX_FRAMES 1024 40 40 #define MAX_ORDER 8 41 41 #define TEST_RUNS 2 42 42 43 const char *test_falloc1(void) 44 { 43 const char *test_falloc1(void) { 44 uintptr_t *frames 45 = (uintptr_t *) malloc(MAX_FRAMES * sizeof(uintptr_t), 0); 46 int results[MAX_ORDER + 1]; 47 48 int i, order, run; 49 int allocated; 50 45 51 if (TEST_RUNS < 2) 46 52 return "Test is compiled with TEST_RUNS < 2"; 47 53 48 uintptr_t *frames = (uintptr_t *)49 malloc(MAX_FRAMES * sizeof(uintptr_t), 0);50 54 if (frames == NULL) 51 55 return "Unable to allocate frames"; 52 56 53 unsigned int results[MAX_ORDER + 1]; 54 for (unsigned int run = 0; run < TEST_RUNS; run++) { 55 for (unsigned int order = 0; order <= MAX_ORDER; order++) { 56 TPRINTF("Allocating %u frames blocks ... ", 1 << order); 57 for (run = 0; run < TEST_RUNS; run++) { 58 for (order = 0; order <= MAX_ORDER; order++) { 59 TPRINTF("Allocating %d frames blocks ... ", 1 << order); 57 60 58 unsigned int allocated = 0; 59 for (unsigned int i = 0; i < (MAX_FRAMES >> order); i++) { 60 frames[allocated] = (uintptr_t) 61 frame_alloc(order, FRAME_ATOMIC | FRAME_KA); 61 allocated = 0; 62 for (i = 0; i < MAX_FRAMES >> order; i++) { 63 frames[allocated] = (uintptr_t) frame_alloc(order, FRAME_ATOMIC | FRAME_KA); 62 64 63 if (ALIGN_UP(frames[allocated], FRAME_SIZE << order) != 64 frames[allocated]) { 65 TPRINTF("Block at address %p (size %u) is not aligned\n", 65 if (ALIGN_UP(frames[allocated], FRAME_SIZE << order) != frames[allocated]) { 66 TPRINTF("Block at address %p (size %dK) is not aligned\n", 66 67 (void *) frames[allocated], (FRAME_SIZE << order) >> 10); 67 68 return "Test failed"; … … 86 87 TPRINTF("Deallocating ... "); 87 88 88 for ( unsigned inti = 0; i < allocated; i++)89 for (i = 0; i < allocated; i++) 89 90 frame_free(KA2PA(frames[i])); 90 91 -
kernel/test/mm/falloc2.c
r380553c r2616a75b 40 40 #include <arch.h> 41 41 42 #define MAX_FRAMES 256 U42 #define MAX_FRAMES 256 43 43 #define MAX_ORDER 8 44 44 … … 51 51 static void falloc(void *arg) 52 52 { 53 int order, run, allocated, i; 53 54 uint8_t val = THREAD->tid % THREADS; 55 size_t k; 54 56 55 void **frames = (void **) 56 malloc(MAX_FRAMES * sizeof(void *), FRAME_ATOMIC); 57 void **frames = (void **) malloc(MAX_FRAMES * sizeof(void *), FRAME_ATOMIC); 57 58 if (frames == NULL) { 58 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 59 "Unable to allocate frames\n", THREAD->tid, CPU->id); 59 TPRINTF("Thread #%" PRIu64 " (cpu%u): Unable to allocate frames\n", THREAD->tid, CPU->id); 60 60 atomic_inc(&thread_fail); 61 61 atomic_dec(&thread_count); … … 65 65 thread_detach(THREAD); 66 66 67 for (unsigned int run = 0; run < THREAD_RUNS; run++) { 68 for (unsigned int order = 0; order <= MAX_ORDER; order++) { 69 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 70 "Allocating %u frames blocks ... \n", THREAD->tid, 71 CPU->id, 1 << order); 67 for (run = 0; run < THREAD_RUNS; run++) { 68 for (order = 0; order <= MAX_ORDER; order++) { 69 TPRINTF("Thread #%" PRIu64 " (cpu%u): Allocating %d frames blocks ... \n", THREAD->tid, CPU->id, 1 << order); 72 70 73 unsigned int allocated = 0; 74 for (unsigned int i = 0; i < (MAX_FRAMES >> order); i++) { 75 frames[allocated] = 76 frame_alloc(order, FRAME_ATOMIC | FRAME_KA); 71 allocated = 0; 72 for (i = 0; i < (MAX_FRAMES >> order); i++) { 73 frames[allocated] = frame_alloc(order, FRAME_ATOMIC | FRAME_KA); 77 74 if (frames[allocated]) { 78 75 memsetb(frames[allocated], FRAME_SIZE << order, val); … … 82 79 } 83 80 84 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 85 "%u blocks allocated.\n", THREAD->tid, CPU->id, 86 allocated); 87 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 88 "Deallocating ... \n", THREAD->tid, CPU->id); 81 TPRINTF("Thread #%" PRIu64 " (cpu%u): %d blocks allocated.\n", THREAD->tid, CPU->id, allocated); 82 TPRINTF("Thread #%" PRIu64 " (cpu%u): Deallocating ... \n", THREAD->tid, CPU->id); 89 83 90 for (unsigned int i = 0; i < allocated; i++) { 91 for (size_t k = 0; k <= (((size_t) FRAME_SIZE << order) - 1); 92 k++) { 84 for (i = 0; i < allocated; i++) { 85 for (k = 0; k <= (((size_t) FRAME_SIZE << order) - 1); k++) { 93 86 if (((uint8_t *) frames[i])[k] != val) { 94 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 95 "Unexpected data (%c) in block %p offset %zu\n", 96 THREAD->tid, CPU->id, ((char *) frames[i])[k], 97 frames[i], k); 87 TPRINTF("Thread #%" PRIu64 " (cpu%u): Unexpected data (%c) in block %p offset %zu\n", 88 THREAD->tid, CPU->id, ((char *) frames[i])[k], frames[i], k); 98 89 atomic_inc(&thread_fail); 99 90 goto cleanup; … … 103 94 } 104 95 105 TPRINTF("Thread #%" PRIu64 " (cpu%u): " 106 "Finished run.\n", THREAD->tid, CPU->id); 96 TPRINTF("Thread #%" PRIu64 " (cpu%u): Finished run.\n", THREAD->tid, CPU->id); 107 97 } 108 98 } … … 111 101 free(frames); 112 102 113 TPRINTF("Thread #%" PRIu64 " (cpu%u): Exiting\n", 114 THREAD->tid, CPU->id); 103 TPRINTF("Thread #%" PRIu64 " (cpu%u): Exiting\n", THREAD->tid, CPU->id); 115 104 atomic_dec(&thread_count); 116 105 } … … 118 107 const char *test_falloc2(void) 119 108 { 109 unsigned int i; 110 120 111 atomic_set(&thread_count, THREADS); 121 112 atomic_set(&thread_fail, 0); 122 113 123 for (unsigned int i = 0; i < THREADS; i++) { 124 thread_t *thrd = thread_create(falloc, NULL, TASK, 125 THREAD_FLAG_NONE, "falloc2"); 114 for (i = 0; i < THREADS; i++) { 115 thread_t * thrd = thread_create(falloc, NULL, TASK, 0, "falloc", false); 126 116 if (!thrd) { 127 117 TPRINTF("Could not create thread %u\n", i); … … 132 122 133 123 while (atomic_get(&thread_count) > 0) { 134 TPRINTF("Threads left: %" PRIua "\n", 135 atomic_get(&thread_count)); 124 TPRINTF("Threads left: %" PRIua "\n", atomic_get(&thread_count)); 136 125 thread_sleep(1); 137 126 } -
kernel/test/mm/slab1.c
r380553c r2616a75b 155 155 156 156 semaphore_initialize(&thr_sem, 0); 157 for (i = 0; i < THREADS; i++) { 158 if (!(t = thread_create(slabtest, (void *) (sysarg_t) i, TASK, THREAD_FLAG_NONE, "slabtest"))) {157 for (i = 0; i < THREADS; i++) { 158 if (!(t = thread_create(slabtest, (void *) (sysarg_t) i, TASK, 0, "slabtest", false))) { 159 159 TPRINTF("Could not create thread %d\n", i); 160 160 } else -
kernel/test/mm/slab2.c
r380553c r2616a75b 52 52 void *olddata1 = NULL, *olddata2 = NULL; 53 53 54 cache1 = slab_cache_create(" test_cache1", ITEM_SIZE, 0, NULL, NULL, 0);55 cache2 = slab_cache_create(" test_cache2", ITEM_SIZE, 0, NULL, NULL, 0);54 cache1 = slab_cache_create("cache1_tst", ITEM_SIZE, 0, NULL, NULL, 0); 55 cache2 = slab_cache_create("cache2_tst", ITEM_SIZE, 0, NULL, NULL, 0); 56 56 57 57 TPRINTF("Allocating..."); … … 210 210 thr_cache = slab_cache_create("thread_cache", size, 0, NULL, NULL, 0); 211 211 semaphore_initialize(&thr_sem,0); 212 for (i = 0; i < THREADS; i++) { 213 if (!(t = thread_create(slabtest, NULL, TASK, THREAD_FLAG_NONE, "slabtest"))) {212 for (i = 0; i < THREADS; i++) { 213 if (!(t = thread_create(slabtest, NULL, TASK, 0, "slabtest", false))) { 214 214 TPRINTF("Could not create thread %d\n", i); 215 215 } else -
kernel/test/synch/semaphore1.c
r380553c r2616a75b 93 93 for (j = 0; j < (CONSUMERS + PRODUCERS) / 2; j++) { 94 94 for (k = 0; k < i; k++) { 95 thrd = thread_create(consumer, NULL, TASK, 96 THREAD_FLAG_NONE, "consumer"); 95 thrd = thread_create(consumer, NULL, TASK, 0, "consumer", false); 97 96 if (thrd) 98 97 thread_ready(thrd); … … 101 100 } 102 101 for (k = 0; k < (4 - i); k++) { 103 thrd = thread_create(producer, NULL, TASK, 104 THREAD_FLAG_NONE, "producer"); 102 thrd = thread_create(producer, NULL, TASK, 0, "producer", false); 105 103 if (thrd) 106 104 thread_ready(thrd); -
kernel/test/synch/semaphore2.c
r380553c r2616a75b 93 93 TPRINTF("Creating %" PRIu32 " consumers\n", k); 94 94 for (i = 0; i < k; i++) { 95 thrd = thread_create(consumer, NULL, TASK, 96 THREAD_FLAG_NONE, "consumer"); 95 thrd = thread_create(consumer, NULL, TASK, 0, "consumer", false); 97 96 if (thrd) 98 97 thread_ready(thrd); -
kernel/test/thread/thread1.c
r380553c r2616a75b 63 63 for (i = 0; i < THREADS; i++) { 64 64 thread_t *t; 65 if (!(t = thread_create(threadtest, NULL, TASK, 66 THREAD_FLAG_NONE, "threadtest"))) { 65 if (!(t = thread_create(threadtest, NULL, TASK, 0, "threadtest", false))) { 67 66 TPRINTF("Could not create thread %d\n", i); 68 67 break; -
tools/toolchain.sh
r380553c r2616a75b 55 55 BINUTILS_VERSION="2.22" 56 56 BINUTILS_RELEASE="" 57 GCC_VERSION="4.7. 1"57 GCC_VERSION="4.7.0" 58 58 GDB_VERSION="7.4" 59 59 … … 274 274 275 275 download_fetch "${BINUTILS_SOURCE}" "${BINUTILS}" "ee0f10756c84979622b992a4a61ea3f5" 276 download_fetch "${GCC_SOURCE}" "${GCC}" " 933e6f15f51c031060af64a9e14149ff"276 download_fetch "${GCC_SOURCE}" "${GCC}" "2a0f1d99fda235c29d40b561f81d9a77" 277 277 download_fetch "${GDB_SOURCE}" "${GDB}" "95a9a8305fed4d30a30a6dc28ff9d060" 278 278 }
Note:
See TracChangeset
for help on using the changeset viewer.