Changes in kernel/generic/src/udebug/udebug_ops.c [96b02eb9:ae5aa90] in mainline
- File:
-
- 1 edited
Legend:
- Unmodified
- Added
- Removed
-
kernel/generic/src/udebug/udebug_ops.c
r96b02eb9 rae5aa90 33 33 /** 34 34 * @file 35 * @brief 35 * @brief Udebug operations. 36 36 * 37 37 * Udebug operations on tasks and threads are implemented here. The … … 39 39 * when servicing udebug IPC messages. 40 40 */ 41 41 42 42 #include <debug.h> 43 43 #include <proc/task.h> … … 46 46 #include <errno.h> 47 47 #include <print.h> 48 #include <str.h>49 48 #include <syscall/copy.h> 50 49 #include <ipc/ipc.h> 51 50 #include <udebug/udebug.h> 52 51 #include <udebug/udebug_ops.h> 53 #include <memstr.h> 54 55 /** Prepare a thread for a debugging operation.52 53 /** 54 * Prepare a thread for a debugging operation. 56 55 * 57 56 * Simply put, return thread t with t->udebug.lock held, … … 72 71 * the t->lock spinlock to the t->udebug.lock mutex. 73 72 * 74 * @param t hreadPointer, need not at all be valid.75 * @param being_go 73 * @param t Pointer, need not at all be valid. 74 * @param being_go Required thread state. 76 75 * 77 76 * Returns EOK if all went well, or an error code otherwise. 78 * 79 */ 80 static int _thread_op_begin(thread_t *thread, bool being_go) 81 { 82 mutex_lock(&TASK->udebug.lock); 83 77 */ 78 static int _thread_op_begin(thread_t *t, bool being_go) 79 { 80 task_id_t taskid; 81 ipl_t ipl; 82 83 taskid = TASK->taskid; 84 85 mutex_lock(&TASK->udebug.lock); 86 84 87 /* thread_exists() must be called with threads_lock held */ 85 irq_spinlock_lock(&threads_lock, true); 86 87 if (!thread_exists(thread)) { 88 irq_spinlock_unlock(&threads_lock, true); 88 ipl = interrupts_disable(); 89 spinlock_lock(&threads_lock); 90 91 if (!thread_exists(t)) { 92 spinlock_unlock(&threads_lock); 93 interrupts_restore(ipl); 89 94 mutex_unlock(&TASK->udebug.lock); 90 95 return ENOENT; 91 96 } 92 93 /* thread->lock is enough to ensure the thread's existence */ 94 irq_spinlock_exchange(&threads_lock, &thread->lock); 95 96 /* Verify that 'thread' is a userspace thread. */ 97 if ((thread->flags & THREAD_FLAG_USPACE) == 0) { 97 98 /* t->lock is enough to ensure the thread's existence */ 99 spinlock_lock(&t->lock); 100 spinlock_unlock(&threads_lock); 101 102 /* Verify that 't' is a userspace thread. */ 103 if ((t->flags & THREAD_FLAG_USPACE) == 0) { 98 104 /* It's not, deny its existence */ 99 irq_spinlock_unlock(&thread->lock, true); 105 spinlock_unlock(&t->lock); 106 interrupts_restore(ipl); 100 107 mutex_unlock(&TASK->udebug.lock); 101 108 return ENOENT; 102 109 } 103 110 104 111 /* Verify debugging state. */ 105 if (t hread->udebug.active != true) {112 if (t->udebug.active != true) { 106 113 /* Not in debugging session or undesired GO state */ 107 irq_spinlock_unlock(&thread->lock, true); 114 spinlock_unlock(&t->lock); 115 interrupts_restore(ipl); 108 116 mutex_unlock(&TASK->udebug.lock); 109 117 return ENOENT; 110 118 } 111 119 112 120 /* 113 121 * Since the thread has active == true, TASK->udebug.lock 114 122 * is enough to ensure its existence and that active remains 115 123 * true. 116 *117 124 */ 118 irq_spinlock_unlock(&thread->lock, true); 119 125 spinlock_unlock(&t->lock); 126 interrupts_restore(ipl); 127 120 128 /* Only mutex TASK->udebug.lock left. */ 121 129 122 130 /* Now verify that the thread belongs to the current task. */ 123 if (t hread->task != TASK) {131 if (t->task != TASK) { 124 132 /* No such thread belonging this task*/ 125 133 mutex_unlock(&TASK->udebug.lock); 126 134 return ENOENT; 127 135 } 128 136 129 137 /* 130 138 * Now we need to grab the thread's debug lock for synchronization 131 139 * of the threads stoppability/stop state. 132 *133 140 */ 134 mutex_lock(&t hread->udebug.lock);135 141 mutex_lock(&t->udebug.lock); 142 136 143 /* The big task mutex is no longer needed. */ 137 144 mutex_unlock(&TASK->udebug.lock); 138 139 if (t hread->udebug.go != being_go) {145 146 if (t->udebug.go != being_go) { 140 147 /* Not in debugging session or undesired GO state. */ 141 mutex_unlock(&t hread->udebug.lock);148 mutex_unlock(&t->udebug.lock); 142 149 return EINVAL; 143 150 } 144 145 /* Only t hread->udebug.lock left. */146 147 return EOK; 151 152 /* Only t->udebug.lock left. */ 153 154 return EOK; /* All went well. */ 148 155 } 149 156 150 157 /** End debugging operation on a thread. */ 151 static void _thread_op_end(thread_t *t hread)152 { 153 mutex_unlock(&t hread->udebug.lock);158 static void _thread_op_end(thread_t *t) 159 { 160 mutex_unlock(&t->udebug.lock); 154 161 } 155 162 … … 165 172 * all the threads become stoppable (i.e. they can be considered stopped). 166 173 * 167 * @param call The BEGIN call we are servicing. 168 * 169 * @return 0 (OK, but not done yet), 1 (done) or negative error code. 170 * 174 * @param call The BEGIN call we are servicing. 175 * @return 0 (OK, but not done yet), 1 (done) or negative error code. 171 176 */ 172 177 int udebug_begin(call_t *call) 173 178 { 174 LOG("Debugging task %" PRIu64, TASK->taskid); 175 176 mutex_lock(&TASK->udebug.lock); 177 179 int reply; 180 181 thread_t *t; 182 link_t *cur; 183 184 LOG("Debugging task %llu", TASK->taskid); 185 mutex_lock(&TASK->udebug.lock); 186 178 187 if (TASK->udebug.dt_state != UDEBUG_TS_INACTIVE) { 179 188 mutex_unlock(&TASK->udebug.lock); 180 189 return EBUSY; 181 190 } 182 191 183 192 TASK->udebug.dt_state = UDEBUG_TS_BEGINNING; 184 193 TASK->udebug.begin_call = call; 185 194 TASK->udebug.debugger = call->sender; 186 187 int reply; 188 195 189 196 if (TASK->udebug.not_stoppable_count == 0) { 190 197 TASK->udebug.dt_state = UDEBUG_TS_ACTIVE; 191 198 TASK->udebug.begin_call = NULL; 192 reply = 1; /* immediate reply */ 193 } else 194 reply = 0; /* no reply */ 199 reply = 1; /* immediate reply */ 200 } else { 201 reply = 0; /* no reply */ 202 } 195 203 196 204 /* Set udebug.active on all of the task's userspace threads. */ 197 198 link_t *cur; 205 199 206 for (cur = TASK->th_head.next; cur != &TASK->th_head; cur = cur->next) { 200 thread_t *thread = list_get_instance(cur, thread_t, th_link); 201 202 mutex_lock(&thread->udebug.lock); 203 if ((thread->flags & THREAD_FLAG_USPACE) != 0) { 204 thread->udebug.active = true; 205 mutex_unlock(&thread->udebug.lock); 206 condvar_broadcast(&thread->udebug.active_cv); 207 } else 208 mutex_unlock(&thread->udebug.lock); 209 } 210 207 t = list_get_instance(cur, thread_t, th_link); 208 209 mutex_lock(&t->udebug.lock); 210 if ((t->flags & THREAD_FLAG_USPACE) != 0) 211 t->udebug.active = true; 212 mutex_unlock(&t->udebug.lock); 213 } 214 211 215 mutex_unlock(&TASK->udebug.lock); 212 216 return reply; … … 216 220 * 217 221 * Closes the debugging session for the current task. 218 *219 222 * @return Zero on success or negative error code. 220 *221 223 */ 222 224 int udebug_end(void) 223 225 { 226 int rc; 227 224 228 LOG("Task %" PRIu64, TASK->taskid); 225 226 mutex_lock(&TASK->udebug.lock); 227 intrc = udebug_task_cleanup(TASK);228 mutex_unlock(&TASK->udebug.lock); 229 229 230 mutex_lock(&TASK->udebug.lock); 231 rc = udebug_task_cleanup(TASK); 232 mutex_unlock(&TASK->udebug.lock); 233 230 234 return rc; 231 235 } … … 235 239 * Sets the event mask that determines which events are enabled. 236 240 * 237 * @param mask Or combination of events that should be enabled. 238 * 239 * @return Zero on success or negative error code. 240 * 241 * @param mask Or combination of events that should be enabled. 242 * @return Zero on success or negative error code. 241 243 */ 242 244 int udebug_set_evmask(udebug_evmask_t mask) 243 245 { 244 246 LOG("mask = 0x%x", mask); 245 246 mutex_lock(&TASK->udebug.lock); 247 247 248 mutex_lock(&TASK->udebug.lock); 249 248 250 if (TASK->udebug.dt_state != UDEBUG_TS_ACTIVE) { 249 251 mutex_unlock(&TASK->udebug.lock); 250 252 return EINVAL; 251 253 } 252 254 253 255 TASK->udebug.evmask = mask; 254 256 mutex_unlock(&TASK->udebug.lock); 255 257 256 258 return 0; 257 259 } … … 263 265 * a debugging event or STOP occurs, at which point the thread loses GO. 264 266 * 265 * @param thread The thread to operate on (unlocked and need not be valid). 266 * @param call The GO call that we are servicing. 267 * 268 */ 269 int udebug_go(thread_t *thread, call_t *call) 270 { 271 /* On success, this will lock thread->udebug.lock. */ 272 int rc = _thread_op_begin(thread, false); 273 if (rc != EOK) 267 * @param t The thread to operate on (unlocked and need not be valid). 268 * @param call The GO call that we are servicing. 269 */ 270 int udebug_go(thread_t *t, call_t *call) 271 { 272 int rc; 273 274 /* On success, this will lock t->udebug.lock. */ 275 rc = _thread_op_begin(t, false); 276 if (rc != EOK) { 274 277 return rc; 275 276 thread->udebug.go_call = call; 277 thread->udebug.go = true; 278 thread->udebug.cur_event = 0; /* none */ 279 278 } 279 280 t->udebug.go_call = call; 281 t->udebug.go = true; 282 t->udebug.cur_event = 0; /* none */ 283 280 284 /* 281 * Neither thread's lock nor threads_lock may be held during wakeup. 282 * 285 * Neither t's lock nor threads_lock may be held during wakeup. 283 286 */ 284 waitq_wakeup(&t hread->udebug.go_wq, WAKEUP_FIRST);285 286 _thread_op_end(t hread);287 287 waitq_wakeup(&t->udebug.go_wq, WAKEUP_FIRST); 288 289 _thread_op_end(t); 290 288 291 return 0; 289 292 } … … 294 297 * can be considered stopped). 295 298 * 296 * @param thread The thread to operate on (unlocked and need not be valid). 297 * @param call The GO call that we are servicing. 298 * 299 */ 300 int udebug_stop(thread_t *thread, call_t *call) 301 { 299 * @param t The thread to operate on (unlocked and need not be valid). 300 * @param call The GO call that we are servicing. 301 */ 302 int udebug_stop(thread_t *t, call_t *call) 303 { 304 int rc; 305 302 306 LOG("udebug_stop()"); 303 307 304 308 /* 305 * On success, this will lock thread->udebug.lock. Note that this 306 * makes sure the thread is not stopped. 307 * 309 * On success, this will lock t->udebug.lock. Note that this makes sure 310 * the thread is not stopped. 308 311 */ 309 int rc = _thread_op_begin(thread, true);310 if (rc != EOK) 312 rc = _thread_op_begin(t, true); 313 if (rc != EOK) { 311 314 return rc; 312 315 } 316 313 317 /* Take GO away from the thread. */ 314 t hread->udebug.go = false;315 316 if (t hread->udebug.stoppable != true) {318 t->udebug.go = false; 319 320 if (t->udebug.stoppable != true) { 317 321 /* Answer will be sent when the thread becomes stoppable. */ 318 _thread_op_end(t hread);322 _thread_op_end(t); 319 323 return 0; 320 324 } 321 325 322 326 /* 323 327 * Answer GO call. 324 *325 328 */ 326 329 327 330 /* Make sure nobody takes this call away from us. */ 328 call = t hread->udebug.go_call;329 t hread->udebug.go_call = NULL;330 331 call = t->udebug.go_call; 332 t->udebug.go_call = NULL; 333 331 334 IPC_SET_RETVAL(call->data, 0); 332 335 IPC_SET_ARG1(call->data, UDEBUG_EVENT_STOP); 333 336 334 337 THREAD->udebug.cur_event = UDEBUG_EVENT_STOP; 335 336 _thread_op_end(t hread);337 338 339 _thread_op_end(t); 340 338 341 mutex_lock(&TASK->udebug.lock); 339 342 ipc_answer(&TASK->answerbox, call); 340 343 mutex_unlock(&TASK->udebug.lock); 341 344 342 345 return 0; 343 346 } … … 351 354 * 352 355 * If the sequence is longer than @a buf_size bytes, only as much hashes 353 * as can fit are copied. The number of bytes copied is stored in @a stored. 354 * The total number of thread bytes that could have been saved had there been 355 * enough space is stored in @a needed. 356 * as can fit are copied. The number of thread hashes copied is stored 357 * in @a n. 356 358 * 357 359 * The rationale for having @a buf_size is that this function is only … … 359 361 * a maximum size for the userspace buffer. 360 362 * 361 * @param buffer The buffer for storing thread hashes. 362 * @param buf_size Buffer size in bytes. 363 * @param stored The actual number of bytes copied will be stored here. 364 * @param needed Total number of hashes that could have been saved. 365 * 366 */ 367 int udebug_thread_read(void **buffer, size_t buf_size, size_t *stored, 368 size_t *needed) 369 { 363 * @param buffer The buffer for storing thread hashes. 364 * @param buf_size Buffer size in bytes. 365 * @param n The actual number of hashes copied will be stored here. 366 */ 367 int udebug_thread_read(void **buffer, size_t buf_size, size_t *n) 368 { 369 thread_t *t; 370 link_t *cur; 371 unative_t tid; 372 unsigned copied_ids; 373 ipl_t ipl; 374 unative_t *id_buffer; 375 int flags; 376 size_t max_ids; 377 370 378 LOG("udebug_thread_read()"); 371 379 372 380 /* Allocate a buffer to hold thread IDs */ 373 sysarg_t *id_buffer = malloc(buf_size + 1, 0);374 375 mutex_lock(&TASK->udebug.lock); 376 381 id_buffer = malloc(buf_size, 0); 382 383 mutex_lock(&TASK->udebug.lock); 384 377 385 /* Verify task state */ 378 386 if (TASK->udebug.dt_state != UDEBUG_TS_ACTIVE) { … … 380 388 return EINVAL; 381 389 } 382 383 i rq_spinlock_lock(&TASK->lock, true);384 390 391 ipl = interrupts_disable(); 392 spinlock_lock(&TASK->lock); 385 393 /* Copy down the thread IDs */ 386 387 size_t max_ids = buf_size / sizeof(sysarg_t); 388 size_t copied_ids = 0; 389 size_t extra_ids = 0; 390 394 395 max_ids = buf_size / sizeof(unative_t); 396 copied_ids = 0; 397 391 398 /* FIXME: make sure the thread isn't past debug shutdown... */ 392 link_t *cur;393 399 for (cur = TASK->th_head.next; cur != &TASK->th_head; cur = cur->next) { 394 thread_t *thread = list_get_instance(cur, thread_t, th_link); 395 396 irq_spinlock_lock(&thread->lock, false); 397 int flags = thread->flags; 398 irq_spinlock_unlock(&thread->lock, false); 399 400 /* Do not write past end of buffer */ 401 if (copied_ids >= max_ids) break; 402 403 t = list_get_instance(cur, thread_t, th_link); 404 405 spinlock_lock(&t->lock); 406 flags = t->flags; 407 spinlock_unlock(&t->lock); 408 400 409 /* Not interested in kernel threads. */ 401 if ((flags & THREAD_FLAG_USPACE) == 0) 402 continue; 403 404 if (copied_ids < max_ids) { 410 if ((flags & THREAD_FLAG_USPACE) != 0) { 405 411 /* Using thread struct pointer as identification hash */ 406 id_buffer[copied_ids++] = (sysarg_t) thread; 407 } else 408 extra_ids++; 409 } 410 411 irq_spinlock_unlock(&TASK->lock, true); 412 413 mutex_unlock(&TASK->udebug.lock); 414 412 tid = (unative_t) t; 413 id_buffer[copied_ids++] = tid; 414 } 415 } 416 417 spinlock_unlock(&TASK->lock); 418 interrupts_restore(ipl); 419 420 mutex_unlock(&TASK->udebug.lock); 421 415 422 *buffer = id_buffer; 416 *stored = copied_ids * sizeof(sysarg_t); 417 *needed = (copied_ids + extra_ids) * sizeof(sysarg_t); 418 419 return 0; 420 } 421 422 /** Read task name. 423 * 424 * Returns task name as non-terminated string in a newly allocated buffer. 425 * Also returns the size of the data. 426 * 427 * @param data Place to store pointer to newly allocated block. 428 * @param data_size Place to store size of the data. 429 * 430 * @return EOK. 431 * 432 */ 433 int udebug_name_read(char **data, size_t *data_size) 434 { 435 size_t name_size = str_size(TASK->name) + 1; 436 437 *data = malloc(name_size, 0); 438 *data_size = name_size; 439 440 memcpy(*data, TASK->name, name_size); 441 423 *n = copied_ids * sizeof(unative_t); 424 442 425 return 0; 443 426 } … … 453 436 * this function will fail with an EINVAL error code. 454 437 * 455 * @param thread Thread where call arguments are to be read. 456 * @param buffer Place to store pointer to new buffer. 457 * 458 * @return EOK on success, ENOENT if @a t is invalid, EINVAL 459 * if thread state is not valid for this operation. 460 * 461 */ 462 int udebug_args_read(thread_t *thread, void **buffer) 463 { 438 * @param buffer The buffer for storing thread hashes. 439 */ 440 int udebug_args_read(thread_t *t, void **buffer) 441 { 442 int rc; 443 unative_t *arg_buffer; 444 464 445 /* Prepare a buffer to hold the arguments. */ 465 sysarg_t *arg_buffer = malloc(6 * sizeof(sysarg_t), 0);466 446 arg_buffer = malloc(6 * sizeof(unative_t), 0); 447 467 448 /* On success, this will lock t->udebug.lock. */ 468 int rc = _thread_op_begin(thread, false);469 if (rc != EOK) 449 rc = _thread_op_begin(t, false); 450 if (rc != EOK) { 470 451 return rc; 471 452 } 453 472 454 /* Additionally we need to verify that we are inside a syscall. */ 473 if ( (thread->udebug.cur_event != UDEBUG_EVENT_SYSCALL_B)&&474 (thread->udebug.cur_event != UDEBUG_EVENT_SYSCALL_E)) {475 _thread_op_end(t hread);455 if (t->udebug.cur_event != UDEBUG_EVENT_SYSCALL_B && 456 t->udebug.cur_event != UDEBUG_EVENT_SYSCALL_E) { 457 _thread_op_end(t); 476 458 return EINVAL; 477 459 } 478 460 479 461 /* Copy to a local buffer before releasing the lock. */ 480 memcpy(arg_buffer, t hread->udebug.syscall_args, 6 * sizeof(sysarg_t));481 482 _thread_op_end(t hread);483 462 memcpy(arg_buffer, t->udebug.syscall_args, 6 * sizeof(unative_t)); 463 464 _thread_op_end(t); 465 484 466 *buffer = arg_buffer; 485 return 0;486 }487 488 /** Read the register state of the thread.489 *490 * The contents of the thread's istate structure are copied to a newly491 * allocated buffer and a pointer to it is written to @a buffer. The size of492 * the buffer will be sizeof(istate_t).493 *494 * Currently register state cannot be read if the thread is inside a system495 * call (as opposed to an exception). This is an implementation limit.496 *497 * @param thread Thread whose state is to be read.498 * @param buffer Place to store pointer to new buffer.499 *500 * @return EOK on success, ENOENT if @a t is invalid, EINVAL501 * if thread is not in valid state, EBUSY if istate502 * is not available.503 *504 */505 int udebug_regs_read(thread_t *thread, void **buffer)506 {507 /* Prepare a buffer to hold the data. */508 istate_t *state_buf = malloc(sizeof(istate_t), 0);509 510 /* On success, this will lock t->udebug.lock */511 int rc = _thread_op_begin(thread, false);512 if (rc != EOK)513 return rc;514 515 istate_t *state = thread->udebug.uspace_state;516 if (state == NULL) {517 _thread_op_end(thread);518 return EBUSY;519 }520 521 /* Copy to the allocated buffer */522 memcpy(state_buf, state, sizeof(istate_t));523 524 _thread_op_end(thread);525 526 *buffer = (void *) state_buf;527 467 return 0; 528 468 } … … 534 474 * and a pointer to it is written into @a buffer. 535 475 * 536 * @param uspace_addr Address from where to start reading. 537 * @param n Number of bytes to read. 538 * @param buffer For storing a pointer to the allocated buffer. 539 * 540 */ 541 int udebug_mem_read(sysarg_t uspace_addr, size_t n, void **buffer) 542 { 476 * @param uspace_addr Address from where to start reading. 477 * @param n Number of bytes to read. 478 * @param buffer For storing a pointer to the allocated buffer. 479 */ 480 int udebug_mem_read(unative_t uspace_addr, size_t n, void **buffer) 481 { 482 void *data_buffer; 483 int rc; 484 543 485 /* Verify task state */ 544 486 mutex_lock(&TASK->udebug.lock); 545 487 546 488 if (TASK->udebug.dt_state != UDEBUG_TS_ACTIVE) { 547 489 mutex_unlock(&TASK->udebug.lock); 548 490 return EBUSY; 549 491 } 550 551 void *data_buffer = malloc(n, 0); 552 553 /* 554 * NOTE: this is not strictly from a syscall... but that shouldn't 555 * be a problem 556 * 557 */ 558 int rc = copy_from_uspace(data_buffer, (void *) uspace_addr, n); 559 mutex_unlock(&TASK->udebug.lock); 560 561 if (rc != 0) 562 return rc; 563 492 493 data_buffer = malloc(n, 0); 494 495 /* NOTE: this is not strictly from a syscall... but that shouldn't 496 * be a problem */ 497 rc = copy_from_uspace(data_buffer, (void *)uspace_addr, n); 498 mutex_unlock(&TASK->udebug.lock); 499 500 if (rc != 0) return rc; 501 564 502 *buffer = data_buffer; 565 503 return 0;
Note:
See TracChangeset
for help on using the changeset viewer.