Kyle Swenson | 8d8f654 | 2021-03-15 11:02:55 -0600 | [diff] [blame^] | 1 | /* |
| 2 | * This program is free software; you can redistribute it and/or modify it |
| 3 | * under the terms of the GNU General Public License as published by the |
| 4 | * Free Software Foundation; either version 2, or (at your option) any |
| 5 | * later version. |
| 6 | * |
| 7 | * This program is distributed in the hope that it will be useful, but |
| 8 | * WITHOUT ANY WARRANTY; without even the implied warranty of |
| 9 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| 10 | * General Public License for more details. |
| 11 | * |
| 12 | */ |
| 13 | |
| 14 | /* |
| 15 | * Copyright (C) 2004 Amit S. Kale <amitkale@linsyssoft.com> |
| 16 | * Copyright (C) 2000-2001 VERITAS Software Corporation. |
| 17 | * Copyright (C) 2002 Andi Kleen, SuSE Labs |
| 18 | * Copyright (C) 2004 LinSysSoft Technologies Pvt. Ltd. |
| 19 | * Copyright (C) 2007 MontaVista Software, Inc. |
| 20 | * Copyright (C) 2007-2008 Jason Wessel, Wind River Systems, Inc. |
| 21 | */ |
| 22 | /**************************************************************************** |
| 23 | * Contributor: Lake Stevens Instrument Division$ |
| 24 | * Written by: Glenn Engel $ |
| 25 | * Updated by: Amit Kale<akale@veritas.com> |
| 26 | * Updated by: Tom Rini <trini@kernel.crashing.org> |
| 27 | * Updated by: Jason Wessel <jason.wessel@windriver.com> |
| 28 | * Modified for 386 by Jim Kingdon, Cygnus Support. |
| 29 | * Origianl kgdb, compatibility with 2.1.xx kernel by |
| 30 | * David Grothe <dave@gcom.com> |
| 31 | * Integrated into 2.2.5 kernel by Tigran Aivazian <tigran@sco.com> |
| 32 | * X86_64 changes from Andi Kleen's patch merged by Jim Houston |
| 33 | */ |
| 34 | #include <linux/spinlock.h> |
| 35 | #include <linux/kdebug.h> |
| 36 | #include <linux/string.h> |
| 37 | #include <linux/kernel.h> |
| 38 | #include <linux/ptrace.h> |
| 39 | #include <linux/sched.h> |
| 40 | #include <linux/delay.h> |
| 41 | #include <linux/kgdb.h> |
| 42 | #include <linux/smp.h> |
| 43 | #include <linux/nmi.h> |
| 44 | #include <linux/hw_breakpoint.h> |
| 45 | #include <linux/uaccess.h> |
| 46 | #include <linux/memory.h> |
| 47 | |
| 48 | #include <asm/debugreg.h> |
| 49 | #include <asm/apicdef.h> |
| 50 | #include <asm/apic.h> |
| 51 | #include <asm/nmi.h> |
| 52 | |
| 53 | struct dbg_reg_def_t dbg_reg_def[DBG_MAX_REG_NUM] = |
| 54 | { |
| 55 | #ifdef CONFIG_X86_32 |
| 56 | { "ax", 4, offsetof(struct pt_regs, ax) }, |
| 57 | { "cx", 4, offsetof(struct pt_regs, cx) }, |
| 58 | { "dx", 4, offsetof(struct pt_regs, dx) }, |
| 59 | { "bx", 4, offsetof(struct pt_regs, bx) }, |
| 60 | { "sp", 4, offsetof(struct pt_regs, sp) }, |
| 61 | { "bp", 4, offsetof(struct pt_regs, bp) }, |
| 62 | { "si", 4, offsetof(struct pt_regs, si) }, |
| 63 | { "di", 4, offsetof(struct pt_regs, di) }, |
| 64 | { "ip", 4, offsetof(struct pt_regs, ip) }, |
| 65 | { "flags", 4, offsetof(struct pt_regs, flags) }, |
| 66 | { "cs", 4, offsetof(struct pt_regs, cs) }, |
| 67 | { "ss", 4, offsetof(struct pt_regs, ss) }, |
| 68 | { "ds", 4, offsetof(struct pt_regs, ds) }, |
| 69 | { "es", 4, offsetof(struct pt_regs, es) }, |
| 70 | #else |
| 71 | { "ax", 8, offsetof(struct pt_regs, ax) }, |
| 72 | { "bx", 8, offsetof(struct pt_regs, bx) }, |
| 73 | { "cx", 8, offsetof(struct pt_regs, cx) }, |
| 74 | { "dx", 8, offsetof(struct pt_regs, dx) }, |
| 75 | { "si", 8, offsetof(struct pt_regs, si) }, |
| 76 | { "di", 8, offsetof(struct pt_regs, di) }, |
| 77 | { "bp", 8, offsetof(struct pt_regs, bp) }, |
| 78 | { "sp", 8, offsetof(struct pt_regs, sp) }, |
| 79 | { "r8", 8, offsetof(struct pt_regs, r8) }, |
| 80 | { "r9", 8, offsetof(struct pt_regs, r9) }, |
| 81 | { "r10", 8, offsetof(struct pt_regs, r10) }, |
| 82 | { "r11", 8, offsetof(struct pt_regs, r11) }, |
| 83 | { "r12", 8, offsetof(struct pt_regs, r12) }, |
| 84 | { "r13", 8, offsetof(struct pt_regs, r13) }, |
| 85 | { "r14", 8, offsetof(struct pt_regs, r14) }, |
| 86 | { "r15", 8, offsetof(struct pt_regs, r15) }, |
| 87 | { "ip", 8, offsetof(struct pt_regs, ip) }, |
| 88 | { "flags", 4, offsetof(struct pt_regs, flags) }, |
| 89 | { "cs", 4, offsetof(struct pt_regs, cs) }, |
| 90 | { "ss", 4, offsetof(struct pt_regs, ss) }, |
| 91 | { "ds", 4, -1 }, |
| 92 | { "es", 4, -1 }, |
| 93 | #endif |
| 94 | { "fs", 4, -1 }, |
| 95 | { "gs", 4, -1 }, |
| 96 | }; |
| 97 | |
| 98 | int dbg_set_reg(int regno, void *mem, struct pt_regs *regs) |
| 99 | { |
| 100 | if ( |
| 101 | #ifdef CONFIG_X86_32 |
| 102 | regno == GDB_SS || regno == GDB_FS || regno == GDB_GS || |
| 103 | #endif |
| 104 | regno == GDB_SP || regno == GDB_ORIG_AX) |
| 105 | return 0; |
| 106 | |
| 107 | if (dbg_reg_def[regno].offset != -1) |
| 108 | memcpy((void *)regs + dbg_reg_def[regno].offset, mem, |
| 109 | dbg_reg_def[regno].size); |
| 110 | return 0; |
| 111 | } |
| 112 | |
| 113 | char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs) |
| 114 | { |
| 115 | if (regno == GDB_ORIG_AX) { |
| 116 | memcpy(mem, ®s->orig_ax, sizeof(regs->orig_ax)); |
| 117 | return "orig_ax"; |
| 118 | } |
| 119 | if (regno >= DBG_MAX_REG_NUM || regno < 0) |
| 120 | return NULL; |
| 121 | |
| 122 | if (dbg_reg_def[regno].offset != -1) |
| 123 | memcpy(mem, (void *)regs + dbg_reg_def[regno].offset, |
| 124 | dbg_reg_def[regno].size); |
| 125 | |
| 126 | #ifdef CONFIG_X86_32 |
| 127 | switch (regno) { |
| 128 | case GDB_SS: |
| 129 | if (!user_mode(regs)) |
| 130 | *(unsigned long *)mem = __KERNEL_DS; |
| 131 | break; |
| 132 | case GDB_SP: |
| 133 | if (!user_mode(regs)) |
| 134 | *(unsigned long *)mem = kernel_stack_pointer(regs); |
| 135 | break; |
| 136 | case GDB_GS: |
| 137 | case GDB_FS: |
| 138 | *(unsigned long *)mem = 0xFFFF; |
| 139 | break; |
| 140 | } |
| 141 | #endif |
| 142 | return dbg_reg_def[regno].name; |
| 143 | } |
| 144 | |
| 145 | /** |
| 146 | * sleeping_thread_to_gdb_regs - Convert ptrace regs to GDB regs |
| 147 | * @gdb_regs: A pointer to hold the registers in the order GDB wants. |
| 148 | * @p: The &struct task_struct of the desired process. |
| 149 | * |
| 150 | * Convert the register values of the sleeping process in @p to |
| 151 | * the format that GDB expects. |
| 152 | * This function is called when kgdb does not have access to the |
| 153 | * &struct pt_regs and therefore it should fill the gdb registers |
| 154 | * @gdb_regs with what has been saved in &struct thread_struct |
| 155 | * thread field during switch_to. |
| 156 | */ |
| 157 | void sleeping_thread_to_gdb_regs(unsigned long *gdb_regs, struct task_struct *p) |
| 158 | { |
| 159 | #ifndef CONFIG_X86_32 |
| 160 | u32 *gdb_regs32 = (u32 *)gdb_regs; |
| 161 | #endif |
| 162 | gdb_regs[GDB_AX] = 0; |
| 163 | gdb_regs[GDB_BX] = 0; |
| 164 | gdb_regs[GDB_CX] = 0; |
| 165 | gdb_regs[GDB_DX] = 0; |
| 166 | gdb_regs[GDB_SI] = 0; |
| 167 | gdb_regs[GDB_DI] = 0; |
| 168 | gdb_regs[GDB_BP] = *(unsigned long *)p->thread.sp; |
| 169 | #ifdef CONFIG_X86_32 |
| 170 | gdb_regs[GDB_DS] = __KERNEL_DS; |
| 171 | gdb_regs[GDB_ES] = __KERNEL_DS; |
| 172 | gdb_regs[GDB_PS] = 0; |
| 173 | gdb_regs[GDB_CS] = __KERNEL_CS; |
| 174 | gdb_regs[GDB_PC] = p->thread.ip; |
| 175 | gdb_regs[GDB_SS] = __KERNEL_DS; |
| 176 | gdb_regs[GDB_FS] = 0xFFFF; |
| 177 | gdb_regs[GDB_GS] = 0xFFFF; |
| 178 | #else |
| 179 | gdb_regs32[GDB_PS] = *(unsigned long *)(p->thread.sp + 8); |
| 180 | gdb_regs32[GDB_CS] = __KERNEL_CS; |
| 181 | gdb_regs32[GDB_SS] = __KERNEL_DS; |
| 182 | gdb_regs[GDB_PC] = 0; |
| 183 | gdb_regs[GDB_R8] = 0; |
| 184 | gdb_regs[GDB_R9] = 0; |
| 185 | gdb_regs[GDB_R10] = 0; |
| 186 | gdb_regs[GDB_R11] = 0; |
| 187 | gdb_regs[GDB_R12] = 0; |
| 188 | gdb_regs[GDB_R13] = 0; |
| 189 | gdb_regs[GDB_R14] = 0; |
| 190 | gdb_regs[GDB_R15] = 0; |
| 191 | #endif |
| 192 | gdb_regs[GDB_SP] = p->thread.sp; |
| 193 | } |
| 194 | |
| 195 | static struct hw_breakpoint { |
| 196 | unsigned enabled; |
| 197 | unsigned long addr; |
| 198 | int len; |
| 199 | int type; |
| 200 | struct perf_event * __percpu *pev; |
| 201 | } breakinfo[HBP_NUM]; |
| 202 | |
| 203 | static unsigned long early_dr7; |
| 204 | |
| 205 | static void kgdb_correct_hw_break(void) |
| 206 | { |
| 207 | int breakno; |
| 208 | |
| 209 | for (breakno = 0; breakno < HBP_NUM; breakno++) { |
| 210 | struct perf_event *bp; |
| 211 | struct arch_hw_breakpoint *info; |
| 212 | int val; |
| 213 | int cpu = raw_smp_processor_id(); |
| 214 | if (!breakinfo[breakno].enabled) |
| 215 | continue; |
| 216 | if (dbg_is_early) { |
| 217 | set_debugreg(breakinfo[breakno].addr, breakno); |
| 218 | early_dr7 |= encode_dr7(breakno, |
| 219 | breakinfo[breakno].len, |
| 220 | breakinfo[breakno].type); |
| 221 | set_debugreg(early_dr7, 7); |
| 222 | continue; |
| 223 | } |
| 224 | bp = *per_cpu_ptr(breakinfo[breakno].pev, cpu); |
| 225 | info = counter_arch_bp(bp); |
| 226 | if (bp->attr.disabled != 1) |
| 227 | continue; |
| 228 | bp->attr.bp_addr = breakinfo[breakno].addr; |
| 229 | bp->attr.bp_len = breakinfo[breakno].len; |
| 230 | bp->attr.bp_type = breakinfo[breakno].type; |
| 231 | info->address = breakinfo[breakno].addr; |
| 232 | info->len = breakinfo[breakno].len; |
| 233 | info->type = breakinfo[breakno].type; |
| 234 | val = arch_install_hw_breakpoint(bp); |
| 235 | if (!val) |
| 236 | bp->attr.disabled = 0; |
| 237 | } |
| 238 | if (!dbg_is_early) |
| 239 | hw_breakpoint_restore(); |
| 240 | } |
| 241 | |
| 242 | static int hw_break_reserve_slot(int breakno) |
| 243 | { |
| 244 | int cpu; |
| 245 | int cnt = 0; |
| 246 | struct perf_event **pevent; |
| 247 | |
| 248 | if (dbg_is_early) |
| 249 | return 0; |
| 250 | |
| 251 | for_each_online_cpu(cpu) { |
| 252 | cnt++; |
| 253 | pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); |
| 254 | if (dbg_reserve_bp_slot(*pevent)) |
| 255 | goto fail; |
| 256 | } |
| 257 | |
| 258 | return 0; |
| 259 | |
| 260 | fail: |
| 261 | for_each_online_cpu(cpu) { |
| 262 | cnt--; |
| 263 | if (!cnt) |
| 264 | break; |
| 265 | pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); |
| 266 | dbg_release_bp_slot(*pevent); |
| 267 | } |
| 268 | return -1; |
| 269 | } |
| 270 | |
| 271 | static int hw_break_release_slot(int breakno) |
| 272 | { |
| 273 | struct perf_event **pevent; |
| 274 | int cpu; |
| 275 | |
| 276 | if (dbg_is_early) |
| 277 | return 0; |
| 278 | |
| 279 | for_each_online_cpu(cpu) { |
| 280 | pevent = per_cpu_ptr(breakinfo[breakno].pev, cpu); |
| 281 | if (dbg_release_bp_slot(*pevent)) |
| 282 | /* |
| 283 | * The debugger is responsible for handing the retry on |
| 284 | * remove failure. |
| 285 | */ |
| 286 | return -1; |
| 287 | } |
| 288 | return 0; |
| 289 | } |
| 290 | |
| 291 | static int |
| 292 | kgdb_remove_hw_break(unsigned long addr, int len, enum kgdb_bptype bptype) |
| 293 | { |
| 294 | int i; |
| 295 | |
| 296 | for (i = 0; i < HBP_NUM; i++) |
| 297 | if (breakinfo[i].addr == addr && breakinfo[i].enabled) |
| 298 | break; |
| 299 | if (i == HBP_NUM) |
| 300 | return -1; |
| 301 | |
| 302 | if (hw_break_release_slot(i)) { |
| 303 | printk(KERN_ERR "Cannot remove hw breakpoint at %lx\n", addr); |
| 304 | return -1; |
| 305 | } |
| 306 | breakinfo[i].enabled = 0; |
| 307 | |
| 308 | return 0; |
| 309 | } |
| 310 | |
| 311 | static void kgdb_remove_all_hw_break(void) |
| 312 | { |
| 313 | int i; |
| 314 | int cpu = raw_smp_processor_id(); |
| 315 | struct perf_event *bp; |
| 316 | |
| 317 | for (i = 0; i < HBP_NUM; i++) { |
| 318 | if (!breakinfo[i].enabled) |
| 319 | continue; |
| 320 | bp = *per_cpu_ptr(breakinfo[i].pev, cpu); |
| 321 | if (!bp->attr.disabled) { |
| 322 | arch_uninstall_hw_breakpoint(bp); |
| 323 | bp->attr.disabled = 1; |
| 324 | continue; |
| 325 | } |
| 326 | if (dbg_is_early) |
| 327 | early_dr7 &= ~encode_dr7(i, breakinfo[i].len, |
| 328 | breakinfo[i].type); |
| 329 | else if (hw_break_release_slot(i)) |
| 330 | printk(KERN_ERR "KGDB: hw bpt remove failed %lx\n", |
| 331 | breakinfo[i].addr); |
| 332 | breakinfo[i].enabled = 0; |
| 333 | } |
| 334 | } |
| 335 | |
| 336 | static int |
| 337 | kgdb_set_hw_break(unsigned long addr, int len, enum kgdb_bptype bptype) |
| 338 | { |
| 339 | int i; |
| 340 | |
| 341 | for (i = 0; i < HBP_NUM; i++) |
| 342 | if (!breakinfo[i].enabled) |
| 343 | break; |
| 344 | if (i == HBP_NUM) |
| 345 | return -1; |
| 346 | |
| 347 | switch (bptype) { |
| 348 | case BP_HARDWARE_BREAKPOINT: |
| 349 | len = 1; |
| 350 | breakinfo[i].type = X86_BREAKPOINT_EXECUTE; |
| 351 | break; |
| 352 | case BP_WRITE_WATCHPOINT: |
| 353 | breakinfo[i].type = X86_BREAKPOINT_WRITE; |
| 354 | break; |
| 355 | case BP_ACCESS_WATCHPOINT: |
| 356 | breakinfo[i].type = X86_BREAKPOINT_RW; |
| 357 | break; |
| 358 | default: |
| 359 | return -1; |
| 360 | } |
| 361 | switch (len) { |
| 362 | case 1: |
| 363 | breakinfo[i].len = X86_BREAKPOINT_LEN_1; |
| 364 | break; |
| 365 | case 2: |
| 366 | breakinfo[i].len = X86_BREAKPOINT_LEN_2; |
| 367 | break; |
| 368 | case 4: |
| 369 | breakinfo[i].len = X86_BREAKPOINT_LEN_4; |
| 370 | break; |
| 371 | #ifdef CONFIG_X86_64 |
| 372 | case 8: |
| 373 | breakinfo[i].len = X86_BREAKPOINT_LEN_8; |
| 374 | break; |
| 375 | #endif |
| 376 | default: |
| 377 | return -1; |
| 378 | } |
| 379 | breakinfo[i].addr = addr; |
| 380 | if (hw_break_reserve_slot(i)) { |
| 381 | breakinfo[i].addr = 0; |
| 382 | return -1; |
| 383 | } |
| 384 | breakinfo[i].enabled = 1; |
| 385 | |
| 386 | return 0; |
| 387 | } |
| 388 | |
| 389 | /** |
| 390 | * kgdb_disable_hw_debug - Disable hardware debugging while we in kgdb. |
| 391 | * @regs: Current &struct pt_regs. |
| 392 | * |
| 393 | * This function will be called if the particular architecture must |
| 394 | * disable hardware debugging while it is processing gdb packets or |
| 395 | * handling exception. |
| 396 | */ |
| 397 | static void kgdb_disable_hw_debug(struct pt_regs *regs) |
| 398 | { |
| 399 | int i; |
| 400 | int cpu = raw_smp_processor_id(); |
| 401 | struct perf_event *bp; |
| 402 | |
| 403 | /* Disable hardware debugging while we are in kgdb: */ |
| 404 | set_debugreg(0UL, 7); |
| 405 | for (i = 0; i < HBP_NUM; i++) { |
| 406 | if (!breakinfo[i].enabled) |
| 407 | continue; |
| 408 | if (dbg_is_early) { |
| 409 | early_dr7 &= ~encode_dr7(i, breakinfo[i].len, |
| 410 | breakinfo[i].type); |
| 411 | continue; |
| 412 | } |
| 413 | bp = *per_cpu_ptr(breakinfo[i].pev, cpu); |
| 414 | if (bp->attr.disabled == 1) |
| 415 | continue; |
| 416 | arch_uninstall_hw_breakpoint(bp); |
| 417 | bp->attr.disabled = 1; |
| 418 | } |
| 419 | } |
| 420 | |
| 421 | #ifdef CONFIG_SMP |
| 422 | /** |
| 423 | * kgdb_roundup_cpus - Get other CPUs into a holding pattern |
| 424 | * @flags: Current IRQ state |
| 425 | * |
| 426 | * On SMP systems, we need to get the attention of the other CPUs |
| 427 | * and get them be in a known state. This should do what is needed |
| 428 | * to get the other CPUs to call kgdb_wait(). Note that on some arches, |
| 429 | * the NMI approach is not used for rounding up all the CPUs. For example, |
| 430 | * in case of MIPS, smp_call_function() is used to roundup CPUs. In |
| 431 | * this case, we have to make sure that interrupts are enabled before |
| 432 | * calling smp_call_function(). The argument to this function is |
| 433 | * the flags that will be used when restoring the interrupts. There is |
| 434 | * local_irq_save() call before kgdb_roundup_cpus(). |
| 435 | * |
| 436 | * On non-SMP systems, this is not called. |
| 437 | */ |
| 438 | void kgdb_roundup_cpus(unsigned long flags) |
| 439 | { |
| 440 | apic->send_IPI_allbutself(APIC_DM_NMI); |
| 441 | } |
| 442 | #endif |
| 443 | |
| 444 | /** |
| 445 | * kgdb_arch_handle_exception - Handle architecture specific GDB packets. |
| 446 | * @e_vector: The error vector of the exception that happened. |
| 447 | * @signo: The signal number of the exception that happened. |
| 448 | * @err_code: The error code of the exception that happened. |
| 449 | * @remcomInBuffer: The buffer of the packet we have read. |
| 450 | * @remcomOutBuffer: The buffer of %BUFMAX bytes to write a packet into. |
| 451 | * @linux_regs: The &struct pt_regs of the current process. |
| 452 | * |
| 453 | * This function MUST handle the 'c' and 's' command packets, |
| 454 | * as well packets to set / remove a hardware breakpoint, if used. |
| 455 | * If there are additional packets which the hardware needs to handle, |
| 456 | * they are handled here. The code should return -1 if it wants to |
| 457 | * process more packets, and a %0 or %1 if it wants to exit from the |
| 458 | * kgdb callback. |
| 459 | */ |
| 460 | int kgdb_arch_handle_exception(int e_vector, int signo, int err_code, |
| 461 | char *remcomInBuffer, char *remcomOutBuffer, |
| 462 | struct pt_regs *linux_regs) |
| 463 | { |
| 464 | unsigned long addr; |
| 465 | char *ptr; |
| 466 | |
| 467 | switch (remcomInBuffer[0]) { |
| 468 | case 'c': |
| 469 | case 's': |
| 470 | /* try to read optional parameter, pc unchanged if no parm */ |
| 471 | ptr = &remcomInBuffer[1]; |
| 472 | if (kgdb_hex2long(&ptr, &addr)) |
| 473 | linux_regs->ip = addr; |
| 474 | case 'D': |
| 475 | case 'k': |
| 476 | /* clear the trace bit */ |
| 477 | linux_regs->flags &= ~X86_EFLAGS_TF; |
| 478 | atomic_set(&kgdb_cpu_doing_single_step, -1); |
| 479 | |
| 480 | /* set the trace bit if we're stepping */ |
| 481 | if (remcomInBuffer[0] == 's') { |
| 482 | linux_regs->flags |= X86_EFLAGS_TF; |
| 483 | atomic_set(&kgdb_cpu_doing_single_step, |
| 484 | raw_smp_processor_id()); |
| 485 | } |
| 486 | |
| 487 | return 0; |
| 488 | } |
| 489 | |
| 490 | /* this means that we do not want to exit from the handler: */ |
| 491 | return -1; |
| 492 | } |
| 493 | |
| 494 | static inline int |
| 495 | single_step_cont(struct pt_regs *regs, struct die_args *args) |
| 496 | { |
| 497 | /* |
| 498 | * Single step exception from kernel space to user space so |
| 499 | * eat the exception and continue the process: |
| 500 | */ |
| 501 | printk(KERN_ERR "KGDB: trap/step from kernel to user space, " |
| 502 | "resuming...\n"); |
| 503 | kgdb_arch_handle_exception(args->trapnr, args->signr, |
| 504 | args->err, "c", "", regs); |
| 505 | /* |
| 506 | * Reset the BS bit in dr6 (pointed by args->err) to |
| 507 | * denote completion of processing |
| 508 | */ |
| 509 | (*(unsigned long *)ERR_PTR(args->err)) &= ~DR_STEP; |
| 510 | |
| 511 | return NOTIFY_STOP; |
| 512 | } |
| 513 | |
| 514 | static DECLARE_BITMAP(was_in_debug_nmi, NR_CPUS); |
| 515 | |
| 516 | static int kgdb_nmi_handler(unsigned int cmd, struct pt_regs *regs) |
| 517 | { |
| 518 | int cpu; |
| 519 | |
| 520 | switch (cmd) { |
| 521 | case NMI_LOCAL: |
| 522 | if (atomic_read(&kgdb_active) != -1) { |
| 523 | /* KGDB CPU roundup */ |
| 524 | cpu = raw_smp_processor_id(); |
| 525 | kgdb_nmicallback(cpu, regs); |
| 526 | set_bit(cpu, was_in_debug_nmi); |
| 527 | touch_nmi_watchdog(); |
| 528 | |
| 529 | return NMI_HANDLED; |
| 530 | } |
| 531 | break; |
| 532 | |
| 533 | case NMI_UNKNOWN: |
| 534 | cpu = raw_smp_processor_id(); |
| 535 | |
| 536 | if (__test_and_clear_bit(cpu, was_in_debug_nmi)) |
| 537 | return NMI_HANDLED; |
| 538 | |
| 539 | break; |
| 540 | default: |
| 541 | /* do nothing */ |
| 542 | break; |
| 543 | } |
| 544 | return NMI_DONE; |
| 545 | } |
| 546 | |
| 547 | static int __kgdb_notify(struct die_args *args, unsigned long cmd) |
| 548 | { |
| 549 | struct pt_regs *regs = args->regs; |
| 550 | |
| 551 | switch (cmd) { |
| 552 | case DIE_DEBUG: |
| 553 | if (atomic_read(&kgdb_cpu_doing_single_step) != -1) { |
| 554 | if (user_mode(regs)) |
| 555 | return single_step_cont(regs, args); |
| 556 | break; |
| 557 | } else if (test_thread_flag(TIF_SINGLESTEP)) |
| 558 | /* This means a user thread is single stepping |
| 559 | * a system call which should be ignored |
| 560 | */ |
| 561 | return NOTIFY_DONE; |
| 562 | /* fall through */ |
| 563 | default: |
| 564 | if (user_mode(regs)) |
| 565 | return NOTIFY_DONE; |
| 566 | } |
| 567 | |
| 568 | if (kgdb_handle_exception(args->trapnr, args->signr, cmd, regs)) |
| 569 | return NOTIFY_DONE; |
| 570 | |
| 571 | /* Must touch watchdog before return to normal operation */ |
| 572 | touch_nmi_watchdog(); |
| 573 | return NOTIFY_STOP; |
| 574 | } |
| 575 | |
| 576 | int kgdb_ll_trap(int cmd, const char *str, |
| 577 | struct pt_regs *regs, long err, int trap, int sig) |
| 578 | { |
| 579 | struct die_args args = { |
| 580 | .regs = regs, |
| 581 | .str = str, |
| 582 | .err = err, |
| 583 | .trapnr = trap, |
| 584 | .signr = sig, |
| 585 | |
| 586 | }; |
| 587 | |
| 588 | if (!kgdb_io_module_registered) |
| 589 | return NOTIFY_DONE; |
| 590 | |
| 591 | return __kgdb_notify(&args, cmd); |
| 592 | } |
| 593 | |
| 594 | static int |
| 595 | kgdb_notify(struct notifier_block *self, unsigned long cmd, void *ptr) |
| 596 | { |
| 597 | unsigned long flags; |
| 598 | int ret; |
| 599 | |
| 600 | local_irq_save(flags); |
| 601 | ret = __kgdb_notify(ptr, cmd); |
| 602 | local_irq_restore(flags); |
| 603 | |
| 604 | return ret; |
| 605 | } |
| 606 | |
| 607 | static struct notifier_block kgdb_notifier = { |
| 608 | .notifier_call = kgdb_notify, |
| 609 | }; |
| 610 | |
| 611 | /** |
| 612 | * kgdb_arch_init - Perform any architecture specific initalization. |
| 613 | * |
| 614 | * This function will handle the initalization of any architecture |
| 615 | * specific callbacks. |
| 616 | */ |
| 617 | int kgdb_arch_init(void) |
| 618 | { |
| 619 | int retval; |
| 620 | |
| 621 | retval = register_die_notifier(&kgdb_notifier); |
| 622 | if (retval) |
| 623 | goto out; |
| 624 | |
| 625 | retval = register_nmi_handler(NMI_LOCAL, kgdb_nmi_handler, |
| 626 | 0, "kgdb"); |
| 627 | if (retval) |
| 628 | goto out1; |
| 629 | |
| 630 | retval = register_nmi_handler(NMI_UNKNOWN, kgdb_nmi_handler, |
| 631 | 0, "kgdb"); |
| 632 | |
| 633 | if (retval) |
| 634 | goto out2; |
| 635 | |
| 636 | return retval; |
| 637 | |
| 638 | out2: |
| 639 | unregister_nmi_handler(NMI_LOCAL, "kgdb"); |
| 640 | out1: |
| 641 | unregister_die_notifier(&kgdb_notifier); |
| 642 | out: |
| 643 | return retval; |
| 644 | } |
| 645 | |
| 646 | static void kgdb_hw_overflow_handler(struct perf_event *event, |
| 647 | struct perf_sample_data *data, struct pt_regs *regs) |
| 648 | { |
| 649 | struct task_struct *tsk = current; |
| 650 | int i; |
| 651 | |
| 652 | for (i = 0; i < 4; i++) |
| 653 | if (breakinfo[i].enabled) |
| 654 | tsk->thread.debugreg6 |= (DR_TRAP0 << i); |
| 655 | } |
| 656 | |
| 657 | void kgdb_arch_late(void) |
| 658 | { |
| 659 | int i, cpu; |
| 660 | struct perf_event_attr attr; |
| 661 | struct perf_event **pevent; |
| 662 | |
| 663 | /* |
| 664 | * Pre-allocate the hw breakpoint structions in the non-atomic |
| 665 | * portion of kgdb because this operation requires mutexs to |
| 666 | * complete. |
| 667 | */ |
| 668 | hw_breakpoint_init(&attr); |
| 669 | attr.bp_addr = (unsigned long)kgdb_arch_init; |
| 670 | attr.bp_len = HW_BREAKPOINT_LEN_1; |
| 671 | attr.bp_type = HW_BREAKPOINT_W; |
| 672 | attr.disabled = 1; |
| 673 | for (i = 0; i < HBP_NUM; i++) { |
| 674 | if (breakinfo[i].pev) |
| 675 | continue; |
| 676 | breakinfo[i].pev = register_wide_hw_breakpoint(&attr, NULL, NULL); |
| 677 | if (IS_ERR((void * __force)breakinfo[i].pev)) { |
| 678 | printk(KERN_ERR "kgdb: Could not allocate hw" |
| 679 | "breakpoints\nDisabling the kernel debugger\n"); |
| 680 | breakinfo[i].pev = NULL; |
| 681 | kgdb_arch_exit(); |
| 682 | return; |
| 683 | } |
| 684 | for_each_online_cpu(cpu) { |
| 685 | pevent = per_cpu_ptr(breakinfo[i].pev, cpu); |
| 686 | pevent[0]->hw.sample_period = 1; |
| 687 | pevent[0]->overflow_handler = kgdb_hw_overflow_handler; |
| 688 | if (pevent[0]->destroy != NULL) { |
| 689 | pevent[0]->destroy = NULL; |
| 690 | release_bp_slot(*pevent); |
| 691 | } |
| 692 | } |
| 693 | } |
| 694 | } |
| 695 | |
| 696 | /** |
| 697 | * kgdb_arch_exit - Perform any architecture specific uninitalization. |
| 698 | * |
| 699 | * This function will handle the uninitalization of any architecture |
| 700 | * specific callbacks, for dynamic registration and unregistration. |
| 701 | */ |
| 702 | void kgdb_arch_exit(void) |
| 703 | { |
| 704 | int i; |
| 705 | for (i = 0; i < 4; i++) { |
| 706 | if (breakinfo[i].pev) { |
| 707 | unregister_wide_hw_breakpoint(breakinfo[i].pev); |
| 708 | breakinfo[i].pev = NULL; |
| 709 | } |
| 710 | } |
| 711 | unregister_nmi_handler(NMI_UNKNOWN, "kgdb"); |
| 712 | unregister_nmi_handler(NMI_LOCAL, "kgdb"); |
| 713 | unregister_die_notifier(&kgdb_notifier); |
| 714 | } |
| 715 | |
| 716 | /** |
| 717 | * |
| 718 | * kgdb_skipexception - Bail out of KGDB when we've been triggered. |
| 719 | * @exception: Exception vector number |
| 720 | * @regs: Current &struct pt_regs. |
| 721 | * |
| 722 | * On some architectures we need to skip a breakpoint exception when |
| 723 | * it occurs after a breakpoint has been removed. |
| 724 | * |
| 725 | * Skip an int3 exception when it occurs after a breakpoint has been |
| 726 | * removed. Backtrack eip by 1 since the int3 would have caused it to |
| 727 | * increment by 1. |
| 728 | */ |
| 729 | int kgdb_skipexception(int exception, struct pt_regs *regs) |
| 730 | { |
| 731 | if (exception == 3 && kgdb_isremovedbreak(regs->ip - 1)) { |
| 732 | regs->ip -= 1; |
| 733 | return 1; |
| 734 | } |
| 735 | return 0; |
| 736 | } |
| 737 | |
| 738 | unsigned long kgdb_arch_pc(int exception, struct pt_regs *regs) |
| 739 | { |
| 740 | if (exception == 3) |
| 741 | return instruction_pointer(regs) - 1; |
| 742 | return instruction_pointer(regs); |
| 743 | } |
| 744 | |
| 745 | void kgdb_arch_set_pc(struct pt_regs *regs, unsigned long ip) |
| 746 | { |
| 747 | regs->ip = ip; |
| 748 | } |
| 749 | |
| 750 | int kgdb_arch_set_breakpoint(struct kgdb_bkpt *bpt) |
| 751 | { |
| 752 | int err; |
| 753 | #ifdef CONFIG_DEBUG_RODATA |
| 754 | char opc[BREAK_INSTR_SIZE]; |
| 755 | #endif /* CONFIG_DEBUG_RODATA */ |
| 756 | |
| 757 | bpt->type = BP_BREAKPOINT; |
| 758 | err = probe_kernel_read(bpt->saved_instr, (char *)bpt->bpt_addr, |
| 759 | BREAK_INSTR_SIZE); |
| 760 | if (err) |
| 761 | return err; |
| 762 | err = probe_kernel_write((char *)bpt->bpt_addr, |
| 763 | arch_kgdb_ops.gdb_bpt_instr, BREAK_INSTR_SIZE); |
| 764 | #ifdef CONFIG_DEBUG_RODATA |
| 765 | if (!err) |
| 766 | return err; |
| 767 | /* |
| 768 | * It is safe to call text_poke() because normal kernel execution |
| 769 | * is stopped on all cores, so long as the text_mutex is not locked. |
| 770 | */ |
| 771 | if (mutex_is_locked(&text_mutex)) |
| 772 | return -EBUSY; |
| 773 | text_poke((void *)bpt->bpt_addr, arch_kgdb_ops.gdb_bpt_instr, |
| 774 | BREAK_INSTR_SIZE); |
| 775 | err = probe_kernel_read(opc, (char *)bpt->bpt_addr, BREAK_INSTR_SIZE); |
| 776 | if (err) |
| 777 | return err; |
| 778 | if (memcmp(opc, arch_kgdb_ops.gdb_bpt_instr, BREAK_INSTR_SIZE)) |
| 779 | return -EINVAL; |
| 780 | bpt->type = BP_POKE_BREAKPOINT; |
| 781 | #endif /* CONFIG_DEBUG_RODATA */ |
| 782 | return err; |
| 783 | } |
| 784 | |
| 785 | int kgdb_arch_remove_breakpoint(struct kgdb_bkpt *bpt) |
| 786 | { |
| 787 | #ifdef CONFIG_DEBUG_RODATA |
| 788 | int err; |
| 789 | char opc[BREAK_INSTR_SIZE]; |
| 790 | |
| 791 | if (bpt->type != BP_POKE_BREAKPOINT) |
| 792 | goto knl_write; |
| 793 | /* |
| 794 | * It is safe to call text_poke() because normal kernel execution |
| 795 | * is stopped on all cores, so long as the text_mutex is not locked. |
| 796 | */ |
| 797 | if (mutex_is_locked(&text_mutex)) |
| 798 | goto knl_write; |
| 799 | text_poke((void *)bpt->bpt_addr, bpt->saved_instr, BREAK_INSTR_SIZE); |
| 800 | err = probe_kernel_read(opc, (char *)bpt->bpt_addr, BREAK_INSTR_SIZE); |
| 801 | if (err || memcmp(opc, bpt->saved_instr, BREAK_INSTR_SIZE)) |
| 802 | goto knl_write; |
| 803 | return err; |
| 804 | knl_write: |
| 805 | #endif /* CONFIG_DEBUG_RODATA */ |
| 806 | return probe_kernel_write((char *)bpt->bpt_addr, |
| 807 | (char *)bpt->saved_instr, BREAK_INSTR_SIZE); |
| 808 | } |
| 809 | |
| 810 | struct kgdb_arch arch_kgdb_ops = { |
| 811 | /* Breakpoint instruction: */ |
| 812 | .gdb_bpt_instr = { 0xcc }, |
| 813 | .flags = KGDB_HW_BREAKPOINT, |
| 814 | .set_hw_breakpoint = kgdb_set_hw_break, |
| 815 | .remove_hw_breakpoint = kgdb_remove_hw_break, |
| 816 | .disable_hw_break = kgdb_disable_hw_debug, |
| 817 | .remove_all_hw_break = kgdb_remove_all_hw_break, |
| 818 | .correct_hw_break = kgdb_correct_hw_break, |
| 819 | }; |