| File: | kern/sys_process.c |
| Warning: | line 651, column 11 Copies out a struct with uncleared padding (>= 4 bytes) |
| 1 | /*- | |||
| 2 | * Copyright (c) 1994, Sean Eric Fagan | |||
| 3 | * All rights reserved. | |||
| 4 | * | |||
| 5 | * Redistribution and use in source and binary forms, with or without | |||
| 6 | * modification, are permitted provided that the following conditions | |||
| 7 | * are met: | |||
| 8 | * 1. Redistributions of source code must retain the above copyright | |||
| 9 | * notice, this list of conditions and the following disclaimer. | |||
| 10 | * 2. Redistributions in binary form must reproduce the above copyright | |||
| 11 | * notice, this list of conditions and the following disclaimer in the | |||
| 12 | * documentation and/or other materials provided with the distribution. | |||
| 13 | * 3. All advertising materials mentioning features or use of this software | |||
| 14 | * must display the following acknowledgement: | |||
| 15 | * This product includes software developed by Sean Eric Fagan. | |||
| 16 | * 4. The name of the author may not be used to endorse or promote products | |||
| 17 | * derived from this software without specific prior written permission. | |||
| 18 | * | |||
| 19 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND | |||
| 20 | * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||
| 21 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||
| 22 | * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE | |||
| 23 | * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||
| 24 | * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS | |||
| 25 | * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) | |||
| 26 | * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT | |||
| 27 | * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY | |||
| 28 | * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF | |||
| 29 | * SUCH DAMAGE. | |||
| 30 | */ | |||
| 31 | ||||
| 32 | #include <sys/cdefs.h> | |||
| 33 | __FBSDID("$FreeBSD: releng/11.0/sys/kern/sys_process.c 303197 2016-07-22 17:34:58Z kib $")__asm__(".ident\t\"" "$FreeBSD: releng/11.0/sys/kern/sys_process.c 303197 2016-07-22 17:34:58Z kib $" "\""); | |||
| 34 | ||||
| 35 | #include "opt_compat.h" | |||
| 36 | ||||
| 37 | #include <sys/param.h> | |||
| 38 | #include <sys/systm.h> | |||
| 39 | #include <sys/lock.h> | |||
| 40 | #include <sys/mutex.h> | |||
| 41 | #include <sys/syscallsubr.h> | |||
| 42 | #include <sys/sysent.h> | |||
| 43 | #include <sys/sysproto.h> | |||
| 44 | #include <sys/pioctl.h> | |||
| 45 | #include <sys/priv.h> | |||
| 46 | #include <sys/proc.h> | |||
| 47 | #include <sys/vnode.h> | |||
| 48 | #include <sys/ptrace.h> | |||
| 49 | #include <sys/rwlock.h> | |||
| 50 | #include <sys/sx.h> | |||
| 51 | #include <sys/malloc.h> | |||
| 52 | #include <sys/signalvar.h> | |||
| 53 | ||||
| 54 | #include <machine/reg.h> | |||
| 55 | ||||
| 56 | #include <security/audit/audit.h> | |||
| 57 | ||||
| 58 | #include <vm/vm.h> | |||
| 59 | #include <vm/pmap.h> | |||
| 60 | #include <vm/vm_extern.h> | |||
| 61 | #include <vm/vm_map.h> | |||
| 62 | #include <vm/vm_kern.h> | |||
| 63 | #include <vm/vm_object.h> | |||
| 64 | #include <vm/vm_page.h> | |||
| 65 | #include <vm/vm_param.h> | |||
| 66 | ||||
| 67 | #ifdef COMPAT_FREEBSD321 | |||
| 68 | #include <sys/procfs.h> | |||
| 69 | #include <compat/freebsd32/freebsd32_signal.h> | |||
| 70 | ||||
| 71 | struct ptrace_io_desc32 { | |||
| 72 | int piod_op; | |||
| 73 | uint32_t piod_offs; | |||
| 74 | uint32_t piod_addr; | |||
| 75 | uint32_t piod_len; | |||
| 76 | }; | |||
| 77 | ||||
| 78 | struct ptrace_vm_entry32 { | |||
| 79 | int pve_entry; | |||
| 80 | int pve_timestamp; | |||
| 81 | uint32_t pve_start; | |||
| 82 | uint32_t pve_end; | |||
| 83 | uint32_t pve_offset; | |||
| 84 | u_int pve_prot; | |||
| 85 | u_int pve_pathlen; | |||
| 86 | int32_t pve_fileid; | |||
| 87 | u_int pve_fsid; | |||
| 88 | uint32_t pve_path; | |||
| 89 | }; | |||
| 90 | ||||
| 91 | struct ptrace_lwpinfo32 { | |||
| 92 | lwpid_t pl_lwpid; /* LWP described. */ | |||
| 93 | int pl_event; /* Event that stopped the LWP. */ | |||
| 94 | int pl_flags; /* LWP flags. */ | |||
| 95 | sigset_t pl_sigmask; /* LWP signal mask */ | |||
| 96 | sigset_t pl_siglist; /* LWP pending signal */ | |||
| 97 | struct siginfo32 pl_siginfo; /* siginfo for signal */ | |||
| 98 | char pl_tdname[MAXCOMLEN19 + 1]; /* LWP name. */ | |||
| 99 | pid_t pl_child_pid; /* New child pid */ | |||
| 100 | u_int pl_syscall_code; | |||
| 101 | u_int pl_syscall_narg; | |||
| 102 | }; | |||
| 103 | ||||
| 104 | #endif | |||
| 105 | ||||
| 106 | /* | |||
| 107 | * Functions implemented using PROC_ACTION(): | |||
| 108 | * | |||
| 109 | * proc_read_regs(proc, regs) | |||
| 110 | * Get the current user-visible register set from the process | |||
| 111 | * and copy it into the regs structure (<machine/reg.h>). | |||
| 112 | * The process is stopped at the time read_regs is called. | |||
| 113 | * | |||
| 114 | * proc_write_regs(proc, regs) | |||
| 115 | * Update the current register set from the passed in regs | |||
| 116 | * structure. Take care to avoid clobbering special CPU | |||
| 117 | * registers or privileged bits in the PSL. | |||
| 118 | * Depending on the architecture this may have fix-up work to do, | |||
| 119 | * especially if the IAR or PCW are modified. | |||
| 120 | * The process is stopped at the time write_regs is called. | |||
| 121 | * | |||
| 122 | * proc_read_fpregs, proc_write_fpregs | |||
| 123 | * deal with the floating point register set, otherwise as above. | |||
| 124 | * | |||
| 125 | * proc_read_dbregs, proc_write_dbregs | |||
| 126 | * deal with the processor debug register set, otherwise as above. | |||
| 127 | * | |||
| 128 | * proc_sstep(proc) | |||
| 129 | * Arrange for the process to trap after executing a single instruction. | |||
| 130 | */ | |||
| 131 | ||||
| 132 | #define PROC_ACTION(action)do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (action); return ( error); } while(0) do { \ | |||
| 133 | int error; \ | |||
| 134 | \ | |||
| 135 | PROC_LOCK_ASSERT(td->td_proc, MA_OWNED)(void)0; \ | |||
| 136 | if ((td->td_proc->p_flag & P_INMEM0x10000000) == 0) \ | |||
| 137 | error = EIO5; \ | |||
| 138 | else \ | |||
| 139 | error = (action); \ | |||
| 140 | return (error); \ | |||
| 141 | } while(0) | |||
| 142 | ||||
| 143 | int | |||
| 144 | proc_read_regs(struct thread *td, struct reg *regs) | |||
| 145 | { | |||
| 146 | ||||
| 147 | PROC_ACTION(fill_regs(td, regs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_regs(td, regs )); return (error); } while(0); | |||
| 148 | } | |||
| 149 | ||||
| 150 | int | |||
| 151 | proc_write_regs(struct thread *td, struct reg *regs) | |||
| 152 | { | |||
| 153 | ||||
| 154 | PROC_ACTION(set_regs(td, regs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_regs(td, regs )); return (error); } while(0); | |||
| 155 | } | |||
| 156 | ||||
| 157 | int | |||
| 158 | proc_read_dbregs(struct thread *td, struct dbreg *dbregs) | |||
| 159 | { | |||
| 160 | ||||
| 161 | PROC_ACTION(fill_dbregs(td, dbregs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_dbregs(td, dbregs )); return (error); } while(0); | |||
| 162 | } | |||
| 163 | ||||
| 164 | int | |||
| 165 | proc_write_dbregs(struct thread *td, struct dbreg *dbregs) | |||
| 166 | { | |||
| 167 | ||||
| 168 | PROC_ACTION(set_dbregs(td, dbregs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_dbregs(td, dbregs )); return (error); } while(0); | |||
| 169 | } | |||
| 170 | ||||
| 171 | /* | |||
| 172 | * Ptrace doesn't support fpregs at all, and there are no security holes | |||
| 173 | * or translations for fpregs, so we can just copy them. | |||
| 174 | */ | |||
| 175 | int | |||
| 176 | proc_read_fpregs(struct thread *td, struct fpreg *fpregs) | |||
| 177 | { | |||
| 178 | ||||
| 179 | PROC_ACTION(fill_fpregs(td, fpregs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_fpregs(td, fpregs )); return (error); } while(0); | |||
| 180 | } | |||
| 181 | ||||
| 182 | int | |||
| 183 | proc_write_fpregs(struct thread *td, struct fpreg *fpregs) | |||
| 184 | { | |||
| 185 | ||||
| 186 | PROC_ACTION(set_fpregs(td, fpregs))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_fpregs(td, fpregs )); return (error); } while(0); | |||
| 187 | } | |||
| 188 | ||||
| 189 | #ifdef COMPAT_FREEBSD321 | |||
| 190 | /* For 32 bit binaries, we need to expose the 32 bit regs layouts. */ | |||
| 191 | int | |||
| 192 | proc_read_regs32(struct thread *td, struct reg32 *regs32) | |||
| 193 | { | |||
| 194 | ||||
| 195 | PROC_ACTION(fill_regs32(td, regs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_regs32(td, regs32 )); return (error); } while(0); | |||
| 196 | } | |||
| 197 | ||||
| 198 | int | |||
| 199 | proc_write_regs32(struct thread *td, struct reg32 *regs32) | |||
| 200 | { | |||
| 201 | ||||
| 202 | PROC_ACTION(set_regs32(td, regs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_regs32(td, regs32 )); return (error); } while(0); | |||
| 203 | } | |||
| 204 | ||||
| 205 | int | |||
| 206 | proc_read_dbregs32(struct thread *td, struct dbreg32 *dbregs32) | |||
| 207 | { | |||
| 208 | ||||
| 209 | PROC_ACTION(fill_dbregs32(td, dbregs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_dbregs32(td, dbregs32)); return (error); } while(0); | |||
| 210 | } | |||
| 211 | ||||
| 212 | int | |||
| 213 | proc_write_dbregs32(struct thread *td, struct dbreg32 *dbregs32) | |||
| 214 | { | |||
| 215 | ||||
| 216 | PROC_ACTION(set_dbregs32(td, dbregs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_dbregs32(td, dbregs32 )); return (error); } while(0); | |||
| 217 | } | |||
| 218 | ||||
| 219 | int | |||
| 220 | proc_read_fpregs32(struct thread *td, struct fpreg32 *fpregs32) | |||
| 221 | { | |||
| 222 | ||||
| 223 | PROC_ACTION(fill_fpregs32(td, fpregs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (fill_fpregs32(td, fpregs32)); return (error); } while(0); | |||
| 224 | } | |||
| 225 | ||||
| 226 | int | |||
| 227 | proc_write_fpregs32(struct thread *td, struct fpreg32 *fpregs32) | |||
| 228 | { | |||
| 229 | ||||
| 230 | PROC_ACTION(set_fpregs32(td, fpregs32))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (set_fpregs32(td, fpregs32 )); return (error); } while(0); | |||
| 231 | } | |||
| 232 | #endif | |||
| 233 | ||||
| 234 | int | |||
| 235 | proc_sstep(struct thread *td) | |||
| 236 | { | |||
| 237 | ||||
| 238 | PROC_ACTION(ptrace_single_step(td))do { int error; (void)0; if ((td->td_proc->p_flag & 0x10000000) == 0) error = 5; else error = (ptrace_single_step (td)); return (error); } while(0); | |||
| 239 | } | |||
| 240 | ||||
| 241 | int | |||
| 242 | proc_rwmem(struct proc *p, struct uio *uio) | |||
| 243 | { | |||
| 244 | vm_map_t map; | |||
| 245 | vm_offset_t pageno; /* page number */ | |||
| 246 | vm_prot_t reqprot; | |||
| 247 | int error, fault_flags, page_offset, writing; | |||
| 248 | ||||
| 249 | /* | |||
| 250 | * Assert that someone has locked this vmspace. (Should be | |||
| 251 | * curthread but we can't assert that.) This keeps the process | |||
| 252 | * from exiting out from under us until this operation completes. | |||
| 253 | */ | |||
| 254 | PROC_ASSERT_HELD(p)do { do { } while (0); } while (0); | |||
| 255 | PROC_LOCK_ASSERT(p, MA_NOTOWNED)(void)0; | |||
| 256 | ||||
| 257 | /* | |||
| 258 | * The map we want... | |||
| 259 | */ | |||
| 260 | map = &p->p_vmspace->vm_map; | |||
| 261 | ||||
| 262 | /* | |||
| 263 | * If we are writing, then we request vm_fault() to create a private | |||
| 264 | * copy of each page. Since these copies will not be writeable by the | |||
| 265 | * process, we must explicity request that they be dirtied. | |||
| 266 | */ | |||
| 267 | writing = uio->uio_rw == UIO_WRITE; | |||
| 268 | reqprot = writing ? VM_PROT_COPY((vm_prot_t) 0x08) | VM_PROT_READ((vm_prot_t) 0x01) : VM_PROT_READ((vm_prot_t) 0x01); | |||
| 269 | fault_flags = writing ? VM_FAULT_DIRTY2 : VM_FAULT_NORMAL0; | |||
| 270 | ||||
| 271 | /* | |||
| 272 | * Only map in one page at a time. We don't have to, but it | |||
| 273 | * makes things easier. This way is trivial - right? | |||
| 274 | */ | |||
| 275 | do { | |||
| 276 | vm_offset_t uva; | |||
| 277 | u_int len; | |||
| 278 | vm_page_t m; | |||
| 279 | ||||
| 280 | uva = (vm_offset_t)uio->uio_offset; | |||
| 281 | ||||
| 282 | /* | |||
| 283 | * Get the page number of this segment. | |||
| 284 | */ | |||
| 285 | pageno = trunc_page(uva)((unsigned long)(uva) & ~(((1<<12)-1))); | |||
| 286 | page_offset = uva - pageno; | |||
| 287 | ||||
| 288 | /* | |||
| 289 | * How many bytes to copy | |||
| 290 | */ | |||
| 291 | len = min(PAGE_SIZE(1<<12) - page_offset, uio->uio_resid); | |||
| 292 | ||||
| 293 | /* | |||
| 294 | * Fault and hold the page on behalf of the process. | |||
| 295 | */ | |||
| 296 | error = vm_fault_hold(map, pageno, reqprot, fault_flags, &m); | |||
| 297 | if (error != KERN_SUCCESS0) { | |||
| 298 | if (error == KERN_RESOURCE_SHORTAGE6) | |||
| 299 | error = ENOMEM12; | |||
| 300 | else | |||
| 301 | error = EFAULT14; | |||
| 302 | break; | |||
| 303 | } | |||
| 304 | ||||
| 305 | /* | |||
| 306 | * Now do the i/o move. | |||
| 307 | */ | |||
| 308 | error = uiomove_fromphys(&m, page_offset, len, uio); | |||
| 309 | ||||
| 310 | /* Make the I-cache coherent for breakpoints. */ | |||
| 311 | if (writing && error == 0) { | |||
| 312 | vm_map_lock_read(map)_vm_map_lock_read(map, ((void *)0), 0); | |||
| 313 | if (vm_map_check_protection(map, pageno, pageno + | |||
| 314 | PAGE_SIZE(1<<12), VM_PROT_EXECUTE((vm_prot_t) 0x04))) | |||
| 315 | vm_sync_icache(map, uva, len); | |||
| 316 | vm_map_unlock_read(map)_vm_map_unlock_read(map, ((void *)0), 0); | |||
| 317 | } | |||
| 318 | ||||
| 319 | /* | |||
| 320 | * Release the page. | |||
| 321 | */ | |||
| 322 | vm_page_lock(m)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( (((struct mtx *)(&pa_lock[((((((m)))->phys_addr)) >> 21) % 256])))))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&((((((((struct mtx *)(&pa_lock[((((((m)))->phys_addr )) >> 21) % 256]))))))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&((((((((struct mtx *)(&pa_lock [((((((m)))->phys_addr)) >> 21) % 256]))))))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) ((((((struct mtx *)(&pa_lock[((((((m)))-> phys_addr)) >> 21) % 256])))))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 323 | vm_page_unhold(m); | |||
| 324 | vm_page_unlock(m)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( ((struct mtx *)(&pa_lock[((((((m)))->phys_addr)) >> 21) % 256])))))))->lock_object.lo_data == 0) do { (void)0 ; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) ((((((struct mtx *)(&pa_lock[((((((m )))->phys_addr)) >> 21) % 256])))))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if ((((((((struct mtx *)(&pa_lock[((((((m)))-> phys_addr)) >> 21) % 256])))))))->mtx_lock != _tid || !atomic_cmpset_long(&((((((((struct mtx *)(&pa_lock[ ((((((m)))->phys_addr)) >> 21) % 256]))))))))->mtx_lock , (_tid), 0x00000004)) __mtx_unlock_sleep(&((((((((struct mtx *)(&pa_lock[((((((m)))->phys_addr)) >> 21) % 256]))))))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 325 | ||||
| 326 | } while (error == 0 && uio->uio_resid > 0); | |||
| 327 | ||||
| 328 | return (error); | |||
| 329 | } | |||
| 330 | ||||
| 331 | static ssize_t | |||
| 332 | proc_iop(struct thread *td, struct proc *p, vm_offset_t va, void *buf, | |||
| 333 | size_t len, enum uio_rw rw) | |||
| 334 | { | |||
| 335 | struct iovec iov; | |||
| 336 | struct uio uio; | |||
| 337 | ssize_t slen; | |||
| 338 | int error; | |||
| 339 | ||||
| 340 | MPASS(len < SSIZE_MAX)do { } while (0); | |||
| 341 | slen = (ssize_t)len; | |||
| 342 | ||||
| 343 | iov.iov_base = (caddr_t)buf; | |||
| 344 | iov.iov_len = len; | |||
| 345 | uio.uio_iov = &iov; | |||
| 346 | uio.uio_iovcnt = 1; | |||
| 347 | uio.uio_offset = va; | |||
| 348 | uio.uio_resid = slen; | |||
| 349 | uio.uio_segflg = UIO_SYSSPACE; | |||
| 350 | uio.uio_rw = rw; | |||
| 351 | uio.uio_td = td; | |||
| 352 | error = proc_rwmem(p, &uio); | |||
| 353 | if (uio.uio_resid == slen) | |||
| 354 | return (-1); | |||
| 355 | return (slen - uio.uio_resid); | |||
| 356 | } | |||
| 357 | ||||
| 358 | ssize_t | |||
| 359 | proc_readmem(struct thread *td, struct proc *p, vm_offset_t va, void *buf, | |||
| 360 | size_t len) | |||
| 361 | { | |||
| 362 | ||||
| 363 | return (proc_iop(td, p, va, buf, len, UIO_READ)); | |||
| 364 | } | |||
| 365 | ||||
| 366 | ssize_t | |||
| 367 | proc_writemem(struct thread *td, struct proc *p, vm_offset_t va, void *buf, | |||
| 368 | size_t len) | |||
| 369 | { | |||
| 370 | ||||
| 371 | return (proc_iop(td, p, va, buf, len, UIO_WRITE)); | |||
| 372 | } | |||
| 373 | ||||
| 374 | static int | |||
| 375 | ptrace_vm_entry(struct thread *td, struct proc *p, struct ptrace_vm_entry *pve) | |||
| 376 | { | |||
| 377 | struct vattr vattr; | |||
| 378 | vm_map_t map; | |||
| 379 | vm_map_entry_t entry; | |||
| 380 | vm_object_t obj, tobj, lobj; | |||
| 381 | struct vmspace *vm; | |||
| 382 | struct vnode *vp; | |||
| 383 | char *freepath, *fullpath; | |||
| 384 | u_int pathlen; | |||
| 385 | int error, index; | |||
| 386 | ||||
| 387 | error = 0; | |||
| 388 | obj = NULL((void *)0); | |||
| 389 | ||||
| 390 | vm = vmspace_acquire_ref(p); | |||
| 391 | map = &vm->vm_map; | |||
| 392 | vm_map_lock_read(map)_vm_map_lock_read(map, ((void *)0), 0); | |||
| 393 | ||||
| 394 | do { | |||
| 395 | entry = map->header.next; | |||
| 396 | index = 0; | |||
| 397 | while (index < pve->pve_entry && entry != &map->header) { | |||
| 398 | entry = entry->next; | |||
| 399 | index++; | |||
| 400 | } | |||
| 401 | if (index != pve->pve_entry) { | |||
| 402 | error = EINVAL22; | |||
| 403 | break; | |||
| 404 | } | |||
| 405 | while (entry != &map->header && | |||
| 406 | (entry->eflags & MAP_ENTRY_IS_SUB_MAP0x0002) != 0) { | |||
| 407 | entry = entry->next; | |||
| 408 | index++; | |||
| 409 | } | |||
| 410 | if (entry == &map->header) { | |||
| 411 | error = ENOENT2; | |||
| 412 | break; | |||
| 413 | } | |||
| 414 | ||||
| 415 | /* We got an entry. */ | |||
| 416 | pve->pve_entry = index + 1; | |||
| 417 | pve->pve_timestamp = map->timestamp; | |||
| 418 | pve->pve_start = entry->start; | |||
| 419 | pve->pve_end = entry->end - 1; | |||
| 420 | pve->pve_offset = entry->offset; | |||
| 421 | pve->pve_prot = entry->protection; | |||
| 422 | ||||
| 423 | /* Backing object's path needed? */ | |||
| 424 | if (pve->pve_pathlen == 0) | |||
| 425 | break; | |||
| 426 | ||||
| 427 | pathlen = pve->pve_pathlen; | |||
| 428 | pve->pve_pathlen = 0; | |||
| 429 | ||||
| 430 | obj = entry->object.vm_object; | |||
| 431 | if (obj != NULL((void *)0)) | |||
| 432 | VM_OBJECT_RLOCK(obj)__rw_rlock(&((&(obj)->lock))->rw_lock, ((void * )0), 0); | |||
| 433 | } while (0); | |||
| 434 | ||||
| 435 | vm_map_unlock_read(map)_vm_map_unlock_read(map, ((void *)0), 0); | |||
| 436 | ||||
| 437 | pve->pve_fsid = VNOVAL(-1); | |||
| 438 | pve->pve_fileid = VNOVAL(-1); | |||
| 439 | ||||
| 440 | if (error == 0 && obj != NULL((void *)0)) { | |||
| 441 | lobj = obj; | |||
| 442 | for (tobj = obj; tobj != NULL((void *)0); tobj = tobj->backing_object) { | |||
| 443 | if (tobj != obj) | |||
| 444 | VM_OBJECT_RLOCK(tobj)__rw_rlock(&((&(tobj)->lock))->rw_lock, ((void * )0), 0); | |||
| 445 | if (lobj != obj) | |||
| 446 | VM_OBJECT_RUNLOCK(lobj)_rw_runlock_cookie(&((&(lobj)->lock))->rw_lock, ((void *)0), 0); | |||
| 447 | lobj = tobj; | |||
| 448 | pve->pve_offset += tobj->backing_object_offset; | |||
| 449 | } | |||
| 450 | vp = vm_object_vnode(lobj); | |||
| 451 | if (vp != NULL((void *)0)) | |||
| 452 | vref(vp); | |||
| 453 | if (lobj != obj) | |||
| 454 | VM_OBJECT_RUNLOCK(lobj)_rw_runlock_cookie(&((&(lobj)->lock))->rw_lock, ((void *)0), 0); | |||
| 455 | VM_OBJECT_RUNLOCK(obj)_rw_runlock_cookie(&((&(obj)->lock))->rw_lock, ( (void *)0), 0); | |||
| 456 | ||||
| 457 | if (vp != NULL((void *)0)) { | |||
| 458 | freepath = NULL((void *)0); | |||
| 459 | fullpath = NULL((void *)0); | |||
| 460 | vn_fullpath(td, vp, &fullpath, &freepath); | |||
| 461 | vn_lock(vp, LK_SHARED | LK_RETRY)_vn_lock(vp, 0x200000 | 0x000400, "/usr/src/sys/kern/sys_process.c" , 461); | |||
| 462 | if (VOP_GETATTR(vp, &vattr, td->td_ucred) == 0) { | |||
| 463 | pve->pve_fileid = vattr.va_fileid; | |||
| 464 | pve->pve_fsid = vattr.va_fsid; | |||
| 465 | } | |||
| 466 | vput(vp); | |||
| 467 | ||||
| 468 | if (fullpath != NULL((void *)0)) { | |||
| 469 | pve->pve_pathlen = strlen(fullpath) + 1; | |||
| 470 | if (pve->pve_pathlen <= pathlen) { | |||
| 471 | error = copyout(fullpath, pve->pve_path, | |||
| 472 | pve->pve_pathlen); | |||
| 473 | } else | |||
| 474 | error = ENAMETOOLONG63; | |||
| 475 | } | |||
| 476 | if (freepath != NULL((void *)0)) | |||
| 477 | free(freepath, M_TEMP); | |||
| 478 | } | |||
| 479 | } | |||
| 480 | vmspace_free(vm); | |||
| 481 | if (error == 0) | |||
| 482 | CTR3(KTR_PTRACE, "PT_VM_ENTRY: pid %d, entry %d, start %p",(void)0 | |||
| 483 | p->p_pid, pve->pve_entry, pve->pve_start)(void)0; | |||
| 484 | ||||
| 485 | return (error); | |||
| 486 | } | |||
| 487 | ||||
| 488 | #ifdef COMPAT_FREEBSD321 | |||
| 489 | static int | |||
| 490 | ptrace_vm_entry32(struct thread *td, struct proc *p, | |||
| 491 | struct ptrace_vm_entry32 *pve32) | |||
| 492 | { | |||
| 493 | struct ptrace_vm_entry pve; | |||
| 494 | int error; | |||
| 495 | ||||
| 496 | pve.pve_entry = pve32->pve_entry; | |||
| 497 | pve.pve_pathlen = pve32->pve_pathlen; | |||
| 498 | pve.pve_path = (void *)(uintptr_t)pve32->pve_path; | |||
| 499 | ||||
| 500 | error = ptrace_vm_entry(td, p, &pve); | |||
| 501 | if (error == 0) { | |||
| 502 | pve32->pve_entry = pve.pve_entry; | |||
| 503 | pve32->pve_timestamp = pve.pve_timestamp; | |||
| 504 | pve32->pve_start = pve.pve_start; | |||
| 505 | pve32->pve_end = pve.pve_end; | |||
| 506 | pve32->pve_offset = pve.pve_offset; | |||
| 507 | pve32->pve_prot = pve.pve_prot; | |||
| 508 | pve32->pve_fileid = pve.pve_fileid; | |||
| 509 | pve32->pve_fsid = pve.pve_fsid; | |||
| 510 | } | |||
| 511 | ||||
| 512 | pve32->pve_pathlen = pve.pve_pathlen; | |||
| 513 | return (error); | |||
| 514 | } | |||
| 515 | ||||
| 516 | static void | |||
| 517 | ptrace_lwpinfo_to32(const struct ptrace_lwpinfo *pl, | |||
| 518 | struct ptrace_lwpinfo32 *pl32) | |||
| 519 | { | |||
| 520 | ||||
| 521 | pl32->pl_lwpid = pl->pl_lwpid; | |||
| 522 | pl32->pl_event = pl->pl_event; | |||
| 523 | pl32->pl_flags = pl->pl_flags; | |||
| 524 | pl32->pl_sigmask = pl->pl_sigmask; | |||
| 525 | pl32->pl_siglist = pl->pl_siglist; | |||
| 526 | siginfo_to_siginfo32(&pl->pl_siginfo, &pl32->pl_siginfo); | |||
| 527 | strcpy(pl32->pl_tdname, pl->pl_tdname); | |||
| 528 | pl32->pl_child_pid = pl->pl_child_pid; | |||
| 529 | pl32->pl_syscall_code = pl->pl_syscall_code; | |||
| 530 | pl32->pl_syscall_narg = pl->pl_syscall_narg; | |||
| 531 | } | |||
| 532 | #endif /* COMPAT_FREEBSD32 */ | |||
| 533 | ||||
| 534 | /* | |||
| 535 | * Process debugging system call. | |||
| 536 | */ | |||
| 537 | #ifndef _SYS_SYSPROTO_H_ | |||
| 538 | struct ptrace_args { | |||
| 539 | int req; | |||
| 540 | pid_t pid; | |||
| 541 | caddr_t addr; | |||
| 542 | int data; | |||
| 543 | }; | |||
| 544 | #endif | |||
| 545 | ||||
| 546 | #ifdef COMPAT_FREEBSD321 | |||
| 547 | /* | |||
| 548 | * This CPP subterfuge is to try and reduce the number of ifdefs in | |||
| 549 | * the body of the code. | |||
| 550 | * COPYIN(uap->addr, &r.reg, sizeof r.reg); | |||
| 551 | * becomes either: | |||
| 552 | * copyin(uap->addr, &r.reg, sizeof r.reg); | |||
| 553 | * or | |||
| 554 | * copyin(uap->addr, &r.reg32, sizeof r.reg32); | |||
| 555 | * .. except this is done at runtime. | |||
| 556 | */ | |||
| 557 | #define COPYIN(u, k, s) wrap32 ? \ | |||
| 558 | copyin(u, k ## 32, s ## 32) : \ | |||
| 559 | copyin(u, k, s) | |||
| 560 | #define COPYOUT(k, u, s) wrap32 ? \ | |||
| 561 | copyout(k ## 32, u, s ## 32) : \ | |||
| 562 | copyout(k, u, s) | |||
| 563 | #else | |||
| 564 | #define COPYIN(u, k, s) copyin(u, k, s) | |||
| 565 | #define COPYOUT(k, u, s) copyout(k, u, s) | |||
| 566 | #endif | |||
| 567 | int | |||
| 568 | sys_ptrace(struct thread *td, struct ptrace_args *uap) | |||
| 569 | { | |||
| 570 | /* | |||
| 571 | * XXX this obfuscation is to reduce stack usage, but the register | |||
| 572 | * structs may be too large to put on the stack anyway. | |||
| 573 | */ | |||
| 574 | union { | |||
| 575 | struct ptrace_io_desc piod; | |||
| 576 | struct ptrace_lwpinfo pl; | |||
| 577 | struct ptrace_vm_entry pve; | |||
| 578 | struct dbreg dbreg; | |||
| 579 | struct fpreg fpreg; | |||
| 580 | struct reg reg; | |||
| 581 | #ifdef COMPAT_FREEBSD321 | |||
| 582 | struct dbreg32 dbreg32; | |||
| 583 | struct fpreg32 fpreg32; | |||
| 584 | struct reg32 reg32; | |||
| 585 | struct ptrace_io_desc32 piod32; | |||
| 586 | struct ptrace_lwpinfo32 pl32; | |||
| 587 | struct ptrace_vm_entry32 pve32; | |||
| 588 | #endif | |||
| 589 | } r; | |||
| 590 | void *addr; | |||
| 591 | int error = 0; | |||
| 592 | #ifdef COMPAT_FREEBSD321 | |||
| 593 | int wrap32 = 0; | |||
| 594 | ||||
| 595 | if (SV_CURPROC_FLAG(SV_ILP32)((((__curthread())->td_proc))->p_sysent->sv_flags & (0x000100))) | |||
| ||||
| 596 | wrap32 = 1; | |||
| 597 | #endif | |||
| 598 | AUDIT_ARG_PID(uap->pid)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_pid ((uap->pid)); } while (0); | |||
| 599 | AUDIT_ARG_CMD(uap->req)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_cmd ((uap->req)); } while (0); | |||
| 600 | AUDIT_ARG_VALUE(uap->data)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_value ((uap->data)); } while (0); | |||
| 601 | addr = &r; | |||
| 602 | switch (uap->req) { | |||
| 603 | case PT_GETREGS33: | |||
| 604 | case PT_GETFPREGS35: | |||
| 605 | case PT_GETDBREGS37: | |||
| 606 | case PT_LWPINFO13: | |||
| 607 | break; | |||
| 608 | case PT_SETREGS34: | |||
| 609 | error = COPYIN(uap->addr, &r.reg, sizeof r.reg); | |||
| 610 | break; | |||
| 611 | case PT_SETFPREGS36: | |||
| 612 | error = COPYIN(uap->addr, &r.fpreg, sizeof r.fpreg); | |||
| 613 | break; | |||
| 614 | case PT_SETDBREGS38: | |||
| 615 | error = COPYIN(uap->addr, &r.dbreg, sizeof r.dbreg); | |||
| 616 | break; | |||
| 617 | case PT_IO12: | |||
| 618 | error = COPYIN(uap->addr, &r.piod, sizeof r.piod); | |||
| 619 | break; | |||
| 620 | case PT_VM_ENTRY41: | |||
| 621 | error = COPYIN(uap->addr, &r.pve, sizeof r.pve); | |||
| 622 | break; | |||
| 623 | default: | |||
| 624 | addr = uap->addr; | |||
| 625 | break; | |||
| 626 | } | |||
| 627 | if (error) | |||
| 628 | return (error); | |||
| 629 | ||||
| 630 | error = kern_ptrace(td, uap->req, uap->pid, addr, uap->data); | |||
| 631 | if (error) | |||
| 632 | return (error); | |||
| 633 | ||||
| 634 | switch (uap->req) { | |||
| 635 | case PT_VM_ENTRY41: | |||
| 636 | error = COPYOUT(&r.pve, uap->addr, sizeof r.pve); | |||
| 637 | break; | |||
| 638 | case PT_IO12: | |||
| 639 | error = COPYOUT(&r.piod, uap->addr, sizeof r.piod); | |||
| 640 | break; | |||
| 641 | case PT_GETREGS33: | |||
| 642 | error = COPYOUT(&r.reg, uap->addr, sizeof r.reg); | |||
| 643 | break; | |||
| 644 | case PT_GETFPREGS35: | |||
| 645 | error = COPYOUT(&r.fpreg, uap->addr, sizeof r.fpreg); | |||
| 646 | break; | |||
| 647 | case PT_GETDBREGS37: | |||
| 648 | error = COPYOUT(&r.dbreg, uap->addr, sizeof r.dbreg); | |||
| 649 | break; | |||
| 650 | case PT_LWPINFO13: | |||
| 651 | error = copyout(&r.pl, uap->addr, uap->data); | |||
| ||||
| 652 | break; | |||
| 653 | } | |||
| 654 | ||||
| 655 | return (error); | |||
| 656 | } | |||
| 657 | #undef COPYIN | |||
| 658 | #undef COPYOUT | |||
| 659 | ||||
| 660 | #ifdef COMPAT_FREEBSD321 | |||
| 661 | /* | |||
| 662 | * PROC_READ(regs, td2, addr); | |||
| 663 | * becomes either: | |||
| 664 | * proc_read_regs(td2, addr); | |||
| 665 | * or | |||
| 666 | * proc_read_regs32(td2, addr); | |||
| 667 | * .. except this is done at runtime. There is an additional | |||
| 668 | * complication in that PROC_WRITE disallows 32 bit consumers | |||
| 669 | * from writing to 64 bit address space targets. | |||
| 670 | */ | |||
| 671 | #define PROC_READ(w, t, a) wrap32 ? \ | |||
| 672 | proc_read_ ## w ## 32(t, a) : \ | |||
| 673 | proc_read_ ## w (t, a) | |||
| 674 | #define PROC_WRITE(w, t, a) wrap32 ? \ | |||
| 675 | (safe ? proc_write_ ## w ## 32(t, a) : EINVAL22 ) : \ | |||
| 676 | proc_write_ ## w (t, a) | |||
| 677 | #else | |||
| 678 | #define PROC_READ(w, t, a) proc_read_ ## w (t, a) | |||
| 679 | #define PROC_WRITE(w, t, a) proc_write_ ## w (t, a) | |||
| 680 | #endif | |||
| 681 | ||||
| 682 | int | |||
| 683 | kern_ptrace(struct thread *td, int req, pid_t pid, void *addr, int data) | |||
| 684 | { | |||
| 685 | struct iovec iov; | |||
| 686 | struct uio uio; | |||
| 687 | struct proc *curp, *p, *pp; | |||
| 688 | struct thread *td2 = NULL((void *)0), *td3; | |||
| 689 | struct ptrace_io_desc *piod = NULL((void *)0); | |||
| 690 | struct ptrace_lwpinfo *pl; | |||
| 691 | int error, num, tmp; | |||
| 692 | int proctree_locked = 0; | |||
| 693 | lwpid_t tid = 0, *buf; | |||
| 694 | #ifdef COMPAT_FREEBSD321 | |||
| 695 | int wrap32 = 0, safe = 0; | |||
| 696 | struct ptrace_io_desc32 *piod32 = NULL((void *)0); | |||
| 697 | struct ptrace_lwpinfo32 *pl32 = NULL((void *)0); | |||
| 698 | struct ptrace_lwpinfo plr; | |||
| 699 | #endif | |||
| 700 | ||||
| 701 | curp = td->td_proc; | |||
| 702 | ||||
| 703 | /* Lock proctree before locking the process. */ | |||
| 704 | switch (req) { | |||
| 705 | case PT_TRACE_ME0: | |||
| 706 | case PT_ATTACH10: | |||
| 707 | case PT_STEP9: | |||
| 708 | case PT_CONTINUE7: | |||
| 709 | case PT_TO_SCE20: | |||
| 710 | case PT_TO_SCX21: | |||
| 711 | case PT_SYSCALL22: | |||
| 712 | case PT_FOLLOW_FORK23: | |||
| 713 | case PT_LWP_EVENTS24: | |||
| 714 | case PT_DETACH11: | |||
| 715 | sx_xlock(&proctree_lock)(void)__sx_xlock(((&proctree_lock)), (__curthread()), 0, ( ((void *)0)), (0)); | |||
| 716 | proctree_locked = 1; | |||
| 717 | break; | |||
| 718 | default: | |||
| 719 | break; | |||
| 720 | } | |||
| 721 | ||||
| 722 | if (req == PT_TRACE_ME0) { | |||
| 723 | p = td->td_proc; | |||
| 724 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 725 | } else { | |||
| 726 | if (pid <= PID_MAX99999) { | |||
| 727 | if ((p = pfind(pid)) == NULL((void *)0)) { | |||
| 728 | if (proctree_locked) | |||
| 729 | sx_xunlock(&proctree_lock)__sx_xunlock(((&proctree_lock)), (__curthread()), (((void *)0)), (0)); | |||
| 730 | return (ESRCH3); | |||
| 731 | } | |||
| 732 | } else { | |||
| 733 | td2 = tdfind(pid, -1); | |||
| 734 | if (td2 == NULL((void *)0)) { | |||
| 735 | if (proctree_locked) | |||
| 736 | sx_xunlock(&proctree_lock)__sx_xunlock(((&proctree_lock)), (__curthread()), (((void *)0)), (0)); | |||
| 737 | return (ESRCH3); | |||
| 738 | } | |||
| 739 | p = td2->td_proc; | |||
| 740 | tid = pid; | |||
| 741 | pid = p->p_pid; | |||
| 742 | } | |||
| 743 | } | |||
| 744 | AUDIT_ARG_PROCESS(p)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_process ((p)); } while (0); | |||
| 745 | ||||
| 746 | if ((p->p_flag & P_WEXIT0x02000) != 0) { | |||
| 747 | error = ESRCH3; | |||
| 748 | goto fail; | |||
| 749 | } | |||
| 750 | if ((error = p_cansee(td, p)) != 0) | |||
| 751 | goto fail; | |||
| 752 | ||||
| 753 | if ((error = p_candebug(td, p)) != 0) | |||
| 754 | goto fail; | |||
| 755 | ||||
| 756 | /* | |||
| 757 | * System processes can't be debugged. | |||
| 758 | */ | |||
| 759 | if ((p->p_flag & P_SYSTEM0x00200) != 0) { | |||
| 760 | error = EINVAL22; | |||
| 761 | goto fail; | |||
| 762 | } | |||
| 763 | ||||
| 764 | if (tid == 0) { | |||
| 765 | if ((p->p_flag & P_STOPPED_TRACE0x40000) != 0) { | |||
| 766 | KASSERT(p->p_xthread != NULL, ("NULL p_xthread"))do { } while (0); | |||
| 767 | td2 = p->p_xthread; | |||
| 768 | } else { | |||
| 769 | td2 = FIRST_THREAD_IN_PROC(p)((&(p)->p_threads)->tqh_first); | |||
| 770 | } | |||
| 771 | tid = td2->td_tid; | |||
| 772 | } | |||
| 773 | ||||
| 774 | #ifdef COMPAT_FREEBSD321 | |||
| 775 | /* | |||
| 776 | * Test if we're a 32 bit client and what the target is. | |||
| 777 | * Set the wrap controls accordingly. | |||
| 778 | */ | |||
| 779 | if (SV_CURPROC_FLAG(SV_ILP32)((((__curthread())->td_proc))->p_sysent->sv_flags & (0x000100))) { | |||
| 780 | if (SV_PROC_FLAG(td2->td_proc, SV_ILP32)((td2->td_proc)->p_sysent->sv_flags & (0x000100) )) | |||
| 781 | safe = 1; | |||
| 782 | wrap32 = 1; | |||
| 783 | } | |||
| 784 | #endif | |||
| 785 | /* | |||
| 786 | * Permissions check | |||
| 787 | */ | |||
| 788 | switch (req) { | |||
| 789 | case PT_TRACE_ME0: | |||
| 790 | /* | |||
| 791 | * Always legal, when there is a parent process which | |||
| 792 | * could trace us. Otherwise, reject. | |||
| 793 | */ | |||
| 794 | if ((p->p_flag & P_TRACED0x00800) != 0) { | |||
| 795 | error = EBUSY16; | |||
| 796 | goto fail; | |||
| 797 | } | |||
| 798 | if (p->p_pptr == initproc) { | |||
| 799 | error = EPERM1; | |||
| 800 | goto fail; | |||
| 801 | } | |||
| 802 | break; | |||
| 803 | ||||
| 804 | case PT_ATTACH10: | |||
| 805 | /* Self */ | |||
| 806 | if (p == td->td_proc) { | |||
| 807 | error = EINVAL22; | |||
| 808 | goto fail; | |||
| 809 | } | |||
| 810 | ||||
| 811 | /* Already traced */ | |||
| 812 | if (p->p_flag & P_TRACED0x00800) { | |||
| 813 | error = EBUSY16; | |||
| 814 | goto fail; | |||
| 815 | } | |||
| 816 | ||||
| 817 | /* Can't trace an ancestor if you're being traced. */ | |||
| 818 | if (curp->p_flag & P_TRACED0x00800) { | |||
| 819 | for (pp = curp->p_pptr; pp != NULL((void *)0); pp = pp->p_pptr) { | |||
| 820 | if (pp == p) { | |||
| 821 | error = EINVAL22; | |||
| 822 | goto fail; | |||
| 823 | } | |||
| 824 | } | |||
| 825 | } | |||
| 826 | ||||
| 827 | ||||
| 828 | /* OK */ | |||
| 829 | break; | |||
| 830 | ||||
| 831 | case PT_CLEARSTEP16: | |||
| 832 | /* Allow thread to clear single step for itself */ | |||
| 833 | if (td->td_tid == tid) | |||
| 834 | break; | |||
| 835 | ||||
| 836 | /* FALLTHROUGH */ | |||
| 837 | default: | |||
| 838 | /* not being traced... */ | |||
| 839 | if ((p->p_flag & P_TRACED0x00800) == 0) { | |||
| 840 | error = EPERM1; | |||
| 841 | goto fail; | |||
| 842 | } | |||
| 843 | ||||
| 844 | /* not being traced by YOU */ | |||
| 845 | if (p->p_pptr != td->td_proc) { | |||
| 846 | error = EBUSY16; | |||
| 847 | goto fail; | |||
| 848 | } | |||
| 849 | ||||
| 850 | /* not currently stopped */ | |||
| 851 | if ((p->p_flag & (P_STOPPED_SIG0x20000 | P_STOPPED_TRACE0x40000)) == 0 || | |||
| 852 | p->p_suspcount != p->p_numthreads || | |||
| 853 | (p->p_flag & P_WAITED0x01000) == 0) { | |||
| 854 | error = EBUSY16; | |||
| 855 | goto fail; | |||
| 856 | } | |||
| 857 | ||||
| 858 | if ((p->p_flag & P_STOPPED_TRACE0x40000) == 0) { | |||
| 859 | static int count = 0; | |||
| 860 | if (count++ == 0) | |||
| 861 | printf("P_STOPPED_TRACE not set.\n"); | |||
| 862 | } | |||
| 863 | ||||
| 864 | /* OK */ | |||
| 865 | break; | |||
| 866 | } | |||
| 867 | ||||
| 868 | /* Keep this process around until we finish this request. */ | |||
| 869 | _PHOLD(p)do { (void)0; do { } while (0); (p)->p_lock++; if (((p)-> p_flag & 0x10000000) == 0) faultin((p)); } while (0); | |||
| 870 | ||||
| 871 | #ifdef FIX_SSTEP | |||
| 872 | /* | |||
| 873 | * Single step fixup ala procfs | |||
| 874 | */ | |||
| 875 | FIX_SSTEP(td2); | |||
| 876 | #endif | |||
| 877 | ||||
| 878 | /* | |||
| 879 | * Actually do the requests | |||
| 880 | */ | |||
| 881 | ||||
| 882 | td->td_retvaltd_uretoff.tdu_retval[0] = 0; | |||
| 883 | ||||
| 884 | switch (req) { | |||
| 885 | case PT_TRACE_ME0: | |||
| 886 | /* set my trace flag and "owner" so it can read/write me */ | |||
| 887 | p->p_flag |= P_TRACED0x00800; | |||
| 888 | if (p->p_flag & P_PPWAIT0x00010) | |||
| 889 | p->p_flag |= P_PPTRACE0x80000000; | |||
| 890 | p->p_oppid = p->p_pptr->p_pid; | |||
| 891 | CTR1(KTR_PTRACE, "PT_TRACE_ME: pid %d", p->p_pid)(void)0; | |||
| 892 | break; | |||
| 893 | ||||
| 894 | case PT_ATTACH10: | |||
| 895 | /* security check done above */ | |||
| 896 | /* | |||
| 897 | * It would be nice if the tracing relationship was separate | |||
| 898 | * from the parent relationship but that would require | |||
| 899 | * another set of links in the proc struct or for "wait" | |||
| 900 | * to scan the entire proc table. To make life easier, | |||
| 901 | * we just re-parent the process we're trying to trace. | |||
| 902 | * The old parent is remembered so we can put things back | |||
| 903 | * on a "detach". | |||
| 904 | */ | |||
| 905 | p->p_flag |= P_TRACED0x00800; | |||
| 906 | p->p_oppid = p->p_pptr->p_pid; | |||
| 907 | if (p->p_pptr != td->td_proc) { | |||
| 908 | proc_reparent(p, td->td_proc); | |||
| 909 | } | |||
| 910 | data = SIGSTOP17; | |||
| 911 | CTR2(KTR_PTRACE, "PT_ATTACH: pid %d, oppid %d", p->p_pid,(void)0 | |||
| 912 | p->p_oppid)(void)0; | |||
| 913 | goto sendsig; /* in PT_CONTINUE below */ | |||
| 914 | ||||
| 915 | case PT_CLEARSTEP16: | |||
| 916 | CTR2(KTR_PTRACE, "PT_CLEARSTEP: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 917 | p->p_pid)(void)0; | |||
| 918 | error = ptrace_clear_single_step(td2); | |||
| 919 | break; | |||
| 920 | ||||
| 921 | case PT_SETSTEP17: | |||
| 922 | CTR2(KTR_PTRACE, "PT_SETSTEP: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 923 | p->p_pid)(void)0; | |||
| 924 | error = ptrace_single_step(td2); | |||
| 925 | break; | |||
| 926 | ||||
| 927 | case PT_SUSPEND18: | |||
| 928 | CTR2(KTR_PTRACE, "PT_SUSPEND: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 929 | p->p_pid)(void)0; | |||
| 930 | td2->td_dbgflags |= TDB_SUSPEND0x00000001; | |||
| 931 | thread_lock(td2)thread_lock_flags_((td2), 0, "/usr/src/sys/kern/sys_process.c" , 931); | |||
| 932 | td2->td_flags |= TDF_NEEDSUSPCHK0x00008000; | |||
| 933 | thread_unlock(td2)do { if ((((((((td2)->td_lock)))))->lock_object.lo_data != 0)) (((((td2)->td_lock))))->lock_object.lo_data--; else { do { (void)0; do { if (__builtin_expect((sdt_lockstat___spin__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___spin__release-> id, (uintptr_t) ((((td2)->td_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); atomic_store_rel_long(&((((((td2)->td_lock)))))->mtx_lock , 0x00000004); } spinlock_exit(); } while (0); | |||
| 934 | break; | |||
| 935 | ||||
| 936 | case PT_RESUME19: | |||
| 937 | CTR2(KTR_PTRACE, "PT_RESUME: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 938 | p->p_pid)(void)0; | |||
| 939 | td2->td_dbgflags &= ~TDB_SUSPEND0x00000001; | |||
| 940 | break; | |||
| 941 | ||||
| 942 | case PT_FOLLOW_FORK23: | |||
| 943 | CTR3(KTR_PTRACE, "PT_FOLLOW_FORK: pid %d %s -> %s", p->p_pid,(void)0 | |||
| 944 | p->p_flag & P_FOLLOWFORK ? "enabled" : "disabled",(void)0 | |||
| 945 | data ? "enabled" : "disabled")(void)0; | |||
| 946 | if (data) | |||
| 947 | p->p_flag |= P_FOLLOWFORK0x00008; | |||
| 948 | else | |||
| 949 | p->p_flag &= ~P_FOLLOWFORK0x00008; | |||
| 950 | break; | |||
| 951 | ||||
| 952 | case PT_LWP_EVENTS24: | |||
| 953 | CTR3(KTR_PTRACE, "PT_LWP_EVENTS: pid %d %s -> %s", p->p_pid,(void)0 | |||
| 954 | p->p_flag2 & P2_LWP_EVENTS ? "enabled" : "disabled",(void)0 | |||
| 955 | data ? "enabled" : "disabled")(void)0; | |||
| 956 | if (data) | |||
| 957 | p->p_flag2 |= P2_LWP_EVENTS0x00000010; | |||
| 958 | else | |||
| 959 | p->p_flag2 &= ~P2_LWP_EVENTS0x00000010; | |||
| 960 | break; | |||
| 961 | ||||
| 962 | case PT_STEP9: | |||
| 963 | case PT_CONTINUE7: | |||
| 964 | case PT_TO_SCE20: | |||
| 965 | case PT_TO_SCX21: | |||
| 966 | case PT_SYSCALL22: | |||
| 967 | case PT_DETACH11: | |||
| 968 | /* Zero means do not send any signal */ | |||
| 969 | if (data < 0 || data > _SIG_MAXSIG128) { | |||
| 970 | error = EINVAL22; | |||
| 971 | break; | |||
| 972 | } | |||
| 973 | ||||
| 974 | switch (req) { | |||
| 975 | case PT_STEP9: | |||
| 976 | CTR2(KTR_PTRACE, "PT_STEP: tid %d (pid %d)",(void)0 | |||
| 977 | td2->td_tid, p->p_pid)(void)0; | |||
| 978 | error = ptrace_single_step(td2); | |||
| 979 | if (error) | |||
| 980 | goto out; | |||
| 981 | break; | |||
| 982 | case PT_CONTINUE7: | |||
| 983 | case PT_TO_SCE20: | |||
| 984 | case PT_TO_SCX21: | |||
| 985 | case PT_SYSCALL22: | |||
| 986 | if (addr != (void *)1) { | |||
| 987 | error = ptrace_set_pc(td2, | |||
| 988 | (u_long)(uintfptr_t)addr); | |||
| 989 | if (error) | |||
| 990 | goto out; | |||
| 991 | } | |||
| 992 | switch (req) { | |||
| 993 | case PT_TO_SCE20: | |||
| 994 | p->p_stops |= S_PT_SCE0x000010000; | |||
| 995 | CTR4(KTR_PTRACE,(void)0 | |||
| 996 | "PT_TO_SCE: pid %d, stops = %#x, PC = %#lx, sig = %d",(void)0 | |||
| 997 | p->p_pid, p->p_stops,(void)0 | |||
| 998 | (u_long)(uintfptr_t)addr, data)(void)0; | |||
| 999 | break; | |||
| 1000 | case PT_TO_SCX21: | |||
| 1001 | p->p_stops |= S_PT_SCX0x000020000; | |||
| 1002 | CTR4(KTR_PTRACE,(void)0 | |||
| 1003 | "PT_TO_SCX: pid %d, stops = %#x, PC = %#lx, sig = %d",(void)0 | |||
| 1004 | p->p_pid, p->p_stops,(void)0 | |||
| 1005 | (u_long)(uintfptr_t)addr, data)(void)0; | |||
| 1006 | break; | |||
| 1007 | case PT_SYSCALL22: | |||
| 1008 | p->p_stops |= S_PT_SCE0x000010000 | S_PT_SCX0x000020000; | |||
| 1009 | CTR4(KTR_PTRACE,(void)0 | |||
| 1010 | "PT_SYSCALL: pid %d, stops = %#x, PC = %#lx, sig = %d",(void)0 | |||
| 1011 | p->p_pid, p->p_stops,(void)0 | |||
| 1012 | (u_long)(uintfptr_t)addr, data)(void)0; | |||
| 1013 | break; | |||
| 1014 | case PT_CONTINUE7: | |||
| 1015 | CTR3(KTR_PTRACE,(void)0 | |||
| 1016 | "PT_CONTINUE: pid %d, PC = %#lx, sig = %d",(void)0 | |||
| 1017 | p->p_pid, (u_long)(uintfptr_t)addr, data)(void)0; | |||
| 1018 | break; | |||
| 1019 | } | |||
| 1020 | break; | |||
| 1021 | case PT_DETACH11: | |||
| 1022 | /* | |||
| 1023 | * Reset the process parent. | |||
| 1024 | * | |||
| 1025 | * NB: This clears P_TRACED before reparenting | |||
| 1026 | * a detached process back to its original | |||
| 1027 | * parent. Otherwise the debugee will be set | |||
| 1028 | * as an orphan of the debugger. | |||
| 1029 | */ | |||
| 1030 | p->p_flag &= ~(P_TRACED0x00800 | P_WAITED0x01000 | P_FOLLOWFORK0x00008); | |||
| 1031 | if (p->p_oppid != p->p_pptr->p_pid) { | |||
| 1032 | PROC_LOCK(p->p_pptr)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p->p_pptr)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&(((((&(p->p_pptr)->p_mtx) ))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep(& (((((&(p->p_pptr)->p_mtx)))))->mtx_lock, _tid, ( ((0))), ((((void *)0))), ((0))); else do { (void)0; do { if ( __builtin_expect((sdt_lockstat___adaptive__acquire->id), 0 )) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire->id, (uintptr_t) (((&(p->p_pptr)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0) ; } while (0); } while (0); | |||
| 1033 | sigqueue_take(p->p_ksi); | |||
| 1034 | PROC_UNLOCK(p->p_pptr)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p->p_pptr)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(p->p_pptr)->p_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); if (((((&(p->p_pptr)->p_mtx))))-> mtx_lock != _tid || !atomic_cmpset_long(&(((((&(p-> p_pptr)->p_mtx)))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep (&(((((&(p->p_pptr)->p_mtx)))))->mtx_lock, ( ((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1035 | ||||
| 1036 | pp = proc_realparent(p); | |||
| 1037 | proc_reparent(p, pp); | |||
| 1038 | if (pp == initproc) | |||
| 1039 | p->p_sigparent = SIGCHLD20; | |||
| 1040 | CTR3(KTR_PTRACE,(void)0 | |||
| 1041 | "PT_DETACH: pid %d reparented to pid %d, sig %d",(void)0 | |||
| 1042 | p->p_pid, pp->p_pid, data)(void)0; | |||
| 1043 | } else | |||
| 1044 | CTR2(KTR_PTRACE, "PT_DETACH: pid %d, sig %d",(void)0 | |||
| 1045 | p->p_pid, data)(void)0; | |||
| 1046 | p->p_oppid = 0; | |||
| 1047 | p->p_stops = 0; | |||
| 1048 | ||||
| 1049 | /* should we send SIGCHLD? */ | |||
| 1050 | /* childproc_continued(p); */ | |||
| 1051 | break; | |||
| 1052 | } | |||
| 1053 | ||||
| 1054 | sendsig: | |||
| 1055 | if (proctree_locked) { | |||
| 1056 | sx_xunlock(&proctree_lock)__sx_xunlock(((&proctree_lock)), (__curthread()), (((void *)0)), (0)); | |||
| 1057 | proctree_locked = 0; | |||
| 1058 | } | |||
| 1059 | p->p_xsig = data; | |||
| 1060 | p->p_xthread = NULL((void *)0); | |||
| 1061 | if ((p->p_flag & (P_STOPPED_SIG0x20000 | P_STOPPED_TRACE0x40000)) != 0) { | |||
| 1062 | /* deliver or queue signal */ | |||
| 1063 | td2->td_dbgflags &= ~TDB_XSIG0x00000002; | |||
| 1064 | td2->td_xsig = data; | |||
| 1065 | ||||
| 1066 | if (req == PT_DETACH11) { | |||
| 1067 | FOREACH_THREAD_IN_PROC(p, td3)for (((td3)) = (((&(p)->p_threads))->tqh_first); (( td3)); ((td3)) = ((((td3)))->td_plist.tqe_next)) | |||
| 1068 | td3->td_dbgflags &= ~TDB_SUSPEND0x00000001; | |||
| 1069 | } | |||
| 1070 | /* | |||
| 1071 | * unsuspend all threads, to not let a thread run, | |||
| 1072 | * you should use PT_SUSPEND to suspend it before | |||
| 1073 | * continuing process. | |||
| 1074 | */ | |||
| 1075 | PROC_SLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); spinlock_enter (); if ((((((&(p)->p_slock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&(((((&(p)->p_slock)))))-> mtx_lock, 0x00000004, (_tid)))) { if (((((&(p)->p_slock ))))->mtx_lock == _tid) ((((&(p)->p_slock))))->lock_object .lo_data++; else _mtx_lock_spin_cookie(&(((((&(p)-> p_slock)))))->mtx_lock, _tid, (((0))), ((((void *)0))), (( 0))); } else do { (void)0; do { if (__builtin_expect((sdt_lockstat___spin__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___spin__acquire-> id, (uintptr_t) (((&(p)->p_slock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1076 | p->p_flag &= ~(P_STOPPED_TRACE0x40000|P_STOPPED_SIG0x20000|P_WAITED0x01000); | |||
| 1077 | thread_unsuspend(p); | |||
| 1078 | PROC_SUNLOCK(p)do { if (((((((&(p)->p_slock)))))->lock_object.lo_data != 0)) ((((&(p)->p_slock))))->lock_object.lo_data-- ; else { do { (void)0; do { if (__builtin_expect((sdt_lockstat___spin__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___spin__release-> id, (uintptr_t) (((&(p)->p_slock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); atomic_store_rel_long(&(((((&(p)->p_slock)))))-> mtx_lock, 0x00000004); } spinlock_exit(); } while (0); | |||
| 1079 | if (req == PT_ATTACH10) | |||
| 1080 | kern_psignal(p, data); | |||
| 1081 | } else { | |||
| 1082 | if (data) | |||
| 1083 | kern_psignal(p, data); | |||
| 1084 | } | |||
| 1085 | break; | |||
| 1086 | ||||
| 1087 | case PT_WRITE_I4: | |||
| 1088 | case PT_WRITE_D5: | |||
| 1089 | td2->td_dbgflags |= TDB_USERWR0x00000004; | |||
| 1090 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1091 | error = 0; | |||
| 1092 | if (proc_writemem(td, p, (off_t)(uintptr_t)addr, &data, | |||
| 1093 | sizeof(int)) != sizeof(int)) | |||
| 1094 | error = ENOMEM12; | |||
| 1095 | else | |||
| 1096 | CTR3(KTR_PTRACE, "PT_WRITE: pid %d: %p <= %#x",(void)0 | |||
| 1097 | p->p_pid, addr, data)(void)0; | |||
| 1098 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1099 | break; | |||
| 1100 | ||||
| 1101 | case PT_READ_I1: | |||
| 1102 | case PT_READ_D2: | |||
| 1103 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1104 | error = tmp = 0; | |||
| 1105 | if (proc_readmem(td, p, (off_t)(uintptr_t)addr, &tmp, | |||
| 1106 | sizeof(int)) != sizeof(int)) | |||
| 1107 | error = ENOMEM12; | |||
| 1108 | else | |||
| 1109 | CTR3(KTR_PTRACE, "PT_READ: pid %d: %p >= %#x",(void)0 | |||
| 1110 | p->p_pid, addr, tmp)(void)0; | |||
| 1111 | td->td_retvaltd_uretoff.tdu_retval[0] = tmp; | |||
| 1112 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1113 | break; | |||
| 1114 | ||||
| 1115 | case PT_IO12: | |||
| 1116 | #ifdef COMPAT_FREEBSD321 | |||
| 1117 | if (wrap32) { | |||
| 1118 | piod32 = addr; | |||
| 1119 | iov.iov_base = (void *)(uintptr_t)piod32->piod_addr; | |||
| 1120 | iov.iov_len = piod32->piod_len; | |||
| 1121 | uio.uio_offset = (off_t)(uintptr_t)piod32->piod_offs; | |||
| 1122 | uio.uio_resid = piod32->piod_len; | |||
| 1123 | } else | |||
| 1124 | #endif | |||
| 1125 | { | |||
| 1126 | piod = addr; | |||
| 1127 | iov.iov_base = piod->piod_addr; | |||
| 1128 | iov.iov_len = piod->piod_len; | |||
| 1129 | uio.uio_offset = (off_t)(uintptr_t)piod->piod_offs; | |||
| 1130 | uio.uio_resid = piod->piod_len; | |||
| 1131 | } | |||
| 1132 | uio.uio_iov = &iov; | |||
| 1133 | uio.uio_iovcnt = 1; | |||
| 1134 | uio.uio_segflg = UIO_USERSPACE; | |||
| 1135 | uio.uio_td = td; | |||
| 1136 | #ifdef COMPAT_FREEBSD321 | |||
| 1137 | tmp = wrap32 ? piod32->piod_op : piod->piod_op; | |||
| 1138 | #else | |||
| 1139 | tmp = piod->piod_op; | |||
| 1140 | #endif | |||
| 1141 | switch (tmp) { | |||
| 1142 | case PIOD_READ_D1: | |||
| 1143 | case PIOD_READ_I3: | |||
| 1144 | CTR3(KTR_PTRACE, "PT_IO: pid %d: READ (%p, %#x)",(void)0 | |||
| 1145 | p->p_pid, (uintptr_t)uio.uio_offset, uio.uio_resid)(void)0; | |||
| 1146 | uio.uio_rw = UIO_READ; | |||
| 1147 | break; | |||
| 1148 | case PIOD_WRITE_D2: | |||
| 1149 | case PIOD_WRITE_I4: | |||
| 1150 | CTR3(KTR_PTRACE, "PT_IO: pid %d: WRITE (%p, %#x)",(void)0 | |||
| 1151 | p->p_pid, (uintptr_t)uio.uio_offset, uio.uio_resid)(void)0; | |||
| 1152 | td2->td_dbgflags |= TDB_USERWR0x00000004; | |||
| 1153 | uio.uio_rw = UIO_WRITE; | |||
| 1154 | break; | |||
| 1155 | default: | |||
| 1156 | error = EINVAL22; | |||
| 1157 | goto out; | |||
| 1158 | } | |||
| 1159 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1160 | error = proc_rwmem(p, &uio); | |||
| 1161 | #ifdef COMPAT_FREEBSD321 | |||
| 1162 | if (wrap32) | |||
| 1163 | piod32->piod_len -= uio.uio_resid; | |||
| 1164 | else | |||
| 1165 | #endif | |||
| 1166 | piod->piod_len -= uio.uio_resid; | |||
| 1167 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1168 | break; | |||
| 1169 | ||||
| 1170 | case PT_KILL8: | |||
| 1171 | CTR1(KTR_PTRACE, "PT_KILL: pid %d", p->p_pid)(void)0; | |||
| 1172 | data = SIGKILL9; | |||
| 1173 | goto sendsig; /* in PT_CONTINUE above */ | |||
| 1174 | ||||
| 1175 | case PT_SETREGS34: | |||
| 1176 | CTR2(KTR_PTRACE, "PT_SETREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1177 | p->p_pid)(void)0; | |||
| 1178 | td2->td_dbgflags |= TDB_USERWR0x00000004; | |||
| 1179 | error = PROC_WRITE(regs, td2, addr); | |||
| 1180 | break; | |||
| 1181 | ||||
| 1182 | case PT_GETREGS33: | |||
| 1183 | CTR2(KTR_PTRACE, "PT_GETREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1184 | p->p_pid)(void)0; | |||
| 1185 | error = PROC_READ(regs, td2, addr); | |||
| 1186 | break; | |||
| 1187 | ||||
| 1188 | case PT_SETFPREGS36: | |||
| 1189 | CTR2(KTR_PTRACE, "PT_SETFPREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1190 | p->p_pid)(void)0; | |||
| 1191 | td2->td_dbgflags |= TDB_USERWR0x00000004; | |||
| 1192 | error = PROC_WRITE(fpregs, td2, addr); | |||
| 1193 | break; | |||
| 1194 | ||||
| 1195 | case PT_GETFPREGS35: | |||
| 1196 | CTR2(KTR_PTRACE, "PT_GETFPREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1197 | p->p_pid)(void)0; | |||
| 1198 | error = PROC_READ(fpregs, td2, addr); | |||
| 1199 | break; | |||
| 1200 | ||||
| 1201 | case PT_SETDBREGS38: | |||
| 1202 | CTR2(KTR_PTRACE, "PT_SETDBREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1203 | p->p_pid)(void)0; | |||
| 1204 | td2->td_dbgflags |= TDB_USERWR0x00000004; | |||
| 1205 | error = PROC_WRITE(dbregs, td2, addr); | |||
| 1206 | break; | |||
| 1207 | ||||
| 1208 | case PT_GETDBREGS37: | |||
| 1209 | CTR2(KTR_PTRACE, "PT_GETDBREGS: tid %d (pid %d)", td2->td_tid,(void)0 | |||
| 1210 | p->p_pid)(void)0; | |||
| 1211 | error = PROC_READ(dbregs, td2, addr); | |||
| 1212 | break; | |||
| 1213 | ||||
| 1214 | case PT_LWPINFO13: | |||
| 1215 | if (data <= 0 || | |||
| 1216 | #ifdef COMPAT_FREEBSD321 | |||
| 1217 | (!wrap32 && data > sizeof(*pl)) || | |||
| 1218 | (wrap32 && data > sizeof(*pl32))) { | |||
| 1219 | #else | |||
| 1220 | data > sizeof(*pl)) { | |||
| 1221 | #endif | |||
| 1222 | error = EINVAL22; | |||
| 1223 | break; | |||
| 1224 | } | |||
| 1225 | #ifdef COMPAT_FREEBSD321 | |||
| 1226 | if (wrap32) { | |||
| 1227 | pl = &plr; | |||
| 1228 | pl32 = addr; | |||
| 1229 | } else | |||
| 1230 | #endif | |||
| 1231 | pl = addr; | |||
| 1232 | pl->pl_lwpid = td2->td_tid; | |||
| 1233 | pl->pl_event = PL_EVENT_NONE0; | |||
| 1234 | pl->pl_flags = 0; | |||
| 1235 | if (td2->td_dbgflags & TDB_XSIG0x00000002) { | |||
| 1236 | pl->pl_event = PL_EVENT_SIGNAL1; | |||
| 1237 | if (td2->td_dbgksi.ksi_signoksi_info.si_signo != 0 && | |||
| 1238 | #ifdef COMPAT_FREEBSD321 | |||
| 1239 | ((!wrap32 && data >= offsetof(struct ptrace_lwpinfo,__builtin_offsetof(struct ptrace_lwpinfo, pl_siginfo) | |||
| 1240 | pl_siginfo)__builtin_offsetof(struct ptrace_lwpinfo, pl_siginfo) + sizeof(pl->pl_siginfo)) || | |||
| 1241 | (wrap32 && data >= offsetof(struct ptrace_lwpinfo32,__builtin_offsetof(struct ptrace_lwpinfo32, pl_siginfo) | |||
| 1242 | pl_siginfo)__builtin_offsetof(struct ptrace_lwpinfo32, pl_siginfo) + sizeof(struct siginfo32))) | |||
| 1243 | #else | |||
| 1244 | data >= offsetof(struct ptrace_lwpinfo, pl_siginfo)__builtin_offsetof(struct ptrace_lwpinfo, pl_siginfo) | |||
| 1245 | + sizeof(pl->pl_siginfo) | |||
| 1246 | #endif | |||
| 1247 | ){ | |||
| 1248 | pl->pl_flags |= PL_FLAG_SI0x20; | |||
| 1249 | pl->pl_siginfo = td2->td_dbgksi.ksi_info; | |||
| 1250 | } | |||
| 1251 | } | |||
| 1252 | if ((pl->pl_flags & PL_FLAG_SI0x20) == 0) | |||
| 1253 | bzero(&pl->pl_siginfo, sizeof(pl->pl_siginfo)); | |||
| 1254 | if (td2->td_dbgflags & TDB_SCE0x00000008) | |||
| 1255 | pl->pl_flags |= PL_FLAG_SCE0x04; | |||
| 1256 | else if (td2->td_dbgflags & TDB_SCX0x00000010) | |||
| 1257 | pl->pl_flags |= PL_FLAG_SCX0x08; | |||
| 1258 | if (td2->td_dbgflags & TDB_EXEC0x00000020) | |||
| 1259 | pl->pl_flags |= PL_FLAG_EXEC0x10; | |||
| 1260 | if (td2->td_dbgflags & TDB_FORK0x00000040) { | |||
| 1261 | pl->pl_flags |= PL_FLAG_FORKED0x40; | |||
| 1262 | pl->pl_child_pid = td2->td_dbg_forked; | |||
| 1263 | } | |||
| 1264 | if (td2->td_dbgflags & TDB_CHILD0x00000100) | |||
| 1265 | pl->pl_flags |= PL_FLAG_CHILD0x80; | |||
| 1266 | if (td2->td_dbgflags & TDB_BORN0x00000200) | |||
| 1267 | pl->pl_flags |= PL_FLAG_BORN0x100; | |||
| 1268 | if (td2->td_dbgflags & TDB_EXIT0x00000400) | |||
| 1269 | pl->pl_flags |= PL_FLAG_EXITED0x200; | |||
| 1270 | pl->pl_sigmask = td2->td_sigmask; | |||
| 1271 | pl->pl_siglist = td2->td_siglisttd_sigqueue.sq_signals; | |||
| 1272 | strcpy(pl->pl_tdname, td2->td_name); | |||
| 1273 | if ((td2->td_dbgflags & (TDB_SCE0x00000008 | TDB_SCX0x00000010)) != 0) { | |||
| 1274 | pl->pl_syscall_code = td2->td_dbg_sc_code; | |||
| 1275 | pl->pl_syscall_narg = td2->td_dbg_sc_narg; | |||
| 1276 | } else { | |||
| 1277 | pl->pl_syscall_code = 0; | |||
| 1278 | pl->pl_syscall_narg = 0; | |||
| 1279 | } | |||
| 1280 | #ifdef COMPAT_FREEBSD321 | |||
| 1281 | if (wrap32) | |||
| 1282 | ptrace_lwpinfo_to32(pl, pl32); | |||
| 1283 | #endif | |||
| 1284 | CTR6(KTR_PTRACE,(void)0 | |||
| 1285 | "PT_LWPINFO: tid %d (pid %d) event %d flags %#x child pid %d syscall %d",(void)0 | |||
| 1286 | td2->td_tid, p->p_pid, pl->pl_event, pl->pl_flags,(void)0 | |||
| 1287 | pl->pl_child_pid, pl->pl_syscall_code)(void)0; | |||
| 1288 | break; | |||
| 1289 | ||||
| 1290 | case PT_GETNUMLWPS14: | |||
| 1291 | CTR2(KTR_PTRACE, "PT_GETNUMLWPS: pid %d: %d threads", p->p_pid,(void)0 | |||
| 1292 | p->p_numthreads)(void)0; | |||
| 1293 | td->td_retvaltd_uretoff.tdu_retval[0] = p->p_numthreads; | |||
| 1294 | break; | |||
| 1295 | ||||
| 1296 | case PT_GETLWPLIST15: | |||
| 1297 | CTR3(KTR_PTRACE, "PT_GETLWPLIST: pid %d: data %d, actual %d",(void)0 | |||
| 1298 | p->p_pid, data, p->p_numthreads)(void)0; | |||
| 1299 | if (data <= 0) { | |||
| 1300 | error = EINVAL22; | |||
| 1301 | break; | |||
| 1302 | } | |||
| 1303 | num = imin(p->p_numthreads, data); | |||
| 1304 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1305 | buf = malloc(num * sizeof(lwpid_t), M_TEMP, M_WAITOK0x0002); | |||
| 1306 | tmp = 0; | |||
| 1307 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1308 | FOREACH_THREAD_IN_PROC(p, td2)for (((td2)) = (((&(p)->p_threads))->tqh_first); (( td2)); ((td2)) = ((((td2)))->td_plist.tqe_next)) { | |||
| 1309 | if (tmp >= num) | |||
| 1310 | break; | |||
| 1311 | buf[tmp++] = td2->td_tid; | |||
| 1312 | } | |||
| 1313 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1314 | error = copyout(buf, addr, tmp * sizeof(lwpid_t)); | |||
| 1315 | free(buf, M_TEMP); | |||
| 1316 | if (!error) | |||
| 1317 | td->td_retvaltd_uretoff.tdu_retval[0] = tmp; | |||
| 1318 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1319 | break; | |||
| 1320 | ||||
| 1321 | case PT_VM_TIMESTAMP40: | |||
| 1322 | CTR2(KTR_PTRACE, "PT_VM_TIMESTAMP: pid %d: timestamp %d",(void)0 | |||
| 1323 | p->p_pid, p->p_vmspace->vm_map.timestamp)(void)0; | |||
| 1324 | td->td_retvaltd_uretoff.tdu_retval[0] = p->p_vmspace->vm_map.timestamp; | |||
| 1325 | break; | |||
| 1326 | ||||
| 1327 | case PT_VM_ENTRY41: | |||
| 1328 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1329 | #ifdef COMPAT_FREEBSD321 | |||
| 1330 | if (wrap32) | |||
| 1331 | error = ptrace_vm_entry32(td, p, addr); | |||
| 1332 | else | |||
| 1333 | #endif | |||
| 1334 | error = ptrace_vm_entry(td, p, addr); | |||
| 1335 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1336 | break; | |||
| 1337 | ||||
| 1338 | default: | |||
| 1339 | #ifdef __HAVE_PTRACE_MACHDEP | |||
| 1340 | if (req >= PT_FIRSTMACH64) { | |||
| 1341 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1342 | error = cpu_ptrace(td2, req, addr, data); | |||
| 1343 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1344 | } else | |||
| 1345 | #endif | |||
| 1346 | /* Unknown request. */ | |||
| 1347 | error = EINVAL22; | |||
| 1348 | break; | |||
| 1349 | } | |||
| 1350 | ||||
| 1351 | out: | |||
| 1352 | /* Drop our hold on this process now that the request has completed. */ | |||
| 1353 | _PRELE(p)do { (void)0; do { do { } while (0); } while (0); (--(p)-> p_lock); if (((p)->p_flag & 0x02000) && (p)-> p_lock == 0) wakeup(&(p)->p_lock); } while (0); | |||
| 1354 | fail: | |||
| 1355 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1356 | if (proctree_locked) | |||
| 1357 | sx_xunlock(&proctree_lock)__sx_xunlock(((&proctree_lock)), (__curthread()), (((void *)0)), (0)); | |||
| 1358 | return (error); | |||
| 1359 | } | |||
| 1360 | #undef PROC_READ | |||
| 1361 | #undef PROC_WRITE | |||
| 1362 | ||||
| 1363 | /* | |||
| 1364 | * Stop a process because of a debugging event; | |||
| 1365 | * stay stopped until p->p_step is cleared | |||
| 1366 | * (cleared by PIOCCONT in procfs). | |||
| 1367 | */ | |||
| 1368 | void | |||
| 1369 | stopevent(struct proc *p, unsigned int event, unsigned int val) | |||
| 1370 | { | |||
| 1371 | ||||
| 1372 | PROC_LOCK_ASSERT(p, MA_OWNED)(void)0; | |||
| 1373 | p->p_step = 1; | |||
| 1374 | CTR3(KTR_PTRACE, "stopevent: pid %d event %u val %u", p->p_pid, event,(void)0 | |||
| 1375 | val)(void)0; | |||
| 1376 | do { | |||
| 1377 | if (event != S_EXIT0x00000020) | |||
| 1378 | p->p_xsig = val; | |||
| 1379 | p->p_xthread = NULL((void *)0); | |||
| 1380 | p->p_stype = event; /* Which event caused the stop? */ | |||
| 1381 | wakeup(&p->p_stype); /* Wake up any PIOCWAIT'ing procs */ | |||
| 1382 | msleep(&p->p_step, &p->p_mtx, PWAIT, "stopevent", 0)_sleep((&p->p_step), &(&p->p_mtx)->lock_object , (((80) + 28)), ("stopevent"), tick_sbt * (0), 0, 0x0100); | |||
| 1383 | } while (p->p_step); | |||
| 1384 | } |