| File: | kern/kern_descrip.c |
| Warning: | line 1369, column 11 Copies out a struct with uncleared padding (>= 4 bytes) |
| 1 | /*- | |||
| 2 | * Copyright (c) 1982, 1986, 1989, 1991, 1993 | |||
| 3 | * The Regents of the University of California. All rights reserved. | |||
| 4 | * (c) UNIX System Laboratories, Inc. | |||
| 5 | * All or some portions of this file are derived from material licensed | |||
| 6 | * to the University of California by American Telephone and Telegraph | |||
| 7 | * Co. or Unix System Laboratories, Inc. and are reproduced herein with | |||
| 8 | * the permission of UNIX System Laboratories, Inc. | |||
| 9 | * | |||
| 10 | * Redistribution and use in source and binary forms, with or without | |||
| 11 | * modification, are permitted provided that the following conditions | |||
| 12 | * are met: | |||
| 13 | * 1. Redistributions of source code must retain the above copyright | |||
| 14 | * notice, this list of conditions and the following disclaimer. | |||
| 15 | * 2. Redistributions in binary form must reproduce the above copyright | |||
| 16 | * notice, this list of conditions and the following disclaimer in the | |||
| 17 | * documentation and/or other materials provided with the distribution. | |||
| 18 | * 4. Neither the name of the University nor the names of its contributors | |||
| 19 | * may be used to endorse or promote products derived from this software | |||
| 20 | * without specific prior written permission. | |||
| 21 | * | |||
| 22 | * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND | |||
| 23 | * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | |||
| 24 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | |||
| 25 | * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE | |||
| 26 | * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL | |||
| 27 | * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS | |||
| 28 | * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) | |||
| 29 | * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT | |||
| 30 | * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY | |||
| 31 | * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF | |||
| 32 | * SUCH DAMAGE. | |||
| 33 | * | |||
| 34 | * @(#)kern_descrip.c 8.6 (Berkeley) 4/19/94 | |||
| 35 | */ | |||
| 36 | ||||
| 37 | #include <sys/cdefs.h> | |||
| 38 | __FBSDID("$FreeBSD: releng/11.0/sys/kern/kern_descrip.c 300852 2016-05-27 17:00:15Z mjg $")__asm__(".ident\t\"" "$FreeBSD: releng/11.0/sys/kern/kern_descrip.c 300852 2016-05-27 17:00:15Z mjg $" "\""); | |||
| 39 | ||||
| 40 | #include "opt_capsicum.h" | |||
| 41 | #include "opt_compat.h" | |||
| 42 | #include "opt_ddb.h" | |||
| 43 | #include "opt_ktrace.h" | |||
| 44 | ||||
| 45 | #include <sys/param.h> | |||
| 46 | #include <sys/systm.h> | |||
| 47 | ||||
| 48 | #include <sys/capsicum.h> | |||
| 49 | #include <sys/conf.h> | |||
| 50 | #include <sys/fcntl.h> | |||
| 51 | #include <sys/file.h> | |||
| 52 | #include <sys/filedesc.h> | |||
| 53 | #include <sys/filio.h> | |||
| 54 | #include <sys/jail.h> | |||
| 55 | #include <sys/kernel.h> | |||
| 56 | #include <sys/limits.h> | |||
| 57 | #include <sys/lock.h> | |||
| 58 | #include <sys/malloc.h> | |||
| 59 | #include <sys/mount.h> | |||
| 60 | #include <sys/mutex.h> | |||
| 61 | #include <sys/namei.h> | |||
| 62 | #include <sys/selinfo.h> | |||
| 63 | #include <sys/priv.h> | |||
| 64 | #include <sys/proc.h> | |||
| 65 | #include <sys/protosw.h> | |||
| 66 | #include <sys/racct.h> | |||
| 67 | #include <sys/resourcevar.h> | |||
| 68 | #include <sys/sbuf.h> | |||
| 69 | #include <sys/signalvar.h> | |||
| 70 | #include <sys/socketvar.h> | |||
| 71 | #include <sys/kdb.h> | |||
| 72 | #include <sys/stat.h> | |||
| 73 | #include <sys/sx.h> | |||
| 74 | #include <sys/syscallsubr.h> | |||
| 75 | #include <sys/sysctl.h> | |||
| 76 | #include <sys/sysproto.h> | |||
| 77 | #include <sys/unistd.h> | |||
| 78 | #include <sys/user.h> | |||
| 79 | #include <sys/vnode.h> | |||
| 80 | #ifdef KTRACE1 | |||
| 81 | #include <sys/ktrace.h> | |||
| 82 | #endif | |||
| 83 | ||||
| 84 | #include <net/vnet.h> | |||
| 85 | ||||
| 86 | #include <security/audit/audit.h> | |||
| 87 | ||||
| 88 | #include <vm/uma.h> | |||
| 89 | #include <vm/vm.h> | |||
| 90 | ||||
| 91 | #include <ddb/ddb.h> | |||
| 92 | ||||
| 93 | static MALLOC_DEFINE(M_FILEDESC, "filedesc", "Open file descriptor table")struct malloc_type M_FILEDESC[1] = { { ((void *)0), 877983977 , "filedesc", ((void *)0) } }; static struct sysinit M_FILEDESC_init_sys_init = { SI_SUB_KMEM, SI_ORDER_THIRD, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_init, ((void *)(M_FILEDESC)) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_M_FILEDESC_init_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(M_FILEDESC_init_sys_init); static struct sysinit M_FILEDESC_uninit_sys_uninit = { SI_SUB_KMEM, SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_uninit, ((void *)(M_FILEDESC)) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_M_FILEDESC_uninit_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(M_FILEDESC_uninit_sys_uninit); | |||
| 94 | static MALLOC_DEFINE(M_FILEDESC_TO_LEADER, "filedesc_to_leader",struct malloc_type M_FILEDESC_TO_LEADER[1] = { { ((void *)0), 877983977, "filedesc_to_leader", ((void *)0) } }; static struct sysinit M_FILEDESC_TO_LEADER_init_sys_init = { SI_SUB_KMEM, SI_ORDER_THIRD , (sysinit_cfunc_t)(sysinit_nfunc_t)malloc_init, ((void *)(M_FILEDESC_TO_LEADER )) }; __asm__(".globl " "__start_set_sysinit_set"); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_M_FILEDESC_TO_LEADER_init_sys_init __attribute__((__section__("set_" "sysinit_set"))) __attribute__ ((__used__)) = &(M_FILEDESC_TO_LEADER_init_sys_init); static struct sysinit M_FILEDESC_TO_LEADER_uninit_sys_uninit = { SI_SUB_KMEM , SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t)malloc_uninit , ((void *)(M_FILEDESC_TO_LEADER)) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_M_FILEDESC_TO_LEADER_uninit_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(M_FILEDESC_TO_LEADER_uninit_sys_uninit) | |||
| 95 | "file desc to leader structures")struct malloc_type M_FILEDESC_TO_LEADER[1] = { { ((void *)0), 877983977, "filedesc_to_leader", ((void *)0) } }; static struct sysinit M_FILEDESC_TO_LEADER_init_sys_init = { SI_SUB_KMEM, SI_ORDER_THIRD , (sysinit_cfunc_t)(sysinit_nfunc_t)malloc_init, ((void *)(M_FILEDESC_TO_LEADER )) }; __asm__(".globl " "__start_set_sysinit_set"); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_M_FILEDESC_TO_LEADER_init_sys_init __attribute__((__section__("set_" "sysinit_set"))) __attribute__ ((__used__)) = &(M_FILEDESC_TO_LEADER_init_sys_init); static struct sysinit M_FILEDESC_TO_LEADER_uninit_sys_uninit = { SI_SUB_KMEM , SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t)malloc_uninit , ((void *)(M_FILEDESC_TO_LEADER)) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_M_FILEDESC_TO_LEADER_uninit_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(M_FILEDESC_TO_LEADER_uninit_sys_uninit); | |||
| 96 | static MALLOC_DEFINE(M_SIGIO, "sigio", "sigio structures")struct malloc_type M_SIGIO[1] = { { ((void *)0), 877983977, "sigio" , ((void *)0) } }; static struct sysinit M_SIGIO_init_sys_init = { SI_SUB_KMEM, SI_ORDER_THIRD, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_init, ((void *)(M_SIGIO)) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_M_SIGIO_init_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(M_SIGIO_init_sys_init); static struct sysinit M_SIGIO_uninit_sys_uninit = { SI_SUB_KMEM, SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_uninit, ((void *)(M_SIGIO)) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_M_SIGIO_uninit_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(M_SIGIO_uninit_sys_uninit); | |||
| 97 | MALLOC_DEFINE(M_FILECAPS, "filecaps", "descriptor capabilities")struct malloc_type M_FILECAPS[1] = { { ((void *)0), 877983977 , "filecaps", ((void *)0) } }; static struct sysinit M_FILECAPS_init_sys_init = { SI_SUB_KMEM, SI_ORDER_THIRD, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_init, ((void *)(M_FILECAPS)) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_M_FILECAPS_init_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(M_FILECAPS_init_sys_init); static struct sysinit M_FILECAPS_uninit_sys_uninit = { SI_SUB_KMEM, SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t )malloc_uninit, ((void *)(M_FILECAPS)) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_M_FILECAPS_uninit_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(M_FILECAPS_uninit_sys_uninit); | |||
| 98 | ||||
| 99 | MALLOC_DECLARE(M_FADVISE)extern struct malloc_type M_FADVISE[1]; | |||
| 100 | ||||
| 101 | static uma_zone_t file_zone; | |||
| 102 | static uma_zone_t filedesc0_zone; | |||
| 103 | ||||
| 104 | static int closefp(struct filedesc *fdp, int fd, struct file *fp, | |||
| 105 | struct thread *td, int holdleaders); | |||
| 106 | static int fd_first_free(struct filedesc *fdp, int low, int size); | |||
| 107 | static int fd_last_used(struct filedesc *fdp, int size); | |||
| 108 | static void fdgrowtable(struct filedesc *fdp, int nfd); | |||
| 109 | static void fdgrowtable_exp(struct filedesc *fdp, int nfd); | |||
| 110 | static void fdunused(struct filedesc *fdp, int fd); | |||
| 111 | static void fdused(struct filedesc *fdp, int fd); | |||
| 112 | static int getmaxfd(struct thread *td); | |||
| 113 | ||||
| 114 | /* | |||
| 115 | * Each process has: | |||
| 116 | * | |||
| 117 | * - An array of open file descriptors (fd_ofiles) | |||
| 118 | * - An array of file flags (fd_ofileflags) | |||
| 119 | * - A bitmap recording which descriptors are in use (fd_map) | |||
| 120 | * | |||
| 121 | * A process starts out with NDFILE descriptors. The value of NDFILE has | |||
| 122 | * been selected based the historical limit of 20 open files, and an | |||
| 123 | * assumption that the majority of processes, especially short-lived | |||
| 124 | * processes like shells, will never need more. | |||
| 125 | * | |||
| 126 | * If this initial allocation is exhausted, a larger descriptor table and | |||
| 127 | * map are allocated dynamically, and the pointers in the process's struct | |||
| 128 | * filedesc are updated to point to those. This is repeated every time | |||
| 129 | * the process runs out of file descriptors (provided it hasn't hit its | |||
| 130 | * resource limit). | |||
| 131 | * | |||
| 132 | * Since threads may hold references to individual descriptor table | |||
| 133 | * entries, the tables are never freed. Instead, they are placed on a | |||
| 134 | * linked list and freed only when the struct filedesc is released. | |||
| 135 | */ | |||
| 136 | #define NDFILE20 20 | |||
| 137 | #define NDSLOTSIZEsizeof(u_long) sizeof(NDSLOTTYPEu_long) | |||
| 138 | #define NDENTRIES(sizeof(u_long) * 8) (NDSLOTSIZEsizeof(u_long) * __CHAR_BIT8) | |||
| 139 | #define NDSLOT(x)((x) / (sizeof(u_long) * 8)) ((x) / NDENTRIES(sizeof(u_long) * 8)) | |||
| 140 | #define NDBIT(x)((u_long)1 << ((x) % (sizeof(u_long) * 8))) ((NDSLOTTYPEu_long)1 << ((x) % NDENTRIES(sizeof(u_long) * 8))) | |||
| 141 | #define NDSLOTS(x)(((x) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8)) (((x) + NDENTRIES(sizeof(u_long) * 8) - 1) / NDENTRIES(sizeof(u_long) * 8)) | |||
| 142 | ||||
| 143 | /* | |||
| 144 | * SLIST entry used to keep track of ofiles which must be reclaimed when | |||
| 145 | * the process exits. | |||
| 146 | */ | |||
| 147 | struct freetable { | |||
| 148 | struct fdescenttbl *ft_table; | |||
| 149 | SLIST_ENTRY(freetable)struct { struct freetable *sle_next; } ft_next; | |||
| 150 | }; | |||
| 151 | ||||
| 152 | /* | |||
| 153 | * Initial allocation: a filedesc structure + the head of SLIST used to | |||
| 154 | * keep track of old ofiles + enough space for NDFILE descriptors. | |||
| 155 | */ | |||
| 156 | ||||
| 157 | struct fdescenttbl0 { | |||
| 158 | int fdt_nfiles; | |||
| 159 | struct filedescent fdt_ofiles[NDFILE20]; | |||
| 160 | }; | |||
| 161 | ||||
| 162 | struct filedesc0 { | |||
| 163 | struct filedesc fd_fd; | |||
| 164 | SLIST_HEAD(, freetable)struct { struct freetable *slh_first; } fd_free; | |||
| 165 | struct fdescenttbl0 fd_dfiles; | |||
| 166 | NDSLOTTYPEu_long fd_dmap[NDSLOTS(NDFILE)(((20) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8))]; | |||
| 167 | }; | |||
| 168 | ||||
| 169 | /* | |||
| 170 | * Descriptor management. | |||
| 171 | */ | |||
| 172 | volatile int openfiles; /* actual number of open files */ | |||
| 173 | struct mtx sigio_lock; /* mtx to protect pointers to sigio */ | |||
| 174 | void (*mq_fdclose)(struct thread *td, int fd, struct file *fp); | |||
| 175 | ||||
| 176 | /* | |||
| 177 | * If low >= size, just return low. Otherwise find the first zero bit in the | |||
| 178 | * given bitmap, starting at low and not exceeding size - 1. Return size if | |||
| 179 | * not found. | |||
| 180 | */ | |||
| 181 | static int | |||
| 182 | fd_first_free(struct filedesc *fdp, int low, int size) | |||
| 183 | { | |||
| 184 | NDSLOTTYPEu_long *map = fdp->fd_map; | |||
| 185 | NDSLOTTYPEu_long mask; | |||
| 186 | int off, maxoff; | |||
| 187 | ||||
| 188 | if (low >= size) | |||
| 189 | return (low); | |||
| 190 | ||||
| 191 | off = NDSLOT(low)((low) / (sizeof(u_long) * 8)); | |||
| 192 | if (low % NDENTRIES(sizeof(u_long) * 8)) { | |||
| 193 | mask = ~(~(NDSLOTTYPEu_long)0 >> (NDENTRIES(sizeof(u_long) * 8) - (low % NDENTRIES(sizeof(u_long) * 8)))); | |||
| 194 | if ((mask &= ~map[off]) != 0UL) | |||
| 195 | return (off * NDENTRIES(sizeof(u_long) * 8) + ffsl(mask) - 1); | |||
| 196 | ++off; | |||
| 197 | } | |||
| 198 | for (maxoff = NDSLOTS(size)(((size) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8)); off < maxoff; ++off) | |||
| 199 | if (map[off] != ~0UL) | |||
| 200 | return (off * NDENTRIES(sizeof(u_long) * 8) + ffsl(~map[off]) - 1); | |||
| 201 | return (size); | |||
| 202 | } | |||
| 203 | ||||
| 204 | /* | |||
| 205 | * Find the highest non-zero bit in the given bitmap, starting at 0 and | |||
| 206 | * not exceeding size - 1. Return -1 if not found. | |||
| 207 | */ | |||
| 208 | static int | |||
| 209 | fd_last_used(struct filedesc *fdp, int size) | |||
| 210 | { | |||
| 211 | NDSLOTTYPEu_long *map = fdp->fd_map; | |||
| 212 | NDSLOTTYPEu_long mask; | |||
| 213 | int off, minoff; | |||
| 214 | ||||
| 215 | off = NDSLOT(size)((size) / (sizeof(u_long) * 8)); | |||
| 216 | if (size % NDENTRIES(sizeof(u_long) * 8)) { | |||
| 217 | mask = ~(~(NDSLOTTYPEu_long)0 << (size % NDENTRIES(sizeof(u_long) * 8))); | |||
| 218 | if ((mask &= map[off]) != 0) | |||
| 219 | return (off * NDENTRIES(sizeof(u_long) * 8) + flsl(mask) - 1); | |||
| 220 | --off; | |||
| 221 | } | |||
| 222 | for (minoff = NDSLOT(0)((0) / (sizeof(u_long) * 8)); off >= minoff; --off) | |||
| 223 | if (map[off] != 0) | |||
| 224 | return (off * NDENTRIES(sizeof(u_long) * 8) + flsl(map[off]) - 1); | |||
| 225 | return (-1); | |||
| 226 | } | |||
| 227 | ||||
| 228 | static int | |||
| 229 | fdisused(struct filedesc *fdp, int fd) | |||
| 230 | { | |||
| 231 | ||||
| 232 | KASSERT(fd >= 0 && fd < fdp->fd_nfiles,do { } while (0) | |||
| 233 | ("file descriptor %d out of range (0, %d)", fd, fdp->fd_nfiles))do { } while (0); | |||
| 234 | ||||
| 235 | return ((fdp->fd_map[NDSLOT(fd)((fd) / (sizeof(u_long) * 8))] & NDBIT(fd)((u_long)1 << ((fd) % (sizeof(u_long) * 8)))) != 0); | |||
| 236 | } | |||
| 237 | ||||
| 238 | /* | |||
| 239 | * Mark a file descriptor as used. | |||
| 240 | */ | |||
| 241 | static void | |||
| 242 | fdused_init(struct filedesc *fdp, int fd) | |||
| 243 | { | |||
| 244 | ||||
| 245 | KASSERT(!fdisused(fdp, fd), ("fd=%d is already used", fd))do { } while (0); | |||
| 246 | ||||
| 247 | fdp->fd_map[NDSLOT(fd)((fd) / (sizeof(u_long) * 8))] |= NDBIT(fd)((u_long)1 << ((fd) % (sizeof(u_long) * 8))); | |||
| 248 | } | |||
| 249 | ||||
| 250 | static void | |||
| 251 | fdused(struct filedesc *fdp, int fd) | |||
| 252 | { | |||
| 253 | ||||
| 254 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 255 | ||||
| 256 | fdused_init(fdp, fd); | |||
| 257 | if (fd > fdp->fd_lastfile) | |||
| 258 | fdp->fd_lastfile = fd; | |||
| 259 | if (fd == fdp->fd_freefile) | |||
| 260 | fdp->fd_freefile = fd_first_free(fdp, fd, fdp->fd_nfilesfd_files->fdt_nfiles); | |||
| 261 | } | |||
| 262 | ||||
| 263 | /* | |||
| 264 | * Mark a file descriptor as unused. | |||
| 265 | */ | |||
| 266 | static void | |||
| 267 | fdunused(struct filedesc *fdp, int fd) | |||
| 268 | { | |||
| 269 | ||||
| 270 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 271 | ||||
| 272 | KASSERT(fdisused(fdp, fd), ("fd=%d is already unused", fd))do { } while (0); | |||
| 273 | KASSERT(fdp->fd_ofiles[fd].fde_file == NULL,do { } while (0) | |||
| 274 | ("fd=%d is still in use", fd))do { } while (0); | |||
| 275 | ||||
| 276 | fdp->fd_map[NDSLOT(fd)((fd) / (sizeof(u_long) * 8))] &= ~NDBIT(fd)((u_long)1 << ((fd) % (sizeof(u_long) * 8))); | |||
| 277 | if (fd < fdp->fd_freefile) | |||
| 278 | fdp->fd_freefile = fd; | |||
| 279 | if (fd == fdp->fd_lastfile) | |||
| 280 | fdp->fd_lastfile = fd_last_used(fdp, fd); | |||
| 281 | } | |||
| 282 | ||||
| 283 | /* | |||
| 284 | * Free a file descriptor. | |||
| 285 | * | |||
| 286 | * Avoid some work if fdp is about to be destroyed. | |||
| 287 | */ | |||
| 288 | static inline void | |||
| 289 | fdefree_last(struct filedescent *fde) | |||
| 290 | { | |||
| 291 | ||||
| 292 | filecaps_free(&fde->fde_caps); | |||
| 293 | } | |||
| 294 | ||||
| 295 | static inline void | |||
| 296 | fdfree(struct filedesc *fdp, int fd) | |||
| 297 | { | |||
| 298 | struct filedescent *fde; | |||
| 299 | ||||
| 300 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[fd]; | |||
| 301 | #ifdef CAPABILITIES1 | |||
| 302 | seq_write_begin(&fde->fde_seq); | |||
| 303 | #endif | |||
| 304 | fdefree_last(fde); | |||
| 305 | fde->fde_file = NULL((void *)0); | |||
| 306 | fdunused(fdp, fd); | |||
| 307 | #ifdef CAPABILITIES1 | |||
| 308 | seq_write_end(&fde->fde_seq); | |||
| 309 | #endif | |||
| 310 | } | |||
| 311 | ||||
| 312 | void | |||
| 313 | pwd_ensure_dirs(void) | |||
| 314 | { | |||
| 315 | struct filedesc *fdp; | |||
| 316 | ||||
| 317 | fdp = curproc((__curthread())->td_proc)->p_fd; | |||
| 318 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 319 | if (fdp->fd_cdir == NULL((void *)0)) { | |||
| 320 | fdp->fd_cdir = rootvnode; | |||
| 321 | VREF(rootvnode)vref(rootvnode); | |||
| 322 | } | |||
| 323 | if (fdp->fd_rdir == NULL((void *)0)) { | |||
| 324 | fdp->fd_rdir = rootvnode; | |||
| 325 | VREF(rootvnode)vref(rootvnode); | |||
| 326 | } | |||
| 327 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 328 | } | |||
| 329 | ||||
| 330 | /* | |||
| 331 | * System calls on descriptors. | |||
| 332 | */ | |||
| 333 | #ifndef _SYS_SYSPROTO_H_ | |||
| 334 | struct getdtablesize_args { | |||
| 335 | int dummy; | |||
| 336 | }; | |||
| 337 | #endif | |||
| 338 | /* ARGSUSED */ | |||
| 339 | int | |||
| 340 | sys_getdtablesize(struct thread *td, struct getdtablesize_args *uap) | |||
| 341 | { | |||
| 342 | #ifdef RACCT1 | |||
| 343 | uint64_t lim; | |||
| 344 | #endif | |||
| 345 | ||||
| 346 | td->td_retvaltd_uretoff.tdu_retval[0] = | |||
| 347 | min((int)lim_cur(td, RLIMIT_NOFILE8), maxfilesperproc); | |||
| 348 | #ifdef RACCT1 | |||
| 349 | PROC_LOCK(td->td_proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(td->td_proc)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&(((((&(td->td_proc)->p_mtx )))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep(& (((((&(td->td_proc)->p_mtx)))))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire->id , (uintptr_t) (((&(td->td_proc)->p_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); } while (0); | |||
| 350 | lim = racct_get_limit(td->td_proc, RACCT_NOFILE7); | |||
| 351 | PROC_UNLOCK(td->td_proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (td->td_proc)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(td->td_proc)->p_mtx))), ( uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(td->td_proc)->p_mtx ))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((& (td->td_proc)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(td->td_proc)->p_mtx )))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 352 | if (lim < td->td_retvaltd_uretoff.tdu_retval[0]) | |||
| 353 | td->td_retvaltd_uretoff.tdu_retval[0] = lim; | |||
| 354 | #endif | |||
| 355 | return (0); | |||
| 356 | } | |||
| 357 | ||||
| 358 | /* | |||
| 359 | * Duplicate a file descriptor to a particular value. | |||
| 360 | * | |||
| 361 | * Note: keep in mind that a potential race condition exists when closing | |||
| 362 | * descriptors from a shared descriptor table (via rfork). | |||
| 363 | */ | |||
| 364 | #ifndef _SYS_SYSPROTO_H_ | |||
| 365 | struct dup2_args { | |||
| 366 | u_int from; | |||
| 367 | u_int to; | |||
| 368 | }; | |||
| 369 | #endif | |||
| 370 | /* ARGSUSED */ | |||
| 371 | int | |||
| 372 | sys_dup2(struct thread *td, struct dup2_args *uap) | |||
| 373 | { | |||
| 374 | ||||
| 375 | return (kern_dup(td, FDDUP_FIXED, 0, (int)uap->from, (int)uap->to)); | |||
| 376 | } | |||
| 377 | ||||
| 378 | /* | |||
| 379 | * Duplicate a file descriptor. | |||
| 380 | */ | |||
| 381 | #ifndef _SYS_SYSPROTO_H_ | |||
| 382 | struct dup_args { | |||
| 383 | u_int fd; | |||
| 384 | }; | |||
| 385 | #endif | |||
| 386 | /* ARGSUSED */ | |||
| 387 | int | |||
| 388 | sys_dup(struct thread *td, struct dup_args *uap) | |||
| 389 | { | |||
| 390 | ||||
| 391 | return (kern_dup(td, FDDUP_NORMAL, 0, (int)uap->fd, 0)); | |||
| 392 | } | |||
| 393 | ||||
| 394 | /* | |||
| 395 | * The file control system call. | |||
| 396 | */ | |||
| 397 | #ifndef _SYS_SYSPROTO_H_ | |||
| 398 | struct fcntl_args { | |||
| 399 | int fd; | |||
| 400 | int cmd; | |||
| 401 | long arg; | |||
| 402 | }; | |||
| 403 | #endif | |||
| 404 | /* ARGSUSED */ | |||
| 405 | int | |||
| 406 | sys_fcntl(struct thread *td, struct fcntl_args *uap) | |||
| 407 | { | |||
| 408 | ||||
| 409 | return (kern_fcntl_freebsd(td, uap->fd, uap->cmd, uap->arg)); | |||
| 410 | } | |||
| 411 | ||||
| 412 | int | |||
| 413 | kern_fcntl_freebsd(struct thread *td, int fd, int cmd, long arg) | |||
| 414 | { | |||
| 415 | struct flock fl; | |||
| 416 | struct __oflock ofl; | |||
| 417 | intptr_t arg1; | |||
| 418 | int error, newcmd; | |||
| 419 | ||||
| 420 | error = 0; | |||
| 421 | newcmd = cmd; | |||
| 422 | switch (cmd) { | |||
| 423 | case F_OGETLK7: | |||
| 424 | case F_OSETLK8: | |||
| 425 | case F_OSETLKW9: | |||
| 426 | /* | |||
| 427 | * Convert old flock structure to new. | |||
| 428 | */ | |||
| 429 | error = copyin((void *)(intptr_t)arg, &ofl, sizeof(ofl)); | |||
| 430 | fl.l_start = ofl.l_start; | |||
| 431 | fl.l_len = ofl.l_len; | |||
| 432 | fl.l_pid = ofl.l_pid; | |||
| 433 | fl.l_type = ofl.l_type; | |||
| 434 | fl.l_whence = ofl.l_whence; | |||
| 435 | fl.l_sysid = 0; | |||
| 436 | ||||
| 437 | switch (cmd) { | |||
| 438 | case F_OGETLK7: | |||
| 439 | newcmd = F_GETLK11; | |||
| 440 | break; | |||
| 441 | case F_OSETLK8: | |||
| 442 | newcmd = F_SETLK12; | |||
| 443 | break; | |||
| 444 | case F_OSETLKW9: | |||
| 445 | newcmd = F_SETLKW13; | |||
| 446 | break; | |||
| 447 | } | |||
| 448 | arg1 = (intptr_t)&fl; | |||
| 449 | break; | |||
| 450 | case F_GETLK11: | |||
| 451 | case F_SETLK12: | |||
| 452 | case F_SETLKW13: | |||
| 453 | case F_SETLK_REMOTE14: | |||
| 454 | error = copyin((void *)(intptr_t)arg, &fl, sizeof(fl)); | |||
| 455 | arg1 = (intptr_t)&fl; | |||
| 456 | break; | |||
| 457 | default: | |||
| 458 | arg1 = arg; | |||
| 459 | break; | |||
| 460 | } | |||
| 461 | if (error) | |||
| 462 | return (error); | |||
| 463 | error = kern_fcntl(td, fd, newcmd, arg1); | |||
| 464 | if (error) | |||
| 465 | return (error); | |||
| 466 | if (cmd == F_OGETLK7) { | |||
| 467 | ofl.l_start = fl.l_start; | |||
| 468 | ofl.l_len = fl.l_len; | |||
| 469 | ofl.l_pid = fl.l_pid; | |||
| 470 | ofl.l_type = fl.l_type; | |||
| 471 | ofl.l_whence = fl.l_whence; | |||
| 472 | error = copyout(&ofl, (void *)(intptr_t)arg, sizeof(ofl)); | |||
| 473 | } else if (cmd == F_GETLK11) { | |||
| 474 | error = copyout(&fl, (void *)(intptr_t)arg, sizeof(fl)); | |||
| 475 | } | |||
| 476 | return (error); | |||
| 477 | } | |||
| 478 | ||||
| 479 | int | |||
| 480 | kern_fcntl(struct thread *td, int fd, int cmd, intptr_t arg) | |||
| 481 | { | |||
| 482 | struct filedesc *fdp; | |||
| 483 | struct flock *flp; | |||
| 484 | struct file *fp, *fp2; | |||
| 485 | struct filedescent *fde; | |||
| 486 | struct proc *p; | |||
| 487 | struct vnode *vp; | |||
| 488 | cap_rights_t rights; | |||
| 489 | int error, flg, tmp; | |||
| 490 | uint64_t bsize; | |||
| 491 | off_t foffset; | |||
| 492 | ||||
| 493 | error = 0; | |||
| 494 | flg = F_POSIX0x040; | |||
| 495 | p = td->td_proc; | |||
| 496 | fdp = p->p_fd; | |||
| 497 | ||||
| 498 | switch (cmd) { | |||
| 499 | case F_DUPFD0: | |||
| 500 | tmp = arg; | |||
| 501 | error = kern_dup(td, FDDUP_FCNTL, 0, fd, tmp); | |||
| 502 | break; | |||
| 503 | ||||
| 504 | case F_DUPFD_CLOEXEC17: | |||
| 505 | tmp = arg; | |||
| 506 | error = kern_dup(td, FDDUP_FCNTL, FDDUP_FLAG_CLOEXEC0x1, fd, tmp); | |||
| 507 | break; | |||
| 508 | ||||
| 509 | case F_DUP2FD10: | |||
| 510 | tmp = arg; | |||
| 511 | error = kern_dup(td, FDDUP_FIXED, 0, fd, tmp); | |||
| 512 | break; | |||
| 513 | ||||
| 514 | case F_DUP2FD_CLOEXEC18: | |||
| 515 | tmp = arg; | |||
| 516 | error = kern_dup(td, FDDUP_FIXED, FDDUP_FLAG_CLOEXEC0x1, fd, tmp); | |||
| 517 | break; | |||
| 518 | ||||
| 519 | case F_GETFD1: | |||
| 520 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 521 | if (fget_locked(fdp, fd) == NULL((void *)0)) { | |||
| 522 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 523 | error = EBADF9; | |||
| 524 | break; | |||
| 525 | } | |||
| 526 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[fd]; | |||
| 527 | td->td_retvaltd_uretoff.tdu_retval[0] = | |||
| 528 | (fde->fde_flags & UF_EXCLOSE0x01) ? FD_CLOEXEC1 : 0; | |||
| 529 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 530 | break; | |||
| 531 | ||||
| 532 | case F_SETFD2: | |||
| 533 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 534 | if (fget_locked(fdp, fd) == NULL((void *)0)) { | |||
| 535 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 536 | error = EBADF9; | |||
| 537 | break; | |||
| 538 | } | |||
| 539 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[fd]; | |||
| 540 | fde->fde_flags = (fde->fde_flags & ~UF_EXCLOSE0x01) | | |||
| 541 | (arg & FD_CLOEXEC1 ? UF_EXCLOSE0x01 : 0); | |||
| 542 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 543 | break; | |||
| 544 | ||||
| 545 | case F_GETFL3: | |||
| 546 | error = fget_fcntl(td, fd, | |||
| 547 | cap_rights_init(&rights, CAP_FCNTL)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000008000ULL)), 0ULL), F_GETFL3, &fp); | |||
| 548 | if (error != 0) | |||
| 549 | break; | |||
| 550 | td->td_retvaltd_uretoff.tdu_retval[0] = OFLAGS(fp->f_flag)((fp->f_flag) & 0x00040000 ? (fp->f_flag) : (fp-> f_flag) - 1); | |||
| 551 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 552 | break; | |||
| 553 | ||||
| 554 | case F_SETFL4: | |||
| 555 | error = fget_fcntl(td, fd, | |||
| 556 | cap_rights_init(&rights, CAP_FCNTL)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000008000ULL)), 0ULL), F_SETFL4, &fp); | |||
| 557 | if (error != 0) | |||
| 558 | break; | |||
| 559 | do { | |||
| 560 | tmp = flg = fp->f_flag; | |||
| 561 | tmp &= ~FCNTLFLAGS(0x0008|0x0040|0x0080|0x0004|0x0100|0x0200| 0x00010000); | |||
| 562 | tmp |= FFLAGS(arg & ~O_ACCMODE)((arg & ~0x0003) & 0x00040000 ? (arg & ~0x0003) : (arg & ~0x0003) + 1) & FCNTLFLAGS(0x0008|0x0040|0x0080|0x0004|0x0100|0x0200| 0x00010000); | |||
| 563 | } while(atomic_cmpset_int(&fp->f_flag, flg, tmp) == 0); | |||
| 564 | tmp = fp->f_flag & FNONBLOCK0x0004; | |||
| 565 | error = fo_ioctl(fp, FIONBIO((unsigned long) ((0x80000000) | (((sizeof(int)) & ((1 << 13) - 1)) << 16) | ((('f')) << 8) | ((126)))), &tmp, td->td_ucred, td); | |||
| 566 | if (error != 0) { | |||
| 567 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 568 | break; | |||
| 569 | } | |||
| 570 | tmp = fp->f_flag & FASYNC0x0040; | |||
| 571 | error = fo_ioctl(fp, FIOASYNC((unsigned long) ((0x80000000) | (((sizeof(int)) & ((1 << 13) - 1)) << 16) | ((('f')) << 8) | ((125)))), &tmp, td->td_ucred, td); | |||
| 572 | if (error == 0) { | |||
| 573 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 574 | break; | |||
| 575 | } | |||
| 576 | atomic_clear_int(&fp->f_flag, FNONBLOCK0x0004); | |||
| 577 | tmp = 0; | |||
| 578 | (void)fo_ioctl(fp, FIONBIO((unsigned long) ((0x80000000) | (((sizeof(int)) & ((1 << 13) - 1)) << 16) | ((('f')) << 8) | ((126)))), &tmp, td->td_ucred, td); | |||
| 579 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 580 | break; | |||
| 581 | ||||
| 582 | case F_GETOWN5: | |||
| 583 | error = fget_fcntl(td, fd, | |||
| 584 | cap_rights_init(&rights, CAP_FCNTL)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000008000ULL)), 0ULL), F_GETOWN5, &fp); | |||
| 585 | if (error != 0) | |||
| 586 | break; | |||
| 587 | error = fo_ioctl(fp, FIOGETOWN((unsigned long) ((0x40000000) | (((sizeof(int)) & ((1 << 13) - 1)) << 16) | ((('f')) << 8) | ((123)))), &tmp, td->td_ucred, td); | |||
| 588 | if (error == 0) | |||
| 589 | td->td_retvaltd_uretoff.tdu_retval[0] = tmp; | |||
| 590 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 591 | break; | |||
| 592 | ||||
| 593 | case F_SETOWN6: | |||
| 594 | error = fget_fcntl(td, fd, | |||
| 595 | cap_rights_init(&rights, CAP_FCNTL)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000008000ULL)), 0ULL), F_SETOWN6, &fp); | |||
| 596 | if (error != 0) | |||
| 597 | break; | |||
| 598 | tmp = arg; | |||
| 599 | error = fo_ioctl(fp, FIOSETOWN((unsigned long) ((0x80000000) | (((sizeof(int)) & ((1 << 13) - 1)) << 16) | ((('f')) << 8) | ((124)))), &tmp, td->td_ucred, td); | |||
| 600 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 601 | break; | |||
| 602 | ||||
| 603 | case F_SETLK_REMOTE14: | |||
| 604 | error = priv_check(td, PRIV_NFS_LOCKD291); | |||
| 605 | if (error) | |||
| 606 | return (error); | |||
| 607 | flg = F_REMOTE0x080; | |||
| 608 | goto do_setlk; | |||
| 609 | ||||
| 610 | case F_SETLKW13: | |||
| 611 | flg |= F_WAIT0x010; | |||
| 612 | /* FALLTHROUGH F_SETLK */ | |||
| 613 | ||||
| 614 | case F_SETLK12: | |||
| 615 | do_setlk: | |||
| 616 | cap_rights_init(&rights, CAP_FLOCK)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000010000ULL)), 0ULL); | |||
| 617 | error = fget_unlocked(fdp, fd, &rights, &fp, NULL((void *)0)); | |||
| 618 | if (error != 0) | |||
| 619 | break; | |||
| 620 | if (fp->f_type != DTYPE_VNODE1) { | |||
| 621 | error = EBADF9; | |||
| 622 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 623 | break; | |||
| 624 | } | |||
| 625 | ||||
| 626 | flp = (struct flock *)arg; | |||
| 627 | if (flp->l_whence == SEEK_CUR1) { | |||
| 628 | foffset = foffset_get(fp); | |||
| 629 | if (foffset < 0 || | |||
| 630 | (flp->l_start > 0 && | |||
| 631 | foffset > OFF_MAX0x7fffffffffffffff - flp->l_start)) { | |||
| 632 | error = EOVERFLOW84; | |||
| 633 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 634 | break; | |||
| 635 | } | |||
| 636 | flp->l_start += foffset; | |||
| 637 | } | |||
| 638 | ||||
| 639 | vp = fp->f_vnode; | |||
| 640 | switch (flp->l_type) { | |||
| 641 | case F_RDLCK1: | |||
| 642 | if ((fp->f_flag & FREAD0x0001) == 0) { | |||
| 643 | error = EBADF9; | |||
| 644 | break; | |||
| 645 | } | |||
| 646 | PROC_LOCK(p->p_leader)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p->p_leader)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&(((((&(p->p_leader)->p_mtx )))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep(& (((((&(p->p_leader)->p_mtx)))))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire->id , (uintptr_t) (((&(p->p_leader)->p_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); } while (0); | |||
| 647 | p->p_leader->p_flag |= P_ADVLOCK0x00001; | |||
| 648 | PROC_UNLOCK(p->p_leader)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p->p_leader)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(p->p_leader)->p_mtx))), ( uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p->p_leader)->p_mtx ))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((& (p->p_leader)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p->p_leader)->p_mtx )))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 649 | error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, F_SETLK12, | |||
| 650 | flp, flg); | |||
| 651 | break; | |||
| 652 | case F_WRLCK3: | |||
| 653 | if ((fp->f_flag & FWRITE0x0002) == 0) { | |||
| 654 | error = EBADF9; | |||
| 655 | break; | |||
| 656 | } | |||
| 657 | PROC_LOCK(p->p_leader)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p->p_leader)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&(((((&(p->p_leader)->p_mtx )))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep(& (((((&(p->p_leader)->p_mtx)))))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire->id , (uintptr_t) (((&(p->p_leader)->p_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); } while (0); | |||
| 658 | p->p_leader->p_flag |= P_ADVLOCK0x00001; | |||
| 659 | PROC_UNLOCK(p->p_leader)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p->p_leader)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(p->p_leader)->p_mtx))), ( uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p->p_leader)->p_mtx ))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((& (p->p_leader)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p->p_leader)->p_mtx )))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 660 | error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, F_SETLK12, | |||
| 661 | flp, flg); | |||
| 662 | break; | |||
| 663 | case F_UNLCK2: | |||
| 664 | error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, F_UNLCK2, | |||
| 665 | flp, flg); | |||
| 666 | break; | |||
| 667 | case F_UNLCKSYS4: | |||
| 668 | /* | |||
| 669 | * Temporary api for testing remote lock | |||
| 670 | * infrastructure. | |||
| 671 | */ | |||
| 672 | if (flg != F_REMOTE0x080) { | |||
| 673 | error = EINVAL22; | |||
| 674 | break; | |||
| 675 | } | |||
| 676 | error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, | |||
| 677 | F_UNLCKSYS4, flp, flg); | |||
| 678 | break; | |||
| 679 | default: | |||
| 680 | error = EINVAL22; | |||
| 681 | break; | |||
| 682 | } | |||
| 683 | if (error != 0 || flp->l_type == F_UNLCK2 || | |||
| 684 | flp->l_type == F_UNLCKSYS4) { | |||
| 685 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 686 | break; | |||
| 687 | } | |||
| 688 | ||||
| 689 | /* | |||
| 690 | * Check for a race with close. | |||
| 691 | * | |||
| 692 | * The vnode is now advisory locked (or unlocked, but this case | |||
| 693 | * is not really important) as the caller requested. | |||
| 694 | * We had to drop the filedesc lock, so we need to recheck if | |||
| 695 | * the descriptor is still valid, because if it was closed | |||
| 696 | * in the meantime we need to remove advisory lock from the | |||
| 697 | * vnode - close on any descriptor leading to an advisory | |||
| 698 | * locked vnode, removes that lock. | |||
| 699 | * We will return 0 on purpose in that case, as the result of | |||
| 700 | * successful advisory lock might have been externally visible | |||
| 701 | * already. This is fine - effectively we pretend to the caller | |||
| 702 | * that the closing thread was a bit slower and that the | |||
| 703 | * advisory lock succeeded before the close. | |||
| 704 | */ | |||
| 705 | error = fget_unlocked(fdp, fd, &rights, &fp2, NULL((void *)0)); | |||
| 706 | if (error != 0) { | |||
| 707 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 708 | break; | |||
| 709 | } | |||
| 710 | if (fp != fp2) { | |||
| 711 | flp->l_whence = SEEK_SET0; | |||
| 712 | flp->l_start = 0; | |||
| 713 | flp->l_len = 0; | |||
| 714 | flp->l_type = F_UNLCK2; | |||
| 715 | (void) VOP_ADVLOCK(vp, (caddr_t)p->p_leader, | |||
| 716 | F_UNLCK2, flp, F_POSIX0x040); | |||
| 717 | } | |||
| 718 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 719 | fdrop(fp2, td)(refcount_release(&(fp2)->f_count) ? _fdrop((fp2), (td )) : _fnoop()); | |||
| 720 | break; | |||
| 721 | ||||
| 722 | case F_GETLK11: | |||
| 723 | error = fget_unlocked(fdp, fd, | |||
| 724 | cap_rights_init(&rights, CAP_FLOCK)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000010000ULL)), 0ULL), &fp, NULL((void *)0)); | |||
| 725 | if (error != 0) | |||
| 726 | break; | |||
| 727 | if (fp->f_type != DTYPE_VNODE1) { | |||
| 728 | error = EBADF9; | |||
| 729 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 730 | break; | |||
| 731 | } | |||
| 732 | flp = (struct flock *)arg; | |||
| 733 | if (flp->l_type != F_RDLCK1 && flp->l_type != F_WRLCK3 && | |||
| 734 | flp->l_type != F_UNLCK2) { | |||
| 735 | error = EINVAL22; | |||
| 736 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 737 | break; | |||
| 738 | } | |||
| 739 | if (flp->l_whence == SEEK_CUR1) { | |||
| 740 | foffset = foffset_get(fp); | |||
| 741 | if ((flp->l_start > 0 && | |||
| 742 | foffset > OFF_MAX0x7fffffffffffffff - flp->l_start) || | |||
| 743 | (flp->l_start < 0 && | |||
| 744 | foffset < OFF_MIN(-0x7fffffffffffffff - 1) - flp->l_start)) { | |||
| 745 | error = EOVERFLOW84; | |||
| 746 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 747 | break; | |||
| 748 | } | |||
| 749 | flp->l_start += foffset; | |||
| 750 | } | |||
| 751 | vp = fp->f_vnode; | |||
| 752 | error = VOP_ADVLOCK(vp, (caddr_t)p->p_leader, F_GETLK11, flp, | |||
| 753 | F_POSIX0x040); | |||
| 754 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 755 | break; | |||
| 756 | ||||
| 757 | case F_RDAHEAD16: | |||
| 758 | arg = arg ? 128 * 1024: 0; | |||
| 759 | /* FALLTHROUGH */ | |||
| 760 | case F_READAHEAD15: | |||
| 761 | error = fget_unlocked(fdp, fd, | |||
| 762 | cap_rights_init(&rights)__cap_rights_init(0, &rights, 0ULL), &fp, NULL((void *)0)); | |||
| 763 | if (error != 0) | |||
| 764 | break; | |||
| 765 | if (fp->f_type != DTYPE_VNODE1) { | |||
| 766 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 767 | error = EBADF9; | |||
| 768 | break; | |||
| 769 | } | |||
| 770 | vp = fp->f_vnode; | |||
| 771 | /* | |||
| 772 | * Exclusive lock synchronizes against f_seqcount reads and | |||
| 773 | * writes in sequential_heuristic(). | |||
| 774 | */ | |||
| 775 | error = vn_lock(vp, LK_EXCLUSIVE)_vn_lock(vp, 0x080000, "/usr/src/sys/kern/kern_descrip.c", 775 ); | |||
| 776 | if (error != 0) { | |||
| 777 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 778 | break; | |||
| 779 | } | |||
| 780 | if (arg >= 0) { | |||
| 781 | bsize = fp->f_vnode->v_mount->mnt_stat.f_iosize; | |||
| 782 | fp->f_seqcount = (arg + bsize - 1) / bsize; | |||
| 783 | atomic_set_int(&fp->f_flag, FRDAHEAD0x0200); | |||
| 784 | } else { | |||
| 785 | atomic_clear_int(&fp->f_flag, FRDAHEAD0x0200); | |||
| 786 | } | |||
| 787 | VOP_UNLOCK(vp, 0); | |||
| 788 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 789 | break; | |||
| 790 | ||||
| 791 | default: | |||
| 792 | error = EINVAL22; | |||
| 793 | break; | |||
| 794 | } | |||
| 795 | return (error); | |||
| 796 | } | |||
| 797 | ||||
| 798 | static int | |||
| 799 | getmaxfd(struct thread *td) | |||
| 800 | { | |||
| 801 | ||||
| 802 | return (min((int)lim_cur(td, RLIMIT_NOFILE8), maxfilesperproc)); | |||
| 803 | } | |||
| 804 | ||||
| 805 | /* | |||
| 806 | * Common code for dup, dup2, fcntl(F_DUPFD) and fcntl(F_DUP2FD). | |||
| 807 | */ | |||
| 808 | int | |||
| 809 | kern_dup(struct thread *td, u_int mode, int flags, int old, int new) | |||
| 810 | { | |||
| 811 | struct filedesc *fdp; | |||
| 812 | struct filedescent *oldfde, *newfde; | |||
| 813 | struct proc *p; | |||
| 814 | struct file *delfp; | |||
| 815 | int error, maxfd; | |||
| 816 | ||||
| 817 | p = td->td_proc; | |||
| 818 | fdp = p->p_fd; | |||
| 819 | ||||
| 820 | MPASS((flags & ~(FDDUP_FLAG_CLOEXEC)) == 0)do { } while (0); | |||
| 821 | MPASS(mode < FDDUP_LASTMODE)do { } while (0); | |||
| 822 | ||||
| 823 | /* | |||
| 824 | * Verify we have a valid descriptor to dup from and possibly to | |||
| 825 | * dup to. Unlike dup() and dup2(), fcntl()'s F_DUPFD should | |||
| 826 | * return EINVAL when the new descriptor is out of bounds. | |||
| 827 | */ | |||
| 828 | if (old < 0) | |||
| 829 | return (EBADF9); | |||
| 830 | if (new < 0) | |||
| 831 | return (mode == FDDUP_FCNTL ? EINVAL22 : EBADF9); | |||
| 832 | maxfd = getmaxfd(td); | |||
| 833 | if (new >= maxfd) | |||
| 834 | return (mode == FDDUP_FCNTL ? EINVAL22 : EBADF9); | |||
| 835 | ||||
| 836 | error = EBADF9; | |||
| 837 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 838 | if (fget_locked(fdp, old) == NULL((void *)0)) | |||
| 839 | goto unlock; | |||
| 840 | if ((mode == FDDUP_FIXED || mode == FDDUP_MUSTREPLACE) && old == new) { | |||
| 841 | td->td_retvaltd_uretoff.tdu_retval[0] = new; | |||
| 842 | if (flags & FDDUP_FLAG_CLOEXEC0x1) | |||
| 843 | fdp->fd_ofilesfd_files->fdt_ofiles[new].fde_flags |= UF_EXCLOSE0x01; | |||
| 844 | error = 0; | |||
| 845 | goto unlock; | |||
| 846 | } | |||
| 847 | ||||
| 848 | /* | |||
| 849 | * If the caller specified a file descriptor, make sure the file | |||
| 850 | * table is large enough to hold it, and grab it. Otherwise, just | |||
| 851 | * allocate a new descriptor the usual way. | |||
| 852 | */ | |||
| 853 | switch (mode) { | |||
| 854 | case FDDUP_NORMAL: | |||
| 855 | case FDDUP_FCNTL: | |||
| 856 | if ((error = fdalloc(td, new, &new)) != 0) | |||
| 857 | goto unlock; | |||
| 858 | break; | |||
| 859 | case FDDUP_MUSTREPLACE: | |||
| 860 | /* Target file descriptor must exist. */ | |||
| 861 | if (fget_locked(fdp, new) == NULL((void *)0)) | |||
| 862 | goto unlock; | |||
| 863 | break; | |||
| 864 | case FDDUP_FIXED: | |||
| 865 | if (new >= fdp->fd_nfilesfd_files->fdt_nfiles) { | |||
| 866 | /* | |||
| 867 | * The resource limits are here instead of e.g. | |||
| 868 | * fdalloc(), because the file descriptor table may be | |||
| 869 | * shared between processes, so we can't really use | |||
| 870 | * racct_add()/racct_sub(). Instead of counting the | |||
| 871 | * number of actually allocated descriptors, just put | |||
| 872 | * the limit on the size of the file descriptor table. | |||
| 873 | */ | |||
| 874 | #ifdef RACCT1 | |||
| 875 | if (racct_enable) { | |||
| 876 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 877 | error = racct_set(p, RACCT_NOFILE7, new + 1); | |||
| 878 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 879 | if (error != 0) { | |||
| 880 | error = EMFILE24; | |||
| 881 | goto unlock; | |||
| 882 | } | |||
| 883 | } | |||
| 884 | #endif | |||
| 885 | fdgrowtable_exp(fdp, new + 1); | |||
| 886 | } | |||
| 887 | if (!fdisused(fdp, new)) | |||
| 888 | fdused(fdp, new); | |||
| 889 | break; | |||
| 890 | default: | |||
| 891 | KASSERT(0, ("%s unsupported mode %d", __func__, mode))do { } while (0); | |||
| 892 | } | |||
| 893 | ||||
| 894 | KASSERT(old != new, ("new fd is same as old"))do { } while (0); | |||
| 895 | ||||
| 896 | oldfde = &fdp->fd_ofilesfd_files->fdt_ofiles[old]; | |||
| 897 | fhold(oldfde->fde_file)(refcount_acquire(&(oldfde->fde_file)->f_count)); | |||
| 898 | newfde = &fdp->fd_ofilesfd_files->fdt_ofiles[new]; | |||
| 899 | delfp = newfde->fde_file; | |||
| 900 | ||||
| 901 | /* | |||
| 902 | * Duplicate the source descriptor. | |||
| 903 | */ | |||
| 904 | #ifdef CAPABILITIES1 | |||
| 905 | seq_write_begin(&newfde->fde_seq); | |||
| 906 | #endif | |||
| 907 | filecaps_free(&newfde->fde_caps); | |||
| 908 | memcpy(newfde, oldfde, fde_change_size(__builtin_offsetof(struct filedescent, fde_seq))); | |||
| 909 | filecaps_copy(&oldfde->fde_caps, &newfde->fde_caps, true1); | |||
| 910 | if ((flags & FDDUP_FLAG_CLOEXEC0x1) != 0) | |||
| 911 | newfde->fde_flags = oldfde->fde_flags | UF_EXCLOSE0x01; | |||
| 912 | else | |||
| 913 | newfde->fde_flags = oldfde->fde_flags & ~UF_EXCLOSE0x01; | |||
| 914 | #ifdef CAPABILITIES1 | |||
| 915 | seq_write_end(&newfde->fde_seq); | |||
| 916 | #endif | |||
| 917 | td->td_retvaltd_uretoff.tdu_retval[0] = new; | |||
| 918 | ||||
| 919 | error = 0; | |||
| 920 | ||||
| 921 | if (delfp != NULL((void *)0)) { | |||
| 922 | (void) closefp(fdp, new, delfp, td, 1); | |||
| 923 | FILEDESC_UNLOCK_ASSERT(fdp)(void)0; | |||
| 924 | } else { | |||
| 925 | unlock: | |||
| 926 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 927 | } | |||
| 928 | ||||
| 929 | return (error); | |||
| 930 | } | |||
| 931 | ||||
| 932 | /* | |||
| 933 | * If sigio is on the list associated with a process or process group, | |||
| 934 | * disable signalling from the device, remove sigio from the list and | |||
| 935 | * free sigio. | |||
| 936 | */ | |||
| 937 | void | |||
| 938 | funsetown(struct sigio **sigiop) | |||
| 939 | { | |||
| 940 | struct sigio *sigio; | |||
| 941 | ||||
| 942 | SIGIO_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &sigio_lock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&sigio_lock)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 943 | sigio = *sigiop; | |||
| 944 | if (sigio == NULL((void *)0)) { | |||
| 945 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 946 | return; | |||
| 947 | } | |||
| 948 | *(sigio->sio_myref) = NULL((void *)0); | |||
| 949 | if ((sigio)->sio_pgid < 0) { | |||
| 950 | struct pgrp *pg = (sigio)->sio_pgrpsio_u.siu_pgrp; | |||
| 951 | PGRP_LOCK(pg)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(pg)->pg_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(pg)->pg_mtx)))))->mtx_lock, 0x00000004 , (_tid)))) __mtx_lock_sleep(&(((((&(pg)->pg_mtx)) )))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(pg)->pg_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 952 | SLIST_REMOVE(&sigio->sio_pgrp->pg_sigiolst, sigio,do { ; if ((((&sigio->sio_u.siu_pgrp->pg_sigiolst)) ->slh_first) == (sigio)) { do { ((((&sigio->sio_u.siu_pgrp ->pg_sigiolst)))->slh_first) = ((((((&sigio->sio_u .siu_pgrp->pg_sigiolst)))->slh_first))->sio_pgsigio. sle_next); } while (0); } else { struct sigio *curelm = ((& sigio->sio_u.siu_pgrp->pg_sigiolst)->slh_first); while (((curelm)->sio_pgsigio.sle_next) != (sigio)) curelm = (( curelm)->sio_pgsigio.sle_next); do { ((curelm)->sio_pgsigio .sle_next) = ((((curelm)->sio_pgsigio.sle_next))->sio_pgsigio .sle_next); } while (0); } ; } while (0) | |||
| 953 | sigio, sio_pgsigio)do { ; if ((((&sigio->sio_u.siu_pgrp->pg_sigiolst)) ->slh_first) == (sigio)) { do { ((((&sigio->sio_u.siu_pgrp ->pg_sigiolst)))->slh_first) = ((((((&sigio->sio_u .siu_pgrp->pg_sigiolst)))->slh_first))->sio_pgsigio. sle_next); } while (0); } else { struct sigio *curelm = ((& sigio->sio_u.siu_pgrp->pg_sigiolst)->slh_first); while (((curelm)->sio_pgsigio.sle_next) != (sigio)) curelm = (( curelm)->sio_pgsigio.sle_next); do { ((curelm)->sio_pgsigio .sle_next) = ((((curelm)->sio_pgsigio.sle_next))->sio_pgsigio .sle_next); } while (0); } ; } while (0); | |||
| 954 | PGRP_UNLOCK(pg)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (pg)->pg_mtx))))->lock_object.lo_data == 0) do { (void) 0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(pg)->pg_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(pg)->pg_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(pg)->pg_mtx)))))-> mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(&(((((& (pg)->pg_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ( (0))); } while (0); | |||
| 955 | } else { | |||
| 956 | struct proc *p = (sigio)->sio_procsio_u.siu_proc; | |||
| 957 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 958 | SLIST_REMOVE(&sigio->sio_proc->p_sigiolst, sigio,do { ; if ((((&sigio->sio_u.siu_proc->p_sigiolst))-> slh_first) == (sigio)) { do { ((((&sigio->sio_u.siu_proc ->p_sigiolst)))->slh_first) = ((((((&sigio->sio_u .siu_proc->p_sigiolst)))->slh_first))->sio_pgsigio.sle_next ); } while (0); } else { struct sigio *curelm = ((&sigio-> sio_u.siu_proc->p_sigiolst)->slh_first); while (((curelm )->sio_pgsigio.sle_next) != (sigio)) curelm = ((curelm)-> sio_pgsigio.sle_next); do { ((curelm)->sio_pgsigio.sle_next ) = ((((curelm)->sio_pgsigio.sle_next))->sio_pgsigio.sle_next ); } while (0); } ; } while (0) | |||
| 959 | sigio, sio_pgsigio)do { ; if ((((&sigio->sio_u.siu_proc->p_sigiolst))-> slh_first) == (sigio)) { do { ((((&sigio->sio_u.siu_proc ->p_sigiolst)))->slh_first) = ((((((&sigio->sio_u .siu_proc->p_sigiolst)))->slh_first))->sio_pgsigio.sle_next ); } while (0); } else { struct sigio *curelm = ((&sigio-> sio_u.siu_proc->p_sigiolst)->slh_first); while (((curelm )->sio_pgsigio.sle_next) != (sigio)) curelm = ((curelm)-> sio_pgsigio.sle_next); do { ((curelm)->sio_pgsigio.sle_next ) = ((((curelm)->sio_pgsigio.sle_next))->sio_pgsigio.sle_next ); } while (0); } ; } while (0); | |||
| 960 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 961 | } | |||
| 962 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 963 | crfree(sigio->sio_ucred); | |||
| 964 | free(sigio, M_SIGIO); | |||
| 965 | } | |||
| 966 | ||||
| 967 | /* | |||
| 968 | * Free a list of sigio structures. | |||
| 969 | * We only need to lock the SIGIO_LOCK because we have made ourselves | |||
| 970 | * inaccessible to callers of fsetown and therefore do not need to lock | |||
| 971 | * the proc or pgrp struct for the list manipulation. | |||
| 972 | */ | |||
| 973 | void | |||
| 974 | funsetownlst(struct sigiolst *sigiolst) | |||
| 975 | { | |||
| 976 | struct proc *p; | |||
| 977 | struct pgrp *pg; | |||
| 978 | struct sigio *sigio; | |||
| 979 | ||||
| 980 | sigio = SLIST_FIRST(sigiolst)((sigiolst)->slh_first); | |||
| 981 | if (sigio == NULL((void *)0)) | |||
| 982 | return; | |||
| 983 | p = NULL((void *)0); | |||
| 984 | pg = NULL((void *)0); | |||
| 985 | ||||
| 986 | /* | |||
| 987 | * Every entry of the list should belong | |||
| 988 | * to a single proc or pgrp. | |||
| 989 | */ | |||
| 990 | if (sigio->sio_pgid < 0) { | |||
| 991 | pg = sigio->sio_pgrpsio_u.siu_pgrp; | |||
| 992 | PGRP_LOCK_ASSERT(pg, MA_NOTOWNED)(void)0; | |||
| 993 | } else /* if (sigio->sio_pgid > 0) */ { | |||
| 994 | p = sigio->sio_procsio_u.siu_proc; | |||
| 995 | PROC_LOCK_ASSERT(p, MA_NOTOWNED)(void)0; | |||
| 996 | } | |||
| 997 | ||||
| 998 | SIGIO_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &sigio_lock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&sigio_lock)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 999 | while ((sigio = SLIST_FIRST(sigiolst)((sigiolst)->slh_first)) != NULL((void *)0)) { | |||
| 1000 | *(sigio->sio_myref) = NULL((void *)0); | |||
| 1001 | if (pg != NULL((void *)0)) { | |||
| 1002 | KASSERT(sigio->sio_pgid < 0,do { } while (0) | |||
| 1003 | ("Proc sigio in pgrp sigio list"))do { } while (0); | |||
| 1004 | KASSERT(sigio->sio_pgrp == pg,do { } while (0) | |||
| 1005 | ("Bogus pgrp in sigio list"))do { } while (0); | |||
| 1006 | PGRP_LOCK(pg)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(pg)->pg_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(pg)->pg_mtx)))))->mtx_lock, 0x00000004 , (_tid)))) __mtx_lock_sleep(&(((((&(pg)->pg_mtx)) )))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(pg)->pg_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1007 | SLIST_REMOVE(&pg->pg_sigiolst, sigio, sigio,do { ; if ((((&pg->pg_sigiolst))->slh_first) == (sigio )) { do { ((((&pg->pg_sigiolst)))->slh_first) = ((( (((&pg->pg_sigiolst)))->slh_first))->sio_pgsigio .sle_next); } while (0); } else { struct sigio *curelm = ((& pg->pg_sigiolst)->slh_first); while (((curelm)->sio_pgsigio .sle_next) != (sigio)) curelm = ((curelm)->sio_pgsigio.sle_next ); do { ((curelm)->sio_pgsigio.sle_next) = ((((curelm)-> sio_pgsigio.sle_next))->sio_pgsigio.sle_next); } while (0) ; } ; } while (0) | |||
| 1008 | sio_pgsigio)do { ; if ((((&pg->pg_sigiolst))->slh_first) == (sigio )) { do { ((((&pg->pg_sigiolst)))->slh_first) = ((( (((&pg->pg_sigiolst)))->slh_first))->sio_pgsigio .sle_next); } while (0); } else { struct sigio *curelm = ((& pg->pg_sigiolst)->slh_first); while (((curelm)->sio_pgsigio .sle_next) != (sigio)) curelm = ((curelm)->sio_pgsigio.sle_next ); do { ((curelm)->sio_pgsigio.sle_next) = ((((curelm)-> sio_pgsigio.sle_next))->sio_pgsigio.sle_next); } while (0) ; } ; } while (0); | |||
| 1009 | PGRP_UNLOCK(pg)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (pg)->pg_mtx))))->lock_object.lo_data == 0) do { (void) 0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(pg)->pg_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(pg)->pg_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(pg)->pg_mtx)))))-> mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(&(((((& (pg)->pg_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ( (0))); } while (0); | |||
| 1010 | } else /* if (p != NULL) */ { | |||
| 1011 | KASSERT(sigio->sio_pgid > 0,do { } while (0) | |||
| 1012 | ("Pgrp sigio in proc sigio list"))do { } while (0); | |||
| 1013 | KASSERT(sigio->sio_proc == p,do { } while (0) | |||
| 1014 | ("Bogus proc in sigio list"))do { } while (0); | |||
| 1015 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1016 | SLIST_REMOVE(&p->p_sigiolst, sigio, sigio,do { ; if ((((&p->p_sigiolst))->slh_first) == (sigio )) { do { ((((&p->p_sigiolst)))->slh_first) = ((((( (&p->p_sigiolst)))->slh_first))->sio_pgsigio.sle_next ); } while (0); } else { struct sigio *curelm = ((&p-> p_sigiolst)->slh_first); while (((curelm)->sio_pgsigio. sle_next) != (sigio)) curelm = ((curelm)->sio_pgsigio.sle_next ); do { ((curelm)->sio_pgsigio.sle_next) = ((((curelm)-> sio_pgsigio.sle_next))->sio_pgsigio.sle_next); } while (0) ; } ; } while (0) | |||
| 1017 | sio_pgsigio)do { ; if ((((&p->p_sigiolst))->slh_first) == (sigio )) { do { ((((&p->p_sigiolst)))->slh_first) = ((((( (&p->p_sigiolst)))->slh_first))->sio_pgsigio.sle_next ); } while (0); } else { struct sigio *curelm = ((&p-> p_sigiolst)->slh_first); while (((curelm)->sio_pgsigio. sle_next) != (sigio)) curelm = ((curelm)->sio_pgsigio.sle_next ); do { ((curelm)->sio_pgsigio.sle_next) = ((((curelm)-> sio_pgsigio.sle_next))->sio_pgsigio.sle_next); } while (0) ; } ; } while (0); | |||
| 1018 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1019 | } | |||
| 1020 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1021 | crfree(sigio->sio_ucred); | |||
| 1022 | free(sigio, M_SIGIO); | |||
| 1023 | SIGIO_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &sigio_lock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&sigio_lock)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1024 | } | |||
| 1025 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1026 | } | |||
| 1027 | ||||
| 1028 | /* | |||
| 1029 | * This is common code for FIOSETOWN ioctl called by fcntl(fd, F_SETOWN, arg). | |||
| 1030 | * | |||
| 1031 | * After permission checking, add a sigio structure to the sigio list for | |||
| 1032 | * the process or process group. | |||
| 1033 | */ | |||
| 1034 | int | |||
| 1035 | fsetown(pid_t pgid, struct sigio **sigiop) | |||
| 1036 | { | |||
| 1037 | struct proc *proc; | |||
| 1038 | struct pgrp *pgrp; | |||
| 1039 | struct sigio *sigio; | |||
| 1040 | int ret; | |||
| 1041 | ||||
| 1042 | if (pgid == 0) { | |||
| 1043 | funsetown(sigiop); | |||
| 1044 | return (0); | |||
| 1045 | } | |||
| 1046 | ||||
| 1047 | ret = 0; | |||
| 1048 | ||||
| 1049 | /* Allocate and fill in the new sigio out of locks. */ | |||
| 1050 | sigio = malloc(sizeof(struct sigio), M_SIGIO, M_WAITOK0x0002); | |||
| 1051 | sigio->sio_pgid = pgid; | |||
| 1052 | sigio->sio_ucred = crhold(curthread(__curthread())->td_ucred); | |||
| 1053 | sigio->sio_myref = sigiop; | |||
| 1054 | ||||
| 1055 | sx_slock(&proctree_lock)(void)__sx_slock(((&proctree_lock)), 0, (((void *)0)), (0 )); | |||
| 1056 | if (pgid > 0) { | |||
| 1057 | proc = pfind(pgid); | |||
| 1058 | if (proc == NULL((void *)0)) { | |||
| 1059 | ret = ESRCH3; | |||
| 1060 | goto fail; | |||
| 1061 | } | |||
| 1062 | ||||
| 1063 | /* | |||
| 1064 | * Policy - Don't allow a process to FSETOWN a process | |||
| 1065 | * in another session. | |||
| 1066 | * | |||
| 1067 | * Remove this test to allow maximum flexibility or | |||
| 1068 | * restrict FSETOWN to the current process or process | |||
| 1069 | * group for maximum safety. | |||
| 1070 | */ | |||
| 1071 | PROC_UNLOCK(proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (proc)->p_mtx))))->lock_object.lo_data == 0) do { (void )0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(proc)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0) ; } while (0); if (((((&(proc)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(proc)->p_mtx )))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(& (((((&(proc)->p_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1072 | if (proc->p_sessionp_pgrp->pg_session != curthread(__curthread())->td_proc->p_sessionp_pgrp->pg_session) { | |||
| 1073 | ret = EPERM1; | |||
| 1074 | goto fail; | |||
| 1075 | } | |||
| 1076 | ||||
| 1077 | pgrp = NULL((void *)0); | |||
| 1078 | } else /* if (pgid < 0) */ { | |||
| 1079 | pgrp = pgfind(-pgid); | |||
| 1080 | if (pgrp == NULL((void *)0)) { | |||
| 1081 | ret = ESRCH3; | |||
| 1082 | goto fail; | |||
| 1083 | } | |||
| 1084 | PGRP_UNLOCK(pgrp)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (pgrp)->pg_mtx))))->lock_object.lo_data == 0) do { (void )0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(pgrp)->pg_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); if (((((&(pgrp)->pg_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(pgrp)->pg_mtx )))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(& (((((&(pgrp)->pg_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1085 | ||||
| 1086 | /* | |||
| 1087 | * Policy - Don't allow a process to FSETOWN a process | |||
| 1088 | * in another session. | |||
| 1089 | * | |||
| 1090 | * Remove this test to allow maximum flexibility or | |||
| 1091 | * restrict FSETOWN to the current process or process | |||
| 1092 | * group for maximum safety. | |||
| 1093 | */ | |||
| 1094 | if (pgrp->pg_session != curthread(__curthread())->td_proc->p_sessionp_pgrp->pg_session) { | |||
| 1095 | ret = EPERM1; | |||
| 1096 | goto fail; | |||
| 1097 | } | |||
| 1098 | ||||
| 1099 | proc = NULL((void *)0); | |||
| 1100 | } | |||
| 1101 | funsetown(sigiop); | |||
| 1102 | if (pgid > 0) { | |||
| 1103 | PROC_LOCK(proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(proc)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(proc)->p_mtx)))))->mtx_lock, 0x00000004 , (_tid)))) __mtx_lock_sleep(&(((((&(proc)->p_mtx) ))))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(proc)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0) ; } while (0); } while (0); | |||
| 1104 | /* | |||
| 1105 | * Since funsetownlst() is called without the proctree | |||
| 1106 | * locked, we need to check for P_WEXIT. | |||
| 1107 | * XXX: is ESRCH correct? | |||
| 1108 | */ | |||
| 1109 | if ((proc->p_flag & P_WEXIT0x02000) != 0) { | |||
| 1110 | PROC_UNLOCK(proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (proc)->p_mtx))))->lock_object.lo_data == 0) do { (void )0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(proc)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0) ; } while (0); if (((((&(proc)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(proc)->p_mtx )))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(& (((((&(proc)->p_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1111 | ret = ESRCH3; | |||
| 1112 | goto fail; | |||
| 1113 | } | |||
| 1114 | SLIST_INSERT_HEAD(&proc->p_sigiolst, sigio, sio_pgsigio)do { (((sigio))->sio_pgsigio.sle_next) = (((&proc-> p_sigiolst))->slh_first); (((&proc->p_sigiolst))-> slh_first) = (sigio); } while (0); | |||
| 1115 | sigio->sio_procsio_u.siu_proc = proc; | |||
| 1116 | PROC_UNLOCK(proc)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (proc)->p_mtx))))->lock_object.lo_data == 0) do { (void )0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(proc)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0) ; } while (0); if (((((&(proc)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(proc)->p_mtx )))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(& (((((&(proc)->p_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1117 | } else { | |||
| 1118 | PGRP_LOCK(pgrp)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(pgrp)->pg_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(pgrp)->pg_mtx)))))->mtx_lock, 0x00000004 , (_tid)))) __mtx_lock_sleep(&(((((&(pgrp)->pg_mtx )))))->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(pgrp)->pg_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); } while (0); | |||
| 1119 | SLIST_INSERT_HEAD(&pgrp->pg_sigiolst, sigio, sio_pgsigio)do { (((sigio))->sio_pgsigio.sle_next) = (((&pgrp-> pg_sigiolst))->slh_first); (((&pgrp->pg_sigiolst))-> slh_first) = (sigio); } while (0); | |||
| 1120 | sigio->sio_pgrpsio_u.siu_pgrp = pgrp; | |||
| 1121 | PGRP_UNLOCK(pgrp)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (pgrp)->pg_mtx))))->lock_object.lo_data == 0) do { (void )0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&(pgrp)->pg_mtx))), (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0 ); } while (0); if (((((&(pgrp)->pg_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&(pgrp)->pg_mtx )))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(& (((((&(pgrp)->pg_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1122 | } | |||
| 1123 | sx_sunlock(&proctree_lock)__sx_sunlock(((&proctree_lock)), (((void *)0)), (0)); | |||
| 1124 | SIGIO_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &sigio_lock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&sigio_lock)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1125 | *sigiop = sigio; | |||
| 1126 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1127 | return (0); | |||
| 1128 | ||||
| 1129 | fail: | |||
| 1130 | sx_sunlock(&proctree_lock)__sx_sunlock(((&proctree_lock)), (((void *)0)), (0)); | |||
| 1131 | crfree(sigio->sio_ucred); | |||
| 1132 | free(sigio, M_SIGIO); | |||
| 1133 | return (ret); | |||
| 1134 | } | |||
| 1135 | ||||
| 1136 | /* | |||
| 1137 | * This is common code for FIOGETOWN ioctl called by fcntl(fd, F_GETOWN, arg). | |||
| 1138 | */ | |||
| 1139 | pid_t | |||
| 1140 | fgetown(sigiop) | |||
| 1141 | struct sigio **sigiop; | |||
| 1142 | { | |||
| 1143 | pid_t pgid; | |||
| 1144 | ||||
| 1145 | SIGIO_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &sigio_lock))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&sigio_lock)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1146 | pgid = (*sigiop != NULL((void *)0)) ? (*sigiop)->sio_pgid : 0; | |||
| 1147 | SIGIO_UNLOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& sigio_lock))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&sigio_lock))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&sigio_lock))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&sigio_lock)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&sigio_lock)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1148 | return (pgid); | |||
| 1149 | } | |||
| 1150 | ||||
| 1151 | /* | |||
| 1152 | * Function drops the filedesc lock on return. | |||
| 1153 | */ | |||
| 1154 | static int | |||
| 1155 | closefp(struct filedesc *fdp, int fd, struct file *fp, struct thread *td, | |||
| 1156 | int holdleaders) | |||
| 1157 | { | |||
| 1158 | int error; | |||
| 1159 | ||||
| 1160 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1161 | ||||
| 1162 | if (holdleaders) { | |||
| 1163 | if (td->td_proc->p_fdtol != NULL((void *)0)) { | |||
| 1164 | /* | |||
| 1165 | * Ask fdfree() to sleep to ensure that all relevant | |||
| 1166 | * process leaders can be traversed in closef(). | |||
| 1167 | */ | |||
| 1168 | fdp->fd_holdleaderscount++; | |||
| 1169 | } else { | |||
| 1170 | holdleaders = 0; | |||
| 1171 | } | |||
| 1172 | } | |||
| 1173 | ||||
| 1174 | /* | |||
| 1175 | * We now hold the fp reference that used to be owned by the | |||
| 1176 | * descriptor array. We have to unlock the FILEDESC *AFTER* | |||
| 1177 | * knote_fdclose to prevent a race of the fd getting opened, a knote | |||
| 1178 | * added, and deleteing a knote for the new fd. | |||
| 1179 | */ | |||
| 1180 | knote_fdclose(td, fd); | |||
| 1181 | ||||
| 1182 | /* | |||
| 1183 | * We need to notify mqueue if the object is of type mqueue. | |||
| 1184 | */ | |||
| 1185 | if (fp->f_type == DTYPE_MQUEUE7) | |||
| 1186 | mq_fdclose(td, fd, fp); | |||
| 1187 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 1188 | ||||
| 1189 | error = closef(fp, td); | |||
| 1190 | if (holdleaders) { | |||
| 1191 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 1192 | fdp->fd_holdleaderscount--; | |||
| 1193 | if (fdp->fd_holdleaderscount == 0 && | |||
| 1194 | fdp->fd_holdleaderswakeup != 0) { | |||
| 1195 | fdp->fd_holdleaderswakeup = 0; | |||
| 1196 | wakeup(&fdp->fd_holdleaderscount); | |||
| 1197 | } | |||
| 1198 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 1199 | } | |||
| 1200 | return (error); | |||
| 1201 | } | |||
| 1202 | ||||
| 1203 | /* | |||
| 1204 | * Close a file descriptor. | |||
| 1205 | */ | |||
| 1206 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1207 | struct close_args { | |||
| 1208 | int fd; | |||
| 1209 | }; | |||
| 1210 | #endif | |||
| 1211 | /* ARGSUSED */ | |||
| 1212 | int | |||
| 1213 | sys_close(struct thread *td, struct close_args *uap) | |||
| 1214 | { | |||
| 1215 | ||||
| 1216 | return (kern_close(td, uap->fd)); | |||
| 1217 | } | |||
| 1218 | ||||
| 1219 | int | |||
| 1220 | kern_close(struct thread *td, int fd) | |||
| 1221 | { | |||
| 1222 | struct filedesc *fdp; | |||
| 1223 | struct file *fp; | |||
| 1224 | ||||
| 1225 | fdp = td->td_proc->p_fd; | |||
| 1226 | ||||
| 1227 | AUDIT_SYSCLOSE(td, fd)do { if (td->td_pflags & 0x01000000) audit_sysclose(td , fd); } while (0); | |||
| 1228 | ||||
| 1229 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 1230 | if ((fp = fget_locked(fdp, fd)) == NULL((void *)0)) { | |||
| 1231 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 1232 | return (EBADF9); | |||
| 1233 | } | |||
| 1234 | fdfree(fdp, fd); | |||
| 1235 | ||||
| 1236 | /* closefp() drops the FILEDESC lock for us. */ | |||
| 1237 | return (closefp(fdp, fd, fp, td, 1)); | |||
| 1238 | } | |||
| 1239 | ||||
| 1240 | /* | |||
| 1241 | * Close open file descriptors. | |||
| 1242 | */ | |||
| 1243 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1244 | struct closefrom_args { | |||
| 1245 | int lowfd; | |||
| 1246 | }; | |||
| 1247 | #endif | |||
| 1248 | /* ARGSUSED */ | |||
| 1249 | int | |||
| 1250 | sys_closefrom(struct thread *td, struct closefrom_args *uap) | |||
| 1251 | { | |||
| 1252 | struct filedesc *fdp; | |||
| 1253 | int fd; | |||
| 1254 | ||||
| 1255 | fdp = td->td_proc->p_fd; | |||
| 1256 | AUDIT_ARG_FD(uap->lowfd)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_fd ((uap->lowfd)); } while (0); | |||
| 1257 | ||||
| 1258 | /* | |||
| 1259 | * Treat negative starting file descriptor values identical to | |||
| 1260 | * closefrom(0) which closes all files. | |||
| 1261 | */ | |||
| 1262 | if (uap->lowfd < 0) | |||
| 1263 | uap->lowfd = 0; | |||
| 1264 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 1265 | for (fd = uap->lowfd; fd <= fdp->fd_lastfile; fd++) { | |||
| 1266 | if (fdp->fd_ofilesfd_files->fdt_ofiles[fd].fde_file != NULL((void *)0)) { | |||
| 1267 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 1268 | (void)kern_close(td, fd); | |||
| 1269 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 1270 | } | |||
| 1271 | } | |||
| 1272 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 1273 | return (0); | |||
| 1274 | } | |||
| 1275 | ||||
| 1276 | #if defined(COMPAT_43) | |||
| 1277 | /* | |||
| 1278 | * Return status information about a file descriptor. | |||
| 1279 | */ | |||
| 1280 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1281 | struct ofstat_args { | |||
| 1282 | int fd; | |||
| 1283 | struct ostat *sb; | |||
| 1284 | }; | |||
| 1285 | #endif | |||
| 1286 | /* ARGSUSED */ | |||
| 1287 | int | |||
| 1288 | ofstat(struct thread *td, struct ofstat_args *uap) | |||
| 1289 | { | |||
| 1290 | struct ostat oub; | |||
| 1291 | struct stat ub; | |||
| 1292 | int error; | |||
| 1293 | ||||
| 1294 | error = kern_fstat(td, uap->fd, &ub); | |||
| 1295 | if (error == 0) { | |||
| 1296 | cvtstat(&ub, &oub); | |||
| 1297 | error = copyout(&oub, uap->sb, sizeof(oub)); | |||
| 1298 | } | |||
| 1299 | return (error); | |||
| 1300 | } | |||
| 1301 | #endif /* COMPAT_43 */ | |||
| 1302 | ||||
| 1303 | /* | |||
| 1304 | * Return status information about a file descriptor. | |||
| 1305 | */ | |||
| 1306 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1307 | struct fstat_args { | |||
| 1308 | int fd; | |||
| 1309 | struct stat *sb; | |||
| 1310 | }; | |||
| 1311 | #endif | |||
| 1312 | /* ARGSUSED */ | |||
| 1313 | int | |||
| 1314 | sys_fstat(struct thread *td, struct fstat_args *uap) | |||
| 1315 | { | |||
| 1316 | struct stat ub; | |||
| 1317 | int error; | |||
| 1318 | ||||
| 1319 | error = kern_fstat(td, uap->fd, &ub); | |||
| 1320 | if (error == 0) | |||
| 1321 | error = copyout(&ub, uap->sb, sizeof(ub)); | |||
| 1322 | return (error); | |||
| 1323 | } | |||
| 1324 | ||||
| 1325 | int | |||
| 1326 | kern_fstat(struct thread *td, int fd, struct stat *sbp) | |||
| 1327 | { | |||
| 1328 | struct file *fp; | |||
| 1329 | cap_rights_t rights; | |||
| 1330 | int error; | |||
| 1331 | ||||
| 1332 | AUDIT_ARG_FD(fd)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_fd ((fd)); } while (0); | |||
| 1333 | ||||
| 1334 | error = fget(td, fd, cap_rights_init(&rights, CAP_FSTAT)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000080000ULL)), 0ULL), &fp); | |||
| 1335 | if (error != 0) | |||
| 1336 | return (error); | |||
| 1337 | ||||
| 1338 | AUDIT_ARG_FILE(td->td_proc, fp)do { if ((((__curthread()))->td_pflags & 0x01000000)) audit_arg_file ((td->td_proc), (fp)); } while (0); | |||
| 1339 | ||||
| 1340 | error = fo_stat(fp, sbp, td->td_ucred, td); | |||
| 1341 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 1342 | #ifdef KTRACE1 | |||
| 1343 | if (error == 0 && KTRPOINT(td, KTR_STRUCT)((((td))->td_proc->p_traceflag & (1 << (8))) && !((td)->td_pflags & 0x00000004))) | |||
| 1344 | ktrstat(sbp)ktrstruct("stat", (sbp), sizeof(struct stat)); | |||
| 1345 | #endif | |||
| 1346 | return (error); | |||
| 1347 | } | |||
| 1348 | ||||
| 1349 | /* | |||
| 1350 | * Return status information about a file descriptor. | |||
| 1351 | */ | |||
| 1352 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1353 | struct nfstat_args { | |||
| 1354 | int fd; | |||
| 1355 | struct nstat *sb; | |||
| 1356 | }; | |||
| 1357 | #endif | |||
| 1358 | /* ARGSUSED */ | |||
| 1359 | int | |||
| 1360 | sys_nfstat(struct thread *td, struct nfstat_args *uap) | |||
| 1361 | { | |||
| 1362 | struct nstat nub; | |||
| 1363 | struct stat ub; | |||
| 1364 | int error; | |||
| 1365 | ||||
| 1366 | error = kern_fstat(td, uap->fd, &ub); | |||
| 1367 | if (error == 0) { | |||
| ||||
| 1368 | cvtnstat(&ub, &nub); | |||
| 1369 | error = copyout(&nub, uap->sb, sizeof(nub)); | |||
| ||||
| 1370 | } | |||
| 1371 | return (error); | |||
| 1372 | } | |||
| 1373 | ||||
| 1374 | /* | |||
| 1375 | * Return pathconf information about a file descriptor. | |||
| 1376 | */ | |||
| 1377 | #ifndef _SYS_SYSPROTO_H_ | |||
| 1378 | struct fpathconf_args { | |||
| 1379 | int fd; | |||
| 1380 | int name; | |||
| 1381 | }; | |||
| 1382 | #endif | |||
| 1383 | /* ARGSUSED */ | |||
| 1384 | int | |||
| 1385 | sys_fpathconf(struct thread *td, struct fpathconf_args *uap) | |||
| 1386 | { | |||
| 1387 | struct file *fp; | |||
| 1388 | struct vnode *vp; | |||
| 1389 | cap_rights_t rights; | |||
| 1390 | int error; | |||
| 1391 | ||||
| 1392 | error = fget(td, uap->fd, cap_rights_init(&rights, CAP_FPATHCONF)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000020000ULL)), 0ULL), &fp); | |||
| 1393 | if (error != 0) | |||
| 1394 | return (error); | |||
| 1395 | ||||
| 1396 | if (uap->name == _PC_ASYNC_IO53) { | |||
| 1397 | td->td_retvaltd_uretoff.tdu_retval[0] = _POSIX_ASYNCHRONOUS_IO200112L; | |||
| 1398 | goto out; | |||
| 1399 | } | |||
| 1400 | vp = fp->f_vnode; | |||
| 1401 | if (vp != NULL((void *)0)) { | |||
| 1402 | vn_lock(vp, LK_SHARED | LK_RETRY)_vn_lock(vp, 0x200000 | 0x000400, "/usr/src/sys/kern/kern_descrip.c" , 1402); | |||
| 1403 | error = VOP_PATHCONF(vp, uap->name, td->td_retvaltd_uretoff.tdu_retval); | |||
| 1404 | VOP_UNLOCK(vp, 0); | |||
| 1405 | } else if (fp->f_type == DTYPE_PIPE3 || fp->f_type == DTYPE_SOCKET2) { | |||
| 1406 | if (uap->name != _PC_PIPE_BUF6) { | |||
| 1407 | error = EINVAL22; | |||
| 1408 | } else { | |||
| 1409 | td->td_retvaltd_uretoff.tdu_retval[0] = PIPE_BUF512; | |||
| 1410 | error = 0; | |||
| 1411 | } | |||
| 1412 | } else { | |||
| 1413 | error = EOPNOTSUPP45; | |||
| 1414 | } | |||
| 1415 | out: | |||
| 1416 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 1417 | return (error); | |||
| 1418 | } | |||
| 1419 | ||||
| 1420 | /* | |||
| 1421 | * Initialize filecaps structure. | |||
| 1422 | */ | |||
| 1423 | void | |||
| 1424 | filecaps_init(struct filecaps *fcaps) | |||
| 1425 | { | |||
| 1426 | ||||
| 1427 | bzero(fcaps, sizeof(*fcaps)); | |||
| 1428 | fcaps->fc_nioctls = -1; | |||
| 1429 | } | |||
| 1430 | ||||
| 1431 | /* | |||
| 1432 | * Copy filecaps structure allocating memory for ioctls array if needed. | |||
| 1433 | * | |||
| 1434 | * The last parameter indicates whether the fdtable is locked. If it is not and | |||
| 1435 | * ioctls are encountered, copying fails and the caller must lock the table. | |||
| 1436 | * | |||
| 1437 | * Note that if the table was not locked, the caller has to check the relevant | |||
| 1438 | * sequence counter to determine whether the operation was successful. | |||
| 1439 | */ | |||
| 1440 | int | |||
| 1441 | filecaps_copy(const struct filecaps *src, struct filecaps *dst, bool locked) | |||
| 1442 | { | |||
| 1443 | size_t size; | |||
| 1444 | ||||
| 1445 | *dst = *src; | |||
| 1446 | if (src->fc_ioctls == NULL((void *)0)) | |||
| 1447 | return (0); | |||
| 1448 | if (!locked) | |||
| 1449 | return (1); | |||
| 1450 | ||||
| 1451 | KASSERT(src->fc_nioctls > 0,do { } while (0) | |||
| 1452 | ("fc_ioctls != NULL, but fc_nioctls=%hd", src->fc_nioctls))do { } while (0); | |||
| 1453 | ||||
| 1454 | size = sizeof(src->fc_ioctls[0]) * src->fc_nioctls; | |||
| 1455 | dst->fc_ioctls = malloc(size, M_FILECAPS, M_WAITOK0x0002); | |||
| 1456 | bcopy(src->fc_ioctls, dst->fc_ioctls, size); | |||
| 1457 | return (0); | |||
| 1458 | } | |||
| 1459 | ||||
| 1460 | /* | |||
| 1461 | * Move filecaps structure to the new place and clear the old place. | |||
| 1462 | */ | |||
| 1463 | void | |||
| 1464 | filecaps_move(struct filecaps *src, struct filecaps *dst) | |||
| 1465 | { | |||
| 1466 | ||||
| 1467 | *dst = *src; | |||
| 1468 | bzero(src, sizeof(*src)); | |||
| 1469 | } | |||
| 1470 | ||||
| 1471 | /* | |||
| 1472 | * Fill the given filecaps structure with full rights. | |||
| 1473 | */ | |||
| 1474 | static void | |||
| 1475 | filecaps_fill(struct filecaps *fcaps) | |||
| 1476 | { | |||
| 1477 | ||||
| 1478 | CAP_ALL(&fcaps->fc_rights)do { (&fcaps->fc_rights)->cr_rights[0] = ((uint64_t )0 << 62) | ((1ULL << (57 + (0))) | (0x000007FFFFFFFFFFULL )); (&fcaps->fc_rights)->cr_rights[1] = ((1ULL << (57 + (1))) | (0x00000000001FFFFFULL)); } while (0); | |||
| 1479 | fcaps->fc_ioctls = NULL((void *)0); | |||
| 1480 | fcaps->fc_nioctls = -1; | |||
| 1481 | fcaps->fc_fcntls = CAP_FCNTL_ALL((1 << 3) | (1 << 4) | (1 << 5) | (1 << 6)); | |||
| 1482 | } | |||
| 1483 | ||||
| 1484 | /* | |||
| 1485 | * Free memory allocated within filecaps structure. | |||
| 1486 | */ | |||
| 1487 | void | |||
| 1488 | filecaps_free(struct filecaps *fcaps) | |||
| 1489 | { | |||
| 1490 | ||||
| 1491 | free(fcaps->fc_ioctls, M_FILECAPS); | |||
| 1492 | bzero(fcaps, sizeof(*fcaps)); | |||
| 1493 | } | |||
| 1494 | ||||
| 1495 | /* | |||
| 1496 | * Validate the given filecaps structure. | |||
| 1497 | */ | |||
| 1498 | static void | |||
| 1499 | filecaps_validate(const struct filecaps *fcaps, const char *func) | |||
| 1500 | { | |||
| 1501 | ||||
| 1502 | KASSERT(cap_rights_is_valid(&fcaps->fc_rights),do { } while (0) | |||
| 1503 | ("%s: invalid rights", func))do { } while (0); | |||
| 1504 | KASSERT((fcaps->fc_fcntls & ~CAP_FCNTL_ALL) == 0,do { } while (0) | |||
| 1505 | ("%s: invalid fcntls", func))do { } while (0); | |||
| 1506 | KASSERT(fcaps->fc_fcntls == 0 ||do { } while (0) | |||
| 1507 | cap_rights_is_set(&fcaps->fc_rights, CAP_FCNTL),do { } while (0) | |||
| 1508 | ("%s: fcntls without CAP_FCNTL", func))do { } while (0); | |||
| 1509 | KASSERT(fcaps->fc_ioctls != NULL ? fcaps->fc_nioctls > 0 :do { } while (0) | |||
| 1510 | (fcaps->fc_nioctls == -1 || fcaps->fc_nioctls == 0),do { } while (0) | |||
| 1511 | ("%s: invalid ioctls", func))do { } while (0); | |||
| 1512 | KASSERT(fcaps->fc_nioctls == 0 ||do { } while (0) | |||
| 1513 | cap_rights_is_set(&fcaps->fc_rights, CAP_IOCTL),do { } while (0) | |||
| 1514 | ("%s: ioctls without CAP_IOCTL", func))do { } while (0); | |||
| 1515 | } | |||
| 1516 | ||||
| 1517 | static void | |||
| 1518 | fdgrowtable_exp(struct filedesc *fdp, int nfd) | |||
| 1519 | { | |||
| 1520 | int nfd1; | |||
| 1521 | ||||
| 1522 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1523 | ||||
| 1524 | nfd1 = fdp->fd_nfilesfd_files->fdt_nfiles * 2; | |||
| 1525 | if (nfd1 < nfd) | |||
| 1526 | nfd1 = nfd; | |||
| 1527 | fdgrowtable(fdp, nfd1); | |||
| 1528 | } | |||
| 1529 | ||||
| 1530 | /* | |||
| 1531 | * Grow the file table to accommodate (at least) nfd descriptors. | |||
| 1532 | */ | |||
| 1533 | static void | |||
| 1534 | fdgrowtable(struct filedesc *fdp, int nfd) | |||
| 1535 | { | |||
| 1536 | struct filedesc0 *fdp0; | |||
| 1537 | struct freetable *ft; | |||
| 1538 | struct fdescenttbl *ntable; | |||
| 1539 | struct fdescenttbl *otable; | |||
| 1540 | int nnfiles, onfiles; | |||
| 1541 | NDSLOTTYPEu_long *nmap, *omap; | |||
| 1542 | ||||
| 1543 | /* | |||
| 1544 | * If lastfile is -1 this struct filedesc was just allocated and we are | |||
| 1545 | * growing it to accommodate for the one we are going to copy from. There | |||
| 1546 | * is no need to have a lock on this one as it's not visible to anyone. | |||
| 1547 | */ | |||
| 1548 | if (fdp->fd_lastfile != -1) | |||
| 1549 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1550 | ||||
| 1551 | KASSERT(fdp->fd_nfiles > 0, ("zero-length file table"))do { } while (0); | |||
| 1552 | ||||
| 1553 | /* save old values */ | |||
| 1554 | onfiles = fdp->fd_nfilesfd_files->fdt_nfiles; | |||
| 1555 | otable = fdp->fd_files; | |||
| 1556 | omap = fdp->fd_map; | |||
| 1557 | ||||
| 1558 | /* compute the size of the new table */ | |||
| 1559 | nnfiles = NDSLOTS(nfd)(((nfd) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8)) * NDENTRIES(sizeof(u_long) * 8); /* round up */ | |||
| 1560 | if (nnfiles <= onfiles) | |||
| 1561 | /* the table is already large enough */ | |||
| 1562 | return; | |||
| 1563 | ||||
| 1564 | /* | |||
| 1565 | * Allocate a new table. We need enough space for the number of | |||
| 1566 | * entries, file entries themselves and the struct freetable we will use | |||
| 1567 | * when we decommission the table and place it on the freelist. | |||
| 1568 | * We place the struct freetable in the middle so we don't have | |||
| 1569 | * to worry about padding. | |||
| 1570 | */ | |||
| 1571 | ntable = malloc(offsetof(struct fdescenttbl, fdt_ofiles)__builtin_offsetof(struct fdescenttbl, fdt_ofiles) + | |||
| 1572 | nnfiles * sizeof(ntable->fdt_ofiles[0]) + | |||
| 1573 | sizeof(struct freetable), | |||
| 1574 | M_FILEDESC, M_ZERO0x0100 | M_WAITOK0x0002); | |||
| 1575 | /* copy the old data */ | |||
| 1576 | ntable->fdt_nfiles = nnfiles; | |||
| 1577 | memcpy(ntable->fdt_ofiles, otable->fdt_ofiles, | |||
| 1578 | onfiles * sizeof(ntable->fdt_ofiles[0])); | |||
| 1579 | ||||
| 1580 | /* | |||
| 1581 | * Allocate a new map only if the old is not large enough. It will | |||
| 1582 | * grow at a slower rate than the table as it can map more | |||
| 1583 | * entries than the table can hold. | |||
| 1584 | */ | |||
| 1585 | if (NDSLOTS(nnfiles)(((nnfiles) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8 )) > NDSLOTS(onfiles)(((onfiles) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8 ))) { | |||
| 1586 | nmap = malloc(NDSLOTS(nnfiles)(((nnfiles) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8 )) * NDSLOTSIZEsizeof(u_long), M_FILEDESC, | |||
| 1587 | M_ZERO0x0100 | M_WAITOK0x0002); | |||
| 1588 | /* copy over the old data and update the pointer */ | |||
| 1589 | memcpy(nmap, omap, NDSLOTS(onfiles)(((onfiles) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8 )) * sizeof(*omap)); | |||
| 1590 | fdp->fd_map = nmap; | |||
| 1591 | } | |||
| 1592 | ||||
| 1593 | /* | |||
| 1594 | * Make sure that ntable is correctly initialized before we replace | |||
| 1595 | * fd_files poiner. Otherwise fget_unlocked() may see inconsistent | |||
| 1596 | * data. | |||
| 1597 | */ | |||
| 1598 | atomic_store_rel_ptratomic_store_rel_long((volatile void *)&fdp->fd_files, (uintptr_t)ntable); | |||
| 1599 | ||||
| 1600 | /* | |||
| 1601 | * Do not free the old file table, as some threads may still | |||
| 1602 | * reference entries within it. Instead, place it on a freelist | |||
| 1603 | * which will be processed when the struct filedesc is released. | |||
| 1604 | * | |||
| 1605 | * Note that if onfiles == NDFILE, we're dealing with the original | |||
| 1606 | * static allocation contained within (struct filedesc0 *)fdp, | |||
| 1607 | * which must not be freed. | |||
| 1608 | */ | |||
| 1609 | if (onfiles > NDFILE20) { | |||
| 1610 | ft = (struct freetable *)&otable->fdt_ofiles[onfiles]; | |||
| 1611 | fdp0 = (struct filedesc0 *)fdp; | |||
| 1612 | ft->ft_table = otable; | |||
| 1613 | SLIST_INSERT_HEAD(&fdp0->fd_free, ft, ft_next)do { (((ft))->ft_next.sle_next) = (((&fdp0->fd_free ))->slh_first); (((&fdp0->fd_free))->slh_first) = (ft); } while (0); | |||
| 1614 | } | |||
| 1615 | /* | |||
| 1616 | * The map does not have the same possibility of threads still | |||
| 1617 | * holding references to it. So always free it as long as it | |||
| 1618 | * does not reference the original static allocation. | |||
| 1619 | */ | |||
| 1620 | if (NDSLOTS(onfiles)(((onfiles) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8 )) > NDSLOTS(NDFILE)(((20) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8))) | |||
| 1621 | free(omap, M_FILEDESC); | |||
| 1622 | } | |||
| 1623 | ||||
| 1624 | /* | |||
| 1625 | * Allocate a file descriptor for the process. | |||
| 1626 | */ | |||
| 1627 | int | |||
| 1628 | fdalloc(struct thread *td, int minfd, int *result) | |||
| 1629 | { | |||
| 1630 | struct proc *p = td->td_proc; | |||
| 1631 | struct filedesc *fdp = p->p_fd; | |||
| 1632 | int fd, maxfd, allocfd; | |||
| 1633 | #ifdef RACCT1 | |||
| 1634 | int error; | |||
| 1635 | #endif | |||
| 1636 | ||||
| 1637 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1638 | ||||
| 1639 | if (fdp->fd_freefile > minfd) | |||
| 1640 | minfd = fdp->fd_freefile; | |||
| 1641 | ||||
| 1642 | maxfd = getmaxfd(td); | |||
| 1643 | ||||
| 1644 | /* | |||
| 1645 | * Search the bitmap for a free descriptor starting at minfd. | |||
| 1646 | * If none is found, grow the file table. | |||
| 1647 | */ | |||
| 1648 | fd = fd_first_free(fdp, minfd, fdp->fd_nfilesfd_files->fdt_nfiles); | |||
| 1649 | if (fd >= maxfd) | |||
| 1650 | return (EMFILE24); | |||
| 1651 | if (fd >= fdp->fd_nfilesfd_files->fdt_nfiles) { | |||
| 1652 | allocfd = min(fd * 2, maxfd); | |||
| 1653 | #ifdef RACCT1 | |||
| 1654 | if (racct_enable) { | |||
| 1655 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 1656 | error = racct_set(p, RACCT_NOFILE7, allocfd); | |||
| 1657 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 1658 | if (error != 0) | |||
| 1659 | return (EMFILE24); | |||
| 1660 | } | |||
| 1661 | #endif | |||
| 1662 | /* | |||
| 1663 | * fd is already equal to first free descriptor >= minfd, so | |||
| 1664 | * we only need to grow the table and we are done. | |||
| 1665 | */ | |||
| 1666 | fdgrowtable_exp(fdp, allocfd); | |||
| 1667 | } | |||
| 1668 | ||||
| 1669 | /* | |||
| 1670 | * Perform some sanity checks, then mark the file descriptor as | |||
| 1671 | * used and return it to the caller. | |||
| 1672 | */ | |||
| 1673 | KASSERT(fd >= 0 && fd < min(maxfd, fdp->fd_nfiles),do { } while (0) | |||
| 1674 | ("invalid descriptor %d", fd))do { } while (0); | |||
| 1675 | KASSERT(!fdisused(fdp, fd),do { } while (0) | |||
| 1676 | ("fd_first_free() returned non-free descriptor"))do { } while (0); | |||
| 1677 | KASSERT(fdp->fd_ofiles[fd].fde_file == NULL,do { } while (0) | |||
| 1678 | ("file descriptor isn't free"))do { } while (0); | |||
| 1679 | fdused(fdp, fd); | |||
| 1680 | *result = fd; | |||
| 1681 | return (0); | |||
| 1682 | } | |||
| 1683 | ||||
| 1684 | /* | |||
| 1685 | * Allocate n file descriptors for the process. | |||
| 1686 | */ | |||
| 1687 | int | |||
| 1688 | fdallocn(struct thread *td, int minfd, int *fds, int n) | |||
| 1689 | { | |||
| 1690 | struct proc *p = td->td_proc; | |||
| 1691 | struct filedesc *fdp = p->p_fd; | |||
| 1692 | int i; | |||
| 1693 | ||||
| 1694 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1695 | ||||
| 1696 | for (i = 0; i < n; i++) | |||
| 1697 | if (fdalloc(td, 0, &fds[i]) != 0) | |||
| 1698 | break; | |||
| 1699 | ||||
| 1700 | if (i < n) { | |||
| 1701 | for (i--; i >= 0; i--) | |||
| 1702 | fdunused(fdp, fds[i]); | |||
| 1703 | return (EMFILE24); | |||
| 1704 | } | |||
| 1705 | ||||
| 1706 | return (0); | |||
| 1707 | } | |||
| 1708 | ||||
| 1709 | /* | |||
| 1710 | * Create a new open file structure and allocate a file descriptor for the | |||
| 1711 | * process that refers to it. We add one reference to the file for the | |||
| 1712 | * descriptor table and one reference for resultfp. This is to prevent us | |||
| 1713 | * being preempted and the entry in the descriptor table closed after we | |||
| 1714 | * release the FILEDESC lock. | |||
| 1715 | */ | |||
| 1716 | int | |||
| 1717 | falloc_caps(struct thread *td, struct file **resultfp, int *resultfd, int flags, | |||
| 1718 | struct filecaps *fcaps) | |||
| 1719 | { | |||
| 1720 | struct file *fp; | |||
| 1721 | int error, fd; | |||
| 1722 | ||||
| 1723 | error = falloc_noinstall(td, &fp); | |||
| 1724 | if (error) | |||
| 1725 | return (error); /* no reference held on error */ | |||
| 1726 | ||||
| 1727 | error = finstall(td, fp, &fd, flags, fcaps); | |||
| 1728 | if (error) { | |||
| 1729 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); /* one reference (fp only) */ | |||
| 1730 | return (error); | |||
| 1731 | } | |||
| 1732 | ||||
| 1733 | if (resultfp != NULL((void *)0)) | |||
| 1734 | *resultfp = fp; /* copy out result */ | |||
| 1735 | else | |||
| 1736 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); /* release local reference */ | |||
| 1737 | ||||
| 1738 | if (resultfd != NULL((void *)0)) | |||
| 1739 | *resultfd = fd; | |||
| 1740 | ||||
| 1741 | return (0); | |||
| 1742 | } | |||
| 1743 | ||||
| 1744 | /* | |||
| 1745 | * Create a new open file structure without allocating a file descriptor. | |||
| 1746 | */ | |||
| 1747 | int | |||
| 1748 | falloc_noinstall(struct thread *td, struct file **resultfp) | |||
| 1749 | { | |||
| 1750 | struct file *fp; | |||
| 1751 | int maxuserfiles = maxfiles - (maxfiles / 20); | |||
| 1752 | static struct timeval lastfail; | |||
| 1753 | static int curfail; | |||
| 1754 | ||||
| 1755 | KASSERT(resultfp != NULL, ("%s: resultfp == NULL", __func__))do { } while (0); | |||
| 1756 | ||||
| 1757 | if ((openfiles >= maxuserfiles && | |||
| 1758 | priv_check(td, PRIV_MAXFILES3) != 0) || | |||
| 1759 | openfiles >= maxfiles) { | |||
| 1760 | if (ppsratecheck(&lastfail, &curfail, 1)) { | |||
| 1761 | printf("kern.maxfiles limit exceeded by uid %i, " | |||
| 1762 | "please see tuning(7).\n", td->td_ucred->cr_ruid); | |||
| 1763 | } | |||
| 1764 | return (ENFILE23); | |||
| 1765 | } | |||
| 1766 | atomic_add_int(&openfiles, 1); | |||
| 1767 | fp = uma_zalloc(file_zone, M_WAITOK0x0002 | M_ZERO0x0100); | |||
| 1768 | refcount_init(&fp->f_count, 1); | |||
| 1769 | fp->f_cred = crhold(td->td_ucred); | |||
| 1770 | fp->f_ops = &badfileops; | |||
| 1771 | *resultfp = fp; | |||
| 1772 | return (0); | |||
| 1773 | } | |||
| 1774 | ||||
| 1775 | /* | |||
| 1776 | * Install a file in a file descriptor table. | |||
| 1777 | */ | |||
| 1778 | void | |||
| 1779 | _finstall(struct filedesc *fdp, struct file *fp, int fd, int flags, | |||
| 1780 | struct filecaps *fcaps) | |||
| 1781 | { | |||
| 1782 | struct filedescent *fde; | |||
| 1783 | ||||
| 1784 | MPASS(fp != NULL)do { } while (0); | |||
| 1785 | if (fcaps != NULL((void *)0)) | |||
| 1786 | filecaps_validate(fcaps, __func__); | |||
| 1787 | FILEDESC_XLOCK_ASSERT(fdp)(void)0; | |||
| 1788 | ||||
| 1789 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[fd]; | |||
| 1790 | #ifdef CAPABILITIES1 | |||
| 1791 | seq_write_begin(&fde->fde_seq); | |||
| 1792 | #endif | |||
| 1793 | fde->fde_file = fp; | |||
| 1794 | fde->fde_flags = (flags & O_CLOEXEC0x00100000) != 0 ? UF_EXCLOSE0x01 : 0; | |||
| 1795 | if (fcaps != NULL((void *)0)) | |||
| 1796 | filecaps_move(fcaps, &fde->fde_caps); | |||
| 1797 | else | |||
| 1798 | filecaps_fill(&fde->fde_caps); | |||
| 1799 | #ifdef CAPABILITIES1 | |||
| 1800 | seq_write_end(&fde->fde_seq); | |||
| 1801 | #endif | |||
| 1802 | } | |||
| 1803 | ||||
| 1804 | int | |||
| 1805 | finstall(struct thread *td, struct file *fp, int *fd, int flags, | |||
| 1806 | struct filecaps *fcaps) | |||
| 1807 | { | |||
| 1808 | struct filedesc *fdp = td->td_proc->p_fd; | |||
| 1809 | int error; | |||
| 1810 | ||||
| 1811 | MPASS(fd != NULL)do { } while (0); | |||
| 1812 | ||||
| 1813 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 1814 | if ((error = fdalloc(td, 0, fd))) { | |||
| 1815 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 1816 | return (error); | |||
| 1817 | } | |||
| 1818 | fhold(fp)(refcount_acquire(&(fp)->f_count)); | |||
| 1819 | _finstall(fdp, fp, *fd, flags, fcaps); | |||
| 1820 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 1821 | return (0); | |||
| 1822 | } | |||
| 1823 | ||||
| 1824 | /* | |||
| 1825 | * Build a new filedesc structure from another. | |||
| 1826 | * Copy the current, root, and jail root vnode references. | |||
| 1827 | * | |||
| 1828 | * If fdp is not NULL, return with it shared locked. | |||
| 1829 | */ | |||
| 1830 | struct filedesc * | |||
| 1831 | fdinit(struct filedesc *fdp, bool prepfiles) | |||
| 1832 | { | |||
| 1833 | struct filedesc0 *newfdp0; | |||
| 1834 | struct filedesc *newfdp; | |||
| 1835 | ||||
| 1836 | newfdp0 = uma_zalloc(filedesc0_zone, M_WAITOK0x0002 | M_ZERO0x0100); | |||
| 1837 | newfdp = &newfdp0->fd_fd; | |||
| 1838 | ||||
| 1839 | /* Create the file descriptor table. */ | |||
| 1840 | FILEDESC_LOCK_INIT(newfdp)sx_init_flags((&(newfdp)->fd_sx), ("filedesc structure" ), 0); | |||
| 1841 | refcount_init(&newfdp->fd_refcnt, 1); | |||
| 1842 | refcount_init(&newfdp->fd_holdcnt, 1); | |||
| 1843 | newfdp->fd_cmask = CMASK022; | |||
| 1844 | newfdp->fd_map = newfdp0->fd_dmap; | |||
| 1845 | newfdp->fd_lastfile = -1; | |||
| 1846 | newfdp->fd_files = (struct fdescenttbl *)&newfdp0->fd_dfiles; | |||
| 1847 | newfdp->fd_files->fdt_nfiles = NDFILE20; | |||
| 1848 | ||||
| 1849 | if (fdp == NULL((void *)0)) | |||
| 1850 | return (newfdp); | |||
| 1851 | ||||
| 1852 | if (prepfiles && fdp->fd_lastfile >= newfdp->fd_nfilesfd_files->fdt_nfiles) | |||
| 1853 | fdgrowtable(newfdp, fdp->fd_lastfile + 1); | |||
| 1854 | ||||
| 1855 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 1856 | newfdp->fd_cdir = fdp->fd_cdir; | |||
| 1857 | if (newfdp->fd_cdir) | |||
| 1858 | VREF(newfdp->fd_cdir)vref(newfdp->fd_cdir); | |||
| 1859 | newfdp->fd_rdir = fdp->fd_rdir; | |||
| 1860 | if (newfdp->fd_rdir) | |||
| 1861 | VREF(newfdp->fd_rdir)vref(newfdp->fd_rdir); | |||
| 1862 | newfdp->fd_jdir = fdp->fd_jdir; | |||
| 1863 | if (newfdp->fd_jdir) | |||
| 1864 | VREF(newfdp->fd_jdir)vref(newfdp->fd_jdir); | |||
| 1865 | ||||
| 1866 | if (!prepfiles) { | |||
| 1867 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 1868 | } else { | |||
| 1869 | while (fdp->fd_lastfile >= newfdp->fd_nfilesfd_files->fdt_nfiles) { | |||
| 1870 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 1871 | fdgrowtable(newfdp, fdp->fd_lastfile + 1); | |||
| 1872 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 1873 | } | |||
| 1874 | } | |||
| 1875 | ||||
| 1876 | return (newfdp); | |||
| 1877 | } | |||
| 1878 | ||||
| 1879 | static struct filedesc * | |||
| 1880 | fdhold(struct proc *p) | |||
| 1881 | { | |||
| 1882 | struct filedesc *fdp; | |||
| 1883 | ||||
| 1884 | PROC_LOCK_ASSERT(p, MA_OWNED)(void)0; | |||
| 1885 | fdp = p->p_fd; | |||
| 1886 | if (fdp != NULL((void *)0)) | |||
| 1887 | refcount_acquire(&fdp->fd_holdcnt); | |||
| 1888 | return (fdp); | |||
| 1889 | } | |||
| 1890 | ||||
| 1891 | static void | |||
| 1892 | fddrop(struct filedesc *fdp) | |||
| 1893 | { | |||
| 1894 | ||||
| 1895 | if (fdp->fd_holdcnt > 1) { | |||
| 1896 | if (refcount_release(&fdp->fd_holdcnt) == 0) | |||
| 1897 | return; | |||
| 1898 | } | |||
| 1899 | ||||
| 1900 | FILEDESC_LOCK_DESTROY(fdp)sx_destroy(&(fdp)->fd_sx); | |||
| 1901 | uma_zfree(filedesc0_zone, fdp); | |||
| 1902 | } | |||
| 1903 | ||||
| 1904 | /* | |||
| 1905 | * Share a filedesc structure. | |||
| 1906 | */ | |||
| 1907 | struct filedesc * | |||
| 1908 | fdshare(struct filedesc *fdp) | |||
| 1909 | { | |||
| 1910 | ||||
| 1911 | refcount_acquire(&fdp->fd_refcnt); | |||
| 1912 | return (fdp); | |||
| 1913 | } | |||
| 1914 | ||||
| 1915 | /* | |||
| 1916 | * Unshare a filedesc structure, if necessary by making a copy | |||
| 1917 | */ | |||
| 1918 | void | |||
| 1919 | fdunshare(struct thread *td) | |||
| 1920 | { | |||
| 1921 | struct filedesc *tmp; | |||
| 1922 | struct proc *p = td->td_proc; | |||
| 1923 | ||||
| 1924 | if (p->p_fd->fd_refcnt == 1) | |||
| 1925 | return; | |||
| 1926 | ||||
| 1927 | tmp = fdcopy(p->p_fd); | |||
| 1928 | fdescfree(td); | |||
| 1929 | p->p_fd = tmp; | |||
| 1930 | } | |||
| 1931 | ||||
| 1932 | void | |||
| 1933 | fdinstall_remapped(struct thread *td, struct filedesc *fdp) | |||
| 1934 | { | |||
| 1935 | ||||
| 1936 | fdescfree(td); | |||
| 1937 | td->td_proc->p_fd = fdp; | |||
| 1938 | } | |||
| 1939 | ||||
| 1940 | /* | |||
| 1941 | * Copy a filedesc structure. A NULL pointer in returns a NULL reference, | |||
| 1942 | * this is to ease callers, not catch errors. | |||
| 1943 | */ | |||
| 1944 | struct filedesc * | |||
| 1945 | fdcopy(struct filedesc *fdp) | |||
| 1946 | { | |||
| 1947 | struct filedesc *newfdp; | |||
| 1948 | struct filedescent *nfde, *ofde; | |||
| 1949 | int i; | |||
| 1950 | ||||
| 1951 | MPASS(fdp != NULL)do { } while (0); | |||
| 1952 | ||||
| 1953 | newfdp = fdinit(fdp, true1); | |||
| 1954 | /* copy all passable descriptors (i.e. not kqueue) */ | |||
| 1955 | newfdp->fd_freefile = -1; | |||
| 1956 | for (i = 0; i <= fdp->fd_lastfile; ++i) { | |||
| 1957 | ofde = &fdp->fd_ofilesfd_files->fdt_ofiles[i]; | |||
| 1958 | if (ofde->fde_file == NULL((void *)0) || | |||
| 1959 | (ofde->fde_file->f_ops->fo_flags & DFLAG_PASSABLE0x01) == 0) { | |||
| 1960 | if (newfdp->fd_freefile == -1) | |||
| 1961 | newfdp->fd_freefile = i; | |||
| 1962 | continue; | |||
| 1963 | } | |||
| 1964 | nfde = &newfdp->fd_ofilesfd_files->fdt_ofiles[i]; | |||
| 1965 | *nfde = *ofde; | |||
| 1966 | filecaps_copy(&ofde->fde_caps, &nfde->fde_caps, true1); | |||
| 1967 | fhold(nfde->fde_file)(refcount_acquire(&(nfde->fde_file)->f_count)); | |||
| 1968 | fdused_init(newfdp, i); | |||
| 1969 | newfdp->fd_lastfile = i; | |||
| 1970 | } | |||
| 1971 | if (newfdp->fd_freefile == -1) | |||
| 1972 | newfdp->fd_freefile = i; | |||
| 1973 | newfdp->fd_cmask = fdp->fd_cmask; | |||
| 1974 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 1975 | return (newfdp); | |||
| 1976 | } | |||
| 1977 | ||||
| 1978 | /* | |||
| 1979 | * Copies a filedesc structure, while remapping all file descriptors | |||
| 1980 | * stored inside using a translation table. | |||
| 1981 | * | |||
| 1982 | * File descriptors are copied over to the new file descriptor table, | |||
| 1983 | * regardless of whether the close-on-exec flag is set. | |||
| 1984 | */ | |||
| 1985 | int | |||
| 1986 | fdcopy_remapped(struct filedesc *fdp, const int *fds, size_t nfds, | |||
| 1987 | struct filedesc **ret) | |||
| 1988 | { | |||
| 1989 | struct filedesc *newfdp; | |||
| 1990 | struct filedescent *nfde, *ofde; | |||
| 1991 | int error, i; | |||
| 1992 | ||||
| 1993 | MPASS(fdp != NULL)do { } while (0); | |||
| 1994 | ||||
| 1995 | newfdp = fdinit(fdp, true1); | |||
| 1996 | if (nfds > fdp->fd_lastfile + 1) { | |||
| 1997 | /* New table cannot be larger than the old one. */ | |||
| 1998 | error = E2BIG7; | |||
| 1999 | goto bad; | |||
| 2000 | } | |||
| 2001 | /* Copy all passable descriptors (i.e. not kqueue). */ | |||
| 2002 | newfdp->fd_freefile = nfds; | |||
| 2003 | for (i = 0; i < nfds; ++i) { | |||
| 2004 | if (fds[i] < 0 || fds[i] > fdp->fd_lastfile) { | |||
| 2005 | /* File descriptor out of bounds. */ | |||
| 2006 | error = EBADF9; | |||
| 2007 | goto bad; | |||
| 2008 | } | |||
| 2009 | ofde = &fdp->fd_ofilesfd_files->fdt_ofiles[fds[i]]; | |||
| 2010 | if (ofde->fde_file == NULL((void *)0)) { | |||
| 2011 | /* Unused file descriptor. */ | |||
| 2012 | error = EBADF9; | |||
| 2013 | goto bad; | |||
| 2014 | } | |||
| 2015 | if ((ofde->fde_file->f_ops->fo_flags & DFLAG_PASSABLE0x01) == 0) { | |||
| 2016 | /* File descriptor cannot be passed. */ | |||
| 2017 | error = EINVAL22; | |||
| 2018 | goto bad; | |||
| 2019 | } | |||
| 2020 | nfde = &newfdp->fd_ofilesfd_files->fdt_ofiles[i]; | |||
| 2021 | *nfde = *ofde; | |||
| 2022 | filecaps_copy(&ofde->fde_caps, &nfde->fde_caps, true1); | |||
| 2023 | fhold(nfde->fde_file)(refcount_acquire(&(nfde->fde_file)->f_count)); | |||
| 2024 | fdused_init(newfdp, i); | |||
| 2025 | newfdp->fd_lastfile = i; | |||
| 2026 | } | |||
| 2027 | newfdp->fd_cmask = fdp->fd_cmask; | |||
| 2028 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 2029 | *ret = newfdp; | |||
| 2030 | return (0); | |||
| 2031 | bad: | |||
| 2032 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 2033 | fdescfree_remapped(newfdp); | |||
| 2034 | return (error); | |||
| 2035 | } | |||
| 2036 | ||||
| 2037 | /* | |||
| 2038 | * Clear POSIX style locks. This is only used when fdp looses a reference (i.e. | |||
| 2039 | * one of processes using it exits) and the table used to be shared. | |||
| 2040 | */ | |||
| 2041 | static void | |||
| 2042 | fdclearlocks(struct thread *td) | |||
| 2043 | { | |||
| 2044 | struct filedesc *fdp; | |||
| 2045 | struct filedesc_to_leader *fdtol; | |||
| 2046 | struct flock lf; | |||
| 2047 | struct file *fp; | |||
| 2048 | struct proc *p; | |||
| 2049 | struct vnode *vp; | |||
| 2050 | int i; | |||
| 2051 | ||||
| 2052 | p = td->td_proc; | |||
| 2053 | fdp = p->p_fd; | |||
| 2054 | fdtol = p->p_fdtol; | |||
| 2055 | MPASS(fdtol != NULL)do { } while (0); | |||
| 2056 | ||||
| 2057 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2058 | KASSERT(fdtol->fdl_refcount > 0,do { } while (0) | |||
| 2059 | ("filedesc_to_refcount botch: fdl_refcount=%d",do { } while (0) | |||
| 2060 | fdtol->fdl_refcount))do { } while (0); | |||
| 2061 | if (fdtol->fdl_refcount == 1 && | |||
| 2062 | (p->p_leader->p_flag & P_ADVLOCK0x00001) != 0) { | |||
| 2063 | for (i = 0; i <= fdp->fd_lastfile; i++) { | |||
| 2064 | fp = fdp->fd_ofilesfd_files->fdt_ofiles[i].fde_file; | |||
| 2065 | if (fp == NULL((void *)0) || fp->f_type != DTYPE_VNODE1) | |||
| 2066 | continue; | |||
| 2067 | fhold(fp)(refcount_acquire(&(fp)->f_count)); | |||
| 2068 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2069 | lf.l_whence = SEEK_SET0; | |||
| 2070 | lf.l_start = 0; | |||
| 2071 | lf.l_len = 0; | |||
| 2072 | lf.l_type = F_UNLCK2; | |||
| 2073 | vp = fp->f_vnode; | |||
| 2074 | (void) VOP_ADVLOCK(vp, | |||
| 2075 | (caddr_t)p->p_leader, F_UNLCK2, | |||
| 2076 | &lf, F_POSIX0x040); | |||
| 2077 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2078 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2079 | } | |||
| 2080 | } | |||
| 2081 | retry: | |||
| 2082 | if (fdtol->fdl_refcount == 1) { | |||
| 2083 | if (fdp->fd_holdleaderscount > 0 && | |||
| 2084 | (p->p_leader->p_flag & P_ADVLOCK0x00001) != 0) { | |||
| 2085 | /* | |||
| 2086 | * close() or kern_dup() has cleared a reference | |||
| 2087 | * in a shared file descriptor table. | |||
| 2088 | */ | |||
| 2089 | fdp->fd_holdleaderswakeup = 1; | |||
| 2090 | sx_sleep(&fdp->fd_holdleaderscount,_sleep((&fdp->fd_holdleaderscount), &((&(fdp)-> fd_sx))->lock_object, (((80) + 32)), ("fdlhold"), tick_sbt * (0), 0, 0x0100) | |||
| 2091 | FILEDESC_LOCK(fdp), PLOCK, "fdlhold", 0)_sleep((&fdp->fd_holdleaderscount), &((&(fdp)-> fd_sx))->lock_object, (((80) + 32)), ("fdlhold"), tick_sbt * (0), 0, 0x0100); | |||
| 2092 | goto retry; | |||
| 2093 | } | |||
| 2094 | if (fdtol->fdl_holdcount > 0) { | |||
| 2095 | /* | |||
| 2096 | * Ensure that fdtol->fdl_leader remains | |||
| 2097 | * valid in closef(). | |||
| 2098 | */ | |||
| 2099 | fdtol->fdl_wakeup = 1; | |||
| 2100 | sx_sleep(fdtol, FILEDESC_LOCK(fdp), PLOCK,_sleep((fdtol), &((&(fdp)->fd_sx))->lock_object , (((80) + 32)), ("fdlhold"), tick_sbt * (0), 0, 0x0100) | |||
| 2101 | "fdlhold", 0)_sleep((fdtol), &((&(fdp)->fd_sx))->lock_object , (((80) + 32)), ("fdlhold"), tick_sbt * (0), 0, 0x0100); | |||
| 2102 | goto retry; | |||
| 2103 | } | |||
| 2104 | } | |||
| 2105 | fdtol->fdl_refcount--; | |||
| 2106 | if (fdtol->fdl_refcount == 0 && | |||
| 2107 | fdtol->fdl_holdcount == 0) { | |||
| 2108 | fdtol->fdl_next->fdl_prev = fdtol->fdl_prev; | |||
| 2109 | fdtol->fdl_prev->fdl_next = fdtol->fdl_next; | |||
| 2110 | } else | |||
| 2111 | fdtol = NULL((void *)0); | |||
| 2112 | p->p_fdtol = NULL((void *)0); | |||
| 2113 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2114 | if (fdtol != NULL((void *)0)) | |||
| 2115 | free(fdtol, M_FILEDESC_TO_LEADER); | |||
| 2116 | } | |||
| 2117 | ||||
| 2118 | /* | |||
| 2119 | * Release a filedesc structure. | |||
| 2120 | */ | |||
| 2121 | static void | |||
| 2122 | fdescfree_fds(struct thread *td, struct filedesc *fdp, bool needclose) | |||
| 2123 | { | |||
| 2124 | struct filedesc0 *fdp0; | |||
| 2125 | struct freetable *ft, *tft; | |||
| 2126 | struct filedescent *fde; | |||
| 2127 | struct file *fp; | |||
| 2128 | int i; | |||
| 2129 | ||||
| 2130 | for (i = 0; i <= fdp->fd_lastfile; i++) { | |||
| 2131 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[i]; | |||
| 2132 | fp = fde->fde_file; | |||
| 2133 | if (fp != NULL((void *)0)) { | |||
| 2134 | fdefree_last(fde); | |||
| 2135 | if (needclose) | |||
| 2136 | (void) closef(fp, td); | |||
| 2137 | else | |||
| 2138 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2139 | } | |||
| 2140 | } | |||
| 2141 | ||||
| 2142 | if (NDSLOTS(fdp->fd_nfiles)(((fdp->fd_files->fdt_nfiles) + (sizeof(u_long) * 8) - 1 ) / (sizeof(u_long) * 8)) > NDSLOTS(NDFILE)(((20) + (sizeof(u_long) * 8) - 1) / (sizeof(u_long) * 8))) | |||
| 2143 | free(fdp->fd_map, M_FILEDESC); | |||
| 2144 | if (fdp->fd_nfilesfd_files->fdt_nfiles > NDFILE20) | |||
| 2145 | free(fdp->fd_files, M_FILEDESC); | |||
| 2146 | ||||
| 2147 | fdp0 = (struct filedesc0 *)fdp; | |||
| 2148 | SLIST_FOREACH_SAFE(ft, &fdp0->fd_free, ft_next, tft)for ((ft) = (((&fdp0->fd_free))->slh_first); (ft) && ((tft) = (((ft))->ft_next.sle_next), 1); (ft) = (tft)) | |||
| 2149 | free(ft->ft_table, M_FILEDESC); | |||
| 2150 | ||||
| 2151 | fddrop(fdp); | |||
| 2152 | } | |||
| 2153 | ||||
| 2154 | void | |||
| 2155 | fdescfree(struct thread *td) | |||
| 2156 | { | |||
| 2157 | struct proc *p; | |||
| 2158 | struct filedesc *fdp; | |||
| 2159 | struct vnode *cdir, *jdir, *rdir; | |||
| 2160 | ||||
| 2161 | p = td->td_proc; | |||
| 2162 | fdp = p->p_fd; | |||
| 2163 | MPASS(fdp != NULL)do { } while (0); | |||
| 2164 | ||||
| 2165 | #ifdef RACCT1 | |||
| 2166 | if (racct_enable) { | |||
| 2167 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 2168 | racct_set(p, RACCT_NOFILE7, 0); | |||
| 2169 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 2170 | } | |||
| 2171 | #endif | |||
| 2172 | ||||
| 2173 | if (p->p_fdtol != NULL((void *)0)) | |||
| 2174 | fdclearlocks(td); | |||
| 2175 | ||||
| 2176 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 2177 | p->p_fd = NULL((void *)0); | |||
| 2178 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 2179 | ||||
| 2180 | if (refcount_release(&fdp->fd_refcnt) == 0) | |||
| 2181 | return; | |||
| 2182 | ||||
| 2183 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2184 | cdir = fdp->fd_cdir; | |||
| 2185 | fdp->fd_cdir = NULL((void *)0); | |||
| 2186 | rdir = fdp->fd_rdir; | |||
| 2187 | fdp->fd_rdir = NULL((void *)0); | |||
| 2188 | jdir = fdp->fd_jdir; | |||
| 2189 | fdp->fd_jdir = NULL((void *)0); | |||
| 2190 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2191 | ||||
| 2192 | if (cdir != NULL((void *)0)) | |||
| 2193 | vrele(cdir); | |||
| 2194 | if (rdir != NULL((void *)0)) | |||
| 2195 | vrele(rdir); | |||
| 2196 | if (jdir != NULL((void *)0)) | |||
| 2197 | vrele(jdir); | |||
| 2198 | ||||
| 2199 | fdescfree_fds(td, fdp, 1); | |||
| 2200 | } | |||
| 2201 | ||||
| 2202 | void | |||
| 2203 | fdescfree_remapped(struct filedesc *fdp) | |||
| 2204 | { | |||
| 2205 | ||||
| 2206 | if (fdp->fd_cdir != NULL((void *)0)) | |||
| 2207 | vrele(fdp->fd_cdir); | |||
| 2208 | if (fdp->fd_rdir != NULL((void *)0)) | |||
| 2209 | vrele(fdp->fd_rdir); | |||
| 2210 | if (fdp->fd_jdir != NULL((void *)0)) | |||
| 2211 | vrele(fdp->fd_jdir); | |||
| 2212 | ||||
| 2213 | fdescfree_fds(curthread(__curthread()), fdp, 0); | |||
| 2214 | } | |||
| 2215 | ||||
| 2216 | /* | |||
| 2217 | * For setugid programs, we don't want to people to use that setugidness | |||
| 2218 | * to generate error messages which write to a file which otherwise would | |||
| 2219 | * otherwise be off-limits to the process. We check for filesystems where | |||
| 2220 | * the vnode can change out from under us after execve (like [lin]procfs). | |||
| 2221 | * | |||
| 2222 | * Since fdsetugidsafety calls this only for fd 0, 1 and 2, this check is | |||
| 2223 | * sufficient. We also don't check for setugidness since we know we are. | |||
| 2224 | */ | |||
| 2225 | static bool | |||
| 2226 | is_unsafe(struct file *fp) | |||
| 2227 | { | |||
| 2228 | struct vnode *vp; | |||
| 2229 | ||||
| 2230 | if (fp->f_type != DTYPE_VNODE1) | |||
| 2231 | return (false0); | |||
| 2232 | ||||
| 2233 | vp = fp->f_vnode; | |||
| 2234 | return ((vp->v_vflag & VV_PROCDEP0x0100) != 0); | |||
| 2235 | } | |||
| 2236 | ||||
| 2237 | /* | |||
| 2238 | * Make this setguid thing safe, if at all possible. | |||
| 2239 | */ | |||
| 2240 | void | |||
| 2241 | fdsetugidsafety(struct thread *td) | |||
| 2242 | { | |||
| 2243 | struct filedesc *fdp; | |||
| 2244 | struct file *fp; | |||
| 2245 | int i; | |||
| 2246 | ||||
| 2247 | fdp = td->td_proc->p_fd; | |||
| 2248 | KASSERT(fdp->fd_refcnt == 1, ("the fdtable should not be shared"))do { } while (0); | |||
| 2249 | MPASS(fdp->fd_nfiles >= 3)do { } while (0); | |||
| 2250 | for (i = 0; i <= 2; i++) { | |||
| 2251 | fp = fdp->fd_ofilesfd_files->fdt_ofiles[i].fde_file; | |||
| 2252 | if (fp != NULL((void *)0) && is_unsafe(fp)) { | |||
| 2253 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2254 | knote_fdclose(td, i); | |||
| 2255 | /* | |||
| 2256 | * NULL-out descriptor prior to close to avoid | |||
| 2257 | * a race while close blocks. | |||
| 2258 | */ | |||
| 2259 | fdfree(fdp, i); | |||
| 2260 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2261 | (void) closef(fp, td); | |||
| 2262 | } | |||
| 2263 | } | |||
| 2264 | } | |||
| 2265 | ||||
| 2266 | /* | |||
| 2267 | * If a specific file object occupies a specific file descriptor, close the | |||
| 2268 | * file descriptor entry and drop a reference on the file object. This is a | |||
| 2269 | * convenience function to handle a subsequent error in a function that calls | |||
| 2270 | * falloc() that handles the race that another thread might have closed the | |||
| 2271 | * file descriptor out from under the thread creating the file object. | |||
| 2272 | */ | |||
| 2273 | void | |||
| 2274 | fdclose(struct thread *td, struct file *fp, int idx) | |||
| 2275 | { | |||
| 2276 | struct filedesc *fdp = td->td_proc->p_fd; | |||
| 2277 | ||||
| 2278 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2279 | if (fdp->fd_ofilesfd_files->fdt_ofiles[idx].fde_file == fp) { | |||
| 2280 | fdfree(fdp, idx); | |||
| 2281 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2282 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2283 | } else | |||
| 2284 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2285 | } | |||
| 2286 | ||||
| 2287 | /* | |||
| 2288 | * Close any files on exec? | |||
| 2289 | */ | |||
| 2290 | void | |||
| 2291 | fdcloseexec(struct thread *td) | |||
| 2292 | { | |||
| 2293 | struct filedesc *fdp; | |||
| 2294 | struct filedescent *fde; | |||
| 2295 | struct file *fp; | |||
| 2296 | int i; | |||
| 2297 | ||||
| 2298 | fdp = td->td_proc->p_fd; | |||
| 2299 | KASSERT(fdp->fd_refcnt == 1, ("the fdtable should not be shared"))do { } while (0); | |||
| 2300 | for (i = 0; i <= fdp->fd_lastfile; i++) { | |||
| 2301 | fde = &fdp->fd_ofilesfd_files->fdt_ofiles[i]; | |||
| 2302 | fp = fde->fde_file; | |||
| 2303 | if (fp != NULL((void *)0) && (fp->f_type == DTYPE_MQUEUE7 || | |||
| 2304 | (fde->fde_flags & UF_EXCLOSE0x01))) { | |||
| 2305 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2306 | fdfree(fdp, i); | |||
| 2307 | (void) closefp(fdp, i, fp, td, 0); | |||
| 2308 | FILEDESC_UNLOCK_ASSERT(fdp)(void)0; | |||
| 2309 | } | |||
| 2310 | } | |||
| 2311 | } | |||
| 2312 | ||||
| 2313 | /* | |||
| 2314 | * It is unsafe for set[ug]id processes to be started with file | |||
| 2315 | * descriptors 0..2 closed, as these descriptors are given implicit | |||
| 2316 | * significance in the Standard C library. fdcheckstd() will create a | |||
| 2317 | * descriptor referencing /dev/null for each of stdin, stdout, and | |||
| 2318 | * stderr that is not already open. | |||
| 2319 | */ | |||
| 2320 | int | |||
| 2321 | fdcheckstd(struct thread *td) | |||
| 2322 | { | |||
| 2323 | struct filedesc *fdp; | |||
| 2324 | register_t save; | |||
| 2325 | int i, error, devnull; | |||
| 2326 | ||||
| 2327 | fdp = td->td_proc->p_fd; | |||
| 2328 | KASSERT(fdp->fd_refcnt == 1, ("the fdtable should not be shared"))do { } while (0); | |||
| 2329 | MPASS(fdp->fd_nfiles >= 3)do { } while (0); | |||
| 2330 | devnull = -1; | |||
| 2331 | for (i = 0; i <= 2; i++) { | |||
| 2332 | if (fdp->fd_ofilesfd_files->fdt_ofiles[i].fde_file != NULL((void *)0)) | |||
| 2333 | continue; | |||
| 2334 | ||||
| 2335 | save = td->td_retvaltd_uretoff.tdu_retval[0]; | |||
| 2336 | if (devnull != -1) { | |||
| 2337 | error = kern_dup(td, FDDUP_FIXED, 0, devnull, i); | |||
| 2338 | } else { | |||
| 2339 | error = kern_openat(td, AT_FDCWD-100, "/dev/null", | |||
| 2340 | UIO_SYSSPACE, O_RDWR0x0002, 0); | |||
| 2341 | if (error == 0) { | |||
| 2342 | devnull = td->td_retvaltd_uretoff.tdu_retval[0]; | |||
| 2343 | KASSERT(devnull == i, ("we didn't get our fd"))do { } while (0); | |||
| 2344 | } | |||
| 2345 | } | |||
| 2346 | td->td_retvaltd_uretoff.tdu_retval[0] = save; | |||
| 2347 | if (error != 0) | |||
| 2348 | return (error); | |||
| 2349 | } | |||
| 2350 | return (0); | |||
| 2351 | } | |||
| 2352 | ||||
| 2353 | /* | |||
| 2354 | * Internal form of close. Decrement reference count on file structure. | |||
| 2355 | * Note: td may be NULL when closing a file that was being passed in a | |||
| 2356 | * message. | |||
| 2357 | * | |||
| 2358 | * XXXRW: Giant is not required for the caller, but often will be held; this | |||
| 2359 | * makes it moderately likely the Giant will be recursed in the VFS case. | |||
| 2360 | */ | |||
| 2361 | int | |||
| 2362 | closef(struct file *fp, struct thread *td) | |||
| 2363 | { | |||
| 2364 | struct vnode *vp; | |||
| 2365 | struct flock lf; | |||
| 2366 | struct filedesc_to_leader *fdtol; | |||
| 2367 | struct filedesc *fdp; | |||
| 2368 | ||||
| 2369 | /* | |||
| 2370 | * POSIX record locking dictates that any close releases ALL | |||
| 2371 | * locks owned by this process. This is handled by setting | |||
| 2372 | * a flag in the unlock to free ONLY locks obeying POSIX | |||
| 2373 | * semantics, and not to free BSD-style file locks. | |||
| 2374 | * If the descriptor was in a message, POSIX-style locks | |||
| 2375 | * aren't passed with the descriptor, and the thread pointer | |||
| 2376 | * will be NULL. Callers should be careful only to pass a | |||
| 2377 | * NULL thread pointer when there really is no owning | |||
| 2378 | * context that might have locks, or the locks will be | |||
| 2379 | * leaked. | |||
| 2380 | */ | |||
| 2381 | if (fp->f_type == DTYPE_VNODE1 && td != NULL((void *)0)) { | |||
| 2382 | vp = fp->f_vnode; | |||
| 2383 | if ((td->td_proc->p_leader->p_flag & P_ADVLOCK0x00001) != 0) { | |||
| 2384 | lf.l_whence = SEEK_SET0; | |||
| 2385 | lf.l_start = 0; | |||
| 2386 | lf.l_len = 0; | |||
| 2387 | lf.l_type = F_UNLCK2; | |||
| 2388 | (void) VOP_ADVLOCK(vp, (caddr_t)td->td_proc->p_leader, | |||
| 2389 | F_UNLCK2, &lf, F_POSIX0x040); | |||
| 2390 | } | |||
| 2391 | fdtol = td->td_proc->p_fdtol; | |||
| 2392 | if (fdtol != NULL((void *)0)) { | |||
| 2393 | /* | |||
| 2394 | * Handle special case where file descriptor table is | |||
| 2395 | * shared between multiple process leaders. | |||
| 2396 | */ | |||
| 2397 | fdp = td->td_proc->p_fd; | |||
| 2398 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2399 | for (fdtol = fdtol->fdl_next; | |||
| 2400 | fdtol != td->td_proc->p_fdtol; | |||
| 2401 | fdtol = fdtol->fdl_next) { | |||
| 2402 | if ((fdtol->fdl_leader->p_flag & | |||
| 2403 | P_ADVLOCK0x00001) == 0) | |||
| 2404 | continue; | |||
| 2405 | fdtol->fdl_holdcount++; | |||
| 2406 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2407 | lf.l_whence = SEEK_SET0; | |||
| 2408 | lf.l_start = 0; | |||
| 2409 | lf.l_len = 0; | |||
| 2410 | lf.l_type = F_UNLCK2; | |||
| 2411 | vp = fp->f_vnode; | |||
| 2412 | (void) VOP_ADVLOCK(vp, | |||
| 2413 | (caddr_t)fdtol->fdl_leader, F_UNLCK2, &lf, | |||
| 2414 | F_POSIX0x040); | |||
| 2415 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2416 | fdtol->fdl_holdcount--; | |||
| 2417 | if (fdtol->fdl_holdcount == 0 && | |||
| 2418 | fdtol->fdl_wakeup != 0) { | |||
| 2419 | fdtol->fdl_wakeup = 0; | |||
| 2420 | wakeup(fdtol); | |||
| 2421 | } | |||
| 2422 | } | |||
| 2423 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2424 | } | |||
| 2425 | } | |||
| 2426 | return (fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop())); | |||
| 2427 | } | |||
| 2428 | ||||
| 2429 | /* | |||
| 2430 | * Initialize the file pointer with the specified properties. | |||
| 2431 | * | |||
| 2432 | * The ops are set with release semantics to be certain that the flags, type, | |||
| 2433 | * and data are visible when ops is. This is to prevent ops methods from being | |||
| 2434 | * called with bad data. | |||
| 2435 | */ | |||
| 2436 | void | |||
| 2437 | finit(struct file *fp, u_int flag, short type, void *data, struct fileops *ops) | |||
| 2438 | { | |||
| 2439 | fp->f_data = data; | |||
| 2440 | fp->f_flag = flag; | |||
| 2441 | fp->f_type = type; | |||
| 2442 | atomic_store_rel_ptratomic_store_rel_long((volatile uintptr_t *)&fp->f_ops, (uintptr_t)ops); | |||
| 2443 | } | |||
| 2444 | ||||
| 2445 | int | |||
| 2446 | fget_unlocked(struct filedesc *fdp, int fd, cap_rights_t *needrightsp, | |||
| 2447 | struct file **fpp, seq_t *seqp) | |||
| 2448 | { | |||
| 2449 | #ifdef CAPABILITIES1 | |||
| 2450 | struct filedescent *fde; | |||
| 2451 | #endif | |||
| 2452 | struct fdescenttbl *fdt; | |||
| 2453 | struct file *fp; | |||
| 2454 | u_int count; | |||
| 2455 | #ifdef CAPABILITIES1 | |||
| 2456 | seq_t seq; | |||
| 2457 | cap_rights_t haverights; | |||
| 2458 | int error; | |||
| 2459 | #endif | |||
| 2460 | ||||
| 2461 | fdt = fdp->fd_files; | |||
| 2462 | if ((u_int)fd >= fdt->fdt_nfiles) | |||
| 2463 | return (EBADF9); | |||
| 2464 | /* | |||
| 2465 | * Fetch the descriptor locklessly. We avoid fdrop() races by | |||
| 2466 | * never raising a refcount above 0. To accomplish this we have | |||
| 2467 | * to use a cmpset loop rather than an atomic_add. The descriptor | |||
| 2468 | * must be re-verified once we acquire a reference to be certain | |||
| 2469 | * that the identity is still correct and we did not lose a race | |||
| 2470 | * due to preemption. | |||
| 2471 | */ | |||
| 2472 | for (;;) { | |||
| 2473 | #ifdef CAPABILITIES1 | |||
| 2474 | seq = seq_read(fd_seq(fdt, fd)(&(fdt)->fdt_ofiles[(fd)].fde_seq)); | |||
| 2475 | fde = &fdt->fdt_ofiles[fd]; | |||
| 2476 | haverights = *cap_rights_fde(fde); | |||
| 2477 | fp = fde->fde_file; | |||
| 2478 | if (!seq_consistent(fd_seq(fdt, fd)(&(fdt)->fdt_ofiles[(fd)].fde_seq), seq)) { | |||
| 2479 | cpu_spinwait()ia32_pause(); | |||
| 2480 | continue; | |||
| 2481 | } | |||
| 2482 | #else | |||
| 2483 | fp = fdt->fdt_ofiles[fd].fde_file; | |||
| 2484 | #endif | |||
| 2485 | if (fp == NULL((void *)0)) | |||
| 2486 | return (EBADF9); | |||
| 2487 | #ifdef CAPABILITIES1 | |||
| 2488 | error = cap_check(&haverights, needrightsp); | |||
| 2489 | if (error != 0) | |||
| 2490 | return (error); | |||
| 2491 | #endif | |||
| 2492 | retry: | |||
| 2493 | count = fp->f_count; | |||
| 2494 | if (count == 0) { | |||
| 2495 | /* | |||
| 2496 | * Force a reload. Other thread could reallocate the | |||
| 2497 | * table before this fd was closed, so it possible that | |||
| 2498 | * there is a stale fp pointer in cached version. | |||
| 2499 | */ | |||
| 2500 | fdt = *(struct fdescenttbl * volatile *)&(fdp->fd_files); | |||
| 2501 | continue; | |||
| 2502 | } | |||
| 2503 | /* | |||
| 2504 | * Use an acquire barrier to force re-reading of fdt so it is | |||
| 2505 | * refreshed for verification. | |||
| 2506 | */ | |||
| 2507 | if (atomic_cmpset_acq_intatomic_cmpset_int(&fp->f_count, count, count + 1) == 0) | |||
| 2508 | goto retry; | |||
| 2509 | fdt = fdp->fd_files; | |||
| 2510 | #ifdef CAPABILITIES1 | |||
| 2511 | if (seq_consistent_nomb(fd_seq(fdt, fd)(&(fdt)->fdt_ofiles[(fd)].fde_seq), seq)) | |||
| 2512 | #else | |||
| 2513 | if (fp == fdt->fdt_ofiles[fd].fde_file) | |||
| 2514 | #endif | |||
| 2515 | break; | |||
| 2516 | fdrop(fp, curthread)(refcount_release(&(fp)->f_count) ? _fdrop((fp), ((__curthread ()))) : _fnoop()); | |||
| 2517 | } | |||
| 2518 | *fpp = fp; | |||
| 2519 | if (seqp != NULL((void *)0)) { | |||
| 2520 | #ifdef CAPABILITIES1 | |||
| 2521 | *seqp = seq; | |||
| 2522 | #endif | |||
| 2523 | } | |||
| 2524 | return (0); | |||
| 2525 | } | |||
| 2526 | ||||
| 2527 | /* | |||
| 2528 | * Extract the file pointer associated with the specified descriptor for the | |||
| 2529 | * current user process. | |||
| 2530 | * | |||
| 2531 | * If the descriptor doesn't exist or doesn't match 'flags', EBADF is | |||
| 2532 | * returned. | |||
| 2533 | * | |||
| 2534 | * File's rights will be checked against the capability rights mask. | |||
| 2535 | * | |||
| 2536 | * If an error occurred the non-zero error is returned and *fpp is set to | |||
| 2537 | * NULL. Otherwise *fpp is held and set and zero is returned. Caller is | |||
| 2538 | * responsible for fdrop(). | |||
| 2539 | */ | |||
| 2540 | static __inline int | |||
| 2541 | _fget(struct thread *td, int fd, struct file **fpp, int flags, | |||
| 2542 | cap_rights_t *needrightsp, seq_t *seqp) | |||
| 2543 | { | |||
| 2544 | struct filedesc *fdp; | |||
| 2545 | struct file *fp; | |||
| 2546 | int error; | |||
| 2547 | ||||
| 2548 | *fpp = NULL((void *)0); | |||
| 2549 | fdp = td->td_proc->p_fd; | |||
| 2550 | error = fget_unlocked(fdp, fd, needrightsp, &fp, seqp); | |||
| 2551 | if (error != 0) | |||
| 2552 | return (error); | |||
| 2553 | if (fp->f_ops == &badfileops) { | |||
| 2554 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2555 | return (EBADF9); | |||
| 2556 | } | |||
| 2557 | ||||
| 2558 | /* | |||
| 2559 | * FREAD and FWRITE failure return EBADF as per POSIX. | |||
| 2560 | */ | |||
| 2561 | error = 0; | |||
| 2562 | switch (flags) { | |||
| 2563 | case FREAD0x0001: | |||
| 2564 | case FWRITE0x0002: | |||
| 2565 | if ((fp->f_flag & flags) == 0) | |||
| 2566 | error = EBADF9; | |||
| 2567 | break; | |||
| 2568 | case FEXEC0x00040000: | |||
| 2569 | if ((fp->f_flag & (FREAD0x0001 | FEXEC0x00040000)) == 0 || | |||
| 2570 | ((fp->f_flag & FWRITE0x0002) != 0)) | |||
| 2571 | error = EBADF9; | |||
| 2572 | break; | |||
| 2573 | case 0: | |||
| 2574 | break; | |||
| 2575 | default: | |||
| 2576 | KASSERT(0, ("wrong flags"))do { } while (0); | |||
| 2577 | } | |||
| 2578 | ||||
| 2579 | if (error != 0) { | |||
| 2580 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2581 | return (error); | |||
| 2582 | } | |||
| 2583 | ||||
| 2584 | *fpp = fp; | |||
| 2585 | return (0); | |||
| 2586 | } | |||
| 2587 | ||||
| 2588 | int | |||
| 2589 | fget(struct thread *td, int fd, cap_rights_t *rightsp, struct file **fpp) | |||
| 2590 | { | |||
| 2591 | ||||
| 2592 | return (_fget(td, fd, fpp, 0, rightsp, NULL((void *)0))); | |||
| 2593 | } | |||
| 2594 | ||||
| 2595 | int | |||
| 2596 | fget_mmap(struct thread *td, int fd, cap_rights_t *rightsp, u_char *maxprotp, | |||
| 2597 | struct file **fpp) | |||
| 2598 | { | |||
| 2599 | int error; | |||
| 2600 | #ifndef CAPABILITIES1 | |||
| 2601 | error = _fget(td, fd, fpp, 0, rightsp, NULL((void *)0)); | |||
| 2602 | if (maxprotp != NULL((void *)0)) | |||
| 2603 | *maxprotp = VM_PROT_ALL(((vm_prot_t) 0x01)|((vm_prot_t) 0x02)|((vm_prot_t) 0x04)); | |||
| 2604 | #else | |||
| 2605 | struct filedesc *fdp = td->td_proc->p_fd; | |||
| 2606 | seq_t seq; | |||
| 2607 | ||||
| 2608 | MPASS(cap_rights_is_set(rightsp, CAP_MMAP))do { } while (0); | |||
| 2609 | for (;;) { | |||
| 2610 | error = _fget(td, fd, fpp, 0, rightsp, &seq); | |||
| 2611 | if (error != 0) | |||
| 2612 | return (error); | |||
| 2613 | /* | |||
| 2614 | * If requested, convert capability rights to access flags. | |||
| 2615 | */ | |||
| 2616 | if (maxprotp != NULL((void *)0)) | |||
| 2617 | *maxprotp = cap_rights_to_vmprot(cap_rights(fdp, fd)); | |||
| 2618 | if (!fd_modified(fdp, fd, seq)) | |||
| 2619 | break; | |||
| 2620 | fdrop(*fpp, td)(refcount_release(&(*fpp)->f_count) ? _fdrop((*fpp), ( td)) : _fnoop()); | |||
| 2621 | } | |||
| 2622 | #endif | |||
| 2623 | return (error); | |||
| 2624 | } | |||
| 2625 | ||||
| 2626 | int | |||
| 2627 | fget_read(struct thread *td, int fd, cap_rights_t *rightsp, struct file **fpp) | |||
| 2628 | { | |||
| 2629 | ||||
| 2630 | return (_fget(td, fd, fpp, FREAD0x0001, rightsp, NULL((void *)0))); | |||
| 2631 | } | |||
| 2632 | ||||
| 2633 | int | |||
| 2634 | fget_write(struct thread *td, int fd, cap_rights_t *rightsp, struct file **fpp) | |||
| 2635 | { | |||
| 2636 | ||||
| 2637 | return (_fget(td, fd, fpp, FWRITE0x0002, rightsp, NULL((void *)0))); | |||
| 2638 | } | |||
| 2639 | ||||
| 2640 | int | |||
| 2641 | fget_fcntl(struct thread *td, int fd, cap_rights_t *rightsp, int needfcntl, | |||
| 2642 | struct file **fpp) | |||
| 2643 | { | |||
| 2644 | struct filedesc *fdp = td->td_proc->p_fd; | |||
| 2645 | #ifndef CAPABILITIES1 | |||
| 2646 | return (fget_unlocked(fdp, fd, rightsp, fpp, NULL((void *)0))); | |||
| 2647 | #else | |||
| 2648 | int error; | |||
| 2649 | seq_t seq; | |||
| 2650 | ||||
| 2651 | MPASS(cap_rights_is_set(rightsp, CAP_FCNTL))do { } while (0); | |||
| 2652 | for (;;) { | |||
| 2653 | error = fget_unlocked(fdp, fd, rightsp, fpp, &seq); | |||
| 2654 | if (error != 0) | |||
| 2655 | return (error); | |||
| 2656 | error = cap_fcntl_check(fdp, fd, needfcntl); | |||
| 2657 | if (!fd_modified(fdp, fd, seq)) | |||
| 2658 | break; | |||
| 2659 | fdrop(*fpp, td)(refcount_release(&(*fpp)->f_count) ? _fdrop((*fpp), ( td)) : _fnoop()); | |||
| 2660 | } | |||
| 2661 | if (error != 0) { | |||
| 2662 | fdrop(*fpp, td)(refcount_release(&(*fpp)->f_count) ? _fdrop((*fpp), ( td)) : _fnoop()); | |||
| 2663 | *fpp = NULL((void *)0); | |||
| 2664 | } | |||
| 2665 | return (error); | |||
| 2666 | #endif | |||
| 2667 | } | |||
| 2668 | ||||
| 2669 | /* | |||
| 2670 | * Like fget() but loads the underlying vnode, or returns an error if the | |||
| 2671 | * descriptor does not represent a vnode. Note that pipes use vnodes but | |||
| 2672 | * never have VM objects. The returned vnode will be vref()'d. | |||
| 2673 | * | |||
| 2674 | * XXX: what about the unused flags ? | |||
| 2675 | */ | |||
| 2676 | static __inline int | |||
| 2677 | _fgetvp(struct thread *td, int fd, int flags, cap_rights_t *needrightsp, | |||
| 2678 | struct vnode **vpp) | |||
| 2679 | { | |||
| 2680 | struct file *fp; | |||
| 2681 | int error; | |||
| 2682 | ||||
| 2683 | *vpp = NULL((void *)0); | |||
| 2684 | error = _fget(td, fd, &fp, flags, needrightsp, NULL((void *)0)); | |||
| 2685 | if (error != 0) | |||
| 2686 | return (error); | |||
| 2687 | if (fp->f_vnode == NULL((void *)0)) { | |||
| 2688 | error = EINVAL22; | |||
| 2689 | } else { | |||
| 2690 | *vpp = fp->f_vnode; | |||
| 2691 | vref(*vpp); | |||
| 2692 | } | |||
| 2693 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2694 | ||||
| 2695 | return (error); | |||
| 2696 | } | |||
| 2697 | ||||
| 2698 | int | |||
| 2699 | fgetvp(struct thread *td, int fd, cap_rights_t *rightsp, struct vnode **vpp) | |||
| 2700 | { | |||
| 2701 | ||||
| 2702 | return (_fgetvp(td, fd, 0, rightsp, vpp)); | |||
| 2703 | } | |||
| 2704 | ||||
| 2705 | int | |||
| 2706 | fgetvp_rights(struct thread *td, int fd, cap_rights_t *needrightsp, | |||
| 2707 | struct filecaps *havecaps, struct vnode **vpp) | |||
| 2708 | { | |||
| 2709 | struct filedesc *fdp; | |||
| 2710 | struct file *fp; | |||
| 2711 | #ifdef CAPABILITIES1 | |||
| 2712 | int error; | |||
| 2713 | #endif | |||
| 2714 | ||||
| 2715 | fdp = td->td_proc->p_fd; | |||
| 2716 | fp = fget_locked(fdp, fd); | |||
| 2717 | if (fp == NULL((void *)0) || fp->f_ops == &badfileops) | |||
| 2718 | return (EBADF9); | |||
| 2719 | ||||
| 2720 | #ifdef CAPABILITIES1 | |||
| 2721 | error = cap_check(cap_rights(fdp, fd), needrightsp); | |||
| 2722 | if (error != 0) | |||
| 2723 | return (error); | |||
| 2724 | #endif | |||
| 2725 | ||||
| 2726 | if (fp->f_vnode == NULL((void *)0)) | |||
| 2727 | return (EINVAL22); | |||
| 2728 | ||||
| 2729 | *vpp = fp->f_vnode; | |||
| 2730 | vref(*vpp); | |||
| 2731 | filecaps_copy(&fdp->fd_ofilesfd_files->fdt_ofiles[fd].fde_caps, havecaps, true1); | |||
| 2732 | ||||
| 2733 | return (0); | |||
| 2734 | } | |||
| 2735 | ||||
| 2736 | int | |||
| 2737 | fgetvp_read(struct thread *td, int fd, cap_rights_t *rightsp, struct vnode **vpp) | |||
| 2738 | { | |||
| 2739 | ||||
| 2740 | return (_fgetvp(td, fd, FREAD0x0001, rightsp, vpp)); | |||
| 2741 | } | |||
| 2742 | ||||
| 2743 | int | |||
| 2744 | fgetvp_exec(struct thread *td, int fd, cap_rights_t *rightsp, struct vnode **vpp) | |||
| 2745 | { | |||
| 2746 | ||||
| 2747 | return (_fgetvp(td, fd, FEXEC0x00040000, rightsp, vpp)); | |||
| 2748 | } | |||
| 2749 | ||||
| 2750 | #ifdef notyet | |||
| 2751 | int | |||
| 2752 | fgetvp_write(struct thread *td, int fd, cap_rights_t *rightsp, | |||
| 2753 | struct vnode **vpp) | |||
| 2754 | { | |||
| 2755 | ||||
| 2756 | return (_fgetvp(td, fd, FWRITE0x0002, rightsp, vpp)); | |||
| 2757 | } | |||
| 2758 | #endif | |||
| 2759 | ||||
| 2760 | /* | |||
| 2761 | * Like fget() but loads the underlying socket, or returns an error if the | |||
| 2762 | * descriptor does not represent a socket. | |||
| 2763 | * | |||
| 2764 | * We bump the ref count on the returned socket. XXX Also obtain the SX lock | |||
| 2765 | * in the future. | |||
| 2766 | * | |||
| 2767 | * Note: fgetsock() and fputsock() are deprecated, as consumers should rely | |||
| 2768 | * on their file descriptor reference to prevent the socket from being free'd | |||
| 2769 | * during use. | |||
| 2770 | */ | |||
| 2771 | int | |||
| 2772 | fgetsock(struct thread *td, int fd, cap_rights_t *rightsp, struct socket **spp, | |||
| 2773 | u_int *fflagp) | |||
| 2774 | { | |||
| 2775 | struct file *fp; | |||
| 2776 | int error; | |||
| 2777 | ||||
| 2778 | *spp = NULL((void *)0); | |||
| 2779 | if (fflagp != NULL((void *)0)) | |||
| 2780 | *fflagp = 0; | |||
| 2781 | if ((error = _fget(td, fd, &fp, 0, rightsp, NULL((void *)0))) != 0) | |||
| 2782 | return (error); | |||
| 2783 | if (fp->f_type != DTYPE_SOCKET2) { | |||
| 2784 | error = ENOTSOCK38; | |||
| 2785 | } else { | |||
| 2786 | *spp = fp->f_data; | |||
| 2787 | if (fflagp) | |||
| 2788 | *fflagp = fp->f_flag; | |||
| 2789 | SOCK_LOCK(*spp)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( (&(&(*spp)->so_rcv)->sb_mtx)))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&((((((&(&(*spp )->so_rcv)->sb_mtx))))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&((((((&(&(*spp)->so_rcv )->sb_mtx))))))->mtx_lock, _tid, (((0))), ((((void *)0) )), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) ((((&(&(*spp)->so_rcv)->sb_mtx )))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t ) 0); } while (0); } while (0); } while (0); | |||
| 2790 | soref(*spp)do { (void)0; ++(*spp)->so_count; } while (0); | |||
| 2791 | SOCK_UNLOCK(*spp)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(&(*spp)->so_rcv)->sb_mtx)))))->lock_object .lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) ((((&(&(*spp)->so_rcv)->sb_mtx )))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t ) 0); } while (0); } while (0); if ((((((&(&(*spp)-> so_rcv)->sb_mtx)))))->mtx_lock != _tid || !atomic_cmpset_long (&((((((&(&(*spp)->so_rcv)->sb_mtx))))))-> mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(&(((((( &(&(*spp)->so_rcv)->sb_mtx))))))->mtx_lock, ( ((0))), ((((void *)0))), ((0))); } while (0); | |||
| 2792 | } | |||
| 2793 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2794 | ||||
| 2795 | return (error); | |||
| 2796 | } | |||
| 2797 | ||||
| 2798 | /* | |||
| 2799 | * Drop the reference count on the socket and XXX release the SX lock in the | |||
| 2800 | * future. The last reference closes the socket. | |||
| 2801 | * | |||
| 2802 | * Note: fputsock() is deprecated, see comment for fgetsock(). | |||
| 2803 | */ | |||
| 2804 | void | |||
| 2805 | fputsock(struct socket *so) | |||
| 2806 | { | |||
| 2807 | ||||
| 2808 | ACCEPT_LOCK()do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &accept_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&accept_mtx)))))->mtx_lock, 0x00000004, (_tid )))) __mtx_lock_sleep(&(((((&accept_mtx)))))->mtx_lock , _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire-> id, (uintptr_t) (((&accept_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 2809 | SOCK_LOCK(so)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( (&(&(so)->so_rcv)->sb_mtx)))))->mtx_lock != 0x00000004 || !atomic_cmpset_long(&((((((&(&(so)->so_rcv )->sb_mtx))))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep (&((((((&(&(so)->so_rcv)->sb_mtx))))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) ((((&(&(so)->so_rcv)->sb_mtx )))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t ) 0); } while (0); } while (0); } while (0); | |||
| 2810 | CURVNET_SET(so->so_vnet); | |||
| 2811 | sorele(so)do { (void)0; (void)0; if ((so)->so_count <= 0) panic("sorele" ); if (--(so)->so_count == 0) sofree(so); else { do { uintptr_t _tid = (uintptr_t)((__curthread())); if ((((((&(&(so )->so_rcv)->sb_mtx)))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) ((((&(&(so)->so_rcv)->sb_mtx )))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t ) 0); } while (0); } while (0); if ((((((&(&(so)-> so_rcv)->sb_mtx)))))->mtx_lock != _tid || !atomic_cmpset_long (&((((((&(&(so)->so_rcv)->sb_mtx))))))-> mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(&(((((( &(&(so)->so_rcv)->sb_mtx))))))->mtx_lock, (( (0))), ((((void *)0))), ((0))); } while (0); do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((&accept_mtx))))-> lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect ((sdt_lockstat___adaptive__release->id), 0)) (*sdt_probe_func )(sdt_lockstat___adaptive__release->id, (uintptr_t) (((& accept_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, ( uintptr_t) 0); } while (0); } while (0); if (((((&accept_mtx ))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((& accept_mtx)))))->mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep (&(((((&accept_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ((0))); } while (0); } } while (0); | |||
| 2812 | CURVNET_RESTORE(); | |||
| 2813 | } | |||
| 2814 | ||||
| 2815 | /* | |||
| 2816 | * Handle the last reference to a file being closed. | |||
| 2817 | */ | |||
| 2818 | int | |||
| 2819 | _fdrop(struct file *fp, struct thread *td) | |||
| 2820 | { | |||
| 2821 | int error; | |||
| 2822 | ||||
| 2823 | if (fp->f_count != 0) | |||
| 2824 | panic("fdrop: count %d", fp->f_count); | |||
| 2825 | error = fo_close(fp, td); | |||
| 2826 | atomic_subtract_int(&openfiles, 1); | |||
| 2827 | crfree(fp->f_cred); | |||
| 2828 | free(fp->f_advicef_vnun.fvn_advice, M_FADVISE); | |||
| 2829 | uma_zfree(file_zone, fp); | |||
| 2830 | ||||
| 2831 | return (error); | |||
| 2832 | } | |||
| 2833 | ||||
| 2834 | /* | |||
| 2835 | * Apply an advisory lock on a file descriptor. | |||
| 2836 | * | |||
| 2837 | * Just attempt to get a record lock of the requested type on the entire file | |||
| 2838 | * (l_whence = SEEK_SET, l_start = 0, l_len = 0). | |||
| 2839 | */ | |||
| 2840 | #ifndef _SYS_SYSPROTO_H_ | |||
| 2841 | struct flock_args { | |||
| 2842 | int fd; | |||
| 2843 | int how; | |||
| 2844 | }; | |||
| 2845 | #endif | |||
| 2846 | /* ARGSUSED */ | |||
| 2847 | int | |||
| 2848 | sys_flock(struct thread *td, struct flock_args *uap) | |||
| 2849 | { | |||
| 2850 | struct file *fp; | |||
| 2851 | struct vnode *vp; | |||
| 2852 | struct flock lf; | |||
| 2853 | cap_rights_t rights; | |||
| 2854 | int error; | |||
| 2855 | ||||
| 2856 | error = fget(td, uap->fd, cap_rights_init(&rights, CAP_FLOCK)__cap_rights_init(0, &rights, ((1ULL << (57 + (0))) | (0x0000000000010000ULL)), 0ULL), &fp); | |||
| 2857 | if (error != 0) | |||
| 2858 | return (error); | |||
| 2859 | if (fp->f_type != DTYPE_VNODE1) { | |||
| 2860 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2861 | return (EOPNOTSUPP45); | |||
| 2862 | } | |||
| 2863 | ||||
| 2864 | vp = fp->f_vnode; | |||
| 2865 | lf.l_whence = SEEK_SET0; | |||
| 2866 | lf.l_start = 0; | |||
| 2867 | lf.l_len = 0; | |||
| 2868 | if (uap->how & LOCK_UN0x08) { | |||
| 2869 | lf.l_type = F_UNLCK2; | |||
| 2870 | atomic_clear_int(&fp->f_flag, FHASLOCK0x4000); | |||
| 2871 | error = VOP_ADVLOCK(vp, (caddr_t)fp, F_UNLCK2, &lf, F_FLOCK0x020); | |||
| 2872 | goto done2; | |||
| 2873 | } | |||
| 2874 | if (uap->how & LOCK_EX0x02) | |||
| 2875 | lf.l_type = F_WRLCK3; | |||
| 2876 | else if (uap->how & LOCK_SH0x01) | |||
| 2877 | lf.l_type = F_RDLCK1; | |||
| 2878 | else { | |||
| 2879 | error = EBADF9; | |||
| 2880 | goto done2; | |||
| 2881 | } | |||
| 2882 | atomic_set_int(&fp->f_flag, FHASLOCK0x4000); | |||
| 2883 | error = VOP_ADVLOCK(vp, (caddr_t)fp, F_SETLK12, &lf, | |||
| 2884 | (uap->how & LOCK_NB0x04) ? F_FLOCK0x020 : F_FLOCK0x020 | F_WAIT0x010); | |||
| 2885 | done2: | |||
| 2886 | fdrop(fp, td)(refcount_release(&(fp)->f_count) ? _fdrop((fp), (td)) : _fnoop()); | |||
| 2887 | return (error); | |||
| 2888 | } | |||
| 2889 | /* | |||
| 2890 | * Duplicate the specified descriptor to a free descriptor. | |||
| 2891 | */ | |||
| 2892 | int | |||
| 2893 | dupfdopen(struct thread *td, struct filedesc *fdp, int dfd, int mode, | |||
| 2894 | int openerror, int *indxp) | |||
| 2895 | { | |||
| 2896 | struct filedescent *newfde, *oldfde; | |||
| 2897 | struct file *fp; | |||
| 2898 | int error, indx; | |||
| 2899 | ||||
| 2900 | KASSERT(openerror == ENODEV || openerror == ENXIO,do { } while (0) | |||
| 2901 | ("unexpected error %d in %s", openerror, __func__))do { } while (0); | |||
| 2902 | ||||
| 2903 | /* | |||
| 2904 | * If the to-be-dup'd fd number is greater than the allowed number | |||
| 2905 | * of file descriptors, or the fd to be dup'd has already been | |||
| 2906 | * closed, then reject. | |||
| 2907 | */ | |||
| 2908 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 2909 | if ((fp = fget_locked(fdp, dfd)) == NULL((void *)0)) { | |||
| 2910 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2911 | return (EBADF9); | |||
| 2912 | } | |||
| 2913 | ||||
| 2914 | error = fdalloc(td, 0, &indx); | |||
| 2915 | if (error != 0) { | |||
| 2916 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2917 | return (error); | |||
| 2918 | } | |||
| 2919 | ||||
| 2920 | /* | |||
| 2921 | * There are two cases of interest here. | |||
| 2922 | * | |||
| 2923 | * For ENODEV simply dup (dfd) to file descriptor (indx) and return. | |||
| 2924 | * | |||
| 2925 | * For ENXIO steal away the file structure from (dfd) and store it in | |||
| 2926 | * (indx). (dfd) is effectively closed by this operation. | |||
| 2927 | */ | |||
| 2928 | switch (openerror) { | |||
| 2929 | case ENODEV19: | |||
| 2930 | /* | |||
| 2931 | * Check that the mode the file is being opened for is a | |||
| 2932 | * subset of the mode of the existing descriptor. | |||
| 2933 | */ | |||
| 2934 | if (((mode & (FREAD0x0001|FWRITE0x0002)) | fp->f_flag) != fp->f_flag) { | |||
| 2935 | fdunused(fdp, indx); | |||
| 2936 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2937 | return (EACCES13); | |||
| 2938 | } | |||
| 2939 | fhold(fp)(refcount_acquire(&(fp)->f_count)); | |||
| 2940 | newfde = &fdp->fd_ofilesfd_files->fdt_ofiles[indx]; | |||
| 2941 | oldfde = &fdp->fd_ofilesfd_files->fdt_ofiles[dfd]; | |||
| 2942 | #ifdef CAPABILITIES1 | |||
| 2943 | seq_write_begin(&newfde->fde_seq); | |||
| 2944 | #endif | |||
| 2945 | memcpy(newfde, oldfde, fde_change_size(__builtin_offsetof(struct filedescent, fde_seq))); | |||
| 2946 | filecaps_copy(&oldfde->fde_caps, &newfde->fde_caps, true1); | |||
| 2947 | #ifdef CAPABILITIES1 | |||
| 2948 | seq_write_end(&newfde->fde_seq); | |||
| 2949 | #endif | |||
| 2950 | break; | |||
| 2951 | case ENXIO6: | |||
| 2952 | /* | |||
| 2953 | * Steal away the file pointer from dfd and stuff it into indx. | |||
| 2954 | */ | |||
| 2955 | newfde = &fdp->fd_ofilesfd_files->fdt_ofiles[indx]; | |||
| 2956 | oldfde = &fdp->fd_ofilesfd_files->fdt_ofiles[dfd]; | |||
| 2957 | #ifdef CAPABILITIES1 | |||
| 2958 | seq_write_begin(&newfde->fde_seq); | |||
| 2959 | #endif | |||
| 2960 | memcpy(newfde, oldfde, fde_change_size(__builtin_offsetof(struct filedescent, fde_seq))); | |||
| 2961 | oldfde->fde_file = NULL((void *)0); | |||
| 2962 | fdunused(fdp, dfd); | |||
| 2963 | #ifdef CAPABILITIES1 | |||
| 2964 | seq_write_end(&newfde->fde_seq); | |||
| 2965 | #endif | |||
| 2966 | break; | |||
| 2967 | } | |||
| 2968 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 2969 | *indxp = indx; | |||
| 2970 | return (0); | |||
| 2971 | } | |||
| 2972 | ||||
| 2973 | /* | |||
| 2974 | * This sysctl determines if we will allow a process to chroot(2) if it | |||
| 2975 | * has a directory open: | |||
| 2976 | * 0: disallowed for all processes. | |||
| 2977 | * 1: allowed for processes that were not already chroot(2)'ed. | |||
| 2978 | * 2: allowed for all processes. | |||
| 2979 | */ | |||
| 2980 | ||||
| 2981 | static int chroot_allow_open_directories = 1; | |||
| 2982 | ||||
| 2983 | SYSCTL_INT(_kern, OID_AUTO, chroot_allow_open_directories, CTLFLAG_RW,static struct sysctl_oid sysctl___kern_chroot_allow_open_directories = { .oid_parent = ((&(&sysctl___kern)->oid_children )), .oid_children = { ((void *)0) }, .oid_number = ((-1)), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = ( &chroot_allow_open_directories), .oid_arg2 = (0), .oid_name = ("chroot_allow_open_directories"), .oid_handler = (sysctl_handle_int ), .oid_fmt = ("I"), .oid_descr = "Allow a process to chroot(2) if it has a directory open" }; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_chroot_allow_open_directories __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_chroot_allow_open_directories ); _Static_assert(((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000|0x40000000)) & 0) == 2) && sizeof (int) == sizeof(*(&chroot_allow_open_directories)), "compile-time assertion failed" ) | |||
| 2984 | &chroot_allow_open_directories, 0,static struct sysctl_oid sysctl___kern_chroot_allow_open_directories = { .oid_parent = ((&(&sysctl___kern)->oid_children )), .oid_children = { ((void *)0) }, .oid_number = ((-1)), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = ( &chroot_allow_open_directories), .oid_arg2 = (0), .oid_name = ("chroot_allow_open_directories"), .oid_handler = (sysctl_handle_int ), .oid_fmt = ("I"), .oid_descr = "Allow a process to chroot(2) if it has a directory open" }; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_chroot_allow_open_directories __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_chroot_allow_open_directories ); _Static_assert(((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000|0x40000000)) & 0) == 2) && sizeof (int) == sizeof(*(&chroot_allow_open_directories)), "compile-time assertion failed" ) | |||
| 2985 | "Allow a process to chroot(2) if it has a directory open")static struct sysctl_oid sysctl___kern_chroot_allow_open_directories = { .oid_parent = ((&(&sysctl___kern)->oid_children )), .oid_children = { ((void *)0) }, .oid_number = ((-1)), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = ( &chroot_allow_open_directories), .oid_arg2 = (0), .oid_name = ("chroot_allow_open_directories"), .oid_handler = (sysctl_handle_int ), .oid_fmt = ("I"), .oid_descr = "Allow a process to chroot(2) if it has a directory open" }; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_chroot_allow_open_directories __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_chroot_allow_open_directories ); _Static_assert(((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000|0x40000000)) & 0) == 2) && sizeof (int) == sizeof(*(&chroot_allow_open_directories)), "compile-time assertion failed" ); | |||
| 2986 | ||||
| 2987 | /* | |||
| 2988 | * Helper function for raised chroot(2) security function: Refuse if | |||
| 2989 | * any filedescriptors are open directories. | |||
| 2990 | */ | |||
| 2991 | static int | |||
| 2992 | chroot_refuse_vdir_fds(struct filedesc *fdp) | |||
| 2993 | { | |||
| 2994 | struct vnode *vp; | |||
| 2995 | struct file *fp; | |||
| 2996 | int fd; | |||
| 2997 | ||||
| 2998 | FILEDESC_LOCK_ASSERT(fdp)(void)0; | |||
| 2999 | ||||
| 3000 | for (fd = 0; fd <= fdp->fd_lastfile; fd++) { | |||
| 3001 | fp = fget_locked(fdp, fd); | |||
| 3002 | if (fp == NULL((void *)0)) | |||
| 3003 | continue; | |||
| 3004 | if (fp->f_type == DTYPE_VNODE1) { | |||
| 3005 | vp = fp->f_vnode; | |||
| 3006 | if (vp->v_type == VDIR) | |||
| 3007 | return (EPERM1); | |||
| 3008 | } | |||
| 3009 | } | |||
| 3010 | return (0); | |||
| 3011 | } | |||
| 3012 | ||||
| 3013 | /* | |||
| 3014 | * Common routine for kern_chroot() and jail_attach(). The caller is | |||
| 3015 | * responsible for invoking priv_check() and mac_vnode_check_chroot() to | |||
| 3016 | * authorize this operation. | |||
| 3017 | */ | |||
| 3018 | int | |||
| 3019 | pwd_chroot(struct thread *td, struct vnode *vp) | |||
| 3020 | { | |||
| 3021 | struct filedesc *fdp; | |||
| 3022 | struct vnode *oldvp; | |||
| 3023 | int error; | |||
| 3024 | ||||
| 3025 | fdp = td->td_proc->p_fd; | |||
| 3026 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 3027 | if (chroot_allow_open_directories == 0 || | |||
| 3028 | (chroot_allow_open_directories == 1 && fdp->fd_rdir != rootvnode)) { | |||
| 3029 | error = chroot_refuse_vdir_fds(fdp); | |||
| 3030 | if (error != 0) { | |||
| 3031 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 3032 | return (error); | |||
| 3033 | } | |||
| 3034 | } | |||
| 3035 | oldvp = fdp->fd_rdir; | |||
| 3036 | VREF(vp)vref(vp); | |||
| 3037 | fdp->fd_rdir = vp; | |||
| 3038 | if (fdp->fd_jdir == NULL((void *)0)) { | |||
| 3039 | VREF(vp)vref(vp); | |||
| 3040 | fdp->fd_jdir = vp; | |||
| 3041 | } | |||
| 3042 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 3043 | vrele(oldvp); | |||
| 3044 | return (0); | |||
| 3045 | } | |||
| 3046 | ||||
| 3047 | void | |||
| 3048 | pwd_chdir(struct thread *td, struct vnode *vp) | |||
| 3049 | { | |||
| 3050 | struct filedesc *fdp; | |||
| 3051 | struct vnode *oldvp; | |||
| 3052 | ||||
| 3053 | fdp = td->td_proc->p_fd; | |||
| 3054 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 3055 | VNASSERT(vp->v_usecount > 0, vp,do { } while (0) | |||
| 3056 | ("chdir to a vnode with zero usecount"))do { } while (0); | |||
| 3057 | oldvp = fdp->fd_cdir; | |||
| 3058 | fdp->fd_cdir = vp; | |||
| 3059 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 3060 | vrele(oldvp); | |||
| 3061 | } | |||
| 3062 | ||||
| 3063 | /* | |||
| 3064 | * Scan all active processes and prisons to see if any of them have a current | |||
| 3065 | * or root directory of `olddp'. If so, replace them with the new mount point. | |||
| 3066 | */ | |||
| 3067 | void | |||
| 3068 | mountcheckdirs(struct vnode *olddp, struct vnode *newdp) | |||
| 3069 | { | |||
| 3070 | struct filedesc *fdp; | |||
| 3071 | struct prison *pr; | |||
| 3072 | struct proc *p; | |||
| 3073 | int nrele; | |||
| 3074 | ||||
| 3075 | if (vrefcnt(olddp) == 1) | |||
| 3076 | return; | |||
| 3077 | nrele = 0; | |||
| 3078 | sx_slock(&allproc_lock)(void)__sx_slock(((&allproc_lock)), 0, (((void *)0)), (0) ); | |||
| 3079 | FOREACH_PROC_IN_SYSTEM(p)for (((p)) = (((&allproc))->lh_first); ((p)); ((p)) = ( (((p)))->p_list.le_next)) { | |||
| 3080 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 3081 | fdp = fdhold(p); | |||
| 3082 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3083 | if (fdp == NULL((void *)0)) | |||
| 3084 | continue; | |||
| 3085 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 3086 | if (fdp->fd_cdir == olddp) { | |||
| 3087 | vref(newdp); | |||
| 3088 | fdp->fd_cdir = newdp; | |||
| 3089 | nrele++; | |||
| 3090 | } | |||
| 3091 | if (fdp->fd_rdir == olddp) { | |||
| 3092 | vref(newdp); | |||
| 3093 | fdp->fd_rdir = newdp; | |||
| 3094 | nrele++; | |||
| 3095 | } | |||
| 3096 | if (fdp->fd_jdir == olddp) { | |||
| 3097 | vref(newdp); | |||
| 3098 | fdp->fd_jdir = newdp; | |||
| 3099 | nrele++; | |||
| 3100 | } | |||
| 3101 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 3102 | fddrop(fdp); | |||
| 3103 | } | |||
| 3104 | sx_sunlock(&allproc_lock)__sx_sunlock(((&allproc_lock)), (((void *)0)), (0)); | |||
| 3105 | if (rootvnode == olddp) { | |||
| 3106 | vref(newdp); | |||
| 3107 | rootvnode = newdp; | |||
| 3108 | nrele++; | |||
| 3109 | } | |||
| 3110 | mtx_lock(&prison0.pr_mtx)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &prison0.pr_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&prison0.pr_mtx)))))->mtx_lock, 0x00000004, (_tid)))) __mtx_lock_sleep(&(((((&prison0.pr_mtx)))) )->mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&prison0.pr_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 3111 | if (prison0.pr_root == olddp) { | |||
| 3112 | vref(newdp); | |||
| 3113 | prison0.pr_root = newdp; | |||
| 3114 | nrele++; | |||
| 3115 | } | |||
| 3116 | mtx_unlock(&prison0.pr_mtx)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& prison0.pr_mtx))))->lock_object.lo_data == 0) do { (void)0 ; do { if (__builtin_expect((sdt_lockstat___adaptive__release ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release ->id, (uintptr_t) (((&prison0.pr_mtx))), (uintptr_t) 0 , (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&prison0.pr_mtx))))->mtx_lock != _tid || !atomic_cmpset_long(&(((((&prison0.pr_mtx)))))-> mtx_lock, (_tid), 0x00000004)) __mtx_unlock_sleep(&(((((& prison0.pr_mtx)))))->mtx_lock, (((0))), ((((void *)0))), ( (0))); } while (0); | |||
| 3117 | sx_slock(&allprison_lock)(void)__sx_slock(((&allprison_lock)), 0, (((void *)0)), ( 0)); | |||
| 3118 | TAILQ_FOREACH(pr, &allprison, pr_list)for ((pr) = (((&allprison))->tqh_first); (pr); (pr) = ( ((pr))->pr_list.tqe_next)) { | |||
| 3119 | mtx_lock(&pr->pr_mtx)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &pr->pr_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&pr->pr_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&pr->pr_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&pr->pr_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 3120 | if (pr->pr_root == olddp) { | |||
| 3121 | vref(newdp); | |||
| 3122 | pr->pr_root = newdp; | |||
| 3123 | nrele++; | |||
| 3124 | } | |||
| 3125 | mtx_unlock(&pr->pr_mtx)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& pr->pr_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&pr->pr_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&pr->pr_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&pr->pr_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&pr->pr_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3126 | } | |||
| 3127 | sx_sunlock(&allprison_lock)__sx_sunlock(((&allprison_lock)), (((void *)0)), (0)); | |||
| 3128 | while (nrele--) | |||
| 3129 | vrele(olddp); | |||
| 3130 | } | |||
| 3131 | ||||
| 3132 | struct filedesc_to_leader * | |||
| 3133 | filedesc_to_leader_alloc(struct filedesc_to_leader *old, struct filedesc *fdp, struct proc *leader) | |||
| 3134 | { | |||
| 3135 | struct filedesc_to_leader *fdtol; | |||
| 3136 | ||||
| 3137 | fdtol = malloc(sizeof(struct filedesc_to_leader), | |||
| 3138 | M_FILEDESC_TO_LEADER, M_WAITOK0x0002); | |||
| 3139 | fdtol->fdl_refcount = 1; | |||
| 3140 | fdtol->fdl_holdcount = 0; | |||
| 3141 | fdtol->fdl_wakeup = 0; | |||
| 3142 | fdtol->fdl_leader = leader; | |||
| 3143 | if (old != NULL((void *)0)) { | |||
| 3144 | FILEDESC_XLOCK(fdp)(void)__sx_xlock(((&(fdp)->fd_sx)), (__curthread()), 0 , (((void *)0)), (0)); | |||
| 3145 | fdtol->fdl_next = old->fdl_next; | |||
| 3146 | fdtol->fdl_prev = old; | |||
| 3147 | old->fdl_next = fdtol; | |||
| 3148 | fdtol->fdl_next->fdl_prev = fdtol; | |||
| 3149 | FILEDESC_XUNLOCK(fdp)__sx_xunlock(((&(fdp)->fd_sx)), (__curthread()), (((void *)0)), (0)); | |||
| 3150 | } else { | |||
| 3151 | fdtol->fdl_next = fdtol; | |||
| 3152 | fdtol->fdl_prev = fdtol; | |||
| 3153 | } | |||
| 3154 | return (fdtol); | |||
| 3155 | } | |||
| 3156 | ||||
| 3157 | static int | |||
| 3158 | sysctl_kern_proc_nfds(SYSCTL_HANDLER_ARGSstruct sysctl_oid *oidp, void *arg1, intmax_t arg2, struct sysctl_req *req) | |||
| 3159 | { | |||
| 3160 | struct filedesc *fdp; | |||
| 3161 | int i, count, slots; | |||
| 3162 | ||||
| 3163 | if (*(int *)arg1 != 0) | |||
| 3164 | return (EINVAL22); | |||
| 3165 | ||||
| 3166 | fdp = curproc((__curthread())->td_proc)->p_fd; | |||
| 3167 | count = 0; | |||
| 3168 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3169 | slots = NDSLOTS(fdp->fd_lastfile + 1)(((fdp->fd_lastfile + 1) + (sizeof(u_long) * 8) - 1) / (sizeof (u_long) * 8)); | |||
| 3170 | for (i = 0; i < slots; i++) | |||
| 3171 | count += bitcountl(fdp->fd_map[i])__bitcount64((unsigned long)((u_long)(fdp->fd_map[i]))); | |||
| 3172 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3173 | ||||
| 3174 | return (SYSCTL_OUT(req, &count, sizeof(count))(req->oldfunc)(req, &count, sizeof(count))); | |||
| 3175 | } | |||
| 3176 | ||||
| 3177 | static SYSCTL_NODE(_kern_proc, KERN_PROC_NFDS, nfds,struct sysctl_oid sysctl___kern_proc_nfds = { .oid_parent = ( (&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (43), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("nfds"), .oid_handler = (sysctl_kern_proc_nfds), .oid_fmt = ("N"), .oid_descr = "Number of open file descriptors" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_nfds __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_nfds); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3178 | CTLFLAG_RD|CTLFLAG_MPSAFE, sysctl_kern_proc_nfds,struct sysctl_oid sysctl___kern_proc_nfds = { .oid_parent = ( (&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (43), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("nfds"), .oid_handler = (sysctl_kern_proc_nfds), .oid_fmt = ("N"), .oid_descr = "Number of open file descriptors" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_nfds __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_nfds); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3179 | "Number of open file descriptors")struct sysctl_oid sysctl___kern_proc_nfds = { .oid_parent = ( (&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (43), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("nfds"), .oid_handler = (sysctl_kern_proc_nfds), .oid_fmt = ("N"), .oid_descr = "Number of open file descriptors" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_nfds __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_nfds); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed"); | |||
| 3180 | ||||
| 3181 | /* | |||
| 3182 | * Get file structures globally. | |||
| 3183 | */ | |||
| 3184 | static int | |||
| 3185 | sysctl_kern_file(SYSCTL_HANDLER_ARGSstruct sysctl_oid *oidp, void *arg1, intmax_t arg2, struct sysctl_req *req) | |||
| 3186 | { | |||
| 3187 | struct xfile xf; | |||
| 3188 | struct filedesc *fdp; | |||
| 3189 | struct file *fp; | |||
| 3190 | struct proc *p; | |||
| 3191 | int error, n; | |||
| 3192 | ||||
| 3193 | error = sysctl_wire_old_buffer(req, 0); | |||
| 3194 | if (error != 0) | |||
| 3195 | return (error); | |||
| 3196 | if (req->oldptr == NULL((void *)0)) { | |||
| 3197 | n = 0; | |||
| 3198 | sx_slock(&allproc_lock)(void)__sx_slock(((&allproc_lock)), 0, (((void *)0)), (0) ); | |||
| 3199 | FOREACH_PROC_IN_SYSTEM(p)for (((p)) = (((&allproc))->lh_first); ((p)); ((p)) = ( (((p)))->p_list.le_next)) { | |||
| 3200 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 3201 | if (p->p_state == PRS_NEW) { | |||
| 3202 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3203 | continue; | |||
| 3204 | } | |||
| 3205 | fdp = fdhold(p); | |||
| 3206 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3207 | if (fdp == NULL((void *)0)) | |||
| 3208 | continue; | |||
| 3209 | /* overestimates sparse tables. */ | |||
| 3210 | if (fdp->fd_lastfile > 0) | |||
| 3211 | n += fdp->fd_lastfile; | |||
| 3212 | fddrop(fdp); | |||
| 3213 | } | |||
| 3214 | sx_sunlock(&allproc_lock)__sx_sunlock(((&allproc_lock)), (((void *)0)), (0)); | |||
| 3215 | return (SYSCTL_OUT(req, 0, n * sizeof(xf))(req->oldfunc)(req, 0, n * sizeof(xf))); | |||
| 3216 | } | |||
| 3217 | error = 0; | |||
| 3218 | bzero(&xf, sizeof(xf)); | |||
| 3219 | xf.xf_size = sizeof(xf); | |||
| 3220 | sx_slock(&allproc_lock)(void)__sx_slock(((&allproc_lock)), 0, (((void *)0)), (0) ); | |||
| 3221 | FOREACH_PROC_IN_SYSTEM(p)for (((p)) = (((&allproc))->lh_first); ((p)); ((p)) = ( (((p)))->p_list.le_next)) { | |||
| 3222 | PROC_LOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((( &(p)->p_mtx))))->mtx_lock != 0x00000004 || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, 0x00000004, ( _tid)))) __mtx_lock_sleep(&(((((&(p)->p_mtx)))))-> mtx_lock, _tid, (((0))), ((((void *)0))), ((0))); else do { ( void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__acquire ->id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__acquire ->id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); } while (0); | |||
| 3223 | if (p->p_state == PRS_NEW) { | |||
| 3224 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3225 | continue; | |||
| 3226 | } | |||
| 3227 | if (p_cansee(req->td, p) != 0) { | |||
| 3228 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3229 | continue; | |||
| 3230 | } | |||
| 3231 | xf.xf_pid = p->p_pid; | |||
| 3232 | xf.xf_uid = p->p_ucred->cr_uid; | |||
| 3233 | fdp = fdhold(p); | |||
| 3234 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3235 | if (fdp == NULL((void *)0)) | |||
| 3236 | continue; | |||
| 3237 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3238 | for (n = 0; fdp->fd_refcnt > 0 && n <= fdp->fd_lastfile; ++n) { | |||
| 3239 | if ((fp = fdp->fd_ofilesfd_files->fdt_ofiles[n].fde_file) == NULL((void *)0)) | |||
| 3240 | continue; | |||
| 3241 | xf.xf_fd = n; | |||
| 3242 | xf.xf_file = fp; | |||
| 3243 | xf.xf_data = fp->f_data; | |||
| 3244 | xf.xf_vnode = fp->f_vnode; | |||
| 3245 | xf.xf_type = fp->f_type; | |||
| 3246 | xf.xf_count = fp->f_count; | |||
| 3247 | xf.xf_msgcount = 0; | |||
| 3248 | xf.xf_offset = foffset_get(fp); | |||
| 3249 | xf.xf_flag = fp->f_flag; | |||
| 3250 | error = SYSCTL_OUT(req, &xf, sizeof(xf))(req->oldfunc)(req, &xf, sizeof(xf)); | |||
| 3251 | if (error) | |||
| 3252 | break; | |||
| 3253 | } | |||
| 3254 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3255 | fddrop(fdp); | |||
| 3256 | if (error) | |||
| 3257 | break; | |||
| 3258 | } | |||
| 3259 | sx_sunlock(&allproc_lock)__sx_sunlock(((&allproc_lock)), (((void *)0)), (0)); | |||
| 3260 | return (error); | |||
| 3261 | } | |||
| 3262 | ||||
| 3263 | SYSCTL_PROC(_kern, KERN_FILE, file, CTLTYPE_OPAQUE|CTLFLAG_RD|CTLFLAG_MPSAFE,static struct sysctl_oid sysctl___kern_file = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (15), .oid_kind = ((5|0x80000000 |0x00040000)), .oid_arg1 = (0), .oid_arg2 = (0), .oid_name = ( "file"), .oid_handler = (sysctl_kern_file), .oid_fmt = ("S,xfile" ), .oid_descr = "Entire file table" }; __asm__(".globl " "__start_set_sysctl_set" ); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_file __attribute__ ((__section__("set_" "sysctl_set"))) __attribute__((__used__) ) = &(sysctl___kern_file); _Static_assert(((5|0x80000000| 0x00040000) & 0xf) != 0, "compile-time assertion failed") | |||
| 3264 | 0, 0, sysctl_kern_file, "S,xfile", "Entire file table")static struct sysctl_oid sysctl___kern_file = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (15), .oid_kind = ((5|0x80000000 |0x00040000)), .oid_arg1 = (0), .oid_arg2 = (0), .oid_name = ( "file"), .oid_handler = (sysctl_kern_file), .oid_fmt = ("S,xfile" ), .oid_descr = "Entire file table" }; __asm__(".globl " "__start_set_sysctl_set" ); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_file __attribute__ ((__section__("set_" "sysctl_set"))) __attribute__((__used__) ) = &(sysctl___kern_file); _Static_assert(((5|0x80000000| 0x00040000) & 0xf) != 0, "compile-time assertion failed"); | |||
| 3265 | ||||
| 3266 | #ifdef KINFO_FILE_SIZE1392 | |||
| 3267 | CTASSERT(sizeof(struct kinfo_file) == KINFO_FILE_SIZE)_Static_assert(sizeof(struct kinfo_file) == 1392, "compile-time assertion failed" ); | |||
| 3268 | #endif | |||
| 3269 | ||||
| 3270 | static int | |||
| 3271 | xlate_fflags(int fflags) | |||
| 3272 | { | |||
| 3273 | static const struct { | |||
| 3274 | int fflag; | |||
| 3275 | int kf_fflag; | |||
| 3276 | } fflags_table[] = { | |||
| 3277 | { FAPPEND0x0008, KF_FLAG_APPEND0x00000004 }, | |||
| 3278 | { FASYNC0x0040, KF_FLAG_ASYNC0x00000008 }, | |||
| 3279 | { FFSYNC0x0080, KF_FLAG_FSYNC0x00000010 }, | |||
| 3280 | { FHASLOCK0x4000, KF_FLAG_HASLOCK0x00000080 }, | |||
| 3281 | { FNONBLOCK0x0004, KF_FLAG_NONBLOCK0x00000020 }, | |||
| 3282 | { FREAD0x0001, KF_FLAG_READ0x00000001 }, | |||
| 3283 | { FWRITE0x0002, KF_FLAG_WRITE0x00000002 }, | |||
| 3284 | { O_CREAT0x0200, KF_FLAG_CREAT0x00000800 }, | |||
| 3285 | { O_DIRECT0x00010000, KF_FLAG_DIRECT0x00000040 }, | |||
| 3286 | { O_EXCL0x0800, KF_FLAG_EXCL0x00002000 }, | |||
| 3287 | { O_EXEC0x00040000, KF_FLAG_EXEC0x00004000 }, | |||
| 3288 | { O_EXLOCK0x0020, KF_FLAG_EXLOCK0x00000200 }, | |||
| 3289 | { O_NOFOLLOW0x0100, KF_FLAG_NOFOLLOW0x00000400 }, | |||
| 3290 | { O_SHLOCK0x0010, KF_FLAG_SHLOCK0x00000100 }, | |||
| 3291 | { O_TRUNC0x0400, KF_FLAG_TRUNC0x00001000 } | |||
| 3292 | }; | |||
| 3293 | unsigned int i; | |||
| 3294 | int kflags; | |||
| 3295 | ||||
| 3296 | kflags = 0; | |||
| 3297 | for (i = 0; i < nitems(fflags_table)(sizeof((fflags_table)) / sizeof((fflags_table)[0])); i++) | |||
| 3298 | if (fflags & fflags_table[i].fflag) | |||
| 3299 | kflags |= fflags_table[i].kf_fflag; | |||
| 3300 | return (kflags); | |||
| 3301 | } | |||
| 3302 | ||||
| 3303 | /* Trim unused data from kf_path by truncating the structure size. */ | |||
| 3304 | static void | |||
| 3305 | pack_kinfo(struct kinfo_file *kif) | |||
| 3306 | { | |||
| 3307 | ||||
| 3308 | kif->kf_structsize = offsetof(struct kinfo_file, kf_path)__builtin_offsetof(struct kinfo_file, kf_path) + | |||
| 3309 | strlen(kif->kf_path) + 1; | |||
| 3310 | kif->kf_structsize = roundup(kif->kf_structsize, sizeof(uint64_t))((((kif->kf_structsize)+((sizeof(uint64_t))-1))/(sizeof(uint64_t )))*(sizeof(uint64_t))); | |||
| 3311 | } | |||
| 3312 | ||||
| 3313 | static void | |||
| 3314 | export_file_to_kinfo(struct file *fp, int fd, cap_rights_t *rightsp, | |||
| 3315 | struct kinfo_file *kif, struct filedesc *fdp, int flags) | |||
| 3316 | { | |||
| 3317 | int error; | |||
| 3318 | ||||
| 3319 | bzero(kif, sizeof(*kif)); | |||
| 3320 | ||||
| 3321 | /* Set a default type to allow for empty fill_kinfo() methods. */ | |||
| 3322 | kif->kf_type = KF_TYPE_UNKNOWN255; | |||
| 3323 | kif->kf_flags = xlate_fflags(fp->f_flag); | |||
| 3324 | if (rightsp != NULL((void *)0)) | |||
| 3325 | kif->kf_cap_rights = *rightsp; | |||
| 3326 | else | |||
| 3327 | cap_rights_init(&kif->kf_cap_rights)__cap_rights_init(0, &kif->kf_cap_rights, 0ULL); | |||
| 3328 | kif->kf_fd = fd; | |||
| 3329 | kif->kf_ref_count = fp->f_count; | |||
| 3330 | kif->kf_offset = foffset_get(fp); | |||
| 3331 | ||||
| 3332 | /* | |||
| 3333 | * This may drop the filedesc lock, so the 'fp' cannot be | |||
| 3334 | * accessed after this call. | |||
| 3335 | */ | |||
| 3336 | error = fo_fill_kinfo(fp, kif, fdp); | |||
| 3337 | if (error == 0) | |||
| 3338 | kif->kf_status |= KF_ATTR_VALID0x0001; | |||
| 3339 | if ((flags & KERN_FILEDESC_PACK_KINFO0x00000001U) != 0) | |||
| 3340 | pack_kinfo(kif); | |||
| 3341 | else | |||
| 3342 | kif->kf_structsize = roundup2(sizeof(*kif), sizeof(uint64_t))(((sizeof(*kif))+((sizeof(uint64_t))-1))&(~((sizeof(uint64_t ))-1))); | |||
| 3343 | } | |||
| 3344 | ||||
| 3345 | static void | |||
| 3346 | export_vnode_to_kinfo(struct vnode *vp, int fd, int fflags, | |||
| 3347 | struct kinfo_file *kif, int flags) | |||
| 3348 | { | |||
| 3349 | int error; | |||
| 3350 | ||||
| 3351 | bzero(kif, sizeof(*kif)); | |||
| 3352 | ||||
| 3353 | kif->kf_type = KF_TYPE_VNODE1; | |||
| 3354 | error = vn_fill_kinfo_vnode(vp, kif); | |||
| 3355 | if (error == 0) | |||
| 3356 | kif->kf_status |= KF_ATTR_VALID0x0001; | |||
| 3357 | kif->kf_flags = xlate_fflags(fflags); | |||
| 3358 | cap_rights_init(&kif->kf_cap_rights)__cap_rights_init(0, &kif->kf_cap_rights, 0ULL); | |||
| 3359 | kif->kf_fd = fd; | |||
| 3360 | kif->kf_ref_count = -1; | |||
| 3361 | kif->kf_offset = -1; | |||
| 3362 | if ((flags & KERN_FILEDESC_PACK_KINFO0x00000001U) != 0) | |||
| 3363 | pack_kinfo(kif); | |||
| 3364 | else | |||
| 3365 | kif->kf_structsize = roundup2(sizeof(*kif), sizeof(uint64_t))(((sizeof(*kif))+((sizeof(uint64_t))-1))&(~((sizeof(uint64_t ))-1))); | |||
| 3366 | vrele(vp); | |||
| 3367 | } | |||
| 3368 | ||||
| 3369 | struct export_fd_buf { | |||
| 3370 | struct filedesc *fdp; | |||
| 3371 | struct sbuf *sb; | |||
| 3372 | ssize_t remainder; | |||
| 3373 | struct kinfo_file kif; | |||
| 3374 | int flags; | |||
| 3375 | }; | |||
| 3376 | ||||
| 3377 | static int | |||
| 3378 | export_kinfo_to_sb(struct export_fd_buf *efbuf) | |||
| 3379 | { | |||
| 3380 | struct kinfo_file *kif; | |||
| 3381 | ||||
| 3382 | kif = &efbuf->kif; | |||
| 3383 | if (efbuf->remainder != -1) { | |||
| 3384 | if (efbuf->remainder < kif->kf_structsize) { | |||
| 3385 | /* Terminate export. */ | |||
| 3386 | efbuf->remainder = 0; | |||
| 3387 | return (0); | |||
| 3388 | } | |||
| 3389 | efbuf->remainder -= kif->kf_structsize; | |||
| 3390 | } | |||
| 3391 | return (sbuf_bcat(efbuf->sb, kif, kif->kf_structsize) == 0 ? 0 : ENOMEM12); | |||
| 3392 | } | |||
| 3393 | ||||
| 3394 | static int | |||
| 3395 | export_file_to_sb(struct file *fp, int fd, cap_rights_t *rightsp, | |||
| 3396 | struct export_fd_buf *efbuf) | |||
| 3397 | { | |||
| 3398 | int error; | |||
| 3399 | ||||
| 3400 | if (efbuf->remainder == 0) | |||
| 3401 | return (0); | |||
| 3402 | export_file_to_kinfo(fp, fd, rightsp, &efbuf->kif, efbuf->fdp, | |||
| 3403 | efbuf->flags); | |||
| 3404 | FILEDESC_SUNLOCK(efbuf->fdp)__sx_sunlock(((&(efbuf->fdp)->fd_sx)), (((void *)0) ), (0)); | |||
| 3405 | error = export_kinfo_to_sb(efbuf); | |||
| 3406 | FILEDESC_SLOCK(efbuf->fdp)(void)__sx_slock(((&(efbuf->fdp)->fd_sx)), 0, (((void *)0)), (0)); | |||
| 3407 | return (error); | |||
| 3408 | } | |||
| 3409 | ||||
| 3410 | static int | |||
| 3411 | export_vnode_to_sb(struct vnode *vp, int fd, int fflags, | |||
| 3412 | struct export_fd_buf *efbuf) | |||
| 3413 | { | |||
| 3414 | int error; | |||
| 3415 | ||||
| 3416 | if (efbuf->remainder == 0) | |||
| 3417 | return (0); | |||
| 3418 | if (efbuf->fdp != NULL((void *)0)) | |||
| 3419 | FILEDESC_SUNLOCK(efbuf->fdp)__sx_sunlock(((&(efbuf->fdp)->fd_sx)), (((void *)0) ), (0)); | |||
| 3420 | export_vnode_to_kinfo(vp, fd, fflags, &efbuf->kif, efbuf->flags); | |||
| 3421 | error = export_kinfo_to_sb(efbuf); | |||
| 3422 | if (efbuf->fdp != NULL((void *)0)) | |||
| 3423 | FILEDESC_SLOCK(efbuf->fdp)(void)__sx_slock(((&(efbuf->fdp)->fd_sx)), 0, (((void *)0)), (0)); | |||
| 3424 | return (error); | |||
| 3425 | } | |||
| 3426 | ||||
| 3427 | /* | |||
| 3428 | * Store a process file descriptor information to sbuf. | |||
| 3429 | * | |||
| 3430 | * Takes a locked proc as argument, and returns with the proc unlocked. | |||
| 3431 | */ | |||
| 3432 | int | |||
| 3433 | kern_proc_filedesc_out(struct proc *p, struct sbuf *sb, ssize_t maxlen, | |||
| 3434 | int flags) | |||
| 3435 | { | |||
| 3436 | struct file *fp; | |||
| 3437 | struct filedesc *fdp; | |||
| 3438 | struct export_fd_buf *efbuf; | |||
| 3439 | struct vnode *cttyvp, *textvp, *tracevp; | |||
| 3440 | int error, i; | |||
| 3441 | cap_rights_t rights; | |||
| 3442 | ||||
| 3443 | PROC_LOCK_ASSERT(p, MA_OWNED)(void)0; | |||
| 3444 | ||||
| 3445 | /* ktrace vnode */ | |||
| 3446 | tracevp = p->p_tracevp; | |||
| 3447 | if (tracevp != NULL((void *)0)) | |||
| 3448 | vref(tracevp); | |||
| 3449 | /* text vnode */ | |||
| 3450 | textvp = p->p_textvp; | |||
| 3451 | if (textvp != NULL((void *)0)) | |||
| 3452 | vref(textvp); | |||
| 3453 | /* Controlling tty. */ | |||
| 3454 | cttyvp = NULL((void *)0); | |||
| 3455 | if (p->p_pgrp != NULL((void *)0) && p->p_pgrp->pg_session != NULL((void *)0)) { | |||
| 3456 | cttyvp = p->p_pgrp->pg_session->s_ttyvp; | |||
| 3457 | if (cttyvp != NULL((void *)0)) | |||
| 3458 | vref(cttyvp); | |||
| 3459 | } | |||
| 3460 | fdp = fdhold(p); | |||
| 3461 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3462 | efbuf = malloc(sizeof(*efbuf), M_TEMP, M_WAITOK0x0002); | |||
| 3463 | efbuf->fdp = NULL((void *)0); | |||
| 3464 | efbuf->sb = sb; | |||
| 3465 | efbuf->remainder = maxlen; | |||
| 3466 | efbuf->flags = flags; | |||
| 3467 | if (tracevp != NULL((void *)0)) | |||
| 3468 | export_vnode_to_sb(tracevp, KF_FD_TYPE_TRACE-4, FREAD0x0001 | FWRITE0x0002, | |||
| 3469 | efbuf); | |||
| 3470 | if (textvp != NULL((void *)0)) | |||
| 3471 | export_vnode_to_sb(textvp, KF_FD_TYPE_TEXT-5, FREAD0x0001, efbuf); | |||
| 3472 | if (cttyvp != NULL((void *)0)) | |||
| 3473 | export_vnode_to_sb(cttyvp, KF_FD_TYPE_CTTY-6, FREAD0x0001 | FWRITE0x0002, | |||
| 3474 | efbuf); | |||
| 3475 | error = 0; | |||
| 3476 | if (fdp == NULL((void *)0)) | |||
| 3477 | goto fail; | |||
| 3478 | efbuf->fdp = fdp; | |||
| 3479 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3480 | /* working directory */ | |||
| 3481 | if (fdp->fd_cdir != NULL((void *)0)) { | |||
| 3482 | vref(fdp->fd_cdir); | |||
| 3483 | export_vnode_to_sb(fdp->fd_cdir, KF_FD_TYPE_CWD-1, FREAD0x0001, efbuf); | |||
| 3484 | } | |||
| 3485 | /* root directory */ | |||
| 3486 | if (fdp->fd_rdir != NULL((void *)0)) { | |||
| 3487 | vref(fdp->fd_rdir); | |||
| 3488 | export_vnode_to_sb(fdp->fd_rdir, KF_FD_TYPE_ROOT-2, FREAD0x0001, efbuf); | |||
| 3489 | } | |||
| 3490 | /* jail directory */ | |||
| 3491 | if (fdp->fd_jdir != NULL((void *)0)) { | |||
| 3492 | vref(fdp->fd_jdir); | |||
| 3493 | export_vnode_to_sb(fdp->fd_jdir, KF_FD_TYPE_JAIL-3, FREAD0x0001, efbuf); | |||
| 3494 | } | |||
| 3495 | for (i = 0; fdp->fd_refcnt > 0 && i <= fdp->fd_lastfile; i++) { | |||
| 3496 | if ((fp = fdp->fd_ofilesfd_files->fdt_ofiles[i].fde_file) == NULL((void *)0)) | |||
| 3497 | continue; | |||
| 3498 | #ifdef CAPABILITIES1 | |||
| 3499 | rights = *cap_rights(fdp, i); | |||
| 3500 | #else /* !CAPABILITIES */ | |||
| 3501 | cap_rights_init(&rights)__cap_rights_init(0, &rights, 0ULL); | |||
| 3502 | #endif | |||
| 3503 | /* | |||
| 3504 | * Create sysctl entry. It is OK to drop the filedesc | |||
| 3505 | * lock inside of export_file_to_sb() as we will | |||
| 3506 | * re-validate and re-evaluate its properties when the | |||
| 3507 | * loop continues. | |||
| 3508 | */ | |||
| 3509 | error = export_file_to_sb(fp, i, &rights, efbuf); | |||
| 3510 | if (error != 0 || efbuf->remainder == 0) | |||
| 3511 | break; | |||
| 3512 | } | |||
| 3513 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3514 | fddrop(fdp); | |||
| 3515 | fail: | |||
| 3516 | free(efbuf, M_TEMP); | |||
| 3517 | return (error); | |||
| 3518 | } | |||
| 3519 | ||||
| 3520 | #define FILEDESC_SBUF_SIZE(sizeof(struct kinfo_file) * 5) (sizeof(struct kinfo_file) * 5) | |||
| 3521 | ||||
| 3522 | /* | |||
| 3523 | * Get per-process file descriptors for use by procstat(1), et al. | |||
| 3524 | */ | |||
| 3525 | static int | |||
| 3526 | sysctl_kern_proc_filedesc(SYSCTL_HANDLER_ARGSstruct sysctl_oid *oidp, void *arg1, intmax_t arg2, struct sysctl_req *req) | |||
| 3527 | { | |||
| 3528 | struct sbuf sb; | |||
| 3529 | struct proc *p; | |||
| 3530 | ssize_t maxlen; | |||
| 3531 | int error, error2, *name; | |||
| 3532 | ||||
| 3533 | name = (int *)arg1; | |||
| 3534 | ||||
| 3535 | sbuf_new_for_sysctl(&sb, NULL((void *)0), FILEDESC_SBUF_SIZE(sizeof(struct kinfo_file) * 5), req); | |||
| 3536 | sbuf_clear_flags(&sb, SBUF_INCLUDENUL0x00000002); | |||
| 3537 | error = pget((pid_t)name[0], PGET_CANDEBUG0x00004 | PGET_NOTWEXIT0x00010, &p); | |||
| 3538 | if (error != 0) { | |||
| 3539 | sbuf_delete(&sb); | |||
| 3540 | return (error); | |||
| 3541 | } | |||
| 3542 | maxlen = req->oldptr != NULL((void *)0) ? req->oldlen : -1; | |||
| 3543 | error = kern_proc_filedesc_out(p, &sb, maxlen, | |||
| 3544 | KERN_FILEDESC_PACK_KINFO0x00000001U); | |||
| 3545 | error2 = sbuf_finish(&sb); | |||
| 3546 | sbuf_delete(&sb); | |||
| 3547 | return (error != 0 ? error : error2); | |||
| 3548 | } | |||
| 3549 | ||||
| 3550 | #ifdef KINFO_OFILE_SIZE1328 | |||
| 3551 | CTASSERT(sizeof(struct kinfo_ofile) == KINFO_OFILE_SIZE)_Static_assert(sizeof(struct kinfo_ofile) == 1328, "compile-time assertion failed" ); | |||
| 3552 | #endif | |||
| 3553 | ||||
| 3554 | #ifdef COMPAT_FREEBSD71 | |||
| 3555 | static void | |||
| 3556 | kinfo_to_okinfo(struct kinfo_file *kif, struct kinfo_ofile *okif) | |||
| 3557 | { | |||
| 3558 | ||||
| 3559 | okif->kf_structsize = sizeof(*okif); | |||
| 3560 | okif->kf_type = kif->kf_type; | |||
| 3561 | okif->kf_fd = kif->kf_fd; | |||
| 3562 | okif->kf_ref_count = kif->kf_ref_count; | |||
| 3563 | okif->kf_flags = kif->kf_flags & (KF_FLAG_READ0x00000001 | KF_FLAG_WRITE0x00000002 | | |||
| 3564 | KF_FLAG_APPEND0x00000004 | KF_FLAG_ASYNC0x00000008 | KF_FLAG_FSYNC0x00000010 | KF_FLAG_NONBLOCK0x00000020 | | |||
| 3565 | KF_FLAG_DIRECT0x00000040 | KF_FLAG_HASLOCK0x00000080); | |||
| 3566 | okif->kf_offset = kif->kf_offset; | |||
| 3567 | okif->kf_vnode_type = kif->kf_vnode_type; | |||
| 3568 | okif->kf_sock_domain = kif->kf_sock_domain; | |||
| 3569 | okif->kf_sock_type = kif->kf_sock_type; | |||
| 3570 | okif->kf_sock_protocol = kif->kf_sock_protocol; | |||
| 3571 | strlcpy(okif->kf_path, kif->kf_path, sizeof(okif->kf_path)); | |||
| 3572 | okif->kf_sa_local = kif->kf_sa_local; | |||
| 3573 | okif->kf_sa_peer = kif->kf_sa_peer; | |||
| 3574 | } | |||
| 3575 | ||||
| 3576 | static int | |||
| 3577 | export_vnode_for_osysctl(struct vnode *vp, int type, struct kinfo_file *kif, | |||
| 3578 | struct kinfo_ofile *okif, struct filedesc *fdp, struct sysctl_req *req) | |||
| 3579 | { | |||
| 3580 | int error; | |||
| 3581 | ||||
| 3582 | vref(vp); | |||
| 3583 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3584 | export_vnode_to_kinfo(vp, type, 0, kif, KERN_FILEDESC_PACK_KINFO0x00000001U); | |||
| 3585 | kinfo_to_okinfo(kif, okif); | |||
| 3586 | error = SYSCTL_OUT(req, okif, sizeof(*okif))(req->oldfunc)(req, okif, sizeof(*okif)); | |||
| 3587 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3588 | return (error); | |||
| 3589 | } | |||
| 3590 | ||||
| 3591 | /* | |||
| 3592 | * Get per-process file descriptors for use by procstat(1), et al. | |||
| 3593 | */ | |||
| 3594 | static int | |||
| 3595 | sysctl_kern_proc_ofiledesc(SYSCTL_HANDLER_ARGSstruct sysctl_oid *oidp, void *arg1, intmax_t arg2, struct sysctl_req *req) | |||
| 3596 | { | |||
| 3597 | struct kinfo_ofile *okif; | |||
| 3598 | struct kinfo_file *kif; | |||
| 3599 | struct filedesc *fdp; | |||
| 3600 | int error, i, *name; | |||
| 3601 | struct file *fp; | |||
| 3602 | struct proc *p; | |||
| 3603 | ||||
| 3604 | name = (int *)arg1; | |||
| 3605 | error = pget((pid_t)name[0], PGET_CANDEBUG0x00004 | PGET_NOTWEXIT0x00010, &p); | |||
| 3606 | if (error != 0) | |||
| 3607 | return (error); | |||
| 3608 | fdp = fdhold(p); | |||
| 3609 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3610 | if (fdp == NULL((void *)0)) | |||
| 3611 | return (ENOENT2); | |||
| 3612 | kif = malloc(sizeof(*kif), M_TEMP, M_WAITOK0x0002); | |||
| 3613 | okif = malloc(sizeof(*okif), M_TEMP, M_WAITOK0x0002); | |||
| 3614 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3615 | if (fdp->fd_cdir != NULL((void *)0)) | |||
| 3616 | export_vnode_for_osysctl(fdp->fd_cdir, KF_FD_TYPE_CWD-1, kif, | |||
| 3617 | okif, fdp, req); | |||
| 3618 | if (fdp->fd_rdir != NULL((void *)0)) | |||
| 3619 | export_vnode_for_osysctl(fdp->fd_rdir, KF_FD_TYPE_ROOT-2, kif, | |||
| 3620 | okif, fdp, req); | |||
| 3621 | if (fdp->fd_jdir != NULL((void *)0)) | |||
| 3622 | export_vnode_for_osysctl(fdp->fd_jdir, KF_FD_TYPE_JAIL-3, kif, | |||
| 3623 | okif, fdp, req); | |||
| 3624 | for (i = 0; fdp->fd_refcnt > 0 && i <= fdp->fd_lastfile; i++) { | |||
| 3625 | if ((fp = fdp->fd_ofilesfd_files->fdt_ofiles[i].fde_file) == NULL((void *)0)) | |||
| 3626 | continue; | |||
| 3627 | export_file_to_kinfo(fp, i, NULL((void *)0), kif, fdp, | |||
| 3628 | KERN_FILEDESC_PACK_KINFO0x00000001U); | |||
| 3629 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3630 | kinfo_to_okinfo(kif, okif); | |||
| 3631 | error = SYSCTL_OUT(req, okif, sizeof(*okif))(req->oldfunc)(req, okif, sizeof(*okif)); | |||
| 3632 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3633 | if (error) | |||
| 3634 | break; | |||
| 3635 | } | |||
| 3636 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3637 | fddrop(fdp); | |||
| 3638 | free(kif, M_TEMP); | |||
| 3639 | free(okif, M_TEMP); | |||
| 3640 | return (0); | |||
| 3641 | } | |||
| 3642 | ||||
| 3643 | static SYSCTL_NODE(_kern_proc, KERN_PROC_OFILEDESC, ofiledesc,struct sysctl_oid sysctl___kern_proc_ofiledesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (14), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("ofiledesc"), .oid_handler = (sysctl_kern_proc_ofiledesc) , .oid_fmt = ("N"), .oid_descr = "Process ofiledesc entries" } ; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_ofiledesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_ofiledesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3644 | CTLFLAG_RD|CTLFLAG_MPSAFE, sysctl_kern_proc_ofiledesc,struct sysctl_oid sysctl___kern_proc_ofiledesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (14), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("ofiledesc"), .oid_handler = (sysctl_kern_proc_ofiledesc) , .oid_fmt = ("N"), .oid_descr = "Process ofiledesc entries" } ; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_ofiledesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_ofiledesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3645 | "Process ofiledesc entries")struct sysctl_oid sysctl___kern_proc_ofiledesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (14), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("ofiledesc"), .oid_handler = (sysctl_kern_proc_ofiledesc) , .oid_fmt = ("N"), .oid_descr = "Process ofiledesc entries" } ; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_ofiledesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_ofiledesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed"); | |||
| 3646 | #endif /* COMPAT_FREEBSD7 */ | |||
| 3647 | ||||
| 3648 | int | |||
| 3649 | vntype_to_kinfo(int vtype) | |||
| 3650 | { | |||
| 3651 | struct { | |||
| 3652 | int vtype; | |||
| 3653 | int kf_vtype; | |||
| 3654 | } vtypes_table[] = { | |||
| 3655 | { VBAD, KF_VTYPE_VBAD8 }, | |||
| 3656 | { VBLK, KF_VTYPE_VBLK3 }, | |||
| 3657 | { VCHR, KF_VTYPE_VCHR4 }, | |||
| 3658 | { VDIR, KF_VTYPE_VDIR2 }, | |||
| 3659 | { VFIFO, KF_VTYPE_VFIFO7 }, | |||
| 3660 | { VLNK, KF_VTYPE_VLNK5 }, | |||
| 3661 | { VNON, KF_VTYPE_VNON0 }, | |||
| 3662 | { VREG, KF_VTYPE_VREG1 }, | |||
| 3663 | { VSOCK, KF_VTYPE_VSOCK6 } | |||
| 3664 | }; | |||
| 3665 | unsigned int i; | |||
| 3666 | ||||
| 3667 | /* | |||
| 3668 | * Perform vtype translation. | |||
| 3669 | */ | |||
| 3670 | for (i = 0; i < nitems(vtypes_table)(sizeof((vtypes_table)) / sizeof((vtypes_table)[0])); i++) | |||
| 3671 | if (vtypes_table[i].vtype == vtype) | |||
| 3672 | return (vtypes_table[i].kf_vtype); | |||
| 3673 | ||||
| 3674 | return (KF_VTYPE_UNKNOWN255); | |||
| 3675 | } | |||
| 3676 | ||||
| 3677 | static SYSCTL_NODE(_kern_proc, KERN_PROC_FILEDESC, filedesc,struct sysctl_oid sysctl___kern_proc_filedesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (33), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("filedesc"), .oid_handler = (sysctl_kern_proc_filedesc), . oid_fmt = ("N"), .oid_descr = "Process filedesc entries" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_filedesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_filedesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3678 | CTLFLAG_RD|CTLFLAG_MPSAFE, sysctl_kern_proc_filedesc,struct sysctl_oid sysctl___kern_proc_filedesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (33), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("filedesc"), .oid_handler = (sysctl_kern_proc_filedesc), . oid_fmt = ("N"), .oid_descr = "Process filedesc entries" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_filedesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_filedesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3679 | "Process filedesc entries")struct sysctl_oid sysctl___kern_proc_filedesc = { .oid_parent = ((&(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (33), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("filedesc"), .oid_handler = (sysctl_kern_proc_filedesc), . oid_fmt = ("N"), .oid_descr = "Process filedesc entries" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_filedesc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_filedesc); _Static_assert (((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed"); | |||
| 3680 | ||||
| 3681 | /* | |||
| 3682 | * Store a process current working directory information to sbuf. | |||
| 3683 | * | |||
| 3684 | * Takes a locked proc as argument, and returns with the proc unlocked. | |||
| 3685 | */ | |||
| 3686 | int | |||
| 3687 | kern_proc_cwd_out(struct proc *p, struct sbuf *sb, ssize_t maxlen) | |||
| 3688 | { | |||
| 3689 | struct filedesc *fdp; | |||
| 3690 | struct export_fd_buf *efbuf; | |||
| 3691 | int error; | |||
| 3692 | ||||
| 3693 | PROC_LOCK_ASSERT(p, MA_OWNED)(void)0; | |||
| 3694 | ||||
| 3695 | fdp = fdhold(p); | |||
| 3696 | PROC_UNLOCK(p)do { uintptr_t _tid = (uintptr_t)((__curthread())); if (((((& (p)->p_mtx))))->lock_object.lo_data == 0) do { (void)0; do { if (__builtin_expect((sdt_lockstat___adaptive__release-> id), 0)) (*sdt_probe_func)(sdt_lockstat___adaptive__release-> id, (uintptr_t) (((&(p)->p_mtx))), (uintptr_t) 0, (uintptr_t ) 0, (uintptr_t) 0, (uintptr_t) 0); } while (0); } while (0); if (((((&(p)->p_mtx))))->mtx_lock != _tid || !atomic_cmpset_long (&(((((&(p)->p_mtx)))))->mtx_lock, (_tid), 0x00000004 )) __mtx_unlock_sleep(&(((((&(p)->p_mtx)))))->mtx_lock , (((0))), ((((void *)0))), ((0))); } while (0); | |||
| 3697 | if (fdp == NULL((void *)0)) | |||
| 3698 | return (EINVAL22); | |||
| 3699 | ||||
| 3700 | efbuf = malloc(sizeof(*efbuf), M_TEMP, M_WAITOK0x0002); | |||
| 3701 | efbuf->fdp = fdp; | |||
| 3702 | efbuf->sb = sb; | |||
| 3703 | efbuf->remainder = maxlen; | |||
| 3704 | ||||
| 3705 | FILEDESC_SLOCK(fdp)(void)__sx_slock(((&(fdp)->fd_sx)), 0, (((void *)0)), ( 0)); | |||
| 3706 | if (fdp->fd_cdir == NULL((void *)0)) | |||
| 3707 | error = EINVAL22; | |||
| 3708 | else { | |||
| 3709 | vref(fdp->fd_cdir); | |||
| 3710 | error = export_vnode_to_sb(fdp->fd_cdir, KF_FD_TYPE_CWD-1, | |||
| 3711 | FREAD0x0001, efbuf); | |||
| 3712 | } | |||
| 3713 | FILEDESC_SUNLOCK(fdp)__sx_sunlock(((&(fdp)->fd_sx)), (((void *)0)), (0)); | |||
| 3714 | fddrop(fdp); | |||
| 3715 | free(efbuf, M_TEMP); | |||
| 3716 | return (error); | |||
| 3717 | } | |||
| 3718 | ||||
| 3719 | /* | |||
| 3720 | * Get per-process current working directory. | |||
| 3721 | */ | |||
| 3722 | static int | |||
| 3723 | sysctl_kern_proc_cwd(SYSCTL_HANDLER_ARGSstruct sysctl_oid *oidp, void *arg1, intmax_t arg2, struct sysctl_req *req) | |||
| 3724 | { | |||
| 3725 | struct sbuf sb; | |||
| 3726 | struct proc *p; | |||
| 3727 | ssize_t maxlen; | |||
| 3728 | int error, error2, *name; | |||
| 3729 | ||||
| 3730 | name = (int *)arg1; | |||
| 3731 | ||||
| 3732 | sbuf_new_for_sysctl(&sb, NULL((void *)0), sizeof(struct kinfo_file), req); | |||
| 3733 | sbuf_clear_flags(&sb, SBUF_INCLUDENUL0x00000002); | |||
| 3734 | error = pget((pid_t)name[0], PGET_CANDEBUG0x00004 | PGET_NOTWEXIT0x00010, &p); | |||
| 3735 | if (error != 0) { | |||
| 3736 | sbuf_delete(&sb); | |||
| 3737 | return (error); | |||
| 3738 | } | |||
| 3739 | maxlen = req->oldptr != NULL((void *)0) ? req->oldlen : -1; | |||
| 3740 | error = kern_proc_cwd_out(p, &sb, maxlen); | |||
| 3741 | error2 = sbuf_finish(&sb); | |||
| 3742 | sbuf_delete(&sb); | |||
| 3743 | return (error != 0 ? error : error2); | |||
| 3744 | } | |||
| 3745 | ||||
| 3746 | static SYSCTL_NODE(_kern_proc, KERN_PROC_CWD, cwd, CTLFLAG_RD|CTLFLAG_MPSAFE,struct sysctl_oid sysctl___kern_proc_cwd = { .oid_parent = (( &(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (42), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("cwd"), .oid_handler = (sysctl_kern_proc_cwd), .oid_fmt = ("N"), .oid_descr = "Process current working directory" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_cwd __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_cwd); _Static_assert( ((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed") | |||
| 3747 | sysctl_kern_proc_cwd, "Process current working directory")struct sysctl_oid sysctl___kern_proc_cwd = { .oid_parent = (( &(&sysctl___kern_proc)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (42), .oid_kind = (1|(0x80000000 |0x00040000)), .oid_arg1 = (((void *)0)), .oid_arg2 = (0), .oid_name = ("cwd"), .oid_handler = (sysctl_kern_proc_cwd), .oid_fmt = ("N"), .oid_descr = "Process current working directory" }; __asm__ (".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set" ); static void const * const __set_sysctl_set_sym_sysctl___kern_proc_cwd __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_proc_cwd); _Static_assert( ((0x80000000|0x00040000) & 0xf) == 0 || ((0x80000000|0x00040000 ) & 0) == 1, "compile-time assertion failed"); | |||
| 3748 | ||||
| 3749 | #ifdef DDB | |||
| 3750 | /* | |||
| 3751 | * For the purposes of debugging, generate a human-readable string for the | |||
| 3752 | * file type. | |||
| 3753 | */ | |||
| 3754 | static const char * | |||
| 3755 | file_type_to_name(short type) | |||
| 3756 | { | |||
| 3757 | ||||
| 3758 | switch (type) { | |||
| 3759 | case 0: | |||
| 3760 | return ("zero"); | |||
| 3761 | case DTYPE_VNODE1: | |||
| 3762 | return ("vnod"); | |||
| 3763 | case DTYPE_SOCKET2: | |||
| 3764 | return ("sock"); | |||
| 3765 | case DTYPE_PIPE3: | |||
| 3766 | return ("pipe"); | |||
| 3767 | case DTYPE_FIFO4: | |||
| 3768 | return ("fifo"); | |||
| 3769 | case DTYPE_KQUEUE5: | |||
| 3770 | return ("kque"); | |||
| 3771 | case DTYPE_CRYPTO6: | |||
| 3772 | return ("crpt"); | |||
| 3773 | case DTYPE_MQUEUE7: | |||
| 3774 | return ("mque"); | |||
| 3775 | case DTYPE_SHM8: | |||
| 3776 | return ("shm"); | |||
| 3777 | case DTYPE_SEM9: | |||
| 3778 | return ("ksem"); | |||
| 3779 | default: | |||
| 3780 | return ("unkn"); | |||
| 3781 | } | |||
| 3782 | } | |||
| 3783 | ||||
| 3784 | /* | |||
| 3785 | * For the purposes of debugging, identify a process (if any, perhaps one of | |||
| 3786 | * many) that references the passed file in its file descriptor array. Return | |||
| 3787 | * NULL if none. | |||
| 3788 | */ | |||
| 3789 | static struct proc * | |||
| 3790 | file_to_first_proc(struct file *fp) | |||
| 3791 | { | |||
| 3792 | struct filedesc *fdp; | |||
| 3793 | struct proc *p; | |||
| 3794 | int n; | |||
| 3795 | ||||
| 3796 | FOREACH_PROC_IN_SYSTEM(p)for (((p)) = (((&allproc))->lh_first); ((p)); ((p)) = ( (((p)))->p_list.le_next)) { | |||
| 3797 | if (p->p_state == PRS_NEW) | |||
| 3798 | continue; | |||
| 3799 | fdp = p->p_fd; | |||
| 3800 | if (fdp == NULL((void *)0)) | |||
| 3801 | continue; | |||
| 3802 | for (n = 0; n <= fdp->fd_lastfile; n++) { | |||
| 3803 | if (fp == fdp->fd_ofilesfd_files->fdt_ofiles[n].fde_file) | |||
| 3804 | return (p); | |||
| 3805 | } | |||
| 3806 | } | |||
| 3807 | return (NULL((void *)0)); | |||
| 3808 | } | |||
| 3809 | ||||
| 3810 | static void | |||
| 3811 | db_print_file(struct file *fp, int header) | |||
| 3812 | { | |||
| 3813 | struct proc *p; | |||
| 3814 | ||||
| 3815 | if (header) | |||
| 3816 | db_printf("%8s %4s %8s %8s %4s %5s %6s %8s %5s %12s\n", | |||
| 3817 | "File", "Type", "Data", "Flag", "GCFl", "Count", | |||
| 3818 | "MCount", "Vnode", "FPID", "FCmd"); | |||
| 3819 | p = file_to_first_proc(fp); | |||
| 3820 | db_printf("%8p %4s %8p %08x %04x %5d %6d %8p %5d %12s\n", fp, | |||
| 3821 | file_type_to_name(fp->f_type), fp->f_data, fp->f_flag, | |||
| 3822 | 0, fp->f_count, 0, fp->f_vnode, | |||
| 3823 | p != NULL((void *)0) ? p->p_pid : -1, p != NULL((void *)0) ? p->p_comm : "-"); | |||
| 3824 | } | |||
| 3825 | ||||
| 3826 | DB_SHOW_COMMAND(file, db_show_file)static db_cmdfcn_t db_show_file; static struct command file_show = { .name = "file", .fcn = db_show_file, .flag = 0, .more = ( (void *)0) }; static void file_show_add(void *arg __attribute__ ((__unused__))) { db_command_register(&db_show_table, & file_show); } static struct sysinit file_show_sys_init = { SI_SUB_KLD , SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t)file_show_add , ((void *)(((void *)0))) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_file_show_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(file_show_sys_init); static void file_show_del(void *arg __attribute__((__unused__))) { db_command_unregister(& db_show_table, &file_show); } static struct sysinit file_show_sys_uninit = { SI_SUB_KLD, SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t )file_show_del, ((void *)(((void *)0))) }; __asm__(".globl " "__start_set_sysuninit_set" ); __asm__(".globl " "__stop_set_sysuninit_set"); static void const * const __set_sysuninit_set_sym_file_show_sys_uninit __attribute__ ((__section__("set_" "sysuninit_set"))) __attribute__((__used__ )) = &(file_show_sys_uninit);; static void db_show_file(db_expr_t addr, bool have_addr, db_expr_t count, char *modif) | |||
| 3827 | { | |||
| 3828 | struct file *fp; | |||
| 3829 | ||||
| 3830 | if (!have_addr) { | |||
| 3831 | db_printf("usage: show file <addr>\n"); | |||
| 3832 | return; | |||
| 3833 | } | |||
| 3834 | fp = (struct file *)addr; | |||
| 3835 | db_print_file(fp, 1); | |||
| 3836 | } | |||
| 3837 | ||||
| 3838 | DB_SHOW_COMMAND(files, db_show_files)static db_cmdfcn_t db_show_files; static struct command files_show = { .name = "files", .fcn = db_show_files, .flag = 0, .more = ((void *)0) }; static void files_show_add(void *arg __attribute__ ((__unused__))) { db_command_register(&db_show_table, & files_show); } static struct sysinit files_show_sys_init = { SI_SUB_KLD , SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t)files_show_add , ((void *)(((void *)0))) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_files_show_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(files_show_sys_init); static void files_show_del(void *arg __attribute__((__unused__))) { db_command_unregister(& db_show_table, &files_show); } static struct sysinit files_show_sys_uninit = { SI_SUB_KLD, SI_ORDER_ANY, (sysinit_cfunc_t)(sysinit_nfunc_t )files_show_del, ((void *)(((void *)0))) }; __asm__(".globl " "__start_set_sysuninit_set"); __asm__(".globl " "__stop_set_sysuninit_set" ); static void const * const __set_sysuninit_set_sym_files_show_sys_uninit __attribute__((__section__("set_" "sysuninit_set"))) __attribute__ ((__used__)) = &(files_show_sys_uninit);; static void db_show_files (db_expr_t addr, bool have_addr, db_expr_t count, char *modif ) | |||
| 3839 | { | |||
| 3840 | struct filedesc *fdp; | |||
| 3841 | struct file *fp; | |||
| 3842 | struct proc *p; | |||
| 3843 | int header; | |||
| 3844 | int n; | |||
| 3845 | ||||
| 3846 | header = 1; | |||
| 3847 | FOREACH_PROC_IN_SYSTEM(p)for (((p)) = (((&allproc))->lh_first); ((p)); ((p)) = ( (((p)))->p_list.le_next)) { | |||
| 3848 | if (p->p_state == PRS_NEW) | |||
| 3849 | continue; | |||
| 3850 | if ((fdp = p->p_fd) == NULL((void *)0)) | |||
| 3851 | continue; | |||
| 3852 | for (n = 0; n <= fdp->fd_lastfile; ++n) { | |||
| 3853 | if ((fp = fdp->fd_ofilesfd_files->fdt_ofiles[n].fde_file) == NULL((void *)0)) | |||
| 3854 | continue; | |||
| 3855 | db_print_file(fp, header); | |||
| 3856 | header = 0; | |||
| 3857 | } | |||
| 3858 | } | |||
| 3859 | } | |||
| 3860 | #endif | |||
| 3861 | ||||
| 3862 | SYSCTL_INT(_kern, KERN_MAXFILESPERPROC, maxfilesperproc, CTLFLAG_RW,static struct sysctl_oid sysctl___kern_maxfilesperproc = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (27), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = (&maxfilesperproc ), .oid_arg2 = (0), .oid_name = ("maxfilesperproc"), .oid_handler = (sysctl_handle_int), .oid_fmt = ("I"), .oid_descr = "Maximum files allowed open per process" }; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_maxfilesperproc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_maxfilesperproc); _Static_assert (((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000 |0x40000000)) & 0) == 2) && sizeof(int) == sizeof (*(&maxfilesperproc)), "compile-time assertion failed") | |||
| 3863 | &maxfilesperproc, 0, "Maximum files allowed open per process")static struct sysctl_oid sysctl___kern_maxfilesperproc = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (27), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = (&maxfilesperproc ), .oid_arg2 = (0), .oid_name = ("maxfilesperproc"), .oid_handler = (sysctl_handle_int), .oid_fmt = ("I"), .oid_descr = "Maximum files allowed open per process" }; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_maxfilesperproc __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_maxfilesperproc); _Static_assert (((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000 |0x40000000)) & 0) == 2) && sizeof(int) == sizeof (*(&maxfilesperproc)), "compile-time assertion failed"); | |||
| 3864 | ||||
| 3865 | SYSCTL_INT(_kern, KERN_MAXFILES, maxfiles, CTLFLAG_RW,static struct sysctl_oid sysctl___kern_maxfiles = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (7), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = (&maxfiles), . oid_arg2 = (0), .oid_name = ("maxfiles"), .oid_handler = (sysctl_handle_int ), .oid_fmt = ("I"), .oid_descr = "Maximum number of files" } ; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_maxfiles __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_maxfiles); _Static_assert( ((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000| 0x40000000)) & 0) == 2) && sizeof(int) == sizeof( *(&maxfiles)), "compile-time assertion failed") | |||
| 3866 | &maxfiles, 0, "Maximum number of files")static struct sysctl_oid sysctl___kern_maxfiles = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = (7), .oid_kind = (2 | 0x00040000 | ((0x80000000|0x40000000))), .oid_arg1 = (&maxfiles), . oid_arg2 = (0), .oid_name = ("maxfiles"), .oid_handler = (sysctl_handle_int ), .oid_fmt = ("I"), .oid_descr = "Maximum number of files" } ; __asm__(".globl " "__start_set_sysctl_set"); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_maxfiles __attribute__((__section__("set_" "sysctl_set"))) __attribute__ ((__used__)) = &(sysctl___kern_maxfiles); _Static_assert( ((((0x80000000|0x40000000)) & 0xf) == 0 || (((0x80000000| 0x40000000)) & 0) == 2) && sizeof(int) == sizeof( *(&maxfiles)), "compile-time assertion failed"); | |||
| 3867 | ||||
| 3868 | SYSCTL_INT(_kern, OID_AUTO, openfiles, CTLFLAG_RD,static struct sysctl_oid sysctl___kern_openfiles = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = ((-1)), .oid_kind = (2 | 0x00040000 | (0x80000000)), .oid_arg1 = (((int *)(__uintptr_t)(volatile void *)(&openfiles))), .oid_arg2 = (0), .oid_name = ("openfiles" ), .oid_handler = (sysctl_handle_int), .oid_fmt = ("I"), .oid_descr = "System-wide number of open files" }; __asm__(".globl " "__start_set_sysctl_set" ); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_openfiles __attribute__ ((__section__("set_" "sysctl_set"))) __attribute__((__used__) ) = &(sysctl___kern_openfiles); _Static_assert((((0x80000000 ) & 0xf) == 0 || ((0x80000000) & 0) == 2) && sizeof (int) == sizeof(*(((int *)(__uintptr_t)(volatile void *)(& openfiles)))), "compile-time assertion failed") | |||
| 3869 | __DEVOLATILE(int *, &openfiles), 0, "System-wide number of open files")static struct sysctl_oid sysctl___kern_openfiles = { .oid_parent = ((&(&sysctl___kern)->oid_children)), .oid_children = { ((void *)0) }, .oid_number = ((-1)), .oid_kind = (2 | 0x00040000 | (0x80000000)), .oid_arg1 = (((int *)(__uintptr_t)(volatile void *)(&openfiles))), .oid_arg2 = (0), .oid_name = ("openfiles" ), .oid_handler = (sysctl_handle_int), .oid_fmt = ("I"), .oid_descr = "System-wide number of open files" }; __asm__(".globl " "__start_set_sysctl_set" ); __asm__(".globl " "__stop_set_sysctl_set"); static void const * const __set_sysctl_set_sym_sysctl___kern_openfiles __attribute__ ((__section__("set_" "sysctl_set"))) __attribute__((__used__) ) = &(sysctl___kern_openfiles); _Static_assert((((0x80000000 ) & 0xf) == 0 || ((0x80000000) & 0) == 2) && sizeof (int) == sizeof(*(((int *)(__uintptr_t)(volatile void *)(& openfiles)))), "compile-time assertion failed"); | |||
| 3870 | ||||
| 3871 | /* ARGSUSED*/ | |||
| 3872 | static void | |||
| 3873 | filelistinit(void *dummy) | |||
| 3874 | { | |||
| 3875 | ||||
| 3876 | file_zone = uma_zcreate("Files", sizeof(struct file), NULL((void *)0), NULL((void *)0), | |||
| 3877 | NULL((void *)0), NULL((void *)0), UMA_ALIGN_PTR(sizeof(void *) - 1), UMA_ZONE_NOFREE0x0020); | |||
| 3878 | filedesc0_zone = uma_zcreate("filedesc0", sizeof(struct filedesc0), | |||
| 3879 | NULL((void *)0), NULL((void *)0), NULL((void *)0), NULL((void *)0), UMA_ALIGN_PTR(sizeof(void *) - 1), 0); | |||
| 3880 | mtx_init(&sigio_lock, "sigio lock", NULL, MTX_DEF)_mtx_init(&(&sigio_lock)->mtx_lock, "sigio lock", ( (void *)0), 0x00000000); | |||
| 3881 | } | |||
| 3882 | SYSINIT(select, SI_SUB_LOCK, SI_ORDER_FIRST, filelistinit, NULL)static struct sysinit select_sys_init = { SI_SUB_LOCK, SI_ORDER_FIRST , (sysinit_cfunc_t)(sysinit_nfunc_t)filelistinit, ((void *)(( (void *)0))) }; __asm__(".globl " "__start_set_sysinit_set"); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_select_sys_init __attribute__( (__section__("set_" "sysinit_set"))) __attribute__((__used__) ) = &(select_sys_init); | |||
| 3883 | ||||
| 3884 | /*-------------------------------------------------------------------*/ | |||
| 3885 | ||||
| 3886 | static int | |||
| 3887 | badfo_readwrite(struct file *fp, struct uio *uio, struct ucred *active_cred, | |||
| 3888 | int flags, struct thread *td) | |||
| 3889 | { | |||
| 3890 | ||||
| 3891 | return (EBADF9); | |||
| 3892 | } | |||
| 3893 | ||||
| 3894 | static int | |||
| 3895 | badfo_truncate(struct file *fp, off_t length, struct ucred *active_cred, | |||
| 3896 | struct thread *td) | |||
| 3897 | { | |||
| 3898 | ||||
| 3899 | return (EINVAL22); | |||
| 3900 | } | |||
| 3901 | ||||
| 3902 | static int | |||
| 3903 | badfo_ioctl(struct file *fp, u_long com, void *data, struct ucred *active_cred, | |||
| 3904 | struct thread *td) | |||
| 3905 | { | |||
| 3906 | ||||
| 3907 | return (EBADF9); | |||
| 3908 | } | |||
| 3909 | ||||
| 3910 | static int | |||
| 3911 | badfo_poll(struct file *fp, int events, struct ucred *active_cred, | |||
| 3912 | struct thread *td) | |||
| 3913 | { | |||
| 3914 | ||||
| 3915 | return (0); | |||
| 3916 | } | |||
| 3917 | ||||
| 3918 | static int | |||
| 3919 | badfo_kqfilter(struct file *fp, struct knote *kn) | |||
| 3920 | { | |||
| 3921 | ||||
| 3922 | return (EBADF9); | |||
| 3923 | } | |||
| 3924 | ||||
| 3925 | static int | |||
| 3926 | badfo_stat(struct file *fp, struct stat *sb, struct ucred *active_cred, | |||
| 3927 | struct thread *td) | |||
| 3928 | { | |||
| 3929 | ||||
| 3930 | return (EBADF9); | |||
| 3931 | } | |||
| 3932 | ||||
| 3933 | static int | |||
| 3934 | badfo_close(struct file *fp, struct thread *td) | |||
| 3935 | { | |||
| 3936 | ||||
| 3937 | return (0); | |||
| 3938 | } | |||
| 3939 | ||||
| 3940 | static int | |||
| 3941 | badfo_chmod(struct file *fp, mode_t mode, struct ucred *active_cred, | |||
| 3942 | struct thread *td) | |||
| 3943 | { | |||
| 3944 | ||||
| 3945 | return (EBADF9); | |||
| 3946 | } | |||
| 3947 | ||||
| 3948 | static int | |||
| 3949 | badfo_chown(struct file *fp, uid_t uid, gid_t gid, struct ucred *active_cred, | |||
| 3950 | struct thread *td) | |||
| 3951 | { | |||
| 3952 | ||||
| 3953 | return (EBADF9); | |||
| 3954 | } | |||
| 3955 | ||||
| 3956 | static int | |||
| 3957 | badfo_sendfile(struct file *fp, int sockfd, struct uio *hdr_uio, | |||
| 3958 | struct uio *trl_uio, off_t offset, size_t nbytes, off_t *sent, int flags, | |||
| 3959 | struct thread *td) | |||
| 3960 | { | |||
| 3961 | ||||
| 3962 | return (EBADF9); | |||
| 3963 | } | |||
| 3964 | ||||
| 3965 | static int | |||
| 3966 | badfo_fill_kinfo(struct file *fp, struct kinfo_file *kif, struct filedesc *fdp) | |||
| 3967 | { | |||
| 3968 | ||||
| 3969 | return (0); | |||
| 3970 | } | |||
| 3971 | ||||
| 3972 | struct fileops badfileops = { | |||
| 3973 | .fo_read = badfo_readwrite, | |||
| 3974 | .fo_write = badfo_readwrite, | |||
| 3975 | .fo_truncate = badfo_truncate, | |||
| 3976 | .fo_ioctl = badfo_ioctl, | |||
| 3977 | .fo_poll = badfo_poll, | |||
| 3978 | .fo_kqfilter = badfo_kqfilter, | |||
| 3979 | .fo_stat = badfo_stat, | |||
| 3980 | .fo_close = badfo_close, | |||
| 3981 | .fo_chmod = badfo_chmod, | |||
| 3982 | .fo_chown = badfo_chown, | |||
| 3983 | .fo_sendfile = badfo_sendfile, | |||
| 3984 | .fo_fill_kinfo = badfo_fill_kinfo, | |||
| 3985 | }; | |||
| 3986 | ||||
| 3987 | int | |||
| 3988 | invfo_rdwr(struct file *fp, struct uio *uio, struct ucred *active_cred, | |||
| 3989 | int flags, struct thread *td) | |||
| 3990 | { | |||
| 3991 | ||||
| 3992 | return (EOPNOTSUPP45); | |||
| 3993 | } | |||
| 3994 | ||||
| 3995 | int | |||
| 3996 | invfo_truncate(struct file *fp, off_t length, struct ucred *active_cred, | |||
| 3997 | struct thread *td) | |||
| 3998 | { | |||
| 3999 | ||||
| 4000 | return (EINVAL22); | |||
| 4001 | } | |||
| 4002 | ||||
| 4003 | int | |||
| 4004 | invfo_ioctl(struct file *fp, u_long com, void *data, | |||
| 4005 | struct ucred *active_cred, struct thread *td) | |||
| 4006 | { | |||
| 4007 | ||||
| 4008 | return (ENOTTY25); | |||
| 4009 | } | |||
| 4010 | ||||
| 4011 | int | |||
| 4012 | invfo_poll(struct file *fp, int events, struct ucred *active_cred, | |||
| 4013 | struct thread *td) | |||
| 4014 | { | |||
| 4015 | ||||
| 4016 | return (poll_no_poll(events)); | |||
| 4017 | } | |||
| 4018 | ||||
| 4019 | int | |||
| 4020 | invfo_kqfilter(struct file *fp, struct knote *kn) | |||
| 4021 | { | |||
| 4022 | ||||
| 4023 | return (EINVAL22); | |||
| 4024 | } | |||
| 4025 | ||||
| 4026 | int | |||
| 4027 | invfo_chmod(struct file *fp, mode_t mode, struct ucred *active_cred, | |||
| 4028 | struct thread *td) | |||
| 4029 | { | |||
| 4030 | ||||
| 4031 | return (EINVAL22); | |||
| 4032 | } | |||
| 4033 | ||||
| 4034 | int | |||
| 4035 | invfo_chown(struct file *fp, uid_t uid, gid_t gid, struct ucred *active_cred, | |||
| 4036 | struct thread *td) | |||
| 4037 | { | |||
| 4038 | ||||
| 4039 | return (EINVAL22); | |||
| 4040 | } | |||
| 4041 | ||||
| 4042 | int | |||
| 4043 | invfo_sendfile(struct file *fp, int sockfd, struct uio *hdr_uio, | |||
| 4044 | struct uio *trl_uio, off_t offset, size_t nbytes, off_t *sent, int flags, | |||
| 4045 | struct thread *td) | |||
| 4046 | { | |||
| 4047 | ||||
| 4048 | return (EINVAL22); | |||
| 4049 | } | |||
| 4050 | ||||
| 4051 | /*-------------------------------------------------------------------*/ | |||
| 4052 | ||||
| 4053 | /* | |||
| 4054 | * File Descriptor pseudo-device driver (/dev/fd/). | |||
| 4055 | * | |||
| 4056 | * Opening minor device N dup()s the file (if any) connected to file | |||
| 4057 | * descriptor N belonging to the calling process. Note that this driver | |||
| 4058 | * consists of only the ``open()'' routine, because all subsequent | |||
| 4059 | * references to this file will be direct to the other driver. | |||
| 4060 | * | |||
| 4061 | * XXX: we could give this one a cloning event handler if necessary. | |||
| 4062 | */ | |||
| 4063 | ||||
| 4064 | /* ARGSUSED */ | |||
| 4065 | static int | |||
| 4066 | fdopen(struct cdev *dev, int mode, int type, struct thread *td) | |||
| 4067 | { | |||
| 4068 | ||||
| 4069 | /* | |||
| 4070 | * XXX Kludge: set curthread->td_dupfd to contain the value of the | |||
| 4071 | * the file descriptor being sought for duplication. The error | |||
| 4072 | * return ensures that the vnode for this device will be released | |||
| 4073 | * by vn_open. Open will detect this special error and take the | |||
| 4074 | * actions in dupfdopen below. Other callers of vn_open or VOP_OPEN | |||
| 4075 | * will simply report the error. | |||
| 4076 | */ | |||
| 4077 | td->td_dupfd = dev2unit(dev)((dev)->si_drv0); | |||
| 4078 | return (ENODEV19); | |||
| 4079 | } | |||
| 4080 | ||||
| 4081 | static struct cdevsw fildesc_cdevsw = { | |||
| 4082 | .d_version = D_VERSION0x17122009, | |||
| 4083 | .d_open = fdopen, | |||
| 4084 | .d_name = "FD", | |||
| 4085 | }; | |||
| 4086 | ||||
| 4087 | static void | |||
| 4088 | fildesc_drvinit(void *unused) | |||
| 4089 | { | |||
| 4090 | struct cdev *dev; | |||
| 4091 | ||||
| 4092 | dev = make_dev_credf(MAKEDEV_ETERNAL0x10, &fildesc_cdevsw, 0, NULL((void *)0), | |||
| 4093 | UID_ROOT0, GID_WHEEL0, 0666, "fd/0"); | |||
| 4094 | make_dev_alias(dev, "stdin"); | |||
| 4095 | dev = make_dev_credf(MAKEDEV_ETERNAL0x10, &fildesc_cdevsw, 1, NULL((void *)0), | |||
| 4096 | UID_ROOT0, GID_WHEEL0, 0666, "fd/1"); | |||
| 4097 | make_dev_alias(dev, "stdout"); | |||
| 4098 | dev = make_dev_credf(MAKEDEV_ETERNAL0x10, &fildesc_cdevsw, 2, NULL((void *)0), | |||
| 4099 | UID_ROOT0, GID_WHEEL0, 0666, "fd/2"); | |||
| 4100 | make_dev_alias(dev, "stderr"); | |||
| 4101 | } | |||
| 4102 | ||||
| 4103 | SYSINIT(fildescdev, SI_SUB_DRIVERS, SI_ORDER_MIDDLE, fildesc_drvinit, NULL)static struct sysinit fildescdev_sys_init = { SI_SUB_DRIVERS, SI_ORDER_MIDDLE, (sysinit_cfunc_t)(sysinit_nfunc_t)fildesc_drvinit , ((void *)(((void *)0))) }; __asm__(".globl " "__start_set_sysinit_set" ); __asm__(".globl " "__stop_set_sysinit_set"); static void const * const __set_sysinit_set_sym_fildescdev_sys_init __attribute__ ((__section__("set_" "sysinit_set"))) __attribute__((__used__ )) = &(fildescdev_sys_init); |