4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
23 * Copyright 2009 Sun Microsystems, Inc. All rights reserved.
24 * Use is subject to license terms.
28 * Copyright (c) 2015, Joyent, Inc. All rights reserved.
31 #include <sys/atomic.h>
32 #include <sys/errno.h>
34 #include <sys/modctl.h>
36 #include <sys/systm.h>
38 #include <sys/sunddi.h>
39 #include <sys/cpuvar.h>
41 #include <sys/strsubr.h>
42 #include <sys/fasttrap.h>
43 #include <sys/fasttrap_impl.h>
44 #include <sys/fasttrap_isa.h>
45 #include <sys/dtrace.h>
46 #include <sys/dtrace_impl.h>
47 #include <sys/sysmacros.h>
50 #include <sys/policy.h>
51 #include <util/qsort.h>
54 * User-Land Trap-Based Tracing
55 * ----------------------------
57 * The fasttrap provider allows DTrace consumers to instrument any user-level
58 * instruction to gather data; this includes probes with semantic
59 * signifigance like entry and return as well as simple offsets into the
60 * function. While the specific techniques used are very ISA specific, the
61 * methodology is generalizable to any architecture.
64 * The General Methodology
65 * -----------------------
67 * With the primary goal of tracing every user-land instruction and the
68 * limitation that we can't trust user space so don't want to rely on much
69 * information there, we begin by replacing the instructions we want to trace
70 * with trap instructions. Each instruction we overwrite is saved into a hash
71 * table keyed by process ID and pc address. When we enter the kernel due to
72 * this trap instruction, we need the effects of the replaced instruction to
73 * appear to have occurred before we proceed with the user thread's
76 * Each user level thread is represented by a ulwp_t structure which is
77 * always easily accessible through a register. The most basic way to produce
78 * the effects of the instruction we replaced is to copy that instruction out
79 * to a bit of scratch space reserved in the user thread's ulwp_t structure
80 * (a sort of kernel-private thread local storage), set the PC to that
81 * scratch space and single step. When we reenter the kernel after single
82 * stepping the instruction we must then adjust the PC to point to what would
83 * normally be the next instruction. Of course, special care must be taken
84 * for branches and jumps, but these represent such a small fraction of any
85 * instruction set that writing the code to emulate these in the kernel is
88 * Return probes may require several tracepoints to trace every return site,
89 * and, conversely, each tracepoint may activate several probes (the entry
90 * and offset 0 probes, for example). To solve this muliplexing problem,
91 * tracepoints contain lists of probes to activate and probes contain lists
92 * of tracepoints to enable. If a probe is activated, it adds its ID to
93 * existing tracepoints or creates new ones as necessary.
95 * Most probes are activated _before_ the instruction is executed, but return
96 * probes are activated _after_ the effects of the last instruction of the
97 * function are visible. Return probes must be fired _after_ we have
98 * single-stepped the instruction whereas all other probes are fired
105 * The lock ordering below -- both internally and with respect to the DTrace
106 * framework -- is a little tricky and bears some explanation. Each provider
107 * has a lock (ftp_mtx) that protects its members including reference counts
108 * for enabled probes (ftp_rcount), consumers actively creating probes
109 * (ftp_ccount) and USDT consumers (ftp_mcount); all three prevent a provider
110 * from being freed. A provider is looked up by taking the bucket lock for the
111 * provider hash table, and is returned with its lock held. The provider lock
112 * may be taken in functions invoked by the DTrace framework, but may not be
113 * held while calling functions in the DTrace framework.
115 * To ensure consistency over multiple calls to the DTrace framework, the
116 * creation lock (ftp_cmtx) should be held. Naturally, the creation lock may
117 * not be taken when holding the provider lock as that would create a cyclic
118 * lock ordering. In situations where one would naturally take the provider
119 * lock and then the creation lock, we instead up a reference count to prevent
120 * the provider from disappearing, drop the provider lock, and acquire the
124 * bucket lock before provider lock
125 * DTrace before provider lock
126 * creation lock before DTrace
127 * never hold the provider lock and creation lock simultaneously
130 static dev_info_t
*fasttrap_devi
;
131 static dtrace_meta_provider_id_t fasttrap_meta_id
;
133 static timeout_id_t fasttrap_timeout
;
134 static kmutex_t fasttrap_cleanup_mtx
;
135 static uint_t fasttrap_cleanup_work
;
138 * Generation count on modifications to the global tracepoint lookup table.
140 static volatile uint64_t fasttrap_mod_gen
;
143 * When the fasttrap provider is loaded, fasttrap_max is set to either
144 * FASTTRAP_MAX_DEFAULT or the value for fasttrap-max-probes in the
145 * fasttrap.conf file. Each time a probe is created, fasttrap_total is
146 * incremented by the number of tracepoints that may be associated with that
147 * probe; fasttrap_total is capped at fasttrap_max.
149 #define FASTTRAP_MAX_DEFAULT 250000
150 static uint32_t fasttrap_max
;
151 static uint32_t fasttrap_total
;
154 #define FASTTRAP_TPOINTS_DEFAULT_SIZE 0x4000
155 #define FASTTRAP_PROVIDERS_DEFAULT_SIZE 0x100
156 #define FASTTRAP_PROCS_DEFAULT_SIZE 0x100
158 #define FASTTRAP_PID_NAME "pid"
160 fasttrap_hash_t fasttrap_tpoints
;
161 static fasttrap_hash_t fasttrap_provs
;
162 static fasttrap_hash_t fasttrap_procs
;
164 static uint64_t fasttrap_pid_count
; /* pid ref count */
165 static kmutex_t fasttrap_count_mtx
; /* lock on ref count */
167 #define FASTTRAP_ENABLE_FAIL 1
168 #define FASTTRAP_ENABLE_PARTIAL 2
170 static int fasttrap_tracepoint_enable(proc_t
*, fasttrap_probe_t
*, uint_t
);
171 static void fasttrap_tracepoint_disable(proc_t
*, fasttrap_probe_t
*, uint_t
);
173 static fasttrap_provider_t
*fasttrap_provider_lookup(pid_t
, const char *,
174 const dtrace_pattr_t
*);
175 static void fasttrap_provider_retire(pid_t
, const char *, int);
176 static void fasttrap_provider_free(fasttrap_provider_t
*);
178 static fasttrap_proc_t
*fasttrap_proc_lookup(pid_t
);
179 static void fasttrap_proc_release(fasttrap_proc_t
*);
181 #define FASTTRAP_PROVS_INDEX(pid, name) \
182 ((fasttrap_hash_str(name) + (pid)) & fasttrap_provs.fth_mask)
184 #define FASTTRAP_PROCS_INDEX(pid) ((pid) & fasttrap_procs.fth_mask)
187 fasttrap_highbit(ulong_t i
)
194 if (i
& 0xffffffff00000000ul
) {
198 if (i
& 0xffff0000) {
217 fasttrap_hash_str(const char *p
)
223 hval
= (hval
<< 4) + *p
++;
224 if ((g
= (hval
& 0xf0000000)) != 0)
232 fasttrap_sigtrap(proc_t
*p
, kthread_t
*t
, uintptr_t pc
)
234 sigqueue_t
*sqp
= kmem_zalloc(sizeof (sigqueue_t
), KM_SLEEP
);
236 sqp
->sq_info
.si_signo
= SIGTRAP
;
237 sqp
->sq_info
.si_code
= TRAP_DTRACE
;
238 sqp
->sq_info
.si_addr
= (caddr_t
)pc
;
240 mutex_enter(&p
->p_lock
);
242 mutex_exit(&p
->p_lock
);
249 * This function ensures that no threads are actively using the memory
250 * associated with probes that were formerly live.
253 fasttrap_mod_barrier(uint64_t gen
)
257 if (gen
< fasttrap_mod_gen
)
262 for (i
= 0; i
< NCPU
; i
++) {
263 mutex_enter(&cpu_core
[i
].cpuc_pid_lock
);
264 mutex_exit(&cpu_core
[i
].cpuc_pid_lock
);
269 * This is the timeout's callback for cleaning up the providers and their
274 fasttrap_pid_cleanup_cb(void *data
)
276 fasttrap_provider_t
**fpp
, *fp
;
277 fasttrap_bucket_t
*bucket
;
278 dtrace_provider_id_t provid
;
281 static volatile int in
= 0;
285 mutex_enter(&fasttrap_cleanup_mtx
);
286 while (fasttrap_cleanup_work
) {
287 fasttrap_cleanup_work
= 0;
288 mutex_exit(&fasttrap_cleanup_mtx
);
293 * Iterate over all the providers trying to remove the marked
294 * ones. If a provider is marked but not retired, we just
295 * have to take a crack at removing it -- it's no big deal if
298 for (i
= 0; i
< fasttrap_provs
.fth_nent
; i
++) {
299 bucket
= &fasttrap_provs
.fth_table
[i
];
300 mutex_enter(&bucket
->ftb_mtx
);
301 fpp
= (fasttrap_provider_t
**)&bucket
->ftb_data
;
303 while ((fp
= *fpp
) != NULL
) {
304 if (!fp
->ftp_marked
) {
309 mutex_enter(&fp
->ftp_mtx
);
312 * If this provider has consumers actively
313 * creating probes (ftp_ccount) or is a USDT
314 * provider (ftp_mcount), we can't unregister
317 if (fp
->ftp_ccount
!= 0 ||
318 fp
->ftp_mcount
!= 0) {
319 mutex_exit(&fp
->ftp_mtx
);
324 if (!fp
->ftp_retired
|| fp
->ftp_rcount
!= 0)
327 mutex_exit(&fp
->ftp_mtx
);
330 * If we successfully unregister this
331 * provider we can remove it from the hash
332 * chain and free the memory. If our attempt
333 * to unregister fails and this is a retired
334 * provider, increment our flag to try again
335 * pretty soon. If we've consumed more than
336 * half of our total permitted number of
337 * probes call dtrace_condense() to try to
338 * clean out the unenabled probes.
340 provid
= fp
->ftp_provid
;
341 if ((rval
= dtrace_unregister(provid
)) != 0) {
342 if (fasttrap_total
> fasttrap_max
/ 2)
343 (void) dtrace_condense(provid
);
348 later
+= fp
->ftp_marked
;
352 fasttrap_provider_free(fp
);
355 mutex_exit(&bucket
->ftb_mtx
);
358 mutex_enter(&fasttrap_cleanup_mtx
);
361 ASSERT(fasttrap_timeout
!= 0);
364 * If we were unable to remove a retired provider, try again after
365 * a second. This situation can occur in certain circumstances where
366 * providers cannot be unregistered even though they have no probes
367 * enabled because of an execution of dtrace -l or something similar.
368 * If the timeout has been disabled (set to 1 because we're trying
369 * to detach), we set fasttrap_cleanup_work to ensure that we'll
370 * get a chance to do that work if and when the timeout is reenabled
374 if (fasttrap_timeout
!= (timeout_id_t
)1) {
376 timeout(&fasttrap_pid_cleanup_cb
, NULL
, hz
);
379 fasttrap_cleanup_work
= 1;
381 fasttrap_timeout
= 0;
384 mutex_exit(&fasttrap_cleanup_mtx
);
389 * Activates the asynchronous cleanup mechanism.
392 fasttrap_pid_cleanup(void)
394 mutex_enter(&fasttrap_cleanup_mtx
);
395 fasttrap_cleanup_work
= 1;
396 if (fasttrap_timeout
== 0)
397 fasttrap_timeout
= timeout(&fasttrap_pid_cleanup_cb
, NULL
, 1);
398 mutex_exit(&fasttrap_cleanup_mtx
);
402 * This is called from cfork() via dtrace_fasttrap_fork(). The child
403 * process's address space is (roughly) a copy of the parent process's so
404 * we have to remove all the instrumentation we had previously enabled in the
408 fasttrap_fork(proc_t
*p
, proc_t
*cp
)
410 pid_t ppid
= p
->p_pid
;
413 ASSERT(curproc
== p
);
414 ASSERT(p
->p_proc_flag
& P_PR_LOCK
);
415 ASSERT(p
->p_dtrace_count
> 0);
416 ASSERT(cp
->p_dtrace_count
== 0);
419 * This would be simpler and faster if we maintained per-process
420 * hash tables of enabled tracepoints. It could, however, potentially
421 * slow down execution of a tracepoint since we'd need to go
422 * through two levels of indirection. In the future, we should
423 * consider either maintaining per-process ancillary lists of
424 * enabled tracepoints or hanging a pointer to a per-process hash
425 * table of enabled tracepoints off the proc structure.
429 * We don't have to worry about the child process disappearing
430 * because we're in fork().
432 mutex_enter(&cp
->p_lock
);
434 mutex_exit(&cp
->p_lock
);
437 * Iterate over every tracepoint looking for ones that belong to the
438 * parent process, and remove each from the child process.
440 for (i
= 0; i
< fasttrap_tpoints
.fth_nent
; i
++) {
441 fasttrap_tracepoint_t
*tp
;
442 fasttrap_bucket_t
*bucket
= &fasttrap_tpoints
.fth_table
[i
];
444 mutex_enter(&bucket
->ftb_mtx
);
445 for (tp
= bucket
->ftb_data
; tp
!= NULL
; tp
= tp
->ftt_next
) {
446 if (tp
->ftt_pid
== ppid
&&
447 tp
->ftt_proc
->ftpc_acount
!= 0) {
448 int ret
= fasttrap_tracepoint_remove(cp
, tp
);
452 * The count of active providers can only be
453 * decremented (i.e. to zero) during exec,
454 * exit, and removal of a meta provider so it
455 * should be impossible to drop the count
458 ASSERT(tp
->ftt_proc
->ftpc_acount
!= 0);
461 mutex_exit(&bucket
->ftb_mtx
);
464 mutex_enter(&cp
->p_lock
);
469 * This is called from proc_exit() or from exec_common() if p_dtrace_probes
470 * is set on the proc structure to indicate that there is a pid provider
471 * associated with this process.
474 fasttrap_exec_exit(proc_t
*p
)
476 ASSERT(p
== curproc
);
477 ASSERT(MUTEX_HELD(&p
->p_lock
));
479 mutex_exit(&p
->p_lock
);
482 * We clean up the pid provider for this process here; user-land
483 * static probes are handled by the meta-provider remove entry point.
485 fasttrap_provider_retire(p
->p_pid
, FASTTRAP_PID_NAME
, 0);
487 mutex_enter(&p
->p_lock
);
493 fasttrap_pid_provide(void *arg
, const dtrace_probedesc_t
*desc
)
496 * There are no "default" pid probes.
501 fasttrap_tracepoint_enable(proc_t
*p
, fasttrap_probe_t
*probe
, uint_t index
)
503 fasttrap_tracepoint_t
*tp
, *new_tp
= NULL
;
504 fasttrap_bucket_t
*bucket
;
509 ASSERT(index
< probe
->ftp_ntps
);
511 pid
= probe
->ftp_pid
;
512 pc
= probe
->ftp_tps
[index
].fit_tp
->ftt_pc
;
513 id
= &probe
->ftp_tps
[index
].fit_id
;
515 ASSERT(probe
->ftp_tps
[index
].fit_tp
->ftt_pid
== pid
);
517 ASSERT(!(p
->p_flag
& SVFORK
));
520 * Before we make any modifications, make sure we've imposed a barrier
521 * on the generation in which this probe was last modified.
523 fasttrap_mod_barrier(probe
->ftp_gen
);
525 bucket
= &fasttrap_tpoints
.fth_table
[FASTTRAP_TPOINTS_INDEX(pid
, pc
)];
528 * If the tracepoint has already been enabled, just add our id to the
529 * list of interested probes. This may be our second time through
530 * this path in which case we'll have constructed the tracepoint we'd
531 * like to install. If we can't find a match, and have an allocated
532 * tracepoint ready to go, enable that one now.
534 * A tracepoint whose process is defunct is also considered defunct.
537 mutex_enter(&bucket
->ftb_mtx
);
538 for (tp
= bucket
->ftb_data
; tp
!= NULL
; tp
= tp
->ftt_next
) {
540 * Note that it's safe to access the active count on the
541 * associated proc structure because we know that at least one
542 * provider (this one) will still be around throughout this
545 if (tp
->ftt_pid
!= pid
|| tp
->ftt_pc
!= pc
||
546 tp
->ftt_proc
->ftpc_acount
== 0)
550 * Now that we've found a matching tracepoint, it would be
551 * a decent idea to confirm that the tracepoint is still
552 * enabled and the trap instruction hasn't been overwritten.
553 * Since this is a little hairy, we'll punt for now.
557 * This can't be the first interested probe. We don't have
558 * to worry about another thread being in the midst of
559 * deleting this tracepoint (which would be the only valid
560 * reason for a tracepoint to have no interested probes)
561 * since we're holding P_PR_LOCK for this process.
563 ASSERT(tp
->ftt_ids
!= NULL
|| tp
->ftt_retids
!= NULL
);
565 switch (id
->fti_ptype
) {
568 case DTFTP_IS_ENABLED
:
569 id
->fti_next
= tp
->ftt_ids
;
576 case DTFTP_POST_OFFSETS
:
577 id
->fti_next
= tp
->ftt_retids
;
587 mutex_exit(&bucket
->ftb_mtx
);
589 if (new_tp
!= NULL
) {
590 new_tp
->ftt_ids
= NULL
;
591 new_tp
->ftt_retids
= NULL
;
598 * If we have a good tracepoint ready to go, install it now while
599 * we have the lock held and no one can screw with us.
601 if (new_tp
!= NULL
) {
604 new_tp
->ftt_next
= bucket
->ftb_data
;
606 bucket
->ftb_data
= new_tp
;
608 mutex_exit(&bucket
->ftb_mtx
);
611 * Activate the tracepoint in the ISA-specific manner.
612 * If this fails, we need to report the failure, but
613 * indicate that this tracepoint must still be disabled
614 * by calling fasttrap_tracepoint_disable().
616 if (fasttrap_tracepoint_install(p
, new_tp
) != 0)
617 rc
= FASTTRAP_ENABLE_PARTIAL
;
620 * Increment the count of the number of tracepoints active in
621 * the victim process.
623 ASSERT(p
->p_proc_flag
& P_PR_LOCK
);
629 mutex_exit(&bucket
->ftb_mtx
);
632 * Initialize the tracepoint that's been preallocated with the probe.
634 new_tp
= probe
->ftp_tps
[index
].fit_tp
;
636 ASSERT(new_tp
->ftt_pid
== pid
);
637 ASSERT(new_tp
->ftt_pc
== pc
);
638 ASSERT(new_tp
->ftt_proc
== probe
->ftp_prov
->ftp_proc
);
639 ASSERT(new_tp
->ftt_ids
== NULL
);
640 ASSERT(new_tp
->ftt_retids
== NULL
);
642 switch (id
->fti_ptype
) {
645 case DTFTP_IS_ENABLED
:
647 new_tp
->ftt_ids
= id
;
651 case DTFTP_POST_OFFSETS
:
653 new_tp
->ftt_retids
= id
;
661 * If the ISA-dependent initialization goes to plan, go back to the
662 * beginning and try to install this freshly made tracepoint.
664 if (fasttrap_tracepoint_init(p
, new_tp
, pc
, id
->fti_ptype
) == 0)
667 new_tp
->ftt_ids
= NULL
;
668 new_tp
->ftt_retids
= NULL
;
670 return (FASTTRAP_ENABLE_FAIL
);
674 fasttrap_tracepoint_disable(proc_t
*p
, fasttrap_probe_t
*probe
, uint_t index
)
676 fasttrap_bucket_t
*bucket
;
677 fasttrap_provider_t
*provider
= probe
->ftp_prov
;
678 fasttrap_tracepoint_t
**pp
, *tp
;
679 fasttrap_id_t
*id
, **idp
;
683 ASSERT(index
< probe
->ftp_ntps
);
685 pid
= probe
->ftp_pid
;
686 pc
= probe
->ftp_tps
[index
].fit_tp
->ftt_pc
;
687 id
= &probe
->ftp_tps
[index
].fit_id
;
689 ASSERT(probe
->ftp_tps
[index
].fit_tp
->ftt_pid
== pid
);
692 * Find the tracepoint and make sure that our id is one of the
693 * ones registered with it.
695 bucket
= &fasttrap_tpoints
.fth_table
[FASTTRAP_TPOINTS_INDEX(pid
, pc
)];
696 mutex_enter(&bucket
->ftb_mtx
);
697 for (tp
= bucket
->ftb_data
; tp
!= NULL
; tp
= tp
->ftt_next
) {
698 if (tp
->ftt_pid
== pid
&& tp
->ftt_pc
== pc
&&
699 tp
->ftt_proc
== provider
->ftp_proc
)
704 * If we somehow lost this tracepoint, we're in a world of hurt.
708 switch (id
->fti_ptype
) {
711 case DTFTP_IS_ENABLED
:
712 ASSERT(tp
->ftt_ids
!= NULL
);
717 case DTFTP_POST_OFFSETS
:
718 ASSERT(tp
->ftt_retids
!= NULL
);
719 idp
= &tp
->ftt_retids
;
726 while ((*idp
)->fti_probe
!= probe
) {
727 idp
= &(*idp
)->fti_next
;
728 ASSERT(*idp
!= NULL
);
735 ASSERT(id
->fti_probe
== probe
);
738 * If there are other registered enablings of this tracepoint, we're
739 * all done, but if this was the last probe assocated with this
740 * this tracepoint, we need to remove and free it.
742 if (tp
->ftt_ids
!= NULL
|| tp
->ftt_retids
!= NULL
) {
745 * If the current probe's tracepoint is in use, swap it
746 * for an unused tracepoint.
748 if (tp
== probe
->ftp_tps
[index
].fit_tp
) {
749 fasttrap_probe_t
*tmp_probe
;
750 fasttrap_tracepoint_t
**tmp_tp
;
753 if (tp
->ftt_ids
!= NULL
) {
754 tmp_probe
= tp
->ftt_ids
->fti_probe
;
755 /* LINTED - alignment */
756 tmp_index
= FASTTRAP_ID_INDEX(tp
->ftt_ids
);
757 tmp_tp
= &tmp_probe
->ftp_tps
[tmp_index
].fit_tp
;
759 tmp_probe
= tp
->ftt_retids
->fti_probe
;
760 /* LINTED - alignment */
761 tmp_index
= FASTTRAP_ID_INDEX(tp
->ftt_retids
);
762 tmp_tp
= &tmp_probe
->ftp_tps
[tmp_index
].fit_tp
;
765 ASSERT(*tmp_tp
!= NULL
);
766 ASSERT(*tmp_tp
!= probe
->ftp_tps
[index
].fit_tp
);
767 ASSERT((*tmp_tp
)->ftt_ids
== NULL
);
768 ASSERT((*tmp_tp
)->ftt_retids
== NULL
);
770 probe
->ftp_tps
[index
].fit_tp
= *tmp_tp
;
774 mutex_exit(&bucket
->ftb_mtx
);
777 * Tag the modified probe with the generation in which it was
780 probe
->ftp_gen
= fasttrap_mod_gen
;
784 mutex_exit(&bucket
->ftb_mtx
);
787 * We can't safely remove the tracepoint from the set of active
788 * tracepoints until we've actually removed the fasttrap instruction
789 * from the process's text. We can, however, operate on this
790 * tracepoint secure in the knowledge that no other thread is going to
791 * be looking at it since we hold P_PR_LOCK on the process if it's
792 * live or we hold the provider lock on the process if it's dead and
797 * We only need to remove the actual instruction if we're looking
798 * at an existing process
802 * If we fail to restore the instruction we need to kill
803 * this process since it's in a completely unrecoverable
806 if (fasttrap_tracepoint_remove(p
, tp
) != 0)
807 fasttrap_sigtrap(p
, NULL
, pc
);
810 * Decrement the count of the number of tracepoints active
811 * in the victim process.
813 ASSERT(p
->p_proc_flag
& P_PR_LOCK
);
818 * Remove the probe from the hash table of active tracepoints.
820 mutex_enter(&bucket
->ftb_mtx
);
821 pp
= (fasttrap_tracepoint_t
**)&bucket
->ftb_data
;
824 pp
= &(*pp
)->ftt_next
;
831 mutex_exit(&bucket
->ftb_mtx
);
834 * Tag the modified probe with the generation in which it was changed.
836 probe
->ftp_gen
= fasttrap_mod_gen
;
840 fasttrap_enable_callbacks(void)
843 * We don't have to play the rw lock game here because we're
844 * providing something rather than taking something away --
845 * we can be sure that no threads have tried to follow this
846 * function pointer yet.
848 mutex_enter(&fasttrap_count_mtx
);
849 if (fasttrap_pid_count
== 0) {
850 ASSERT(dtrace_pid_probe_ptr
== NULL
);
851 ASSERT(dtrace_return_probe_ptr
== NULL
);
852 dtrace_pid_probe_ptr
= &fasttrap_pid_probe
;
853 dtrace_return_probe_ptr
= &fasttrap_return_probe
;
855 ASSERT(dtrace_pid_probe_ptr
== &fasttrap_pid_probe
);
856 ASSERT(dtrace_return_probe_ptr
== &fasttrap_return_probe
);
857 fasttrap_pid_count
++;
858 mutex_exit(&fasttrap_count_mtx
);
862 fasttrap_disable_callbacks(void)
864 ASSERT(MUTEX_HELD(&cpu_lock
));
866 mutex_enter(&fasttrap_count_mtx
);
867 ASSERT(fasttrap_pid_count
> 0);
868 fasttrap_pid_count
--;
869 if (fasttrap_pid_count
== 0) {
870 cpu_t
*cur
, *cpu
= CPU
;
872 for (cur
= cpu
->cpu_next_onln
; cur
!= cpu
;
873 cur
= cur
->cpu_next_onln
) {
874 rw_enter(&cur
->cpu_ft_lock
, RW_WRITER
);
877 dtrace_pid_probe_ptr
= NULL
;
878 dtrace_return_probe_ptr
= NULL
;
880 for (cur
= cpu
->cpu_next_onln
; cur
!= cpu
;
881 cur
= cur
->cpu_next_onln
) {
882 rw_exit(&cur
->cpu_ft_lock
);
885 mutex_exit(&fasttrap_count_mtx
);
890 fasttrap_pid_enable(void *arg
, dtrace_id_t id
, void *parg
)
892 fasttrap_probe_t
*probe
= parg
;
896 ASSERT(probe
!= NULL
);
897 ASSERT(!probe
->ftp_enabled
);
898 ASSERT(id
== probe
->ftp_id
);
899 ASSERT(MUTEX_HELD(&cpu_lock
));
902 * Increment the count of enabled probes on this probe's provider;
903 * the provider can't go away while the probe still exists. We
904 * must increment this even if we aren't able to properly enable
907 mutex_enter(&probe
->ftp_prov
->ftp_mtx
);
908 probe
->ftp_prov
->ftp_rcount
++;
909 mutex_exit(&probe
->ftp_prov
->ftp_mtx
);
912 * If this probe's provider is retired (meaning it was valid in a
913 * previously exec'ed incarnation of this address space), bail out. The
914 * provider can't go away while we're in this code path.
916 if (probe
->ftp_prov
->ftp_retired
)
920 * If we can't find the process, it may be that we're in the context of
921 * a fork in which the traced process is being born and we're copying
922 * USDT probes. Otherwise, the process is gone so bail.
924 if ((p
= sprlock(probe
->ftp_pid
)) == NULL
) {
925 if ((curproc
->p_flag
& SFORKING
) == 0)
928 mutex_enter(&pidlock
);
929 p
= prfind(probe
->ftp_pid
);
933 * So it's not that the target process is being born,
934 * it's that it isn't there at all (and we simply
935 * happen to be forking). Anyway, we know that the
936 * target is definitely gone, so bail out.
938 mutex_exit(&pidlock
);
943 * Confirm that curproc is indeed forking the process in which
944 * we're trying to enable probes.
946 ASSERT(p
->p_parent
== curproc
);
947 ASSERT(p
->p_stat
== SIDL
);
949 mutex_enter(&p
->p_lock
);
950 mutex_exit(&pidlock
);
955 ASSERT(!(p
->p_flag
& SVFORK
));
956 mutex_exit(&p
->p_lock
);
959 * We have to enable the trap entry point before any user threads have
960 * the chance to execute the trap instruction we're about to place
961 * in their process's text.
963 fasttrap_enable_callbacks();
966 * Enable all the tracepoints and add this probe's id to each
967 * tracepoint's list of active probes.
969 for (i
= 0; i
< probe
->ftp_ntps
; i
++) {
970 if ((rc
= fasttrap_tracepoint_enable(p
, probe
, i
)) != 0) {
972 * If enabling the tracepoint failed completely,
973 * we don't have to disable it; if the failure
974 * was only partial we must disable it.
976 if (rc
== FASTTRAP_ENABLE_FAIL
)
979 ASSERT(rc
== FASTTRAP_ENABLE_PARTIAL
);
982 * Back up and pull out all the tracepoints we've
983 * created so far for this probe.
986 fasttrap_tracepoint_disable(p
, probe
, i
);
990 mutex_enter(&p
->p_lock
);
994 * Since we're not actually enabling this probe,
995 * drop our reference on the trap table entry.
997 fasttrap_disable_callbacks();
1002 mutex_enter(&p
->p_lock
);
1005 probe
->ftp_enabled
= 1;
1011 fasttrap_pid_disable(void *arg
, dtrace_id_t id
, void *parg
)
1013 fasttrap_probe_t
*probe
= parg
;
1014 fasttrap_provider_t
*provider
= probe
->ftp_prov
;
1018 ASSERT(id
== probe
->ftp_id
);
1021 * We won't be able to acquire a /proc-esque lock on the process
1022 * iff the process is dead and gone. In this case, we rely on the
1023 * provider lock as a point of mutual exclusion to prevent other
1024 * DTrace consumers from disabling this probe.
1026 if ((p
= sprlock(probe
->ftp_pid
)) != NULL
) {
1027 ASSERT(!(p
->p_flag
& SVFORK
));
1028 mutex_exit(&p
->p_lock
);
1031 mutex_enter(&provider
->ftp_mtx
);
1034 * Disable all the associated tracepoints (for fully enabled probes).
1036 if (probe
->ftp_enabled
) {
1037 for (i
= 0; i
< probe
->ftp_ntps
; i
++) {
1038 fasttrap_tracepoint_disable(p
, probe
, i
);
1042 ASSERT(provider
->ftp_rcount
> 0);
1043 provider
->ftp_rcount
--;
1047 * Even though we may not be able to remove it entirely, we
1048 * mark this retired provider to get a chance to remove some
1049 * of the associated probes.
1051 if (provider
->ftp_retired
&& !provider
->ftp_marked
)
1052 whack
= provider
->ftp_marked
= 1;
1053 mutex_exit(&provider
->ftp_mtx
);
1055 mutex_enter(&p
->p_lock
);
1059 * If the process is dead, we're just waiting for the
1060 * last probe to be disabled to be able to free it.
1062 if (provider
->ftp_rcount
== 0 && !provider
->ftp_marked
)
1063 whack
= provider
->ftp_marked
= 1;
1064 mutex_exit(&provider
->ftp_mtx
);
1068 fasttrap_pid_cleanup();
1070 if (!probe
->ftp_enabled
)
1073 probe
->ftp_enabled
= 0;
1075 ASSERT(MUTEX_HELD(&cpu_lock
));
1076 fasttrap_disable_callbacks();
1081 fasttrap_pid_getargdesc(void *arg
, dtrace_id_t id
, void *parg
,
1082 dtrace_argdesc_t
*desc
)
1084 fasttrap_probe_t
*probe
= parg
;
1088 desc
->dtargd_native
[0] = '\0';
1089 desc
->dtargd_xlate
[0] = '\0';
1091 if (probe
->ftp_prov
->ftp_retired
!= 0 ||
1092 desc
->dtargd_ndx
>= probe
->ftp_nargs
) {
1093 desc
->dtargd_ndx
= DTRACE_ARGNONE
;
1097 ndx
= (probe
->ftp_argmap
!= NULL
) ?
1098 probe
->ftp_argmap
[desc
->dtargd_ndx
] : desc
->dtargd_ndx
;
1100 str
= probe
->ftp_ntypes
;
1101 for (i
= 0; i
< ndx
; i
++) {
1102 str
+= strlen(str
) + 1;
1105 ASSERT(strlen(str
+ 1) < sizeof (desc
->dtargd_native
));
1106 (void) strcpy(desc
->dtargd_native
, str
);
1108 if (probe
->ftp_xtypes
== NULL
)
1111 str
= probe
->ftp_xtypes
;
1112 for (i
= 0; i
< desc
->dtargd_ndx
; i
++) {
1113 str
+= strlen(str
) + 1;
1116 ASSERT(strlen(str
+ 1) < sizeof (desc
->dtargd_xlate
));
1117 (void) strcpy(desc
->dtargd_xlate
, str
);
1122 fasttrap_pid_destroy(void *arg
, dtrace_id_t id
, void *parg
)
1124 fasttrap_probe_t
*probe
= parg
;
1128 ASSERT(probe
!= NULL
);
1129 ASSERT(!probe
->ftp_enabled
);
1130 ASSERT(fasttrap_total
>= probe
->ftp_ntps
);
1132 atomic_add_32(&fasttrap_total
, -probe
->ftp_ntps
);
1133 size
= offsetof(fasttrap_probe_t
, ftp_tps
[probe
->ftp_ntps
]);
1135 if (probe
->ftp_gen
+ 1 >= fasttrap_mod_gen
)
1136 fasttrap_mod_barrier(probe
->ftp_gen
);
1138 for (i
= 0; i
< probe
->ftp_ntps
; i
++) {
1139 kmem_free(probe
->ftp_tps
[i
].fit_tp
,
1140 sizeof (fasttrap_tracepoint_t
));
1143 kmem_free(probe
, size
);
1147 static const dtrace_pattr_t pid_attr
= {
1148 { DTRACE_STABILITY_EVOLVING
, DTRACE_STABILITY_EVOLVING
, DTRACE_CLASS_ISA
},
1149 { DTRACE_STABILITY_PRIVATE
, DTRACE_STABILITY_PRIVATE
, DTRACE_CLASS_UNKNOWN
},
1150 { DTRACE_STABILITY_PRIVATE
, DTRACE_STABILITY_PRIVATE
, DTRACE_CLASS_UNKNOWN
},
1151 { DTRACE_STABILITY_EVOLVING
, DTRACE_STABILITY_EVOLVING
, DTRACE_CLASS_ISA
},
1152 { DTRACE_STABILITY_PRIVATE
, DTRACE_STABILITY_PRIVATE
, DTRACE_CLASS_UNKNOWN
},
1155 static dtrace_pops_t pid_pops
= {
1156 fasttrap_pid_provide
,
1158 fasttrap_pid_enable
,
1159 fasttrap_pid_disable
,
1162 fasttrap_pid_getargdesc
,
1163 fasttrap_pid_getarg
,
1165 fasttrap_pid_destroy
1168 static dtrace_pops_t usdt_pops
= {
1169 fasttrap_pid_provide
,
1171 fasttrap_pid_enable
,
1172 fasttrap_pid_disable
,
1175 fasttrap_pid_getargdesc
,
1176 fasttrap_usdt_getarg
,
1178 fasttrap_pid_destroy
1181 static fasttrap_proc_t
*
1182 fasttrap_proc_lookup(pid_t pid
)
1184 fasttrap_bucket_t
*bucket
;
1185 fasttrap_proc_t
*fprc
, *new_fprc
;
1187 bucket
= &fasttrap_procs
.fth_table
[FASTTRAP_PROCS_INDEX(pid
)];
1188 mutex_enter(&bucket
->ftb_mtx
);
1190 for (fprc
= bucket
->ftb_data
; fprc
!= NULL
; fprc
= fprc
->ftpc_next
) {
1191 if (fprc
->ftpc_pid
== pid
&& fprc
->ftpc_acount
!= 0) {
1192 mutex_enter(&fprc
->ftpc_mtx
);
1193 mutex_exit(&bucket
->ftb_mtx
);
1194 fprc
->ftpc_rcount
++;
1195 atomic_inc_64(&fprc
->ftpc_acount
);
1196 ASSERT(fprc
->ftpc_acount
<= fprc
->ftpc_rcount
);
1197 mutex_exit(&fprc
->ftpc_mtx
);
1204 * Drop the bucket lock so we don't try to perform a sleeping
1205 * allocation under it.
1207 mutex_exit(&bucket
->ftb_mtx
);
1209 new_fprc
= kmem_zalloc(sizeof (fasttrap_proc_t
), KM_SLEEP
);
1210 new_fprc
->ftpc_pid
= pid
;
1211 new_fprc
->ftpc_rcount
= 1;
1212 new_fprc
->ftpc_acount
= 1;
1214 mutex_enter(&bucket
->ftb_mtx
);
1217 * Take another lap through the list to make sure a proc hasn't
1218 * been created for this pid while we weren't under the bucket lock.
1220 for (fprc
= bucket
->ftb_data
; fprc
!= NULL
; fprc
= fprc
->ftpc_next
) {
1221 if (fprc
->ftpc_pid
== pid
&& fprc
->ftpc_acount
!= 0) {
1222 mutex_enter(&fprc
->ftpc_mtx
);
1223 mutex_exit(&bucket
->ftb_mtx
);
1224 fprc
->ftpc_rcount
++;
1225 atomic_inc_64(&fprc
->ftpc_acount
);
1226 ASSERT(fprc
->ftpc_acount
<= fprc
->ftpc_rcount
);
1227 mutex_exit(&fprc
->ftpc_mtx
);
1229 kmem_free(new_fprc
, sizeof (fasttrap_proc_t
));
1235 new_fprc
->ftpc_next
= bucket
->ftb_data
;
1236 bucket
->ftb_data
= new_fprc
;
1238 mutex_exit(&bucket
->ftb_mtx
);
1244 fasttrap_proc_release(fasttrap_proc_t
*proc
)
1246 fasttrap_bucket_t
*bucket
;
1247 fasttrap_proc_t
*fprc
, **fprcp
;
1248 pid_t pid
= proc
->ftpc_pid
;
1250 mutex_enter(&proc
->ftpc_mtx
);
1252 ASSERT(proc
->ftpc_rcount
!= 0);
1253 ASSERT(proc
->ftpc_acount
<= proc
->ftpc_rcount
);
1255 if (--proc
->ftpc_rcount
!= 0) {
1256 mutex_exit(&proc
->ftpc_mtx
);
1260 mutex_exit(&proc
->ftpc_mtx
);
1263 * There should definitely be no live providers associated with this
1264 * process at this point.
1266 ASSERT(proc
->ftpc_acount
== 0);
1268 bucket
= &fasttrap_procs
.fth_table
[FASTTRAP_PROCS_INDEX(pid
)];
1269 mutex_enter(&bucket
->ftb_mtx
);
1271 fprcp
= (fasttrap_proc_t
**)&bucket
->ftb_data
;
1272 while ((fprc
= *fprcp
) != NULL
) {
1276 fprcp
= &fprc
->ftpc_next
;
1280 * Something strange has happened if we can't find the proc.
1282 ASSERT(fprc
!= NULL
);
1284 *fprcp
= fprc
->ftpc_next
;
1286 mutex_exit(&bucket
->ftb_mtx
);
1288 kmem_free(fprc
, sizeof (fasttrap_proc_t
));
1292 * Lookup a fasttrap-managed provider based on its name and associated pid.
1293 * If the pattr argument is non-NULL, this function instantiates the provider
1294 * if it doesn't exist otherwise it returns NULL. The provider is returned
1295 * with its lock held.
1297 static fasttrap_provider_t
*
1298 fasttrap_provider_lookup(pid_t pid
, const char *name
,
1299 const dtrace_pattr_t
*pattr
)
1301 fasttrap_provider_t
*fp
, *new_fp
= NULL
;
1302 fasttrap_bucket_t
*bucket
;
1303 char provname
[DTRACE_PROVNAMELEN
];
1307 ASSERT(strlen(name
) < sizeof (fp
->ftp_name
));
1308 ASSERT(pattr
!= NULL
);
1310 bucket
= &fasttrap_provs
.fth_table
[FASTTRAP_PROVS_INDEX(pid
, name
)];
1311 mutex_enter(&bucket
->ftb_mtx
);
1314 * Take a lap through the list and return the match if we find it.
1316 for (fp
= bucket
->ftb_data
; fp
!= NULL
; fp
= fp
->ftp_next
) {
1317 if (fp
->ftp_pid
== pid
&& strcmp(fp
->ftp_name
, name
) == 0 &&
1319 mutex_enter(&fp
->ftp_mtx
);
1320 mutex_exit(&bucket
->ftb_mtx
);
1326 * Drop the bucket lock so we don't try to perform a sleeping
1327 * allocation under it.
1329 mutex_exit(&bucket
->ftb_mtx
);
1332 * Make sure the process exists, isn't a child created as the result
1333 * of a vfork(2), and isn't a zombie (but may be in fork).
1335 mutex_enter(&pidlock
);
1336 if ((p
= prfind(pid
)) == NULL
) {
1337 mutex_exit(&pidlock
);
1340 mutex_enter(&p
->p_lock
);
1341 mutex_exit(&pidlock
);
1342 if (p
->p_flag
& (SVFORK
| SEXITING
)) {
1343 mutex_exit(&p
->p_lock
);
1348 * Increment p_dtrace_probes so that the process knows to inform us
1349 * when it exits or execs. fasttrap_provider_free() decrements this
1350 * when we're done with this provider.
1352 p
->p_dtrace_probes
++;
1355 * Grab the credentials for this process so we have
1356 * something to pass to dtrace_register().
1358 mutex_enter(&p
->p_crlock
);
1361 mutex_exit(&p
->p_crlock
);
1362 mutex_exit(&p
->p_lock
);
1364 new_fp
= kmem_zalloc(sizeof (fasttrap_provider_t
), KM_SLEEP
);
1365 new_fp
->ftp_pid
= pid
;
1366 new_fp
->ftp_proc
= fasttrap_proc_lookup(pid
);
1368 ASSERT(new_fp
->ftp_proc
!= NULL
);
1370 mutex_enter(&bucket
->ftb_mtx
);
1373 * Take another lap through the list to make sure a provider hasn't
1374 * been created for this pid while we weren't under the bucket lock.
1376 for (fp
= bucket
->ftb_data
; fp
!= NULL
; fp
= fp
->ftp_next
) {
1377 if (fp
->ftp_pid
== pid
&& strcmp(fp
->ftp_name
, name
) == 0 &&
1379 mutex_enter(&fp
->ftp_mtx
);
1380 mutex_exit(&bucket
->ftb_mtx
);
1381 fasttrap_provider_free(new_fp
);
1387 (void) strcpy(new_fp
->ftp_name
, name
);
1390 * Fail and return NULL if either the provider name is too long
1391 * or we fail to register this new provider with the DTrace
1392 * framework. Note that this is the only place we ever construct
1393 * the full provider name -- we keep it in pieces in the provider
1396 if (snprintf(provname
, sizeof (provname
), "%s%u", name
, (uint_t
)pid
) >=
1397 sizeof (provname
) ||
1398 dtrace_register(provname
, pattr
,
1399 DTRACE_PRIV_PROC
| DTRACE_PRIV_OWNER
| DTRACE_PRIV_ZONEOWNER
, cred
,
1400 pattr
== &pid_attr
? &pid_pops
: &usdt_pops
, new_fp
,
1401 &new_fp
->ftp_provid
) != 0) {
1402 mutex_exit(&bucket
->ftb_mtx
);
1403 fasttrap_provider_free(new_fp
);
1408 new_fp
->ftp_next
= bucket
->ftb_data
;
1409 bucket
->ftb_data
= new_fp
;
1411 mutex_enter(&new_fp
->ftp_mtx
);
1412 mutex_exit(&bucket
->ftb_mtx
);
1419 fasttrap_provider_free(fasttrap_provider_t
*provider
)
1421 pid_t pid
= provider
->ftp_pid
;
1425 * There need to be no associated enabled probes, no consumers
1426 * creating probes, and no meta providers referencing this provider.
1428 ASSERT(provider
->ftp_rcount
== 0);
1429 ASSERT(provider
->ftp_ccount
== 0);
1430 ASSERT(provider
->ftp_mcount
== 0);
1433 * If this provider hasn't been retired, we need to explicitly drop the
1434 * count of active providers on the associated process structure.
1436 if (!provider
->ftp_retired
) {
1437 atomic_dec_64(&provider
->ftp_proc
->ftpc_acount
);
1438 ASSERT(provider
->ftp_proc
->ftpc_acount
<
1439 provider
->ftp_proc
->ftpc_rcount
);
1442 fasttrap_proc_release(provider
->ftp_proc
);
1444 kmem_free(provider
, sizeof (fasttrap_provider_t
));
1447 * Decrement p_dtrace_probes on the process whose provider we're
1448 * freeing. We don't have to worry about clobbering somone else's
1449 * modifications to it because we have locked the bucket that
1450 * corresponds to this process's hash chain in the provider hash
1451 * table. Don't sweat it if we can't find the process.
1453 mutex_enter(&pidlock
);
1454 if ((p
= prfind(pid
)) == NULL
) {
1455 mutex_exit(&pidlock
);
1459 mutex_enter(&p
->p_lock
);
1460 mutex_exit(&pidlock
);
1462 p
->p_dtrace_probes
--;
1463 mutex_exit(&p
->p_lock
);
1467 fasttrap_provider_retire(pid_t pid
, const char *name
, int mprov
)
1469 fasttrap_provider_t
*fp
;
1470 fasttrap_bucket_t
*bucket
;
1471 dtrace_provider_id_t provid
;
1473 ASSERT(strlen(name
) < sizeof (fp
->ftp_name
));
1475 bucket
= &fasttrap_provs
.fth_table
[FASTTRAP_PROVS_INDEX(pid
, name
)];
1476 mutex_enter(&bucket
->ftb_mtx
);
1478 for (fp
= bucket
->ftb_data
; fp
!= NULL
; fp
= fp
->ftp_next
) {
1479 if (fp
->ftp_pid
== pid
&& strcmp(fp
->ftp_name
, name
) == 0 &&
1485 mutex_exit(&bucket
->ftb_mtx
);
1489 mutex_enter(&fp
->ftp_mtx
);
1490 ASSERT(!mprov
|| fp
->ftp_mcount
> 0);
1491 if (mprov
&& --fp
->ftp_mcount
!= 0) {
1492 mutex_exit(&fp
->ftp_mtx
);
1493 mutex_exit(&bucket
->ftb_mtx
);
1498 * Mark the provider to be removed in our post-processing step, mark it
1499 * retired, and drop the active count on its proc. Marking it indicates
1500 * that we should try to remove it; setting the retired flag indicates
1501 * that we're done with this provider; dropping the active the proc
1502 * releases our hold, and when this reaches zero (as it will during
1503 * exit or exec) the proc and associated providers become defunct.
1505 * We obviously need to take the bucket lock before the provider lock
1506 * to perform the lookup, but we need to drop the provider lock
1507 * before calling into the DTrace framework since we acquire the
1508 * provider lock in callbacks invoked from the DTrace framework. The
1509 * bucket lock therefore protects the integrity of the provider hash
1512 atomic_dec_64(&fp
->ftp_proc
->ftpc_acount
);
1513 ASSERT(fp
->ftp_proc
->ftpc_acount
< fp
->ftp_proc
->ftpc_rcount
);
1515 fp
->ftp_retired
= 1;
1517 provid
= fp
->ftp_provid
;
1518 mutex_exit(&fp
->ftp_mtx
);
1521 * We don't have to worry about invalidating the same provider twice
1522 * since fasttrap_provider_lookup() will ignore provider that have
1523 * been marked as retired.
1525 dtrace_invalidate(provid
);
1527 mutex_exit(&bucket
->ftb_mtx
);
1529 fasttrap_pid_cleanup();
1533 fasttrap_uint32_cmp(const void *ap
, const void *bp
)
1535 return (*(const uint32_t *)ap
- *(const uint32_t *)bp
);
1539 fasttrap_uint64_cmp(const void *ap
, const void *bp
)
1541 return (*(const uint64_t *)ap
- *(const uint64_t *)bp
);
1545 fasttrap_add_probe(fasttrap_probe_spec_t
*pdata
)
1547 fasttrap_provider_t
*provider
;
1548 fasttrap_probe_t
*pp
;
1549 fasttrap_tracepoint_t
*tp
;
1551 int i
, aframes
, whack
;
1554 * There needs to be at least one desired trace point.
1556 if (pdata
->ftps_noffs
== 0)
1559 switch (pdata
->ftps_type
) {
1562 aframes
= FASTTRAP_ENTRY_AFRAMES
;
1566 aframes
= FASTTRAP_RETURN_AFRAMES
;
1575 if ((provider
= fasttrap_provider_lookup(pdata
->ftps_pid
,
1576 FASTTRAP_PID_NAME
, &pid_attr
)) == NULL
)
1580 * Increment this reference count to indicate that a consumer is
1581 * actively adding a new probe associated with this provider. This
1582 * prevents the provider from being deleted -- we'll need to check
1583 * for pending deletions when we drop this reference count.
1585 provider
->ftp_ccount
++;
1586 mutex_exit(&provider
->ftp_mtx
);
1589 * Grab the creation lock to ensure consistency between calls to
1590 * dtrace_probe_lookup() and dtrace_probe_create() in the face of
1591 * other threads creating probes. We must drop the provider lock
1592 * before taking this lock to avoid a three-way deadlock with the
1595 mutex_enter(&provider
->ftp_cmtx
);
1598 for (i
= 0; i
< pdata
->ftps_noffs
; i
++) {
1601 (void) sprintf(name_str
, "%llx",
1602 (unsigned long long)pdata
->ftps_offs
[i
]);
1604 if (dtrace_probe_lookup(provider
->ftp_provid
,
1605 pdata
->ftps_mod
, pdata
->ftps_func
, name_str
) != 0)
1608 atomic_inc_32(&fasttrap_total
);
1610 if (fasttrap_total
> fasttrap_max
) {
1611 atomic_dec_32(&fasttrap_total
);
1615 pp
= kmem_zalloc(sizeof (fasttrap_probe_t
), KM_SLEEP
);
1617 pp
->ftp_prov
= provider
;
1618 pp
->ftp_faddr
= pdata
->ftps_pc
;
1619 pp
->ftp_fsize
= pdata
->ftps_size
;
1620 pp
->ftp_pid
= pdata
->ftps_pid
;
1623 tp
= kmem_zalloc(sizeof (fasttrap_tracepoint_t
),
1626 tp
->ftt_proc
= provider
->ftp_proc
;
1627 tp
->ftt_pc
= pdata
->ftps_offs
[i
] + pdata
->ftps_pc
;
1628 tp
->ftt_pid
= pdata
->ftps_pid
;
1630 pp
->ftp_tps
[0].fit_tp
= tp
;
1631 pp
->ftp_tps
[0].fit_id
.fti_probe
= pp
;
1632 pp
->ftp_tps
[0].fit_id
.fti_ptype
= pdata
->ftps_type
;
1634 pp
->ftp_id
= dtrace_probe_create(provider
->ftp_provid
,
1635 pdata
->ftps_mod
, pdata
->ftps_func
, name_str
,
1636 FASTTRAP_OFFSET_AFRAMES
, pp
);
1639 } else if (dtrace_probe_lookup(provider
->ftp_provid
, pdata
->ftps_mod
,
1640 pdata
->ftps_func
, name
) == 0) {
1641 atomic_add_32(&fasttrap_total
, pdata
->ftps_noffs
);
1643 if (fasttrap_total
> fasttrap_max
) {
1644 atomic_add_32(&fasttrap_total
, -pdata
->ftps_noffs
);
1649 * Make sure all tracepoint program counter values are unique.
1650 * We later assume that each probe has exactly one tracepoint
1653 qsort(pdata
->ftps_offs
, pdata
->ftps_noffs
,
1654 sizeof (uint64_t), fasttrap_uint64_cmp
);
1655 for (i
= 1; i
< pdata
->ftps_noffs
; i
++) {
1656 if (pdata
->ftps_offs
[i
] > pdata
->ftps_offs
[i
- 1])
1659 atomic_add_32(&fasttrap_total
, -pdata
->ftps_noffs
);
1663 ASSERT(pdata
->ftps_noffs
> 0);
1664 pp
= kmem_zalloc(offsetof(fasttrap_probe_t
,
1665 ftp_tps
[pdata
->ftps_noffs
]), KM_SLEEP
);
1667 pp
->ftp_prov
= provider
;
1668 pp
->ftp_faddr
= pdata
->ftps_pc
;
1669 pp
->ftp_fsize
= pdata
->ftps_size
;
1670 pp
->ftp_pid
= pdata
->ftps_pid
;
1671 pp
->ftp_ntps
= pdata
->ftps_noffs
;
1673 for (i
= 0; i
< pdata
->ftps_noffs
; i
++) {
1674 tp
= kmem_zalloc(sizeof (fasttrap_tracepoint_t
),
1677 tp
->ftt_proc
= provider
->ftp_proc
;
1678 tp
->ftt_pc
= pdata
->ftps_offs
[i
] + pdata
->ftps_pc
;
1679 tp
->ftt_pid
= pdata
->ftps_pid
;
1681 pp
->ftp_tps
[i
].fit_tp
= tp
;
1682 pp
->ftp_tps
[i
].fit_id
.fti_probe
= pp
;
1683 pp
->ftp_tps
[i
].fit_id
.fti_ptype
= pdata
->ftps_type
;
1686 pp
->ftp_id
= dtrace_probe_create(provider
->ftp_provid
,
1687 pdata
->ftps_mod
, pdata
->ftps_func
, name
, aframes
, pp
);
1690 mutex_exit(&provider
->ftp_cmtx
);
1693 * We know that the provider is still valid since we incremented the
1694 * creation reference count. If someone tried to clean up this provider
1695 * while we were using it (e.g. because the process called exec(2) or
1696 * exit(2)), take note of that and try to clean it up now.
1698 mutex_enter(&provider
->ftp_mtx
);
1699 provider
->ftp_ccount
--;
1700 whack
= provider
->ftp_retired
;
1701 mutex_exit(&provider
->ftp_mtx
);
1704 fasttrap_pid_cleanup();
1710 * If we've exhausted the allowable resources, we'll try to remove
1711 * this provider to free some up. This is to cover the case where
1712 * the user has accidentally created many more probes than was
1713 * intended (e.g. pid123:::).
1715 mutex_exit(&provider
->ftp_cmtx
);
1716 mutex_enter(&provider
->ftp_mtx
);
1717 provider
->ftp_ccount
--;
1718 provider
->ftp_marked
= 1;
1719 mutex_exit(&provider
->ftp_mtx
);
1721 fasttrap_pid_cleanup();
1728 fasttrap_meta_provide(void *arg
, dtrace_helper_provdesc_t
*dhpv
, pid_t pid
)
1730 fasttrap_provider_t
*provider
;
1733 * A 32-bit unsigned integer (like a pid for example) can be
1734 * expressed in 10 or fewer decimal digits. Make sure that we'll
1735 * have enough space for the provider name.
1737 if (strlen(dhpv
->dthpv_provname
) + 10 >=
1738 sizeof (provider
->ftp_name
)) {
1739 cmn_err(CE_WARN
, "failed to instantiate provider %s: "
1740 "name too long to accomodate pid", dhpv
->dthpv_provname
);
1745 * Don't let folks spoof the true pid provider.
1747 if (strcmp(dhpv
->dthpv_provname
, FASTTRAP_PID_NAME
) == 0) {
1748 cmn_err(CE_WARN
, "failed to instantiate provider %s: "
1749 "%s is an invalid name", dhpv
->dthpv_provname
,
1755 * The highest stability class that fasttrap supports is ISA; cap
1756 * the stability of the new provider accordingly.
1758 if (dhpv
->dthpv_pattr
.dtpa_provider
.dtat_class
> DTRACE_CLASS_ISA
)
1759 dhpv
->dthpv_pattr
.dtpa_provider
.dtat_class
= DTRACE_CLASS_ISA
;
1760 if (dhpv
->dthpv_pattr
.dtpa_mod
.dtat_class
> DTRACE_CLASS_ISA
)
1761 dhpv
->dthpv_pattr
.dtpa_mod
.dtat_class
= DTRACE_CLASS_ISA
;
1762 if (dhpv
->dthpv_pattr
.dtpa_func
.dtat_class
> DTRACE_CLASS_ISA
)
1763 dhpv
->dthpv_pattr
.dtpa_func
.dtat_class
= DTRACE_CLASS_ISA
;
1764 if (dhpv
->dthpv_pattr
.dtpa_name
.dtat_class
> DTRACE_CLASS_ISA
)
1765 dhpv
->dthpv_pattr
.dtpa_name
.dtat_class
= DTRACE_CLASS_ISA
;
1766 if (dhpv
->dthpv_pattr
.dtpa_args
.dtat_class
> DTRACE_CLASS_ISA
)
1767 dhpv
->dthpv_pattr
.dtpa_args
.dtat_class
= DTRACE_CLASS_ISA
;
1769 if ((provider
= fasttrap_provider_lookup(pid
, dhpv
->dthpv_provname
,
1770 &dhpv
->dthpv_pattr
)) == NULL
) {
1771 cmn_err(CE_WARN
, "failed to instantiate provider %s for "
1772 "process %u", dhpv
->dthpv_provname
, (uint_t
)pid
);
1777 * Up the meta provider count so this provider isn't removed until
1778 * the meta provider has been told to remove it.
1780 provider
->ftp_mcount
++;
1782 mutex_exit(&provider
->ftp_mtx
);
1788 * We know a few things about our context here: we know that the probe being
1789 * created doesn't already exist (DTrace won't load DOF at the same address
1790 * twice, even if explicitly told to do so) and we know that we are
1791 * single-threaded with respect to the meta provider machinery. Knowing that
1792 * this is a new probe and that there is no way for us to race with another
1793 * operation on this provider allows us an important optimization: we need not
1794 * lookup a probe before adding it. Saving this lookup is important because
1795 * this code is in the fork path for processes with USDT probes, and lookups
1796 * here are potentially very expensive because of long hash conflicts on
1797 * module, function and name (DTrace doesn't hash on provider name).
1801 fasttrap_meta_create_probe(void *arg
, void *parg
,
1802 dtrace_helper_probedesc_t
*dhpb
)
1804 fasttrap_provider_t
*provider
= parg
;
1805 fasttrap_probe_t
*pp
;
1806 fasttrap_tracepoint_t
*tp
;
1811 * Since the meta provider count is non-zero we don't have to worry
1812 * about this provider disappearing.
1814 ASSERT(provider
->ftp_mcount
> 0);
1817 * The offsets must be unique.
1819 qsort(dhpb
->dthpb_offs
, dhpb
->dthpb_noffs
, sizeof (uint32_t),
1820 fasttrap_uint32_cmp
);
1821 for (i
= 1; i
< dhpb
->dthpb_noffs
; i
++) {
1822 if (dhpb
->dthpb_base
+ dhpb
->dthpb_offs
[i
] <=
1823 dhpb
->dthpb_base
+ dhpb
->dthpb_offs
[i
- 1])
1827 qsort(dhpb
->dthpb_enoffs
, dhpb
->dthpb_nenoffs
, sizeof (uint32_t),
1828 fasttrap_uint32_cmp
);
1829 for (i
= 1; i
< dhpb
->dthpb_nenoffs
; i
++) {
1830 if (dhpb
->dthpb_base
+ dhpb
->dthpb_enoffs
[i
] <=
1831 dhpb
->dthpb_base
+ dhpb
->dthpb_enoffs
[i
- 1])
1835 ntps
= dhpb
->dthpb_noffs
+ dhpb
->dthpb_nenoffs
;
1838 atomic_add_32(&fasttrap_total
, ntps
);
1840 if (fasttrap_total
> fasttrap_max
) {
1841 atomic_add_32(&fasttrap_total
, -ntps
);
1845 pp
= kmem_zalloc(offsetof(fasttrap_probe_t
, ftp_tps
[ntps
]), KM_SLEEP
);
1847 pp
->ftp_prov
= provider
;
1848 pp
->ftp_pid
= provider
->ftp_pid
;
1849 pp
->ftp_ntps
= ntps
;
1850 pp
->ftp_nargs
= dhpb
->dthpb_xargc
;
1851 pp
->ftp_xtypes
= dhpb
->dthpb_xtypes
;
1852 pp
->ftp_ntypes
= dhpb
->dthpb_ntypes
;
1855 * First create a tracepoint for each actual point of interest.
1857 for (i
= 0; i
< dhpb
->dthpb_noffs
; i
++) {
1858 tp
= kmem_zalloc(sizeof (fasttrap_tracepoint_t
), KM_SLEEP
);
1860 tp
->ftt_proc
= provider
->ftp_proc
;
1861 tp
->ftt_pc
= dhpb
->dthpb_base
+ dhpb
->dthpb_offs
[i
];
1862 tp
->ftt_pid
= provider
->ftp_pid
;
1864 pp
->ftp_tps
[i
].fit_tp
= tp
;
1865 pp
->ftp_tps
[i
].fit_id
.fti_probe
= pp
;
1866 pp
->ftp_tps
[i
].fit_id
.fti_ptype
= DTFTP_OFFSETS
;
1870 * Then create a tracepoint for each is-enabled point.
1872 for (j
= 0; i
< ntps
; i
++, j
++) {
1873 tp
= kmem_zalloc(sizeof (fasttrap_tracepoint_t
), KM_SLEEP
);
1875 tp
->ftt_proc
= provider
->ftp_proc
;
1876 tp
->ftt_pc
= dhpb
->dthpb_base
+ dhpb
->dthpb_enoffs
[j
];
1877 tp
->ftt_pid
= provider
->ftp_pid
;
1879 pp
->ftp_tps
[i
].fit_tp
= tp
;
1880 pp
->ftp_tps
[i
].fit_id
.fti_probe
= pp
;
1881 pp
->ftp_tps
[i
].fit_id
.fti_ptype
= DTFTP_IS_ENABLED
;
1885 * If the arguments are shuffled around we set the argument remapping
1886 * table. Later, when the probe fires, we only remap the arguments
1887 * if the table is non-NULL.
1889 for (i
= 0; i
< dhpb
->dthpb_xargc
; i
++) {
1890 if (dhpb
->dthpb_args
[i
] != i
) {
1891 pp
->ftp_argmap
= dhpb
->dthpb_args
;
1897 * The probe is fully constructed -- register it with DTrace.
1899 pp
->ftp_id
= dtrace_probe_create(provider
->ftp_provid
, dhpb
->dthpb_mod
,
1900 dhpb
->dthpb_func
, dhpb
->dthpb_name
, FASTTRAP_OFFSET_AFRAMES
, pp
);
1905 fasttrap_meta_remove(void *arg
, dtrace_helper_provdesc_t
*dhpv
, pid_t pid
)
1908 * Clean up the USDT provider. There may be active consumers of the
1909 * provider busy adding probes, no damage will actually befall the
1910 * provider until that count has dropped to zero. This just puts
1911 * the provider on death row.
1913 fasttrap_provider_retire(pid
, dhpv
->dthpv_provname
, 1);
1916 static dtrace_mops_t fasttrap_mops
= {
1917 fasttrap_meta_create_probe
,
1918 fasttrap_meta_provide
,
1919 fasttrap_meta_remove
1924 fasttrap_open(dev_t
*devp
, int flag
, int otyp
, cred_t
*cred_p
)
1931 fasttrap_ioctl(dev_t dev
, int cmd
, intptr_t arg
, int md
, cred_t
*cr
, int *rv
)
1933 if (!dtrace_attached())
1936 if (cmd
== FASTTRAPIOC_MAKEPROBE
) {
1937 fasttrap_probe_spec_t
*uprobe
= (void *)arg
;
1938 fasttrap_probe_spec_t
*probe
;
1943 if (copyin(&uprobe
->ftps_noffs
, &noffs
,
1944 sizeof (uprobe
->ftps_noffs
)))
1948 * Probes must have at least one tracepoint.
1953 size
= sizeof (fasttrap_probe_spec_t
) +
1954 sizeof (probe
->ftps_offs
[0]) * (noffs
- 1);
1956 if (size
> 1024 * 1024)
1959 probe
= kmem_alloc(size
, KM_SLEEP
);
1961 if (copyin(uprobe
, probe
, size
) != 0 ||
1962 probe
->ftps_noffs
!= noffs
) {
1963 kmem_free(probe
, size
);
1968 * Verify that the function and module strings contain no
1971 if (u8_validate(probe
->ftps_func
, strlen(probe
->ftps_func
),
1972 NULL
, U8_VALIDATE_ENTIRE
, &err
) < 0) {
1977 if (u8_validate(probe
->ftps_mod
, strlen(probe
->ftps_mod
),
1978 NULL
, U8_VALIDATE_ENTIRE
, &err
) < 0) {
1983 if (!PRIV_POLICY_CHOICE(cr
, PRIV_ALL
, B_FALSE
)) {
1985 pid_t pid
= probe
->ftps_pid
;
1987 mutex_enter(&pidlock
);
1989 * Report an error if the process doesn't exist
1990 * or is actively being birthed.
1992 if ((p
= prfind(pid
)) == NULL
|| p
->p_stat
== SIDL
) {
1993 mutex_exit(&pidlock
);
1996 mutex_enter(&p
->p_lock
);
1997 mutex_exit(&pidlock
);
1999 if ((ret
= priv_proc_cred_perm(cr
, p
, NULL
,
2000 VREAD
| VWRITE
)) != 0) {
2001 mutex_exit(&p
->p_lock
);
2005 mutex_exit(&p
->p_lock
);
2008 ret
= fasttrap_add_probe(probe
);
2010 kmem_free(probe
, size
);
2014 } else if (cmd
== FASTTRAPIOC_GETINSTR
) {
2015 fasttrap_instr_query_t instr
;
2016 fasttrap_tracepoint_t
*tp
;
2020 if (copyin((void *)arg
, &instr
, sizeof (instr
)) != 0)
2023 if (!PRIV_POLICY_CHOICE(cr
, PRIV_ALL
, B_FALSE
)) {
2025 pid_t pid
= instr
.ftiq_pid
;
2027 mutex_enter(&pidlock
);
2029 * Report an error if the process doesn't exist
2030 * or is actively being birthed.
2032 if ((p
= prfind(pid
)) == NULL
|| p
->p_stat
== SIDL
) {
2033 mutex_exit(&pidlock
);
2036 mutex_enter(&p
->p_lock
);
2037 mutex_exit(&pidlock
);
2039 if ((ret
= priv_proc_cred_perm(cr
, p
, NULL
,
2041 mutex_exit(&p
->p_lock
);
2045 mutex_exit(&p
->p_lock
);
2048 index
= FASTTRAP_TPOINTS_INDEX(instr
.ftiq_pid
, instr
.ftiq_pc
);
2050 mutex_enter(&fasttrap_tpoints
.fth_table
[index
].ftb_mtx
);
2051 tp
= fasttrap_tpoints
.fth_table
[index
].ftb_data
;
2052 while (tp
!= NULL
) {
2053 if (instr
.ftiq_pid
== tp
->ftt_pid
&&
2054 instr
.ftiq_pc
== tp
->ftt_pc
&&
2055 tp
->ftt_proc
->ftpc_acount
!= 0)
2062 mutex_exit(&fasttrap_tpoints
.fth_table
[index
].ftb_mtx
);
2066 bcopy(&tp
->ftt_instr
, &instr
.ftiq_instr
,
2067 sizeof (instr
.ftiq_instr
));
2068 mutex_exit(&fasttrap_tpoints
.fth_table
[index
].ftb_mtx
);
2070 if (copyout(&instr
, (void *)arg
, sizeof (instr
)) != 0)
2079 static struct cb_ops fasttrap_cb_ops
= {
2080 fasttrap_open
, /* open */
2082 nulldev
, /* strategy */
2083 nulldev
, /* print */
2087 fasttrap_ioctl
, /* ioctl */
2091 nochpoll
, /* poll */
2092 ddi_prop_op
, /* cb_prop_op */
2094 D_NEW
| D_MP
/* Driver compatibility flag */
2099 fasttrap_info(dev_info_t
*dip
, ddi_info_cmd_t infocmd
, void *arg
, void **result
)
2104 case DDI_INFO_DEVT2DEVINFO
:
2105 *result
= (void *)fasttrap_devi
;
2106 error
= DDI_SUCCESS
;
2108 case DDI_INFO_DEVT2INSTANCE
:
2110 error
= DDI_SUCCESS
;
2113 error
= DDI_FAILURE
;
2119 fasttrap_attach(dev_info_t
*devi
, ddi_attach_cmd_t cmd
)
2127 return (DDI_SUCCESS
);
2129 return (DDI_FAILURE
);
2132 if (ddi_create_minor_node(devi
, "fasttrap", S_IFCHR
, 0,
2133 DDI_PSEUDO
, 0) == DDI_FAILURE
) {
2134 ddi_remove_minor_node(devi
, NULL
);
2135 return (DDI_FAILURE
);
2138 ddi_report_dev(devi
);
2139 fasttrap_devi
= devi
;
2142 * Install our hooks into fork(2), exec(2), and exit(2).
2144 dtrace_fasttrap_fork_ptr
= &fasttrap_fork
;
2145 dtrace_fasttrap_exit_ptr
= &fasttrap_exec_exit
;
2146 dtrace_fasttrap_exec_ptr
= &fasttrap_exec_exit
;
2148 fasttrap_max
= ddi_getprop(DDI_DEV_T_ANY
, devi
, DDI_PROP_DONTPASS
,
2149 "fasttrap-max-probes", FASTTRAP_MAX_DEFAULT
);
2153 * Conjure up the tracepoints hashtable...
2155 nent
= ddi_getprop(DDI_DEV_T_ANY
, devi
, DDI_PROP_DONTPASS
,
2156 "fasttrap-hash-size", FASTTRAP_TPOINTS_DEFAULT_SIZE
);
2158 if (nent
== 0 || nent
> 0x1000000)
2159 nent
= FASTTRAP_TPOINTS_DEFAULT_SIZE
;
2162 fasttrap_tpoints
.fth_nent
= nent
;
2164 fasttrap_tpoints
.fth_nent
= 1 << fasttrap_highbit(nent
);
2165 ASSERT(fasttrap_tpoints
.fth_nent
> 0);
2166 fasttrap_tpoints
.fth_mask
= fasttrap_tpoints
.fth_nent
- 1;
2167 fasttrap_tpoints
.fth_table
= kmem_zalloc(fasttrap_tpoints
.fth_nent
*
2168 sizeof (fasttrap_bucket_t
), KM_SLEEP
);
2171 * ... and the providers hash table...
2173 nent
= FASTTRAP_PROVIDERS_DEFAULT_SIZE
;
2175 fasttrap_provs
.fth_nent
= nent
;
2177 fasttrap_provs
.fth_nent
= 1 << fasttrap_highbit(nent
);
2178 ASSERT(fasttrap_provs
.fth_nent
> 0);
2179 fasttrap_provs
.fth_mask
= fasttrap_provs
.fth_nent
- 1;
2180 fasttrap_provs
.fth_table
= kmem_zalloc(fasttrap_provs
.fth_nent
*
2181 sizeof (fasttrap_bucket_t
), KM_SLEEP
);
2184 * ... and the procs hash table.
2186 nent
= FASTTRAP_PROCS_DEFAULT_SIZE
;
2188 fasttrap_procs
.fth_nent
= nent
;
2190 fasttrap_procs
.fth_nent
= 1 << fasttrap_highbit(nent
);
2191 ASSERT(fasttrap_procs
.fth_nent
> 0);
2192 fasttrap_procs
.fth_mask
= fasttrap_procs
.fth_nent
- 1;
2193 fasttrap_procs
.fth_table
= kmem_zalloc(fasttrap_procs
.fth_nent
*
2194 sizeof (fasttrap_bucket_t
), KM_SLEEP
);
2196 (void) dtrace_meta_register("fasttrap", &fasttrap_mops
, NULL
,
2199 return (DDI_SUCCESS
);
2203 fasttrap_detach(dev_info_t
*devi
, ddi_detach_cmd_t cmd
)
2212 return (DDI_SUCCESS
);
2214 return (DDI_FAILURE
);
2218 * Unregister the meta-provider to make sure no new fasttrap-
2219 * managed providers come along while we're trying to close up
2220 * shop. If we fail to detach, we'll need to re-register as a
2221 * meta-provider. We can fail to unregister as a meta-provider
2222 * if providers we manage still exist.
2224 if (fasttrap_meta_id
!= DTRACE_METAPROVNONE
&&
2225 dtrace_meta_unregister(fasttrap_meta_id
) != 0)
2226 return (DDI_FAILURE
);
2229 * Prevent any new timeouts from running by setting fasttrap_timeout
2230 * to a non-zero value, and wait for the current timeout to complete.
2232 mutex_enter(&fasttrap_cleanup_mtx
);
2233 fasttrap_cleanup_work
= 0;
2235 while (fasttrap_timeout
!= (timeout_id_t
)1) {
2236 tmp
= fasttrap_timeout
;
2237 fasttrap_timeout
= (timeout_id_t
)1;
2240 mutex_exit(&fasttrap_cleanup_mtx
);
2241 (void) untimeout(tmp
);
2242 mutex_enter(&fasttrap_cleanup_mtx
);
2246 fasttrap_cleanup_work
= 0;
2247 mutex_exit(&fasttrap_cleanup_mtx
);
2250 * Iterate over all of our providers. If there's still a process
2251 * that corresponds to that pid, fail to detach.
2253 for (i
= 0; i
< fasttrap_provs
.fth_nent
; i
++) {
2254 fasttrap_provider_t
**fpp
, *fp
;
2255 fasttrap_bucket_t
*bucket
= &fasttrap_provs
.fth_table
[i
];
2257 mutex_enter(&bucket
->ftb_mtx
);
2258 fpp
= (fasttrap_provider_t
**)&bucket
->ftb_data
;
2259 while ((fp
= *fpp
) != NULL
) {
2261 * Acquire and release the lock as a simple way of
2262 * waiting for any other consumer to finish with
2263 * this provider. A thread must first acquire the
2264 * bucket lock so there's no chance of another thread
2265 * blocking on the provider's lock.
2267 mutex_enter(&fp
->ftp_mtx
);
2268 mutex_exit(&fp
->ftp_mtx
);
2270 if (dtrace_unregister(fp
->ftp_provid
) != 0) {
2272 fpp
= &fp
->ftp_next
;
2274 *fpp
= fp
->ftp_next
;
2275 fasttrap_provider_free(fp
);
2279 mutex_exit(&bucket
->ftb_mtx
);
2285 * If we're failing to detach, we need to unblock timeouts
2286 * and start a new timeout if any work has accumulated while
2287 * we've been unsuccessfully trying to detach.
2289 mutex_enter(&fasttrap_cleanup_mtx
);
2290 fasttrap_timeout
= 0;
2291 work
= fasttrap_cleanup_work
;
2292 mutex_exit(&fasttrap_cleanup_mtx
);
2295 fasttrap_pid_cleanup();
2297 (void) dtrace_meta_register("fasttrap", &fasttrap_mops
, NULL
,
2300 return (DDI_FAILURE
);
2304 mutex_enter(&fasttrap_count_mtx
);
2305 ASSERT(fasttrap_pid_count
== 0);
2306 mutex_exit(&fasttrap_count_mtx
);
2309 kmem_free(fasttrap_tpoints
.fth_table
,
2310 fasttrap_tpoints
.fth_nent
* sizeof (fasttrap_bucket_t
));
2311 fasttrap_tpoints
.fth_nent
= 0;
2313 kmem_free(fasttrap_provs
.fth_table
,
2314 fasttrap_provs
.fth_nent
* sizeof (fasttrap_bucket_t
));
2315 fasttrap_provs
.fth_nent
= 0;
2317 kmem_free(fasttrap_procs
.fth_table
,
2318 fasttrap_procs
.fth_nent
* sizeof (fasttrap_bucket_t
));
2319 fasttrap_procs
.fth_nent
= 0;
2322 * We know there are no tracepoints in any process anywhere in
2323 * the system so there is no process which has its p_dtrace_count
2324 * greater than zero, therefore we know that no thread can actively
2325 * be executing code in fasttrap_fork(). Similarly for p_dtrace_probes
2326 * and fasttrap_exec() and fasttrap_exit().
2328 ASSERT(dtrace_fasttrap_fork_ptr
== &fasttrap_fork
);
2329 dtrace_fasttrap_fork_ptr
= NULL
;
2331 ASSERT(dtrace_fasttrap_exec_ptr
== &fasttrap_exec_exit
);
2332 dtrace_fasttrap_exec_ptr
= NULL
;
2334 ASSERT(dtrace_fasttrap_exit_ptr
== &fasttrap_exec_exit
);
2335 dtrace_fasttrap_exit_ptr
= NULL
;
2337 ddi_remove_minor_node(devi
, NULL
);
2339 return (DDI_SUCCESS
);
2342 static struct dev_ops fasttrap_ops
= {
2343 DEVO_REV
, /* devo_rev */
2345 fasttrap_info
, /* get_dev_info */
2346 nulldev
, /* identify */
2347 nulldev
, /* probe */
2348 fasttrap_attach
, /* attach */
2349 fasttrap_detach
, /* detach */
2351 &fasttrap_cb_ops
, /* driver operations */
2352 NULL
, /* bus operations */
2353 nodev
, /* dev power */
2354 ddi_quiesce_not_needed
, /* quiesce */
2358 * Module linkage information for the kernel.
2360 static struct modldrv modldrv
= {
2361 &mod_driverops
, /* module type (this is a pseudo driver) */
2362 "Fasttrap Tracing", /* name of module */
2363 &fasttrap_ops
, /* driver ops */
2366 static struct modlinkage modlinkage
= {
2375 return (mod_install(&modlinkage
));
2379 _info(struct modinfo
*modinfop
)
2381 return (mod_info(&modlinkage
, modinfop
));
2387 return (mod_remove(&modlinkage
));