2 * Read-Copy Update tracing for hierarchical implementation.
4 * This program is free software; you can redistribute it and/or modify
5 * it under the terms of the GNU General Public License as published by
6 * the Free Software Foundation; either version 2 of the License, or
7 * (at your option) any later version.
9 * This program is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
14 * You should have received a copy of the GNU General Public License
15 * along with this program; if not, you can access it online at
16 * http://www.gnu.org/licenses/gpl-2.0.html.
18 * Copyright IBM Corporation, 2008
19 * Author: Paul E. McKenney
21 * Papers: http://www.rdrop.com/users/paulmck/RCU
23 * For detailed explanation of Read-Copy Update mechanism see -
27 #include <linux/types.h>
28 #include <linux/kernel.h>
29 #include <linux/init.h>
30 #include <linux/spinlock.h>
31 #include <linux/smp.h>
32 #include <linux/rcupdate.h>
33 #include <linux/interrupt.h>
34 #include <linux/sched.h>
35 #include <linux/atomic.h>
36 #include <linux/bitops.h>
37 #include <linux/completion.h>
38 #include <linux/percpu.h>
39 #include <linux/notifier.h>
40 #include <linux/cpu.h>
41 #include <linux/mutex.h>
42 #include <linux/debugfs.h>
43 #include <linux/seq_file.h>
45 #define RCU_TREE_NONCORE
48 DECLARE_PER_CPU_SHARED_ALIGNED(unsigned long, rcu_qs_ctr
);
50 static int r_open(struct inode
*inode
, struct file
*file
,
51 const struct seq_operations
*op
)
53 int ret
= seq_open(file
, op
);
55 struct seq_file
*m
= (struct seq_file
*)file
->private_data
;
56 m
->private = inode
->i_private
;
61 static void *r_start(struct seq_file
*m
, loff_t
*pos
)
63 struct rcu_state
*rsp
= (struct rcu_state
*)m
->private;
64 *pos
= cpumask_next(*pos
- 1, cpu_possible_mask
);
65 if ((*pos
) < nr_cpu_ids
)
66 return per_cpu_ptr(rsp
->rda
, *pos
);
70 static void *r_next(struct seq_file
*m
, void *v
, loff_t
*pos
)
73 return r_start(m
, pos
);
76 static void r_stop(struct seq_file
*m
, void *v
)
80 static int show_rcubarrier(struct seq_file
*m
, void *v
)
82 struct rcu_state
*rsp
= (struct rcu_state
*)m
->private;
83 seq_printf(m
, "bcc: %d bseq: %lu\n",
84 atomic_read(&rsp
->barrier_cpu_count
),
85 rsp
->barrier_sequence
);
89 static int rcubarrier_open(struct inode
*inode
, struct file
*file
)
91 return single_open(file
, show_rcubarrier
, inode
->i_private
);
94 static const struct file_operations rcubarrier_fops
= {
96 .open
= rcubarrier_open
,
99 .release
= single_release
,
102 #ifdef CONFIG_RCU_BOOST
104 static char convert_kthread_status(unsigned int kthread_status
)
106 if (kthread_status
> RCU_KTHREAD_MAX
)
108 return "SRWOY"[kthread_status
];
111 #endif /* #ifdef CONFIG_RCU_BOOST */
113 static void print_one_rcu_data(struct seq_file
*m
, struct rcu_data
*rdp
)
117 if (!rdp
->beenonline
)
119 seq_printf(m
, "%3d%cc=%ld g=%ld cnq=%d/%d:%d",
121 cpu_is_offline(rdp
->cpu
) ? '!' : ' ',
122 ulong2long(rdp
->completed
), ulong2long(rdp
->gpnum
),
123 rdp
->cpu_no_qs
.b
.norm
,
124 rdp
->rcu_qs_ctr_snap
== per_cpu(rcu_qs_ctr
, rdp
->cpu
),
126 seq_printf(m
, " dt=%d/%llx/%d df=%lu",
127 atomic_read(&rdp
->dynticks
->dynticks
),
128 rdp
->dynticks
->dynticks_nesting
,
129 rdp
->dynticks
->dynticks_nmi_nesting
,
131 seq_printf(m
, " of=%lu", rdp
->offline_fqs
);
132 rcu_nocb_q_lengths(rdp
, &ql
, &qll
);
133 qll
+= rdp
->qlen_lazy
;
135 seq_printf(m
, " ql=%ld/%ld qs=%c%c%c%c",
137 ".N"[rdp
->nxttail
[RCU_NEXT_READY_TAIL
] !=
138 rdp
->nxttail
[RCU_NEXT_TAIL
]],
139 ".R"[rdp
->nxttail
[RCU_WAIT_TAIL
] !=
140 rdp
->nxttail
[RCU_NEXT_READY_TAIL
]],
141 ".W"[rdp
->nxttail
[RCU_DONE_TAIL
] !=
142 rdp
->nxttail
[RCU_WAIT_TAIL
]],
143 ".D"[&rdp
->nxtlist
!= rdp
->nxttail
[RCU_DONE_TAIL
]]);
144 #ifdef CONFIG_RCU_BOOST
145 seq_printf(m
, " kt=%d/%c ktl=%x",
146 per_cpu(rcu_cpu_has_work
, rdp
->cpu
),
147 convert_kthread_status(per_cpu(rcu_cpu_kthread_status
,
149 per_cpu(rcu_cpu_kthread_loops
, rdp
->cpu
) & 0xffff);
150 #endif /* #ifdef CONFIG_RCU_BOOST */
151 seq_printf(m
, " b=%ld", rdp
->blimit
);
152 seq_printf(m
, " ci=%lu nci=%lu co=%lu ca=%lu\n",
153 rdp
->n_cbs_invoked
, rdp
->n_nocbs_invoked
,
154 rdp
->n_cbs_orphaned
, rdp
->n_cbs_adopted
);
157 static int show_rcudata(struct seq_file
*m
, void *v
)
159 print_one_rcu_data(m
, (struct rcu_data
*)v
);
163 static const struct seq_operations rcudate_op
= {
167 .show
= show_rcudata
,
170 static int rcudata_open(struct inode
*inode
, struct file
*file
)
172 return r_open(inode
, file
, &rcudate_op
);
175 static const struct file_operations rcudata_fops
= {
176 .owner
= THIS_MODULE
,
177 .open
= rcudata_open
,
180 .release
= seq_release
,
183 static int show_rcuexp(struct seq_file
*m
, void *v
)
186 struct rcu_state
*rsp
= (struct rcu_state
*)m
->private;
187 struct rcu_data
*rdp
;
188 unsigned long s0
= 0, s1
= 0, s2
= 0, s3
= 0;
190 for_each_possible_cpu(cpu
) {
191 rdp
= per_cpu_ptr(rsp
->rda
, cpu
);
192 s0
+= atomic_long_read(&rdp
->expedited_workdone0
);
193 s1
+= atomic_long_read(&rdp
->expedited_workdone1
);
194 s2
+= atomic_long_read(&rdp
->expedited_workdone2
);
195 s3
+= atomic_long_read(&rdp
->expedited_workdone3
);
197 seq_printf(m
, "s=%lu wd0=%lu wd1=%lu wd2=%lu wd3=%lu n=%lu enq=%d sc=%lu\n",
198 rsp
->expedited_sequence
, s0
, s1
, s2
, s3
,
199 atomic_long_read(&rsp
->expedited_normal
),
200 atomic_read(&rsp
->expedited_need_qs
),
201 rsp
->expedited_sequence
/ 2);
205 static int rcuexp_open(struct inode
*inode
, struct file
*file
)
207 return single_open(file
, show_rcuexp
, inode
->i_private
);
210 static const struct file_operations rcuexp_fops
= {
211 .owner
= THIS_MODULE
,
215 .release
= single_release
,
218 #ifdef CONFIG_RCU_BOOST
220 static void print_one_rcu_node_boost(struct seq_file
*m
, struct rcu_node
*rnp
)
222 seq_printf(m
, "%d:%d tasks=%c%c%c%c kt=%c ntb=%lu neb=%lu nnb=%lu ",
223 rnp
->grplo
, rnp
->grphi
,
224 "T."[list_empty(&rnp
->blkd_tasks
)],
225 "N."[!rnp
->gp_tasks
],
226 "E."[!rnp
->exp_tasks
],
227 "B."[!rnp
->boost_tasks
],
228 convert_kthread_status(rnp
->boost_kthread_status
),
229 rnp
->n_tasks_boosted
, rnp
->n_exp_boosts
,
230 rnp
->n_normal_boosts
);
231 seq_printf(m
, "j=%04x bt=%04x\n",
232 (int)(jiffies
& 0xffff),
233 (int)(rnp
->boost_time
& 0xffff));
234 seq_printf(m
, " balk: nt=%lu egt=%lu bt=%lu nb=%lu ny=%lu nos=%lu\n",
235 rnp
->n_balk_blkd_tasks
,
236 rnp
->n_balk_exp_gp_tasks
,
237 rnp
->n_balk_boost_tasks
,
238 rnp
->n_balk_notblocked
,
243 static int show_rcu_node_boost(struct seq_file
*m
, void *unused
)
245 struct rcu_node
*rnp
;
247 rcu_for_each_leaf_node(&rcu_preempt_state
, rnp
)
248 print_one_rcu_node_boost(m
, rnp
);
252 static int rcu_node_boost_open(struct inode
*inode
, struct file
*file
)
254 return single_open(file
, show_rcu_node_boost
, NULL
);
257 static const struct file_operations rcu_node_boost_fops
= {
258 .owner
= THIS_MODULE
,
259 .open
= rcu_node_boost_open
,
262 .release
= single_release
,
265 #endif /* #ifdef CONFIG_RCU_BOOST */
267 static void print_one_rcu_state(struct seq_file
*m
, struct rcu_state
*rsp
)
271 struct rcu_node
*rnp
;
274 seq_printf(m
, "c=%ld g=%ld s=%d jfq=%ld j=%x ",
275 ulong2long(rsp
->completed
), ulong2long(gpnum
),
277 (long)(rsp
->jiffies_force_qs
- jiffies
),
278 (int)(jiffies
& 0xffff));
279 seq_printf(m
, "nfqs=%lu/nfqsng=%lu(%lu) fqlh=%lu oqlen=%ld/%ld\n",
280 rsp
->n_force_qs
, rsp
->n_force_qs_ngp
,
281 rsp
->n_force_qs
- rsp
->n_force_qs_ngp
,
282 READ_ONCE(rsp
->n_force_qs_lh
), rsp
->qlen_lazy
, rsp
->qlen
);
283 for (rnp
= &rsp
->node
[0]; rnp
- &rsp
->node
[0] < rcu_num_nodes
; rnp
++) {
284 if (rnp
->level
!= level
) {
288 seq_printf(m
, "%lx/%lx->%lx %c%c>%c %d:%d ^%d ",
289 rnp
->qsmask
, rnp
->qsmaskinit
, rnp
->qsmaskinitnext
,
290 ".G"[rnp
->gp_tasks
!= NULL
],
291 ".E"[rnp
->exp_tasks
!= NULL
],
292 ".T"[!list_empty(&rnp
->blkd_tasks
)],
293 rnp
->grplo
, rnp
->grphi
, rnp
->grpnum
);
298 static int show_rcuhier(struct seq_file
*m
, void *v
)
300 struct rcu_state
*rsp
= (struct rcu_state
*)m
->private;
301 print_one_rcu_state(m
, rsp
);
305 static int rcuhier_open(struct inode
*inode
, struct file
*file
)
307 return single_open(file
, show_rcuhier
, inode
->i_private
);
310 static const struct file_operations rcuhier_fops
= {
311 .owner
= THIS_MODULE
,
312 .open
= rcuhier_open
,
315 .release
= single_release
,
318 static void show_one_rcugp(struct seq_file
*m
, struct rcu_state
*rsp
)
321 unsigned long completed
;
325 struct rcu_node
*rnp
= &rsp
->node
[0];
327 raw_spin_lock_irqsave_rcu_node(rnp
, flags
);
328 completed
= READ_ONCE(rsp
->completed
);
329 gpnum
= READ_ONCE(rsp
->gpnum
);
330 if (completed
== gpnum
)
333 gpage
= jiffies
- rsp
->gp_start
;
335 raw_spin_unlock_irqrestore(&rnp
->lock
, flags
);
336 seq_printf(m
, "completed=%ld gpnum=%ld age=%ld max=%ld\n",
337 ulong2long(completed
), ulong2long(gpnum
), gpage
, gpmax
);
340 static int show_rcugp(struct seq_file
*m
, void *v
)
342 struct rcu_state
*rsp
= (struct rcu_state
*)m
->private;
343 show_one_rcugp(m
, rsp
);
347 static int rcugp_open(struct inode
*inode
, struct file
*file
)
349 return single_open(file
, show_rcugp
, inode
->i_private
);
352 static const struct file_operations rcugp_fops
= {
353 .owner
= THIS_MODULE
,
357 .release
= single_release
,
360 static void print_one_rcu_pending(struct seq_file
*m
, struct rcu_data
*rdp
)
362 if (!rdp
->beenonline
)
364 seq_printf(m
, "%3d%cnp=%ld ",
366 cpu_is_offline(rdp
->cpu
) ? '!' : ' ',
368 seq_printf(m
, "qsp=%ld rpq=%ld cbr=%ld cng=%ld ",
369 rdp
->n_rp_core_needs_qs
,
372 rdp
->n_rp_cpu_needs_gp
);
373 seq_printf(m
, "gpc=%ld gps=%ld nn=%ld ndw%ld\n",
374 rdp
->n_rp_gp_completed
,
375 rdp
->n_rp_gp_started
,
376 rdp
->n_rp_nocb_defer_wakeup
,
377 rdp
->n_rp_need_nothing
);
380 static int show_rcu_pending(struct seq_file
*m
, void *v
)
382 print_one_rcu_pending(m
, (struct rcu_data
*)v
);
386 static const struct seq_operations rcu_pending_op
= {
390 .show
= show_rcu_pending
,
393 static int rcu_pending_open(struct inode
*inode
, struct file
*file
)
395 return r_open(inode
, file
, &rcu_pending_op
);
398 static const struct file_operations rcu_pending_fops
= {
399 .owner
= THIS_MODULE
,
400 .open
= rcu_pending_open
,
403 .release
= seq_release
,
406 static int show_rcutorture(struct seq_file
*m
, void *unused
)
408 seq_printf(m
, "rcutorture test sequence: %lu %s\n",
409 rcutorture_testseq
>> 1,
410 (rcutorture_testseq
& 0x1) ? "(test in progress)" : "");
411 seq_printf(m
, "rcutorture update version number: %lu\n",
416 static int rcutorture_open(struct inode
*inode
, struct file
*file
)
418 return single_open(file
, show_rcutorture
, NULL
);
421 static const struct file_operations rcutorture_fops
= {
422 .owner
= THIS_MODULE
,
423 .open
= rcutorture_open
,
426 .release
= single_release
,
429 static struct dentry
*rcudir
;
431 static int __init
rcutree_trace_init(void)
433 struct rcu_state
*rsp
;
434 struct dentry
*retval
;
435 struct dentry
*rspdir
;
437 rcudir
= debugfs_create_dir("rcu", NULL
);
441 for_each_rcu_flavor(rsp
) {
442 rspdir
= debugfs_create_dir(rsp
->name
, rcudir
);
446 retval
= debugfs_create_file("rcudata", 0444,
447 rspdir
, rsp
, &rcudata_fops
);
451 retval
= debugfs_create_file("rcuexp", 0444,
452 rspdir
, rsp
, &rcuexp_fops
);
456 retval
= debugfs_create_file("rcu_pending", 0444,
457 rspdir
, rsp
, &rcu_pending_fops
);
461 retval
= debugfs_create_file("rcubarrier", 0444,
462 rspdir
, rsp
, &rcubarrier_fops
);
466 #ifdef CONFIG_RCU_BOOST
467 if (rsp
== &rcu_preempt_state
) {
468 retval
= debugfs_create_file("rcuboost", 0444,
469 rspdir
, NULL
, &rcu_node_boost_fops
);
475 retval
= debugfs_create_file("rcugp", 0444,
476 rspdir
, rsp
, &rcugp_fops
);
480 retval
= debugfs_create_file("rcuhier", 0444,
481 rspdir
, rsp
, &rcuhier_fops
);
486 retval
= debugfs_create_file("rcutorture", 0444, rcudir
,
487 NULL
, &rcutorture_fops
);
492 debugfs_remove_recursive(rcudir
);
495 device_initcall(rcutree_trace_init
);