Merge tag 'for_linus' of git://git.kernel.org/pub/scm/linux/kernel/git/mst/vhost
[cris-mirror.git] / drivers / block / drbd / drbd_proc.c
blob582caeb0de8665267698836105167da159383c81
1 /*
2 drbd_proc.c
4 This file is part of DRBD by Philipp Reisner and Lars Ellenberg.
6 Copyright (C) 2001-2008, LINBIT Information Technologies GmbH.
7 Copyright (C) 1999-2008, Philipp Reisner <philipp.reisner@linbit.com>.
8 Copyright (C) 2002-2008, Lars Ellenberg <lars.ellenberg@linbit.com>.
10 drbd is free software; you can redistribute it and/or modify
11 it under the terms of the GNU General Public License as published by
12 the Free Software Foundation; either version 2, or (at your option)
13 any later version.
15 drbd is distributed in the hope that it will be useful,
16 but WITHOUT ANY WARRANTY; without even the implied warranty of
17 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 GNU General Public License for more details.
20 You should have received a copy of the GNU General Public License
21 along with drbd; see the file COPYING. If not, write to
22 the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
26 #include <linux/module.h>
28 #include <linux/uaccess.h>
29 #include <linux/fs.h>
30 #include <linux/file.h>
31 #include <linux/proc_fs.h>
32 #include <linux/seq_file.h>
33 #include <linux/drbd.h>
34 #include "drbd_int.h"
36 static int drbd_proc_open(struct inode *inode, struct file *file);
37 static int drbd_proc_release(struct inode *inode, struct file *file);
40 struct proc_dir_entry *drbd_proc;
41 const struct file_operations drbd_proc_fops = {
42 .owner = THIS_MODULE,
43 .open = drbd_proc_open,
44 .read = seq_read,
45 .llseek = seq_lseek,
46 .release = drbd_proc_release,
49 static void seq_printf_with_thousands_grouping(struct seq_file *seq, long v)
51 /* v is in kB/sec. We don't expect TiByte/sec yet. */
52 if (unlikely(v >= 1000000)) {
53 /* cool: > GiByte/s */
54 seq_printf(seq, "%ld,", v / 1000000);
55 v %= 1000000;
56 seq_printf(seq, "%03ld,%03ld", v/1000, v % 1000);
57 } else if (likely(v >= 1000))
58 seq_printf(seq, "%ld,%03ld", v/1000, v % 1000);
59 else
60 seq_printf(seq, "%ld", v);
63 static void drbd_get_syncer_progress(struct drbd_device *device,
64 union drbd_dev_state state, unsigned long *rs_total,
65 unsigned long *bits_left, unsigned int *per_mil_done)
67 /* this is to break it at compile time when we change that, in case we
68 * want to support more than (1<<32) bits on a 32bit arch. */
69 typecheck(unsigned long, device->rs_total);
70 *rs_total = device->rs_total;
72 /* note: both rs_total and rs_left are in bits, i.e. in
73 * units of BM_BLOCK_SIZE.
74 * for the percentage, we don't care. */
76 if (state.conn == C_VERIFY_S || state.conn == C_VERIFY_T)
77 *bits_left = device->ov_left;
78 else
79 *bits_left = drbd_bm_total_weight(device) - device->rs_failed;
80 /* >> 10 to prevent overflow,
81 * +1 to prevent division by zero */
82 if (*bits_left > *rs_total) {
83 /* D'oh. Maybe a logic bug somewhere. More likely just a race
84 * between state change and reset of rs_total.
86 *bits_left = *rs_total;
87 *per_mil_done = *rs_total ? 0 : 1000;
88 } else {
89 /* Make sure the division happens in long context.
90 * We allow up to one petabyte storage right now,
91 * at a granularity of 4k per bit that is 2**38 bits.
92 * After shift right and multiplication by 1000,
93 * this should still fit easily into a 32bit long,
94 * so we don't need a 64bit division on 32bit arch.
95 * Note: currently we don't support such large bitmaps on 32bit
96 * arch anyways, but no harm done to be prepared for it here.
98 unsigned int shift = *rs_total > UINT_MAX ? 16 : 10;
99 unsigned long left = *bits_left >> shift;
100 unsigned long total = 1UL + (*rs_total >> shift);
101 unsigned long tmp = 1000UL - left * 1000UL/total;
102 *per_mil_done = tmp;
107 /*lge
108 * progress bars shamelessly adapted from driver/md/md.c
109 * output looks like
110 * [=====>..............] 33.5% (23456/123456)
111 * finish: 2:20:20 speed: 6,345 (6,456) K/sec
113 static void drbd_syncer_progress(struct drbd_device *device, struct seq_file *seq,
114 union drbd_dev_state state)
116 unsigned long db, dt, dbdt, rt, rs_total, rs_left;
117 unsigned int res;
118 int i, x, y;
119 int stalled = 0;
121 drbd_get_syncer_progress(device, state, &rs_total, &rs_left, &res);
123 x = res/50;
124 y = 20-x;
125 seq_puts(seq, "\t[");
126 for (i = 1; i < x; i++)
127 seq_putc(seq, '=');
128 seq_putc(seq, '>');
129 for (i = 0; i < y; i++)
130 seq_putc(seq, '.');
131 seq_puts(seq, "] ");
133 if (state.conn == C_VERIFY_S || state.conn == C_VERIFY_T)
134 seq_puts(seq, "verified:");
135 else
136 seq_puts(seq, "sync'ed:");
137 seq_printf(seq, "%3u.%u%% ", res / 10, res % 10);
139 /* if more than a few GB, display in MB */
140 if (rs_total > (4UL << (30 - BM_BLOCK_SHIFT)))
141 seq_printf(seq, "(%lu/%lu)M",
142 (unsigned long) Bit2KB(rs_left >> 10),
143 (unsigned long) Bit2KB(rs_total >> 10));
144 else
145 seq_printf(seq, "(%lu/%lu)K",
146 (unsigned long) Bit2KB(rs_left),
147 (unsigned long) Bit2KB(rs_total));
149 seq_puts(seq, "\n\t");
151 /* see drivers/md/md.c
152 * We do not want to overflow, so the order of operands and
153 * the * 100 / 100 trick are important. We do a +1 to be
154 * safe against division by zero. We only estimate anyway.
156 * dt: time from mark until now
157 * db: blocks written from mark until now
158 * rt: remaining time
160 /* Rolling marks. last_mark+1 may just now be modified. last_mark+2 is
161 * at least (DRBD_SYNC_MARKS-2)*DRBD_SYNC_MARK_STEP old, and has at
162 * least DRBD_SYNC_MARK_STEP time before it will be modified. */
163 /* ------------------------ ~18s average ------------------------ */
164 i = (device->rs_last_mark + 2) % DRBD_SYNC_MARKS;
165 dt = (jiffies - device->rs_mark_time[i]) / HZ;
166 if (dt > 180)
167 stalled = 1;
169 if (!dt)
170 dt++;
171 db = device->rs_mark_left[i] - rs_left;
172 rt = (dt * (rs_left / (db/100+1)))/100; /* seconds */
174 seq_printf(seq, "finish: %lu:%02lu:%02lu",
175 rt / 3600, (rt % 3600) / 60, rt % 60);
177 dbdt = Bit2KB(db/dt);
178 seq_puts(seq, " speed: ");
179 seq_printf_with_thousands_grouping(seq, dbdt);
180 seq_puts(seq, " (");
181 /* ------------------------- ~3s average ------------------------ */
182 if (drbd_proc_details >= 1) {
183 /* this is what drbd_rs_should_slow_down() uses */
184 i = (device->rs_last_mark + DRBD_SYNC_MARKS-1) % DRBD_SYNC_MARKS;
185 dt = (jiffies - device->rs_mark_time[i]) / HZ;
186 if (!dt)
187 dt++;
188 db = device->rs_mark_left[i] - rs_left;
189 dbdt = Bit2KB(db/dt);
190 seq_printf_with_thousands_grouping(seq, dbdt);
191 seq_puts(seq, " -- ");
194 /* --------------------- long term average ---------------------- */
195 /* mean speed since syncer started
196 * we do account for PausedSync periods */
197 dt = (jiffies - device->rs_start - device->rs_paused) / HZ;
198 if (dt == 0)
199 dt = 1;
200 db = rs_total - rs_left;
201 dbdt = Bit2KB(db/dt);
202 seq_printf_with_thousands_grouping(seq, dbdt);
203 seq_putc(seq, ')');
205 if (state.conn == C_SYNC_TARGET ||
206 state.conn == C_VERIFY_S) {
207 seq_puts(seq, " want: ");
208 seq_printf_with_thousands_grouping(seq, device->c_sync_rate);
210 seq_printf(seq, " K/sec%s\n", stalled ? " (stalled)" : "");
212 if (drbd_proc_details >= 1) {
213 /* 64 bit:
214 * we convert to sectors in the display below. */
215 unsigned long bm_bits = drbd_bm_bits(device);
216 unsigned long bit_pos;
217 unsigned long long stop_sector = 0;
218 if (state.conn == C_VERIFY_S ||
219 state.conn == C_VERIFY_T) {
220 bit_pos = bm_bits - device->ov_left;
221 if (verify_can_do_stop_sector(device))
222 stop_sector = device->ov_stop_sector;
223 } else
224 bit_pos = device->bm_resync_fo;
225 /* Total sectors may be slightly off for oddly
226 * sized devices. So what. */
227 seq_printf(seq,
228 "\t%3d%% sector pos: %llu/%llu",
229 (int)(bit_pos / (bm_bits/100+1)),
230 (unsigned long long)bit_pos * BM_SECT_PER_BIT,
231 (unsigned long long)bm_bits * BM_SECT_PER_BIT);
232 if (stop_sector != 0 && stop_sector != ULLONG_MAX)
233 seq_printf(seq, " stop sector: %llu", stop_sector);
234 seq_putc(seq, '\n');
238 static int drbd_seq_show(struct seq_file *seq, void *v)
240 int i, prev_i = -1;
241 const char *sn;
242 struct drbd_device *device;
243 struct net_conf *nc;
244 union drbd_dev_state state;
245 char wp;
247 static char write_ordering_chars[] = {
248 [WO_NONE] = 'n',
249 [WO_DRAIN_IO] = 'd',
250 [WO_BDEV_FLUSH] = 'f',
253 seq_printf(seq, "version: " REL_VERSION " (api:%d/proto:%d-%d)\n%s\n",
254 API_VERSION, PRO_VERSION_MIN, PRO_VERSION_MAX, drbd_buildtag());
257 cs .. connection state
258 ro .. node role (local/remote)
259 ds .. disk state (local/remote)
260 protocol
261 various flags
262 ns .. network send
263 nr .. network receive
264 dw .. disk write
265 dr .. disk read
266 al .. activity log write count
267 bm .. bitmap update write count
268 pe .. pending (waiting for ack or data reply)
269 ua .. unack'd (still need to send ack or data reply)
270 ap .. application requests accepted, but not yet completed
271 ep .. number of epochs currently "on the fly", P_BARRIER_ACK pending
272 wo .. write ordering mode currently in use
273 oos .. known out-of-sync kB
276 rcu_read_lock();
277 idr_for_each_entry(&drbd_devices, device, i) {
278 if (prev_i != i - 1)
279 seq_putc(seq, '\n');
280 prev_i = i;
282 state = device->state;
283 sn = drbd_conn_str(state.conn);
285 if (state.conn == C_STANDALONE &&
286 state.disk == D_DISKLESS &&
287 state.role == R_SECONDARY) {
288 seq_printf(seq, "%2d: cs:Unconfigured\n", i);
289 } else {
290 /* reset device->congestion_reason */
291 bdi_rw_congested(device->rq_queue->backing_dev_info);
293 nc = rcu_dereference(first_peer_device(device)->connection->net_conf);
294 wp = nc ? nc->wire_protocol - DRBD_PROT_A + 'A' : ' ';
295 seq_printf(seq,
296 "%2d: cs:%s ro:%s/%s ds:%s/%s %c %c%c%c%c%c%c\n"
297 " ns:%u nr:%u dw:%u dr:%u al:%u bm:%u "
298 "lo:%d pe:%d ua:%d ap:%d ep:%d wo:%c",
299 i, sn,
300 drbd_role_str(state.role),
301 drbd_role_str(state.peer),
302 drbd_disk_str(state.disk),
303 drbd_disk_str(state.pdsk),
305 drbd_suspended(device) ? 's' : 'r',
306 state.aftr_isp ? 'a' : '-',
307 state.peer_isp ? 'p' : '-',
308 state.user_isp ? 'u' : '-',
309 device->congestion_reason ?: '-',
310 test_bit(AL_SUSPENDED, &device->flags) ? 's' : '-',
311 device->send_cnt/2,
312 device->recv_cnt/2,
313 device->writ_cnt/2,
314 device->read_cnt/2,
315 device->al_writ_cnt,
316 device->bm_writ_cnt,
317 atomic_read(&device->local_cnt),
318 atomic_read(&device->ap_pending_cnt) +
319 atomic_read(&device->rs_pending_cnt),
320 atomic_read(&device->unacked_cnt),
321 atomic_read(&device->ap_bio_cnt),
322 first_peer_device(device)->connection->epochs,
323 write_ordering_chars[device->resource->write_ordering]
325 seq_printf(seq, " oos:%llu\n",
326 Bit2KB((unsigned long long)
327 drbd_bm_total_weight(device)));
329 if (state.conn == C_SYNC_SOURCE ||
330 state.conn == C_SYNC_TARGET ||
331 state.conn == C_VERIFY_S ||
332 state.conn == C_VERIFY_T)
333 drbd_syncer_progress(device, seq, state);
335 if (drbd_proc_details >= 1 && get_ldev_if_state(device, D_FAILED)) {
336 lc_seq_printf_stats(seq, device->resync);
337 lc_seq_printf_stats(seq, device->act_log);
338 put_ldev(device);
341 if (drbd_proc_details >= 2)
342 seq_printf(seq, "\tblocked on activity log: %d\n", atomic_read(&device->ap_actlog_cnt));
344 rcu_read_unlock();
346 return 0;
349 static int drbd_proc_open(struct inode *inode, struct file *file)
351 int err;
353 if (try_module_get(THIS_MODULE)) {
354 err = single_open(file, drbd_seq_show, NULL);
355 if (err)
356 module_put(THIS_MODULE);
357 return err;
359 return -ENODEV;
362 static int drbd_proc_release(struct inode *inode, struct file *file)
364 module_put(THIS_MODULE);
365 return single_release(inode, file);
368 /* PROC FS stuff end */