2 /*--------------------------------------------------------------------*/
3 /*--- Error management for Helgrind. ---*/
4 /*--- hg_errors.c ---*/
5 /*--------------------------------------------------------------------*/
8 This file is part of Helgrind, a Valgrind tool for detecting errors
11 Copyright (C) 2007-2013 OpenWorks Ltd
14 This program is free software; you can redistribute it and/or
15 modify it under the terms of the GNU General Public License as
16 published by the Free Software Foundation; either version 2 of the
17 License, or (at your option) any later version.
19 This program is distributed in the hope that it will be useful, but
20 WITHOUT ANY WARRANTY; without even the implied warranty of
21 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
22 General Public License for more details.
24 You should have received a copy of the GNU General Public License
25 along with this program; if not, write to the Free Software
26 Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
29 The GNU General Public License is contained in the file COPYING.
32 #include "pub_tool_basics.h"
33 #include "pub_tool_libcbase.h"
34 #include "pub_tool_libcassert.h"
35 #include "pub_tool_libcprint.h"
36 #include "pub_tool_stacktrace.h"
37 #include "pub_tool_execontext.h"
38 #include "pub_tool_errormgr.h"
39 #include "pub_tool_wordfm.h"
40 #include "pub_tool_xarray.h"
41 #include "pub_tool_debuginfo.h"
42 #include "pub_tool_threadstate.h"
43 #include "pub_tool_options.h" // VG_(clo_xml)
44 #include "pub_tool_aspacemgr.h"
45 #include "pub_tool_addrinfo.h"
47 #include "hg_basics.h"
48 #include "hg_addrdescr.h"
49 #include "hg_wordset.h"
50 #include "hg_lock_n_thread.h"
52 #include "hg_errors.h" /* self */
55 /*----------------------------------------------------------------*/
56 /*--- Error management -- storage ---*/
57 /*----------------------------------------------------------------*/
59 /* maps (by value) strings to a copy of them in ARENA_TOOL */
61 static WordFM
* string_table
= NULL
;
63 ULong
HG_(stats__string_table_queries
) = 0;
65 ULong
HG_(stats__string_table_get_map_size
) ( void ) {
66 return string_table
? (ULong
)VG_(sizeFM
)(string_table
) : 0;
69 static Word
string_table_cmp ( UWord s1
, UWord s2
) {
70 return (Word
)VG_(strcmp
)( (HChar
*)s1
, (HChar
*)s2
);
73 static HChar
* string_table_strdup ( const HChar
* str
) {
75 HG_(stats__string_table_queries
)++;
79 string_table
= VG_(newFM
)( HG_(zalloc
), "hg.sts.1",
80 HG_(free
), string_table_cmp
);
82 if (VG_(lookupFM
)( string_table
,
83 NULL
, (UWord
*)©
, (UWord
)str
)) {
85 if (0) VG_(printf
)("string_table_strdup: %p -> %p\n", str
, copy
);
88 copy
= HG_(strdup
)("hg.sts.2", str
);
89 VG_(addToFM
)( string_table
, (UWord
)copy
, (UWord
)copy
);
94 /* maps from Lock .unique fields to LockP*s */
96 static WordFM
* map_LockN_to_P
= NULL
;
98 ULong
HG_(stats__LockN_to_P_queries
) = 0;
100 ULong
HG_(stats__LockN_to_P_get_map_size
) ( void ) {
101 return map_LockN_to_P
? (ULong
)VG_(sizeFM
)(map_LockN_to_P
) : 0;
104 static Word
lock_unique_cmp ( UWord lk1W
, UWord lk2W
)
106 Lock
* lk1
= (Lock
*)lk1W
;
107 Lock
* lk2
= (Lock
*)lk2W
;
108 tl_assert( HG_(is_sane_LockNorP
)(lk1
) );
109 tl_assert( HG_(is_sane_LockNorP
)(lk2
) );
110 if (lk1
->unique
< lk2
->unique
) return -1;
111 if (lk1
->unique
> lk2
->unique
) return 1;
115 /* Given a normal Lock (LockN), convert it to a persistent Lock
116 (LockP). In some cases the LockN could be invalid (if it's been
117 freed), so we enquire, in hg_main.c's admin_locks list, whether it
118 is in fact valid. If allowed_to_be_invalid is True, then it's OK
119 for the LockN to be invalid, in which case Lock_INVALID is
120 returned. In all other cases, we insist that the LockN is a valid
121 lock, and return its corresponding LockP.
123 Why can LockNs sometimes be invalid? Because they are harvested
124 from locksets that are attached to the OldRef info for conflicting
125 threads. By the time we detect a race, the some of the elements of
126 the lockset may have been destroyed by the client, in which case
127 the corresponding Lock structures we maintain will have been freed.
129 So we check that each LockN is a member of the admin_locks double
130 linked list of all Lock structures. That stops us prodding around
131 in potentially freed-up Lock structures. However, it's not quite a
132 proper check: if a new Lock has been reallocated at the same
133 address as one which was previously freed, we'll wind up copying
134 the new one as the basis for the LockP, which is completely bogus
135 because it is unrelated to the previous Lock that lived there.
136 Let's hope that doesn't happen too often.
138 static Lock
* mk_LockP_from_LockN ( Lock
* lkn
,
139 Bool allowed_to_be_invalid
)
142 HG_(stats__LockN_to_P_queries
)++;
144 /* First off, let's do some sanity checks. If
145 allowed_to_be_invalid is False, we _must_ be able to find 'lkn'
146 in admin_locks; else we must assert. If it is True, it's OK for
147 it not to be findable, but in that case we must return
148 Lock_INVALID right away. */
149 Lock
* lock_list
= HG_(get_admin_locks
)();
151 if (lock_list
== lkn
)
153 lock_list
= lock_list
->admin_next
;
155 if (lock_list
== NULL
) {
156 /* We didn't find it. That possibility has to be OK'd by the
158 tl_assert(allowed_to_be_invalid
);
162 /* So we must be looking at a valid LockN. */
163 tl_assert( HG_(is_sane_LockN
)(lkn
) );
165 if (!map_LockN_to_P
) {
166 map_LockN_to_P
= VG_(newFM
)( HG_(zalloc
), "hg.mLPfLN.1",
167 HG_(free
), lock_unique_cmp
);
169 if (!VG_(lookupFM
)( map_LockN_to_P
, NULL
, (UWord
*)&lkp
, (UWord
)lkn
)) {
170 lkp
= HG_(zalloc
)( "hg.mLPfLN.2", sizeof(Lock
) );
172 lkp
->admin_next
= NULL
;
173 lkp
->admin_prev
= NULL
;
174 lkp
->magic
= LockP_MAGIC
;
175 /* Forget about the bag of lock holders - don't copy that.
176 Also, acquired_at should be NULL whenever heldBy is, and vice
177 versa. Also forget about the associated libhb synch object. */
180 lkp
->acquired_at
= NULL
;
182 VG_(addToFM
)( map_LockN_to_P
, (UWord
)lkp
, (UWord
)lkp
);
184 tl_assert( HG_(is_sane_LockP
)(lkp
) );
188 static Int
sort_by_guestaddr(const void* n1
, const void* n2
)
190 const Lock
* l1
= *(const Lock
*const *)n1
;
191 const Lock
* l2
= *(const Lock
*const *)n2
;
193 Addr a1
= l1
== Lock_INVALID
? 0 : l1
->guestaddr
;
194 Addr a2
= l2
== Lock_INVALID
? 0 : l2
->guestaddr
;
195 if (a1
< a2
) return -1;
196 if (a1
> a2
) return 1;
200 /* Expand a WordSet of LockN*'s into a NULL-terminated vector of
201 LockP*'s. Any LockN's that can't be converted into a LockP
202 (because they have been freed, see comment on mk_LockP_from_LockN)
203 are converted instead into the value Lock_INVALID. Hence the
204 returned vector is a sequence: zero or more (valid LockP* or
205 LockN_INVALID), terminated by a NULL. */
207 Lock
** enumerate_WordSet_into_LockP_vector( WordSetU
* univ_lsets
,
209 Bool allowed_to_be_invalid
)
211 tl_assert(univ_lsets
);
212 tl_assert( HG_(plausibleWS
)(univ_lsets
, lockset
) );
213 UWord nLocks
= HG_(cardinalityWS
)(univ_lsets
, lockset
);
214 Lock
** lockPs
= HG_(zalloc
)( "hg.eWSiLPa",
215 (nLocks
+1) * sizeof(Lock
*) );
216 tl_assert(lockPs
[nLocks
] == NULL
); /* pre-NULL terminated */
217 UWord
* lockNs
= NULL
;
220 /* HG_(getPayloadWS) doesn't assign non-NULL to &lockNs if the
221 lockset is empty; hence the guarding "if". Sigh. */
222 HG_(getPayloadWS
)( &lockNs
, &nLockNs
, univ_lsets
, lockset
);
226 /* Convert to LockPs. */
227 for (i
= 0; i
< nLockNs
; i
++) {
228 lockPs
[i
] = mk_LockP_from_LockN( (Lock
*)lockNs
[i
],
229 allowed_to_be_invalid
);
231 /* Sort the locks by increasing Lock::guestaddr to avoid jitters
233 VG_(ssort
)(lockPs
, nLockNs
, sizeof lockPs
[0], sort_by_guestaddr
);
238 /* Get the number of useful elements in a vector created by
239 enumerate_WordSet_into_LockP_vector. Returns both the total number
240 of elements (not including the terminating NULL) and the number of
241 non-Lock_INVALID elements. */
242 static void count_LockP_vector ( /*OUT*/UWord
* nLocks
,
243 /*OUT*/UWord
* nLocksValid
,
247 *nLocks
= *nLocksValid
= 0;
251 if (vec
[n
] != Lock_INVALID
)
257 /* Find out whether 'lk' is in 'vec'. */
258 static Bool
elem_LockP_vector ( Lock
** vec
, Lock
* lk
)
274 race: program counter
280 FIXME: how does state printing interact with lockset gc?
281 Are the locksets in prev/curr state always valid?
282 Ditto question for the threadsets
283 ThreadSets - probably are always valid if Threads
284 are never thrown away.
285 LockSets - could at least print the lockset elements that
286 correspond to actual locks at the time of printing. Hmm.
292 XE_Race
=1101, // race
293 XE_UnlockUnlocked
, // unlocking a not-locked lock
294 XE_UnlockForeign
, // unlocking a lock held by some other thread
295 XE_UnlockBogus
, // unlocking an address not known to be a lock
296 XE_PthAPIerror
, // error from the POSIX pthreads API
297 XE_LockOrder
, // lock order error
298 XE_Misc
// misc other error (w/ string to describe it)
302 /* Extra contexts for kinds */
310 AddrInfo data_addrinfo
;
314 /* h1_* and h2_* provide some description of a previously
315 observed access with which we are conflicting. */
316 Thread
* h1_ct
; /* non-NULL means h1 info present */
317 ExeContext
* h1_ct_mbsegstartEC
;
318 ExeContext
* h1_ct_mbsegendEC
;
319 Thread
* h2_ct
; /* non-NULL means h2 info present */
320 ExeContext
* h2_ct_accEC
;
323 Lock
** h2_ct_locksHeldW
;
326 Thread
* thr
; /* doing the unlocking */
327 Lock
* lock
; /* lock (that is already unlocked) */
330 Thread
* thr
; /* doing the unlocking */
331 Thread
* owner
; /* thread that actually holds the lock */
332 Lock
* lock
; /* lock (that is held by 'owner') */
335 Thread
* thr
; /* doing the unlocking */
336 Addr lock_ga
; /* purported address of the lock */
340 HChar
* fnname
; /* persistent, in tool-arena */
341 Word err
; /* pth error code */
342 HChar
* errstr
; /* persistent, in tool-arena */
346 /* The first 4 fields describe the previously observed
347 (should-be) ordering. */
348 Lock
* shouldbe_earlier_lk
;
349 Lock
* shouldbe_later_lk
;
350 ExeContext
* shouldbe_earlier_ec
;
351 ExeContext
* shouldbe_later_ec
;
352 /* In principle we need to record two more stacks, from
353 this thread, when acquiring the locks in the "wrong"
354 order. In fact the wallclock-later acquisition by this
355 thread is recorded in the main stack for this error.
356 So we only need a stack for the earlier acquisition by
358 ExeContext
* actual_earlier_ec
;
362 HChar
* errstr
; /* persistent, in tool-arena */
363 HChar
* auxstr
; /* optional, persistent, in tool-arena */
364 ExeContext
* auxctx
; /* optional */
370 static void init_XError ( XError
* xe
) {
371 VG_(memset
)(xe
, 0, sizeof(*xe
) );
372 xe
->tag
= XE_Race
-1; /* bogus */
376 /* Extensions of suppressions */
379 XS_Race
=1201, /* race */
391 /* Updates the copy with address info if necessary. */
392 UInt
HG_(update_extra
) ( const Error
* err
)
394 XError
* xe
= (XError
*)VG_(get_error_extra
)(err
);
396 //if (extra != NULL && Undescribed == extra->addrinfo.akind) {
397 // describe_addr ( VG_(get_error_address)(err), &(extra->addrinfo) );
400 if (xe
->tag
== XE_Race
) {
402 /* Note the set of locks that the thread is (w-)holding.
403 Convert the WordSetID of LockN*'s into a NULL-terminated
404 vector of LockP*'s. We don't expect to encounter any invalid
405 LockNs in this conversion. */
406 tl_assert(xe
->XE
.Race
.thr
);
407 xe
->XE
.Race
.locksHeldW
408 = enumerate_WordSet_into_LockP_vector(
409 HG_(get_univ_lsets
)(),
410 xe
->XE
.Race
.thr
->locksetW
,
411 False
/*!allowed_to_be_invalid*/
414 /* See if we can come up with a source level description of the
415 raced-upon address. This is potentially expensive, which is
416 why it's only done at the update_extra point, not when the
417 error is initially created. */
421 VG_(printf
)("HG_(update_extra): "
422 "%d conflicting-event queries\n", xxx
);
424 HG_(describe_addr
) (xe
->XE
.Race
.data_addr
, &xe
->XE
.Race
.data_addrinfo
);
426 /* And poke around in the conflicting-event map, to see if we
427 can rustle up a plausible-looking conflicting memory access
429 if (HG_(clo_history_level
) >= 2) {
431 ExeContext
* wherep
= NULL
;
432 Addr acc_addr
= xe
->XE
.Race
.data_addr
;
433 Int acc_szB
= xe
->XE
.Race
.szB
;
434 Thr
* acc_thr
= xe
->XE
.Race
.thr
->hbthr
;
435 Bool acc_isW
= xe
->XE
.Race
.isWrite
;
437 Bool conf_isW
= False
;
438 WordSetID conf_locksHeldW
= 0;
439 tl_assert(!xe
->XE
.Race
.h2_ct_accEC
);
440 tl_assert(!xe
->XE
.Race
.h2_ct
);
441 if (libhb_event_map_lookup(
442 &wherep
, &thrp
, &conf_szB
, &conf_isW
, &conf_locksHeldW
,
443 acc_thr
, acc_addr
, acc_szB
, acc_isW
)) {
447 threadp
= libhb_get_Thr_hgthread( thrp
);
449 xe
->XE
.Race
.h2_ct_accEC
= wherep
;
450 xe
->XE
.Race
.h2_ct
= threadp
;
451 xe
->XE
.Race
.h2_ct_accSzB
= (Int
)conf_szB
;
452 xe
->XE
.Race
.h2_ct_accIsW
= conf_isW
;
453 xe
->XE
.Race
.h2_ct_locksHeldW
454 = enumerate_WordSet_into_LockP_vector(
455 HG_(get_univ_lsets
)(),
457 True
/*allowed_to_be_invalid*/
462 // both NULL or both non-NULL
463 tl_assert( (!!xe
->XE
.Race
.h2_ct
) == (!!xe
->XE
.Race
.h2_ct_accEC
) );
466 return sizeof(XError
);
469 void HG_(record_error_Race
) ( Thread
* thr
,
470 Addr data_addr
, Int szB
, Bool isWrite
,
472 ExeContext
* h1_ct_segstart
,
473 ExeContext
* h1_ct_mbsegendEC
)
476 tl_assert( HG_(is_sane_Thread
)(thr
) );
478 # if defined(VGO_linux)
479 /* Skip any races on locations apparently in GOTPLT sections. This
480 is said to be caused by ld.so poking PLT table entries (or
481 whatever) when it writes the resolved address of a dynamically
482 linked routine, into the table (or whatever) when it is called
483 for the first time. */
485 VgSectKind sect
= VG_(DebugInfo_sect_kind
)( NULL
, data_addr
);
486 if (0) VG_(printf
)("XXXXXXXXX RACE on %#lx %s\n",
487 data_addr
, VG_(pp_SectKind
)(sect
));
488 /* SectPLT is required on ???-linux */
489 if (sect
== Vg_SectGOTPLT
) return;
490 /* SectPLT is required on ppc32/64-linux */
491 if (sect
== Vg_SectPLT
) return;
492 /* SectGOT is required on arm-linux */
493 if (sect
== Vg_SectGOT
) return;
499 xe
.XE
.Race
.data_addr
= data_addr
;
500 xe
.XE
.Race
.szB
= szB
;
501 xe
.XE
.Race
.isWrite
= isWrite
;
502 xe
.XE
.Race
.thr
= thr
;
503 tl_assert(isWrite
== False
|| isWrite
== True
);
504 tl_assert(szB
== 8 || szB
== 4 || szB
== 2 || szB
== 1);
505 /* Skip on the detailed description of the raced-on address at this
506 point; it's expensive. Leave it for the update_extra function
507 if we ever make it that far. */
508 xe
.XE
.Race
.data_addrinfo
.tag
= Addr_Undescribed
;
510 // Skip on any of the conflicting-access info at this point.
511 // It's expensive to obtain, and this error is more likely than
512 // not to be discarded. We'll fill these fields in in
513 // HG_(update_extra) just above, assuming the error ever makes
514 // it that far (unlikely).
515 xe
.XE
.Race
.h2_ct_accSzB
= 0;
516 xe
.XE
.Race
.h2_ct_accIsW
= False
;
517 xe
.XE
.Race
.h2_ct_accEC
= NULL
;
518 xe
.XE
.Race
.h2_ct
= NULL
;
519 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
520 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
522 xe
.XE
.Race
.h1_ct
= h1_ct
;
523 xe
.XE
.Race
.h1_ct_mbsegstartEC
= h1_ct_segstart
;
524 xe
.XE
.Race
.h1_ct_mbsegendEC
= h1_ct_mbsegendEC
;
526 VG_(maybe_record_error
)( thr
->coretid
,
527 XE_Race
, data_addr
, NULL
, &xe
);
530 void HG_(record_error_UnlockUnlocked
) ( Thread
* thr
, Lock
* lk
)
533 tl_assert( HG_(is_sane_Thread
)(thr
) );
534 tl_assert( HG_(is_sane_LockN
)(lk
) );
536 xe
.tag
= XE_UnlockUnlocked
;
537 xe
.XE
.UnlockUnlocked
.thr
539 xe
.XE
.UnlockUnlocked
.lock
540 = mk_LockP_from_LockN(lk
, False
/*!allowed_to_be_invalid*/);
542 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
543 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
544 VG_(maybe_record_error
)( thr
->coretid
,
545 XE_UnlockUnlocked
, 0, NULL
, &xe
);
548 void HG_(record_error_UnlockForeign
) ( Thread
* thr
,
549 Thread
* owner
, Lock
* lk
)
552 tl_assert( HG_(is_sane_Thread
)(thr
) );
553 tl_assert( HG_(is_sane_Thread
)(owner
) );
554 tl_assert( HG_(is_sane_LockN
)(lk
) );
556 xe
.tag
= XE_UnlockForeign
;
557 xe
.XE
.UnlockForeign
.thr
= thr
;
558 xe
.XE
.UnlockForeign
.owner
= owner
;
559 xe
.XE
.UnlockForeign
.lock
560 = mk_LockP_from_LockN(lk
, False
/*!allowed_to_be_invalid*/);
562 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
563 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
564 VG_(maybe_record_error
)( thr
->coretid
,
565 XE_UnlockForeign
, 0, NULL
, &xe
);
568 void HG_(record_error_UnlockBogus
) ( Thread
* thr
, Addr lock_ga
)
571 tl_assert( HG_(is_sane_Thread
)(thr
) );
573 xe
.tag
= XE_UnlockBogus
;
574 xe
.XE
.UnlockBogus
.thr
= thr
;
575 xe
.XE
.UnlockBogus
.lock_ga
= lock_ga
;
577 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
578 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
579 VG_(maybe_record_error
)( thr
->coretid
,
580 XE_UnlockBogus
, 0, NULL
, &xe
);
583 void HG_(record_error_LockOrder
)(
585 Lock
* shouldbe_earlier_lk
,
586 Lock
* shouldbe_later_lk
,
587 ExeContext
* shouldbe_earlier_ec
,
588 ExeContext
* shouldbe_later_ec
,
589 ExeContext
* actual_earlier_ec
593 tl_assert( HG_(is_sane_Thread
)(thr
) );
594 tl_assert(HG_(clo_track_lockorders
));
596 xe
.tag
= XE_LockOrder
;
597 xe
.XE
.LockOrder
.thr
= thr
;
598 xe
.XE
.LockOrder
.shouldbe_earlier_lk
599 = mk_LockP_from_LockN(shouldbe_earlier_lk
,
600 False
/*!allowed_to_be_invalid*/);
601 xe
.XE
.LockOrder
.shouldbe_earlier_ec
= shouldbe_earlier_ec
;
602 xe
.XE
.LockOrder
.shouldbe_later_lk
603 = mk_LockP_from_LockN(shouldbe_later_lk
,
604 False
/*!allowed_to_be_invalid*/);
605 xe
.XE
.LockOrder
.shouldbe_later_ec
= shouldbe_later_ec
;
606 xe
.XE
.LockOrder
.actual_earlier_ec
= actual_earlier_ec
;
608 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
609 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
610 VG_(maybe_record_error
)( thr
->coretid
,
611 XE_LockOrder
, 0, NULL
, &xe
);
614 void HG_(record_error_PthAPIerror
) ( Thread
* thr
, const HChar
* fnname
,
615 Word err
, const HChar
* errstr
)
618 tl_assert( HG_(is_sane_Thread
)(thr
) );
622 xe
.tag
= XE_PthAPIerror
;
623 xe
.XE
.PthAPIerror
.thr
= thr
;
624 xe
.XE
.PthAPIerror
.fnname
= string_table_strdup(fnname
);
625 xe
.XE
.PthAPIerror
.err
= err
;
626 xe
.XE
.PthAPIerror
.errstr
= string_table_strdup(errstr
);
628 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
629 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
630 VG_(maybe_record_error
)( thr
->coretid
,
631 XE_PthAPIerror
, 0, NULL
, &xe
);
634 void HG_(record_error_Misc_w_aux
) ( Thread
* thr
, const HChar
* errstr
,
635 const HChar
* auxstr
, ExeContext
* auxctx
)
638 tl_assert( HG_(is_sane_Thread
)(thr
) );
642 xe
.XE
.Misc
.thr
= thr
;
643 xe
.XE
.Misc
.errstr
= string_table_strdup(errstr
);
644 xe
.XE
.Misc
.auxstr
= auxstr
? string_table_strdup(auxstr
) : NULL
;
645 xe
.XE
.Misc
.auxctx
= auxctx
;
647 tl_assert( HG_(is_sane_ThreadId
)(thr
->coretid
) );
648 tl_assert( thr
->coretid
!= VG_INVALID_THREADID
);
649 VG_(maybe_record_error
)( thr
->coretid
,
650 XE_Misc
, 0, NULL
, &xe
);
653 void HG_(record_error_Misc
) ( Thread
* thr
, const HChar
* errstr
)
655 HG_(record_error_Misc_w_aux
)(thr
, errstr
, NULL
, NULL
);
658 Bool
HG_(eq_Error
) ( VgRes not_used
, const Error
* e1
, const Error
* e2
)
662 tl_assert(VG_(get_error_kind
)(e1
) == VG_(get_error_kind
)(e2
));
664 xe1
= (XError
*)VG_(get_error_extra
)(e1
);
665 xe2
= (XError
*)VG_(get_error_extra
)(e2
);
669 switch (VG_(get_error_kind
)(e1
)) {
671 return xe1
->XE
.Race
.szB
== xe2
->XE
.Race
.szB
672 && xe1
->XE
.Race
.isWrite
== xe2
->XE
.Race
.isWrite
673 && (HG_(clo_cmp_race_err_addrs
)
674 ? xe1
->XE
.Race
.data_addr
== xe2
->XE
.Race
.data_addr
676 case XE_UnlockUnlocked
:
677 return xe1
->XE
.UnlockUnlocked
.thr
== xe2
->XE
.UnlockUnlocked
.thr
678 && xe1
->XE
.UnlockUnlocked
.lock
== xe2
->XE
.UnlockUnlocked
.lock
;
679 case XE_UnlockForeign
:
680 return xe1
->XE
.UnlockForeign
.thr
== xe2
->XE
.UnlockForeign
.thr
681 && xe1
->XE
.UnlockForeign
.owner
== xe2
->XE
.UnlockForeign
.owner
682 && xe1
->XE
.UnlockForeign
.lock
== xe2
->XE
.UnlockForeign
.lock
;
684 return xe1
->XE
.UnlockBogus
.thr
== xe2
->XE
.UnlockBogus
.thr
685 && xe1
->XE
.UnlockBogus
.lock_ga
== xe2
->XE
.UnlockBogus
.lock_ga
;
687 return xe1
->XE
.PthAPIerror
.thr
== xe2
->XE
.PthAPIerror
.thr
688 && 0==VG_(strcmp
)(xe1
->XE
.PthAPIerror
.fnname
,
689 xe2
->XE
.PthAPIerror
.fnname
)
690 && xe1
->XE
.PthAPIerror
.err
== xe2
->XE
.PthAPIerror
.err
;
692 return xe1
->XE
.LockOrder
.thr
== xe2
->XE
.LockOrder
.thr
;
694 return xe1
->XE
.Misc
.thr
== xe2
->XE
.Misc
.thr
695 && 0==VG_(strcmp
)(xe1
->XE
.Misc
.errstr
, xe2
->XE
.Misc
.errstr
);
705 /*----------------------------------------------------------------*/
706 /*--- Error management -- printing ---*/
707 /*----------------------------------------------------------------*/
709 /* Do a printf-style operation on either the XML or normal output
710 channel, depending on the setting of VG_(clo_xml).
712 static void emit_WRK ( const HChar
* format
, va_list vargs
)
715 VG_(vprintf_xml
)(format
, vargs
);
717 VG_(vmessage
)(Vg_UserMsg
, format
, vargs
);
720 static void emit ( const HChar
* format
, ... ) PRINTF_CHECK(1, 2);
721 static void emit ( const HChar
* format
, ... )
724 va_start(vargs
, format
);
725 emit_WRK(format
, vargs
);
730 /* Announce (that is, print the point-of-creation) of 'thr'. Only do
731 this once, as we only want to see these announcements once per
732 thread. Returned Bool indicates whether or not an announcement was
735 static Bool
announce_one_thread ( Thread
* thr
)
737 tl_assert(HG_(is_sane_Thread
)(thr
));
738 tl_assert(thr
->errmsg_index
>= 1);
744 VG_(printf_xml
)("<announcethread>\n");
745 VG_(printf_xml
)(" <hthreadid>%d</hthreadid>\n", thr
->errmsg_index
);
746 if (thr
->errmsg_index
== 1) {
747 tl_assert(thr
->created_at
== NULL
);
748 VG_(printf_xml
)(" <isrootthread></isrootthread>\n");
750 tl_assert(thr
->created_at
!= NULL
);
751 VG_(pp_ExeContext
)( thr
->created_at
);
753 VG_(printf_xml
)("</announcethread>\n\n");
757 VG_(umsg
)("---Thread-Announcement----------"
758 "--------------------------------" "\n");
761 if (thr
->errmsg_index
== 1) {
762 tl_assert(thr
->created_at
== NULL
);
763 VG_(message
)(Vg_UserMsg
,
764 "Thread #%d is the program's root thread\n",
767 tl_assert(thr
->created_at
!= NULL
);
768 VG_(message
)(Vg_UserMsg
, "Thread #%d was created\n",
770 VG_(pp_ExeContext
)( thr
->created_at
);
772 VG_(message
)(Vg_UserMsg
, "\n");
776 thr
->announced
= True
;
781 static void announce_LockP ( Lock
* lk
)
784 if (lk
== Lock_INVALID
)
785 return; /* Can't be announced -- we know nothing about it. */
786 tl_assert(lk
->magic
== LockP_MAGIC
);
789 if (lk
->appeared_at
) {
790 emit( " <auxwhat>Lock at %p was first observed</auxwhat>\n",
792 VG_(pp_ExeContext
)( lk
->appeared_at
);
796 if (lk
->appeared_at
) {
797 VG_(umsg
)( " Lock at %p was first observed\n",
798 (void*)lk
->guestaddr
);
799 VG_(pp_ExeContext
)( lk
->appeared_at
);
801 VG_(umsg
)( " Lock at %p : no stacktrace for first observation\n",
802 (void*)lk
->guestaddr
);
804 HG_(get_and_pp_addrdescr
) (lk
->guestaddr
);
809 /* Announce (that is, print point-of-first-observation) for the
810 locks in 'lockvec' and, if non-NULL, 'lockvec2'. */
811 static void announce_combined_LockP_vecs ( Lock
** lockvec
,
816 for (i
= 0; lockvec
[i
]; i
++) {
817 announce_LockP(lockvec
[i
]);
820 for (i
= 0; lockvec2
[i
]; i
++) {
821 Lock
* lk
= lockvec2
[i
];
822 if (!elem_LockP_vector(lockvec
, lk
))
829 static void show_LockP_summary_textmode ( Lock
** locks
, const HChar
* pre
)
833 UWord nLocks
= 0, nLocksValid
= 0;
834 count_LockP_vector(&nLocks
, &nLocksValid
, locks
);
835 tl_assert(nLocksValid
<= nLocks
);
838 VG_(umsg
)( "%sLocks held: none", pre
);
840 VG_(umsg
)( "%sLocks held: %lu, at address%s ",
841 pre
, nLocks
, nLocksValid
== 1 ? "" : "es" );
845 for (i
= 0; i
< nLocks
; i
++) {
846 if (locks
[i
] == Lock_INVALID
)
848 VG_(umsg
)( "%p", (void*)locks
[i
]->guestaddr
);
849 if (locks
[i
+1] != NULL
)
852 if (nLocksValid
< nLocks
)
853 VG_(umsg
)(" (and %lu that can't be shown)", nLocks
- nLocksValid
);
859 /* This is the "this error is due to be printed shortly; so have a
860 look at it any print any preamble you want" function. We use it to
861 announce any previously un-announced threads in the upcoming error
864 void HG_(before_pp_Error
) ( const Error
* err
)
868 xe
= (XError
*)VG_(get_error_extra
)(err
);
871 switch (VG_(get_error_kind
)(err
)) {
873 announce_one_thread( xe
->XE
.Misc
.thr
);
876 announce_one_thread( xe
->XE
.LockOrder
.thr
);
879 announce_one_thread( xe
->XE
.PthAPIerror
.thr
);
882 announce_one_thread( xe
->XE
.UnlockBogus
.thr
);
884 case XE_UnlockForeign
:
885 announce_one_thread( xe
->XE
.UnlockForeign
.thr
);
886 announce_one_thread( xe
->XE
.UnlockForeign
.owner
);
888 case XE_UnlockUnlocked
:
889 announce_one_thread( xe
->XE
.UnlockUnlocked
.thr
);
892 announce_one_thread( xe
->XE
.Race
.thr
);
893 if (xe
->XE
.Race
.h2_ct
)
894 announce_one_thread( xe
->XE
.Race
.h2_ct
);
895 if (xe
->XE
.Race
.h1_ct
)
896 announce_one_thread( xe
->XE
.Race
.h1_ct
);
897 if (xe
->XE
.Race
.data_addrinfo
.Addr
.Block
.alloc_tinfo
.tnr
) {
898 Thread
* thr
= get_admin_threads();
900 if (thr
->errmsg_index
901 == xe
->XE
.Race
.data_addrinfo
.Addr
.Block
.alloc_tinfo
.tnr
) {
902 announce_one_thread (thr
);
914 void HG_(pp_Error
) ( const Error
* err
)
916 const Bool xml
= VG_(clo_xml
); /* a shorthand, that's all */
919 VG_(umsg
)("--------------------------------"
920 "--------------------------------" "\n");
924 XError
*xe
= (XError
*)VG_(get_error_extra
)(err
);
928 emit( " <kind>%s</kind>\n", HG_(get_error_name
)(err
));
930 switch (VG_(get_error_kind
)(err
)) {
933 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.Misc
.thr
) );
937 emit( " <xwhat>\n" );
938 emit( " <text>Thread #%d: %s</text>\n",
939 (Int
)xe
->XE
.Misc
.thr
->errmsg_index
,
940 xe
->XE
.Misc
.errstr
);
941 emit( " <hthreadid>%d</hthreadid>\n",
942 (Int
)xe
->XE
.Misc
.thr
->errmsg_index
);
943 emit( " </xwhat>\n" );
944 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
945 if (xe
->XE
.Misc
.auxstr
) {
946 emit(" <auxwhat>%s</auxwhat>\n", xe
->XE
.Misc
.auxstr
);
947 if (xe
->XE
.Misc
.auxctx
)
948 VG_(pp_ExeContext
)( xe
->XE
.Misc
.auxctx
);
953 emit( "Thread #%d: %s\n",
954 (Int
)xe
->XE
.Misc
.thr
->errmsg_index
,
955 xe
->XE
.Misc
.errstr
);
956 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
957 if (xe
->XE
.Misc
.auxstr
) {
958 emit(" %s\n", xe
->XE
.Misc
.auxstr
);
959 if (xe
->XE
.Misc
.auxctx
)
960 VG_(pp_ExeContext
)( xe
->XE
.Misc
.auxctx
);
968 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.LockOrder
.thr
) );
972 emit( " <xwhat>\n" );
973 emit( " <text>Thread #%d: lock order \"%p before %p\" "
975 (Int
)xe
->XE
.LockOrder
.thr
->errmsg_index
,
976 (void*)xe
->XE
.LockOrder
.shouldbe_earlier_lk
->guestaddr
,
977 (void*)xe
->XE
.LockOrder
.shouldbe_later_lk
->guestaddr
);
978 emit( " <hthreadid>%d</hthreadid>\n",
979 (Int
)xe
->XE
.LockOrder
.thr
->errmsg_index
);
980 emit( " </xwhat>\n" );
981 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
982 if (xe
->XE
.LockOrder
.shouldbe_earlier_ec
983 && xe
->XE
.LockOrder
.shouldbe_later_ec
) {
984 emit( " <auxwhat>Required order was established by "
985 "acquisition of lock at %p</auxwhat>\n",
986 (void*)xe
->XE
.LockOrder
.shouldbe_earlier_lk
->guestaddr
);
987 VG_(pp_ExeContext
)( xe
->XE
.LockOrder
.shouldbe_earlier_ec
);
988 emit( " <auxwhat>followed by a later acquisition "
989 "of lock at %p</auxwhat>\n",
990 (void*)xe
->XE
.LockOrder
.shouldbe_later_lk
->guestaddr
);
991 VG_(pp_ExeContext
)( xe
->XE
.LockOrder
.shouldbe_later_ec
);
993 announce_LockP ( xe
->XE
.LockOrder
.shouldbe_earlier_lk
);
994 announce_LockP ( xe
->XE
.LockOrder
.shouldbe_later_lk
);
998 emit( "Thread #%d: lock order \"%p before %p\" violated\n",
999 (Int
)xe
->XE
.LockOrder
.thr
->errmsg_index
,
1000 (void*)xe
->XE
.LockOrder
.shouldbe_earlier_lk
->guestaddr
,
1001 (void*)xe
->XE
.LockOrder
.shouldbe_later_lk
->guestaddr
);
1003 emit( "Observed (incorrect) order is: "
1004 "acquisition of lock at %p\n",
1005 (void*)xe
->XE
.LockOrder
.shouldbe_later_lk
->guestaddr
);
1006 if (xe
->XE
.LockOrder
.actual_earlier_ec
) {
1007 VG_(pp_ExeContext
)(xe
->XE
.LockOrder
.actual_earlier_ec
);
1009 emit(" (stack unavailable)\n");
1012 emit(" followed by a later acquisition of lock at %p\n",
1013 (void*)xe
->XE
.LockOrder
.shouldbe_earlier_lk
->guestaddr
);
1014 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1015 if (xe
->XE
.LockOrder
.shouldbe_earlier_ec
1016 && xe
->XE
.LockOrder
.shouldbe_later_ec
) {
1018 emit( "Required order was established by "
1019 "acquisition of lock at %p\n",
1020 (void*)xe
->XE
.LockOrder
.shouldbe_earlier_lk
->guestaddr
);
1021 VG_(pp_ExeContext
)( xe
->XE
.LockOrder
.shouldbe_earlier_ec
);
1023 emit( " followed by a later acquisition of lock at %p\n",
1024 (void*)xe
->XE
.LockOrder
.shouldbe_later_lk
->guestaddr
);
1025 VG_(pp_ExeContext
)( xe
->XE
.LockOrder
.shouldbe_later_ec
);
1028 announce_LockP ( xe
->XE
.LockOrder
.shouldbe_earlier_lk
);
1029 announce_LockP ( xe
->XE
.LockOrder
.shouldbe_later_lk
);
1036 case XE_PthAPIerror
: {
1037 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.PthAPIerror
.thr
) );
1041 emit( " <xwhat>\n" );
1043 " <text>Thread #%d's call to %pS failed</text>\n",
1044 (Int
)xe
->XE
.PthAPIerror
.thr
->errmsg_index
,
1045 xe
->XE
.PthAPIerror
.fnname
);
1046 emit( " <hthreadid>%d</hthreadid>\n",
1047 (Int
)xe
->XE
.PthAPIerror
.thr
->errmsg_index
);
1048 emit( " </xwhat>\n" );
1049 emit( " <what>with error code %ld (%s)</what>\n",
1050 xe
->XE
.PthAPIerror
.err
, xe
->XE
.PthAPIerror
.errstr
);
1051 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1055 emit( "Thread #%d's call to %pS failed\n",
1056 (Int
)xe
->XE
.PthAPIerror
.thr
->errmsg_index
,
1057 xe
->XE
.PthAPIerror
.fnname
);
1058 emit( " with error code %ld (%s)\n",
1059 xe
->XE
.PthAPIerror
.err
, xe
->XE
.PthAPIerror
.errstr
);
1060 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1067 case XE_UnlockBogus
: {
1068 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.UnlockBogus
.thr
) );
1072 emit( " <xwhat>\n" );
1073 emit( " <text>Thread #%d unlocked an invalid "
1074 "lock at %p</text>\n",
1075 (Int
)xe
->XE
.UnlockBogus
.thr
->errmsg_index
,
1076 (void*)xe
->XE
.UnlockBogus
.lock_ga
);
1077 emit( " <hthreadid>%d</hthreadid>\n",
1078 (Int
)xe
->XE
.UnlockBogus
.thr
->errmsg_index
);
1079 emit( " </xwhat>\n" );
1080 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1084 emit( "Thread #%d unlocked an invalid lock at %p\n",
1085 (Int
)xe
->XE
.UnlockBogus
.thr
->errmsg_index
,
1086 (void*)xe
->XE
.UnlockBogus
.lock_ga
);
1087 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1094 case XE_UnlockForeign
: {
1095 tl_assert( HG_(is_sane_LockP
)( xe
->XE
.UnlockForeign
.lock
) );
1096 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.UnlockForeign
.owner
) );
1097 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.UnlockForeign
.thr
) );
1101 emit( " <xwhat>\n" );
1102 emit( " <text>Thread #%d unlocked lock at %p "
1103 "currently held by thread #%d</text>\n",
1104 (Int
)xe
->XE
.UnlockForeign
.thr
->errmsg_index
,
1105 (void*)xe
->XE
.UnlockForeign
.lock
->guestaddr
,
1106 (Int
)xe
->XE
.UnlockForeign
.owner
->errmsg_index
);
1107 emit( " <hthreadid>%d</hthreadid>\n",
1108 (Int
)xe
->XE
.UnlockForeign
.thr
->errmsg_index
);
1109 emit( " <hthreadid>%d</hthreadid>\n",
1110 (Int
)xe
->XE
.UnlockForeign
.owner
->errmsg_index
);
1111 emit( " </xwhat>\n" );
1112 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1113 announce_LockP ( xe
->XE
.UnlockForeign
.lock
);
1117 emit( "Thread #%d unlocked lock at %p "
1118 "currently held by thread #%d\n",
1119 (Int
)xe
->XE
.UnlockForeign
.thr
->errmsg_index
,
1120 (void*)xe
->XE
.UnlockForeign
.lock
->guestaddr
,
1121 (Int
)xe
->XE
.UnlockForeign
.owner
->errmsg_index
);
1122 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1123 announce_LockP ( xe
->XE
.UnlockForeign
.lock
);
1130 case XE_UnlockUnlocked
: {
1131 tl_assert( HG_(is_sane_LockP
)( xe
->XE
.UnlockUnlocked
.lock
) );
1132 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.UnlockUnlocked
.thr
) );
1136 emit( " <xwhat>\n" );
1137 emit( " <text>Thread #%d unlocked a "
1138 "not-locked lock at %p</text>\n",
1139 (Int
)xe
->XE
.UnlockUnlocked
.thr
->errmsg_index
,
1140 (void*)xe
->XE
.UnlockUnlocked
.lock
->guestaddr
);
1141 emit( " <hthreadid>%d</hthreadid>\n",
1142 (Int
)xe
->XE
.UnlockUnlocked
.thr
->errmsg_index
);
1143 emit( " </xwhat>\n" );
1144 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1145 announce_LockP ( xe
->XE
.UnlockUnlocked
.lock
);
1149 emit( "Thread #%d unlocked a not-locked lock at %p\n",
1150 (Int
)xe
->XE
.UnlockUnlocked
.thr
->errmsg_index
,
1151 (void*)xe
->XE
.UnlockUnlocked
.lock
->guestaddr
);
1152 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1153 announce_LockP ( xe
->XE
.UnlockUnlocked
.lock
);
1164 what
= xe
->XE
.Race
.isWrite
? "write" : "read";
1165 szB
= xe
->XE
.Race
.szB
;
1166 err_ga
= VG_(get_error_address
)(err
);
1168 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.Race
.thr
));
1169 if (xe
->XE
.Race
.h2_ct
)
1170 tl_assert( HG_(is_sane_Thread
)( xe
->XE
.Race
.h2_ct
));
1174 /* ------ XML ------ */
1175 emit( " <xwhat>\n" );
1176 emit( " <text>Possible data race during %s of size %d "
1177 "at %p by thread #%d</text>\n",
1178 what
, szB
, (void*)err_ga
, (Int
)xe
->XE
.Race
.thr
->errmsg_index
);
1179 emit( " <hthreadid>%d</hthreadid>\n",
1180 (Int
)xe
->XE
.Race
.thr
->errmsg_index
);
1181 emit( " </xwhat>\n" );
1182 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1184 if (xe
->XE
.Race
.h2_ct
) {
1185 tl_assert(xe
->XE
.Race
.h2_ct_accEC
); // assured by update_extra
1186 emit( " <xauxwhat>\n");
1187 emit( " <text>This conflicts with a previous %s of size %d "
1188 "by thread #%d</text>\n",
1189 xe
->XE
.Race
.h2_ct_accIsW
? "write" : "read",
1190 xe
->XE
.Race
.h2_ct_accSzB
,
1191 xe
->XE
.Race
.h2_ct
->errmsg_index
);
1192 emit( " <hthreadid>%d</hthreadid>\n",
1193 xe
->XE
.Race
.h2_ct
->errmsg_index
);
1194 emit(" </xauxwhat>\n");
1195 VG_(pp_ExeContext
)( xe
->XE
.Race
.h2_ct_accEC
);
1198 if (xe
->XE
.Race
.h1_ct
) {
1199 emit( " <xauxwhat>\n");
1200 emit( " <text>This conflicts with a previous access "
1201 "by thread #%d, after</text>\n",
1202 xe
->XE
.Race
.h1_ct
->errmsg_index
);
1203 emit( " <hthreadid>%d</hthreadid>\n",
1204 xe
->XE
.Race
.h1_ct
->errmsg_index
);
1205 emit(" </xauxwhat>\n");
1206 if (xe
->XE
.Race
.h1_ct_mbsegstartEC
) {
1207 VG_(pp_ExeContext
)( xe
->XE
.Race
.h1_ct_mbsegstartEC
);
1209 emit( " <auxwhat>(the start of the thread)</auxwhat>\n" );
1211 emit( " <auxwhat>but before</auxwhat>\n" );
1212 if (xe
->XE
.Race
.h1_ct_mbsegendEC
) {
1213 VG_(pp_ExeContext
)( xe
->XE
.Race
.h1_ct_mbsegendEC
);
1215 emit( " <auxwhat>(the end of the thread)</auxwhat>\n" );
1221 /* ------ Text ------ */
1222 announce_combined_LockP_vecs( xe
->XE
.Race
.locksHeldW
,
1223 xe
->XE
.Race
.h2_ct_locksHeldW
);
1225 emit( "Possible data race during %s of size %d "
1226 "at %p by thread #%d\n",
1227 what
, szB
, (void*)err_ga
, (Int
)xe
->XE
.Race
.thr
->errmsg_index
);
1229 tl_assert(xe
->XE
.Race
.locksHeldW
);
1230 show_LockP_summary_textmode( xe
->XE
.Race
.locksHeldW
, "" );
1231 VG_(pp_ExeContext
)( VG_(get_error_where
)(err
) );
1233 if (xe
->XE
.Race
.h2_ct
) {
1234 tl_assert(xe
->XE
.Race
.h2_ct_accEC
); // assured by update_extra
1235 tl_assert(xe
->XE
.Race
.h2_ct_locksHeldW
);
1237 emit( "This conflicts with a previous %s of size %d "
1239 xe
->XE
.Race
.h2_ct_accIsW
? "write" : "read",
1240 xe
->XE
.Race
.h2_ct_accSzB
,
1241 xe
->XE
.Race
.h2_ct
->errmsg_index
);
1242 show_LockP_summary_textmode( xe
->XE
.Race
.h2_ct_locksHeldW
, "" );
1243 VG_(pp_ExeContext
)( xe
->XE
.Race
.h2_ct_accEC
);
1246 if (xe
->XE
.Race
.h1_ct
) {
1247 emit( " This conflicts with a previous access by thread #%d, "
1249 xe
->XE
.Race
.h1_ct
->errmsg_index
);
1250 if (xe
->XE
.Race
.h1_ct_mbsegstartEC
) {
1251 VG_(pp_ExeContext
)( xe
->XE
.Race
.h1_ct_mbsegstartEC
);
1253 emit( " (the start of the thread)\n" );
1255 emit( " but before\n" );
1256 if (xe
->XE
.Race
.h1_ct_mbsegendEC
) {
1257 VG_(pp_ExeContext
)( xe
->XE
.Race
.h1_ct_mbsegendEC
);
1259 emit( " (the end of the thread)\n" );
1264 VG_(pp_addrinfo
) (err_ga
, &xe
->XE
.Race
.data_addrinfo
);
1265 break; /* case XE_Race */
1266 } /* case XE_Race */
1270 } /* switch (VG_(get_error_kind)(err)) */
1273 void HG_(print_access
) (StackTrace ips
, UInt n_ips
,
1278 WordSetID locksHeldW
)
1282 threadp
= libhb_get_Thr_hgthread( thr_a
);
1284 if (!threadp
->announced
) {
1285 /* This is for interactive use. We announce the thread if needed,
1286 but reset it to not announced afterwards, because we want
1287 the thread to be announced on the error output/log if needed. */
1288 announce_one_thread (threadp
);
1289 threadp
->announced
= False
;
1292 announce_one_thread (threadp
);
1293 VG_(printf
) ("%s of size %d at %p by thread #%d",
1294 isW
? "write" : "read",
1295 (int)SzB
, (void*)ga
, threadp
->errmsg_index
);
1296 if (threadp
->coretid
== VG_INVALID_THREADID
)
1297 VG_(printf
)(" tid (exited)\n");
1299 VG_(printf
)(" tid %d\n", threadp
->coretid
);
1301 Lock
** locksHeldW_P
;
1302 locksHeldW_P
= enumerate_WordSet_into_LockP_vector(
1303 HG_(get_univ_lsets
)(),
1305 True
/*allowed_to_be_invalid*/
1307 show_LockP_summary_textmode( locksHeldW_P
, "" );
1308 HG_(free
) (locksHeldW_P
);
1310 VG_(pp_StackTrace
) (ips
, n_ips
);
1314 const HChar
* HG_(get_error_name
) ( const Error
* err
)
1316 switch (VG_(get_error_kind
)(err
)) {
1317 case XE_Race
: return "Race";
1318 case XE_UnlockUnlocked
: return "UnlockUnlocked";
1319 case XE_UnlockForeign
: return "UnlockForeign";
1320 case XE_UnlockBogus
: return "UnlockBogus";
1321 case XE_PthAPIerror
: return "PthAPIerror";
1322 case XE_LockOrder
: return "LockOrder";
1323 case XE_Misc
: return "Misc";
1324 default: tl_assert(0); /* fill in missing case */
1328 Bool
HG_(recognised_suppression
) ( const HChar
* name
, Supp
*su
)
1330 # define TRY(_name,_xskind) \
1331 if (0 == VG_(strcmp)(name, (_name))) { \
1332 VG_(set_supp_kind)(su, (_xskind)); \
1335 TRY("Race", XS_Race
);
1336 TRY("FreeMemLock", XS_FreeMemLock
);
1337 TRY("UnlockUnlocked", XS_UnlockUnlocked
);
1338 TRY("UnlockForeign", XS_UnlockForeign
);
1339 TRY("UnlockBogus", XS_UnlockBogus
);
1340 TRY("PthAPIerror", XS_PthAPIerror
);
1341 TRY("LockOrder", XS_LockOrder
);
1342 TRY("Misc", XS_Misc
);
1347 Bool
HG_(read_extra_suppression_info
) ( Int fd
, HChar
** bufpp
, SizeT
* nBufp
,
1348 Int
* lineno
, Supp
* su
)
1350 /* do nothing -- no extra suppression info present. Return True to
1351 indicate nothing bad happened. */
1355 Bool
HG_(error_matches_suppression
) ( const Error
* err
, const Supp
* su
)
1357 switch (VG_(get_supp_kind
)(su
)) {
1358 case XS_Race
: return VG_(get_error_kind
)(err
) == XE_Race
;
1359 case XS_UnlockUnlocked
: return VG_(get_error_kind
)(err
) == XE_UnlockUnlocked
;
1360 case XS_UnlockForeign
: return VG_(get_error_kind
)(err
) == XE_UnlockForeign
;
1361 case XS_UnlockBogus
: return VG_(get_error_kind
)(err
) == XE_UnlockBogus
;
1362 case XS_PthAPIerror
: return VG_(get_error_kind
)(err
) == XE_PthAPIerror
;
1363 case XS_LockOrder
: return VG_(get_error_kind
)(err
) == XE_LockOrder
;
1364 case XS_Misc
: return VG_(get_error_kind
)(err
) == XE_Misc
;
1365 //case XS_: return VG_(get_error_kind)(err) == XE_;
1366 default: tl_assert(0); /* fill in missing cases */
1370 SizeT
HG_(get_extra_suppression_info
) ( const Error
* err
,
1371 /*OUT*/HChar
* buf
, Int nBuf
)
1373 tl_assert(nBuf
>= 1);
1379 SizeT
HG_(print_extra_suppression_use
) ( const Supp
* su
,
1380 /*OUT*/HChar
* buf
, Int nBuf
)
1382 tl_assert(nBuf
>= 1);
1388 void HG_(update_extra_suppression_use
) ( const Error
* err
, const Supp
* su
)
1395 /*--------------------------------------------------------------------*/
1396 /*--- end hg_errors.c ---*/
1397 /*--------------------------------------------------------------------*/