mirror of
https://github.com/ioacademy-jikim/debugging
synced 2025-06-08 00:16:11 +00:00
1727 lines
54 KiB
C
1727 lines
54 KiB
C
/*
|
|
This file is part of drd, a thread error detector.
|
|
|
|
Copyright (C) 2006-2015 Bart Van Assche <bvanassche@acm.org>.
|
|
|
|
This program is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU General Public License as
|
|
published by the Free Software Foundation; either version 2 of the
|
|
License, or (at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful, but
|
|
WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA
|
|
02111-1307, USA.
|
|
|
|
The GNU General Public License is contained in the file COPYING.
|
|
*/
|
|
|
|
|
|
#include "drd_error.h"
|
|
#include "drd_barrier.h"
|
|
#include "drd_clientobj.h"
|
|
#include "drd_cond.h"
|
|
#include "drd_mutex.h"
|
|
#include "drd_segment.h"
|
|
#include "drd_semaphore.h"
|
|
#include "drd_suppression.h"
|
|
#include "drd_thread.h"
|
|
#include "pub_tool_vki.h"
|
|
#include "pub_tool_basics.h" // Addr, SizeT
|
|
#include "pub_tool_libcassert.h" // tl_assert()
|
|
#include "pub_tool_libcbase.h" // VG_(strlen)()
|
|
#include "pub_tool_libcprint.h" // VG_(printf)()
|
|
#include "pub_tool_machine.h"
|
|
#include "pub_tool_mallocfree.h" // VG_(malloc)(), VG_(free)()
|
|
#include "pub_tool_options.h" // VG_(clo_backtrace_size)
|
|
#include "pub_tool_threadstate.h" // VG_(get_pthread_id)()
|
|
|
|
|
|
|
|
/* Local functions. */
|
|
|
|
static void thread_append_segment(const DrdThreadId tid, Segment* const sg);
|
|
static void thread_discard_segment(const DrdThreadId tid, Segment* const sg);
|
|
static void thread_compute_conflict_set(struct bitmap** conflict_set,
|
|
const DrdThreadId tid);
|
|
static Bool thread_conflict_set_up_to_date(const DrdThreadId tid);
|
|
|
|
|
|
/* Local variables. */
|
|
|
|
static ULong s_context_switch_count;
|
|
static ULong s_discard_ordered_segments_count;
|
|
static ULong s_compute_conflict_set_count;
|
|
static ULong s_update_conflict_set_count;
|
|
static ULong s_update_conflict_set_new_sg_count;
|
|
static ULong s_update_conflict_set_sync_count;
|
|
static ULong s_update_conflict_set_join_count;
|
|
static ULong s_conflict_set_bitmap_creation_count;
|
|
static ULong s_conflict_set_bitmap2_creation_count;
|
|
static ThreadId s_vg_running_tid = VG_INVALID_THREADID;
|
|
DrdThreadId DRD_(g_drd_running_tid) = DRD_INVALID_THREADID;
|
|
ThreadInfo* DRD_(g_threadinfo);
|
|
struct bitmap* DRD_(g_conflict_set);
|
|
Bool DRD_(verify_conflict_set);
|
|
static Bool s_trace_context_switches = False;
|
|
static Bool s_trace_conflict_set = False;
|
|
static Bool s_trace_conflict_set_bm = False;
|
|
static Bool s_trace_fork_join = False;
|
|
static Bool s_segment_merging = True;
|
|
static Bool s_new_segments_since_last_merge;
|
|
static int s_segment_merge_interval = 10;
|
|
static unsigned s_join_list_vol = 10;
|
|
static unsigned s_deletion_head;
|
|
static unsigned s_deletion_tail;
|
|
#if defined(VGO_solaris)
|
|
Bool DRD_(ignore_thread_creation) = True;
|
|
#else
|
|
Bool DRD_(ignore_thread_creation) = False;
|
|
#endif /* VGO_solaris */
|
|
|
|
|
|
/* Function definitions. */
|
|
|
|
/** Enables/disables context switch tracing. */
|
|
void DRD_(thread_trace_context_switches)(const Bool t)
|
|
{
|
|
tl_assert(t == False || t == True);
|
|
s_trace_context_switches = t;
|
|
}
|
|
|
|
/** Enables/disables conflict set tracing. */
|
|
void DRD_(thread_trace_conflict_set)(const Bool t)
|
|
{
|
|
tl_assert(t == False || t == True);
|
|
s_trace_conflict_set = t;
|
|
}
|
|
|
|
/** Enables/disables conflict set bitmap tracing. */
|
|
void DRD_(thread_trace_conflict_set_bm)(const Bool t)
|
|
{
|
|
tl_assert(t == False || t == True);
|
|
s_trace_conflict_set_bm = t;
|
|
}
|
|
|
|
/** Report whether fork/join tracing is enabled. */
|
|
Bool DRD_(thread_get_trace_fork_join)(void)
|
|
{
|
|
return s_trace_fork_join;
|
|
}
|
|
|
|
/** Enables/disables fork/join tracing. */
|
|
void DRD_(thread_set_trace_fork_join)(const Bool t)
|
|
{
|
|
tl_assert(t == False || t == True);
|
|
s_trace_fork_join = t;
|
|
}
|
|
|
|
/** Enables/disables segment merging. */
|
|
void DRD_(thread_set_segment_merging)(const Bool m)
|
|
{
|
|
tl_assert(m == False || m == True);
|
|
s_segment_merging = m;
|
|
}
|
|
|
|
/** Get the segment merging interval. */
|
|
int DRD_(thread_get_segment_merge_interval)(void)
|
|
{
|
|
return s_segment_merge_interval;
|
|
}
|
|
|
|
/** Set the segment merging interval. */
|
|
void DRD_(thread_set_segment_merge_interval)(const int i)
|
|
{
|
|
s_segment_merge_interval = i;
|
|
}
|
|
|
|
void DRD_(thread_set_join_list_vol)(const int jlv)
|
|
{
|
|
s_join_list_vol = jlv;
|
|
}
|
|
|
|
void DRD_(thread_init)(void)
|
|
{
|
|
DRD_(g_threadinfo) = VG_(malloc)("drd.main.ti.1",
|
|
DRD_N_THREADS * sizeof DRD_(g_threadinfo)[0]);
|
|
for (UInt i = 0; i < DRD_N_THREADS; ++i) {
|
|
static ThreadInfo initval;
|
|
DRD_(g_threadinfo)[i] = initval;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Convert Valgrind's ThreadId into a DrdThreadId.
|
|
*
|
|
* @return DRD thread ID upon success and DRD_INVALID_THREADID if the passed
|
|
* Valgrind ThreadId does not yet exist.
|
|
*/
|
|
DrdThreadId DRD_(VgThreadIdToDrdThreadId)(const ThreadId tid)
|
|
{
|
|
UInt i;
|
|
|
|
if (tid == VG_INVALID_THREADID)
|
|
return DRD_INVALID_THREADID;
|
|
|
|
for (i = 1; i < DRD_N_THREADS; i++)
|
|
{
|
|
if (DRD_(g_threadinfo)[i].vg_thread_exists == True
|
|
&& DRD_(g_threadinfo)[i].vg_threadid == tid)
|
|
{
|
|
return i;
|
|
}
|
|
}
|
|
|
|
return DRD_INVALID_THREADID;
|
|
}
|
|
|
|
/** Allocate a new DRD thread ID for the specified Valgrind thread ID. */
|
|
static DrdThreadId DRD_(VgThreadIdToNewDrdThreadId)(const ThreadId tid)
|
|
{
|
|
UInt i;
|
|
|
|
tl_assert(DRD_(VgThreadIdToDrdThreadId)(tid) == DRD_INVALID_THREADID);
|
|
|
|
for (i = 1; i < DRD_N_THREADS; i++)
|
|
{
|
|
if (!DRD_(g_threadinfo)[i].valid)
|
|
{
|
|
tl_assert(! DRD_(IsValidDrdThreadId)(i));
|
|
|
|
DRD_(g_threadinfo)[i].valid = True;
|
|
DRD_(g_threadinfo)[i].vg_thread_exists = True;
|
|
DRD_(g_threadinfo)[i].vg_threadid = tid;
|
|
DRD_(g_threadinfo)[i].pt_threadid = INVALID_POSIX_THREADID;
|
|
DRD_(g_threadinfo)[i].stack_min = 0;
|
|
DRD_(g_threadinfo)[i].stack_min_min = 0;
|
|
DRD_(g_threadinfo)[i].stack_startup = 0;
|
|
DRD_(g_threadinfo)[i].stack_max = 0;
|
|
DRD_(thread_set_name)(i, "");
|
|
DRD_(g_threadinfo)[i].on_alt_stack = False;
|
|
DRD_(g_threadinfo)[i].is_recording_loads = True;
|
|
DRD_(g_threadinfo)[i].is_recording_stores = True;
|
|
DRD_(g_threadinfo)[i].pthread_create_nesting_level = 0;
|
|
DRD_(g_threadinfo)[i].synchr_nesting = 0;
|
|
DRD_(g_threadinfo)[i].deletion_seq = s_deletion_tail - 1;
|
|
DRD_(g_threadinfo)[i].creator_thread = DRD_INVALID_THREADID;
|
|
#if defined (VGO_solaris)
|
|
DRD_(g_threadinfo)[i].bind_guard_flag = 0;
|
|
#endif /* VGO_solaris */
|
|
|
|
tl_assert(DRD_(g_threadinfo)[i].sg_first == NULL);
|
|
tl_assert(DRD_(g_threadinfo)[i].sg_last == NULL);
|
|
|
|
tl_assert(DRD_(IsValidDrdThreadId)(i));
|
|
|
|
return i;
|
|
}
|
|
}
|
|
|
|
VG_(printf)(
|
|
"\nSorry, but the maximum number of threads supported by DRD has been exceeded."
|
|
"Aborting.\n");
|
|
|
|
tl_assert(False);
|
|
|
|
return DRD_INVALID_THREADID;
|
|
}
|
|
|
|
/** Convert a POSIX thread ID into a DRD thread ID. */
|
|
DrdThreadId DRD_(PtThreadIdToDrdThreadId)(const PThreadId tid)
|
|
{
|
|
UInt i;
|
|
|
|
if (tid != INVALID_POSIX_THREADID)
|
|
{
|
|
for (i = 1; i < DRD_N_THREADS; i++)
|
|
{
|
|
if (DRD_(g_threadinfo)[i].posix_thread_exists
|
|
&& DRD_(g_threadinfo)[i].pt_threadid == tid)
|
|
{
|
|
return i;
|
|
}
|
|
}
|
|
}
|
|
return DRD_INVALID_THREADID;
|
|
}
|
|
|
|
/** Convert a DRD thread ID into a Valgrind thread ID. */
|
|
ThreadId DRD_(DrdThreadIdToVgThreadId)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
return (DRD_(g_threadinfo)[tid].vg_thread_exists
|
|
? DRD_(g_threadinfo)[tid].vg_threadid
|
|
: VG_INVALID_THREADID);
|
|
}
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
/**
|
|
* Sanity check of the doubly linked list of segments referenced by a
|
|
* ThreadInfo struct.
|
|
* @return True if sane, False if not.
|
|
*/
|
|
static Bool DRD_(sane_ThreadInfo)(const ThreadInfo* const ti)
|
|
{
|
|
Segment* p;
|
|
|
|
for (p = ti->sg_first; p; p = p->thr_next) {
|
|
if (p->thr_next && p->thr_next->thr_prev != p)
|
|
return False;
|
|
if (p->thr_next == 0 && p != ti->sg_last)
|
|
return False;
|
|
}
|
|
for (p = ti->sg_last; p; p = p->thr_prev) {
|
|
if (p->thr_prev && p->thr_prev->thr_next != p)
|
|
return False;
|
|
if (p->thr_prev == 0 && p != ti->sg_first)
|
|
return False;
|
|
}
|
|
return True;
|
|
}
|
|
#endif
|
|
|
|
/**
|
|
* Create the first segment for a newly started thread.
|
|
*
|
|
* This function is called from the handler installed via
|
|
* VG_(track_pre_thread_ll_create)(). The Valgrind core invokes this handler
|
|
* from the context of the creator thread, before the new thread has been
|
|
* created.
|
|
*
|
|
* @param[in] creator DRD thread ID of the creator thread.
|
|
* @param[in] vg_created Valgrind thread ID of the created thread.
|
|
*
|
|
* @return DRD thread ID of the created thread.
|
|
*/
|
|
DrdThreadId DRD_(thread_pre_create)(const DrdThreadId creator,
|
|
const ThreadId vg_created)
|
|
{
|
|
DrdThreadId created;
|
|
|
|
tl_assert(DRD_(VgThreadIdToDrdThreadId)(vg_created) == DRD_INVALID_THREADID);
|
|
created = DRD_(VgThreadIdToNewDrdThreadId)(vg_created);
|
|
tl_assert(0 <= (int)created && created < DRD_N_THREADS
|
|
&& created != DRD_INVALID_THREADID);
|
|
|
|
tl_assert(DRD_(g_threadinfo)[created].sg_first == NULL);
|
|
tl_assert(DRD_(g_threadinfo)[created].sg_last == NULL);
|
|
|
|
if (creator != DRD_INVALID_THREADID) {
|
|
if (DRD_(ignore_thread_creation)) {
|
|
tl_assert(DRD_(thread_get_synchr_nesting_count)(created) == 0);
|
|
DRD_(thread_enter_synchr)(created);
|
|
/* Counterpart in DRD_(thread_set_pthreadid)(). */
|
|
}
|
|
}
|
|
DRD_(g_threadinfo)[created].creator_thread = creator;
|
|
|
|
/* Create an initial segment for the newly created thread. */
|
|
thread_append_segment(created, DRD_(sg_new)(creator, created));
|
|
|
|
return created;
|
|
}
|
|
|
|
/**
|
|
* Initialize DRD_(g_threadinfo)[] for a newly created thread. Must be called
|
|
* after the thread has been created and before any client instructions are run
|
|
* on the newly created thread, e.g. from the handler installed via
|
|
* VG_(track_pre_thread_first_insn)().
|
|
*
|
|
* @param[in] vg_created Valgrind thread ID of the newly created thread.
|
|
*
|
|
* @return DRD thread ID for the new thread.
|
|
*/
|
|
DrdThreadId DRD_(thread_post_create)(const ThreadId vg_created)
|
|
{
|
|
const DrdThreadId created = DRD_(VgThreadIdToDrdThreadId)(vg_created);
|
|
|
|
tl_assert(0 <= (int)created && created < DRD_N_THREADS
|
|
&& created != DRD_INVALID_THREADID);
|
|
|
|
DRD_(g_threadinfo)[created].stack_max
|
|
= VG_(thread_get_stack_max)(vg_created);
|
|
DRD_(g_threadinfo)[created].stack_startup
|
|
= DRD_(g_threadinfo)[created].stack_max;
|
|
DRD_(g_threadinfo)[created].stack_min
|
|
= DRD_(g_threadinfo)[created].stack_max;
|
|
DRD_(g_threadinfo)[created].stack_min_min
|
|
= DRD_(g_threadinfo)[created].stack_max;
|
|
DRD_(g_threadinfo)[created].stack_size
|
|
= VG_(thread_get_stack_size)(vg_created);
|
|
tl_assert(DRD_(g_threadinfo)[created].stack_max != 0);
|
|
|
|
return created;
|
|
}
|
|
|
|
static void DRD_(thread_delayed_delete)(const DrdThreadId tid)
|
|
{
|
|
UInt j;
|
|
|
|
DRD_(g_threadinfo)[tid].vg_thread_exists = False;
|
|
DRD_(g_threadinfo)[tid].posix_thread_exists = False;
|
|
DRD_(g_threadinfo)[tid].deletion_seq = s_deletion_head++;
|
|
#if 0
|
|
VG_(message)(Vg_DebugMsg, "Adding thread %d to the deletion list\n", tid);
|
|
#endif
|
|
if (s_deletion_head - s_deletion_tail >= s_join_list_vol) {
|
|
for (j = 0; j < DRD_N_THREADS; ++j) {
|
|
if (DRD_(IsValidDrdThreadId)(j)
|
|
&& DRD_(g_threadinfo)[j].deletion_seq == s_deletion_tail)
|
|
{
|
|
s_deletion_tail++;
|
|
#if 0
|
|
VG_(message)(Vg_DebugMsg, "Delayed delete of thread %d\n", j);
|
|
#endif
|
|
DRD_(thread_delete)(j, False);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Process VG_USERREQ__POST_THREAD_JOIN. This client request is invoked just
|
|
* after thread drd_joiner joined thread drd_joinee.
|
|
*/
|
|
void DRD_(thread_post_join)(DrdThreadId drd_joiner, DrdThreadId drd_joinee)
|
|
{
|
|
tl_assert(DRD_(IsValidDrdThreadId)(drd_joiner));
|
|
tl_assert(DRD_(IsValidDrdThreadId)(drd_joinee));
|
|
|
|
DRD_(thread_new_segment)(drd_joiner);
|
|
DRD_(thread_combine_vc_join)(drd_joiner, drd_joinee);
|
|
DRD_(thread_new_segment)(drd_joinee);
|
|
|
|
if (s_trace_fork_join)
|
|
{
|
|
const ThreadId joiner = DRD_(DrdThreadIdToVgThreadId)(drd_joiner);
|
|
const unsigned msg_size = 256;
|
|
HChar* msg;
|
|
|
|
msg = VG_(malloc)("drd.main.dptj.1", msg_size);
|
|
|
|
VG_(snprintf)(msg, msg_size,
|
|
"drd_post_thread_join joiner = %u, joinee = %u",
|
|
drd_joiner, drd_joinee);
|
|
if (joiner)
|
|
{
|
|
HChar* vc;
|
|
|
|
vc = DRD_(vc_aprint)(DRD_(thread_get_vc)(drd_joiner));
|
|
VG_(snprintf)(msg + VG_(strlen)(msg), msg_size - VG_(strlen)(msg),
|
|
", new vc: %s", vc);
|
|
VG_(free)(vc);
|
|
}
|
|
DRD_(trace_msg)("%pS", msg);
|
|
VG_(free)(msg);
|
|
}
|
|
|
|
if (! DRD_(get_check_stack_accesses)())
|
|
{
|
|
DRD_(finish_suppression)(DRD_(thread_get_stack_max)(drd_joinee)
|
|
- DRD_(thread_get_stack_size)(drd_joinee),
|
|
DRD_(thread_get_stack_max)(drd_joinee));
|
|
}
|
|
DRD_(clientobj_delete_thread)(drd_joinee);
|
|
DRD_(thread_delayed_delete)(drd_joinee);
|
|
}
|
|
|
|
/**
|
|
* NPTL hack: NPTL allocates the 'struct pthread' on top of the stack,
|
|
* and accesses this data structure from multiple threads without locking.
|
|
* Any conflicting accesses in the range stack_startup..stack_max will be
|
|
* ignored.
|
|
*/
|
|
void DRD_(thread_set_stack_startup)(const DrdThreadId tid,
|
|
const Addr stack_startup)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].stack_min <= stack_startup);
|
|
tl_assert(stack_startup <= DRD_(g_threadinfo)[tid].stack_max);
|
|
DRD_(g_threadinfo)[tid].stack_startup = stack_startup;
|
|
}
|
|
|
|
/** Return the stack pointer for the specified thread. */
|
|
Addr DRD_(thread_get_stack_min)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return DRD_(g_threadinfo)[tid].stack_min;
|
|
}
|
|
|
|
/**
|
|
* Return the lowest value that was ever assigned to the stack pointer
|
|
* for the specified thread.
|
|
*/
|
|
Addr DRD_(thread_get_stack_min_min)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return DRD_(g_threadinfo)[tid].stack_min_min;
|
|
}
|
|
|
|
/** Return the top address for the stack of the specified thread. */
|
|
Addr DRD_(thread_get_stack_max)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return DRD_(g_threadinfo)[tid].stack_max;
|
|
}
|
|
|
|
/** Return the maximum stack size for the specified thread. */
|
|
SizeT DRD_(thread_get_stack_size)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return DRD_(g_threadinfo)[tid].stack_size;
|
|
}
|
|
|
|
Bool DRD_(thread_get_on_alt_stack)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return DRD_(g_threadinfo)[tid].on_alt_stack;
|
|
}
|
|
|
|
void DRD_(thread_set_on_alt_stack)(const DrdThreadId tid,
|
|
const Bool on_alt_stack)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(on_alt_stack == !!on_alt_stack);
|
|
DRD_(g_threadinfo)[tid].on_alt_stack = on_alt_stack;
|
|
}
|
|
|
|
Int DRD_(thread_get_threads_on_alt_stack)(void)
|
|
{
|
|
int n = 0;
|
|
|
|
for (UInt i = 1; i < DRD_N_THREADS; i++)
|
|
n += DRD_(g_threadinfo)[i].on_alt_stack;
|
|
return n;
|
|
}
|
|
|
|
/**
|
|
* Clean up thread-specific data structures.
|
|
*/
|
|
void DRD_(thread_delete)(const DrdThreadId tid, const Bool detached)
|
|
{
|
|
Segment* sg;
|
|
Segment* sg_prev;
|
|
|
|
tl_assert(DRD_(IsValidDrdThreadId)(tid));
|
|
|
|
tl_assert(DRD_(g_threadinfo)[tid].synchr_nesting >= 0);
|
|
for (sg = DRD_(g_threadinfo)[tid].sg_last; sg; sg = sg_prev) {
|
|
sg_prev = sg->thr_prev;
|
|
sg->thr_next = NULL;
|
|
sg->thr_prev = NULL;
|
|
DRD_(sg_put)(sg);
|
|
}
|
|
DRD_(g_threadinfo)[tid].valid = False;
|
|
DRD_(g_threadinfo)[tid].vg_thread_exists = False;
|
|
DRD_(g_threadinfo)[tid].posix_thread_exists = False;
|
|
if (detached)
|
|
DRD_(g_threadinfo)[tid].detached_posix_thread = False;
|
|
else
|
|
tl_assert(!DRD_(g_threadinfo)[tid].detached_posix_thread);
|
|
DRD_(g_threadinfo)[tid].sg_first = NULL;
|
|
DRD_(g_threadinfo)[tid].sg_last = NULL;
|
|
|
|
tl_assert(!DRD_(IsValidDrdThreadId)(tid));
|
|
}
|
|
|
|
/**
|
|
* Called after a thread performed its last memory access and before
|
|
* thread_delete() is called. Note: thread_delete() is only called for
|
|
* joinable threads, not for detached threads.
|
|
*/
|
|
void DRD_(thread_finished)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
DRD_(g_threadinfo)[tid].vg_thread_exists = False;
|
|
|
|
if (DRD_(g_threadinfo)[tid].detached_posix_thread)
|
|
{
|
|
/*
|
|
* Once a detached thread has finished, its stack is deallocated and
|
|
* should no longer be taken into account when computing the conflict set.
|
|
*/
|
|
DRD_(g_threadinfo)[tid].stack_min = DRD_(g_threadinfo)[tid].stack_max;
|
|
|
|
/*
|
|
* For a detached thread, calling pthread_exit() invalidates the
|
|
* POSIX thread ID associated with the detached thread. For joinable
|
|
* POSIX threads however, the POSIX thread ID remains live after the
|
|
* pthread_exit() call until pthread_join() is called.
|
|
*/
|
|
DRD_(g_threadinfo)[tid].posix_thread_exists = False;
|
|
}
|
|
}
|
|
|
|
/** Called just after fork() in the child process. */
|
|
void DRD_(drd_thread_atfork_child)(const DrdThreadId tid)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 1; i < DRD_N_THREADS; i++)
|
|
{
|
|
if (i == tid)
|
|
continue;
|
|
if (DRD_(IsValidDrdThreadId(i)))
|
|
DRD_(thread_delete)(i, True);
|
|
tl_assert(!DRD_(IsValidDrdThreadId(i)));
|
|
}
|
|
|
|
DRD_(bm_cleanup)(DRD_(g_conflict_set));
|
|
DRD_(bm_init)(DRD_(g_conflict_set));
|
|
}
|
|
|
|
/** Called just before pthread_cancel(). */
|
|
void DRD_(thread_pre_cancel)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID);
|
|
|
|
if (DRD_(thread_get_trace_fork_join)())
|
|
DRD_(trace_msg)("[%u] drd_thread_pre_cancel %u",
|
|
DRD_(g_drd_running_tid), tid);
|
|
}
|
|
|
|
/**
|
|
* Store the POSIX thread ID for the specified thread.
|
|
*
|
|
* @note This function can be called two times for the same thread -- see also
|
|
* the comment block preceding the pthread_create() wrapper in
|
|
* drd_pthread_intercepts.c.
|
|
*/
|
|
void DRD_(thread_set_pthreadid)(const DrdThreadId tid, const PThreadId ptid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pt_threadid == INVALID_POSIX_THREADID
|
|
|| DRD_(g_threadinfo)[tid].pt_threadid == ptid);
|
|
tl_assert(ptid != INVALID_POSIX_THREADID);
|
|
DRD_(g_threadinfo)[tid].posix_thread_exists = True;
|
|
DRD_(g_threadinfo)[tid].pt_threadid = ptid;
|
|
|
|
if (DRD_(g_threadinfo)[tid].creator_thread != DRD_INVALID_THREADID) {
|
|
if (DRD_(ignore_thread_creation)) {
|
|
DRD_(thread_leave_synchr)(tid);
|
|
tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0);
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Returns true for joinable threads and false for detached threads. */
|
|
Bool DRD_(thread_get_joinable)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
return ! DRD_(g_threadinfo)[tid].detached_posix_thread;
|
|
}
|
|
|
|
/** Store the thread mode: joinable or detached. */
|
|
#if defined(VGP_mips32_linux) || defined(VGP_mips64_linux)
|
|
/* There is a cse related issue in gcc for MIPS. Optimization level
|
|
has to be lowered, so cse related optimizations are not
|
|
included.*/
|
|
__attribute__((optimize("O1")))
|
|
#endif
|
|
void DRD_(thread_set_joinable)(const DrdThreadId tid, const Bool joinable)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert((!! joinable) == joinable);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID);
|
|
|
|
DRD_(g_threadinfo)[tid].detached_posix_thread = ! joinable;
|
|
}
|
|
|
|
/** Tells DRD that the calling thread is about to enter pthread_create(). */
|
|
void DRD_(thread_entering_pthread_create)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pthread_create_nesting_level >= 0);
|
|
|
|
DRD_(g_threadinfo)[tid].pthread_create_nesting_level++;
|
|
|
|
if (DRD_(ignore_thread_creation)) {
|
|
tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0);
|
|
DRD_(thread_enter_synchr)(tid);
|
|
}
|
|
}
|
|
|
|
/** Tells DRD that the calling thread has left pthread_create(). */
|
|
void DRD_(thread_left_pthread_create)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pt_threadid != INVALID_POSIX_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].pthread_create_nesting_level > 0);
|
|
|
|
DRD_(g_threadinfo)[tid].pthread_create_nesting_level--;
|
|
|
|
if (DRD_(ignore_thread_creation)) {
|
|
DRD_(thread_leave_synchr)(tid);
|
|
tl_assert(DRD_(thread_get_synchr_nesting_count)(tid) == 0);
|
|
}
|
|
}
|
|
|
|
#if defined(VGO_solaris)
|
|
/** Handles the bind_guard() intercept. */
|
|
void DRD_(thread_entering_rtld_bind_guard)(const DrdThreadId tid, int flags)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
Int bindflag = (flags & VKI_THR_FLG_RTLD);
|
|
if ((bindflag & DRD_(g_threadinfo)[tid].bind_guard_flag) == 0) {
|
|
DRD_(g_threadinfo)[tid].bind_guard_flag |= bindflag;
|
|
DRD_(thread_enter_synchr)(tid);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Handles the bind_clear() intercept.
|
|
* Call to bind_clear(0) is typically used to determine value of bind_flags.
|
|
*/
|
|
void DRD_(thread_leaving_rtld_bind_clear)(const DrdThreadId tid, int flags)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
Int bindflag = (flags & VKI_THR_FLG_RTLD);
|
|
if ((DRD_(g_threadinfo)[tid].bind_guard_flag & bindflag) != 0) {
|
|
DRD_(g_threadinfo)[tid].bind_guard_flag &= ~bindflag;
|
|
DRD_(thread_leave_synchr)(tid);
|
|
}
|
|
}
|
|
#endif /* VGO_solaris */
|
|
|
|
/** Obtain the thread number and the user-assigned thread name. */
|
|
const HChar* DRD_(thread_get_name)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
return DRD_(g_threadinfo)[tid].name;
|
|
}
|
|
|
|
/** Set the name of the specified thread. */
|
|
void DRD_(thread_set_name)(const DrdThreadId tid, const HChar* const name)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
if (name == NULL || name[0] == 0)
|
|
VG_(snprintf)(DRD_(g_threadinfo)[tid].name,
|
|
sizeof(DRD_(g_threadinfo)[tid].name),
|
|
"Thread %u",
|
|
tid);
|
|
else
|
|
VG_(snprintf)(DRD_(g_threadinfo)[tid].name,
|
|
sizeof(DRD_(g_threadinfo)[tid].name),
|
|
"Thread %u (%s)",
|
|
tid, name);
|
|
DRD_(g_threadinfo)[tid].name[sizeof(DRD_(g_threadinfo)[tid].name) - 1] = 0;
|
|
}
|
|
|
|
/**
|
|
* Update s_vg_running_tid, DRD_(g_drd_running_tid) and recalculate the
|
|
* conflict set.
|
|
*/
|
|
void DRD_(thread_set_vg_running_tid)(const ThreadId vg_tid)
|
|
{
|
|
tl_assert(vg_tid != VG_INVALID_THREADID);
|
|
|
|
if (vg_tid != s_vg_running_tid)
|
|
{
|
|
DRD_(thread_set_running_tid)(vg_tid,
|
|
DRD_(VgThreadIdToDrdThreadId)(vg_tid));
|
|
}
|
|
|
|
tl_assert(s_vg_running_tid != VG_INVALID_THREADID);
|
|
tl_assert(DRD_(g_drd_running_tid) != DRD_INVALID_THREADID);
|
|
}
|
|
|
|
/**
|
|
* Update s_vg_running_tid, DRD_(g_drd_running_tid) and recalculate the
|
|
* conflict set.
|
|
*/
|
|
void DRD_(thread_set_running_tid)(const ThreadId vg_tid,
|
|
const DrdThreadId drd_tid)
|
|
{
|
|
tl_assert(vg_tid != VG_INVALID_THREADID);
|
|
tl_assert(drd_tid != DRD_INVALID_THREADID);
|
|
|
|
if (vg_tid != s_vg_running_tid)
|
|
{
|
|
if (s_trace_context_switches
|
|
&& DRD_(g_drd_running_tid) != DRD_INVALID_THREADID)
|
|
{
|
|
VG_(message)(Vg_DebugMsg,
|
|
"Context switch from thread %u to thread %u;"
|
|
" segments: %llu\n",
|
|
DRD_(g_drd_running_tid), drd_tid,
|
|
DRD_(sg_get_segments_alive_count)());
|
|
}
|
|
s_vg_running_tid = vg_tid;
|
|
DRD_(g_drd_running_tid) = drd_tid;
|
|
thread_compute_conflict_set(&DRD_(g_conflict_set), drd_tid);
|
|
s_context_switch_count++;
|
|
}
|
|
|
|
tl_assert(s_vg_running_tid != VG_INVALID_THREADID);
|
|
tl_assert(DRD_(g_drd_running_tid) != DRD_INVALID_THREADID);
|
|
}
|
|
|
|
/**
|
|
* Increase the synchronization nesting counter. Must be called before the
|
|
* client calls a synchronization function.
|
|
*/
|
|
int DRD_(thread_enter_synchr)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(DRD_(IsValidDrdThreadId)(tid));
|
|
return DRD_(g_threadinfo)[tid].synchr_nesting++;
|
|
}
|
|
|
|
/**
|
|
* Decrease the synchronization nesting counter. Must be called after the
|
|
* client left a synchronization function.
|
|
*/
|
|
int DRD_(thread_leave_synchr)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(DRD_(IsValidDrdThreadId)(tid));
|
|
tl_assert(DRD_(g_threadinfo)[tid].synchr_nesting >= 1);
|
|
return --DRD_(g_threadinfo)[tid].synchr_nesting;
|
|
}
|
|
|
|
/** Returns the synchronization nesting counter. */
|
|
int DRD_(thread_get_synchr_nesting_count)(const DrdThreadId tid)
|
|
{
|
|
tl_assert(DRD_(IsValidDrdThreadId)(tid));
|
|
return DRD_(g_threadinfo)[tid].synchr_nesting;
|
|
}
|
|
|
|
/** Append a new segment at the end of the segment list. */
|
|
static
|
|
void thread_append_segment(const DrdThreadId tid, Segment* const sg)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid]));
|
|
#endif
|
|
|
|
// add at tail
|
|
sg->thr_prev = DRD_(g_threadinfo)[tid].sg_last;
|
|
sg->thr_next = NULL;
|
|
if (DRD_(g_threadinfo)[tid].sg_last)
|
|
DRD_(g_threadinfo)[tid].sg_last->thr_next = sg;
|
|
DRD_(g_threadinfo)[tid].sg_last = sg;
|
|
if (DRD_(g_threadinfo)[tid].sg_first == NULL)
|
|
DRD_(g_threadinfo)[tid].sg_first = sg;
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid]));
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Remove a segment from the segment list of thread threadid, and free the
|
|
* associated memory.
|
|
*/
|
|
static
|
|
void thread_discard_segment(const DrdThreadId tid, Segment* const sg)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid]));
|
|
#endif
|
|
|
|
if (sg->thr_prev)
|
|
sg->thr_prev->thr_next = sg->thr_next;
|
|
if (sg->thr_next)
|
|
sg->thr_next->thr_prev = sg->thr_prev;
|
|
if (sg == DRD_(g_threadinfo)[tid].sg_first)
|
|
DRD_(g_threadinfo)[tid].sg_first = sg->thr_next;
|
|
if (sg == DRD_(g_threadinfo)[tid].sg_last)
|
|
DRD_(g_threadinfo)[tid].sg_last = sg->thr_prev;
|
|
DRD_(sg_put)(sg);
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[tid]));
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Returns a pointer to the vector clock of the most recent segment associated
|
|
* with thread 'tid'.
|
|
*/
|
|
VectorClock* DRD_(thread_get_vc)(const DrdThreadId tid)
|
|
{
|
|
Segment* latest_sg;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
latest_sg = DRD_(g_threadinfo)[tid].sg_last;
|
|
tl_assert(latest_sg);
|
|
return &latest_sg->vc;
|
|
}
|
|
|
|
/**
|
|
* Return the latest segment of thread 'tid' and increment its reference count.
|
|
*/
|
|
void DRD_(thread_get_latest_segment)(Segment** sg, const DrdThreadId tid)
|
|
{
|
|
Segment* latest_sg;
|
|
|
|
tl_assert(sg);
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
latest_sg = DRD_(g_threadinfo)[tid].sg_last;
|
|
tl_assert(latest_sg);
|
|
|
|
DRD_(sg_put)(*sg);
|
|
*sg = DRD_(sg_get)(latest_sg);
|
|
}
|
|
|
|
/**
|
|
* Compute the minimum of all latest vector clocks of all threads
|
|
* (Michiel Ronsse calls this "clock snooping" in his papers about DIOTA).
|
|
*
|
|
* @param vc pointer to a vectorclock, holds result upon return.
|
|
*/
|
|
static void DRD_(thread_compute_minimum_vc)(VectorClock* vc)
|
|
{
|
|
unsigned i;
|
|
Bool first;
|
|
Segment* latest_sg;
|
|
|
|
first = True;
|
|
for (i = 0; i < DRD_N_THREADS; i++)
|
|
{
|
|
latest_sg = DRD_(g_threadinfo)[i].sg_last;
|
|
if (latest_sg) {
|
|
if (first)
|
|
DRD_(vc_assign)(vc, &latest_sg->vc);
|
|
else
|
|
DRD_(vc_min)(vc, &latest_sg->vc);
|
|
first = False;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Compute the maximum of all latest vector clocks of all threads.
|
|
*
|
|
* @param vc pointer to a vectorclock, holds result upon return.
|
|
*/
|
|
static void DRD_(thread_compute_maximum_vc)(VectorClock* vc)
|
|
{
|
|
unsigned i;
|
|
Bool first;
|
|
Segment* latest_sg;
|
|
|
|
first = True;
|
|
for (i = 0; i < DRD_N_THREADS; i++)
|
|
{
|
|
latest_sg = DRD_(g_threadinfo)[i].sg_last;
|
|
if (latest_sg) {
|
|
if (first)
|
|
DRD_(vc_assign)(vc, &latest_sg->vc);
|
|
else
|
|
DRD_(vc_combine)(vc, &latest_sg->vc);
|
|
first = False;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Discard all segments that have a defined order against the latest vector
|
|
* clock of all threads -- these segments can no longer be involved in a
|
|
* data race.
|
|
*/
|
|
static void thread_discard_ordered_segments(void)
|
|
{
|
|
unsigned i;
|
|
VectorClock thread_vc_min;
|
|
|
|
s_discard_ordered_segments_count++;
|
|
|
|
DRD_(vc_init)(&thread_vc_min, 0, 0);
|
|
DRD_(thread_compute_minimum_vc)(&thread_vc_min);
|
|
if (DRD_(sg_get_trace)())
|
|
{
|
|
HChar *vc_min, *vc_max;
|
|
VectorClock thread_vc_max;
|
|
|
|
DRD_(vc_init)(&thread_vc_max, 0, 0);
|
|
DRD_(thread_compute_maximum_vc)(&thread_vc_max);
|
|
vc_min = DRD_(vc_aprint)(&thread_vc_min);
|
|
vc_max = DRD_(vc_aprint)(&thread_vc_max);
|
|
VG_(message)(Vg_DebugMsg,
|
|
"Discarding ordered segments -- min vc is %s, max vc is %s\n",
|
|
vc_min, vc_max);
|
|
VG_(free)(vc_min);
|
|
VG_(free)(vc_max);
|
|
DRD_(vc_cleanup)(&thread_vc_max);
|
|
}
|
|
|
|
for (i = 0; i < DRD_N_THREADS; i++) {
|
|
Segment* sg;
|
|
Segment* sg_next;
|
|
|
|
for (sg = DRD_(g_threadinfo)[i].sg_first;
|
|
sg && (sg_next = sg->thr_next)
|
|
&& DRD_(vc_lte)(&sg->vc, &thread_vc_min);
|
|
sg = sg_next)
|
|
{
|
|
thread_discard_segment(i, sg);
|
|
}
|
|
}
|
|
DRD_(vc_cleanup)(&thread_vc_min);
|
|
}
|
|
|
|
/**
|
|
* An implementation of the property 'equiv(sg1, sg2)' as defined in the paper
|
|
* by Mark Christiaens e.a. The property equiv(sg1, sg2) holds if and only if
|
|
* all segments in the set CS are ordered consistently against both sg1 and
|
|
* sg2. The set CS is defined as the set of segments that can immediately
|
|
* precede future segments via inter-thread synchronization operations. In
|
|
* DRD the set CS consists of the latest segment of each thread combined with
|
|
* all segments for which the reference count is strictly greater than one.
|
|
* The code below is an optimized version of the following:
|
|
*
|
|
* for (i = 0; i < DRD_N_THREADS; i++)
|
|
* {
|
|
* Segment* sg;
|
|
*
|
|
* for (sg = DRD_(g_threadinfo)[i].first; sg; sg = sg->next)
|
|
* {
|
|
* if (sg == DRD_(g_threadinfo)[i].last || DRD_(sg_get_refcnt)(sg) > 1)
|
|
* {
|
|
* if ( DRD_(vc_lte)(&sg1->vc, &sg->vc)
|
|
* != DRD_(vc_lte)(&sg2->vc, &sg->vc)
|
|
* || DRD_(vc_lte)(&sg->vc, &sg1->vc)
|
|
* != DRD_(vc_lte)(&sg->vc, &sg2->vc))
|
|
* {
|
|
* return False;
|
|
* }
|
|
* }
|
|
* }
|
|
* }
|
|
*/
|
|
static Bool thread_consistent_segment_ordering(const DrdThreadId tid,
|
|
Segment* const sg1,
|
|
Segment* const sg2)
|
|
{
|
|
unsigned i;
|
|
|
|
tl_assert(sg1->thr_next);
|
|
tl_assert(sg2->thr_next);
|
|
tl_assert(sg1->thr_next == sg2);
|
|
tl_assert(DRD_(vc_lte)(&sg1->vc, &sg2->vc));
|
|
|
|
for (i = 0; i < DRD_N_THREADS; i++)
|
|
{
|
|
Segment* sg;
|
|
|
|
for (sg = DRD_(g_threadinfo)[i].sg_first; sg; sg = sg->thr_next) {
|
|
if (!sg->thr_next || DRD_(sg_get_refcnt)(sg) > 1) {
|
|
if (DRD_(vc_lte)(&sg2->vc, &sg->vc))
|
|
break;
|
|
if (DRD_(vc_lte)(&sg1->vc, &sg->vc))
|
|
return False;
|
|
}
|
|
}
|
|
for (sg = DRD_(g_threadinfo)[i].sg_last; sg; sg = sg->thr_prev) {
|
|
if (!sg->thr_next || DRD_(sg_get_refcnt)(sg) > 1) {
|
|
if (DRD_(vc_lte)(&sg->vc, &sg1->vc))
|
|
break;
|
|
if (DRD_(vc_lte)(&sg->vc, &sg2->vc))
|
|
return False;
|
|
}
|
|
}
|
|
}
|
|
return True;
|
|
}
|
|
|
|
/**
|
|
* Merge all segments that may be merged without triggering false positives
|
|
* or discarding real data races. For the theoretical background of segment
|
|
* merging, see also the following paper: Mark Christiaens, Michiel Ronsse
|
|
* and Koen De Bosschere. Bounding the number of segment histories during
|
|
* data race detection. Parallel Computing archive, Volume 28, Issue 9,
|
|
* pp 1221-1238, September 2002. This paper contains a proof that merging
|
|
* consecutive segments for which the property equiv(s1,s2) holds can be
|
|
* merged without reducing the accuracy of datarace detection. Furthermore
|
|
* it is also proven that the total number of all segments will never grow
|
|
* unbounded if all segments s1, s2 for which equiv(s1, s2) holds are merged
|
|
* every time a new segment is created. The property equiv(s1, s2) is defined
|
|
* as follows: equiv(s1, s2) <=> for all segments in the set CS, the vector
|
|
* clocks of segments s and s1 are ordered in the same way as those of segments
|
|
* s and s2. The set CS is defined as the set of existing segments s that have
|
|
* the potential to conflict with not yet created segments, either because the
|
|
* segment s is the latest segment of a thread or because it can become the
|
|
* immediate predecessor of a new segment due to a synchronization operation.
|
|
*/
|
|
static void thread_merge_segments(void)
|
|
{
|
|
unsigned i;
|
|
|
|
s_new_segments_since_last_merge = 0;
|
|
|
|
for (i = 0; i < DRD_N_THREADS; i++)
|
|
{
|
|
Segment* sg;
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[i]));
|
|
#endif
|
|
|
|
for (sg = DRD_(g_threadinfo)[i].sg_first; sg; sg = sg->thr_next) {
|
|
if (DRD_(sg_get_refcnt)(sg) == 1 && sg->thr_next) {
|
|
Segment* const sg_next = sg->thr_next;
|
|
if (DRD_(sg_get_refcnt)(sg_next) == 1
|
|
&& sg_next->thr_next
|
|
&& thread_consistent_segment_ordering(i, sg, sg_next))
|
|
{
|
|
/* Merge sg and sg_next into sg. */
|
|
DRD_(sg_merge)(sg, sg_next);
|
|
thread_discard_segment(i, sg_next);
|
|
}
|
|
}
|
|
}
|
|
|
|
#ifdef ENABLE_DRD_CONSISTENCY_CHECKS
|
|
tl_assert(DRD_(sane_ThreadInfo)(&DRD_(g_threadinfo)[i]));
|
|
#endif
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Create a new segment for the specified thread, and discard any segments
|
|
* that cannot cause races anymore.
|
|
*/
|
|
void DRD_(thread_new_segment)(const DrdThreadId tid)
|
|
{
|
|
Segment* last_sg;
|
|
Segment* new_sg;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid)));
|
|
|
|
last_sg = DRD_(g_threadinfo)[tid].sg_last;
|
|
new_sg = DRD_(sg_new)(tid, tid);
|
|
thread_append_segment(tid, new_sg);
|
|
if (tid == DRD_(g_drd_running_tid) && last_sg)
|
|
{
|
|
DRD_(thread_update_conflict_set)(tid, &last_sg->vc);
|
|
s_update_conflict_set_new_sg_count++;
|
|
}
|
|
|
|
tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid)));
|
|
|
|
if (s_segment_merging
|
|
&& ++s_new_segments_since_last_merge >= s_segment_merge_interval)
|
|
{
|
|
thread_discard_ordered_segments();
|
|
thread_merge_segments();
|
|
}
|
|
}
|
|
|
|
/** Call this function after thread 'joiner' joined thread 'joinee'. */
|
|
void DRD_(thread_combine_vc_join)(DrdThreadId joiner, DrdThreadId joinee)
|
|
{
|
|
tl_assert(joiner != joinee);
|
|
tl_assert(0 <= (int)joiner && joiner < DRD_N_THREADS
|
|
&& joiner != DRD_INVALID_THREADID);
|
|
tl_assert(0 <= (int)joinee && joinee < DRD_N_THREADS
|
|
&& joinee != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[joiner].sg_first);
|
|
tl_assert(DRD_(g_threadinfo)[joiner].sg_last);
|
|
tl_assert(DRD_(g_threadinfo)[joinee].sg_first);
|
|
tl_assert(DRD_(g_threadinfo)[joinee].sg_last);
|
|
|
|
if (DRD_(sg_get_trace)())
|
|
{
|
|
HChar *str1, *str2;
|
|
str1 = DRD_(vc_aprint)(DRD_(thread_get_vc)(joiner));
|
|
str2 = DRD_(vc_aprint)(DRD_(thread_get_vc)(joinee));
|
|
VG_(message)(Vg_DebugMsg, "Before join: joiner %s, joinee %s\n",
|
|
str1, str2);
|
|
VG_(free)(str1);
|
|
VG_(free)(str2);
|
|
}
|
|
if (joiner == DRD_(g_drd_running_tid)) {
|
|
VectorClock old_vc;
|
|
|
|
DRD_(vc_copy)(&old_vc, DRD_(thread_get_vc)(joiner));
|
|
DRD_(vc_combine)(DRD_(thread_get_vc)(joiner),
|
|
DRD_(thread_get_vc)(joinee));
|
|
DRD_(thread_update_conflict_set)(joiner, &old_vc);
|
|
s_update_conflict_set_join_count++;
|
|
DRD_(vc_cleanup)(&old_vc);
|
|
} else {
|
|
DRD_(vc_combine)(DRD_(thread_get_vc)(joiner),
|
|
DRD_(thread_get_vc)(joinee));
|
|
}
|
|
|
|
thread_discard_ordered_segments();
|
|
|
|
if (DRD_(sg_get_trace)()) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(DRD_(thread_get_vc)(joiner));
|
|
VG_(message)(Vg_DebugMsg, "After join: %s\n", str);
|
|
VG_(free)(str);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Update the vector clock of the last segment of thread tid with the
|
|
* the vector clock of segment sg.
|
|
*/
|
|
static void thread_combine_vc_sync(DrdThreadId tid, const Segment* sg)
|
|
{
|
|
const VectorClock* const vc = &sg->vc;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(DRD_(g_threadinfo)[tid].sg_first);
|
|
tl_assert(DRD_(g_threadinfo)[tid].sg_last);
|
|
tl_assert(sg);
|
|
tl_assert(vc);
|
|
|
|
if (tid != sg->tid) {
|
|
VectorClock old_vc;
|
|
|
|
DRD_(vc_copy)(&old_vc, DRD_(thread_get_vc)(tid));
|
|
DRD_(vc_combine)(DRD_(thread_get_vc)(tid), vc);
|
|
if (DRD_(sg_get_trace)()) {
|
|
HChar *str1, *str2;
|
|
str1 = DRD_(vc_aprint)(&old_vc);
|
|
str2 = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid));
|
|
VG_(message)(Vg_DebugMsg, "thread %u: vc %s -> %s\n", tid, str1, str2);
|
|
VG_(free)(str1);
|
|
VG_(free)(str2);
|
|
}
|
|
|
|
thread_discard_ordered_segments();
|
|
|
|
DRD_(thread_update_conflict_set)(tid, &old_vc);
|
|
s_update_conflict_set_sync_count++;
|
|
|
|
DRD_(vc_cleanup)(&old_vc);
|
|
} else {
|
|
tl_assert(DRD_(vc_lte)(vc, DRD_(thread_get_vc)(tid)));
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Create a new segment for thread tid and update the vector clock of the last
|
|
* segment of this thread with the vector clock of segment sg. Call this
|
|
* function after thread tid had to wait because of thread synchronization
|
|
* until the memory accesses in the segment sg finished.
|
|
*/
|
|
void DRD_(thread_new_segment_and_combine_vc)(DrdThreadId tid, const Segment* sg)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid)));
|
|
tl_assert(sg);
|
|
|
|
thread_append_segment(tid, DRD_(sg_new)(tid, tid));
|
|
|
|
thread_combine_vc_sync(tid, sg);
|
|
|
|
if (s_segment_merging
|
|
&& ++s_new_segments_since_last_merge >= s_segment_merge_interval)
|
|
{
|
|
thread_discard_ordered_segments();
|
|
thread_merge_segments();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Call this function whenever a thread is no longer using the memory
|
|
* [ a1, a2 [, e.g. because of a call to free() or a stack pointer
|
|
* increase.
|
|
*/
|
|
void DRD_(thread_stop_using_mem)(const Addr a1, const Addr a2)
|
|
{
|
|
Segment* p;
|
|
|
|
for (p = DRD_(g_sg_list); p; p = p->g_next)
|
|
DRD_(bm_clear)(DRD_(sg_bm)(p), a1, a2);
|
|
|
|
DRD_(bm_clear)(DRD_(g_conflict_set), a1, a2);
|
|
}
|
|
|
|
/** Specify whether memory loads should be recorded. */
|
|
void DRD_(thread_set_record_loads)(const DrdThreadId tid, const Bool enabled)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(enabled == !! enabled);
|
|
|
|
DRD_(g_threadinfo)[tid].is_recording_loads = enabled;
|
|
}
|
|
|
|
/** Specify whether memory stores should be recorded. */
|
|
void DRD_(thread_set_record_stores)(const DrdThreadId tid, const Bool enabled)
|
|
{
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(enabled == !! enabled);
|
|
|
|
DRD_(g_threadinfo)[tid].is_recording_stores = enabled;
|
|
}
|
|
|
|
/**
|
|
* Print the segment information for all threads.
|
|
*
|
|
* This function is only used for debugging purposes.
|
|
*/
|
|
void DRD_(thread_print_all)(void)
|
|
{
|
|
UInt i;
|
|
Segment* p;
|
|
|
|
for (i = 0; i < DRD_N_THREADS; i++)
|
|
{
|
|
p = DRD_(g_threadinfo)[i].sg_first;
|
|
if (p) {
|
|
VG_(printf)("**************\n"
|
|
"* thread %3u (%d/%u/%u/%u/0x%lx/%d) *\n"
|
|
"**************\n",
|
|
i,
|
|
DRD_(g_threadinfo)[i].valid,
|
|
DRD_(g_threadinfo)[i].vg_thread_exists,
|
|
DRD_(g_threadinfo)[i].vg_threadid,
|
|
DRD_(g_threadinfo)[i].posix_thread_exists,
|
|
DRD_(g_threadinfo)[i].pt_threadid,
|
|
DRD_(g_threadinfo)[i].detached_posix_thread);
|
|
for ( ; p; p = p->thr_next)
|
|
DRD_(sg_print)(p);
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Show a call stack involved in a data race. */
|
|
static void show_call_stack(const DrdThreadId tid, ExeContext* const callstack)
|
|
{
|
|
const ThreadId vg_tid = DRD_(DrdThreadIdToVgThreadId)(tid);
|
|
|
|
if (vg_tid != VG_INVALID_THREADID) {
|
|
if (callstack)
|
|
VG_(pp_ExeContext)(callstack);
|
|
else
|
|
VG_(get_and_pp_StackTrace)(vg_tid, VG_(clo_backtrace_size));
|
|
} else {
|
|
if (!VG_(clo_xml))
|
|
VG_(message)(Vg_UserMsg,
|
|
" (thread finished, call stack no longer available)\n");
|
|
}
|
|
}
|
|
|
|
/** Print information about the segments involved in a data race. */
|
|
static void
|
|
thread_report_conflicting_segments_segment(const DrdThreadId tid,
|
|
const Addr addr,
|
|
const SizeT size,
|
|
const BmAccessTypeT access_type,
|
|
const Segment* const p)
|
|
{
|
|
unsigned i;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(p);
|
|
|
|
for (i = 0; i < DRD_N_THREADS; i++) {
|
|
if (i != tid) {
|
|
Segment* q;
|
|
|
|
for (q = DRD_(g_threadinfo)[i].sg_last; q; q = q->thr_prev) {
|
|
/*
|
|
* Since q iterates over the segments of thread i in order of
|
|
* decreasing vector clocks, if q->vc <= p->vc, then
|
|
* q->next->vc <= p->vc will also hold. Hence, break out of the
|
|
* loop once this condition is met.
|
|
*/
|
|
if (DRD_(vc_lte)(&q->vc, &p->vc))
|
|
break;
|
|
if (!DRD_(vc_lte)(&p->vc, &q->vc)) {
|
|
if (DRD_(bm_has_conflict_with)(DRD_(sg_bm)(q), addr, addr + size,
|
|
access_type)) {
|
|
Segment* q_next;
|
|
|
|
tl_assert(q->stacktrace);
|
|
if (VG_(clo_xml))
|
|
VG_(printf_xml)(" <other_segment_start>\n");
|
|
else
|
|
VG_(message)(Vg_UserMsg,
|
|
"Other segment start (thread %u)\n", i);
|
|
show_call_stack(i, q->stacktrace);
|
|
if (VG_(clo_xml))
|
|
VG_(printf_xml)(" </other_segment_start>\n"
|
|
" <other_segment_end>\n");
|
|
else
|
|
VG_(message)(Vg_UserMsg,
|
|
"Other segment end (thread %u)\n", i);
|
|
q_next = q->thr_next;
|
|
show_call_stack(i, q_next ? q_next->stacktrace : 0);
|
|
if (VG_(clo_xml))
|
|
VG_(printf_xml)(" </other_segment_end>\n");
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Print information about all segments involved in a data race. */
|
|
void DRD_(thread_report_conflicting_segments)(const DrdThreadId tid,
|
|
const Addr addr,
|
|
const SizeT size,
|
|
const BmAccessTypeT access_type)
|
|
{
|
|
Segment* p;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
|
|
for (p = DRD_(g_threadinfo)[tid].sg_first; p; p = p->thr_next) {
|
|
if (DRD_(bm_has)(DRD_(sg_bm)(p), addr, addr + size, access_type))
|
|
thread_report_conflicting_segments_segment(tid, addr, size,
|
|
access_type, p);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Verify whether the conflict set for thread tid is up to date. Only perform
|
|
* the check if the environment variable DRD_VERIFY_CONFLICT_SET has been set.
|
|
*/
|
|
static Bool thread_conflict_set_up_to_date(const DrdThreadId tid)
|
|
{
|
|
Bool result;
|
|
struct bitmap* computed_conflict_set = 0;
|
|
|
|
if (!DRD_(verify_conflict_set))
|
|
return True;
|
|
|
|
thread_compute_conflict_set(&computed_conflict_set, tid);
|
|
result = DRD_(bm_equal)(DRD_(g_conflict_set), computed_conflict_set);
|
|
if (! result)
|
|
{
|
|
VG_(printf)("actual conflict set:\n");
|
|
DRD_(bm_print)(DRD_(g_conflict_set));
|
|
VG_(printf)("\n");
|
|
VG_(printf)("computed conflict set:\n");
|
|
DRD_(bm_print)(computed_conflict_set);
|
|
VG_(printf)("\n");
|
|
}
|
|
DRD_(bm_delete)(computed_conflict_set);
|
|
return result;
|
|
}
|
|
|
|
/**
|
|
* Compute the conflict set: a bitmap that represents the union of all memory
|
|
* accesses of all segments that are unordered to the current segment of the
|
|
* thread tid.
|
|
*/
|
|
static void thread_compute_conflict_set(struct bitmap** conflict_set,
|
|
const DrdThreadId tid)
|
|
{
|
|
Segment* p;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(tid == DRD_(g_drd_running_tid));
|
|
|
|
s_compute_conflict_set_count++;
|
|
s_conflict_set_bitmap_creation_count
|
|
-= DRD_(bm_get_bitmap_creation_count)();
|
|
s_conflict_set_bitmap2_creation_count
|
|
-= DRD_(bm_get_bitmap2_creation_count)();
|
|
|
|
if (*conflict_set) {
|
|
DRD_(bm_cleanup)(*conflict_set);
|
|
DRD_(bm_init)(*conflict_set);
|
|
} else {
|
|
*conflict_set = DRD_(bm_new)();
|
|
}
|
|
|
|
if (s_trace_conflict_set) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid));
|
|
VG_(message)(Vg_DebugMsg,
|
|
"computing conflict set for thread %u with vc %s\n",
|
|
tid, str);
|
|
VG_(free)(str);
|
|
}
|
|
|
|
p = DRD_(g_threadinfo)[tid].sg_last;
|
|
{
|
|
unsigned j;
|
|
|
|
if (s_trace_conflict_set) {
|
|
HChar* vc;
|
|
|
|
vc = DRD_(vc_aprint)(&p->vc);
|
|
VG_(message)(Vg_DebugMsg, "conflict set: thread [%u] at vc %s\n",
|
|
tid, vc);
|
|
VG_(free)(vc);
|
|
}
|
|
|
|
for (j = 0; j < DRD_N_THREADS; j++) {
|
|
if (j != tid && DRD_(IsValidDrdThreadId)(j)) {
|
|
Segment* q;
|
|
|
|
for (q = DRD_(g_threadinfo)[j].sg_last; q; q = q->thr_prev) {
|
|
if (!DRD_(vc_lte)(&q->vc, &p->vc)
|
|
&& !DRD_(vc_lte)(&p->vc, &q->vc)) {
|
|
if (s_trace_conflict_set) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(&q->vc);
|
|
VG_(message)(Vg_DebugMsg,
|
|
"conflict set: [%u] merging segment %s\n",
|
|
j, str);
|
|
VG_(free)(str);
|
|
}
|
|
DRD_(bm_merge2)(*conflict_set, DRD_(sg_bm)(q));
|
|
} else {
|
|
if (s_trace_conflict_set) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(&q->vc);
|
|
VG_(message)(Vg_DebugMsg,
|
|
"conflict set: [%u] ignoring segment %s\n",
|
|
j, str);
|
|
VG_(free)(str);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
s_conflict_set_bitmap_creation_count
|
|
+= DRD_(bm_get_bitmap_creation_count)();
|
|
s_conflict_set_bitmap2_creation_count
|
|
+= DRD_(bm_get_bitmap2_creation_count)();
|
|
|
|
if (s_trace_conflict_set_bm) {
|
|
VG_(message)(Vg_DebugMsg, "[%u] new conflict set:\n", tid);
|
|
DRD_(bm_print)(*conflict_set);
|
|
VG_(message)(Vg_DebugMsg, "[%u] end of new conflict set.\n", tid);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Update the conflict set after the vector clock of thread tid has been
|
|
* updated from old_vc to its current value, either because a new segment has
|
|
* been created or because of a synchronization operation.
|
|
*/
|
|
void DRD_(thread_update_conflict_set)(const DrdThreadId tid,
|
|
const VectorClock* const old_vc)
|
|
{
|
|
const VectorClock* new_vc;
|
|
Segment* p;
|
|
unsigned j;
|
|
|
|
tl_assert(0 <= (int)tid && tid < DRD_N_THREADS
|
|
&& tid != DRD_INVALID_THREADID);
|
|
tl_assert(old_vc);
|
|
tl_assert(tid == DRD_(g_drd_running_tid));
|
|
tl_assert(DRD_(g_conflict_set));
|
|
|
|
if (s_trace_conflict_set) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(DRD_(thread_get_vc)(tid));
|
|
VG_(message)(Vg_DebugMsg,
|
|
"updating conflict set for thread %u with vc %s\n",
|
|
tid, str);
|
|
VG_(free)(str);
|
|
}
|
|
|
|
new_vc = DRD_(thread_get_vc)(tid);
|
|
tl_assert(DRD_(vc_lte)(old_vc, new_vc));
|
|
|
|
DRD_(bm_unmark)(DRD_(g_conflict_set));
|
|
|
|
for (j = 0; j < DRD_N_THREADS; j++)
|
|
{
|
|
Segment* q;
|
|
|
|
if (j == tid || ! DRD_(IsValidDrdThreadId)(j))
|
|
continue;
|
|
|
|
for (q = DRD_(g_threadinfo)[j].sg_last;
|
|
q && !DRD_(vc_lte)(&q->vc, new_vc);
|
|
q = q->thr_prev) {
|
|
const Bool included_in_old_conflict_set
|
|
= !DRD_(vc_lte)(old_vc, &q->vc);
|
|
const Bool included_in_new_conflict_set
|
|
= !DRD_(vc_lte)(new_vc, &q->vc);
|
|
|
|
if (UNLIKELY(s_trace_conflict_set)) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(&q->vc);
|
|
VG_(message)(Vg_DebugMsg,
|
|
"conflict set: [%u] %s segment %s\n", j,
|
|
included_in_old_conflict_set
|
|
!= included_in_new_conflict_set
|
|
? "merging" : "ignoring", str);
|
|
VG_(free)(str);
|
|
}
|
|
if (included_in_old_conflict_set != included_in_new_conflict_set)
|
|
DRD_(bm_mark)(DRD_(g_conflict_set), DRD_(sg_bm)(q));
|
|
}
|
|
|
|
for ( ; q && !DRD_(vc_lte)(&q->vc, old_vc); q = q->thr_prev) {
|
|
const Bool included_in_old_conflict_set
|
|
= !DRD_(vc_lte)(old_vc, &q->vc);
|
|
const Bool included_in_new_conflict_set
|
|
= !DRD_(vc_lte)(&q->vc, new_vc)
|
|
&& !DRD_(vc_lte)(new_vc, &q->vc);
|
|
|
|
if (UNLIKELY(s_trace_conflict_set)) {
|
|
HChar* str;
|
|
|
|
str = DRD_(vc_aprint)(&q->vc);
|
|
VG_(message)(Vg_DebugMsg,
|
|
"conflict set: [%u] %s segment %s\n", j,
|
|
included_in_old_conflict_set
|
|
!= included_in_new_conflict_set
|
|
? "merging" : "ignoring", str);
|
|
VG_(free)(str);
|
|
}
|
|
if (included_in_old_conflict_set != included_in_new_conflict_set)
|
|
DRD_(bm_mark)(DRD_(g_conflict_set), DRD_(sg_bm)(q));
|
|
}
|
|
}
|
|
|
|
DRD_(bm_clear_marked)(DRD_(g_conflict_set));
|
|
|
|
p = DRD_(g_threadinfo)[tid].sg_last;
|
|
for (j = 0; j < DRD_N_THREADS; j++) {
|
|
if (j != tid && DRD_(IsValidDrdThreadId)(j)) {
|
|
Segment* q;
|
|
for (q = DRD_(g_threadinfo)[j].sg_last;
|
|
q && !DRD_(vc_lte)(&q->vc, &p->vc);
|
|
q = q->thr_prev) {
|
|
if (!DRD_(vc_lte)(&p->vc, &q->vc))
|
|
DRD_(bm_merge2_marked)(DRD_(g_conflict_set), DRD_(sg_bm)(q));
|
|
}
|
|
}
|
|
}
|
|
|
|
DRD_(bm_remove_cleared_marked)(DRD_(g_conflict_set));
|
|
|
|
s_update_conflict_set_count++;
|
|
|
|
if (s_trace_conflict_set_bm)
|
|
{
|
|
VG_(message)(Vg_DebugMsg, "[%u] updated conflict set:\n", tid);
|
|
DRD_(bm_print)(DRD_(g_conflict_set));
|
|
VG_(message)(Vg_DebugMsg, "[%u] end of updated conflict set.\n", tid);
|
|
}
|
|
|
|
tl_assert(thread_conflict_set_up_to_date(DRD_(g_drd_running_tid)));
|
|
}
|
|
|
|
/** Report the number of context switches performed. */
|
|
ULong DRD_(thread_get_context_switch_count)(void)
|
|
{
|
|
return s_context_switch_count;
|
|
}
|
|
|
|
/** Report the number of ordered segments that have been discarded. */
|
|
ULong DRD_(thread_get_discard_ordered_segments_count)(void)
|
|
{
|
|
return s_discard_ordered_segments_count;
|
|
}
|
|
|
|
/** Return how many times the conflict set has been updated entirely. */
|
|
ULong DRD_(thread_get_compute_conflict_set_count)()
|
|
{
|
|
return s_compute_conflict_set_count;
|
|
}
|
|
|
|
/** Return how many times the conflict set has been updated partially. */
|
|
ULong DRD_(thread_get_update_conflict_set_count)(void)
|
|
{
|
|
return s_update_conflict_set_count;
|
|
}
|
|
|
|
/**
|
|
* Return how many times the conflict set has been updated partially
|
|
* because a new segment has been created.
|
|
*/
|
|
ULong DRD_(thread_get_update_conflict_set_new_sg_count)(void)
|
|
{
|
|
return s_update_conflict_set_new_sg_count;
|
|
}
|
|
|
|
/**
|
|
* Return how many times the conflict set has been updated partially
|
|
* because of combining vector clocks due to synchronization operations
|
|
* other than reader/writer lock or barrier operations.
|
|
*/
|
|
ULong DRD_(thread_get_update_conflict_set_sync_count)(void)
|
|
{
|
|
return s_update_conflict_set_sync_count;
|
|
}
|
|
|
|
/**
|
|
* Return how many times the conflict set has been updated partially
|
|
* because of thread joins.
|
|
*/
|
|
ULong DRD_(thread_get_update_conflict_set_join_count)(void)
|
|
{
|
|
return s_update_conflict_set_join_count;
|
|
}
|
|
|
|
/**
|
|
* Return the number of first-level bitmaps that have been created during
|
|
* conflict set updates.
|
|
*/
|
|
ULong DRD_(thread_get_conflict_set_bitmap_creation_count)(void)
|
|
{
|
|
return s_conflict_set_bitmap_creation_count;
|
|
}
|
|
|
|
/**
|
|
* Return the number of second-level bitmaps that have been created during
|
|
* conflict set updates.
|
|
*/
|
|
ULong DRD_(thread_get_conflict_set_bitmap2_creation_count)(void)
|
|
{
|
|
return s_conflict_set_bitmap2_creation_count;
|
|
}
|