mirror of https://gitee.com/openkylin/qemu.git
Queued tcg patches
-----BEGIN PGP SIGNATURE----- iQEcBAABAgAGBQJbq8+KAAoJEGTfOOivfiFfbiQH/if+cTUAU+Fr2Qez96avYt7t jimDOUc7bG+FFrNZYveVNPiP/feKWUIYPJVs9ZoT4jxvT4NOBm/drRkW+BiZO7Tt zqceA+/1Hoc7RlSeo/6AbcIXQLjnTnpFlUW24zNGF0QkG6iS92BPcqezgcR3sRS0 Outf68NxQh7hW/TnHGlL/nxTuHzMfKXZLGiphu6ykzWWXUckrzYmXT4R3tfVVxHV S48nASWsZb8Cga/F1KdCHDv8qYAK8qoEA+01tt//zc/l/ivxfy71HenueN6Dj1Xy 8+HStsh/fRrfq4NSqkXLtBPmeq4bhPsiEx1aCOcnXVIG0hTOe3/QO9Hc+qhkf8Y= =cOeh -----END PGP SIGNATURE----- Merge remote-tracking branch 'remotes/rth/tags/pull-tcg-20180926' into staging Queued tcg patches # gpg: Signature made Wed 26 Sep 2018 19:27:22 BST # gpg: using RSA key 64DF38E8AF7E215F # gpg: Good signature from "Richard Henderson <richard.henderson@linaro.org>" # Primary key fingerprint: 7A48 1E78 868B 4DB6 A85A 05C0 64DF 38E8 AF7E 215F * remotes/rth/tags/pull-tcg-20180926: tcg/i386: fix vector operations on 32-bit hosts qht-bench: add -p flag to precompute hash values qht: constify arguments to some internal functions qht: constify qht_statistics_init qht: constify qht_lookup qht: fix comment in qht_bucket_remove_entry qht: drop ht argument from qht iterators test-qht: speed up + test qht_resize test-qht: test deletion of the last entry in a bucket test-qht: test removal of non-existent entries test-qht: test qht_iter_remove qht: add qht_iter_remove qht: remove unused map param from qht_remove__locked Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
commit
07f426c35e
|
@ -1282,8 +1282,7 @@ void tb_flush(CPUState *cpu)
|
|||
*/
|
||||
#ifdef CONFIG_USER_ONLY
|
||||
|
||||
static void
|
||||
do_tb_invalidate_check(struct qht *ht, void *p, uint32_t hash, void *userp)
|
||||
static void do_tb_invalidate_check(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
TranslationBlock *tb = p;
|
||||
target_ulong addr = *(target_ulong *)userp;
|
||||
|
@ -1304,8 +1303,7 @@ static void tb_invalidate_check(target_ulong address)
|
|||
qht_iter(&tb_ctx.htable, do_tb_invalidate_check, &address);
|
||||
}
|
||||
|
||||
static void
|
||||
do_tb_page_check(struct qht *ht, void *p, uint32_t hash, void *userp)
|
||||
static void do_tb_page_check(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
TranslationBlock *tb = p;
|
||||
int flags1, flags2;
|
||||
|
|
|
@ -43,7 +43,8 @@ struct qht_stats {
|
|||
};
|
||||
|
||||
typedef bool (*qht_lookup_func_t)(const void *obj, const void *userp);
|
||||
typedef void (*qht_iter_func_t)(struct qht *ht, void *p, uint32_t h, void *up);
|
||||
typedef void (*qht_iter_func_t)(void *p, uint32_t h, void *up);
|
||||
typedef bool (*qht_iter_bool_func_t)(void *p, uint32_t h, void *up);
|
||||
|
||||
#define QHT_MODE_AUTO_RESIZE 0x1 /* auto-resize when heavily loaded */
|
||||
#define QHT_MODE_RAW_MUTEXES 0x2 /* bypass the profiler (QSP) */
|
||||
|
@ -103,7 +104,7 @@ bool qht_insert(struct qht *ht, void *p, uint32_t hash, void **existing);
|
|||
* Returns the corresponding pointer when a match is found.
|
||||
* Returns NULL otherwise.
|
||||
*/
|
||||
void *qht_lookup_custom(struct qht *ht, const void *userp, uint32_t hash,
|
||||
void *qht_lookup_custom(const struct qht *ht, const void *userp, uint32_t hash,
|
||||
qht_lookup_func_t func);
|
||||
|
||||
/**
|
||||
|
@ -114,7 +115,7 @@ void *qht_lookup_custom(struct qht *ht, const void *userp, uint32_t hash,
|
|||
*
|
||||
* Calls qht_lookup_custom() using @ht's default comparison function.
|
||||
*/
|
||||
void *qht_lookup(struct qht *ht, const void *userp, uint32_t hash);
|
||||
void *qht_lookup(const struct qht *ht, const void *userp, uint32_t hash);
|
||||
|
||||
/**
|
||||
* qht_remove - remove a pointer from the hash table
|
||||
|
@ -179,9 +180,26 @@ bool qht_resize(struct qht *ht, size_t n_elems);
|
|||
*
|
||||
* Each time it is called, user-provided @func is passed a pointer-hash pair,
|
||||
* plus @userp.
|
||||
*
|
||||
* Note: @ht cannot be accessed from @func
|
||||
* See also: qht_iter_remove()
|
||||
*/
|
||||
void qht_iter(struct qht *ht, qht_iter_func_t func, void *userp);
|
||||
|
||||
/**
|
||||
* qht_iter_remove - Iterate over a QHT, optionally removing entries
|
||||
* @ht: QHT to be iterated over
|
||||
* @func: function to be called for each entry in QHT
|
||||
* @userp: additional pointer to be passed to @func
|
||||
*
|
||||
* Each time it is called, user-provided @func is passed a pointer-hash pair,
|
||||
* plus @userp. If @func returns true, the pointer-hash pair is removed.
|
||||
*
|
||||
* Note: @ht cannot be accessed from @func
|
||||
* See also: qht_iter()
|
||||
*/
|
||||
void qht_iter_remove(struct qht *ht, qht_iter_bool_func_t func, void *userp);
|
||||
|
||||
/**
|
||||
* qht_statistics_init - Gather statistics from a QHT
|
||||
* @ht: QHT to gather statistics from
|
||||
|
@ -193,7 +211,7 @@ void qht_iter(struct qht *ht, qht_iter_func_t func, void *userp);
|
|||
* When done with @stats, pass the struct to qht_statistics_destroy().
|
||||
* Failing to do this will leak memory.
|
||||
*/
|
||||
void qht_statistics_init(struct qht *ht, struct qht_stats *stats);
|
||||
void qht_statistics_init(const struct qht *ht, struct qht_stats *stats);
|
||||
|
||||
/**
|
||||
* qht_statistics_destroy - Destroy a &struct qht_stats
|
||||
|
|
|
@ -302,11 +302,7 @@ static inline int tcg_target_const_match(tcg_target_long val, TCGType type,
|
|||
return 0;
|
||||
}
|
||||
|
||||
#if TCG_TARGET_REG_BITS == 64
|
||||
# define LOWREGMASK(x) ((x) & 7)
|
||||
#else
|
||||
# define LOWREGMASK(x) (x)
|
||||
#endif
|
||||
|
||||
#define P_EXT 0x100 /* 0x0f opcode prefix */
|
||||
#define P_EXT38 0x200 /* 0x0f 0x38 opcode prefix */
|
||||
|
|
|
@ -53,6 +53,7 @@ static unsigned long resize_delay = 1000;
|
|||
static double resize_rate; /* 0.0 to 1.0 */
|
||||
static unsigned int n_rz_threads = 1;
|
||||
static QemuThread *rz_threads;
|
||||
static bool precompute_hash;
|
||||
|
||||
static double update_rate; /* 0.0 to 1.0 */
|
||||
static uint64_t update_threshold;
|
||||
|
@ -101,11 +102,18 @@ static bool is_equal(const void *ap, const void *bp)
|
|||
return *a == *b;
|
||||
}
|
||||
|
||||
static inline uint32_t h(unsigned long v)
|
||||
static uint32_t h(unsigned long v)
|
||||
{
|
||||
return tb_hash_func7(v, 0, 0, 0, 0);
|
||||
}
|
||||
|
||||
static uint32_t hval(unsigned long v)
|
||||
{
|
||||
return v;
|
||||
}
|
||||
|
||||
static uint32_t (*hfunc)(unsigned long v) = h;
|
||||
|
||||
/*
|
||||
* From: https://en.wikipedia.org/wiki/Xorshift
|
||||
* This is faster than rand_r(), and gives us a wider range (RAND_MAX is only
|
||||
|
@ -149,7 +157,7 @@ static void do_rw(struct thread_info *info)
|
|||
bool read;
|
||||
|
||||
p = &keys[info->r & (lookup_range - 1)];
|
||||
hash = h(*p);
|
||||
hash = hfunc(*p);
|
||||
read = qht_lookup(&ht, p, hash);
|
||||
if (read) {
|
||||
stats->rd++;
|
||||
|
@ -158,7 +166,7 @@ static void do_rw(struct thread_info *info)
|
|||
}
|
||||
} else {
|
||||
p = &keys[info->r & (update_range - 1)];
|
||||
hash = h(*p);
|
||||
hash = hfunc(*p);
|
||||
if (info->write_op) {
|
||||
bool written = false;
|
||||
|
||||
|
@ -289,7 +297,9 @@ static void htable_init(void)
|
|||
/* avoid allocating memory later by allocating all the keys now */
|
||||
keys = g_malloc(sizeof(*keys) * n);
|
||||
for (i = 0; i < n; i++) {
|
||||
keys[i] = populate_offset + i;
|
||||
long val = populate_offset + i;
|
||||
|
||||
keys[i] = precompute_hash ? h(val) : hval(val);
|
||||
}
|
||||
|
||||
/* some sanity checks */
|
||||
|
@ -321,7 +331,7 @@ static void htable_init(void)
|
|||
|
||||
r = xorshift64star(r);
|
||||
p = &keys[r & (init_range - 1)];
|
||||
hash = h(*p);
|
||||
hash = hfunc(*p);
|
||||
if (qht_insert(&ht, p, hash, NULL)) {
|
||||
break;
|
||||
}
|
||||
|
@ -412,7 +422,7 @@ static void parse_args(int argc, char *argv[])
|
|||
int c;
|
||||
|
||||
for (;;) {
|
||||
c = getopt(argc, argv, "d:D:g:k:K:l:hn:N:o:r:Rs:S:u:");
|
||||
c = getopt(argc, argv, "d:D:g:k:K:l:hn:N:o:pr:Rs:S:u:");
|
||||
if (c < 0) {
|
||||
break;
|
||||
}
|
||||
|
@ -451,6 +461,10 @@ static void parse_args(int argc, char *argv[])
|
|||
case 'o':
|
||||
populate_offset = atol(optarg);
|
||||
break;
|
||||
case 'p':
|
||||
precompute_hash = true;
|
||||
hfunc = hval;
|
||||
break;
|
||||
case 'r':
|
||||
update_range = pow2ceil(atol(optarg));
|
||||
break;
|
||||
|
|
|
@ -41,7 +41,7 @@ static void insert(int a, int b)
|
|||
}
|
||||
}
|
||||
|
||||
static void rm(int init, int end)
|
||||
static void do_rm(int init, int end, bool exist)
|
||||
{
|
||||
int i;
|
||||
|
||||
|
@ -49,10 +49,24 @@ static void rm(int init, int end)
|
|||
uint32_t hash;
|
||||
|
||||
hash = arr[i];
|
||||
g_assert_true(qht_remove(&ht, &arr[i], hash));
|
||||
if (exist) {
|
||||
g_assert_true(qht_remove(&ht, &arr[i], hash));
|
||||
} else {
|
||||
g_assert_false(qht_remove(&ht, &arr[i], hash));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
static void rm(int init, int end)
|
||||
{
|
||||
do_rm(init, end, true);
|
||||
}
|
||||
|
||||
static void rm_nonexist(int init, int end)
|
||||
{
|
||||
do_rm(init, end, false);
|
||||
}
|
||||
|
||||
static void check(int a, int b, bool expected)
|
||||
{
|
||||
struct qht_stats stats;
|
||||
|
@ -84,7 +98,7 @@ static void check(int a, int b, bool expected)
|
|||
qht_statistics_destroy(&stats);
|
||||
}
|
||||
|
||||
static void count_func(struct qht *ht, void *p, uint32_t hash, void *userp)
|
||||
static void count_func(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
unsigned int *curr = userp;
|
||||
|
||||
|
@ -108,14 +122,79 @@ static void iter_check(unsigned int count)
|
|||
g_assert_cmpuint(curr, ==, count);
|
||||
}
|
||||
|
||||
static void sum_func(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
uint32_t *sum = userp;
|
||||
uint32_t a = *(uint32_t *)p;
|
||||
|
||||
*sum += a;
|
||||
}
|
||||
|
||||
static void iter_sum_check(unsigned int expected)
|
||||
{
|
||||
unsigned int sum = 0;
|
||||
|
||||
qht_iter(&ht, sum_func, &sum);
|
||||
g_assert_cmpuint(sum, ==, expected);
|
||||
}
|
||||
|
||||
static bool rm_mod_func(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
uint32_t a = *(uint32_t *)p;
|
||||
unsigned int mod = *(unsigned int *)userp;
|
||||
|
||||
return a % mod == 0;
|
||||
}
|
||||
|
||||
static void iter_rm_mod(unsigned int mod)
|
||||
{
|
||||
qht_iter_remove(&ht, rm_mod_func, &mod);
|
||||
}
|
||||
|
||||
static void iter_rm_mod_check(unsigned int mod)
|
||||
{
|
||||
unsigned int expected = 0;
|
||||
unsigned int i;
|
||||
|
||||
for (i = 0; i < N; i++) {
|
||||
if (i % mod == 0) {
|
||||
continue;
|
||||
}
|
||||
expected += i;
|
||||
}
|
||||
iter_sum_check(expected);
|
||||
}
|
||||
|
||||
static void qht_do_test(unsigned int mode, size_t init_entries)
|
||||
{
|
||||
/* under KVM we might fetch stats from an uninitialized qht */
|
||||
check_n(0);
|
||||
|
||||
qht_init(&ht, is_equal, 0, mode);
|
||||
rm_nonexist(0, 4);
|
||||
/*
|
||||
* Test that we successfully delete the last element in a bucket.
|
||||
* This is a hard-to-reach code path when resizing is on, but without
|
||||
* resizing we can easily hit it if init_entries <= 1.
|
||||
* Given that the number of elements per bucket can be 4 or 6 depending on
|
||||
* the host's pointer size, test the removal of the 4th and 6th elements.
|
||||
*/
|
||||
insert(0, 4);
|
||||
rm_nonexist(5, 6);
|
||||
rm(3, 4);
|
||||
check_n(3);
|
||||
insert(3, 6);
|
||||
rm(5, 6);
|
||||
check_n(5);
|
||||
rm_nonexist(7, 8);
|
||||
iter_rm_mod(1);
|
||||
|
||||
if (!(mode & QHT_MODE_AUTO_RESIZE)) {
|
||||
qht_resize(&ht, init_entries * 4 + 4);
|
||||
}
|
||||
|
||||
check_n(0);
|
||||
rm_nonexist(0, 10);
|
||||
insert(0, N);
|
||||
check(0, N, true);
|
||||
check_n(N);
|
||||
|
@ -138,8 +217,12 @@ static void qht_do_test(unsigned int mode, size_t init_entries)
|
|||
insert(10, 150);
|
||||
check_n(N);
|
||||
|
||||
rm(1, 2);
|
||||
check_n(N - 1);
|
||||
qht_reset(&ht);
|
||||
insert(0, N);
|
||||
rm_nonexist(N, N + 32);
|
||||
iter_rm_mod(10);
|
||||
iter_rm_mod_check(10);
|
||||
check_n(N * 9 / 10);
|
||||
qht_reset_size(&ht, 0);
|
||||
check_n(0);
|
||||
check(0, N, false);
|
||||
|
|
138
util/qht.c
138
util/qht.c
|
@ -89,6 +89,19 @@
|
|||
#define QHT_BUCKET_ENTRIES 4
|
||||
#endif
|
||||
|
||||
enum qht_iter_type {
|
||||
QHT_ITER_VOID, /* do nothing; use retvoid */
|
||||
QHT_ITER_RM, /* remove element if retbool returns true */
|
||||
};
|
||||
|
||||
struct qht_iter {
|
||||
union {
|
||||
qht_iter_func_t retvoid;
|
||||
qht_iter_bool_func_t retbool;
|
||||
} f;
|
||||
enum qht_iter_type type;
|
||||
};
|
||||
|
||||
/*
|
||||
* Do _not_ use qemu_mutex_[try]lock directly! Use these macros, otherwise
|
||||
* the profiler (QSP) will deadlock.
|
||||
|
@ -223,7 +236,7 @@ static inline void qht_head_init(struct qht_bucket *b)
|
|||
}
|
||||
|
||||
static inline
|
||||
struct qht_bucket *qht_map_to_bucket(struct qht_map *map, uint32_t hash)
|
||||
struct qht_bucket *qht_map_to_bucket(const struct qht_map *map, uint32_t hash)
|
||||
{
|
||||
return &map->buckets[hash & (map->n_buckets - 1)];
|
||||
}
|
||||
|
@ -255,7 +268,8 @@ static void qht_map_unlock_buckets(struct qht_map *map)
|
|||
* Call with at least a bucket lock held.
|
||||
* @map should be the value read before acquiring the lock (or locks).
|
||||
*/
|
||||
static inline bool qht_map_is_stale__locked(struct qht *ht, struct qht_map *map)
|
||||
static inline bool qht_map_is_stale__locked(const struct qht *ht,
|
||||
const struct qht_map *map)
|
||||
{
|
||||
return map != ht->map;
|
||||
}
|
||||
|
@ -324,12 +338,12 @@ struct qht_bucket *qht_bucket_lock__no_stale(struct qht *ht, uint32_t hash,
|
|||
return b;
|
||||
}
|
||||
|
||||
static inline bool qht_map_needs_resize(struct qht_map *map)
|
||||
static inline bool qht_map_needs_resize(const struct qht_map *map)
|
||||
{
|
||||
return atomic_read(&map->n_added_buckets) > map->n_added_buckets_threshold;
|
||||
}
|
||||
|
||||
static inline void qht_chain_destroy(struct qht_bucket *head)
|
||||
static inline void qht_chain_destroy(const struct qht_bucket *head)
|
||||
{
|
||||
struct qht_bucket *curr = head->next;
|
||||
struct qht_bucket *prev;
|
||||
|
@ -469,10 +483,10 @@ bool qht_reset_size(struct qht *ht, size_t n_elems)
|
|||
}
|
||||
|
||||
static inline
|
||||
void *qht_do_lookup(struct qht_bucket *head, qht_lookup_func_t func,
|
||||
void *qht_do_lookup(const struct qht_bucket *head, qht_lookup_func_t func,
|
||||
const void *userp, uint32_t hash)
|
||||
{
|
||||
struct qht_bucket *b = head;
|
||||
const struct qht_bucket *b = head;
|
||||
int i;
|
||||
|
||||
do {
|
||||
|
@ -496,7 +510,7 @@ void *qht_do_lookup(struct qht_bucket *head, qht_lookup_func_t func,
|
|||
}
|
||||
|
||||
static __attribute__((noinline))
|
||||
void *qht_lookup__slowpath(struct qht_bucket *b, qht_lookup_func_t func,
|
||||
void *qht_lookup__slowpath(const struct qht_bucket *b, qht_lookup_func_t func,
|
||||
const void *userp, uint32_t hash)
|
||||
{
|
||||
unsigned int version;
|
||||
|
@ -509,11 +523,11 @@ void *qht_lookup__slowpath(struct qht_bucket *b, qht_lookup_func_t func,
|
|||
return ret;
|
||||
}
|
||||
|
||||
void *qht_lookup_custom(struct qht *ht, const void *userp, uint32_t hash,
|
||||
void *qht_lookup_custom(const struct qht *ht, const void *userp, uint32_t hash,
|
||||
qht_lookup_func_t func)
|
||||
{
|
||||
struct qht_bucket *b;
|
||||
struct qht_map *map;
|
||||
const struct qht_bucket *b;
|
||||
const struct qht_map *map;
|
||||
unsigned int version;
|
||||
void *ret;
|
||||
|
||||
|
@ -532,13 +546,16 @@ void *qht_lookup_custom(struct qht *ht, const void *userp, uint32_t hash,
|
|||
return qht_lookup__slowpath(b, func, userp, hash);
|
||||
}
|
||||
|
||||
void *qht_lookup(struct qht *ht, const void *userp, uint32_t hash)
|
||||
void *qht_lookup(const struct qht *ht, const void *userp, uint32_t hash)
|
||||
{
|
||||
return qht_lookup_custom(ht, userp, hash, ht->cmp);
|
||||
}
|
||||
|
||||
/* call with head->lock held */
|
||||
static void *qht_insert__locked(struct qht *ht, struct qht_map *map,
|
||||
/*
|
||||
* call with head->lock held
|
||||
* @ht is const since it is only used for ht->cmp()
|
||||
*/
|
||||
static void *qht_insert__locked(const struct qht *ht, struct qht_map *map,
|
||||
struct qht_bucket *head, void *p, uint32_t hash,
|
||||
bool *needs_resize)
|
||||
{
|
||||
|
@ -632,7 +649,7 @@ bool qht_insert(struct qht *ht, void *p, uint32_t hash, void **existing)
|
|||
return false;
|
||||
}
|
||||
|
||||
static inline bool qht_entry_is_last(struct qht_bucket *b, int pos)
|
||||
static inline bool qht_entry_is_last(const struct qht_bucket *b, int pos)
|
||||
{
|
||||
if (pos == QHT_BUCKET_ENTRIES - 1) {
|
||||
if (b->next == NULL) {
|
||||
|
@ -658,7 +675,7 @@ qht_entry_move(struct qht_bucket *to, int i, struct qht_bucket *from, int j)
|
|||
}
|
||||
|
||||
/*
|
||||
* Find the last valid entry in @head, and swap it with @orig[pos], which has
|
||||
* Find the last valid entry in @orig, and swap it with @orig[pos], which has
|
||||
* just been invalidated.
|
||||
*/
|
||||
static inline void qht_bucket_remove_entry(struct qht_bucket *orig, int pos)
|
||||
|
@ -692,8 +709,7 @@ static inline void qht_bucket_remove_entry(struct qht_bucket *orig, int pos)
|
|||
|
||||
/* call with b->lock held */
|
||||
static inline
|
||||
bool qht_remove__locked(struct qht_map *map, struct qht_bucket *head,
|
||||
const void *p, uint32_t hash)
|
||||
bool qht_remove__locked(struct qht_bucket *head, const void *p, uint32_t hash)
|
||||
{
|
||||
struct qht_bucket *b = head;
|
||||
int i;
|
||||
|
@ -728,15 +744,16 @@ bool qht_remove(struct qht *ht, const void *p, uint32_t hash)
|
|||
qht_debug_assert(p);
|
||||
|
||||
b = qht_bucket_lock__no_stale(ht, hash, &map);
|
||||
ret = qht_remove__locked(map, b, p, hash);
|
||||
ret = qht_remove__locked(b, p, hash);
|
||||
qht_bucket_debug__locked(b);
|
||||
qemu_spin_unlock(&b->lock);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static inline void qht_bucket_iter(struct qht *ht, struct qht_bucket *b,
|
||||
qht_iter_func_t func, void *userp)
|
||||
static inline void qht_bucket_iter(struct qht_bucket *head,
|
||||
const struct qht_iter *iter, void *userp)
|
||||
{
|
||||
struct qht_bucket *b = head;
|
||||
int i;
|
||||
|
||||
do {
|
||||
|
@ -744,37 +761,83 @@ static inline void qht_bucket_iter(struct qht *ht, struct qht_bucket *b,
|
|||
if (b->pointers[i] == NULL) {
|
||||
return;
|
||||
}
|
||||
func(ht, b->pointers[i], b->hashes[i], userp);
|
||||
switch (iter->type) {
|
||||
case QHT_ITER_VOID:
|
||||
iter->f.retvoid(b->pointers[i], b->hashes[i], userp);
|
||||
break;
|
||||
case QHT_ITER_RM:
|
||||
if (iter->f.retbool(b->pointers[i], b->hashes[i], userp)) {
|
||||
/* replace i with the last valid element in the bucket */
|
||||
seqlock_write_begin(&head->sequence);
|
||||
qht_bucket_remove_entry(b, i);
|
||||
seqlock_write_end(&head->sequence);
|
||||
qht_bucket_debug__locked(b);
|
||||
/* reevaluate i, since it just got replaced */
|
||||
i--;
|
||||
continue;
|
||||
}
|
||||
break;
|
||||
default:
|
||||
g_assert_not_reached();
|
||||
}
|
||||
}
|
||||
b = b->next;
|
||||
} while (b);
|
||||
}
|
||||
|
||||
/* call with all of the map's locks held */
|
||||
static inline void qht_map_iter__all_locked(struct qht *ht, struct qht_map *map,
|
||||
qht_iter_func_t func, void *userp)
|
||||
static inline void qht_map_iter__all_locked(struct qht_map *map,
|
||||
const struct qht_iter *iter,
|
||||
void *userp)
|
||||
{
|
||||
size_t i;
|
||||
|
||||
for (i = 0; i < map->n_buckets; i++) {
|
||||
qht_bucket_iter(ht, &map->buckets[i], func, userp);
|
||||
qht_bucket_iter(&map->buckets[i], iter, userp);
|
||||
}
|
||||
}
|
||||
|
||||
void qht_iter(struct qht *ht, qht_iter_func_t func, void *userp)
|
||||
static inline void
|
||||
do_qht_iter(struct qht *ht, const struct qht_iter *iter, void *userp)
|
||||
{
|
||||
struct qht_map *map;
|
||||
|
||||
map = atomic_rcu_read(&ht->map);
|
||||
qht_map_lock_buckets(map);
|
||||
/* Note: ht here is merely for carrying ht->mode; ht->map won't be read */
|
||||
qht_map_iter__all_locked(ht, map, func, userp);
|
||||
qht_map_iter__all_locked(map, iter, userp);
|
||||
qht_map_unlock_buckets(map);
|
||||
}
|
||||
|
||||
static void qht_map_copy(struct qht *ht, void *p, uint32_t hash, void *userp)
|
||||
void qht_iter(struct qht *ht, qht_iter_func_t func, void *userp)
|
||||
{
|
||||
struct qht_map *new = userp;
|
||||
const struct qht_iter iter = {
|
||||
.f.retvoid = func,
|
||||
.type = QHT_ITER_VOID,
|
||||
};
|
||||
|
||||
do_qht_iter(ht, &iter, userp);
|
||||
}
|
||||
|
||||
void qht_iter_remove(struct qht *ht, qht_iter_bool_func_t func, void *userp)
|
||||
{
|
||||
const struct qht_iter iter = {
|
||||
.f.retbool = func,
|
||||
.type = QHT_ITER_RM,
|
||||
};
|
||||
|
||||
do_qht_iter(ht, &iter, userp);
|
||||
}
|
||||
|
||||
struct qht_map_copy_data {
|
||||
struct qht *ht;
|
||||
struct qht_map *new;
|
||||
};
|
||||
|
||||
static void qht_map_copy(void *p, uint32_t hash, void *userp)
|
||||
{
|
||||
struct qht_map_copy_data *data = userp;
|
||||
struct qht *ht = data->ht;
|
||||
struct qht_map *new = data->new;
|
||||
struct qht_bucket *b = qht_map_to_bucket(new, hash);
|
||||
|
||||
/* no need to acquire b->lock because no thread has seen this map yet */
|
||||
|
@ -788,6 +851,11 @@ static void qht_map_copy(struct qht *ht, void *p, uint32_t hash, void *userp)
|
|||
static void qht_do_resize_reset(struct qht *ht, struct qht_map *new, bool reset)
|
||||
{
|
||||
struct qht_map *old;
|
||||
const struct qht_iter iter = {
|
||||
.f.retvoid = qht_map_copy,
|
||||
.type = QHT_ITER_VOID,
|
||||
};
|
||||
struct qht_map_copy_data data;
|
||||
|
||||
old = ht->map;
|
||||
qht_map_lock_buckets(old);
|
||||
|
@ -802,7 +870,9 @@ static void qht_do_resize_reset(struct qht *ht, struct qht_map *new, bool reset)
|
|||
}
|
||||
|
||||
g_assert(new->n_buckets != old->n_buckets);
|
||||
qht_map_iter__all_locked(ht, old, qht_map_copy, new);
|
||||
data.ht = ht;
|
||||
data.new = new;
|
||||
qht_map_iter__all_locked(old, &iter, &data);
|
||||
qht_map_debug__all_locked(new);
|
||||
|
||||
atomic_rcu_set(&ht->map, new);
|
||||
|
@ -829,9 +899,9 @@ bool qht_resize(struct qht *ht, size_t n_elems)
|
|||
}
|
||||
|
||||
/* pass @stats to qht_statistics_destroy() when done */
|
||||
void qht_statistics_init(struct qht *ht, struct qht_stats *stats)
|
||||
void qht_statistics_init(const struct qht *ht, struct qht_stats *stats)
|
||||
{
|
||||
struct qht_map *map;
|
||||
const struct qht_map *map;
|
||||
int i;
|
||||
|
||||
map = atomic_rcu_read(&ht->map);
|
||||
|
@ -848,8 +918,8 @@ void qht_statistics_init(struct qht *ht, struct qht_stats *stats)
|
|||
stats->head_buckets = map->n_buckets;
|
||||
|
||||
for (i = 0; i < map->n_buckets; i++) {
|
||||
struct qht_bucket *head = &map->buckets[i];
|
||||
struct qht_bucket *b;
|
||||
const struct qht_bucket *head = &map->buckets[i];
|
||||
const struct qht_bucket *b;
|
||||
unsigned int version;
|
||||
size_t buckets;
|
||||
size_t entries;
|
||||
|
|
11
util/qsp.c
11
util/qsp.c
|
@ -533,7 +533,7 @@ static gint qsp_tree_cmp(gconstpointer ap, gconstpointer bp, gpointer up)
|
|||
}
|
||||
}
|
||||
|
||||
static void qsp_sort(struct qht *ht, void *p, uint32_t h, void *userp)
|
||||
static void qsp_sort(void *p, uint32_t h, void *userp)
|
||||
{
|
||||
QSPEntry *e = p;
|
||||
GTree *tree = userp;
|
||||
|
@ -541,7 +541,7 @@ static void qsp_sort(struct qht *ht, void *p, uint32_t h, void *userp)
|
|||
g_tree_insert(tree, e, NULL);
|
||||
}
|
||||
|
||||
static void qsp_aggregate(struct qht *global_ht, void *p, uint32_t h, void *up)
|
||||
static void qsp_aggregate(void *p, uint32_t h, void *up)
|
||||
{
|
||||
struct qht *ht = up;
|
||||
const QSPEntry *e = p;
|
||||
|
@ -553,7 +553,7 @@ static void qsp_aggregate(struct qht *global_ht, void *p, uint32_t h, void *up)
|
|||
qsp_entry_aggregate(agg, e);
|
||||
}
|
||||
|
||||
static void qsp_iter_diff(struct qht *orig, void *p, uint32_t hash, void *htp)
|
||||
static void qsp_iter_diff(void *p, uint32_t hash, void *htp)
|
||||
{
|
||||
struct qht *ht = htp;
|
||||
QSPEntry *old = p;
|
||||
|
@ -583,8 +583,7 @@ static void qsp_diff(struct qht *orig, struct qht *new)
|
|||
qht_iter(orig, qsp_iter_diff, new);
|
||||
}
|
||||
|
||||
static void
|
||||
qsp_iter_callsite_coalesce(struct qht *orig, void *p, uint32_t h, void *htp)
|
||||
static void qsp_iter_callsite_coalesce(void *p, uint32_t h, void *htp)
|
||||
{
|
||||
struct qht *ht = htp;
|
||||
QSPEntry *old = p;
|
||||
|
@ -603,7 +602,7 @@ qsp_iter_callsite_coalesce(struct qht *orig, void *p, uint32_t h, void *htp)
|
|||
e->n_acqs += old->n_acqs;
|
||||
}
|
||||
|
||||
static void qsp_ht_delete(struct qht *ht, void *p, uint32_t h, void *htp)
|
||||
static void qsp_ht_delete(void *p, uint32_t h, void *htp)
|
||||
{
|
||||
g_free(p);
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue