/*
* %CopyrightBegin%
*
* Copyright Ericsson AB 1999-2018. All Rights Reserved.
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*
* %CopyrightEnd%
*/
/*
* Code for process dictionaries.
*
*/
#ifdef HAVE_CONFIG_H
# include "config.h"
#endif
#include "sys.h"
#include "erl_vm.h"
#include "global.h"
#include "erl_process.h" /* Will include erl_process_dict.h */
#include "error.h"
#include "erl_driver.h"
#include "bif.h"
#include "big.h"
#include "dist.h"
#include "erl_version.h"
/* #define HARDDEBUG */
/*
** Utility macros
*/
/* Flags to pd_get_hash */
#define PD_GET_OTHER_PROCESS 1UL
/* Hash constant macros */
#define MAX_HASH 1342177280UL
#define INITIAL_SIZE (erts_pd_initial_size)
/* Hash utility macros */
#define HASH_RANGE(PDict) ((PDict)->usedSlots)
#define MAKE_HASH(Term) \
((is_small(Term)) ? (Uint32) unsigned_val(Term) : \
((is_atom(Term)) ? \
(Uint32) atom_val(Term) : \
make_internal_hash(Term, 0)))
#define PD_SZ2BYTES(Sz) (sizeof(ProcDict) + ((Sz) - 1)*sizeof(Eterm))
#define pd_hash_value(Pdict, Key) \
pd_hash_value_to_ix(Pdict, MAKE_HASH((Key)))
/* Memory allocation macros */
#define PD_ALLOC(Sz) \
erts_alloc(ERTS_ALC_T_PROC_DICT, (Sz))
#define PD_FREE(P, Sz) \
erts_free(ERTS_ALC_T_PROC_DICT, (P))
#define PD_REALLOC(P, OSz, NSz) \
erts_realloc(ERTS_ALC_T_PROC_DICT, (P), (NSz))
#define TCAR(Term) CAR(list_val(Term))
#define TCDR(Term) CDR(list_val(Term))
/* Array access macro */
#define ARRAY_GET(PDict, Index) (ASSERT((Index) < (PDict)->arraySize), \
(PDict)->data[Index])
#define ARRAY_GET_PTR(PDict, Index) (ASSERT((Index) < (PDict)->arraySize), \
&(PDict)->data[Index])
#define ARRAY_PUT(PDict, Index, Val) (ASSERT((Index) < (PDict)->arraySize), \
(PDict)->data[Index] = (Val))
#define IS_POW2(X) ((X) && !((X) & ((X)-1)))
/*
* Forward decalarations
*/
static void pd_hash_erase(Process *p, Eterm id, Eterm *ret);
static void pd_hash_erase_all(Process *p);
static Eterm pd_hash_get_with_hval(Process *p, Eterm bucket, Eterm id);
static Eterm pd_hash_get_keys(Process *p, Eterm value);
static Eterm pd_hash_get_all_keys(Process *p, ProcDict *pd);
static Eterm pd_hash_get_all(Process *p, ProcDict *pd, int keep_dict);
static Eterm pd_hash_put(Process *p, Eterm id, Eterm value);
static void shrink(Process *p, Eterm* ret);
static void grow(Process *p);
static void array_shrink(ProcDict **ppd, unsigned int need);
static void ensure_array_size(ProcDict**, unsigned int size);
static unsigned int pd_hash_value_to_ix(ProcDict *pdict, Uint32 hx);
static unsigned int next_array_size(unsigned int need);
/*
** Debugging prototypes and macros
*/
#ifdef HARDDEBUG
#include <stdarg.h>
static int hdebugf(char *format, ...);
static char *hdebugf_file = "";
static int hdebugf_line;
#define HDEBUGF(X) ((int) hdebugf_file = __FILE__, hdebugf_line = __LINE__, \
hdebugf X)
#ifndef DEBUG
#define DEBUG 1
#endif
#else /* !HARDDEBUG */
#define HDEBUGF(X) /* Nothing */
#endif /* HARDDEBUG (else) */
#ifdef DEBUG
static void pd_check(ProcDict *pd);
#define PD_CHECK(PD) pd_check(PD)
#else /* !DEBUG */
#define PD_CHECK(PD) /* Nothing */
#endif /* DEBUG (else) */
/*
** External interface
*/
int
erts_pd_set_initial_size(int size)
{
if (size <= 0)
return 0;
erts_pd_initial_size = 1 << erts_fit_in_bits_uint(size-1);
return 1;
}
/*
* Called from break handler
*/
void
erts_dictionary_dump(fmtfn_t to, void *to_arg, ProcDict *pd)
{
unsigned int i;
#ifdef DEBUG
/*PD_CHECK(pd);*/
if (pd == NULL)
return;
erts_print(to, to_arg, "(size = %d, usedSlots = %d, "
"splitPosition = %d, numElements = %d)\n",
pd->arraySize, pd->usedSlots,
pd->splitPosition, (unsigned int) pd->numElements);
for (i = 0; i < HASH_RANGE(pd); ++i) {
erts_print(to, to_arg, "%d: %T\n", i, ARRAY_GET(pd, i));
}
#else /* !DEBUG */
int written = 0;
Eterm t;
erts_print(to, to_arg, "[");
if (pd != NULL) {
for (i = 0; i < HASH_RANGE(pd); ++i) {
t = ARRAY_GET(pd, i);
if (is_list(t)) {
for (; t != NIL; t = TCDR(t)) {
erts_print(to, to_arg, written++ ? ",%T" : "%T", TCAR(t));
}
} else if (is_tuple(t)) {
erts_print(to, to_arg, written++ ? ",%T" : "%T", t);
}
}
}
erts_print(to, to_arg, "]");
#endif /* DEBUG (else) */
}
void
erts_deep_dictionary_dump(fmtfn_t to, void *to_arg,
ProcDict* pd, void (*cb)(fmtfn_t, void *, Eterm))
{
unsigned int i;
Eterm t;
if (pd != NULL) {
for (i = 0; i < HASH_RANGE(pd); ++i) {
t = ARRAY_GET(pd, i);
if (is_list(t)) {
for (; t != NIL; t = TCDR(t)) {
(*cb)(to, to_arg, TCAR(t));
}
} else if (is_tuple(t)) {
(*cb)(to, to_arg, t);
}
}
}
}
Uint
erts_dicts_mem_size(Process *p)
{
Uint size = 0;
if (p->dictionary)
size += PD_SZ2BYTES(p->dictionary->arraySize);
return size;
}
void
erts_erase_dicts(Process *p)
{
if (p->dictionary) {
pd_hash_erase_all(p);
p->dictionary = NULL;
}
}
/*
* Called from process_info/1,2.
*/
Eterm erts_dictionary_copy(ErtsHeapFactory *hfact, ProcDict *pd, Uint reserve_size)
{
Eterm res;
unsigned int i, num;
Uint *sz;
Uint szi, rsz = reserve_size;
if (pd == NULL)
return NIL;
PD_CHECK(pd);
num = HASH_RANGE(pd);
sz = (Uint *) erts_alloc(ERTS_ALC_T_TMP, sizeof(Uint) * pd->numElements);
for (i = 0, szi = 0; i < num; ++i) {
Eterm tmp = ARRAY_GET(pd, i);
if (is_boxed(tmp)) {
Uint size;
ASSERT(is_tuple(tmp));
size = size_object(tmp) + 2;
sz[szi++] = size;
rsz += size;
}
else if (is_list(tmp)) {
while (tmp != NIL) {
Uint size = size_object(TCAR(tmp)) + 2;
sz[szi++] = size;
rsz += size;
tmp = TCDR(tmp);
}
}
}
res = NIL;
for (i = 0, szi = 0; i < num; ++i) {
Eterm tmp = ARRAY_GET(pd, i);
if (is_boxed(tmp)) {
Uint size;
Eterm el, *hp;
ASSERT(is_tuple(tmp));
size = sz[szi++];
rsz -= size;
hp = erts_produce_heap(hfact, size, rsz);
el = copy_struct(tmp, size-2, &hp, hfact->off_heap);
res = CONS(hp, el, res);
}
else if (is_list(tmp)) {
while (tmp != NIL) {
Uint size = sz[szi++];
Eterm el, *hp;
rsz -= size;
hp = erts_produce_heap(hfact, size, rsz);
el = copy_struct(TCAR(tmp), size-2, &hp, hfact->off_heap);
res = CONS(hp, el, res);
tmp = TCDR(tmp);
}
}
}
ASSERT(rsz == reserve_size);
erts_free(ERTS_ALC_T_TMP, sz);
return res;
}
/*
** BIF interface
*/
BIF_RETTYPE get_0(BIF_ALIST_0)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = pd_hash_get_all(BIF_P, BIF_P->dictionary, 1);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE get_1(BIF_ALIST_1)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = erts_pd_hash_get(BIF_P, BIF_ARG_1);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE get_keys_0(BIF_ALIST_0)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = pd_hash_get_all_keys(BIF_P,BIF_P->dictionary);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE get_keys_1(BIF_ALIST_1)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = pd_hash_get_keys(BIF_P, BIF_ARG_1);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE put_2(BIF_ALIST_2)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = pd_hash_put(BIF_P, BIF_ARG_1, BIF_ARG_2);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE erase_0(BIF_ALIST_0)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
ret = pd_hash_get_all(BIF_P, BIF_P->dictionary, 0);
pd_hash_erase_all(BIF_P);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
BIF_RETTYPE erase_1(BIF_ALIST_1)
{
Eterm ret;
PD_CHECK(BIF_P->dictionary);
pd_hash_erase(BIF_P, BIF_ARG_1, &ret);
PD_CHECK(BIF_P->dictionary);
BIF_RET(ret);
}
/*
* BIF implementations
*/
static void pd_hash_erase(Process *p, Eterm id, Eterm *ret)
{
unsigned int hval;
Eterm old;
Eterm tmp;
unsigned int range;
*ret = am_undefined;
if (p->dictionary == NULL) {
return;
}
hval = pd_hash_value(p->dictionary, id);
old = ARRAY_GET(p->dictionary, hval);
if (is_boxed(old)) { /* Tuple */
ASSERT(is_tuple(old));
if (EQ(tuple_val(old)[1], id)) {
ARRAY_PUT(p->dictionary, hval, NIL);
--(p->dictionary->numElements);
*ret = tuple_val(old)[2];
}
} else if (is_list(old)) {
/* Find cons cell for identical value */
Eterm* prev = &p->dictionary->data[hval];
for (tmp = *prev; tmp != NIL; prev = &TCDR(tmp), tmp = *prev) {
if (EQ(tuple_val(TCAR(tmp))[1], id)) {
*prev = TCDR(tmp);
*ret = tuple_val(TCAR(tmp))[2];
--(p->dictionary->numElements);
}
}
/* If there is only one element left in the list we must remove the list. */
old = ARRAY_GET(p->dictionary, hval);
ASSERT(is_list(old));
if (is_nil(TCDR(old))) {
ARRAY_PUT(p->dictionary, hval, TCAR(old));
}
} else if (is_not_nil(old)) {
#ifdef DEBUG
erts_fprintf(stderr,
"Process dictionary for process %T is broken, trying to "
"display term found in line %d:\n"
"%T\n", p->common.id, __LINE__, old);
#endif
erts_exit(ERTS_ERROR_EXIT, "Damaged process dictionary found during erase/1.");
}
if ((range = HASH_RANGE(p->dictionary)) > INITIAL_SIZE &&
range / 2 > (p->dictionary->numElements)) {
shrink(p, ret);
}
}
static void pd_hash_erase_all(Process *p)
{
if (p->dictionary != NULL) {
PD_FREE(p->dictionary, PD_SZ2BYTES(p->dictionary->size));
p->dictionary = NULL;
}
}
Uint32 erts_pd_make_hx(Eterm key)
{
return MAKE_HASH(key);
}
Eterm erts_pd_hash_get_with_hx(Process *p, Uint32 hx, Eterm id)
{
unsigned int hval;
ProcDict *pd = p->dictionary;
ASSERT(hx == MAKE_HASH(id));
if (pd == NULL)
return am_undefined;
hval = pd_hash_value_to_ix(pd, hx);
return pd_hash_get_with_hval(p, ARRAY_GET(pd, hval), id);
}
Eterm erts_pd_hash_get(Process *p, Eterm id)
{
unsigned int hval;
ProcDict *pd = p->dictionary;
if (pd == NULL)
return am_undefined;
hval = pd_hash_value(pd, id);
return pd_hash_get_with_hval(p, ARRAY_GET(pd, hval), id);
}
Eterm pd_hash_get_with_hval(Process *p, Eterm bucket, Eterm id)
{
if (is_boxed(bucket)) { /* Tuple */
ASSERT(is_tuple(bucket));
if (EQ(tuple_val(bucket)[1], id)) {
return tuple_val(bucket)[2];
}
} else if (is_list(bucket)) {
for (; bucket != NIL && !EQ(tuple_val(TCAR(bucket))[1], id); bucket = TCDR(bucket)) {
;
}
if (bucket != NIL) {
return tuple_val(TCAR(bucket))[2];
}
} else if (is_not_nil(bucket)) {
#ifdef DEBUG
erts_fprintf(stderr,
"Process dictionary for process %T is broken, trying to "
"display term found in line %d:\n"
"%T\n", p->common.id, __LINE__, bucket);
#endif
erts_exit(ERTS_ERROR_EXIT, "Damaged process dictionary found during get/1.");
}
return am_undefined;
}
#define PD_GET_TKEY(Dst,Src) \
do { \
ASSERT(is_tuple((Src))); \
ASSERT(arityval(*((Eterm*)tuple_val((Src)))) == 2); \
(Dst) = ((Eterm*)tuple_val((Src)))[1]; \
} while(0)
static Eterm pd_hash_get_all_keys(Process *p, ProcDict *pd) {
Eterm* hp;
Eterm res = NIL;
Eterm tmp, tmp2;
unsigned int i;
unsigned int num;
if (pd == NULL) {
return res;
}
num = HASH_RANGE(pd);
hp = HAlloc(p, pd->numElements * 2);
for (i = 0; i < num; ++i) {
tmp = ARRAY_GET(pd, i);
if (is_boxed(tmp)) {
PD_GET_TKEY(tmp,tmp);
res = CONS(hp, tmp, res);
hp += 2;
} else if (is_list(tmp)) {
while (tmp != NIL) {
tmp2 = TCAR(tmp);
PD_GET_TKEY(tmp2,tmp2);
res = CONS(hp, tmp2, res);
hp += 2;
tmp = TCDR(tmp);
}
}
}
return res;
}
#undef PD_GET_TKEY
static Eterm pd_hash_get_keys(Process *p, Eterm value)
{
Eterm *hp;
Eterm res = NIL;
ProcDict *pd = p->dictionary;
unsigned int i, num;
Eterm tmp, tmp2;
if (pd == NULL) {
return res;
}
num = HASH_RANGE(pd);
for (i = 0; i < num; ++i) {
tmp = ARRAY_GET(pd, i);
if (is_boxed(tmp)) {
ASSERT(is_tuple(tmp));
if (EQ(tuple_val(tmp)[2], value)) {
hp = HAlloc(p, 2);
res = CONS(hp, tuple_val(tmp)[1], res);
}
} else if (is_list(tmp)) {
while (tmp != NIL) {
tmp2 = TCAR(tmp);
if (EQ(tuple_val(tmp2)[2], value)) {
hp = HAlloc(p, 2);
res = CONS(hp, tuple_val(tmp2)[1], res);
}
tmp = TCDR(tmp);
}
}
}
return res;
}
static Eterm
pd_hash_get_all(Process *p, ProcDict *pd, int keep_dict)
{
Eterm* hp;
Eterm* tp;
Eterm res = NIL;
Eterm tmp, tmp2;
unsigned int i;
unsigned int num;
Uint need;
if (pd == NULL) {
return res;
}
num = HASH_RANGE(pd);
/*
* If this is not erase/0, then must copy all key-value tuples
* as they may be mutated by put/2.
*/
need = pd->numElements * (keep_dict ? 2+3 : 2);
hp = HAlloc(p, need);
for (i = 0; i < num; ++i) {
tmp = ARRAY_GET(pd, i);
if (is_boxed(tmp)) {
if (keep_dict) {
tp = tuple_val(tmp);
tmp = TUPLE2(hp, tp[1], tp[2]);
hp += 3;
}
res = CONS(hp, tmp, res);
hp += 2;
} else if (is_list(tmp)) {
while (tmp != NIL) {
tmp2 = TCAR(tmp);
if (keep_dict) {
tp = tuple_val(tmp2);
tmp2 = TUPLE2(hp, tp[1], tp[2]);
hp += 3;
}
res = CONS(hp, tmp2, res);
hp += 2;
tmp = TCDR(tmp);
}
}
}
return res;
}
static Eterm pd_hash_put(Process *p, Eterm id, Eterm value)
{
unsigned int hval;
Eterm *hp;
Eterm *tp;
Eterm *bucket;
Eterm tpl;
Eterm old;
Eterm old_val = am_undefined;
Eterm tmp;
int needed;
int new_key = 1;
#ifdef DEBUG
Eterm *hp_limit;
#endif
if (p->dictionary == NULL) {
/* Create it */
ensure_array_size(&p->dictionary, INITIAL_SIZE);
p->dictionary->usedSlots = INITIAL_SIZE;
p->dictionary->sizeMask = INITIAL_SIZE*2 - 1;
p->dictionary->splitPosition = 0;
p->dictionary->numElements = 0;
}
hval = pd_hash_value(p->dictionary, id);
bucket = ARRAY_GET_PTR(p->dictionary, hval);
old = *bucket;
/*
* Calculate the number of heap words needed and garbage
* collect if necessary. (Might be a slight overestimation.)
*/
needed = 3; /* {Key,Value} tuple */
if (is_boxed(old)) {
ASSERT(is_tuple(old));
tp = tuple_val(old);
if (EQ(tp[1], id)) {
old_val = tp[2];
if (is_immed(value)) {
tp[2] = value; /* DESTRUCTIVE HEAP ASSIGNMENT */
return old_val;
}
new_key = 0;
}
else {
needed += 2+2;
}
} else if (is_list(old)) {
Eterm* prev_cdr = bucket;
needed += 2;
for (tmp = old; tmp != NIL; prev_cdr = &TCDR(tmp), tmp = *prev_cdr) {
tp = tuple_val(TCAR(tmp));
if (EQ(tp[1], id)) {
old_val = tp[2];
if (is_immed(value)) {
tp[2] = value; /* DESTRUCTIVE HEAP ASSIGNMENT */
return old_val;
}
new_key = 0;
/* Unlink old {Key,Value} from list */
*prev_cdr = TCDR(tmp); /* maybe DESTRUCTIVE HEAP ASSIGNMENT */
break;
}
}
}
if (HeapWordsLeft(p) < needed) {
Eterm root[3];
root[0] = id;
root[1] = value;
root[2] = old_val;
erts_garbage_collect(p, needed, root, 3);
id = root[0];
value = root[1];
old_val = root[2];
ASSERT(bucket == ARRAY_GET_PTR(p->dictionary, hval));
old = *bucket;
}
#ifdef DEBUG
hp_limit = p->htop + needed;
#endif
/*
* Create the {Key,Value} tuple.
*/
hp = HeapOnlyAlloc(p, 3);
tpl = TUPLE2(hp, id, value);
/*
* Update the dictionary.
*/
if (is_nil(old)) {
*bucket = tpl;
} else if (is_boxed(old)) {
ASSERT(is_tuple(old));
if (!new_key) {
ASSERT(EQ(tuple_val(old)[1],id));
*bucket = tpl;
return old_val;
} else {
hp = HeapOnlyAlloc(p, 4);
tmp = CONS(hp, old, NIL);
hp += 2;
*bucket = CONS(hp, tpl, tmp);
hp += 2;
ASSERT(hp <= hp_limit);
}
} else if (is_list(old)) {
/*
* Simply prepend the tuple to the beginning of the list.
*/
hp = HeapOnlyAlloc(p, 2);
*bucket = CONS(hp, tpl, *bucket);
hp += 2;
ASSERT(hp <= hp_limit);
} else {
#ifdef DEBUG
erts_fprintf(stderr,
"Process dictionary for process %T is broken, trying to "
"display term found in line %d:\n"
"%T\n", p->common.id, __LINE__, old);
#endif
erts_exit(ERTS_ERROR_EXIT, "Damaged process dictionary found during put/2.");
}
p->dictionary->numElements += new_key;
if (HASH_RANGE(p->dictionary) <= p->dictionary->numElements) {
grow(p);
}
return old_val;
}
/*
* Hash table utilities, rehashing
*/
static void shrink(Process *p, Eterm* ret)
{
ProcDict *pd = p->dictionary;
unsigned int range = HASH_RANGE(p->dictionary);
unsigned int steps = (range*3) / 10;
Eterm hi, lo, tmp;
unsigned int i;
Eterm *hp;
#ifdef DEBUG
Eterm *hp_limit;
#endif
if (range - steps < INITIAL_SIZE) {
steps = range - INITIAL_SIZE;
}
for (i = 0; i < steps; ++i) {
if (pd->splitPosition == 0) {
ASSERT(IS_POW2(pd->usedSlots));
pd->sizeMask = pd->usedSlots - 1;
pd->splitPosition = pd->usedSlots / 2;
}
--(pd->splitPosition);
/* Must wait to decrement 'usedSlots' for GC rootset below */
hi = ARRAY_GET(pd, pd->usedSlots - 1);
lo = ARRAY_GET(pd, pd->splitPosition);
if (hi != NIL) {
if (lo == NIL) {
ARRAY_PUT(pd, pd->splitPosition, hi);
} else {
Sint needed = 4;
if (is_list(hi) && is_list(lo)) {
needed = 2*erts_list_length(hi);
}
if (HeapWordsLeft(p) < needed) {
erts_garbage_collect(p, needed, ret, 1);
hi = pd->data[pd->usedSlots - 1];
lo = pd->data[pd->splitPosition];
}
#ifdef DEBUG
hp_limit = p->htop + needed;
#endif
if (is_tuple(lo)) {
if (is_tuple(hi)) {
hp = HeapOnlyAlloc(p, 4);
tmp = CONS(hp, hi, NIL);
hp += 2;
ARRAY_PUT(pd, pd->splitPosition,
CONS(hp,lo,tmp));
hp += 2;
ASSERT(hp <= hp_limit);
} else { /* hi is a list */
hp = HeapOnlyAlloc(p, 2);
ARRAY_PUT(pd, pd->splitPosition,
CONS(hp, lo, hi));
hp += 2;
ASSERT(hp <= hp_limit);
}
} else { /* lo is a list */
if (is_tuple(hi)) {
hp = HeapOnlyAlloc(p, 2);
ARRAY_PUT(pd, pd->splitPosition,
CONS(hp, hi, lo));
hp += 2;
ASSERT(hp <= hp_limit);
} else { /* Two lists */
hp = HeapOnlyAlloc(p, needed);
for (tmp = hi; tmp != NIL; tmp = TCDR(tmp)) {
lo = CONS(hp, TCAR(tmp), lo);
hp += 2;
}
ASSERT(hp <= hp_limit);
ARRAY_PUT(pd, pd->splitPosition, lo);
}
}
}
}
--pd->usedSlots;
ARRAY_PUT(pd, pd->usedSlots, NIL);
}
if (HASH_RANGE(p->dictionary) <= (p->dictionary->arraySize / 4)) {
array_shrink(&(p->dictionary), (HASH_RANGE(p->dictionary) * 3) / 2);
}
}
static void grow(Process *p)
{
unsigned int i,j;
unsigned int steps = (p->dictionary->usedSlots / 4) & 0xf;
Eterm l1,l2;
Eterm l;
Eterm *hp;
unsigned int pos;
unsigned int homeSize;
Sint needed = 0;
ProcDict *pd = p->dictionary;
#ifdef DEBUG
Eterm *hp_limit;
#endif
HDEBUGF(("grow: steps = %d", steps));
if (steps == 0)
steps = 1;
/* Dont grow over MAX_HASH */
if ((MAX_HASH - steps) <= HASH_RANGE(pd)) {
return;
}
ensure_array_size(&p->dictionary, HASH_RANGE(pd) + steps);
pd = p->dictionary;
/*
* Calculate total number of heap words needed, and garbage collect
* if necessary.
*/
pos = pd->splitPosition;
homeSize = pd->usedSlots - pd->splitPosition;
for (i = 0; i < steps; ++i) {
if (pos == homeSize) {
homeSize *= 2;
pos = 0;
}
l = ARRAY_GET(pd, pos);
pos++;
if (is_not_tuple(l)) {
while (l != NIL) {
needed += 2;
l = TCDR(l);
}
}
}
if (HeapWordsLeft(p) < needed) {
erts_garbage_collect(p, needed, 0, 0);
}
#ifdef DEBUG
hp_limit = p->htop + needed;
#endif
/*
* Now grow.
*/
homeSize = pd->usedSlots - pd->splitPosition;
for (i = 0; i < steps; ++i) {
if (pd->splitPosition == homeSize) {
homeSize *= 2;
pd->sizeMask = homeSize*2 - 1;
pd->splitPosition = 0;
}
pos = pd->splitPosition;
++pd->splitPosition; /* For the hashes */
++pd->usedSlots;
ASSERT(pos + homeSize == pd->usedSlots - 1);
l = ARRAY_GET(pd, pos);
if (is_tuple(l)) {
if (pd_hash_value(pd, tuple_val(l)[1]) != pos) {
ARRAY_PUT(pd, pos + homeSize, l);
ARRAY_PUT(pd, pos, NIL);
}
} else {
l2 = NIL;
l1 = l;
for (j = 0; l1 != NIL; l1 = TCDR(l1))
j += 2;
hp = HeapOnlyAlloc(p, j);
while (l != NIL) {
if (pd_hash_value(pd, tuple_val(TCAR(l))[1]) == pos)
l1 = CONS(hp, TCAR(l), l1);
else
l2 = CONS(hp, TCAR(l), l2);
hp += 2;
l = TCDR(l);
}
if (l1 != NIL && TCDR(l1) == NIL)
l1 = TCAR(l1);
if (l2 != NIL && TCDR(l2) == NIL)
l2 = TCAR(l2);
ASSERT(hp <= hp_limit);
ARRAY_PUT(pd, pos, l1);
ARRAY_PUT(pd, pos + homeSize, l2);
}
}
#ifdef HARDDEBUG
dictionary_dump(p->dictionary,CERR);
#endif
}
/*
** Array oriented operations
*/
static void array_shrink(ProcDict **ppd, unsigned int need)
{
unsigned int siz = next_array_size(need);
HDEBUGF(("array_shrink: size = %d, need = %d",
(*ppd)->arraySize, need));
if (siz >= (*ppd)->arraySize)
return; /* Only shrink */
*ppd = PD_REALLOC(((void *) *ppd),
PD_SZ2BYTES((*ppd)->arraySize),
PD_SZ2BYTES(siz));
(*ppd)->arraySize = siz;
}
static void ensure_array_size(ProcDict **ppdict, unsigned int size)
{
ProcDict *pd = *ppdict;
unsigned int i;
if (pd == NULL) {
Uint siz = next_array_size(size);
pd = PD_ALLOC(PD_SZ2BYTES(siz));
for (i = 0; i < siz; ++i)
pd->data[i] = NIL;
pd->arraySize = siz;
*ppdict = pd;
} else if (size > pd->arraySize) {
Uint osize = pd->arraySize;
Uint nsize = next_array_size(size);
pd = PD_REALLOC(((void *) pd),
PD_SZ2BYTES(osize),
PD_SZ2BYTES(nsize));
for (i = osize; i < nsize; ++i)
pd->data[i] = NIL;
pd->arraySize = nsize;
*ppdict = pd;
}
}
/*
** Basic utilities
*/
static unsigned int pd_hash_value_to_ix(ProcDict *pdict, Uint32 hx)
{
Uint high;
ASSERT(IS_POW2(pdict->sizeMask+1));
ASSERT(HASH_RANGE(pdict) >= (pdict->sizeMask >> 1));
ASSERT(HASH_RANGE(pdict) <= (pdict->sizeMask + 1));
high = hx & pdict->sizeMask;
if (high >= HASH_RANGE(pdict))
return hx & (pdict->sizeMask >> 1);
return high;
}
static unsigned int next_array_size(unsigned int need)
{
static unsigned int tab[] =
{
10UL,
20UL,
40UL,
80UL,
160UL,
320UL,
640UL,
1280UL,
2560UL,
5120UL,
10240UL,
20480UL,
40960UL,
81920UL,
163840UL,
327680UL,
655360UL,
1310720UL,
2621440UL,
5242880UL,
10485760UL,
20971520UL,
41943040UL,
83886080UL,
167772160UL,
335544320UL,
671088640UL,
1342177280UL,
2684354560UL
};
int hi = sizeof(tab) / sizeof(Uint) - 1;
int lo = 1;
int cur = 4;
while (hi >= lo) {
if (tab[cur] >= need && tab[cur - 1] < need)
return tab[cur];
if (tab[cur] > need)
hi = cur - 1;
else
lo = cur + 1;
cur = (hi + lo) / 2;
}
return need;
}
/*
** Debug functions
*/
#ifdef DEBUG
static void pd_check(ProcDict *pd)
{
unsigned int i;
unsigned int used;
Uint num;
if (pd == NULL)
return;
used = HASH_RANGE(pd);
ASSERT(pd->arraySize >= used);
ASSERT(HASH_RANGE(pd) <= MAX_HASH);
for (i = 0, num = 0; i < used; ++i) {
Eterm t = pd->data[i];
if (is_nil(t)) {
continue;
} else if (is_tuple(t)) {
++num;
ASSERT(arityval(*tuple_val(t)) == 2);
ASSERT(pd_hash_value(pd, tuple_val(t)[1]) == i);
continue;
} else if (is_list(t)) {
while (t != NIL) {
++num;
ASSERT(is_tuple(TCAR(t)));
ASSERT(arityval(*(tuple_val(TCAR(t)))) == 2);
ASSERT(pd_hash_value(pd, tuple_val(TCAR(t))[1]) == i);
t = TCDR(t);
}
continue;
} else {
erts_exit(ERTS_ERROR_EXIT,
"Found tag 0x%08x in process dictionary at position %d",
(unsigned long) t, (int) i);
}
}
ASSERT(num == pd->numElements);
ASSERT(pd->usedSlots >= pd->splitPosition*2);
}
#endif /* DEBUG */
#ifdef HARDDEBUG
static int hdebugf(char *format, ...)
{
va_list ap;
erts_fprintf(stderr, "DEBUG: %s:%d :", hdebugf_file, hdebugf_line);
va_start(ap, format);
erts_vfprintf(stderr, format, ap);
va_end(ap);
erts_fprintf(stderr, "\n");
return 0;
}
#endif /* HARDDEBUG */