Improve memory usage

By avoiding unnecessary fragmentation in indexes.

With a larger set of dictionaries, I've observed a ~20% improvement.
This commit is contained in:
Přemysl Eric Janouch 2016-10-07 17:28:21 +02:00
parent a1b7e84c6a
commit 0d779d718b
Signed by: p
GPG Key ID: B715679E3A361BE6

View File

@ -374,6 +374,7 @@ struct stardict_dict_private
StardictInfo * info; //!< General information about the dict
GArray * index; //!< Word index
GArray * synonyms; //!< Synonyms
GStringChunk * string_allocator; //!< String allocator (index+synonyms)
// The collated indexes are only permutations of their normal selves.
@ -405,6 +406,7 @@ stardict_dict_finalize (GObject *self)
g_array_free (priv->index, TRUE);
g_array_free (priv->synonyms, TRUE);
g_string_chunk_free (priv->string_allocator);
if (priv->collator)
ucol_close (priv->collator);
@ -475,7 +477,8 @@ load_idx_internal (StardictDict *sd, GInputStream *is, GError **error)
StardictIndexEntry entry;
GError *err = NULL;
// Ignoring "wordcount", just reading as long as we can
while ((entry.name = stream_read_string (dis, &err)))
gchar *name;
while ((name = stream_read_string (dis, &err)))
{
if (priv->info->idx_offset_bits == 32)
entry.data_offset
@ -490,7 +493,9 @@ load_idx_internal (StardictDict *sd, GInputStream *is, GError **error)
if (err)
goto error;
entry.name = g_string_chunk_insert (sd->priv->string_allocator, name);
g_array_append_val (priv->index, entry);
g_free (name);
}
if (err != NULL)
@ -501,7 +506,7 @@ load_idx_internal (StardictDict *sd, GInputStream *is, GError **error)
error:
g_propagate_error (error, err);
g_free (entry.name);
g_free (name);
g_object_unref (dis);
return FALSE;
}
@ -556,18 +561,21 @@ load_syn (StardictDict *sd, const gchar *filename, GError **error)
StardictSynonymEntry entry;
GError *err = NULL;
// Ignoring "synwordcount", just reading as long as we can
while ((entry.word = stream_read_string (dis, &err)))
gchar *word;
while ((word = stream_read_string (dis, &err)))
{
entry.original_word = g_data_input_stream_read_uint32 (dis, NULL, &err);
if (err)
break;
entry.word = g_string_chunk_insert (sd->priv->string_allocator, word);
g_array_append_val (sd->priv->synonyms, entry);
g_free (word);
}
if (err != NULL)
{
g_free (entry.word);
g_free (word);
g_propagate_error (error, err);
}
else
@ -580,22 +588,6 @@ cannot_open:
return ret_val;
}
/// Destroy an index entry.
static void
index_destroy_cb (gpointer sde)
{
StardictIndexEntry *e = sde;
g_free (e->name);
}
/// Destroy a synonym entry.
static void
syn_destroy_cb (gpointer sde)
{
StardictSynonymEntry *e = sde;
g_free (e->word);
}
/// Load StarDict dictionary data.
static gboolean
load_dict (StardictDict *sd, const gchar *filename, gboolean gzipped,
@ -786,9 +778,8 @@ stardict_dict_new_from_info (StardictInfo *sdi, GError **error)
StardictDictPrivate *priv = sd->priv;
priv->info = sdi;
priv->index = g_array_new (FALSE, FALSE, sizeof (StardictIndexEntry));
g_array_set_clear_func (priv->index, index_destroy_cb);
priv->synonyms = g_array_new (FALSE, FALSE, sizeof (StardictSynonymEntry));
g_array_set_clear_func (priv->synonyms, syn_destroy_cb);
priv->string_allocator = g_string_chunk_new ((1 << 15));
const gchar *dot = strrchr (sdi->path, '.');
gchar *base = dot ? g_strndup (sdi->path, dot - sdi->path)