blob: da0d6cd341501edfc13418894690ac87f3d07757 [file] [log] [blame]
/*
* Copyright (C) 2009 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#ifndef PINYINIME_INCLUDE_DICTBUILDER_H__
#define PINYINIME_INCLUDE_DICTBUILDER_H__
#include <stdlib.h>
#include "./utf16char.h"
#include "./dictdef.h"
#include "./dictlist.h"
#include "./spellingtable.h"
#include "./spellingtrie.h"
#include "./splparser.h"
namespace ime_pinyin {
#ifdef ___BUILD_MODEL___
#define ___DO_STATISTICS___
class DictTrie;
class DictBuilder {
private:
// The raw lemma array buffer.
LemmaEntry *lemma_arr_;
size_t lemma_num_;
// Used to store all possible single char items.
// Two items may have the same Hanzi while their spelling ids are different.
SingleCharItem *scis_;
size_t scis_num_;
// In the tree, root's level is -1.
// Lemma nodes for root, and level 0
LmaNodeLE0 *lma_nodes_le0_;
// Lemma nodes for layers whose levels are deeper than 0
LmaNodeGE1 *lma_nodes_ge1_;
// Number of used lemma nodes
size_t lma_nds_used_num_le0_;
size_t lma_nds_used_num_ge1_;
// Used to store homophonies' ids.
LemmaIdType *homo_idx_buf_;
// Number of homophonies each of which only contains one Chinese character.
size_t homo_idx_num_eq1_;
// Number of homophonies each of which contains more than one character.
size_t homo_idx_num_gt1_;
// The items with highest scores.
LemmaEntry *top_lmas_;
size_t top_lmas_num_;
SpellingTable *spl_table_;
SpellingParser *spl_parser_;
#ifdef ___DO_STATISTICS___
size_t max_sonbuf_len_[kMaxLemmaSize];
size_t max_homobuf_len_[kMaxLemmaSize];
size_t total_son_num_[kMaxLemmaSize];
size_t total_node_hasson_[kMaxLemmaSize];
size_t total_sonbuf_num_[kMaxLemmaSize];
size_t total_sonbuf_allnoson_[kMaxLemmaSize];
size_t total_node_in_sonbuf_allnoson_[kMaxLemmaSize];
size_t total_homo_num_[kMaxLemmaSize];
size_t sonbufs_num1_; // Number of son buffer with only 1 son
size_t sonbufs_numgt1_; // Number of son buffer with more 1 son;
size_t total_lma_node_num_;
void stat_init();
void stat_print();
#endif
public:
DictBuilder();
~DictBuilder();
// Build dictionary trie from the file fn_raw. File fn_validhzs provides
// valid chars. If fn_validhzs is NULL, only chars in GB2312 will be
// included.
bool build_dict(const char* fn_raw, const char* fn_validhzs,
DictTrie *dict_trie);
private:
// Fill in the buffer with id. The caller guarantees that the paramters are
// vaild.
void id_to_charbuf(unsigned char *buf, LemmaIdType id);
// Update the offset of sons for a node.
void set_son_offset(LmaNodeGE1 *node, size_t offset);
// Update the offset of homophonies' ids for a node.
void set_homo_id_buf_offset(LmaNodeGE1 *node, size_t offset);
// Format a speling string.
void format_spelling_str(char *spl_str);
// Sort the lemma_arr by the hanzi string, and give each of unique items
// a id. Why we need to sort the lemma list according to their Hanzi string
// is to find items started by a given prefix string to do prediction.
// Actually, the single char items are be in other order, for example,
// in spelling id order, etc.
// Return value is next un-allocated idx available.
LemmaIdType sort_lemmas_by_hz();
// Build the SingleCharItem list, and fill the hanzi_scis_ids in the
// lemma buffer lemma_arr_.
// This function should be called after the lemma array is ready.
// Return the number of unique SingleCharItem elements.
size_t build_scis();
// Construct a subtree using a subset of the spelling array (from
// item_star to item_end)
// parent is the parent node to update the necessary information
// parent can be a member of LmaNodeLE0 or LmaNodeGE1
bool construct_subset(void* parent, LemmaEntry* lemma_arr,
size_t item_start, size_t item_end, size_t level);
// Read valid Chinese Hanzis from the given file.
// num is used to return number of chars.
// The return buffer is sorted and caller needs to free the returned buffer.
char16* read_valid_hanzis(const char *fn_validhzs, size_t *num);
// Read a raw dictionary. max_item is the maximum number of items. If there
// are more items in the ditionary, only the first max_item will be read.
// Returned value is the number of items successfully read from the file.
size_t read_raw_dict(const char* fn_raw, const char *fn_validhzs,
size_t max_item);
// Try to find if a character is in hzs buffer.
bool hz_in_hanzis_list(const char16 *hzs, size_t hzs_len, char16 hz);
// Try to find if all characters in str are in hzs buffer.
bool str_in_hanzis_list(const char16 *hzs, size_t hzs_len,
const char16 *str, size_t str_len);
// Get these lemmas with toppest scores.
void get_top_lemmas();
// Allocate resource to build dictionary.
// lma_num is the number of items to be loaded
bool alloc_resource(size_t lma_num);
// Free resource.
void free_resource();
};
#endif // ___BUILD_MODEL___
}
#endif // PINYINIME_INCLUDE_DICTBUILDER_H__