+// Copyright 2011 Google Inc. All Rights Reserved.
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
package com.hughes.android.dictionary.engine;
import java.util.ArrayList;
+import java.util.Collections;
+import java.util.Comparator;
import java.util.EnumMap;
import java.util.HashSet;
import java.util.List;
import java.util.SortedMap;
import java.util.TreeMap;
-import com.hughes.android.dictionary.Language;
+import com.hughes.android.dictionary.engine.Index.IndexEntry;
+import com.hughes.android.dictionary.parser.DictFileParser;
public class IndexBuilder {
-
- final DictionaryBuilder dictionaryBuilder;
- final Index index;
-
- final SortedMap<String, TokenData> tokenToData;
-
- @SuppressWarnings("unchecked")
- IndexBuilder(final DictionaryBuilder dictionaryBuilder, final String shortName, final String longName, final Language language) {
- this.dictionaryBuilder = dictionaryBuilder;
- index = new Index(dictionaryBuilder.dictionary, shortName, longName, language);
- tokenToData = new TreeMap<String, TokenData>(language.getSortCollator());
- }
-
- public void build() {
- final Set<EntryData> tokenEntryDatas = new HashSet<EntryData>();
- final List<RowBase> rows = index.rows;
- for (final TokenData tokenData : tokenToData.values()) {
- tokenEntryDatas.clear();
- final int indexRow = index.sortedIndexEntries.size();
- index.sortedIndexEntries.add(new Index.IndexEntry(tokenData.token, rows.size()));
- rows.add(new TokenRow(indexRow, rows.size(), index));
- int count = 0;
- for (final List<EntryData> entryDatas : tokenData.typeToEntries.values()) {
- for (final EntryData entryData : entryDatas) {
- if (tokenEntryDatas.add(entryData)) {
- rows.add(new PairEntry.Row(entryData.index(), rows.size(), index));
- ++count;
- }
+
+ final DictionaryBuilder dictionaryBuilder;
+ public final Index index;
+ final Set<String> stoplist;
+
+ final SortedMap<FastCompareString, TokenData> tokenToData;
+
+ IndexBuilder(final DictionaryBuilder dictionaryBuilder, final String shortName, final String longName, final Language language, final String normalizerRules, final Set<String> stoplist, final boolean swapPairEntries) {
+ this.dictionaryBuilder = dictionaryBuilder;
+ index = new Index(dictionaryBuilder.dictionary, shortName, longName, language, normalizerRules, swapPairEntries, stoplist);
+ tokenToData = new TreeMap<>(new FastNormalizeComparator(index.getSortComparator()));
+ this.stoplist = stoplist;
+ }
+
+ public void build() {
+ final Set<IndexedEntry> tokenIndexedEntries = new HashSet<>();
+ final List<RowBase> rows = index.rows;
+ index.mainTokenCount = 0;
+ for (final TokenData tokenData : tokenToData.values()) {
+ tokenIndexedEntries.clear();
+ final int indexIndex = index.sortedIndexEntries.size();
+ final int startRow = rows.size();
+
+ TokenRow tokenRow = null;
+ if (!tokenData.htmlEntries.isEmpty()) {
+ tokenRow = new TokenRow(indexIndex, rows.size(), index, tokenData.hasMainEntry);
+ rows.add(tokenRow);
+ }
+
+// System.out.println("Added TokenRow: " + rows.get(rows.size() - 1));
+
+ int numRows = 0; // off by one--doesn't count the token row!
+// System.out.println("TOKEN: " + tokenData.token);
+ for (final Map.Entry<EntryTypeName, List<IndexedEntry>> typeToIndexedEntries : tokenData.typeToEntries.entrySet()) {
+ for (final IndexedEntry indexedEntry : typeToIndexedEntries.getValue()) {
+ if (!indexedEntry.isValid) {
+ continue;
+ }
+
+ if (tokenRow == null) {
+ tokenRow = new TokenRow(indexIndex, rows.size(), index, tokenData.hasMainEntry);
+ rows.add(tokenRow);
+ }
+
+ if (indexedEntry.entry.index() == -1) {
+ indexedEntry.entry.addToDictionary(dictionaryBuilder.dictionary);
+ assert indexedEntry.entry.index() >= 0;
+ }
+ if (tokenIndexedEntries.add(indexedEntry) && !tokenData.htmlEntries.contains(indexedEntry.entry)) {
+ rows.add(indexedEntry.entry.CreateRow(rows.size(), index));
+ ++indexedEntry.entry.entrySource.numEntries;
+ ++numRows;
+
+// System.out.print(" " + typeToEntry.getKey() + ": ");
+ // rows.get(rows.size() - 1).print(System.out);
+// System.out.println();
+ }
+ }
+ }
+
+ if (tokenRow != null) {
+ if (tokenRow.hasMainEntry) {
+ index.mainTokenCount++;
+ }
+
+ final Index.IndexEntry indexEntry = new Index.IndexEntry(index, tokenData.token, index
+ .normalizer().transliterate(tokenData.token), startRow, numRows);
+ indexEntry.htmlEntries.addAll(tokenData.htmlEntries);
+ index.sortedIndexEntries.add(indexEntry);
+ }
+ }
+
+ final List<IndexEntry> entriesSortedByNumRows = new ArrayList<>(index.sortedIndexEntries);
+ entriesSortedByNumRows.sort((object1, object2) -> object2.numRows - object1.numRows);
+ System.out.println("Most common tokens:");
+ for (int i = 0; i < 50 && i < entriesSortedByNumRows.size(); ++i) {
+ System.out.println(" " + entriesSortedByNumRows.get(i));
}
- }
- System.out.println(count + " ENTRIES FOR TOKEN " + tokenData.token);
}
- }
-
- static class TokenData {
- final String token;
-
- final Map<EntryTypeName, List<EntryData>> typeToEntries = new EnumMap<EntryTypeName, List<EntryData>>(EntryTypeName.class);
-
- TokenData(final String token) {
- assert token.equals(token.trim());
- assert token.length() > 0;
- this.token = token;
+
+ public static class TokenData {
+ final String token;
+
+ final Map<EntryTypeName, List<IndexedEntry>> typeToEntries = new EnumMap<>(EntryTypeName.class);
+ public boolean hasMainEntry = false;
+
+ public final List<HtmlEntry> htmlEntries = new ArrayList<>();
+
+ TokenData(final String token) {
+ assert token.equals(token.trim());
+ assert token.length() > 0;
+ this.token = token;
+ }
+ }
+
+ public TokenData getOrCreateTokenData(final String token) {
+ final FastCompareString c = new FastCompareString(token);
+ TokenData tokenData = tokenToData.get(c);
+ if (tokenData == null) {
+ tokenData = new TokenData(token);
+ tokenToData.put(c, tokenData);
+ }
+ return tokenData;
}
- }
- public TokenData getOrCreateTokenData(final String token) {
- TokenData tokenData = tokenToData.get(token);
- if (tokenData == null) {
- tokenData = new TokenData(token);
- tokenToData.put(token, tokenData);
+ private List<IndexedEntry> getOrCreateEntries(final String token, final EntryTypeName entryTypeName) {
+ final TokenData tokenData = getOrCreateTokenData(token);
+ List<IndexedEntry> entries = tokenData.typeToEntries.get(entryTypeName);
+ if (entryTypeName.mainWord) {
+ tokenData.hasMainEntry = true;
+ }
+ if (entries == null) {
+ entries = new ArrayList<>();
+ tokenData.typeToEntries.put(entryTypeName, entries);
+ }
+ return entries;
+ }
+
+ public void addEntryWithTokens(final IndexedEntry indexedEntry, final Set<String> tokens,
+ final EntryTypeName entryTypeName) {
+ if (indexedEntry == null) {
+ System.out.println("asdfasdf");
+ }
+ assert indexedEntry != null;
+ for (final String token : tokens) {
+ if (entryTypeName.overridesStopList || !stoplist.contains(token)) {
+ getOrCreateEntries(token, entryTypeName).add(indexedEntry);
+ }
+ }
}
- return tokenData;
- }
-
- public List<EntryData> getOrCreateEntries(final String token, final EntryTypeName entryTypeName) {
- final TokenData tokenData = getOrCreateTokenData(token);
- List<EntryData> entries = tokenData.typeToEntries.get(entryTypeName);
- if (entries == null) {
- entries = new ArrayList<EntryData>();
- tokenData.typeToEntries.put(entryTypeName, entries);
+
+ public void addEntryWithString(final IndexedEntry indexedEntry, final String untokenizedString,
+ final EntryTypeName entryTypeName) {
+ final Set<String> tokens = DictFileParser.tokenize(untokenizedString, DictFileParser.NON_CHAR);
+ addEntryWithTokens(indexedEntry, tokens, tokens.size() == 1 ? entryTypeName.singleWordInstance : entryTypeName);
}
- return entries;
- }
-
+ public void addEntryWithStringNoSingle(final IndexedEntry indexedEntry, final String untokenizedString,
+ final EntryTypeName entryTypeName) {
+ final Set<String> tokens = DictFileParser.tokenize(untokenizedString, DictFileParser.NON_CHAR);
+ addEntryWithTokens(indexedEntry, tokens, entryTypeName);
+ }
}