// Copyright 2013 Google Inc. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // // Author: dsites@google.com (Dick Sites) // Updated 2014.01 for dual table lookup // #include "scoreonescriptspan.h" #include "cldutil.h" #include "debug.h" #include "lang_script.h" #include using namespace std; namespace CLD2 { static const int kUnreliablePercentThreshold = 75; void AddLangProb(uint32 langprob, Tote* chunk_tote) { ProcessProbV2Tote(langprob, chunk_tote); } void ZeroPSLang(uint32 langprob, Tote* chunk_tote) { uint8 top1 = (langprob >> 8) & 0xff; chunk_tote->SetScore(top1, 0); } bool SameCloseSet(uint16 lang1, uint16 lang2) { int lang1_close_set = LanguageCloseSet(static_cast(lang1)); if (lang1_close_set == 0) {return false;} int lang2_close_set = LanguageCloseSet(static_cast(lang2)); return (lang1_close_set == lang2_close_set); } bool SameCloseSet(Language lang1, Language lang2) { int lang1_close_set = LanguageCloseSet(lang1); if (lang1_close_set == 0) {return false;} int lang2_close_set = LanguageCloseSet(lang2); return (lang1_close_set == lang2_close_set); } // Needs expected score per 1KB in scoring context void SetChunkSummary(ULScript ulscript, int first_linear_in_chunk, int offset, int len, const ScoringContext* scoringcontext, const Tote* chunk_tote, ChunkSummary* chunksummary) { int key3[3]; chunk_tote->CurrentTopThreeKeys(key3); Language lang1 = FromPerScriptNumber(ulscript, key3[0]); Language lang2 = FromPerScriptNumber(ulscript, key3[1]); int actual_score_per_kb = 0; if (len > 0) { actual_score_per_kb = (chunk_tote->GetScore(key3[0]) << 10) / len; } int expected_subscr = lang1 * 4 + LScript4(ulscript); int expected_score_per_kb = scoringcontext->scoringtables->kExpectedScore[expected_subscr]; chunksummary->offset = offset; chunksummary->chunk_start = first_linear_in_chunk; chunksummary->lang1 = lang1; chunksummary->lang2 = lang2; chunksummary->score1 = chunk_tote->GetScore(key3[0]); chunksummary->score2 = chunk_tote->GetScore(key3[1]); chunksummary->bytes = len; chunksummary->grams = chunk_tote->GetScoreCount(); chunksummary->ulscript = ulscript; chunksummary->reliability_delta = ReliabilityDelta(chunksummary->score1, chunksummary->score2, chunksummary->grams); // If lang1/lang2 in same close set, set delta reliability to 100% if (SameCloseSet(lang1, lang2)) { chunksummary->reliability_delta = 100; } chunksummary->reliability_score = ReliabilityExpected(actual_score_per_kb, expected_score_per_kb); } // Return true if just lang1 is there: lang2=0 and lang3=0 bool IsSingleLang(uint32 langprob) { // Probably a bug -- which end is lang1? But only used to call empty Boost1 return ((langprob & 0x00ffff00) == 0); } // Update scoring context distinct_boost for single language quad void AddDistinctBoost1(uint32 langprob, ScoringContext* scoringcontext) { // Probably keep this empty -- not a good enough signal } // Update scoring context distinct_boost for distinct octagram // Keep last 4 used. Since these are mostly (except at splices) in // hitbuffer, we might be able to just use a subscript and splice void AddDistinctBoost2(uint32 langprob, ScoringContext* scoringcontext) { // this is called 0..n times per chunk with decoded hitbuffer->distinct... LangBoosts* distinct_boost = &scoringcontext->distinct_boost.latn; if (scoringcontext->ulscript != ULScript_Latin) { distinct_boost = &scoringcontext->distinct_boost.othr; } int n = distinct_boost->n; distinct_boost->langprob[n] = langprob; distinct_boost->n = distinct_boost->wrap(n + 1); } // For each chunk, add extra weight for language priors (from content-lang and // meta lang=xx) and distinctive tokens void ScoreBoosts(const ScoringContext* scoringcontext, Tote* chunk_tote) { // Get boosts for current script const LangBoosts* langprior_boost = &scoringcontext->langprior_boost.latn; const LangBoosts* langprior_whack = &scoringcontext->langprior_whack.latn; const LangBoosts* distinct_boost = &scoringcontext->distinct_boost.latn; if (scoringcontext->ulscript != ULScript_Latin) { langprior_boost = &scoringcontext->langprior_boost.othr; langprior_whack = &scoringcontext->langprior_whack.othr; distinct_boost = &scoringcontext->distinct_boost.othr; } for (int k = 0; k < kMaxBoosts; ++k) { uint32 langprob = langprior_boost->langprob[k]; if (langprob > 0) {AddLangProb(langprob, chunk_tote);} } for (int k = 0; k < kMaxBoosts; ++k) { uint32 langprob = distinct_boost->langprob[k]; if (langprob > 0) {AddLangProb(langprob, chunk_tote);} } // boost has a packed set of per-script langs and probabilites // whack has a packed set of per-script lang to be suppressed (zeroed) // When a language in a close set is given as an explicit hint, others in // that set will be whacked here. for (int k = 0; k < kMaxBoosts; ++k) { uint32 langprob = langprior_whack->langprob[k]; if (langprob > 0) {ZeroPSLang(langprob, chunk_tote);} } } // At this point, The chunk is described by // hitbuffer->base[cspan->chunk_base .. cspan->chunk_base + cspan->base_len) // hitbuffer->delta[cspan->chunk_delta ... ) // hitbuffer->distinct[cspan->chunk_distinct ... ) // Scored text is in text[lo..hi) where // lo is 0 or the min of first base/delta/distinct hitbuffer offset and // hi is the min of next base/delta/distinct hitbuffer offset after // base_len, etc. void GetTextSpanOffsets(const ScoringHitBuffer* hitbuffer, const ChunkSpan* cspan, int* lo, int* hi) { // Front of this span int lo_base = hitbuffer->base[cspan->chunk_base].offset; int lo_delta = hitbuffer->delta[cspan->chunk_delta].offset; int lo_distinct = hitbuffer->distinct[cspan->chunk_distinct].offset; // Front of next span int hi_base = hitbuffer->base[cspan->chunk_base + cspan->base_len].offset; int hi_delta = hitbuffer->delta[cspan->chunk_delta + cspan->delta_len].offset; int hi_distinct = hitbuffer->distinct[cspan->chunk_distinct + cspan->distinct_len].offset; *lo = 0; // if (cspan->chunk_base > 0) { // *lo = minint(minint(lo_base, lo_delta), lo_distinct); // } *lo = minint(minint(lo_base, lo_delta), lo_distinct); *hi = minint(minint(hi_base, hi_delta), hi_distinct); } int DiffScore(const CLD2TableSummary* obj, int indirect, uint16 lang1, uint16 lang2) { if (indirect < static_cast(obj->kCLDTableSizeOne)) { // Up to three languages at indirect uint32 langprob = obj->kCLDTableInd[indirect]; return GetLangScore(langprob, lang1) - GetLangScore(langprob, lang2); } else { // Up to six languages at start + 2 * (indirect - start) indirect += (indirect - obj->kCLDTableSizeOne); uint32 langprob = obj->kCLDTableInd[indirect]; uint32 langprob2 = obj->kCLDTableInd[indirect + 1]; return (GetLangScore(langprob, lang1) + GetLangScore(langprob2, lang1)) - (GetLangScore(langprob, lang2) + GetLangScore(langprob2, lang2)); } } // Score all the bases, deltas, distincts, boosts for one chunk into chunk_tote // After last chunk there is always a hitbuffer entry with an offset just off // the end of the text. // Sets delta_len, and distinct_len void ScoreOneChunk(const char* text, ULScript ulscript, const ScoringHitBuffer* hitbuffer, int chunk_i, ScoringContext* scoringcontext, ChunkSpan* cspan, Tote* chunk_tote, ChunkSummary* chunksummary) { int first_linear_in_chunk = hitbuffer->chunk_start[chunk_i]; int first_linear_in_next_chunk = hitbuffer->chunk_start[chunk_i + 1]; chunk_tote->Reinit(); cspan->delta_len = 0; cspan->distinct_len = 0; if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, "
ScoreOneChunk[%d..%d) ", first_linear_in_chunk, first_linear_in_next_chunk); } // 2013.02.05 linear design: just use base and base_len for the span cspan->chunk_base = first_linear_in_chunk; cspan->base_len = first_linear_in_next_chunk - first_linear_in_chunk; for (int i = first_linear_in_chunk; i < first_linear_in_next_chunk; ++i) { uint32 langprob = hitbuffer->linear[i].langprob; AddLangProb(langprob, chunk_tote); if (hitbuffer->linear[i].type <= QUADHIT) { chunk_tote->AddScoreCount(); // Just count quads, not octas } if (hitbuffer->linear[i].type == DISTINCTHIT) { AddDistinctBoost2(langprob, scoringcontext); } } // Score language prior boosts // Score distinct word boost ScoreBoosts(scoringcontext, chunk_tote); int lo = hitbuffer->linear[first_linear_in_chunk].offset; int hi = hitbuffer->linear[first_linear_in_next_chunk].offset; // Chunk_tote: get top langs, scores, etc. and fill in chunk summary SetChunkSummary(ulscript, first_linear_in_chunk, lo, hi - lo, scoringcontext, chunk_tote, chunksummary); bool more_to_come = false; bool score_cjk = false; if (scoringcontext->flags_cld2_html) { // Show one chunk in readable output CLD2_Debug(text, lo, hi, more_to_come, score_cjk, hitbuffer, scoringcontext, cspan, chunksummary); } scoringcontext->prior_chunk_lang = static_cast(chunksummary->lang1); } // Score chunks of text described by hitbuffer, allowing each to be in a // different language, and optionally adjusting the boundaries inbetween. // Set last_cspan to the last chunkspan used void ScoreAllHits(const char* text, ULScript ulscript, bool more_to_come, bool score_cjk, const ScoringHitBuffer* hitbuffer, ScoringContext* scoringcontext, SummaryBuffer* summarybuffer, ChunkSpan* last_cspan) { ChunkSpan prior_cspan = {0, 0, 0, 0, 0, 0}; ChunkSpan cspan = {0, 0, 0, 0, 0, 0}; for (int i = 0; i < hitbuffer->next_chunk_start; ++i) { // Score one chunk // Sets delta_len, and distinct_len Tote chunk_tote; ChunkSummary chunksummary; ScoreOneChunk(text, ulscript, hitbuffer, i, scoringcontext, &cspan, &chunk_tote, &chunksummary); // Put result in summarybuffer if (summarybuffer->n < kMaxSummaries) { summarybuffer->chunksummary[summarybuffer->n] = chunksummary; summarybuffer->n += 1; } prior_cspan = cspan; cspan.chunk_base += cspan.base_len; cspan.chunk_delta += cspan.delta_len; cspan.chunk_distinct += cspan.distinct_len; } // Add one dummy off the end to hold first unused linear_in_chunk int linear_off_end = hitbuffer->next_linear; int offset_off_end = hitbuffer->linear[linear_off_end].offset; ChunkSummary* cs = &summarybuffer->chunksummary[summarybuffer->n]; memset(cs, 0, sizeof(ChunkSummary)); cs->offset = offset_off_end; cs->chunk_start = linear_off_end; *last_cspan = prior_cspan; } void SummaryBufferToDocTote(const SummaryBuffer* summarybuffer, bool more_to_come, DocTote* doc_tote) { int cs_bytes_sum = 0; for (int i = 0; i < summarybuffer->n; ++i) { const ChunkSummary* cs = &summarybuffer->chunksummary[i]; int reliability = minint(cs->reliability_delta, cs->reliability_score); // doc_tote uses full languages doc_tote->Add(cs->lang1, cs->bytes, cs->score1, reliability); cs_bytes_sum += cs->bytes; } } // Turn on for debugging vectors static const bool kShowLettersOriginal = false; // If next chunk language matches last vector language, extend last element // Otherwise add new element to vector void ItemToVector(ScriptScanner* scanner, ResultChunkVector* vec, Language new_lang, int mapped_offset, int mapped_len) { uint16 last_vec_lang = static_cast(UNKNOWN_LANGUAGE); int last_vec_subscr = vec->size() - 1; if (last_vec_subscr >= 0) { ResultChunk* priorrc = &(*vec)[last_vec_subscr]; last_vec_lang = priorrc->lang1; if (new_lang == last_vec_lang) { // Extend prior. Current mapped_offset may be beyond prior end, so do // the arithmetic to include any such gap priorrc->bytes = minint((mapped_offset + mapped_len) - priorrc->offset, kMaxResultChunkBytes); if (kShowLettersOriginal) { // Optionally print the new chunk original text string temp2(&scanner->GetBufferStart()[priorrc->offset], priorrc->bytes); fprintf(stderr, "Item[%d..%d) '%s'
\n", priorrc->offset, priorrc->offset + priorrc->bytes, GetHtmlEscapedText(temp2).c_str()); } return; } } // Add new vector element ResultChunk rc; rc.offset = mapped_offset; rc.bytes = minint(mapped_len, kMaxResultChunkBytes); rc.lang1 = static_cast(new_lang); vec->push_back(rc); if (kShowLettersOriginal) { // Optionally print the new chunk original text string temp2(&scanner->GetBufferStart()[rc.offset], rc.bytes); fprintf(stderr, "Item[%d..%d) '%s'
\n", rc.offset, rc.offset + rc.bytes, GetHtmlEscapedText(temp2).c_str()); } } uint16 PriorVecLang(const ResultChunkVector* vec) { if (vec->empty()) {return static_cast(UNKNOWN_LANGUAGE);} return (*vec)[vec->size() - 1].lang1; } uint16 NextChunkLang(const SummaryBuffer* summarybuffer, int i) { if ((i + 1) >= summarybuffer->n) { return static_cast(UNKNOWN_LANGUAGE); } return summarybuffer->chunksummary[i + 1].lang1; } // Add n elements of summarybuffer to resultchunk vector: // Each element is letters-only text [offset..offset+bytes) // This maps back to original[Back(offset)..Back(offset+bytes)) // // We go out of our way to minimize the variation in the ResultChunkVector, // so that the caller has fewer but more meaningful spans in different // lanaguges, for the likely purpose of translation or spell-check. // // The language of each chunk is lang1, but it might be unreliable for // either of two reasons: its score is relatively too close to the score of // lang2, or its score is too far away from the expected score of real text in // the given language. Unreliable languages are mapped to Unknown. // void SummaryBufferToVector(ScriptScanner* scanner, const char* text, const SummaryBuffer* summarybuffer, bool more_to_come, ResultChunkVector* vec) { if (vec == NULL) {return;} if (kShowLettersOriginal) { fprintf(stderr, "map2original_ "); scanner->map2original_.DumpWindow(); fprintf(stderr, "
\n"); fprintf(stderr, "map2uplow_ "); scanner->map2uplow_.DumpWindow(); fprintf(stderr, "
\n"); } for (int i = 0; i < summarybuffer->n; ++i) { const ChunkSummary* cs = &summarybuffer->chunksummary[i]; int unmapped_offset = cs->offset; int unmapped_len = cs->bytes; if (kShowLettersOriginal) { // Optionally print the chunk lowercase letters/marks text string temp(&text[unmapped_offset], unmapped_len); fprintf(stderr, "Letters [%d..%d) '%s'
\n", unmapped_offset, unmapped_offset + unmapped_len, GetHtmlEscapedText(temp).c_str()); } int mapped_offset = scanner->MapBack(unmapped_offset); // Trim back a little to prefer splicing original at word boundaries if (mapped_offset > 0) { // Size of prior vector entry, if any int prior_size = 0; if (!vec->empty()) { ResultChunk* rc = &(*vec)[vec->size() - 1]; prior_size = rc->bytes; } // Maximum back up size to leave at least 3 bytes in prior, // and not entire buffer, and no more than 12 bytes total backup int n_limit = minint(prior_size - 3, mapped_offset); n_limit = minint(n_limit, 12); // Backscan over letters, stopping if prior byte is < 0x41 // There is some possibility that we will backscan over a different script const char* s = &scanner->GetBufferStart()[mapped_offset]; const unsigned char* us = reinterpret_cast(s); int n = 0; while ((n < n_limit) && (us[-n - 1] >= 0x41)) {++n;} if (n >= n_limit) {n = 0;} // New boundary not found within range // Also back up exactly one leading punctuation character if '"#@ if (n < n_limit) { unsigned char c = us[-n - 1]; if ((c == '\'') || (c == '"') || (c == '#') || (c == '@')) {++n;} } // Shrink the previous chunk slightly if (n > 0) { ResultChunk* rc = &(*vec)[vec->size() - 1]; rc->bytes -= n; mapped_offset -= n; if (kShowLettersOriginal) { fprintf(stderr, "Back up %d bytes
\n", n); // Optionally print the prior chunk original text string temp2(&scanner->GetBufferStart()[rc->offset], rc->bytes); fprintf(stderr, "Prior [%d..%d) '%s'
\n", rc->offset, rc->offset + rc->bytes, GetHtmlEscapedText(temp2).c_str()); } } } int mapped_len = scanner->MapBack(unmapped_offset + unmapped_len) - mapped_offset; if (kShowLettersOriginal) { // Optionally print the chunk original text string temp2(&scanner->GetBufferStart()[mapped_offset], mapped_len); fprintf(stderr, "Original[%d..%d) '%s'
\n", mapped_offset, mapped_offset + mapped_len, GetHtmlEscapedText(temp2).c_str()); } Language new_lang = static_cast(cs->lang1); bool reliability_delta_bad = (cs->reliability_delta < kUnreliablePercentThreshold); bool reliability_score_bad = (cs->reliability_score < kUnreliablePercentThreshold); // If the top language matches last vector, ignore reliability_delta uint16 prior_lang = PriorVecLang(vec); if (prior_lang == cs->lang1) { reliability_delta_bad = false; } // If the top language is in same close set as last vector, set up to merge if (SameCloseSet(cs->lang1, prior_lang)) { new_lang = static_cast(prior_lang); reliability_delta_bad = false; } // If the top two languages are in the same close set and the last vector // language is the second language, set up to merge if (SameCloseSet(cs->lang1, cs->lang2) && (prior_lang == cs->lang2)) { new_lang = static_cast(prior_lang); reliability_delta_bad = false; } // If unreliable and the last and next vector languages are both // the second language, set up to merge uint16 next_lang = NextChunkLang(summarybuffer, i); if (reliability_delta_bad && (prior_lang == cs->lang2) && (next_lang == cs->lang2)) { new_lang = static_cast(prior_lang); reliability_delta_bad = false; } if (reliability_delta_bad || reliability_score_bad) { new_lang = UNKNOWN_LANGUAGE; } ItemToVector(scanner, vec, new_lang, mapped_offset, mapped_len); } } // Add just one element to resultchunk vector: // For RTypeNone or RTypeOne void JustOneItemToVector(ScriptScanner* scanner, const char* text, Language lang1, int unmapped_offset, int unmapped_len, ResultChunkVector* vec) { if (vec == NULL) {return;} if (kShowLettersOriginal) { fprintf(stderr, "map2original_ "); scanner->map2original_.DumpWindow(); fprintf(stderr, "
\n"); fprintf(stderr, "map2uplow_ "); scanner->map2uplow_.DumpWindow(); fprintf(stderr, "
\n"); } if (kShowLettersOriginal) { // Optionally print the chunk lowercase letters/marks text string temp(&text[unmapped_offset], unmapped_len); fprintf(stderr, "Letters1 [%d..%d) '%s'
\n", unmapped_offset, unmapped_offset + unmapped_len, GetHtmlEscapedText(temp).c_str()); } int mapped_offset = scanner->MapBack(unmapped_offset); int mapped_len = scanner->MapBack(unmapped_offset + unmapped_len) - mapped_offset; if (kShowLettersOriginal) { // Optionally print the chunk original text string temp2(&scanner->GetBufferStart()[mapped_offset], mapped_len); fprintf(stderr, "Original1[%d..%d) '%s'
\n", mapped_offset, mapped_offset + mapped_len, GetHtmlEscapedText(temp2).c_str()); } ItemToVector(scanner, vec, lang1, mapped_offset, mapped_len); } // Debugging. Not thread safe. Defined in getonescriptspan char* DisplayPiece(const char* next_byte_, int byte_length_); // If high bit is on, take out high bit and add 2B to make table2 entries easy inline int PrintableIndirect(int x) { if ((x & 0x80000000u) != 0) { return (x & ~0x80000000u) + 2000000000; } return x; } void DumpHitBuffer(FILE* df, const char* text, const ScoringHitBuffer* hitbuffer) { fprintf(df, "
DumpHitBuffer[%s, next_base/delta/distinct %d, %d, %d)
\n", ULScriptCode(hitbuffer->ulscript), hitbuffer->next_base, hitbuffer->next_delta, hitbuffer->next_distinct); for (int i = 0; i < hitbuffer->maxscoringhits; ++i) { if (i < hitbuffer->next_base) { fprintf(df, "Q[%d]%d,%d,%s ", i, hitbuffer->base[i].offset, PrintableIndirect(hitbuffer->base[i].indirect), DisplayPiece(&text[hitbuffer->base[i].offset], 6)); } if (i < hitbuffer->next_delta) { fprintf(df, "DL[%d]%d,%d,%s ", i, hitbuffer->delta[i].offset, hitbuffer->delta[i].indirect, DisplayPiece(&text[hitbuffer->delta[i].offset], 12)); } if (i < hitbuffer->next_distinct) { fprintf(df, "D[%d]%d,%d,%s ", i, hitbuffer->distinct[i].offset, hitbuffer->distinct[i].indirect, DisplayPiece(&text[hitbuffer->distinct[i].offset], 12)); } if (i < hitbuffer->next_base) { fprintf(df, "
\n"); } if (i > 50) {break;} } if (hitbuffer->next_base > 50) { int i = hitbuffer->next_base; fprintf(df, "Q[%d]%d,%d,%s ", i, hitbuffer->base[i].offset, PrintableIndirect(hitbuffer->base[i].indirect), DisplayPiece(&text[hitbuffer->base[i].offset], 6)); } if (hitbuffer->next_delta > 50) { int i = hitbuffer->next_delta; fprintf(df, "DL[%d]%d,%d,%s ", i, hitbuffer->delta[i].offset, hitbuffer->delta[i].indirect, DisplayPiece(&text[hitbuffer->delta[i].offset], 12)); } if (hitbuffer->next_distinct > 50) { int i = hitbuffer->next_distinct; fprintf(df, "D[%d]%d,%d,%s ", i, hitbuffer->distinct[i].offset, hitbuffer->distinct[i].indirect, DisplayPiece(&text[hitbuffer->distinct[i].offset], 12)); } fprintf(df, "
\n"); } void DumpLinearBuffer(FILE* df, const char* text, const ScoringHitBuffer* hitbuffer) { fprintf(df, "
DumpLinearBuffer[%d)
\n", hitbuffer->next_linear); // Include the dummy entry off the end for (int i = 0; i < hitbuffer->next_linear + 1; ++i) { if ((50 < i) && (i < (hitbuffer->next_linear - 1))) {continue;} fprintf(df, "[%d]%d,%c=%08x,%s
\n", i, hitbuffer->linear[i].offset, "UQLD"[hitbuffer->linear[i].type], hitbuffer->linear[i].langprob, DisplayPiece(&text[hitbuffer->linear[i].offset], 6)); } fprintf(df, "
\n"); fprintf(df, "DumpChunkStart[%d]
\n", hitbuffer->next_chunk_start); for (int i = 0; i < hitbuffer->next_chunk_start + 1; ++i) { fprintf(df, "[%d]%d\n", i, hitbuffer->chunk_start[i]); } fprintf(df, "
\n"); } // Move this verbose debugging output to debug.cc eventually void DumpChunkSummary(FILE* df, const ChunkSummary* cs) { // Print chunksummary fprintf(df, "%d lin[%d] %s.%d %s.%d %dB %d# %s %dRd %dRs
\n", cs->offset, cs->chunk_start, LanguageCode(static_cast(cs->lang1)), cs->score1, LanguageCode(static_cast(cs->lang2)), cs->score2, cs->bytes, cs->grams, ULScriptCode(static_cast(cs->ulscript)), cs->reliability_delta, cs->reliability_score); } void DumpSummaryBuffer(FILE* df, const SummaryBuffer* summarybuffer) { fprintf(df, "
DumpSummaryBuffer[%d]
\n", summarybuffer->n); fprintf(df, "[i] offset linear[chunk_start] lang.score1 lang.score2 " "bytesB ngrams# script rel_delta rel_score
\n"); for (int i = 0; i <= summarybuffer->n; ++i) { fprintf(df, "[%d] ", i); DumpChunkSummary(df, &summarybuffer->chunksummary[i]); } fprintf(df, "
\n"); } // Within hitbufer->linear[] // <-- prior chunk --><-- this chunk --> // | | | // linear0 linear1 linear2 // lang0 lang1 // The goal of sharpening is to move this_linear to better separate langs int BetterBoundary(const char* text, ScoringHitBuffer* hitbuffer, ScoringContext* scoringcontext, uint16 pslang0, uint16 pslang1, int linear0, int linear1, int linear2) { // Degenerate case, no change if ((linear2 - linear0) <= 8) {return linear1;} // Each diff gives pslang0 score - pslang1 score // Running diff has four entries + + + + followed by four entries - - - - // so that this value is maximal at the sharpest boundary between pslang0 // (positive diffs) and pslang1 (negative diffs) int running_diff = 0; int diff[8]; // Ring buffer of pslang0-pslang1 differences // Initialize with first 8 diffs for (int i = linear0; i < linear0 + 8; ++i) { int j = i & 7; uint32 langprob = hitbuffer->linear[i].langprob; diff[j] = GetLangScore(langprob, pslang0) - GetLangScore(langprob, pslang1); if (i < linear0 + 4) { // First four diffs pslang0 - pslang1 running_diff += diff[j]; } else { // Second four diffs -(pslang0 - pslang1) running_diff -= diff[j]; } } // Now scan for sharpest boundary. j is at left end of 8 entries // To be a boundary, there must be both >0 and <0 entries in the window int better_boundary_value = 0; int better_boundary = linear1; for (int i = linear0; i < linear2 - 8; ++i) { int j = i & 7; if (better_boundary_value < running_diff) { bool has_plus = false; bool has_minus = false; for (int kk = 0; kk < 8; ++kk) { if (diff[kk] > 0) {has_plus = true;} if (diff[kk] < 0) {has_minus = true;} } if (has_plus && has_minus) { better_boundary_value = running_diff; better_boundary = i + 4; } } // Shift right one entry uint32 langprob = hitbuffer->linear[i + 8].langprob; int newdiff = GetLangScore(langprob, pslang0) - GetLangScore(langprob, pslang1); int middiff = diff[(i + 4) & 7]; int olddiff = diff[j]; diff[j] = newdiff; running_diff -= olddiff; // Remove left running_diff += 2 * middiff; // Convert middle from - to + running_diff -= newdiff; // Insert right } if (scoringcontext->flags_cld2_verbose && (linear1 != better_boundary)) { Language lang0 = FromPerScriptNumber(scoringcontext->ulscript, pslang0); Language lang1 = FromPerScriptNumber(scoringcontext->ulscript, pslang1); fprintf(scoringcontext->debug_file, " Better lin[%d=>%d] %s^^%s
\n", linear1, better_boundary, LanguageCode(lang0), LanguageCode(lang1)); int lin0_off = hitbuffer->linear[linear0].offset; int lin1_off = hitbuffer->linear[linear1].offset; int lin2_off = hitbuffer->linear[linear2].offset; int better_offm1 = hitbuffer->linear[better_boundary - 1].offset; int better_off = hitbuffer->linear[better_boundary].offset; int better_offp1 = hitbuffer->linear[better_boundary + 1].offset; string old0(&text[lin0_off], lin1_off - lin0_off); string old1(&text[lin1_off], lin2_off - lin1_off); string new0(&text[lin0_off], better_offm1 - lin0_off); string new0m1(&text[better_offm1], better_off - better_offm1); string new1(&text[better_off], better_offp1 - better_off); string new1p1(&text[better_offp1], lin2_off - better_offp1); fprintf(scoringcontext->debug_file, "%s^^%s =>
\n%s^%s^^%s^%s
\n", GetHtmlEscapedText(old0).c_str(), GetHtmlEscapedText(old1).c_str(), GetHtmlEscapedText(new0).c_str(), GetHtmlEscapedText(new0m1).c_str(), GetHtmlEscapedText(new1).c_str(), GetHtmlEscapedText(new1p1).c_str()); // Slow picture of differences per linear entry int d; for (int i = linear0; i < linear2; ++i) { if (i == better_boundary) { fprintf(scoringcontext->debug_file, "^^ "); } uint32 langprob = hitbuffer->linear[i].langprob; d = GetLangScore(langprob, pslang0) - GetLangScore(langprob, pslang1); const char* s = "="; //if (d > 2) {s = "\xc2\xaf";} // Macron if (d > 2) {s = "#";} else if (d > 0) {s = "+";} else if (d < -2) {s = "_";} else if (d < 0) {s = "-";} fprintf(scoringcontext->debug_file, "%s ", s); } fprintf(scoringcontext->debug_file, "   (scale: #+=-_)
\n"); } return better_boundary; } // For all but the first summary, if its top language differs from // the previous chunk, refine the boundary // Linearized version void SharpenBoundaries(const char* text, bool more_to_come, ScoringHitBuffer* hitbuffer, ScoringContext* scoringcontext, SummaryBuffer* summarybuffer) { int prior_linear = summarybuffer->chunksummary[0].chunk_start; uint16 prior_lang = summarybuffer->chunksummary[0].lang1; if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, "
SharpenBoundaries
\n"); } for (int i = 1; i < summarybuffer->n; ++i) { ChunkSummary* cs = &summarybuffer->chunksummary[i]; uint16 this_lang = cs->lang1; if (this_lang == prior_lang) { prior_linear = cs->chunk_start; continue; } int this_linear = cs->chunk_start; int next_linear = summarybuffer->chunksummary[i + 1].chunk_start; // If this/prior in same close set, don't move boundary if (SameCloseSet(prior_lang, this_lang)) { prior_linear = this_linear; prior_lang = this_lang; continue; } // Within hitbuffer->linear[] // <-- prior chunk --><-- this chunk --> // | | | // prior_linear this_linear next_linear // prior_lang this_lang // The goal of sharpening is to move this_linear to better separate langs uint8 pslang0 = PerScriptNumber(scoringcontext->ulscript, static_cast(prior_lang)); uint8 pslang1 = PerScriptNumber(scoringcontext->ulscript, static_cast(this_lang)); int better_linear = BetterBoundary(text, hitbuffer, scoringcontext, pslang0, pslang1, prior_linear, this_linear, next_linear); int old_offset = hitbuffer->linear[this_linear].offset; int new_offset = hitbuffer->linear[better_linear].offset; cs->chunk_start = better_linear; cs->offset = new_offset; // If this_linear moved right, make bytes smaller for this, larger for prior // If this_linear moved left, make bytes larger for this, smaller for prior cs->bytes -= (new_offset - old_offset); summarybuffer->chunksummary[i - 1].bytes += (new_offset - old_offset); this_linear = better_linear; // Update so that next chunk doesn't intrude // Consider rescoring the two chunks // Update for next round (note: using pre-updated boundary) prior_linear = this_linear; prior_lang = this_lang; } } // Make a langprob that gives small weight to the default language for ulscript uint32 DefaultLangProb(ULScript ulscript) { Language default_lang = DefaultLanguage(ulscript); return MakeLangProb(default_lang, 1); } // Effectively, do a merge-sort based on text offsets // Look up each indirect value in appropriate scoring table and keep // just the resulting langprobs void LinearizeAll(ScoringContext* scoringcontext, bool score_cjk, ScoringHitBuffer* hitbuffer) { const CLD2TableSummary* base_obj; // unigram or quadgram const CLD2TableSummary* base_obj2; // quadgram dual table const CLD2TableSummary* delta_obj; // bigram or octagram const CLD2TableSummary* distinct_obj; // bigram or octagram uint16 base_hit; if (score_cjk) { base_obj = scoringcontext->scoringtables->unigram_compat_obj; base_obj2 = scoringcontext->scoringtables->unigram_compat_obj; delta_obj = scoringcontext->scoringtables->deltabi_obj; distinct_obj = scoringcontext->scoringtables->distinctbi_obj; base_hit = UNIHIT; } else { base_obj = scoringcontext->scoringtables->quadgram_obj; base_obj2 = scoringcontext->scoringtables->quadgram_obj2; delta_obj = scoringcontext->scoringtables->deltaocta_obj; distinct_obj = scoringcontext->scoringtables->distinctocta_obj; base_hit = QUADHIT; } int base_limit = hitbuffer->next_base; int delta_limit = hitbuffer->next_delta; int distinct_limit = hitbuffer->next_distinct; int base_i = 0; int delta_i = 0; int distinct_i = 0; int linear_i = 0; // Start with an initial base hit for the default language for this script // Inserting this avoids edge effects with no hits at all hitbuffer->linear[linear_i].offset = hitbuffer->lowest_offset; hitbuffer->linear[linear_i].type = base_hit; hitbuffer->linear[linear_i].langprob = DefaultLangProb(scoringcontext->ulscript); ++linear_i; while ((base_i < base_limit) || (delta_i < delta_limit) || (distinct_i < distinct_limit)) { int base_off = hitbuffer->base[base_i].offset; int delta_off = hitbuffer->delta[delta_i].offset; int distinct_off = hitbuffer->distinct[distinct_i].offset; // Do delta and distinct first, so that they are not lost at base_limit if ((delta_i < delta_limit) && (delta_off <= base_off) && (delta_off <= distinct_off)) { // Add delta entry int indirect = hitbuffer->delta[delta_i].indirect; ++delta_i; uint32 langprob = delta_obj->kCLDTableInd[indirect]; if (langprob > 0) { hitbuffer->linear[linear_i].offset = delta_off; hitbuffer->linear[linear_i].type = DELTAHIT; hitbuffer->linear[linear_i].langprob = langprob; ++linear_i; } } else if ((distinct_i < distinct_limit) && (distinct_off <= base_off) && (distinct_off <= delta_off)) { // Add distinct entry int indirect = hitbuffer->distinct[distinct_i].indirect; ++distinct_i; uint32 langprob = distinct_obj->kCLDTableInd[indirect]; if (langprob > 0) { hitbuffer->linear[linear_i].offset = distinct_off; hitbuffer->linear[linear_i].type = DISTINCTHIT; hitbuffer->linear[linear_i].langprob = langprob; ++linear_i; } } else { // Add one or two base entries int indirect = hitbuffer->base[base_i].indirect; // First, get right scoring table const CLD2TableSummary* local_base_obj = base_obj; if ((indirect & 0x80000000u) != 0) { local_base_obj = base_obj2; indirect &= ~0x80000000u; } ++base_i; // One langprob in kQuadInd[0..SingleSize), // two in kQuadInd[SingleSize..Size) if (indirect < static_cast(local_base_obj->kCLDTableSizeOne)) { // Up to three languages at indirect uint32 langprob = local_base_obj->kCLDTableInd[indirect]; if (langprob > 0) { hitbuffer->linear[linear_i].offset = base_off; hitbuffer->linear[linear_i].type = base_hit; hitbuffer->linear[linear_i].langprob = langprob; ++linear_i; } } else { // Up to six languages at start + 2 * (indirect - start) indirect += (indirect - local_base_obj->kCLDTableSizeOne); uint32 langprob = local_base_obj->kCLDTableInd[indirect]; uint32 langprob2 = local_base_obj->kCLDTableInd[indirect + 1]; if (langprob > 0) { hitbuffer->linear[linear_i].offset = base_off; hitbuffer->linear[linear_i].type = base_hit; hitbuffer->linear[linear_i].langprob = langprob; ++linear_i; } if (langprob2 > 0) { hitbuffer->linear[linear_i].offset = base_off; hitbuffer->linear[linear_i].type = base_hit; hitbuffer->linear[linear_i].langprob = langprob2; ++linear_i; } } } } // Update hitbuffer->next_linear = linear_i; // Add a dummy entry off the end, just to capture final offset hitbuffer->linear[linear_i].offset = hitbuffer->base[hitbuffer->next_base].offset; hitbuffer->linear[linear_i].langprob = 0; } // Break linear array into chunks of ~20 quadgram hits or ~50 CJK unigram hits void ChunkAll(int letter_offset, bool score_cjk, ScoringHitBuffer* hitbuffer) { int chunksize; uint16 base_hit; if (score_cjk) { chunksize = kChunksizeUnis; base_hit = UNIHIT; } else { chunksize = kChunksizeQuads; base_hit = QUADHIT; } int linear_i = 0; int linear_off_end = hitbuffer->next_linear; int text_i = letter_offset; // Next unseen text offset int next_chunk_start = 0; int bases_left = hitbuffer->next_base; while (bases_left > 0) { // Linearize one chunk int base_len = chunksize; // Default; may be changed below if (bases_left < (chunksize + (chunksize >> 1))) { // If within 1.5 chunks of the end, avoid runts by using it all base_len = bases_left; } else if (bases_left < (2 * chunksize)) { // Avoid runts by splitting 1.5 to 2 chunks in half (about 3/4 each) base_len = (bases_left + 1) >> 1; } hitbuffer->chunk_start[next_chunk_start] = linear_i; hitbuffer->chunk_offset[next_chunk_start] = text_i; ++next_chunk_start; int base_count = 0; while ((base_count < base_len) && (linear_i < linear_off_end)) { if (hitbuffer->linear[linear_i].type == base_hit) {++base_count;} ++linear_i; } text_i = hitbuffer->linear[linear_i].offset; // Next unseen text offset bases_left -= base_len; } // If no base hits at all, make a single dummy chunk if (next_chunk_start == 0) { hitbuffer->chunk_start[next_chunk_start] = 0; hitbuffer->chunk_offset[next_chunk_start] = hitbuffer->linear[0].offset; ++next_chunk_start; } // Remember the linear array start of dummy entry hitbuffer->next_chunk_start = next_chunk_start; // Add a dummy entry off the end, just to capture final linear subscr hitbuffer->chunk_start[next_chunk_start] = hitbuffer->next_linear; hitbuffer->chunk_offset[next_chunk_start] = text_i; } // Merge-sort the individual hit arrays, go indirect on the scoring subscripts, // break linear array into chunks. // // Input: // hitbuffer base, delta, distinct arrays // Output: // linear array // chunk_start array // void LinearizeHitBuffer(int letter_offset, ScoringContext* scoringcontext, bool more_to_come, bool score_cjk, ScoringHitBuffer* hitbuffer) { LinearizeAll(scoringcontext, score_cjk, hitbuffer); ChunkAll(letter_offset, score_cjk, hitbuffer); } // The hitbuffer is in an awkward form -- three sets of base/delta/distinct // scores, each with an indirect subscript to one of six scoring tables, some // of which can yield two langprobs for six languages, others one langprob for // three languages. The only correlation between base/delta/distinct is their // offsets into the letters-only text buffer. // // SummaryBuffer needs to be built to linear, giving linear offset of start of // each chunk // // So we first do all the langprob lookups and merge-sort by offset to make // a single linear vector, building a side vector of chunk beginnings as we go. // The sharpening is simply moving the beginnings, scoring is a simple linear // sweep, etc. void ProcessHitBuffer(const LangSpan& scriptspan, int letter_offset, ScoringContext* scoringcontext, DocTote* doc_tote, ResultChunkVector* vec, bool more_to_come, bool score_cjk, ScoringHitBuffer* hitbuffer) { if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, "Hitbuffer[) "); DumpHitBuffer(scoringcontext->debug_file, scriptspan.text, hitbuffer); } LinearizeHitBuffer(letter_offset, scoringcontext, more_to_come, score_cjk, hitbuffer); if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, "Linear[) "); DumpLinearBuffer(scoringcontext->debug_file, scriptspan.text, hitbuffer); } SummaryBuffer summarybuffer; summarybuffer.n = 0; ChunkSpan last_cspan; ScoreAllHits(scriptspan.text, scriptspan.ulscript, more_to_come, score_cjk, hitbuffer, scoringcontext, &summarybuffer, &last_cspan); if (scoringcontext->flags_cld2_verbose) { DumpSummaryBuffer(scoringcontext->debug_file, &summarybuffer); } if (vec != NULL) { // Sharpen boundaries of summarybuffer // This is not a high-performance path SharpenBoundaries(scriptspan.text, more_to_come, hitbuffer, scoringcontext, &summarybuffer); // Show after the sharpening // CLD2_Debug2(scriptspan.text, more_to_come, score_cjk, // hitbuffer, scoringcontext, &summarybuffer); if (scoringcontext->flags_cld2_verbose) { DumpSummaryBuffer(scoringcontext->debug_file, &summarybuffer); } } SummaryBufferToDocTote(&summarybuffer, more_to_come, doc_tote); SummaryBufferToVector(scoringcontext->scanner, scriptspan.text, &summarybuffer, more_to_come, vec); } void SpliceHitBuffer(ScoringHitBuffer* hitbuffer, int next_offset) { // Splice hitbuffer and summarybuffer for next round. With big chunks and // distinctive-word state carried across chunks, we might not need to do this. hitbuffer->next_base = 0; hitbuffer->next_delta = 0; hitbuffer->next_distinct = 0; hitbuffer->next_linear = 0; hitbuffer->next_chunk_start = 0; hitbuffer->lowest_offset = next_offset; } // Score RTypeNone or RTypeOne scriptspan into doc_tote and vec, updating // scoringcontext void ScoreEntireScriptSpan(const LangSpan& scriptspan, ScoringContext* scoringcontext, DocTote* doc_tote, ResultChunkVector* vec) { int bytes = scriptspan.text_bytes; // Artificially set score to 1024 per 1KB, or 1 per byte int score = bytes; int reliability = 100; // doc_tote uses full languages Language one_one_lang = DefaultLanguage(scriptspan.ulscript); doc_tote->Add(one_one_lang, bytes, score, reliability); if (scoringcontext->flags_cld2_html) { ChunkSummary chunksummary = { 1, 0, one_one_lang, UNKNOWN_LANGUAGE, score, 1, bytes, 0, scriptspan.ulscript, reliability, reliability }; CLD2_Debug(scriptspan.text, 1, scriptspan.text_bytes, false, false, NULL, scoringcontext, NULL, &chunksummary); } // First byte is always a space JustOneItemToVector(scoringcontext->scanner, scriptspan.text, one_one_lang, 1, bytes - 1, vec); scoringcontext->prior_chunk_lang = UNKNOWN_LANGUAGE; } // Score RTypeCJK scriptspan into doc_tote and vec, updating scoringcontext void ScoreCJKScriptSpan(const LangSpan& scriptspan, ScoringContext* scoringcontext, DocTote* doc_tote, ResultChunkVector* vec) { // Allocate three parallel arrays of scoring hits ScoringHitBuffer* hitbuffer = new ScoringHitBuffer; hitbuffer->init(); hitbuffer->ulscript = scriptspan.ulscript; scoringcontext->prior_chunk_lang = UNKNOWN_LANGUAGE; scoringcontext->oldest_distinct_boost = 0; // Incoming scriptspan has a single leading space at scriptspan.text[0] // and three trailing spaces then NUL at scriptspan.text[text_bytes + 0/1/2/3] int letter_offset = 1; // Skip initial space hitbuffer->lowest_offset = letter_offset; int letter_limit = scriptspan.text_bytes; while (letter_offset < letter_limit) { if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, " ScoreCJKScriptSpan[%d,%d)
\n", letter_offset, letter_limit); } // // Fill up one hitbuffer, possibly splicing onto previous fragment // // NOTE: GetUniHits deals with close repeats // NOTE: After last chunk there is always a hitbuffer entry with an offset // just off the end of the text = next_offset. int next_offset = GetUniHits(scriptspan.text, letter_offset, letter_limit, scoringcontext, hitbuffer); // NOTE: GetBiHitVectors deals with close repeats, // does one hash and two lookups (delta and distinct) per word GetBiHits(scriptspan.text, letter_offset, next_offset, scoringcontext, hitbuffer); // // Score one hitbuffer in chunks to summarybuffer // bool more_to_come = next_offset < letter_limit; bool score_cjk = true; ProcessHitBuffer(scriptspan, letter_offset, scoringcontext, doc_tote, vec, more_to_come, score_cjk, hitbuffer); SpliceHitBuffer(hitbuffer, next_offset); letter_offset = next_offset; } delete hitbuffer; // Context across buffers is not connected yet scoringcontext->prior_chunk_lang = UNKNOWN_LANGUAGE; } // Score RTypeMany scriptspan into doc_tote and vec, updating scoringcontext // We have a scriptspan with all lowercase text in one script. Look up // quadgrams and octagrams, saving the hits in three parallel vectors. // Score from those vectors in chunks, toting each chunk to get a single // language, and combining into the overall document score. The hit vectors // in general are not big enough to handle and entire scriptspan, so // repeat until the entire scriptspan is scored. // Caller deals with minimizing numbr of runt scriptspans // This routine deals with minimizing number of runt chunks. // // Returns updated scoringcontext // Returns updated doc_tote // If vec != NULL, appends to that vector of ResultChunk's void ScoreQuadScriptSpan(const LangSpan& scriptspan, ScoringContext* scoringcontext, DocTote* doc_tote, ResultChunkVector* vec) { // Allocate three parallel arrays of scoring hits ScoringHitBuffer* hitbuffer = new ScoringHitBuffer; hitbuffer->init(); hitbuffer->ulscript = scriptspan.ulscript; scoringcontext->prior_chunk_lang = UNKNOWN_LANGUAGE; scoringcontext->oldest_distinct_boost = 0; // Incoming scriptspan has a single leading space at scriptspan.text[0] // and three trailing spaces then NUL at scriptspan.text[text_bytes + 0/1/2/3] int letter_offset = 1; // Skip initial space hitbuffer->lowest_offset = letter_offset; int letter_limit = scriptspan.text_bytes; while (letter_offset < letter_limit) { // // Fill up one hitbuffer, possibly splicing onto previous fragment // // NOTE: GetQuadHits deals with close repeats // NOTE: After last chunk there is always a hitbuffer entry with an offset // just off the end of the text = next_offset. int next_offset = GetQuadHits(scriptspan.text, letter_offset, letter_limit, scoringcontext, hitbuffer); // If true, there is more text to process in this scriptspan // NOTE: GetOctaHitVectors deals with close repeats, // does one hash and two lookups (delta and distinct) per word GetOctaHits(scriptspan.text, letter_offset, next_offset, scoringcontext, hitbuffer); // // Score one hitbuffer in chunks to summarybuffer // bool more_to_come = next_offset < letter_limit; bool score_cjk = false; ProcessHitBuffer(scriptspan, letter_offset, scoringcontext, doc_tote, vec, more_to_come, score_cjk, hitbuffer); SpliceHitBuffer(hitbuffer, next_offset); letter_offset = next_offset; } delete hitbuffer; } // Score one scriptspan into doc_tote and vec, updating scoringcontext // Inputs: // One scriptspan of perhaps 40-60KB, all same script lower-case letters // and single ASCII spaces. First character is a space to allow simple // begining-of-word detect. End of buffer has three spaces and NUL to // allow easy scan-to-end-of-word. // Scoring context of // scoring tables // flags // running boosts // Outputs: // Updated doc_tote giving overall languages and byte counts // Optional updated chunk vector giving offset, length, language // // Caller initializes flags, boosts, doc_tote and vec. // Caller aggregates across multiple scriptspans // Caller calculates final document result // Caller deals with detecting and triggering suppression of repeated text. // // This top-level routine just chooses the recognition type and calls one of // the next-level-down routines. // void ScoreOneScriptSpan(const LangSpan& scriptspan, ScoringContext* scoringcontext, DocTote* doc_tote, ResultChunkVector* vec) { if (scoringcontext->flags_cld2_verbose) { fprintf(scoringcontext->debug_file, "
ScoreOneScriptSpan(%s,%d) ", ULScriptCode(scriptspan.ulscript), scriptspan.text_bytes); // Optionally print the chunk lowercase letters/marks text string temp(&scriptspan.text[0], scriptspan.text_bytes); fprintf(scoringcontext->debug_file, "'%s'", GetHtmlEscapedText(temp).c_str()); fprintf(scoringcontext->debug_file, "
\n"); } scoringcontext->prior_chunk_lang = UNKNOWN_LANGUAGE; scoringcontext->oldest_distinct_boost = 0; ULScriptRType rtype = ULScriptRecognitionType(scriptspan.ulscript); if (scoringcontext->flags_cld2_score_as_quads && (rtype != RTypeCJK)) { rtype = RTypeMany; } switch (rtype) { case RTypeNone: case RTypeOne: ScoreEntireScriptSpan(scriptspan, scoringcontext, doc_tote, vec); break; case RTypeCJK: ScoreCJKScriptSpan(scriptspan, scoringcontext, doc_tote, vec); break; case RTypeMany: ScoreQuadScriptSpan(scriptspan, scoringcontext, doc_tote, vec); break; } } } // End namespace CLD2