2020-08-11 09:10:23 +00:00
|
|
|
// © 2016 and later: Unicode, Inc. and others.
|
|
|
|
// License & terms of use: http://www.unicode.org/copyright.html
|
|
|
|
/*
|
|
|
|
*******************************************************************************
|
|
|
|
*
|
|
|
|
* Copyright (C) 2005-2016, International Business Machines
|
|
|
|
* Corporation and others. All Rights Reserved.
|
|
|
|
*
|
|
|
|
*******************************************************************************
|
|
|
|
* file name: utext.cpp
|
|
|
|
* encoding: UTF-8
|
|
|
|
* tab size: 8 (not used)
|
|
|
|
* indentation:4
|
|
|
|
*
|
|
|
|
* created on: 2005apr12
|
|
|
|
* created by: Markus W. Scherer
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <cstddef>
|
|
|
|
|
|
|
|
#include "unicode/utypes.h"
|
|
|
|
#include "unicode/ustring.h"
|
|
|
|
#include "unicode/unistr.h"
|
|
|
|
#include "unicode/chariter.h"
|
|
|
|
#include "unicode/utext.h"
|
|
|
|
#include "unicode/utf.h"
|
|
|
|
#include "unicode/utf8.h"
|
|
|
|
#include "unicode/utf16.h"
|
|
|
|
#include "ustr_imp.h"
|
|
|
|
#include "cmemory.h"
|
|
|
|
#include "cstring.h"
|
|
|
|
#include "uassert.h"
|
|
|
|
#include "putilimp.h"
|
|
|
|
|
|
|
|
U_NAMESPACE_USE
|
|
|
|
|
|
|
|
#define I32_FLAG(bitIndex) ((int32_t)1<<(bitIndex))
|
|
|
|
|
|
|
|
|
|
|
|
static UBool
|
|
|
|
utext_access(UText *ut, int64_t index, UBool forward) {
|
|
|
|
return ut->pFuncs->access(ut, index, forward);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UBool U_EXPORT2
|
|
|
|
utext_moveIndex32(UText *ut, int32_t delta) {
|
|
|
|
UChar32 c;
|
|
|
|
if (delta > 0) {
|
|
|
|
do {
|
2022-10-28 06:11:55 +00:00
|
|
|
if(ut->chunkOffset>=ut->chunkLength && !utext_access(ut, ut->chunkNativeLimit, true)) {
|
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
c = ut->chunkContents[ut->chunkOffset];
|
|
|
|
if (U16_IS_SURROGATE(c)) {
|
|
|
|
c = utext_next32(ut);
|
|
|
|
if (c == U_SENTINEL) {
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset++;
|
|
|
|
}
|
|
|
|
} while(--delta>0);
|
|
|
|
|
|
|
|
} else if (delta<0) {
|
|
|
|
do {
|
2022-10-28 06:11:55 +00:00
|
|
|
if(ut->chunkOffset<=0 && !utext_access(ut, ut->chunkNativeStart, false)) {
|
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
c = ut->chunkContents[ut->chunkOffset-1];
|
|
|
|
if (U16_IS_SURROGATE(c)) {
|
|
|
|
c = utext_previous32(ut);
|
|
|
|
if (c == U_SENTINEL) {
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset--;
|
|
|
|
}
|
|
|
|
} while(++delta<0);
|
|
|
|
}
|
|
|
|
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI int64_t U_EXPORT2
|
|
|
|
utext_nativeLength(UText *ut) {
|
|
|
|
return ut->pFuncs->nativeLength(ut);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UBool U_EXPORT2
|
|
|
|
utext_isLengthExpensive(const UText *ut) {
|
|
|
|
UBool r = (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE)) != 0;
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI int64_t U_EXPORT2
|
|
|
|
utext_getNativeIndex(const UText *ut) {
|
|
|
|
if(ut->chunkOffset <= ut->nativeIndexingLimit) {
|
|
|
|
return ut->chunkNativeStart+ut->chunkOffset;
|
|
|
|
} else {
|
|
|
|
return ut->pFuncs->mapOffsetToNative(ut);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI void U_EXPORT2
|
|
|
|
utext_setNativeIndex(UText *ut, int64_t index) {
|
|
|
|
if(index<ut->chunkNativeStart || index>=ut->chunkNativeLimit) {
|
|
|
|
// The desired position is outside of the current chunk.
|
|
|
|
// Access the new position. Assume a forward iteration from here,
|
|
|
|
// which will also be optimimum for a single random access.
|
|
|
|
// Reverse iterations may suffer slightly.
|
2022-10-28 06:11:55 +00:00
|
|
|
ut->pFuncs->access(ut, index, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
} else if((int32_t)(index - ut->chunkNativeStart) <= ut->nativeIndexingLimit) {
|
|
|
|
// utf-16 indexing.
|
|
|
|
ut->chunkOffset=(int32_t)(index-ut->chunkNativeStart);
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset=ut->pFuncs->mapNativeIndexToUTF16(ut, index);
|
|
|
|
}
|
|
|
|
// The convention is that the index must always be on a code point boundary.
|
|
|
|
// Adjust the index position if it is in the middle of a surrogate pair.
|
|
|
|
if (ut->chunkOffset<ut->chunkLength) {
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t c= ut->chunkContents[ut->chunkOffset];
|
2020-08-11 09:10:23 +00:00
|
|
|
if (U16_IS_TRAIL(c)) {
|
|
|
|
if (ut->chunkOffset==0) {
|
2022-10-28 06:11:55 +00:00
|
|
|
ut->pFuncs->access(ut, ut->chunkNativeStart, false);
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
if (ut->chunkOffset>0) {
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t lead = ut->chunkContents[ut->chunkOffset-1];
|
2020-08-11 09:10:23 +00:00
|
|
|
if (U16_IS_LEAD(lead)) {
|
|
|
|
ut->chunkOffset--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI int64_t U_EXPORT2
|
|
|
|
utext_getPreviousNativeIndex(UText *ut) {
|
|
|
|
//
|
|
|
|
// Fast-path the common case.
|
|
|
|
// Common means current position is not at the beginning of a chunk
|
|
|
|
// and the preceding character is not supplementary.
|
|
|
|
//
|
|
|
|
int32_t i = ut->chunkOffset - 1;
|
|
|
|
int64_t result;
|
|
|
|
if (i >= 0) {
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t c = ut->chunkContents[i];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_TRAIL(c) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
if (i <= ut->nativeIndexingLimit) {
|
|
|
|
result = ut->chunkNativeStart + i;
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset = i;
|
|
|
|
result = ut->pFuncs->mapOffsetToNative(ut);
|
|
|
|
ut->chunkOffset++;
|
|
|
|
}
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If at the start of text, simply return 0.
|
|
|
|
if (ut->chunkOffset==0 && ut->chunkNativeStart==0) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Harder, less common cases. We are at a chunk boundary, or on a surrogate.
|
|
|
|
// Keep it simple, use other functions to handle the edges.
|
|
|
|
//
|
|
|
|
utext_previous32(ut);
|
|
|
|
result = UTEXT_GETNATIVEINDEX(ut);
|
|
|
|
utext_next32(ut);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// utext_current32. Get the UChar32 at the current position.
|
|
|
|
// UText iteration position is always on a code point boundary,
|
|
|
|
// never on the trail half of a surrogate pair.
|
|
|
|
//
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_current32(UText *ut) {
|
|
|
|
UChar32 c;
|
|
|
|
if (ut->chunkOffset==ut->chunkLength) {
|
|
|
|
// Current position is just off the end of the chunk.
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, ut->chunkNativeLimit, true) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// Off the end of the text.
|
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c = ut->chunkContents[ut->chunkOffset];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_LEAD(c) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// Normal, non-supplementary case.
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Possible supplementary char.
|
|
|
|
//
|
|
|
|
UChar32 trail = 0;
|
|
|
|
UChar32 supplementaryC = c;
|
|
|
|
if ((ut->chunkOffset+1) < ut->chunkLength) {
|
|
|
|
// The trail surrogate is in the same chunk.
|
|
|
|
trail = ut->chunkContents[ut->chunkOffset+1];
|
|
|
|
} else {
|
|
|
|
// The trail surrogate is in a different chunk.
|
|
|
|
// Because we must maintain the iteration position, we need to switch forward
|
|
|
|
// into the new chunk, get the trail surrogate, then revert the chunk back to the
|
|
|
|
// original one.
|
|
|
|
// An edge case to be careful of: the entire text may end with an unpaired
|
|
|
|
// leading surrogate. The attempt to access the trail will fail, but
|
|
|
|
// the original position before the unpaired lead still needs to be restored.
|
|
|
|
int64_t nativePosition = ut->chunkNativeLimit;
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, nativePosition, true)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
trail = ut->chunkContents[ut->chunkOffset];
|
|
|
|
}
|
2022-10-28 06:11:55 +00:00
|
|
|
UBool r = ut->pFuncs->access(ut, nativePosition, false); // reverse iteration flag loads preceding chunk
|
2023-05-23 00:05:01 +00:00
|
|
|
U_ASSERT(r);
|
|
|
|
// Here we need to restore chunkOffset since the access functions were called with
|
|
|
|
// chunkNativeLimit but that is not where we were (we were 1 code unit before the
|
|
|
|
// limit). Restoring was originally added in ICU-4669 but did not support access
|
|
|
|
// functions that changed the chunk size, the following does.
|
|
|
|
ut->chunkOffset = ut->chunkLength - 1;
|
2020-08-11 09:10:23 +00:00
|
|
|
if(!r) {
|
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (U16_IS_TRAIL(trail)) {
|
|
|
|
supplementaryC = U16_GET_SUPPLEMENTARY(c, trail);
|
|
|
|
}
|
|
|
|
return supplementaryC;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_char32At(UText *ut, int64_t nativeIndex) {
|
|
|
|
UChar32 c = U_SENTINEL;
|
|
|
|
|
|
|
|
// Fast path the common case.
|
|
|
|
if (nativeIndex>=ut->chunkNativeStart && nativeIndex < ut->chunkNativeStart + ut->nativeIndexingLimit) {
|
|
|
|
ut->chunkOffset = (int32_t)(nativeIndex - ut->chunkNativeStart);
|
|
|
|
c = ut->chunkContents[ut->chunkOffset];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_SURROGATE(c) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
return c;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
utext_setNativeIndex(ut, nativeIndex);
|
|
|
|
if (nativeIndex>=ut->chunkNativeStart && ut->chunkOffset<ut->chunkLength) {
|
|
|
|
c = ut->chunkContents[ut->chunkOffset];
|
|
|
|
if (U16_IS_SURROGATE(c)) {
|
|
|
|
// For surrogates, let current32() deal with the complications
|
|
|
|
// of supplementaries that may span chunk boundaries.
|
|
|
|
c = utext_current32(ut);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_next32(UText *ut) {
|
|
|
|
UChar32 c;
|
|
|
|
|
|
|
|
if (ut->chunkOffset >= ut->chunkLength) {
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, ut->chunkNativeLimit, true) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c = ut->chunkContents[ut->chunkOffset++];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_LEAD(c) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// Normal case, not supplementary.
|
|
|
|
// (A trail surrogate seen here is just returned as is, as a surrogate value.
|
|
|
|
// It cannot be part of a pair.)
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ut->chunkOffset >= ut->chunkLength) {
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, ut->chunkNativeLimit, true) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// c is an unpaired lead surrogate at the end of the text.
|
|
|
|
// return it as it is.
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
UChar32 trail = ut->chunkContents[ut->chunkOffset];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_TRAIL(trail) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// c was an unpaired lead surrogate, not at the end of the text.
|
|
|
|
// return it as it is (unpaired). Iteration position is on the
|
|
|
|
// following character, possibly in the next chunk, where the
|
|
|
|
// trail surrogate would have been if it had existed.
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
UChar32 supplementary = U16_GET_SUPPLEMENTARY(c, trail);
|
|
|
|
ut->chunkOffset++; // move iteration position over the trail surrogate.
|
|
|
|
return supplementary;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_previous32(UText *ut) {
|
|
|
|
UChar32 c;
|
|
|
|
|
|
|
|
if (ut->chunkOffset <= 0) {
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, ut->chunkNativeStart, false) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ut->chunkOffset--;
|
|
|
|
c = ut->chunkContents[ut->chunkOffset];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_TRAIL(c) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// Normal case, not supplementary.
|
|
|
|
// (A lead surrogate seen here is just returned as is, as a surrogate value.
|
|
|
|
// It cannot be part of a pair.)
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ut->chunkOffset <= 0) {
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->pFuncs->access(ut, ut->chunkNativeStart, false) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// c is an unpaired trail surrogate at the start of the text.
|
|
|
|
// return it as it is.
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
UChar32 lead = ut->chunkContents[ut->chunkOffset-1];
|
2022-10-28 06:11:55 +00:00
|
|
|
if (U16_IS_LEAD(lead) == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// c was an unpaired trail surrogate, not at the end of the text.
|
|
|
|
// return it as it is (unpaired). Iteration position is at c
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
UChar32 supplementary = U16_GET_SUPPLEMENTARY(lead, c);
|
|
|
|
ut->chunkOffset--; // move iteration position over the lead surrogate.
|
|
|
|
return supplementary;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_next32From(UText *ut, int64_t index) {
|
|
|
|
UChar32 c = U_SENTINEL;
|
|
|
|
|
|
|
|
if(index<ut->chunkNativeStart || index>=ut->chunkNativeLimit) {
|
|
|
|
// Desired position is outside of the current chunk.
|
2022-10-28 06:11:55 +00:00
|
|
|
if(!ut->pFuncs->access(ut, index, true)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// no chunk available here
|
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
} else if (index - ut->chunkNativeStart <= (int64_t)ut->nativeIndexingLimit) {
|
|
|
|
// Desired position is in chunk, with direct 1:1 native to UTF16 indexing
|
|
|
|
ut->chunkOffset = (int32_t)(index - ut->chunkNativeStart);
|
|
|
|
} else {
|
|
|
|
// Desired position is in chunk, with non-UTF16 indexing.
|
|
|
|
ut->chunkOffset = ut->pFuncs->mapNativeIndexToUTF16(ut, index);
|
|
|
|
}
|
|
|
|
|
|
|
|
c = ut->chunkContents[ut->chunkOffset++];
|
|
|
|
if (U16_IS_SURROGATE(c)) {
|
|
|
|
// Surrogates. Many edge cases. Use other functions that already
|
|
|
|
// deal with the problems.
|
|
|
|
utext_setNativeIndex(ut, index);
|
|
|
|
c = utext_next32(ut);
|
|
|
|
}
|
|
|
|
return c;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UChar32 U_EXPORT2
|
|
|
|
utext_previous32From(UText *ut, int64_t index) {
|
|
|
|
//
|
|
|
|
// Return the character preceding the specified index.
|
|
|
|
// Leave the iteration position at the start of the character that was returned.
|
|
|
|
//
|
|
|
|
UChar32 cPrev; // The character preceding cCurr, which is what we will return.
|
|
|
|
|
2021-04-22 12:08:59 +00:00
|
|
|
// Address the chunk containing the position preceding the incoming index
|
2020-08-11 09:10:23 +00:00
|
|
|
// A tricky edge case:
|
|
|
|
// We try to test the requested native index against the chunkNativeStart to determine
|
|
|
|
// whether the character preceding the one at the index is in the current chunk.
|
|
|
|
// BUT, this test can fail with UTF-8 (or any other multibyte encoding), when the
|
|
|
|
// requested index is on something other than the first position of the first char.
|
|
|
|
//
|
|
|
|
if(index<=ut->chunkNativeStart || index>ut->chunkNativeLimit) {
|
|
|
|
// Requested native index is outside of the current chunk.
|
2022-10-28 06:11:55 +00:00
|
|
|
if(!ut->pFuncs->access(ut, index, false)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// no chunk available here
|
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
} else if(index - ut->chunkNativeStart <= (int64_t)ut->nativeIndexingLimit) {
|
|
|
|
// Direct UTF-16 indexing.
|
|
|
|
ut->chunkOffset = (int32_t)(index - ut->chunkNativeStart);
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset=ut->pFuncs->mapNativeIndexToUTF16(ut, index);
|
2022-10-28 06:11:55 +00:00
|
|
|
if (ut->chunkOffset==0 && !ut->pFuncs->access(ut, index, false)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// no chunk available here
|
|
|
|
return U_SENTINEL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Simple case with no surrogates.
|
|
|
|
//
|
|
|
|
ut->chunkOffset--;
|
|
|
|
cPrev = ut->chunkContents[ut->chunkOffset];
|
|
|
|
|
|
|
|
if (U16_IS_SURROGATE(cPrev)) {
|
|
|
|
// Possible supplementary. Many edge cases.
|
|
|
|
// Let other functions do the heavy lifting.
|
|
|
|
utext_setNativeIndex(ut, index);
|
|
|
|
cPrev = utext_previous32(ut);
|
|
|
|
}
|
|
|
|
return cPrev;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI int32_t U_EXPORT2
|
|
|
|
utext_extract(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *status) {
|
|
|
|
return ut->pFuncs->extract(ut, start, limit, dest, destCapacity, status);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UBool U_EXPORT2
|
|
|
|
utext_equals(const UText *a, const UText *b) {
|
2023-05-23 00:05:01 +00:00
|
|
|
if (a==nullptr || b==nullptr ||
|
2020-08-11 09:10:23 +00:00
|
|
|
a->magic != UTEXT_MAGIC ||
|
|
|
|
b->magic != UTEXT_MAGIC) {
|
|
|
|
// Null or invalid arguments don't compare equal to anything.
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (a->pFuncs != b->pFuncs) {
|
|
|
|
// Different types of text providers.
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (a->context != b->context) {
|
|
|
|
// Different sources (different strings)
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
if (utext_getNativeIndex(a) != utext_getNativeIndex(b)) {
|
|
|
|
// Different current position in the string.
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
U_CAPI UBool U_EXPORT2
|
|
|
|
utext_isWritable(const UText *ut)
|
|
|
|
{
|
|
|
|
UBool b = (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_WRITABLE)) != 0;
|
|
|
|
return b;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI void U_EXPORT2
|
|
|
|
utext_freeze(UText *ut) {
|
|
|
|
// Zero out the WRITABLE flag.
|
|
|
|
ut->providerProperties &= ~(I32_FLAG(UTEXT_PROVIDER_WRITABLE));
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UBool U_EXPORT2
|
|
|
|
utext_hasMetaData(const UText *ut)
|
|
|
|
{
|
|
|
|
UBool b = (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_HAS_META_DATA)) != 0;
|
|
|
|
return b;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI int32_t U_EXPORT2
|
|
|
|
utext_replace(UText *ut,
|
|
|
|
int64_t nativeStart, int64_t nativeLimit,
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *replacementText, int32_t replacementLength,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *status)
|
|
|
|
{
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if ((ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_WRITABLE)) == 0) {
|
|
|
|
*status = U_NO_WRITE_PERMISSION;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
int32_t i = ut->pFuncs->replace(ut, nativeStart, nativeLimit, replacementText, replacementLength, status);
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
|
|
|
|
U_CAPI void U_EXPORT2
|
|
|
|
utext_copy(UText *ut,
|
|
|
|
int64_t nativeStart, int64_t nativeLimit,
|
|
|
|
int64_t destIndex,
|
|
|
|
UBool move,
|
|
|
|
UErrorCode *status)
|
|
|
|
{
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if ((ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_WRITABLE)) == 0) {
|
|
|
|
*status = U_NO_WRITE_PERMISSION;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
ut->pFuncs->copy(ut, nativeStart, nativeLimit, destIndex, move, status);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_clone(UText *dest, const UText *src, UBool deep, UBool readOnly, UErrorCode *status) {
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
UText *result = src->pFuncs->clone(dest, src, deep, status);
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return result;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if (result == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_MEMORY_ALLOCATION_ERROR;
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
if (readOnly) {
|
|
|
|
utext_freeze(result);
|
|
|
|
}
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// UText common functions implementation
|
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
//
|
|
|
|
// UText.flags bit definitions
|
|
|
|
//
|
|
|
|
enum {
|
|
|
|
UTEXT_HEAP_ALLOCATED = 1, // 1 if ICU has allocated this UText struct on the heap.
|
|
|
|
// 0 if caller provided storage for the UText.
|
|
|
|
|
|
|
|
UTEXT_EXTRA_HEAP_ALLOCATED = 2, // 1 if ICU has allocated extra storage as a separate
|
|
|
|
// heap block.
|
|
|
|
// 0 if there is no separate allocation. Either no extra
|
|
|
|
// storage was requested, or it is appended to the end
|
|
|
|
// of the main UText storage.
|
|
|
|
|
|
|
|
UTEXT_OPEN = 4 // 1 if this UText is currently open
|
|
|
|
// 0 if this UText is not open.
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// Extended form of a UText. The purpose is to aid in computing the total size required
|
|
|
|
// when a provider asks for a UText to be allocated with extra storage.
|
|
|
|
|
|
|
|
struct ExtendedUText {
|
|
|
|
UText ut;
|
|
|
|
std::max_align_t extension;
|
|
|
|
};
|
|
|
|
|
|
|
|
static const UText emptyText = UTEXT_INITIALIZER;
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_setup(UText *ut, int32_t extraSpace, UErrorCode *status) {
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
if (ut == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// We need to heap-allocate storage for the new UText
|
|
|
|
int32_t spaceRequired = sizeof(UText);
|
|
|
|
if (extraSpace > 0) {
|
|
|
|
spaceRequired = sizeof(ExtendedUText) + extraSpace - sizeof(std::max_align_t);
|
|
|
|
}
|
|
|
|
ut = (UText *)uprv_malloc(spaceRequired);
|
2023-05-23 00:05:01 +00:00
|
|
|
if (ut == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_MEMORY_ALLOCATION_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
|
|
|
*ut = emptyText;
|
|
|
|
ut->flags |= UTEXT_HEAP_ALLOCATED;
|
|
|
|
if (spaceRequired>0) {
|
|
|
|
ut->extraSize = extraSpace;
|
|
|
|
ut->pExtra = &((ExtendedUText *)ut)->extension;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// We have been supplied with an already existing UText.
|
|
|
|
// Verify that it really appears to be a UText.
|
|
|
|
if (ut->magic != UTEXT_MAGIC) {
|
|
|
|
*status = U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
// If the ut is already open and there's a provider supplied close
|
|
|
|
// function, call it.
|
2023-05-23 00:05:01 +00:00
|
|
|
if ((ut->flags & UTEXT_OPEN) && ut->pFuncs->close != nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->pFuncs->close(ut);
|
|
|
|
}
|
|
|
|
ut->flags &= ~UTEXT_OPEN;
|
|
|
|
|
|
|
|
// If extra space was requested by our caller, check whether
|
|
|
|
// sufficient already exists, and allocate new if needed.
|
|
|
|
if (extraSpace > ut->extraSize) {
|
|
|
|
// Need more space. If there is existing separately allocated space,
|
|
|
|
// delete it first, then allocate new space.
|
|
|
|
if (ut->flags & UTEXT_EXTRA_HEAP_ALLOCATED) {
|
|
|
|
uprv_free(ut->pExtra);
|
|
|
|
ut->extraSize = 0;
|
|
|
|
}
|
|
|
|
ut->pExtra = uprv_malloc(extraSpace);
|
2023-05-23 00:05:01 +00:00
|
|
|
if (ut->pExtra == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_MEMORY_ALLOCATION_ERROR;
|
|
|
|
} else {
|
|
|
|
ut->extraSize = extraSpace;
|
|
|
|
ut->flags |= UTEXT_EXTRA_HEAP_ALLOCATED;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (U_SUCCESS(*status)) {
|
|
|
|
ut->flags |= UTEXT_OPEN;
|
|
|
|
|
|
|
|
// Initialize all remaining fields of the UText.
|
|
|
|
//
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->context = nullptr;
|
|
|
|
ut->chunkContents = nullptr;
|
|
|
|
ut->p = nullptr;
|
|
|
|
ut->q = nullptr;
|
|
|
|
ut->r = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->a = 0;
|
|
|
|
ut->b = 0;
|
|
|
|
ut->c = 0;
|
|
|
|
ut->chunkOffset = 0;
|
|
|
|
ut->chunkLength = 0;
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
ut->chunkNativeLimit = 0;
|
|
|
|
ut->nativeIndexingLimit = 0;
|
|
|
|
ut->providerProperties = 0;
|
|
|
|
ut->privA = 0;
|
|
|
|
ut->privB = 0;
|
|
|
|
ut->privC = 0;
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->privP = nullptr;
|
|
|
|
if (ut->pExtra!=nullptr && ut->extraSize>0)
|
2020-08-11 09:10:23 +00:00
|
|
|
uprv_memset(ut->pExtra, 0, ut->extraSize);
|
|
|
|
|
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_close(UText *ut) {
|
2023-05-23 00:05:01 +00:00
|
|
|
if (ut==nullptr ||
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->magic != UTEXT_MAGIC ||
|
|
|
|
(ut->flags & UTEXT_OPEN) == 0)
|
|
|
|
{
|
|
|
|
// The supplied ut is not an open UText.
|
|
|
|
// Do nothing.
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the provider gave us a close function, call it now.
|
|
|
|
// This will clean up anything allocated specifically by the provider.
|
2023-05-23 00:05:01 +00:00
|
|
|
if (ut->pFuncs->close != nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->pFuncs->close(ut);
|
|
|
|
}
|
|
|
|
ut->flags &= ~UTEXT_OPEN;
|
|
|
|
|
|
|
|
// If we (the framework) allocated the UText or subsidiary storage,
|
|
|
|
// delete it.
|
|
|
|
if (ut->flags & UTEXT_EXTRA_HEAP_ALLOCATED) {
|
|
|
|
uprv_free(ut->pExtra);
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->pExtra = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->flags &= ~UTEXT_EXTRA_HEAP_ALLOCATED;
|
|
|
|
ut->extraSize = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Zero out function table of the closed UText. This is a defensive move,
|
2021-10-28 06:15:28 +00:00
|
|
|
// intended to cause applications that inadvertently use a closed
|
2020-08-11 09:10:23 +00:00
|
|
|
// utext to crash with null pointer errors.
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->pFuncs = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
if (ut->flags & UTEXT_HEAP_ALLOCATED) {
|
|
|
|
// This UText was allocated by UText setup. We need to free it.
|
|
|
|
// Clear magic, so we can detect if the user messes up and immediately
|
|
|
|
// tries to reopen another UText using the deleted storage.
|
|
|
|
ut->magic = 0;
|
|
|
|
uprv_free(ut);
|
2023-05-23 00:05:01 +00:00
|
|
|
ut = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// invalidateChunk Reset a chunk to have no contents, so that the next call
|
|
|
|
// to access will cause new data to load.
|
|
|
|
// This is needed when copy/move/replace operate directly on the
|
|
|
|
// backing text, potentially putting it out of sync with the
|
|
|
|
// contents in the chunk.
|
|
|
|
//
|
|
|
|
static void
|
|
|
|
invalidateChunk(UText *ut) {
|
|
|
|
ut->chunkLength = 0;
|
|
|
|
ut->chunkNativeLimit = 0;
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
ut->chunkOffset = 0;
|
|
|
|
ut->nativeIndexingLimit = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// pinIndex Do range pinning on a native index parameter.
|
|
|
|
// 64 bit pinning is done in place.
|
|
|
|
// 32 bit truncated result is returned as a convenience for
|
|
|
|
// use in providers that don't need 64 bits.
|
|
|
|
static int32_t
|
|
|
|
pinIndex(int64_t &index, int64_t limit) {
|
|
|
|
if (index<0) {
|
|
|
|
index = 0;
|
|
|
|
} else if (index > limit) {
|
|
|
|
index = limit;
|
|
|
|
}
|
|
|
|
return (int32_t)index;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
|
|
|
|
//
|
|
|
|
// Pointer relocation function,
|
|
|
|
// a utility used by shallow clone.
|
|
|
|
// Adjust a pointer that refers to something within one UText (the source)
|
|
|
|
// to refer to the same relative offset within a another UText (the target)
|
|
|
|
//
|
|
|
|
static void adjustPointer(UText *dest, const void **destPtr, const UText *src) {
|
|
|
|
// convert all pointers to (char *) so that byte address arithmetic will work.
|
|
|
|
char *dptr = (char *)*destPtr;
|
|
|
|
char *dUText = (char *)dest;
|
|
|
|
char *sUText = (char *)src;
|
|
|
|
|
|
|
|
if (dptr >= (char *)src->pExtra && dptr < ((char*)src->pExtra)+src->extraSize) {
|
|
|
|
// target ptr was to something within the src UText's pExtra storage.
|
|
|
|
// relocate it into the target UText's pExtra region.
|
|
|
|
*destPtr = ((char *)dest->pExtra) + (dptr - (char *)src->pExtra);
|
|
|
|
} else if (dptr>=sUText && dptr < sUText+src->sizeOfStruct) {
|
|
|
|
// target ptr was pointing to somewhere within the source UText itself.
|
|
|
|
// Move it to the same offset within the target UText.
|
|
|
|
*destPtr = dUText + (dptr-sUText);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// Clone. This is a generic copy-the-utext-by-value clone function that can be
|
|
|
|
// used as-is with some utext types, and as a helper by other clones.
|
|
|
|
//
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
shallowTextClone(UText * dest, const UText * src, UErrorCode * status) {
|
|
|
|
if (U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
int32_t srcExtraSize = src->extraSize;
|
|
|
|
|
|
|
|
//
|
|
|
|
// Use the generic text_setup to allocate storage if required.
|
|
|
|
//
|
|
|
|
dest = utext_setup(dest, srcExtraSize, status);
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// flags (how the UText was allocated) and the pointer to the
|
|
|
|
// extra storage must retain the values in the cloned utext that
|
|
|
|
// were set up by utext_setup. Save them separately before
|
|
|
|
// copying the whole struct.
|
|
|
|
//
|
|
|
|
void *destExtra = dest->pExtra;
|
|
|
|
int32_t flags = dest->flags;
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// Copy the whole UText struct by value.
|
|
|
|
// Any "Extra" storage is copied also.
|
|
|
|
//
|
|
|
|
int sizeToCopy = src->sizeOfStruct;
|
|
|
|
if (sizeToCopy > dest->sizeOfStruct) {
|
|
|
|
sizeToCopy = dest->sizeOfStruct;
|
|
|
|
}
|
|
|
|
uprv_memcpy(dest, src, sizeToCopy);
|
|
|
|
dest->pExtra = destExtra;
|
|
|
|
dest->flags = flags;
|
|
|
|
if (srcExtraSize > 0) {
|
|
|
|
uprv_memcpy(dest->pExtra, src->pExtra, srcExtraSize);
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Relocate any pointers in the target that refer to the UText itself
|
|
|
|
// to point to the cloned copy rather than the original source.
|
|
|
|
//
|
|
|
|
adjustPointer(dest, &dest->context, src);
|
|
|
|
adjustPointer(dest, &dest->p, src);
|
|
|
|
adjustPointer(dest, &dest->q, src);
|
|
|
|
adjustPointer(dest, &dest->r, src);
|
|
|
|
adjustPointer(dest, (const void **)&dest->chunkContents, src);
|
|
|
|
|
|
|
|
// The newly shallow-cloned UText does _not_ own the underlying storage for the text.
|
|
|
|
// (The source for the clone may or may not have owned the text.)
|
|
|
|
|
|
|
|
dest->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT);
|
|
|
|
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
U_CDECL_END
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// UText implementation for UTF-8 char * strings (read-only)
|
|
|
|
// Limitation: string length must be <= 0x7fffffff in length.
|
|
|
|
// (length must for in an int32_t variable)
|
|
|
|
//
|
|
|
|
// Use of UText data members:
|
|
|
|
// context pointer to UTF-8 string
|
|
|
|
// utext.b is the input string length (bytes).
|
|
|
|
// utext.c Length scanned so far in string
|
|
|
|
// (for optimizing finding length of zero terminated strings.)
|
|
|
|
// utext.p pointer to the current buffer
|
|
|
|
// utext.q pointer to the other buffer.
|
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
// Chunk size.
|
2023-05-23 00:05:01 +00:00
|
|
|
// Must be less than 85 (256/3), because of byte mapping from char16_t indexes to native indexes.
|
|
|
|
// Worst case is three native bytes to one char16_t. (Supplemenaries are 4 native bytes
|
2020-08-11 09:10:23 +00:00
|
|
|
// to two UChars.)
|
|
|
|
// The longest illegal byte sequence treated as a single error (and converted to U+FFFD)
|
|
|
|
// is a three-byte sequence (truncated four-byte sequence).
|
|
|
|
//
|
|
|
|
enum { UTF8_TEXT_CHUNK_SIZE=32 };
|
|
|
|
|
|
|
|
//
|
|
|
|
// UTF8Buf Two of these structs will be set up in the UText's extra allocated space.
|
2023-05-23 00:05:01 +00:00
|
|
|
// Each contains the char16_t chunk buffer, the to and from native maps, and
|
2020-08-11 09:10:23 +00:00
|
|
|
// header info.
|
|
|
|
//
|
|
|
|
// because backwards iteration fills the buffers starting at the end and
|
|
|
|
// working towards the front, the filled part of the buffers may not begin
|
|
|
|
// at the start of the available storage for the buffers.
|
|
|
|
//
|
|
|
|
// Buffer size is one bigger than the specified UTF8_TEXT_CHUNK_SIZE to allow for
|
|
|
|
// the last character added being a supplementary, and thus requiring a surrogate
|
|
|
|
// pair. Doing this is simpler than checking for the edge case.
|
|
|
|
//
|
|
|
|
|
|
|
|
struct UTF8Buf {
|
2023-05-23 00:05:01 +00:00
|
|
|
int32_t bufNativeStart; // Native index of first char in char16_t buf
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t bufNativeLimit; // Native index following last char in buf.
|
|
|
|
int32_t bufStartIdx; // First filled position in buf.
|
|
|
|
int32_t bufLimitIdx; // Limit of filled range in buf.
|
|
|
|
int32_t bufNILimit; // Limit of native indexing part of buf
|
|
|
|
int32_t toUCharsMapStart; // Native index corresponding to
|
|
|
|
// mapToUChars[0].
|
|
|
|
// Set to bufNativeStart when filling forwards.
|
|
|
|
// Set to computed value when filling backwards.
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t buf[UTF8_TEXT_CHUNK_SIZE+4]; // The char16_t buffer. Requires one extra position beyond the
|
2020-08-11 09:10:23 +00:00
|
|
|
// the chunk size, to allow for surrogate at the end.
|
|
|
|
// Length must be identical to mapToNative array, below,
|
|
|
|
// because of the way indexing works when the array is
|
|
|
|
// filled backwards during a reverse iteration. Thus,
|
|
|
|
// the additional extra size.
|
2023-05-23 00:05:01 +00:00
|
|
|
uint8_t mapToNative[UTF8_TEXT_CHUNK_SIZE+4]; // map char16_t index in buf to
|
2020-08-11 09:10:23 +00:00
|
|
|
// native offset from bufNativeStart.
|
|
|
|
// Requires two extra slots,
|
|
|
|
// one for a supplementary starting in the last normal position,
|
|
|
|
// and one for an entry for the buffer limit position.
|
|
|
|
uint8_t mapToUChars[UTF8_TEXT_CHUNK_SIZE*3+6]; // Map native offset from bufNativeStart to
|
2021-04-22 12:08:59 +00:00
|
|
|
// corresponding offset in filled part of buf.
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t align;
|
|
|
|
};
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
|
|
|
|
//
|
|
|
|
// utf8TextLength
|
|
|
|
//
|
|
|
|
// Get the length of the string. If we don't already know it,
|
|
|
|
// we'll need to scan for the trailing nul.
|
|
|
|
//
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
utf8TextLength(UText *ut) {
|
|
|
|
if (ut->b < 0) {
|
|
|
|
// Zero terminated string, and we haven't scanned to the end yet.
|
|
|
|
// Scan it now.
|
|
|
|
const char *r = (const char *)ut->context + ut->c;
|
|
|
|
while (*r != 0) {
|
|
|
|
r++;
|
|
|
|
}
|
|
|
|
if ((r - (const char *)ut->context) < 0x7fffffff) {
|
|
|
|
ut->b = (int32_t)(r - (const char *)ut->context);
|
|
|
|
} else {
|
|
|
|
// Actual string was bigger (more than 2 gig) than we
|
|
|
|
// can handle. Clip it to 2 GB.
|
|
|
|
ut->b = 0x7fffffff;
|
|
|
|
}
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
return ut->b;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static UBool U_CALLCONV
|
|
|
|
utf8TextAccess(UText *ut, int64_t index, UBool forward) {
|
|
|
|
//
|
|
|
|
// Apologies to those who are allergic to goto statements.
|
|
|
|
// Consider each goto to a labelled block to be the equivalent of
|
|
|
|
// call the named block as if it were a function();
|
|
|
|
// return;
|
|
|
|
//
|
|
|
|
const uint8_t *s8=(const uint8_t *)ut->context;
|
2023-05-23 00:05:01 +00:00
|
|
|
UTF8Buf *u8b = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t length = ut->b; // Length of original utf-8
|
|
|
|
int32_t ix= (int32_t)index; // Requested index, trimmed to 32 bits.
|
|
|
|
int32_t mapIndex = 0;
|
|
|
|
if (index<0) {
|
|
|
|
ix=0;
|
|
|
|
} else if (index > 0x7fffffff) {
|
|
|
|
// Strings with 64 bit lengths not supported by this UTF-8 provider.
|
|
|
|
ix = 0x7fffffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Pin requested index to the string length.
|
|
|
|
if (ix>length) {
|
|
|
|
if (length>=0) {
|
|
|
|
ix=length;
|
|
|
|
} else if (ix>=ut->c) {
|
|
|
|
// Zero terminated string, and requested index is beyond
|
|
|
|
// the region that has already been scanned.
|
|
|
|
// Scan up to either the end of the string or to the
|
|
|
|
// requested position, whichever comes first.
|
|
|
|
while (ut->c<ix && s8[ut->c]!=0) {
|
|
|
|
ut->c++;
|
|
|
|
}
|
|
|
|
// TODO: support for null terminated string length > 32 bits.
|
|
|
|
if (s8[ut->c] == 0) {
|
|
|
|
// We just found the actual length of the string.
|
|
|
|
// Trim the requested index back to that.
|
|
|
|
ix = ut->c;
|
|
|
|
ut->b = ut->c;
|
|
|
|
length = ut->c;
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// Dispatch to the appropriate action for a forward iteration request.
|
|
|
|
//
|
|
|
|
if (forward) {
|
|
|
|
if (ix==ut->chunkNativeLimit) {
|
|
|
|
// Check for normal sequential iteration cases first.
|
|
|
|
if (ix==length) {
|
|
|
|
// Just reached end of string
|
|
|
|
// Don't swap buffers, but do set the
|
|
|
|
// current buffer position.
|
|
|
|
ut->chunkOffset = ut->chunkLength;
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
|
|
|
// End of current buffer.
|
|
|
|
// check whether other buffer already has what we need.
|
|
|
|
UTF8Buf *altB = (UTF8Buf *)ut->q;
|
|
|
|
if (ix>=altB->bufNativeStart && ix<altB->bufNativeLimit) {
|
|
|
|
goto swapBuffers;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// A random access. Desired index could be in either or niether buf.
|
|
|
|
// For optimizing the order of testing, first check for the index
|
|
|
|
// being in the other buffer. This will be the case for uses that
|
|
|
|
// move back and forth over a fairly limited range
|
|
|
|
{
|
|
|
|
u8b = (UTF8Buf *)ut->q; // the alternate buffer
|
|
|
|
if (ix>=u8b->bufNativeStart && ix<u8b->bufNativeLimit) {
|
|
|
|
// Requested index is in the other buffer.
|
|
|
|
goto swapBuffers;
|
|
|
|
}
|
|
|
|
if (ix == length) {
|
|
|
|
// Requested index is end-of-string.
|
|
|
|
// (this is the case of randomly seeking to the end.
|
|
|
|
// The case of iterating off the end is handled earlier.)
|
|
|
|
if (ix == ut->chunkNativeLimit) {
|
|
|
|
// Current buffer extends up to the end of the string.
|
|
|
|
// Leave it as the current buffer.
|
|
|
|
ut->chunkOffset = ut->chunkLength;
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
if (ix == u8b->bufNativeLimit) {
|
|
|
|
// Alternate buffer extends to the end of string.
|
|
|
|
// Swap it in as the current buffer.
|
|
|
|
goto swapBuffersAndFail;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Neither existing buffer extends to the end of the string.
|
|
|
|
goto makeStubBuffer;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ix<ut->chunkNativeStart || ix>=ut->chunkNativeLimit) {
|
|
|
|
// Requested index is in neither buffer.
|
|
|
|
goto fillForward;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Requested index is in this buffer.
|
|
|
|
u8b = (UTF8Buf *)ut->p; // the current buffer
|
|
|
|
mapIndex = ix - u8b->toUCharsMapStart;
|
|
|
|
U_ASSERT(mapIndex < (int32_t)sizeof(UTF8Buf::mapToUChars));
|
|
|
|
ut->chunkOffset = u8b->mapToUChars[mapIndex] - u8b->bufStartIdx;
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// Dispatch to the appropriate action for a
|
2021-10-28 06:15:28 +00:00
|
|
|
// Backwards Direction iteration request.
|
2020-08-11 09:10:23 +00:00
|
|
|
//
|
|
|
|
if (ix==ut->chunkNativeStart) {
|
|
|
|
// Check for normal sequential iteration cases first.
|
|
|
|
if (ix==0) {
|
|
|
|
// Just reached the start of string
|
|
|
|
// Don't swap buffers, but do set the
|
|
|
|
// current buffer position.
|
|
|
|
ut->chunkOffset = 0;
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
|
|
|
// Start of current buffer.
|
|
|
|
// check whether other buffer already has what we need.
|
|
|
|
UTF8Buf *altB = (UTF8Buf *)ut->q;
|
|
|
|
if (ix>altB->bufNativeStart && ix<=altB->bufNativeLimit) {
|
|
|
|
goto swapBuffers;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// A random access. Desired index could be in either or niether buf.
|
|
|
|
// For optimizing the order of testing,
|
|
|
|
// Most likely case: in the other buffer.
|
|
|
|
// Second most likely: in neither buffer.
|
|
|
|
// Unlikely, but must work: in the current buffer.
|
|
|
|
u8b = (UTF8Buf *)ut->q; // the alternate buffer
|
|
|
|
if (ix>u8b->bufNativeStart && ix<=u8b->bufNativeLimit) {
|
|
|
|
// Requested index is in the other buffer.
|
|
|
|
goto swapBuffers;
|
|
|
|
}
|
|
|
|
// Requested index is start-of-string.
|
|
|
|
// (this is the case of randomly seeking to the start.
|
|
|
|
// The case of iterating off the start is handled earlier.)
|
|
|
|
if (ix==0) {
|
|
|
|
if (u8b->bufNativeStart==0) {
|
|
|
|
// Alternate buffer contains the data for the start string.
|
|
|
|
// Make it be the current buffer.
|
|
|
|
goto swapBuffersAndFail;
|
|
|
|
} else {
|
|
|
|
// Request for data before the start of string,
|
|
|
|
// neither buffer is usable.
|
|
|
|
// set up a zero-length buffer.
|
|
|
|
goto makeStubBuffer;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ix<=ut->chunkNativeStart || ix>ut->chunkNativeLimit) {
|
|
|
|
// Requested index is in neither buffer.
|
|
|
|
goto fillReverse;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Requested index is in this buffer.
|
|
|
|
// Set the utf16 buffer index.
|
|
|
|
u8b = (UTF8Buf *)ut->p;
|
|
|
|
mapIndex = ix - u8b->toUCharsMapStart;
|
|
|
|
ut->chunkOffset = u8b->mapToUChars[mapIndex] - u8b->bufStartIdx;
|
|
|
|
if (ut->chunkOffset==0) {
|
|
|
|
// This occurs when the first character in the text is
|
|
|
|
// a multi-byte UTF-8 char, and the requested index is to
|
|
|
|
// one of the trailing bytes. Because there is no preceding ,
|
|
|
|
// character, this access fails. We can't pick up on the
|
|
|
|
// situation sooner because the requested index is not zero.
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
swapBuffers:
|
|
|
|
// The alternate buffer (ut->q) has the string data that was requested.
|
|
|
|
// Swap the primary and alternate buffers, and set the
|
|
|
|
// chunk index into the new primary buffer.
|
|
|
|
{
|
|
|
|
u8b = (UTF8Buf *)ut->q;
|
|
|
|
ut->q = ut->p;
|
|
|
|
ut->p = u8b;
|
|
|
|
ut->chunkContents = &u8b->buf[u8b->bufStartIdx];
|
|
|
|
ut->chunkLength = u8b->bufLimitIdx - u8b->bufStartIdx;
|
|
|
|
ut->chunkNativeStart = u8b->bufNativeStart;
|
|
|
|
ut->chunkNativeLimit = u8b->bufNativeLimit;
|
|
|
|
ut->nativeIndexingLimit = u8b->bufNILimit;
|
|
|
|
|
|
|
|
// Index into the (now current) chunk
|
|
|
|
// Use the map to set the chunk index. It's more trouble than it's worth
|
|
|
|
// to check whether native indexing can be used.
|
|
|
|
U_ASSERT(ix>=u8b->bufNativeStart);
|
|
|
|
U_ASSERT(ix<=u8b->bufNativeLimit);
|
|
|
|
mapIndex = ix - u8b->toUCharsMapStart;
|
|
|
|
U_ASSERT(mapIndex>=0);
|
|
|
|
U_ASSERT(mapIndex<(int32_t)sizeof(u8b->mapToUChars));
|
|
|
|
ut->chunkOffset = u8b->mapToUChars[mapIndex] - u8b->bufStartIdx;
|
|
|
|
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
swapBuffersAndFail:
|
|
|
|
// We got a request for either the start or end of the string,
|
|
|
|
// with iteration continuing in the out-of-bounds direction.
|
|
|
|
// The alternate buffer already contains the data up to the
|
|
|
|
// start/end.
|
|
|
|
// Swap the buffers, then return failure, indicating that we couldn't
|
|
|
|
// make things correct for continuing the iteration in the requested
|
|
|
|
// direction. The position & buffer are correct should the
|
|
|
|
// user decide to iterate in the opposite direction.
|
|
|
|
u8b = (UTF8Buf *)ut->q;
|
|
|
|
ut->q = ut->p;
|
|
|
|
ut->p = u8b;
|
|
|
|
ut->chunkContents = &u8b->buf[u8b->bufStartIdx];
|
|
|
|
ut->chunkLength = u8b->bufLimitIdx - u8b->bufStartIdx;
|
|
|
|
ut->chunkNativeStart = u8b->bufNativeStart;
|
|
|
|
ut->chunkNativeLimit = u8b->bufNativeLimit;
|
|
|
|
ut->nativeIndexingLimit = u8b->bufNILimit;
|
|
|
|
|
|
|
|
// Index into the (now current) chunk
|
|
|
|
// For this function (swapBuffersAndFail), the requested index
|
|
|
|
// will always be at either the start or end of the chunk.
|
|
|
|
if (ix==u8b->bufNativeLimit) {
|
|
|
|
ut->chunkOffset = ut->chunkLength;
|
|
|
|
} else {
|
|
|
|
ut->chunkOffset = 0;
|
|
|
|
U_ASSERT(ix == u8b->bufNativeStart);
|
|
|
|
}
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
makeStubBuffer:
|
|
|
|
// The user has done a seek/access past the start or end
|
|
|
|
// of the string. Rather than loading data that is likely
|
|
|
|
// to never be used, just set up a zero-length buffer at
|
|
|
|
// the position.
|
|
|
|
u8b = (UTF8Buf *)ut->q;
|
|
|
|
u8b->bufNativeStart = ix;
|
|
|
|
u8b->bufNativeLimit = ix;
|
|
|
|
u8b->bufStartIdx = 0;
|
|
|
|
u8b->bufLimitIdx = 0;
|
|
|
|
u8b->bufNILimit = 0;
|
|
|
|
u8b->toUCharsMapStart = ix;
|
|
|
|
u8b->mapToNative[0] = 0;
|
|
|
|
u8b->mapToUChars[0] = 0;
|
|
|
|
goto swapBuffersAndFail;
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
fillForward:
|
|
|
|
{
|
|
|
|
// Move the incoming index to a code point boundary.
|
|
|
|
U8_SET_CP_START(s8, 0, ix);
|
|
|
|
|
|
|
|
// Swap the UText buffers.
|
|
|
|
// We want to fill what was previously the alternate buffer,
|
|
|
|
// and make what was the current buffer be the new alternate.
|
|
|
|
UTF8Buf *u8b_swap = (UTF8Buf *)ut->q;
|
|
|
|
ut->q = ut->p;
|
|
|
|
ut->p = u8b_swap;
|
|
|
|
|
|
|
|
int32_t strLen = ut->b;
|
2022-10-28 06:11:55 +00:00
|
|
|
UBool nulTerminated = false;
|
2020-08-11 09:10:23 +00:00
|
|
|
if (strLen < 0) {
|
|
|
|
strLen = 0x7fffffff;
|
2022-10-28 06:11:55 +00:00
|
|
|
nulTerminated = true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *buf = u8b_swap->buf;
|
2020-08-11 09:10:23 +00:00
|
|
|
uint8_t *mapToNative = u8b_swap->mapToNative;
|
|
|
|
uint8_t *mapToUChars = u8b_swap->mapToUChars;
|
|
|
|
int32_t destIx = 0;
|
|
|
|
int32_t srcIx = ix;
|
2022-10-28 06:11:55 +00:00
|
|
|
UBool seenNonAscii = false;
|
2020-08-11 09:10:23 +00:00
|
|
|
UChar32 c = 0;
|
|
|
|
|
|
|
|
// Fill the chunk buffer and mapping arrays.
|
|
|
|
while (destIx<UTF8_TEXT_CHUNK_SIZE) {
|
|
|
|
c = s8[srcIx];
|
|
|
|
if (c>0 && c<0x80) {
|
|
|
|
// Special case ASCII range for speed.
|
|
|
|
// zero is excluded to simplify bounds checking.
|
2023-05-23 00:05:01 +00:00
|
|
|
buf[destIx] = (char16_t)c;
|
2020-08-11 09:10:23 +00:00
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - ix);
|
|
|
|
mapToUChars[srcIx-ix] = (uint8_t)destIx;
|
|
|
|
srcIx++;
|
|
|
|
destIx++;
|
|
|
|
} else {
|
|
|
|
// General case, handle everything.
|
2022-10-28 06:11:55 +00:00
|
|
|
if (seenNonAscii == false) {
|
|
|
|
seenNonAscii = true;
|
2020-08-11 09:10:23 +00:00
|
|
|
u8b_swap->bufNILimit = destIx;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t cIx = srcIx;
|
|
|
|
int32_t dIx = destIx;
|
|
|
|
int32_t dIxSaved = destIx;
|
|
|
|
U8_NEXT_OR_FFFD(s8, srcIx, strLen, c);
|
|
|
|
if (c==0 && nulTerminated) {
|
|
|
|
srcIx--;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
U16_APPEND_UNSAFE(buf, destIx, c);
|
|
|
|
do {
|
|
|
|
mapToNative[dIx++] = (uint8_t)(cIx - ix);
|
|
|
|
} while (dIx < destIx);
|
|
|
|
|
|
|
|
do {
|
|
|
|
mapToUChars[cIx++ - ix] = (uint8_t)dIxSaved;
|
|
|
|
} while (cIx < srcIx);
|
|
|
|
}
|
|
|
|
if (srcIx>=strLen) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
// store Native <--> Chunk Map entries for the end of the buffer.
|
|
|
|
// There is no actual character here, but the index position is valid.
|
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - ix);
|
|
|
|
mapToUChars[srcIx - ix] = (uint8_t)destIx;
|
|
|
|
|
|
|
|
// fill in Buffer descriptor
|
|
|
|
u8b_swap->bufNativeStart = ix;
|
|
|
|
u8b_swap->bufNativeLimit = srcIx;
|
|
|
|
u8b_swap->bufStartIdx = 0;
|
|
|
|
u8b_swap->bufLimitIdx = destIx;
|
2022-10-28 06:11:55 +00:00
|
|
|
if (seenNonAscii == false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
u8b_swap->bufNILimit = destIx;
|
|
|
|
}
|
|
|
|
u8b_swap->toUCharsMapStart = u8b_swap->bufNativeStart;
|
|
|
|
|
|
|
|
// Set UText chunk to refer to this buffer.
|
|
|
|
ut->chunkContents = buf;
|
|
|
|
ut->chunkOffset = 0;
|
|
|
|
ut->chunkLength = u8b_swap->bufLimitIdx;
|
|
|
|
ut->chunkNativeStart = u8b_swap->bufNativeStart;
|
|
|
|
ut->chunkNativeLimit = u8b_swap->bufNativeLimit;
|
|
|
|
ut->nativeIndexingLimit = u8b_swap->bufNILimit;
|
|
|
|
|
|
|
|
// For zero terminated strings, keep track of the maximum point
|
|
|
|
// scanned so far.
|
|
|
|
if (nulTerminated && srcIx>ut->c) {
|
|
|
|
ut->c = srcIx;
|
|
|
|
if (c==0) {
|
|
|
|
// We scanned to the end.
|
|
|
|
// Remember the actual length.
|
|
|
|
ut->b = srcIx;
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
}
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
fillReverse:
|
|
|
|
{
|
|
|
|
// Move the incoming index to a code point boundary.
|
|
|
|
// Can only do this if the incoming index is somewhere in the interior of the string.
|
|
|
|
// If index is at the end, there is no character there to look at.
|
|
|
|
if (ix != ut->b) {
|
|
|
|
// Note: this function will only move the index back if it is on a trail byte
|
|
|
|
// and there is a preceding lead byte and the sequence from the lead
|
|
|
|
// through this trail could be part of a valid UTF-8 sequence
|
|
|
|
// Otherwise the index remains unchanged.
|
|
|
|
U8_SET_CP_START(s8, 0, ix);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Swap the UText buffers.
|
|
|
|
// We want to fill what was previously the alternate buffer,
|
|
|
|
// and make what was the current buffer be the new alternate.
|
|
|
|
UTF8Buf *u8b_swap = (UTF8Buf *)ut->q;
|
|
|
|
ut->q = ut->p;
|
|
|
|
ut->p = u8b_swap;
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *buf = u8b_swap->buf;
|
2020-08-11 09:10:23 +00:00
|
|
|
uint8_t *mapToNative = u8b_swap->mapToNative;
|
|
|
|
uint8_t *mapToUChars = u8b_swap->mapToUChars;
|
|
|
|
int32_t toUCharsMapStart = ix - sizeof(UTF8Buf::mapToUChars) + 1;
|
|
|
|
// Note that toUCharsMapStart can be negative. Happens when the remaining
|
|
|
|
// text from current position to the beginning is less than the buffer size.
|
|
|
|
// + 1 because mapToUChars must have a slot at the end for the bufNativeLimit entry.
|
|
|
|
int32_t destIx = UTF8_TEXT_CHUNK_SIZE+2; // Start in the overflow region
|
|
|
|
// at end of buffer to leave room
|
|
|
|
// for a surrogate pair at the
|
|
|
|
// buffer start.
|
|
|
|
int32_t srcIx = ix;
|
|
|
|
int32_t bufNILimit = destIx;
|
|
|
|
UChar32 c;
|
|
|
|
|
|
|
|
// Map to/from Native Indexes, fill in for the position at the end of
|
|
|
|
// the buffer.
|
|
|
|
//
|
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - toUCharsMapStart);
|
|
|
|
mapToUChars[srcIx - toUCharsMapStart] = (uint8_t)destIx;
|
|
|
|
|
|
|
|
// Fill the chunk buffer
|
|
|
|
// Work backwards, filling from the end of the buffer towards the front.
|
|
|
|
//
|
|
|
|
while (destIx>2 && (srcIx - toUCharsMapStart > 5) && (srcIx > 0)) {
|
|
|
|
srcIx--;
|
|
|
|
destIx--;
|
|
|
|
|
|
|
|
// Get last byte of the UTF-8 character
|
|
|
|
c = s8[srcIx];
|
|
|
|
if (c<0x80) {
|
|
|
|
// Special case ASCII range for speed.
|
2023-05-23 00:05:01 +00:00
|
|
|
buf[destIx] = (char16_t)c;
|
2020-08-11 09:10:23 +00:00
|
|
|
U_ASSERT(toUCharsMapStart <= srcIx);
|
|
|
|
mapToUChars[srcIx - toUCharsMapStart] = (uint8_t)destIx;
|
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - toUCharsMapStart);
|
|
|
|
} else {
|
|
|
|
// General case, handle everything non-ASCII.
|
|
|
|
|
|
|
|
int32_t sIx = srcIx; // ix of last byte of multi-byte u8 char
|
|
|
|
|
|
|
|
// Get the full character from the UTF8 string.
|
2021-10-28 06:15:28 +00:00
|
|
|
// use code derived from the macros in utf8.h
|
2020-08-11 09:10:23 +00:00
|
|
|
// Leaves srcIx pointing at the first byte of the UTF-8 char.
|
|
|
|
//
|
|
|
|
c=utf8_prevCharSafeBody(s8, 0, &srcIx, c, -3);
|
|
|
|
// leaves srcIx at first byte of the multi-byte char.
|
|
|
|
|
|
|
|
// Store the character in UTF-16 buffer.
|
|
|
|
if (c<0x10000) {
|
2023-05-23 00:05:01 +00:00
|
|
|
buf[destIx] = (char16_t)c;
|
2020-08-11 09:10:23 +00:00
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - toUCharsMapStart);
|
|
|
|
} else {
|
|
|
|
buf[destIx] = U16_TRAIL(c);
|
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - toUCharsMapStart);
|
|
|
|
buf[--destIx] = U16_LEAD(c);
|
|
|
|
mapToNative[destIx] = (uint8_t)(srcIx - toUCharsMapStart);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Fill in the map from native indexes to UChars buf index.
|
|
|
|
do {
|
|
|
|
mapToUChars[sIx-- - toUCharsMapStart] = (uint8_t)destIx;
|
|
|
|
} while (sIx >= srcIx);
|
|
|
|
U_ASSERT(toUCharsMapStart <= (srcIx+1));
|
|
|
|
|
|
|
|
// Set native indexing limit to be the current position.
|
|
|
|
// We are processing a non-ascii, non-native-indexing char now;
|
|
|
|
// the limit will be here if the rest of the chars to be
|
|
|
|
// added to this buffer are ascii.
|
|
|
|
bufNILimit = destIx;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
u8b_swap->bufNativeStart = srcIx;
|
|
|
|
u8b_swap->bufNativeLimit = ix;
|
|
|
|
u8b_swap->bufStartIdx = destIx;
|
|
|
|
u8b_swap->bufLimitIdx = UTF8_TEXT_CHUNK_SIZE+2;
|
|
|
|
u8b_swap->bufNILimit = bufNILimit - u8b_swap->bufStartIdx;
|
|
|
|
u8b_swap->toUCharsMapStart = toUCharsMapStart;
|
|
|
|
|
|
|
|
ut->chunkContents = &buf[u8b_swap->bufStartIdx];
|
|
|
|
ut->chunkLength = u8b_swap->bufLimitIdx - u8b_swap->bufStartIdx;
|
|
|
|
ut->chunkOffset = ut->chunkLength;
|
|
|
|
ut->chunkNativeStart = u8b_swap->bufNativeStart;
|
|
|
|
ut->chunkNativeLimit = u8b_swap->bufNativeLimit;
|
|
|
|
ut->nativeIndexingLimit = u8b_swap->bufNILimit;
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//
|
|
|
|
// This is a slightly modified copy of u_strFromUTF8,
|
|
|
|
// Inserts a Replacement Char rather than failing on invalid UTF-8
|
|
|
|
// Removes unnecessary features.
|
|
|
|
//
|
2023-05-23 00:05:01 +00:00
|
|
|
static char16_t*
|
|
|
|
utext_strFromUTF8(char16_t *dest,
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t destCapacity,
|
|
|
|
int32_t *pDestLength,
|
|
|
|
const char* src,
|
|
|
|
int32_t srcLength, // required. NUL terminated not supported.
|
|
|
|
UErrorCode *pErrorCode
|
|
|
|
)
|
|
|
|
{
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *pDest = dest;
|
|
|
|
char16_t *pDestLimit = (dest!=nullptr)?(dest+destCapacity):nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
UChar32 ch=0;
|
|
|
|
int32_t index = 0;
|
|
|
|
int32_t reqLength = 0;
|
|
|
|
uint8_t* pSrc = (uint8_t*) src;
|
|
|
|
|
|
|
|
|
|
|
|
while((index < srcLength)&&(pDest<pDestLimit)){
|
|
|
|
ch = pSrc[index++];
|
|
|
|
if(ch <=0x7f){
|
2023-05-23 00:05:01 +00:00
|
|
|
*pDest++=(char16_t)ch;
|
2020-08-11 09:10:23 +00:00
|
|
|
}else{
|
|
|
|
ch=utf8_nextCharSafeBody(pSrc, &index, srcLength, ch, -3);
|
|
|
|
if(U_IS_BMP(ch)){
|
2023-05-23 00:05:01 +00:00
|
|
|
*(pDest++)=(char16_t)ch;
|
2020-08-11 09:10:23 +00:00
|
|
|
}else{
|
|
|
|
*(pDest++)=U16_LEAD(ch);
|
|
|
|
if(pDest<pDestLimit){
|
|
|
|
*(pDest++)=U16_TRAIL(ch);
|
|
|
|
}else{
|
|
|
|
reqLength++;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/* donot fill the dest buffer just count the UChars needed */
|
|
|
|
while(index < srcLength){
|
|
|
|
ch = pSrc[index++];
|
|
|
|
if(ch <= 0x7f){
|
|
|
|
reqLength++;
|
|
|
|
}else{
|
|
|
|
ch=utf8_nextCharSafeBody(pSrc, &index, srcLength, ch, -3);
|
|
|
|
reqLength+=U16_LENGTH(ch);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
reqLength+=(int32_t)(pDest - dest);
|
|
|
|
|
|
|
|
if(pDestLength){
|
|
|
|
*pDestLength = reqLength;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Terminate the buffer */
|
|
|
|
u_terminateUChars(dest,destCapacity,reqLength,pErrorCode);
|
|
|
|
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
utf8TextExtract(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *pErrorCode) {
|
|
|
|
if(U_FAILURE(*pErrorCode)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(destCapacity<0 || (dest==nullptr && destCapacity>0)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*pErrorCode=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
int32_t length = ut->b;
|
|
|
|
int32_t start32 = pinIndex(start, length);
|
|
|
|
int32_t limit32 = pinIndex(limit, length);
|
|
|
|
|
|
|
|
if(start32>limit32) {
|
|
|
|
*pErrorCode=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
// adjust the incoming indexes to land on code point boundaries if needed.
|
|
|
|
// adjust by no more than three, because that is the largest number of trail bytes
|
|
|
|
// in a well formed UTF8 character.
|
|
|
|
const uint8_t *buf = (const uint8_t *)ut->context;
|
|
|
|
int i;
|
|
|
|
if (start32 < ut->chunkNativeLimit) {
|
|
|
|
for (i=0; i<3; i++) {
|
|
|
|
if (U8_IS_SINGLE(buf[start32]) || U8_IS_LEAD(buf[start32]) || start32==0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
start32--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (limit32 < ut->chunkNativeLimit) {
|
|
|
|
for (i=0; i<3; i++) {
|
|
|
|
if (U8_IS_SINGLE(buf[limit32]) || U8_IS_LEAD(buf[limit32]) || limit32==0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
limit32--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Do the actual extract.
|
|
|
|
int32_t destLength=0;
|
|
|
|
utext_strFromUTF8(dest, destCapacity, &destLength,
|
|
|
|
(const char *)ut->context+start32, limit32-start32,
|
|
|
|
pErrorCode);
|
2022-10-28 06:11:55 +00:00
|
|
|
utf8TextAccess(ut, limit32, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
return destLength;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
|
|
|
// utf8TextMapOffsetToNative
|
|
|
|
//
|
|
|
|
// Map a chunk (UTF-16) offset to a native index.
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
utf8TextMapOffsetToNative(const UText *ut) {
|
|
|
|
//
|
|
|
|
UTF8Buf *u8b = (UTF8Buf *)ut->p;
|
|
|
|
U_ASSERT(ut->chunkOffset>ut->nativeIndexingLimit && ut->chunkOffset<=ut->chunkLength);
|
|
|
|
int32_t nativeOffset = u8b->mapToNative[ut->chunkOffset + u8b->bufStartIdx] + u8b->toUCharsMapStart;
|
|
|
|
U_ASSERT(nativeOffset >= ut->chunkNativeStart && nativeOffset <= ut->chunkNativeLimit);
|
|
|
|
return nativeOffset;
|
|
|
|
}
|
|
|
|
|
|
|
|
//
|
2021-04-22 12:08:59 +00:00
|
|
|
// Map a native index to the corresponding chunk offset
|
2020-08-11 09:10:23 +00:00
|
|
|
//
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
utf8TextMapIndexToUTF16(const UText *ut, int64_t index64) {
|
|
|
|
U_ASSERT(index64 <= 0x7fffffff);
|
|
|
|
int32_t index = (int32_t)index64;
|
|
|
|
UTF8Buf *u8b = (UTF8Buf *)ut->p;
|
|
|
|
U_ASSERT(index>=ut->chunkNativeStart+ut->nativeIndexingLimit);
|
|
|
|
U_ASSERT(index<=ut->chunkNativeLimit);
|
|
|
|
int32_t mapIndex = index - u8b->toUCharsMapStart;
|
|
|
|
U_ASSERT(mapIndex < (int32_t)sizeof(UTF8Buf::mapToUChars));
|
|
|
|
int32_t offset = u8b->mapToUChars[mapIndex] - u8b->bufStartIdx;
|
|
|
|
U_ASSERT(offset>=0 && offset<=ut->chunkLength);
|
|
|
|
return offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
utf8TextClone(UText *dest, const UText *src, UBool deep, UErrorCode *status)
|
|
|
|
{
|
|
|
|
// First do a generic shallow clone. Does everything needed for the UText struct itself.
|
|
|
|
dest = shallowTextClone(dest, src, status);
|
|
|
|
|
|
|
|
// For deep clones, make a copy of the string.
|
|
|
|
// The copied storage is owned by the newly created clone.
|
|
|
|
//
|
2021-10-28 06:15:28 +00:00
|
|
|
// TODO: There is an issue with using utext_nativeLength().
|
2020-08-11 09:10:23 +00:00
|
|
|
// That function is non-const in cases where the input was NUL terminated
|
|
|
|
// and the length has not yet been determined.
|
|
|
|
// This function (clone()) is const.
|
|
|
|
// There potentially a thread safety issue lurking here.
|
|
|
|
//
|
|
|
|
if (deep && U_SUCCESS(*status)) {
|
|
|
|
int32_t len = (int32_t)utext_nativeLength((UText *)src);
|
|
|
|
char *copyStr = (char *)uprv_malloc(len+1);
|
2023-05-23 00:05:01 +00:00
|
|
|
if (copyStr == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_MEMORY_ALLOCATION_ERROR;
|
|
|
|
} else {
|
|
|
|
uprv_memcpy(copyStr, src->context, len+1);
|
|
|
|
dest->context = copyStr;
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
utf8TextClose(UText *ut) {
|
|
|
|
// Most of the work of close is done by the generic UText framework close.
|
|
|
|
// All that needs to be done here is to delete the UTF8 string if the UText
|
|
|
|
// owns it. This occurs if the UText was created by cloning.
|
|
|
|
if (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT)) {
|
|
|
|
char *s = (char *)ut->context;
|
|
|
|
uprv_free(s);
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->context = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
U_CDECL_END
|
|
|
|
|
|
|
|
|
|
|
|
static const struct UTextFuncs utf8Funcs =
|
|
|
|
{
|
|
|
|
sizeof(UTextFuncs),
|
|
|
|
0, 0, 0, // Reserved alignment padding
|
|
|
|
utf8TextClone,
|
|
|
|
utf8TextLength,
|
|
|
|
utf8TextAccess,
|
|
|
|
utf8TextExtract,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, /* replace*/
|
|
|
|
nullptr, /* copy */
|
2020-08-11 09:10:23 +00:00
|
|
|
utf8TextMapOffsetToNative,
|
|
|
|
utf8TextMapIndexToUTF16,
|
|
|
|
utf8TextClose,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // spare 1
|
|
|
|
nullptr, // spare 2
|
|
|
|
nullptr // spare 3
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
static const char gEmptyString[] = {0};
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_openUTF8(UText *ut, const char *s, int64_t length, UErrorCode *status) {
|
|
|
|
if(U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(s==nullptr && length==0) {
|
2020-08-11 09:10:23 +00:00
|
|
|
s = gEmptyString;
|
|
|
|
}
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
if(s==nullptr || length<-1 || length>INT32_MAX) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status=U_ILLEGAL_ARGUMENT_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
ut = utext_setup(ut, sizeof(UTF8Buf) * 2, status);
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
ut->pFuncs = &utf8Funcs;
|
|
|
|
ut->context = s;
|
|
|
|
ut->b = (int32_t)length;
|
|
|
|
ut->c = (int32_t)length;
|
|
|
|
if (ut->c < 0) {
|
|
|
|
ut->c = 0;
|
|
|
|
ut->providerProperties |= I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
ut->p = ut->pExtra;
|
|
|
|
ut->q = (char *)ut->pExtra + sizeof(UTF8Buf);
|
|
|
|
return ut;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// UText implementation wrapper for Replaceable (read/write)
|
|
|
|
//
|
|
|
|
// Use of UText data members:
|
|
|
|
// context pointer to Replaceable.
|
|
|
|
// p pointer to Replaceable if it is owned by the UText.
|
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
// minimum chunk size for this implementation: 3
|
|
|
|
// to allow for possible trimming for code point boundaries
|
|
|
|
enum { REP_TEXT_CHUNK_SIZE=10 };
|
|
|
|
|
|
|
|
struct ReplExtra {
|
|
|
|
/*
|
|
|
|
* Chunk UChars.
|
|
|
|
* +1 to simplify filling with surrogate pair at the end.
|
|
|
|
*/
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t s[REP_TEXT_CHUNK_SIZE+1];
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
repTextClone(UText *dest, const UText *src, UBool deep, UErrorCode *status) {
|
|
|
|
// First do a generic shallow clone. Does everything needed for the UText struct itself.
|
|
|
|
dest = shallowTextClone(dest, src, status);
|
|
|
|
|
|
|
|
// For deep clones, make a copy of the Replaceable.
|
|
|
|
// The copied Replaceable storage is owned by the newly created UText clone.
|
2023-05-23 00:05:01 +00:00
|
|
|
// A non-nullptr pointer in UText.p is the signal to the close() function to delete
|
2020-08-11 09:10:23 +00:00
|
|
|
// it.
|
|
|
|
//
|
|
|
|
if (deep && U_SUCCESS(*status)) {
|
|
|
|
const Replaceable *replSrc = (const Replaceable *)src->context;
|
|
|
|
dest->context = replSrc->clone();
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT);
|
|
|
|
|
|
|
|
// with deep clone, the copy is writable, even when the source is not.
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_WRITABLE);
|
|
|
|
}
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
repTextClose(UText *ut) {
|
|
|
|
// Most of the work of close is done by the generic UText framework close.
|
|
|
|
// All that needs to be done here is delete the Replaceable if the UText
|
|
|
|
// owns it. This occurs if the UText was created by cloning.
|
|
|
|
if (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT)) {
|
|
|
|
Replaceable *rep = (Replaceable *)ut->context;
|
|
|
|
delete rep;
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->context = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
repTextLength(UText *ut) {
|
|
|
|
const Replaceable *replSrc = (const Replaceable *)ut->context;
|
|
|
|
int32_t len = replSrc->length();
|
|
|
|
return len;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static UBool U_CALLCONV
|
|
|
|
repTextAccess(UText *ut, int64_t index, UBool forward) {
|
|
|
|
const Replaceable *rep=(const Replaceable *)ut->context;
|
|
|
|
int32_t length=rep->length(); // Full length of the input text (bigger than a chunk)
|
|
|
|
|
|
|
|
// clip the requested index to the limits of the text.
|
|
|
|
int32_t index32 = pinIndex(index, length);
|
|
|
|
U_ASSERT(index<=INT32_MAX);
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Compute start/limit boundaries around index, for a segment of text
|
|
|
|
* to be extracted.
|
|
|
|
* To allow for the possibility that our user gave an index to the trailing
|
2023-05-23 00:05:01 +00:00
|
|
|
* half of a surrogate pair, we must request one extra preceding char16_t when
|
2020-08-11 09:10:23 +00:00
|
|
|
* going in the forward direction. This will ensure that the buffer has the
|
|
|
|
* entire code point at the specified index.
|
|
|
|
*/
|
|
|
|
if(forward) {
|
|
|
|
|
|
|
|
if (index32>=ut->chunkNativeStart && index32<ut->chunkNativeLimit) {
|
|
|
|
// Buffer already contains the requested position.
|
|
|
|
ut->chunkOffset = (int32_t)(index - ut->chunkNativeStart);
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
if (index32>=length && ut->chunkNativeLimit==length) {
|
|
|
|
// Request for end of string, and buffer already extends up to it.
|
|
|
|
// Can't get the data, but don't change the buffer.
|
|
|
|
ut->chunkOffset = length - (int32_t)ut->chunkNativeStart;
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
ut->chunkNativeLimit = index + REP_TEXT_CHUNK_SIZE - 1;
|
|
|
|
// Going forward, so we want to have the buffer with stuff at and beyond
|
|
|
|
// the requested index. The -1 gets us one code point before the
|
|
|
|
// requested index also, to handle the case of the index being on
|
|
|
|
// a trail surrogate of a surrogate pair.
|
|
|
|
if(ut->chunkNativeLimit > length) {
|
|
|
|
ut->chunkNativeLimit = length;
|
|
|
|
}
|
|
|
|
// unless buffer ran off end, start is index-1.
|
|
|
|
ut->chunkNativeStart = ut->chunkNativeLimit - REP_TEXT_CHUNK_SIZE;
|
|
|
|
if(ut->chunkNativeStart < 0) {
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// Reverse iteration. Fill buffer with data preceding the requested index.
|
|
|
|
if (index32>ut->chunkNativeStart && index32<=ut->chunkNativeLimit) {
|
|
|
|
// Requested position already in buffer.
|
|
|
|
ut->chunkOffset = index32 - (int32_t)ut->chunkNativeStart;
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
if (index32==0 && ut->chunkNativeStart==0) {
|
|
|
|
// Request for start, buffer already begins at start.
|
|
|
|
// No data, but keep the buffer as is.
|
|
|
|
ut->chunkOffset = 0;
|
2022-10-28 06:11:55 +00:00
|
|
|
return false;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Figure out the bounds of the chunk to extract for reverse iteration.
|
|
|
|
// Need to worry about chunk not splitting surrogate pairs, and while still
|
|
|
|
// containing the data we need.
|
2023-05-23 00:05:01 +00:00
|
|
|
// Fix by requesting a chunk that includes an extra char16_t at the end.
|
2020-08-11 09:10:23 +00:00
|
|
|
// If this turns out to be a lead surrogate, we can lop it off and still have
|
|
|
|
// the data we wanted.
|
|
|
|
ut->chunkNativeStart = index32 + 1 - REP_TEXT_CHUNK_SIZE;
|
|
|
|
if (ut->chunkNativeStart < 0) {
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ut->chunkNativeLimit = index32 + 1;
|
|
|
|
if (ut->chunkNativeLimit > length) {
|
|
|
|
ut->chunkNativeLimit = length;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Extract the new chunk of text from the Replaceable source.
|
|
|
|
ReplExtra *ex = (ReplExtra *)ut->pExtra;
|
|
|
|
// UnicodeString with its buffer a writable alias to the chunk buffer
|
|
|
|
UnicodeString buffer(ex->s, 0 /*buffer length*/, REP_TEXT_CHUNK_SIZE /*buffer capacity*/);
|
|
|
|
rep->extractBetween((int32_t)ut->chunkNativeStart, (int32_t)ut->chunkNativeLimit, buffer);
|
|
|
|
|
|
|
|
ut->chunkContents = ex->s;
|
|
|
|
ut->chunkLength = (int32_t)(ut->chunkNativeLimit - ut->chunkNativeStart);
|
|
|
|
ut->chunkOffset = (int32_t)(index32 - ut->chunkNativeStart);
|
|
|
|
|
|
|
|
// Surrogate pairs from the input text must not span chunk boundaries.
|
|
|
|
// If end of chunk could be the start of a surrogate, trim it off.
|
|
|
|
if (ut->chunkNativeLimit < length &&
|
|
|
|
U16_IS_LEAD(ex->s[ut->chunkLength-1])) {
|
|
|
|
ut->chunkLength--;
|
|
|
|
ut->chunkNativeLimit--;
|
|
|
|
if (ut->chunkOffset > ut->chunkLength) {
|
|
|
|
ut->chunkOffset = ut->chunkLength;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
// if the first char16_t in the chunk could be the trailing half of a surrogate pair,
|
2020-08-11 09:10:23 +00:00
|
|
|
// trim it off.
|
|
|
|
if(ut->chunkNativeStart>0 && U16_IS_TRAIL(ex->s[0])) {
|
|
|
|
++(ut->chunkContents);
|
|
|
|
++(ut->chunkNativeStart);
|
|
|
|
--(ut->chunkLength);
|
|
|
|
--(ut->chunkOffset);
|
|
|
|
}
|
|
|
|
|
|
|
|
// adjust the index/chunkOffset to a code point boundary
|
|
|
|
U16_SET_CP_START(ut->chunkContents, 0, ut->chunkOffset);
|
|
|
|
|
|
|
|
// Use fast indexing for get/setNativeIndex()
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
|
2022-10-28 06:11:55 +00:00
|
|
|
return true;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
repTextExtract(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *status) {
|
|
|
|
const Replaceable *rep=(const Replaceable *)ut->context;
|
|
|
|
int32_t length=rep->length();
|
|
|
|
|
|
|
|
if(U_FAILURE(*status)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(destCapacity<0 || (dest==nullptr && destCapacity>0)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
}
|
|
|
|
if(start>limit) {
|
|
|
|
*status=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t start32 = pinIndex(start, length);
|
|
|
|
int32_t limit32 = pinIndex(limit, length);
|
|
|
|
|
|
|
|
// adjust start, limit if they point to trail half of surrogates
|
|
|
|
if (start32<length && U16_IS_TRAIL(rep->charAt(start32)) &&
|
|
|
|
U_IS_SUPPLEMENTARY(rep->char32At(start32))){
|
|
|
|
start32--;
|
|
|
|
}
|
|
|
|
if (limit32<length && U16_IS_TRAIL(rep->charAt(limit32)) &&
|
|
|
|
U_IS_SUPPLEMENTARY(rep->char32At(limit32))){
|
|
|
|
limit32--;
|
|
|
|
}
|
|
|
|
|
|
|
|
length=limit32-start32;
|
|
|
|
if(length>destCapacity) {
|
|
|
|
limit32 = start32 + destCapacity;
|
|
|
|
}
|
|
|
|
UnicodeString buffer(dest, 0, destCapacity); // writable alias
|
|
|
|
rep->extractBetween(start32, limit32, buffer);
|
2022-10-28 06:11:55 +00:00
|
|
|
repTextAccess(ut, limit32, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
return u_terminateUChars(dest, destCapacity, length, status);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
repTextReplace(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *src, int32_t length,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *status) {
|
|
|
|
Replaceable *rep=(Replaceable *)ut->context;
|
|
|
|
int32_t oldLength;
|
|
|
|
|
|
|
|
if(U_FAILURE(*status)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(src==nullptr && length!=0) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
oldLength=rep->length(); // will subtract from new length
|
|
|
|
if(start>limit ) {
|
|
|
|
*status=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t start32 = pinIndex(start, oldLength);
|
|
|
|
int32_t limit32 = pinIndex(limit, oldLength);
|
|
|
|
|
|
|
|
// Snap start & limit to code point boundaries.
|
|
|
|
if (start32<oldLength && U16_IS_TRAIL(rep->charAt(start32)) &&
|
|
|
|
start32>0 && U16_IS_LEAD(rep->charAt(start32-1)))
|
|
|
|
{
|
|
|
|
start32--;
|
|
|
|
}
|
|
|
|
if (limit32<oldLength && U16_IS_LEAD(rep->charAt(limit32-1)) &&
|
|
|
|
U16_IS_TRAIL(rep->charAt(limit32)))
|
|
|
|
{
|
|
|
|
limit32++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Do the actual replace operation using methods of the Replaceable class
|
|
|
|
UnicodeString replStr((UBool)(length<0), src, length); // read-only alias
|
|
|
|
rep->handleReplaceBetween(start32, limit32, replStr);
|
|
|
|
int32_t newLength = rep->length();
|
|
|
|
int32_t lengthDelta = newLength - oldLength;
|
|
|
|
|
|
|
|
// Is the UText chunk buffer OK?
|
|
|
|
if (ut->chunkNativeLimit > start32) {
|
|
|
|
// this replace operation may have impacted the current chunk.
|
|
|
|
// invalidate it, which will force a reload on the next access.
|
|
|
|
invalidateChunk(ut);
|
|
|
|
}
|
|
|
|
|
|
|
|
// set the iteration position to the end of the newly inserted replacement text.
|
|
|
|
int32_t newIndexPos = limit32 + lengthDelta;
|
2022-10-28 06:11:55 +00:00
|
|
|
repTextAccess(ut, newIndexPos, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
return lengthDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
repTextCopy(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
|
|
|
int64_t destIndex,
|
|
|
|
UBool move,
|
|
|
|
UErrorCode *status)
|
|
|
|
{
|
|
|
|
Replaceable *rep=(Replaceable *)ut->context;
|
|
|
|
int32_t length=rep->length();
|
|
|
|
|
|
|
|
if(U_FAILURE(*status)) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (start>limit || (start<destIndex && destIndex<limit))
|
|
|
|
{
|
|
|
|
*status=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t start32 = pinIndex(start, length);
|
|
|
|
int32_t limit32 = pinIndex(limit, length);
|
|
|
|
int32_t destIndex32 = pinIndex(destIndex, length);
|
|
|
|
|
|
|
|
// TODO: snap input parameters to code point boundaries.
|
|
|
|
|
|
|
|
if(move) {
|
|
|
|
// move: copy to destIndex, then replace original with nothing
|
|
|
|
int32_t segLength=limit32-start32;
|
|
|
|
rep->copy(start32, limit32, destIndex32);
|
|
|
|
if(destIndex32<start32) {
|
|
|
|
start32+=segLength;
|
|
|
|
limit32+=segLength;
|
|
|
|
}
|
|
|
|
rep->handleReplaceBetween(start32, limit32, UnicodeString());
|
|
|
|
} else {
|
|
|
|
// copy
|
|
|
|
rep->copy(start32, limit32, destIndex32);
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the change to the text touched the region in the chunk buffer,
|
|
|
|
// invalidate the buffer.
|
|
|
|
int32_t firstAffectedIndex = destIndex32;
|
|
|
|
if (move && start32<firstAffectedIndex) {
|
|
|
|
firstAffectedIndex = start32;
|
|
|
|
}
|
|
|
|
if (firstAffectedIndex < ut->chunkNativeLimit) {
|
|
|
|
// changes may have affected range covered by the chunk
|
|
|
|
invalidateChunk(ut);
|
|
|
|
}
|
|
|
|
|
|
|
|
// Put iteration position at the newly inserted (moved) block,
|
|
|
|
int32_t nativeIterIndex = destIndex32 + limit32 - start32;
|
|
|
|
if (move && destIndex32>start32) {
|
|
|
|
// moved a block of text towards the end of the string.
|
|
|
|
nativeIterIndex = destIndex32;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Set position, reload chunk if needed.
|
2022-10-28 06:11:55 +00:00
|
|
|
repTextAccess(ut, nativeIterIndex, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static const struct UTextFuncs repFuncs =
|
|
|
|
{
|
|
|
|
sizeof(UTextFuncs),
|
|
|
|
0, 0, 0, // Reserved alignment padding
|
|
|
|
repTextClone,
|
|
|
|
repTextLength,
|
|
|
|
repTextAccess,
|
|
|
|
repTextExtract,
|
|
|
|
repTextReplace,
|
|
|
|
repTextCopy,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // MapOffsetToNative,
|
|
|
|
nullptr, // MapIndexToUTF16,
|
2020-08-11 09:10:23 +00:00
|
|
|
repTextClose,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // spare 1
|
|
|
|
nullptr, // spare 2
|
|
|
|
nullptr // spare 3
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_openReplaceable(UText *ut, Replaceable *rep, UErrorCode *status)
|
|
|
|
{
|
|
|
|
if(U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(rep==nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status=U_ILLEGAL_ARGUMENT_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
ut = utext_setup(ut, sizeof(ReplExtra), status);
|
|
|
|
if(U_FAILURE(*status)) {
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
ut->providerProperties = I32_FLAG(UTEXT_PROVIDER_WRITABLE);
|
|
|
|
if(rep->hasMetaData()) {
|
|
|
|
ut->providerProperties |=I32_FLAG(UTEXT_PROVIDER_HAS_META_DATA);
|
|
|
|
}
|
|
|
|
|
|
|
|
ut->pFuncs = &repFuncs;
|
|
|
|
ut->context = rep;
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
U_CDECL_END
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// UText implementation for UnicodeString (read/write) and
|
|
|
|
// for const UnicodeString (read only)
|
|
|
|
// (same implementation, only the flags are different)
|
|
|
|
//
|
|
|
|
// Use of UText data members:
|
|
|
|
// context pointer to UnicodeString
|
|
|
|
// p pointer to UnicodeString IF this UText owns the string
|
2023-05-23 00:05:01 +00:00
|
|
|
// and it must be deleted on close(). nullptr otherwise.
|
2020-08-11 09:10:23 +00:00
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
|
|
|
|
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
unistrTextClone(UText *dest, const UText *src, UBool deep, UErrorCode *status) {
|
|
|
|
// First do a generic shallow clone. Does everything needed for the UText struct itself.
|
|
|
|
dest = shallowTextClone(dest, src, status);
|
|
|
|
|
|
|
|
// For deep clones, make a copy of the UnicodeSring.
|
|
|
|
// The copied UnicodeString storage is owned by the newly created UText clone.
|
2023-05-23 00:05:01 +00:00
|
|
|
// A non-nullptr pointer in UText.p is the signal to the close() function to delete
|
2020-08-11 09:10:23 +00:00
|
|
|
// the UText.
|
|
|
|
//
|
|
|
|
if (deep && U_SUCCESS(*status)) {
|
|
|
|
const UnicodeString *srcString = (const UnicodeString *)src->context;
|
|
|
|
dest->context = new UnicodeString(*srcString);
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT);
|
|
|
|
|
|
|
|
// with deep clone, the copy is writable, even when the source is not.
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_WRITABLE);
|
|
|
|
}
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
unistrTextClose(UText *ut) {
|
|
|
|
// Most of the work of close is done by the generic UText framework close.
|
|
|
|
// All that needs to be done here is delete the UnicodeString if the UText
|
|
|
|
// owns it. This occurs if the UText was created by cloning.
|
|
|
|
if (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT)) {
|
|
|
|
UnicodeString *str = (UnicodeString *)ut->context;
|
|
|
|
delete str;
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->context = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
unistrTextLength(UText *t) {
|
|
|
|
return ((const UnicodeString *)t->context)->length();
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static UBool U_CALLCONV
|
|
|
|
unistrTextAccess(UText *ut, int64_t index, UBool forward) {
|
|
|
|
int32_t length = ut->chunkLength;
|
|
|
|
ut->chunkOffset = pinIndex(index, length);
|
|
|
|
|
|
|
|
// Check whether request is at the start or end
|
|
|
|
UBool retVal = (forward && index<length) || (!forward && index>0);
|
|
|
|
return retVal;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
unistrTextExtract(UText *t,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *pErrorCode) {
|
|
|
|
const UnicodeString *us=(const UnicodeString *)t->context;
|
|
|
|
int32_t length=us->length();
|
|
|
|
|
|
|
|
if(U_FAILURE(*pErrorCode)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(destCapacity<0 || (dest==nullptr && destCapacity>0)) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*pErrorCode=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
}
|
|
|
|
if(start<0 || start>limit) {
|
|
|
|
*pErrorCode=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t start32 = start<length ? us->getChar32Start((int32_t)start) : length;
|
|
|
|
int32_t limit32 = limit<length ? us->getChar32Start((int32_t)limit) : length;
|
|
|
|
|
|
|
|
length=limit32-start32;
|
2023-05-23 00:05:01 +00:00
|
|
|
if (destCapacity>0 && dest!=nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t trimmedLength = length;
|
|
|
|
if(trimmedLength>destCapacity) {
|
|
|
|
trimmedLength=destCapacity;
|
|
|
|
}
|
|
|
|
us->extract(start32, trimmedLength, dest);
|
|
|
|
t->chunkOffset = start32+trimmedLength;
|
|
|
|
} else {
|
|
|
|
t->chunkOffset = start32;
|
|
|
|
}
|
|
|
|
u_terminateUChars(dest, destCapacity, length, pErrorCode);
|
|
|
|
return length;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
unistrTextReplace(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *src, int32_t length,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *pErrorCode) {
|
|
|
|
UnicodeString *us=(UnicodeString *)ut->context;
|
|
|
|
int32_t oldLength;
|
|
|
|
|
|
|
|
if(U_FAILURE(*pErrorCode)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(src==nullptr && length!=0) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*pErrorCode=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
}
|
|
|
|
if(start>limit) {
|
|
|
|
*pErrorCode=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
oldLength=us->length();
|
|
|
|
int32_t start32 = pinIndex(start, oldLength);
|
|
|
|
int32_t limit32 = pinIndex(limit, oldLength);
|
|
|
|
if (start32 < oldLength) {
|
|
|
|
start32 = us->getChar32Start(start32);
|
|
|
|
}
|
|
|
|
if (limit32 < oldLength) {
|
|
|
|
limit32 = us->getChar32Start(limit32);
|
|
|
|
}
|
|
|
|
|
|
|
|
// replace
|
|
|
|
us->replace(start32, limit32-start32, src, length);
|
|
|
|
int32_t newLength = us->length();
|
|
|
|
|
|
|
|
// Update the chunk description.
|
|
|
|
ut->chunkContents = us->getBuffer();
|
|
|
|
ut->chunkLength = newLength;
|
|
|
|
ut->chunkNativeLimit = newLength;
|
|
|
|
ut->nativeIndexingLimit = newLength;
|
|
|
|
|
|
|
|
// Set iteration position to the point just following the newly inserted text.
|
|
|
|
int32_t lengthDelta = newLength - oldLength;
|
|
|
|
ut->chunkOffset = limit32 + lengthDelta;
|
|
|
|
|
|
|
|
return lengthDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
unistrTextCopy(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
|
|
|
int64_t destIndex,
|
|
|
|
UBool move,
|
|
|
|
UErrorCode *pErrorCode) {
|
|
|
|
UnicodeString *us=(UnicodeString *)ut->context;
|
|
|
|
int32_t length=us->length();
|
|
|
|
|
|
|
|
if(U_FAILURE(*pErrorCode)) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
int32_t start32 = pinIndex(start, length);
|
|
|
|
int32_t limit32 = pinIndex(limit, length);
|
|
|
|
int32_t destIndex32 = pinIndex(destIndex, length);
|
|
|
|
|
|
|
|
if( start32>limit32 || (start32<destIndex32 && destIndex32<limit32)) {
|
|
|
|
*pErrorCode=U_INDEX_OUTOFBOUNDS_ERROR;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if(move) {
|
|
|
|
// move: copy to destIndex, then remove original
|
|
|
|
int32_t segLength=limit32-start32;
|
|
|
|
us->copy(start32, limit32, destIndex32);
|
|
|
|
if(destIndex32<start32) {
|
|
|
|
start32+=segLength;
|
|
|
|
}
|
|
|
|
us->remove(start32, segLength);
|
|
|
|
} else {
|
|
|
|
// copy
|
|
|
|
us->copy(start32, limit32, destIndex32);
|
|
|
|
}
|
|
|
|
|
|
|
|
// update chunk description, set iteration position.
|
|
|
|
ut->chunkContents = us->getBuffer();
|
2022-10-28 06:11:55 +00:00
|
|
|
if (move==false) {
|
2020-08-11 09:10:23 +00:00
|
|
|
// copy operation, string length grows
|
|
|
|
ut->chunkLength += limit32-start32;
|
|
|
|
ut->chunkNativeLimit = ut->chunkLength;
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Iteration position to end of the newly inserted text.
|
|
|
|
ut->chunkOffset = destIndex32+limit32-start32;
|
|
|
|
if (move && destIndex32>start32) {
|
|
|
|
ut->chunkOffset = destIndex32;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct UTextFuncs unistrFuncs =
|
|
|
|
{
|
|
|
|
sizeof(UTextFuncs),
|
|
|
|
0, 0, 0, // Reserved alignment padding
|
|
|
|
unistrTextClone,
|
|
|
|
unistrTextLength,
|
|
|
|
unistrTextAccess,
|
|
|
|
unistrTextExtract,
|
|
|
|
unistrTextReplace,
|
|
|
|
unistrTextCopy,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // MapOffsetToNative,
|
|
|
|
nullptr, // MapIndexToUTF16,
|
2020-08-11 09:10:23 +00:00
|
|
|
unistrTextClose,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // spare 1
|
|
|
|
nullptr, // spare 2
|
|
|
|
nullptr // spare 3
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CDECL_END
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_openUnicodeString(UText *ut, UnicodeString *s, UErrorCode *status) {
|
|
|
|
ut = utext_openConstUnicodeString(ut, s, status);
|
|
|
|
if (U_SUCCESS(*status)) {
|
|
|
|
ut->providerProperties |= I32_FLAG(UTEXT_PROVIDER_WRITABLE);
|
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_openConstUnicodeString(UText *ut, const UnicodeString *s, UErrorCode *status) {
|
|
|
|
if (U_SUCCESS(*status) && s->isBogus()) {
|
|
|
|
// The UnicodeString is bogus, but we still need to detach the UText
|
|
|
|
// from whatever it was hooked to before, if anything.
|
2023-05-23 00:05:01 +00:00
|
|
|
utext_openUChars(ut, nullptr, 0, status);
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
ut = utext_setup(ut, 0, status);
|
|
|
|
// note: use the standard (writable) function table for UnicodeString.
|
|
|
|
// The flag settings disable writing, so having the functions in
|
|
|
|
// the table is harmless.
|
|
|
|
if (U_SUCCESS(*status)) {
|
|
|
|
ut->pFuncs = &unistrFuncs;
|
|
|
|
ut->context = s;
|
|
|
|
ut->providerProperties = I32_FLAG(UTEXT_PROVIDER_STABLE_CHUNKS);
|
|
|
|
ut->chunkContents = s->getBuffer();
|
|
|
|
ut->chunkLength = s->length();
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
ut->chunkNativeLimit = ut->chunkLength;
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
2023-05-23 00:05:01 +00:00
|
|
|
// UText implementation for const char16_t * strings
|
2020-08-11 09:10:23 +00:00
|
|
|
//
|
|
|
|
// Use of UText data members:
|
|
|
|
// context pointer to UnicodeString
|
|
|
|
// a length. -1 if not yet known.
|
|
|
|
//
|
|
|
|
// TODO: support 64 bit lengths.
|
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
|
|
|
|
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
ucstrTextClone(UText *dest, const UText * src, UBool deep, UErrorCode * status) {
|
|
|
|
// First do a generic shallow clone.
|
|
|
|
dest = shallowTextClone(dest, src, status);
|
|
|
|
|
|
|
|
// For deep clones, make a copy of the string.
|
|
|
|
// The copied storage is owned by the newly created clone.
|
2023-05-23 00:05:01 +00:00
|
|
|
// A non-nullptr pointer in UText.p is the signal to the close() function to delete
|
2020-08-11 09:10:23 +00:00
|
|
|
// it.
|
|
|
|
//
|
|
|
|
if (deep && U_SUCCESS(*status)) {
|
|
|
|
U_ASSERT(utext_nativeLength(dest) < INT32_MAX);
|
|
|
|
int32_t len = (int32_t)utext_nativeLength(dest);
|
|
|
|
|
|
|
|
// The cloned string IS going to be NUL terminated, whether or not the original was.
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *srcStr = (const char16_t *)src->context;
|
|
|
|
char16_t *copyStr = (char16_t *)uprv_malloc((len+1) * sizeof(char16_t));
|
|
|
|
if (copyStr == nullptr) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_MEMORY_ALLOCATION_ERROR;
|
|
|
|
} else {
|
|
|
|
int64_t i;
|
|
|
|
for (i=0; i<len; i++) {
|
|
|
|
copyStr[i] = srcStr[i];
|
|
|
|
}
|
|
|
|
copyStr[len] = 0;
|
|
|
|
dest->context = copyStr;
|
|
|
|
dest->providerProperties |= I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void U_CALLCONV
|
|
|
|
ucstrTextClose(UText *ut) {
|
|
|
|
// Most of the work of close is done by the generic UText framework close.
|
|
|
|
// All that needs to be done here is delete the string if the UText
|
|
|
|
// owns it. This occurs if the UText was created by cloning.
|
|
|
|
if (ut->providerProperties & I32_FLAG(UTEXT_PROVIDER_OWNS_TEXT)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *s = (char16_t *)ut->context;
|
2020-08-11 09:10:23 +00:00
|
|
|
uprv_free(s);
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->context = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
ucstrTextLength(UText *ut) {
|
|
|
|
if (ut->a < 0) {
|
2021-10-28 06:15:28 +00:00
|
|
|
// null terminated, we don't yet know the length. Scan for it.
|
2020-08-11 09:10:23 +00:00
|
|
|
// Access is not convenient for doing this
|
2021-10-28 06:15:28 +00:00
|
|
|
// because the current iteration position can't be changed.
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *str = (const char16_t *)ut->context;
|
2020-08-11 09:10:23 +00:00
|
|
|
for (;;) {
|
|
|
|
if (str[ut->chunkNativeLimit] == 0) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
ut->chunkNativeLimit++;
|
|
|
|
}
|
|
|
|
ut->a = ut->chunkNativeLimit;
|
|
|
|
ut->chunkLength = (int32_t)ut->chunkNativeLimit;
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
return ut->a;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static UBool U_CALLCONV
|
|
|
|
ucstrTextAccess(UText *ut, int64_t index, UBool forward) {
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *str = (const char16_t *)ut->context;
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
// pin the requested index to the bounds of the string,
|
|
|
|
// and set current iteration position.
|
|
|
|
if (index<0) {
|
|
|
|
index = 0;
|
|
|
|
} else if (index < ut->chunkNativeLimit) {
|
|
|
|
// The request data is within the chunk as it is known so far.
|
|
|
|
// Put index on a code point boundary.
|
|
|
|
U16_SET_CP_START(str, 0, index);
|
|
|
|
} else if (ut->a >= 0) {
|
|
|
|
// We know the length of this string, and the user is requesting something
|
|
|
|
// at or beyond the length. Pin the requested index to the length.
|
|
|
|
index = ut->a;
|
|
|
|
} else {
|
|
|
|
// Null terminated string, length not yet known, and the requested index
|
|
|
|
// is beyond where we have scanned so far.
|
|
|
|
// Scan to 32 UChars beyond the requested index. The strategy here is
|
|
|
|
// to avoid fully scanning a long string when the caller only wants to
|
|
|
|
// see a few characters at its beginning.
|
|
|
|
int32_t scanLimit = (int32_t)index + 32;
|
|
|
|
if ((index + 32)>INT32_MAX || (index + 32)<0 ) { // note: int64 expression
|
|
|
|
scanLimit = INT32_MAX;
|
|
|
|
}
|
|
|
|
|
|
|
|
int32_t chunkLimit = (int32_t)ut->chunkNativeLimit;
|
|
|
|
for (; chunkLimit<scanLimit; chunkLimit++) {
|
|
|
|
if (str[chunkLimit] == 0) {
|
|
|
|
// We found the end of the string. Remember it, pin the requested index to it,
|
|
|
|
// and bail out of here.
|
|
|
|
ut->a = chunkLimit;
|
|
|
|
ut->chunkLength = chunkLimit;
|
|
|
|
ut->nativeIndexingLimit = chunkLimit;
|
|
|
|
if (index >= chunkLimit) {
|
|
|
|
index = chunkLimit;
|
|
|
|
} else {
|
|
|
|
U16_SET_CP_START(str, 0, index);
|
|
|
|
}
|
|
|
|
|
|
|
|
ut->chunkNativeLimit = chunkLimit;
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
goto breakout;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// We scanned through the next batch of UChars without finding the end.
|
|
|
|
U16_SET_CP_START(str, 0, index);
|
|
|
|
if (chunkLimit == INT32_MAX) {
|
|
|
|
// Scanned to the limit of a 32 bit length.
|
|
|
|
// Forceably trim the overlength string back so length fits in int32
|
|
|
|
// TODO: add support for 64 bit strings.
|
|
|
|
ut->a = chunkLimit;
|
|
|
|
ut->chunkLength = chunkLimit;
|
|
|
|
ut->nativeIndexingLimit = chunkLimit;
|
|
|
|
if (index > chunkLimit) {
|
|
|
|
index = chunkLimit;
|
|
|
|
}
|
|
|
|
ut->chunkNativeLimit = chunkLimit;
|
|
|
|
ut->providerProperties &= ~I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
} else {
|
|
|
|
// The endpoint of a chunk must not be left in the middle of a surrogate pair.
|
|
|
|
// If the current end is on a lead surrogate, back the end up by one.
|
|
|
|
// It doesn't matter if the end char happens to be an unpaired surrogate,
|
|
|
|
// and it's simpler not to worry about it.
|
|
|
|
if (U16_IS_LEAD(str[chunkLimit-1])) {
|
|
|
|
--chunkLimit;
|
|
|
|
}
|
|
|
|
// Null-terminated chunk with end still unknown.
|
|
|
|
// Update the chunk length to reflect what has been scanned thus far.
|
|
|
|
// That the full length is still unknown is (still) flagged by
|
|
|
|
// ut->a being < 0.
|
|
|
|
ut->chunkNativeLimit = chunkLimit;
|
|
|
|
ut->nativeIndexingLimit = chunkLimit;
|
|
|
|
ut->chunkLength = chunkLimit;
|
|
|
|
}
|
|
|
|
|
|
|
|
}
|
|
|
|
breakout:
|
|
|
|
U_ASSERT(index<=INT32_MAX);
|
|
|
|
ut->chunkOffset = (int32_t)index;
|
|
|
|
|
|
|
|
// Check whether request is at the start or end
|
|
|
|
UBool retVal = (forward && index<ut->chunkNativeLimit) || (!forward && index>0);
|
|
|
|
return retVal;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
ucstrTextExtract(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *pErrorCode)
|
|
|
|
{
|
|
|
|
if(U_FAILURE(*pErrorCode)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(destCapacity<0 || (dest==nullptr && destCapacity>0) || start>limit) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*pErrorCode=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
//const char16_t *s=(const char16_t *)ut->context;
|
2020-08-11 09:10:23 +00:00
|
|
|
int32_t si, di;
|
|
|
|
|
|
|
|
int32_t start32;
|
|
|
|
int32_t limit32;
|
|
|
|
|
|
|
|
// Access the start. Does two things we need:
|
|
|
|
// Pins 'start' to the length of the string, if it came in out-of-bounds.
|
|
|
|
// Snaps 'start' to the beginning of a code point.
|
2022-10-28 06:11:55 +00:00
|
|
|
ucstrTextAccess(ut, start, true);
|
2023-05-23 00:05:01 +00:00
|
|
|
const char16_t *s=ut->chunkContents;
|
2020-08-11 09:10:23 +00:00
|
|
|
start32 = ut->chunkOffset;
|
|
|
|
|
|
|
|
int32_t strLength=(int32_t)ut->a;
|
|
|
|
if (strLength >= 0) {
|
|
|
|
limit32 = pinIndex(limit, strLength);
|
|
|
|
} else {
|
|
|
|
limit32 = pinIndex(limit, INT32_MAX);
|
|
|
|
}
|
|
|
|
di = 0;
|
|
|
|
for (si=start32; si<limit32; si++) {
|
|
|
|
if (strLength<0 && s[si]==0) {
|
|
|
|
// Just hit the end of a null-terminated string.
|
|
|
|
ut->a = si; // set string length for this UText
|
|
|
|
ut->chunkNativeLimit = si;
|
|
|
|
ut->chunkLength = si;
|
|
|
|
ut->nativeIndexingLimit = si;
|
|
|
|
strLength = si;
|
|
|
|
limit32 = si;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
U_ASSERT(di>=0); /* to ensure di never exceeds INT32_MAX, which must not happen logically */
|
|
|
|
if (di<destCapacity) {
|
|
|
|
// only store if there is space.
|
|
|
|
dest[di] = s[si];
|
|
|
|
} else {
|
|
|
|
if (strLength>=0) {
|
|
|
|
// We have filled the destination buffer, and the string length is known.
|
|
|
|
// Cut the loop short. There is no need to scan string termination.
|
|
|
|
di = limit32 - start32;
|
|
|
|
si = limit32;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
di++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If the limit index points to a lead surrogate of a pair,
|
|
|
|
// add the corresponding trail surrogate to the destination.
|
|
|
|
if (si>0 && U16_IS_LEAD(s[si-1]) &&
|
|
|
|
((si<strLength || strLength<0) && U16_IS_TRAIL(s[si])))
|
|
|
|
{
|
|
|
|
if (di<destCapacity) {
|
|
|
|
// store only if there is space in the output buffer.
|
|
|
|
dest[di++] = s[si];
|
|
|
|
}
|
|
|
|
si++;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Put iteration position at the point just following the extracted text
|
|
|
|
if (si <= ut->chunkNativeLimit) {
|
|
|
|
ut->chunkOffset = si;
|
|
|
|
} else {
|
2022-10-28 06:11:55 +00:00
|
|
|
ucstrTextAccess(ut, si, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Add a terminating NUL if space in the buffer permits,
|
|
|
|
// and set the error status as required.
|
|
|
|
u_terminateUChars(dest, destCapacity, di, pErrorCode);
|
|
|
|
return di;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct UTextFuncs ucstrFuncs =
|
|
|
|
{
|
|
|
|
sizeof(UTextFuncs),
|
|
|
|
0, 0, 0, // Reserved alignment padding
|
|
|
|
ucstrTextClone,
|
|
|
|
ucstrTextLength,
|
|
|
|
ucstrTextAccess,
|
|
|
|
ucstrTextExtract,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // Replace
|
|
|
|
nullptr, // Copy
|
|
|
|
nullptr, // MapOffsetToNative,
|
|
|
|
nullptr, // MapIndexToUTF16,
|
2020-08-11 09:10:23 +00:00
|
|
|
ucstrTextClose,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // spare 1
|
|
|
|
nullptr, // spare 2
|
|
|
|
nullptr, // spare 3
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
U_CDECL_END
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
static const char16_t gEmptyUString[] = {0};
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
2023-05-23 00:05:01 +00:00
|
|
|
utext_openUChars(UText *ut, const char16_t *s, int64_t length, UErrorCode *status) {
|
2020-08-11 09:10:23 +00:00
|
|
|
if (U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(s==nullptr && length==0) {
|
2020-08-11 09:10:23 +00:00
|
|
|
s = gEmptyUString;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if (s==nullptr || length < -1 || length>INT32_MAX) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status = U_ILLEGAL_ARGUMENT_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
ut = utext_setup(ut, 0, status);
|
|
|
|
if (U_SUCCESS(*status)) {
|
|
|
|
ut->pFuncs = &ucstrFuncs;
|
|
|
|
ut->context = s;
|
|
|
|
ut->providerProperties = I32_FLAG(UTEXT_PROVIDER_STABLE_CHUNKS);
|
|
|
|
if (length==-1) {
|
|
|
|
ut->providerProperties |= I32_FLAG(UTEXT_PROVIDER_LENGTH_IS_EXPENSIVE);
|
|
|
|
}
|
|
|
|
ut->a = length;
|
|
|
|
ut->chunkContents = s;
|
|
|
|
ut->chunkNativeStart = 0;
|
|
|
|
ut->chunkNativeLimit = length>=0? length : 0;
|
|
|
|
ut->chunkLength = (int32_t)ut->chunkNativeLimit;
|
|
|
|
ut->chunkOffset = 0;
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
//
|
|
|
|
// UText implementation for text from ICU CharacterIterators
|
|
|
|
//
|
|
|
|
// Use of UText data members:
|
|
|
|
// context pointer to the CharacterIterator
|
|
|
|
// a length of the full text.
|
|
|
|
// p pointer to buffer 1
|
|
|
|
// b start index of local buffer 1 contents
|
|
|
|
// q pointer to buffer 2
|
|
|
|
// c start index of local buffer 2 contents
|
|
|
|
// r pointer to the character iterator if the UText owns it.
|
|
|
|
// Null otherwise.
|
|
|
|
//
|
|
|
|
//------------------------------------------------------------------------------
|
|
|
|
#define CIBufSize 16
|
|
|
|
|
|
|
|
U_CDECL_BEGIN
|
|
|
|
static void U_CALLCONV
|
|
|
|
charIterTextClose(UText *ut) {
|
|
|
|
// Most of the work of close is done by the generic UText framework close.
|
|
|
|
// All that needs to be done here is delete the CharacterIterator if the UText
|
|
|
|
// owns it. This occurs if the UText was created by cloning.
|
|
|
|
CharacterIterator *ci = (CharacterIterator *)ut->r;
|
|
|
|
delete ci;
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->r = nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int64_t U_CALLCONV
|
|
|
|
charIterTextLength(UText *ut) {
|
|
|
|
return (int32_t)ut->a;
|
|
|
|
}
|
|
|
|
|
|
|
|
static UBool U_CALLCONV
|
|
|
|
charIterTextAccess(UText *ut, int64_t index, UBool forward) {
|
|
|
|
CharacterIterator *ci = (CharacterIterator *)ut->context;
|
|
|
|
|
|
|
|
int32_t clippedIndex = (int32_t)index;
|
|
|
|
if (clippedIndex<0) {
|
|
|
|
clippedIndex=0;
|
|
|
|
} else if (clippedIndex>=ut->a) {
|
|
|
|
clippedIndex=(int32_t)ut->a;
|
|
|
|
}
|
|
|
|
int32_t neededIndex = clippedIndex;
|
|
|
|
if (!forward && neededIndex>0) {
|
|
|
|
// reverse iteration, want the position just before what was asked for.
|
|
|
|
neededIndex--;
|
|
|
|
} else if (forward && neededIndex==ut->a && neededIndex>0) {
|
|
|
|
// Forward iteration, don't ask for something past the end of the text.
|
|
|
|
neededIndex--;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Find the native index of the start of the buffer containing what we want.
|
|
|
|
neededIndex -= neededIndex % CIBufSize;
|
|
|
|
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *buf = nullptr;
|
2022-10-28 06:11:55 +00:00
|
|
|
UBool needChunkSetup = true;
|
2020-08-11 09:10:23 +00:00
|
|
|
int i;
|
|
|
|
if (ut->chunkNativeStart == neededIndex) {
|
|
|
|
// The buffer we want is already the current chunk.
|
2022-10-28 06:11:55 +00:00
|
|
|
needChunkSetup = false;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else if (ut->b == neededIndex) {
|
|
|
|
// The first buffer (buffer p) has what we need.
|
2023-05-23 00:05:01 +00:00
|
|
|
buf = (char16_t *)ut->p;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else if (ut->c == neededIndex) {
|
|
|
|
// The second buffer (buffer q) has what we need.
|
2023-05-23 00:05:01 +00:00
|
|
|
buf = (char16_t *)ut->q;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
|
|
|
// Neither buffer already has what we need.
|
|
|
|
// Load new data from the character iterator.
|
|
|
|
// Use the buf that is not the current buffer.
|
2023-05-23 00:05:01 +00:00
|
|
|
buf = (char16_t *)ut->p;
|
2020-08-11 09:10:23 +00:00
|
|
|
if (ut->p == ut->chunkContents) {
|
2023-05-23 00:05:01 +00:00
|
|
|
buf = (char16_t *)ut->q;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
ci->setIndex(neededIndex);
|
|
|
|
for (i=0; i<CIBufSize; i++) {
|
|
|
|
buf[i] = ci->nextPostInc();
|
|
|
|
if (i+neededIndex > ut->a) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We have a buffer with the data we need.
|
|
|
|
// Set it up as the current chunk, if it wasn't already.
|
|
|
|
if (needChunkSetup) {
|
|
|
|
ut->chunkContents = buf;
|
|
|
|
ut->chunkLength = CIBufSize;
|
|
|
|
ut->chunkNativeStart = neededIndex;
|
|
|
|
ut->chunkNativeLimit = neededIndex + CIBufSize;
|
|
|
|
if (ut->chunkNativeLimit > ut->a) {
|
|
|
|
ut->chunkNativeLimit = ut->a;
|
|
|
|
ut->chunkLength = (int32_t)(ut->chunkNativeLimit)-(int32_t)(ut->chunkNativeStart);
|
|
|
|
}
|
|
|
|
ut->nativeIndexingLimit = ut->chunkLength;
|
|
|
|
U_ASSERT(ut->chunkOffset>=0 && ut->chunkOffset<=CIBufSize);
|
|
|
|
}
|
|
|
|
ut->chunkOffset = clippedIndex - (int32_t)ut->chunkNativeStart;
|
|
|
|
UBool success = (forward? ut->chunkOffset<ut->chunkLength : ut->chunkOffset>0);
|
|
|
|
return success;
|
|
|
|
}
|
|
|
|
|
|
|
|
static UText * U_CALLCONV
|
|
|
|
charIterTextClone(UText *dest, const UText *src, UBool deep, UErrorCode * status) {
|
|
|
|
if (U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (deep) {
|
|
|
|
// There is no CharacterIterator API for cloning the underlying text storage.
|
|
|
|
*status = U_UNSUPPORTED_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
} else {
|
|
|
|
CharacterIterator *srcCI =(CharacterIterator *)src->context;
|
|
|
|
srcCI = srcCI->clone();
|
|
|
|
dest = utext_openCharacterIterator(dest, srcCI, status);
|
|
|
|
if (U_FAILURE(*status)) {
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
// cast off const on getNativeIndex.
|
|
|
|
// For CharacterIterator based UTexts, this is safe, the operation is const.
|
|
|
|
int64_t ix = utext_getNativeIndex((UText *)src);
|
|
|
|
utext_setNativeIndex(dest, ix);
|
|
|
|
dest->r = srcCI; // flags that this UText owns the CharacterIterator
|
|
|
|
}
|
|
|
|
return dest;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int32_t U_CALLCONV
|
|
|
|
charIterTextExtract(UText *ut,
|
|
|
|
int64_t start, int64_t limit,
|
2023-05-23 00:05:01 +00:00
|
|
|
char16_t *dest, int32_t destCapacity,
|
2020-08-11 09:10:23 +00:00
|
|
|
UErrorCode *status)
|
|
|
|
{
|
|
|
|
if(U_FAILURE(*status)) {
|
|
|
|
return 0;
|
|
|
|
}
|
2023-05-23 00:05:01 +00:00
|
|
|
if(destCapacity<0 || (dest==nullptr && destCapacity>0) || start>limit) {
|
2020-08-11 09:10:23 +00:00
|
|
|
*status=U_ILLEGAL_ARGUMENT_ERROR;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
int32_t length = (int32_t)ut->a;
|
|
|
|
int32_t start32 = pinIndex(start, length);
|
|
|
|
int32_t limit32 = pinIndex(limit, length);
|
|
|
|
int32_t desti = 0;
|
|
|
|
int32_t srci;
|
|
|
|
int32_t copyLimit;
|
|
|
|
|
|
|
|
CharacterIterator *ci = (CharacterIterator *)ut->context;
|
|
|
|
ci->setIndex32(start32); // Moves ix to lead of surrogate pair, if needed.
|
|
|
|
srci = ci->getIndex();
|
|
|
|
copyLimit = srci;
|
|
|
|
while (srci<limit32) {
|
|
|
|
UChar32 c = ci->next32PostInc();
|
|
|
|
int32_t len = U16_LENGTH(c);
|
|
|
|
U_ASSERT(desti+len>0); /* to ensure desti+len never exceeds MAX_INT32, which must not happen logically */
|
|
|
|
if (desti+len <= destCapacity) {
|
|
|
|
U16_APPEND_UNSAFE(dest, desti, c);
|
|
|
|
copyLimit = srci+len;
|
|
|
|
} else {
|
|
|
|
desti += len;
|
|
|
|
*status = U_BUFFER_OVERFLOW_ERROR;
|
|
|
|
}
|
|
|
|
srci += len;
|
|
|
|
}
|
|
|
|
|
2022-10-28 06:11:55 +00:00
|
|
|
charIterTextAccess(ut, copyLimit, true);
|
2020-08-11 09:10:23 +00:00
|
|
|
|
|
|
|
u_terminateUChars(dest, destCapacity, desti, status);
|
|
|
|
return desti;
|
|
|
|
}
|
|
|
|
|
|
|
|
static const struct UTextFuncs charIterFuncs =
|
|
|
|
{
|
|
|
|
sizeof(UTextFuncs),
|
|
|
|
0, 0, 0, // Reserved alignment padding
|
|
|
|
charIterTextClone,
|
|
|
|
charIterTextLength,
|
|
|
|
charIterTextAccess,
|
|
|
|
charIterTextExtract,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // Replace
|
|
|
|
nullptr, // Copy
|
|
|
|
nullptr, // MapOffsetToNative,
|
|
|
|
nullptr, // MapIndexToUTF16,
|
2020-08-11 09:10:23 +00:00
|
|
|
charIterTextClose,
|
2023-05-23 00:05:01 +00:00
|
|
|
nullptr, // spare 1
|
|
|
|
nullptr, // spare 2
|
|
|
|
nullptr // spare 3
|
2020-08-11 09:10:23 +00:00
|
|
|
};
|
|
|
|
U_CDECL_END
|
|
|
|
|
|
|
|
|
|
|
|
U_CAPI UText * U_EXPORT2
|
|
|
|
utext_openCharacterIterator(UText *ut, CharacterIterator *ci, UErrorCode *status) {
|
|
|
|
if (U_FAILURE(*status)) {
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (ci->startIndex() > 0) {
|
|
|
|
// No support for CharacterIterators that do not start indexing from zero.
|
|
|
|
*status = U_UNSUPPORTED_ERROR;
|
2023-05-23 00:05:01 +00:00
|
|
|
return nullptr;
|
2020-08-11 09:10:23 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Extra space in UText for 2 buffers of CIBufSize UChars each.
|
2023-05-23 00:05:01 +00:00
|
|
|
int32_t extraSpace = 2 * CIBufSize * sizeof(char16_t);
|
2020-08-11 09:10:23 +00:00
|
|
|
ut = utext_setup(ut, extraSpace, status);
|
|
|
|
if (U_SUCCESS(*status)) {
|
|
|
|
ut->pFuncs = &charIterFuncs;
|
|
|
|
ut->context = ci;
|
|
|
|
ut->providerProperties = 0;
|
|
|
|
ut->a = ci->endIndex(); // Length of text
|
|
|
|
ut->p = ut->pExtra; // First buffer
|
|
|
|
ut->b = -1; // Native index of first buffer contents
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->q = (char16_t*)ut->pExtra+CIBufSize; // Second buffer
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->c = -1; // Native index of second buffer contents
|
|
|
|
|
|
|
|
// Initialize current chunk contents to be empty.
|
|
|
|
// First access will fault something in.
|
|
|
|
// Note: The initial nativeStart and chunkOffset must sum to zero
|
|
|
|
// so that getNativeIndex() will correctly compute to zero
|
|
|
|
// if no call to Access() has ever been made. They can't be both
|
|
|
|
// zero without Access() thinking that the chunk is valid.
|
2023-05-23 00:05:01 +00:00
|
|
|
ut->chunkContents = (char16_t *)ut->p;
|
2020-08-11 09:10:23 +00:00
|
|
|
ut->chunkNativeStart = -1;
|
|
|
|
ut->chunkOffset = 1;
|
|
|
|
ut->chunkNativeLimit = 0;
|
|
|
|
ut->chunkLength = 0;
|
|
|
|
ut->nativeIndexingLimit = ut->chunkOffset; // enables native indexing
|
|
|
|
}
|
|
|
|
return ut;
|
|
|
|
}
|