字符串编码转换

字符串编码转换波及宽字节表示法与UTF-8表示法之间的转换、宽字节表示法与UTF-16表示法之间的转换、UTF-8表示法与UTF-16表示法之间的转换、UTF-16表示法于ASCII表示法之间的转换、ASCII表示法宽字节表示法之间的转换。

相干文件

  • base/strings/utf\_string\_conversions.h // 字符串编码转换定义
  • base/strings/utf\_string\_conversions.cc // 字符串编码转换实现
  • base/strings/utf\_string\_conversions\_fuzzer.cc // 字符串编码转换实现

办法定义

// base/strings/utf_string_conversions.h namespace base {// These convert between UTF-8, -16, and -32 strings. They are potentially slow,// so avoid unnecessary conversions. The low-level versions return a boolean// indicating whether the conversion was 100% valid. In this case, it will still// do the best it can and put the result in the output buffer. The versions that// return strings ignore this error and just return the best conversion// possible.BASE_EXPORT bool WideToUTF8(const wchar_t* src, size_t src_len,                            std::string* output);BASE_EXPORT std::string WideToUTF8(WStringPiece wide) WARN_UNUSED_RESULT;BASE_EXPORT bool UTF8ToWide(const char* src, size_t src_len,                            std::wstring* output);BASE_EXPORT std::wstring UTF8ToWide(StringPiece utf8) WARN_UNUSED_RESULT;BASE_EXPORT bool WideToUTF16(const wchar_t* src,                             size_t src_len,                             std::u16string* output);BASE_EXPORT std::u16string WideToUTF16(WStringPiece wide) WARN_UNUSED_RESULT;BASE_EXPORT bool UTF16ToWide(const char16_t* src,                             size_t src_len,                             std::wstring* output);BASE_EXPORT std::wstring UTF16ToWide(StringPiece16 utf16) WARN_UNUSED_RESULT;BASE_EXPORT bool UTF8ToUTF16(const char* src,                             size_t src_len,                             std::u16string* output);BASE_EXPORT std::u16string UTF8ToUTF16(StringPiece utf8) WARN_UNUSED_RESULT;BASE_EXPORT bool UTF16ToUTF8(const char16_t* src,                             size_t src_len,                             std::string* output);BASE_EXPORT std::string UTF16ToUTF8(StringPiece16 utf16) WARN_UNUSED_RESULT;// This converts an ASCII string, typically a hardcoded constant, to a UTF16// string.BASE_EXPORT std::u16string ASCIIToUTF16(StringPiece ascii) WARN_UNUSED_RESULT;// Converts to 7-bit ASCII by truncating. The result must be known to be ASCII// beforehand.BASE_EXPORT std::string UTF16ToASCII(StringPiece16 utf16) WARN_UNUSED_RESULT;#if defined(WCHAR_T_IS_UTF16)// This converts an ASCII string, typically a hardcoded constant, to a wide// string.BASE_EXPORT std::wstring ASCIIToWide(StringPiece ascii) WARN_UNUSED_RESULT;// Converts to 7-bit ASCII by truncating. The result must be known to be ASCII// beforehand.BASE_EXPORT std::string WideToASCII(WStringPiece wide) WARN_UNUSED_RESULT;#endif  // defined(WCHAR_T_IS_UTF16)// The conversion functions in this file should not be used to convert string// literals. Instead, the corresponding prefixes (e.g. u"" for UTF16 or L"" for// Wide) should be used. Deleting the overloads here catches these cases at// compile time.template <size_t N>std::u16string WideToUTF16(const wchar_t (&str)[N]) {  static_assert(N == 0, "Error: Use the u\"...\" prefix instead.");  return std::u16string();}// TODO(crbug.com/1189439): Also disallow passing string constants in tests.#if !defined(UNIT_TEST)template <size_t N>std::u16string ASCIIToUTF16(const char (&str)[N]) {  static_assert(N == 0, "Error: Use the u\"...\" prefix instead.");  return std::u16string();}// Mutable character arrays are usually only populated during runtime. Continue// to allow this conversion.template <size_t N>std::u16string ASCIIToUTF16(char (&str)[N]) {  return ASCIIToUTF16(StringPiece(str));}#endif}  // namespace base

办法实现

// base/strings/utf_string_conversions.ccnamespace base {namespace {constexpr int32_t kErrorCodePoint = 0xFFFD;// Size coefficient ----------------------------------------------------------// The maximum number of codeunits in the destination encoding corresponding to// one codeunit in the source encoding.template <typename SrcChar, typename DestChar>struct SizeCoefficient {  static_assert(sizeof(SrcChar) < sizeof(DestChar),                "Default case: from a smaller encoding to the bigger one");  // ASCII symbols are encoded by one codeunit in all encodings.  static constexpr int value = 1;};template <>struct SizeCoefficient<char16_t, char> {  // One UTF-16 codeunit corresponds to at most 3 codeunits in UTF-8.  static constexpr int value = 3;};#if defined(WCHAR_T_IS_UTF32)template <>struct SizeCoefficient<wchar_t, char> {  // UTF-8 uses at most 4 codeunits per character.  static constexpr int value = 4;};template <>struct SizeCoefficient<wchar_t, char16_t> {  // UTF-16 uses at most 2 codeunits per character.  static constexpr int value = 2;};#endif  // defined(WCHAR_T_IS_UTF32)template <typename SrcChar, typename DestChar>constexpr int size_coefficient_v =    SizeCoefficient<std::decay_t<SrcChar>, std::decay_t<DestChar>>::value;// UnicodeAppendUnsafe --------------------------------------------------------// Function overloads that write code_point to the output string. Output string// has to have enough space for the codepoint.// Convenience typedef that checks whether the passed in type is integral (i.e.// bool, char, int or their extended versions) and is of the correct size.template <typename Char, size_t N>using EnableIfBitsAre = std::enable_if_t<std::is_integral<Char>::value &&                                             CHAR_BIT * sizeof(Char) == N,                                         bool>;template <typename Char, EnableIfBitsAre<Char, 8> = true>void UnicodeAppendUnsafe(Char* out, int32_t* size, uint32_t code_point) {  CBU8_APPEND_UNSAFE(out, *size, code_point);}template <typename Char, EnableIfBitsAre<Char, 16> = true>void UnicodeAppendUnsafe(Char* out, int32_t* size, uint32_t code_point) {  CBU16_APPEND_UNSAFE(out, *size, code_point);}template <typename Char, EnableIfBitsAre<Char, 32> = true>void UnicodeAppendUnsafe(Char* out, int32_t* size, uint32_t code_point) {  out[(*size)++] = code_point;}// DoUTFConversion ------------------------------------------------------------// Main driver of UTFConversion specialized for different Src encodings.// dest has to have enough room for the converted text.template <typename DestChar>bool DoUTFConversion(const char* src,                     int32_t src_len,                     DestChar* dest,                     int32_t* dest_len) {  bool success = true;  for (int32_t i = 0; i < src_len;) {    int32_t code_point;    CBU8_NEXT(src, i, src_len, code_point);    if (!IsValidCodepoint(code_point)) {      success = false;      code_point = kErrorCodePoint;    }    UnicodeAppendUnsafe(dest, dest_len, code_point);  }  return success;}template <typename DestChar>bool DoUTFConversion(const char16_t* src,                     int32_t src_len,                     DestChar* dest,                     int32_t* dest_len) {  bool success = true;  auto ConvertSingleChar = [&success](char16_t in) -> int32_t {    if (!CBU16_IS_SINGLE(in) || !IsValidCodepoint(in)) {      success = false;      return kErrorCodePoint;    }    return in;  };  int32_t i = 0;  // Always have another symbol in order to avoid checking boundaries in the  // middle of the surrogate pair.  while (i < src_len - 1) {    int32_t code_point;    if (CBU16_IS_LEAD(src[i]) && CBU16_IS_TRAIL(src[i + 1])) {      code_point = CBU16_GET_SUPPLEMENTARY(src[i], src[i + 1]);      if (!IsValidCodepoint(code_point)) {        code_point = kErrorCodePoint;        success = false;      }      i += 2;    } else {      code_point = ConvertSingleChar(src[i]);      ++i;    }    UnicodeAppendUnsafe(dest, dest_len, code_point);  }  if (i < src_len)    UnicodeAppendUnsafe(dest, dest_len, ConvertSingleChar(src[i]));  return success;}#if defined(WCHAR_T_IS_UTF32)template <typename DestChar>bool DoUTFConversion(const wchar_t* src,                     int32_t src_len,                     DestChar* dest,                     int32_t* dest_len) {  bool success = true;  for (int32_t i = 0; i < src_len; ++i) {    int32_t code_point = src[i];    if (!IsValidCodepoint(code_point)) {      success = false;      code_point = kErrorCodePoint;    }    UnicodeAppendUnsafe(dest, dest_len, code_point);  }  return success;}#endif  // defined(WCHAR_T_IS_UTF32)// UTFConversion --------------------------------------------------------------// Function template for generating all UTF conversions.template <typename InputString, typename DestString>bool UTFConversion(const InputString& src_str, DestString* dest_str) {  if (IsStringASCII(src_str)) {    dest_str->assign(src_str.begin(), src_str.end());    return true;  }  dest_str->resize(src_str.length() *                   size_coefficient_v<typename InputString::value_type,                                      typename DestString::value_type>);  // Empty string is ASCII => it OK to call operator[].  auto* dest = &(*dest_str)[0];  // ICU requires 32 bit numbers.  int32_t src_len32 = static_cast<int32_t>(src_str.length());  int32_t dest_len32 = 0;  bool res = DoUTFConversion(src_str.data(), src_len32, dest, &dest_len32);  dest_str->resize(dest_len32);  dest_str->shrink_to_fit();  return res;}}  // namespace// UTF16 <-> UTF8 --------------------------------------------------------------bool UTF8ToUTF16(const char* src, size_t src_len, std::u16string* output) {  return UTFConversion(StringPiece(src, src_len), output);}std::u16string UTF8ToUTF16(StringPiece utf8) {  std::u16string ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  UTF8ToUTF16(utf8.data(), utf8.size(), &ret);  return ret;}bool UTF16ToUTF8(const char16_t* src, size_t src_len, std::string* output) {  return UTFConversion(StringPiece16(src, src_len), output);}std::string UTF16ToUTF8(StringPiece16 utf16) {  std::string ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  UTF16ToUTF8(utf16.data(), utf16.length(), &ret);  return ret;}// UTF-16 <-> Wide -------------------------------------------------------------#if defined(WCHAR_T_IS_UTF16)// When wide == UTF-16 the conversions are a NOP.bool WideToUTF16(const wchar_t* src, size_t src_len, std::u16string* output) {  output->assign(src, src + src_len);  return true;}std::u16string WideToUTF16(WStringPiece wide) {  return std::u16string(wide.begin(), wide.end());}bool UTF16ToWide(const char16_t* src, size_t src_len, std::wstring* output) {  output->assign(src, src + src_len);  return true;}std::wstring UTF16ToWide(StringPiece16 utf16) {  return std::wstring(utf16.begin(), utf16.end());}#elif defined(WCHAR_T_IS_UTF32)bool WideToUTF16(const wchar_t* src, size_t src_len, std::u16string* output) {  return UTFConversion(base::WStringPiece(src, src_len), output);}std::u16string WideToUTF16(WStringPiece wide) {  std::u16string ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  WideToUTF16(wide.data(), wide.length(), &ret);  return ret;}bool UTF16ToWide(const char16_t* src, size_t src_len, std::wstring* output) {  return UTFConversion(StringPiece16(src, src_len), output);}std::wstring UTF16ToWide(StringPiece16 utf16) {  std::wstring ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  UTF16ToWide(utf16.data(), utf16.length(), &ret);  return ret;}#endif  // defined(WCHAR_T_IS_UTF32)// UTF-8 <-> Wide --------------------------------------------------------------// UTF8ToWide is the same code, regardless of whether wide is 16 or 32 bitsbool UTF8ToWide(const char* src, size_t src_len, std::wstring* output) {  return UTFConversion(StringPiece(src, src_len), output);}std::wstring UTF8ToWide(StringPiece utf8) {  std::wstring ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  UTF8ToWide(utf8.data(), utf8.length(), &ret);  return ret;}#if defined(WCHAR_T_IS_UTF16)// Easy case since we can use the "utf" versions we already wrote above.bool WideToUTF8(const wchar_t* src, size_t src_len, std::string* output) {  return UTF16ToUTF8(as_u16cstr(src), src_len, output);}std::string WideToUTF8(WStringPiece wide) {  return UTF16ToUTF8(StringPiece16(as_u16cstr(wide), wide.size()));}#elif defined(WCHAR_T_IS_UTF32)bool WideToUTF8(const wchar_t* src, size_t src_len, std::string* output) {  return UTFConversion(WStringPiece(src, src_len), output);}std::string WideToUTF8(WStringPiece wide) {  std::string ret;  // Ignore the success flag of this call, it will do the best it can for  // invalid input, which is what we want here.  WideToUTF8(wide.data(), wide.length(), &ret);  return ret;}#endif  // defined(WCHAR_T_IS_UTF32)std::u16string ASCIIToUTF16(StringPiece ascii) {  DCHECK(IsStringASCII(ascii)) << ascii;  return std::u16string(ascii.begin(), ascii.end());}std::string UTF16ToASCII(StringPiece16 utf16) {  DCHECK(IsStringASCII(utf16)) << UTF16ToUTF8(utf16);  return std::string(utf16.begin(), utf16.end());}#if defined(WCHAR_T_IS_UTF16)std::wstring ASCIIToWide(StringPiece ascii) {  DCHECK(IsStringASCII(ascii)) << ascii;  return std::wstring(ascii.begin(), ascii.end());}std::string WideToASCII(WStringPiece wide) {  DCHECK(IsStringASCII(wide)) << wide;  return std::string(wide.begin(), wide.end());}#endif  // defined(WCHAR_T_IS_UTF16)}  // namespace base

fuzzer

// Entry point for LibFuzzer.extern "C" int LLVMFuzzerTestOneInput(const uint8_t* data, size_t size) {  base::StringPiece string_piece_input(reinterpret_cast<const char*>(data),                                       size);  ignore_result(base::UTF8ToWide(string_piece_input));  base::UTF8ToWide(reinterpret_cast<const char*>(data), size,                   &output_std_wstring);  ignore_result(base::UTF8ToUTF16(string_piece_input));  base::UTF8ToUTF16(reinterpret_cast<const char*>(data), size,                    &output_string16);  // Test for char16_t.  if (size % 2 == 0) {    base::StringPiece16 string_piece_input16(        reinterpret_cast<const char16_t*>(data), size / 2);    ignore_result(base::UTF16ToWide(output_string16));    base::UTF16ToWide(reinterpret_cast<const char16_t*>(data), size / 2,                      &output_std_wstring);    ignore_result(base::UTF16ToUTF8(string_piece_input16));    base::UTF16ToUTF8(reinterpret_cast<const char16_t*>(data), size / 2,                      &output_std_string);  }  // Test for wchar_t.  size_t wchar_t_size = sizeof(wchar_t);  if (size % wchar_t_size == 0) {    ignore_result(base::WideToUTF8(output_std_wstring));    base::WideToUTF8(reinterpret_cast<const wchar_t*>(data),                     size / wchar_t_size, &output_std_string);    ignore_result(base::WideToUTF16(output_std_wstring));    base::WideToUTF16(reinterpret_cast<const wchar_t*>(data),                      size / wchar_t_size, &output_string16);  }  // Test for ASCII. This condition is needed to avoid hitting instant CHECK  // failures.  if (base::IsStringASCII(string_piece_input)) {    output_string16 = base::ASCIIToUTF16(string_piece_input);    base::StringPiece16 string_piece_input16(output_string16);    ignore_result(base::UTF16ToASCII(string_piece_input16));  }  return 0;}