mirror of
https://github.com/CloverHackyColor/CloverBootloader.git
synced 2024-12-28 17:08:18 +01:00
127 lines
3.5 KiB
C++
Executable File
127 lines
3.5 KiB
C++
Executable File
#include <Platform.h>
|
|
#include "../cpp_foundation/XString.h"
|
|
#include "../cpp_foundation/utf8Conversion.h"
|
|
#include "global_test.h"
|
|
|
|
|
|
//#include <wchar.h>
|
|
|
|
|
|
int XString_tests()
|
|
{
|
|
|
|
#ifdef JIEF_DEBUG
|
|
// DebugLog(2, "XStringW_tests -> Enter\n");
|
|
#endif
|
|
|
|
if ( global_str1 != "global_str1" ) return 1;
|
|
if ( global_str2 != "global_str2" ) return 2;
|
|
|
|
#ifdef XSTRINGW_HAS_CTOR_LITTERAL
|
|
XStringW str("1");
|
|
if ( str != "1" ) return 3;
|
|
str.StrCat("2");
|
|
if ( str != "12" ) return 4;
|
|
#endif
|
|
|
|
XString str;
|
|
str.takeValueFrom("12");
|
|
XString str2;
|
|
if ( !str2.isEmpty() ) return 10;
|
|
str2.StrnCpy(str.data(), 2);
|
|
if ( str2 != "12" ) return 11;
|
|
str2.StrnCat("345", 2);
|
|
if ( str2 != "1234" ) return 12;
|
|
str2.Insert(1, str);
|
|
if ( str2 != "112234" ) return 13;
|
|
str2 += "6";
|
|
if ( str2 != "1122346" ) return 14;
|
|
|
|
str2 = "123";
|
|
str2.Insert(0, "x");
|
|
if ( str2 != "x123" ) return 15;
|
|
str2.Insert(4, "y");
|
|
if ( str2 != "x123y" ) return 16;
|
|
|
|
//wchar_t c2 = L'Ň';
|
|
//printf("1=%lc\n", c2);
|
|
//const char* s1 = "𐌾";
|
|
|
|
str2.SPrintf("%c", 'a'); // signle UTF8 ascii char
|
|
if ( str2 != "a" ) return 20;
|
|
str2.takeValueFrom("ab"); // UTF16(32) string containing ascii char
|
|
if ( str2 != "ab" ) return 21;
|
|
#ifdef _MSC_VER
|
|
// IMPORTANT : you can't pass a litteral char in a vararg function with Visual Studio (Microsoft strikes again :-).
|
|
// At least, you got a warning C4066
|
|
// IMPORTANT2 : Litteral string containing UTF16 char are WRONG. And you don't get a warning !!! If litteral is only ascii, it's ok.
|
|
// Maybe it's compilation option butI didn't find them.
|
|
wchar_t c = 'Ň'; // using an imtermediary var for Microsoft.
|
|
|
|
|
|
wchar_t s[2]; // wchar_t s2[] = "Ň";
|
|
s[0] = 'Ň';
|
|
s[1] = 0;
|
|
|
|
str2.SPrintf("%lc", c); // UTF16(32) char. (2 bytes in total if UTF16)
|
|
if (str2 != s) return 22;
|
|
str2.takeValueFrom("");
|
|
if (str2.length() != 0) return 221;
|
|
str2.takeValueFrom(s); // this is a UTF8 string 2 bytes long
|
|
if (str2 != s) return 23;
|
|
#else
|
|
str2.SPrintf("%lc", L'Ň'); // signe UTF16(32) char. (2 bytes in total if UTF16)
|
|
if ( str2 != "Ň" ) return 22;
|
|
str2.takeValueFrom("");
|
|
if (str2.length() != 0) return 221;
|
|
#ifdef XSTRING_HAS_CTOR_LITTERAL
|
|
str2.takeValueFrom("Ň"); // this is a UTF8 string 2 bytes long
|
|
if (str2 != "Ň") return 23; // utf8 litteral are converted to an XStringW if ctor is available.
|
|
#endif
|
|
str2.takeValueFrom("");
|
|
if (str2.length() != 0) return 231;
|
|
#ifdef XSTRING_HAS_CTOR_LITTERAL
|
|
str2.takeValueFrom("Ň"); // this is a UTF8 string 2 bytes long
|
|
if (str2 != "Ň") return 24;
|
|
#endif
|
|
#endif
|
|
|
|
#if __WCHAR_MAX__ > 0xFFFFu
|
|
str2.SPrintf("%lc", L'𐌾'); // L'𐌾' // this char cannot convert to an UTF16 char. So it doesn't compile with -fshort-wchar
|
|
if ( str2 != L'𐌾' ) return 30;
|
|
#endif
|
|
|
|
|
|
#ifndef _MSC_VER
|
|
// "𐌾" in UTF16 is 2 char : 0xd800, 0xdf3e
|
|
|
|
str2.takeValueFrom("𐌾"); // this is a UTF8 string 4 bytes long
|
|
if ( str2 != "𐌾" ) return 31;
|
|
str2.takeValueFrom("𐌾"); // this is a UTF16 or UTF32 string (depending of -fshort-wchar)
|
|
if ( str2 != "𐌾" ) return 32;
|
|
|
|
#ifdef XSTRINGW_HAS_CTOR_LITTERAL
|
|
{
|
|
XStringW str3("a");
|
|
if ( str3 != "a" ) return 40;
|
|
XStringW str4("aŇ𐌾");
|
|
if ( str4 != "aŇ𐌾" ) return 41;
|
|
}
|
|
#endif
|
|
#endif
|
|
|
|
/* Stil doesn't work as VS doesn't encode correctly litteral */
|
|
#ifndef _MSC_VER
|
|
|
|
// int i1 = sizeof(wchar_t);
|
|
// int i2 = sizeof(char16_t);
|
|
// const wchar_t* ls = L"Выход";
|
|
// const char16_t* us = u"Выход";
|
|
// const char32_t* Us = U"Выход";
|
|
XString str3(L"Выход");
|
|
if ( str3 != "Выход" ) return 50;
|
|
#endif
|
|
|
|
return 0;
|
|
}
|