diff options
author | Michael Mair-Keimberger <mmk@levelnine.at> | 2022-03-12 19:45:00 +0100 |
---|---|---|
committer | David Seifert <soap@gentoo.org> | 2022-03-12 19:45:00 +0100 |
commit | 341a76663bacc1a3582b75528c547f3c16f42da3 (patch) | |
tree | 5175142f2316ab676ed6fd02bcd0f915a455bedd /app-i18n/mozc | |
parent | dev-util/nvidia-cuda-sdk: drop 10.2.89-r1 (diff) | |
download | gentoo-341a76663bacc1a3582b75528c547f3c16f42da3.tar.gz gentoo-341a76663bacc1a3582b75528c547f3c16f42da3.tar.bz2 gentoo-341a76663bacc1a3582b75528c547f3c16f42da3.zip |
app-i18n/mozc: remove unused patches
Closes: https://github.com/gentoo/gentoo/pull/24515
Package-Manager: Portage-3.0.30, Repoman-3.0.3
Signed-off-by: Michael Mair-Keimberger <mmk@levelnine.at>
Signed-off-by: David Seifert <soap@gentoo.org>
Diffstat (limited to 'app-i18n/mozc')
12 files changed, 0 insertions, 3009 deletions
diff --git a/app-i18n/mozc/files/mozc-2.20.2673.102-tests_build.patch b/app-i18n/mozc/files/mozc-2.20.2673.102-tests_build.patch deleted file mode 100644 index b8b4477437f8..000000000000 --- a/app-i18n/mozc/files/mozc-2.20.2673.102-tests_build.patch +++ /dev/null @@ -1,27 +0,0 @@ ---- /src/net/json_util_test.cc -+++ /src/net/json_util_test.cc -@@ -784,13 +784,13 @@ - for (size_t i = 0; i < arraysize(kNumS32ValueKeys); ++i) { - { - Json::Value json_value; -- json_value[kNumS32ValueKeys[i]] = -2147483649ll; -+ json_value[kNumS32ValueKeys[i]] = static_cast<Json::Value::Int64>(-2147483649ll); - TestMsg msg; - EXPECT_FALSE(JsonUtil::JsonValueToProtobufMessage(json_value, &msg)); - } - { - Json::Value json_value; -- json_value[kNumS32ValueKeys[i]] = 2147483648ull; -+ json_value[kNumS32ValueKeys[i]] = static_cast<Json::Value::UInt64>(2147483648ull); - TestMsg msg; - EXPECT_FALSE(JsonUtil::JsonValueToProtobufMessage(json_value, &msg)); - } -@@ -805,7 +805,7 @@ - } - { - Json::Value json_value; -- json_value[kNumU32ValueKeys[i]] = 4294967296ull; -+ json_value[kNumU32ValueKeys[i]] = static_cast<Json::Value::UInt64>(4294967296ull); - TestMsg msg; - EXPECT_FALSE(JsonUtil::JsonValueToProtobufMessage(json_value, &msg)); - } diff --git a/app-i18n/mozc/files/mozc-2.20.2673.102-tests_skipping.patch b/app-i18n/mozc/files/mozc-2.20.2673.102-tests_skipping.patch deleted file mode 100644 index 28c5486de50d..000000000000 --- a/app-i18n/mozc/files/mozc-2.20.2673.102-tests_skipping.patch +++ /dev/null @@ -1,70 +0,0 @@ -Disable test leaving mozc_server orphan process. - ---- /src/unix/ibus/mozc_engine_test.cc -+++ /src/unix/ibus/mozc_engine_test.cc -@@ -41,65 +41,5 @@ - namespace mozc { - namespace ibus { - --class LaunchToolTest : public testing::Test { -- public: -- LaunchToolTest() { -- g_type_init(); -- } -- -- protected: -- virtual void SetUp() { -- mozc_engine_.reset(new MozcEngine()); -- -- mock_ = new client::ClientMock(); -- mock_->ClearFunctionCounter(); -- mozc_engine_->client_.reset(mock_); -- } -- -- virtual void TearDown() { -- mozc_engine_.reset(); -- } -- -- client::ClientMock* mock_; -- unique_ptr<MozcEngine> mozc_engine_; -- -- private: -- DISALLOW_COPY_AND_ASSIGN(LaunchToolTest); --}; -- --TEST_F(LaunchToolTest, LaunchToolTest) { -- commands::Output output; -- -- // Launch config dialog -- mock_->ClearFunctionCounter(); -- mock_->SetBoolFunctionReturn("LaunchToolWithProtoBuf", true); -- output.set_launch_tool_mode(commands::Output::CONFIG_DIALOG); -- EXPECT_TRUE(mozc_engine_->LaunchTool(output)); -- -- // Launch dictionary tool -- mock_->ClearFunctionCounter(); -- mock_->SetBoolFunctionReturn("LaunchToolWithProtoBuf", true); -- output.set_launch_tool_mode(commands::Output::DICTIONARY_TOOL); -- EXPECT_TRUE(mozc_engine_->LaunchTool(output)); -- -- // Launch word register dialog -- mock_->ClearFunctionCounter(); -- mock_->SetBoolFunctionReturn("LaunchToolWithProtoBuf", true); -- output.set_launch_tool_mode(commands::Output::WORD_REGISTER_DIALOG); -- EXPECT_TRUE(mozc_engine_->LaunchTool(output)); -- -- // Launch no tool(means do nothing) -- mock_->ClearFunctionCounter(); -- mock_->SetBoolFunctionReturn("LaunchToolWithProtoBuf", false); -- output.set_launch_tool_mode(commands::Output::NO_TOOL); -- EXPECT_FALSE(mozc_engine_->LaunchTool(output)); -- -- // Something occurring in client::Client::LaunchTool -- mock_->ClearFunctionCounter(); -- mock_->SetBoolFunctionReturn("LaunchToolWithProtoBuf", false); -- output.set_launch_tool_mode(commands::Output::CONFIG_DIALOG); -- EXPECT_FALSE(mozc_engine_->LaunchTool(output)); --} -- - } // namespace ibus - } // namespace mozc diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-environmental_variables.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-environmental_variables.patch deleted file mode 100644 index 02e522a32e9e..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-environmental_variables.patch +++ /dev/null @@ -1,132 +0,0 @@ -https://github.com/google/mozc/issues/470 - ---- /src/base/system_util.cc -+++ /src/base/system_util.cc -@@ -208,28 +208,39 @@ - dir_ = "/"; - return; - #else // MOZC_USE_PEPPER_FILE_IO -+ const char *configuration_directory_env; - string dir; - - #ifdef OS_WIN -- DCHECK(SUCCEEDED(Singleton<LocalAppDataDirectoryCache>::get()->result())); -- dir = Singleton<LocalAppDataDirectoryCache>::get()->path(); -+ configuration_directory_env = ::getenv("MOZC_CONFIGURATION_DIRECTORY"); -+ if (configuration_directory_env) { -+ dir = configuration_directory_env; -+ } else { -+ DCHECK(SUCCEEDED(Singleton<LocalAppDataDirectoryCache>::get()->result())); -+ dir = Singleton<LocalAppDataDirectoryCache>::get()->path(); - #ifdef GOOGLE_JAPANESE_INPUT_BUILD -- dir = FileUtil::JoinPath(dir, kCompanyNameInEnglish); -- FileUtil::CreateDirectory(dir); -+ dir = FileUtil::JoinPath(dir, kCompanyNameInEnglish); -+ FileUtil::CreateDirectory(dir); - #endif // GOOGLE_JAPANESE_INPUT_BUILD -- dir = FileUtil::JoinPath(dir, kProductNameInEnglish); -+ dir = FileUtil::JoinPath(dir, kProductNameInEnglish); -+ } - - #elif defined(OS_MACOSX) -- dir = MacUtil::GetApplicationSupportDirectory(); -+ configuration_directory_env = ::getenv("MOZC_CONFIGURATION_DIRECTORY"); -+ if (configuration_directory_env) { -+ dir = configuration_directory_env; -+ } else { -+ dir = MacUtil::GetApplicationSupportDirectory(); - #ifdef GOOGLE_JAPANESE_INPUT_BUILD -- dir = FileUtil::JoinPath(dir, "Google"); -- // The permission of ~/Library/Application Support/Google seems to be 0755. -- // TODO(komatsu): nice to make a wrapper function. -- ::mkdir(dir.c_str(), 0755); -- dir = FileUtil::JoinPath(dir, "JapaneseInput"); -+ dir = FileUtil::JoinPath(dir, "Google"); -+ // The permission of ~/Library/Application Support/Google seems to be 0755. -+ // TODO(komatsu): nice to make a wrapper function. -+ ::mkdir(dir.c_str(), 0755); -+ dir = FileUtil::JoinPath(dir, "JapaneseInput"); - #else // GOOGLE_JAPANESE_INPUT_BUILD -- dir = FileUtil::JoinPath(dir, "Mozc"); -+ dir = FileUtil::JoinPath(dir, "Mozc"); - #endif // GOOGLE_JAPANESE_INPUT_BUILD -+ } - - #elif defined(OS_ANDROID) - // For android, we do nothing here because user profile directory, -@@ -237,14 +248,24 @@ - // is injected from Java layer. - - #else // !OS_WIN && !OS_MACOSX && !OS_ANDROID -- char buf[1024]; -- struct passwd pw, *ppw; -- const uid_t uid = geteuid(); -- CHECK_EQ(0, getpwuid_r(uid, &pw, buf, sizeof(buf), &ppw)) -- << "Can't get passwd entry for uid " << uid << "."; -- CHECK_LT(0, strlen(pw.pw_dir)) -- << "Home directory for uid " << uid << " is not set."; -- dir = FileUtil::JoinPath(pw.pw_dir, ".mozc"); -+ configuration_directory_env = ::getenv("MOZC_CONFIGURATION_DIRECTORY"); -+ if (configuration_directory_env) { -+ dir = configuration_directory_env; -+ } else { -+ const char *home_env = ::getenv("HOME"); -+ if (home_env) { -+ dir = FileUtil::JoinPath(home_env, ".mozc"); -+ } else { -+ char buf[1024]; -+ struct passwd pw, *ppw; -+ const uid_t uid = geteuid(); -+ CHECK_EQ(0, getpwuid_r(uid, &pw, buf, sizeof(buf), &ppw)) -+ << "Can't get passwd entry for uid " << uid << "."; -+ CHECK_LT(0, strlen(pw.pw_dir)) -+ << "Home directory for uid " << uid << " is not set."; -+ dir = FileUtil::JoinPath(pw.pw_dir, ".mozc"); -+ } -+ } - #endif // !OS_WIN && !OS_MACOSX && !OS_ANDROID - - FileUtil::CreateDirectory(dir); -@@ -356,6 +377,10 @@ - #endif // OS_WIN - - string SystemUtil::GetServerDirectory() { -+ const char *server_directory_env = ::getenv("MOZC_SERVER_DIRECTORY"); -+ if (server_directory_env) { -+ return server_directory_env; -+ } - #ifdef OS_WIN - DCHECK(SUCCEEDED(Singleton<ProgramFilesX86Cache>::get()->result())); - #if defined(GOOGLE_JAPANESE_INPUT_BUILD) -@@ -409,6 +434,10 @@ - } - - string SystemUtil::GetDocumentDirectory() { -+ const char *documents_directory_env = ::getenv("MOZC_DOCUMENTS_DIRECTORY"); -+ if (documents_directory_env) { -+ return documents_directory_env; -+ } - #if defined(OS_MACOSX) - return GetServerDirectory(); - #elif defined(MOZC_DOCUMENT_DIRECTORY) ---- /src/handwriting/zinnia_handwriting.cc -+++ /src/handwriting/zinnia_handwriting.cc -@@ -31,6 +31,7 @@ - - #include "handwriting/zinnia_handwriting.h" - -+#include <cstdlib> - #include <memory> - #include <string> - -@@ -48,6 +49,10 @@ - - // static - string ZinniaHandwriting::GetModelFileName() { -+ const char *zinnia_model_file_env = ::getenv("MOZC_ZINNIA_MODEL_FILE"); -+ if (zinnia_model_file_env) { -+ return zinnia_model_file_env; -+ } - #if defined(MOZC_BUILD) - return MOZC_ZINNIA_MODEL_FILE; - #else diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-gcc-8.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-gcc-8.patch deleted file mode 100644 index 07514048e7a4..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-gcc-8.patch +++ /dev/null @@ -1,22 +0,0 @@ -https://github.com/google/mozc/issues/441 - ---- /src/prediction/zero_query_dict.h -+++ /src/prediction/zero_query_dict.h -@@ -147,6 +147,17 @@ - return iter; - } - -+ iterator &operator--() { -+ ptr_ -= kTokenByteSize; -+ return *this; -+ } -+ -+ iterator operator--(int) { -+ const iterator tmp(ptr_, string_array_); -+ ptr_ -= kTokenByteSize; -+ return tmp; -+ } -+ - iterator &operator-=(ptrdiff_t n) { - ptr_ -= n * kTokenByteSize; - return *this; diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-protobuf-3.18.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-protobuf-3.18.patch deleted file mode 100644 index 15077bb1752a..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-protobuf-3.18.patch +++ /dev/null @@ -1,13 +0,0 @@ -Fix building with Protocol Buffers >=3.18. - ---- /src/dictionary/user_dictionary_storage.cc -+++ /src/dictionary/user_dictionary_storage.cc -@@ -108,7 +108,7 @@ - // wants to use more than 512MB. - mozc::protobuf::io::IstreamInputStream zero_copy_input(&ifs); - mozc::protobuf::io::CodedInputStream decoder(&zero_copy_input); -- decoder.SetTotalBytesLimit(kDefaultTotalBytesLimit, -1); -+ decoder.SetTotalBytesLimit(kDefaultTotalBytesLimit); - if (!user_dictionary_storage_base.ParseFromCodedStream(&decoder)) { - LOG(ERROR) << "Failed to parse"; - if (!decoder.ConsumedEntireMessage() || !ifs.eof()) { diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch deleted file mode 100644 index 2b9bbd720cde..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch +++ /dev/null @@ -1,621 +0,0 @@ -https://github.com/google/mozc/issues/462 - ---- /src/base/gen_character_set.py -+++ /src/base/gen_character_set.py -@@ -33,7 +33,6 @@ - import itertools - import optparse - import re --import string - import sys - - -@@ -89,7 +88,8 @@ - @staticmethod - def _LoadTable(filename, column_index, pattern, validater): - result = set() -- for line in open(filename): -+ fh = open(filename) -+ for line in fh: - if line.startswith('#'): - # Skip a comment line. - continue -@@ -100,6 +100,7 @@ - ucs = int(match.group(1), 16) - if validater(ucs): - result.add(ucs) -+ fh.close() - - return result - -@@ -250,7 +251,7 @@ - # (at most) four code points. - bit_list = [] - for _, group in itertools.groupby(enumerate(category_list), -- lambda (codepoint, _): codepoint / 4): -+ lambda x: x[0] // 4): - # Fill bits from LSB to MSB for each group. - bits = 0 - for index, (_, category) in enumerate(group): -@@ -263,7 +264,7 @@ - - # Output the content. Each line would have (at most) 16 bytes. - for _, group in itertools.groupby(enumerate(bit_list), -- lambda (index, _): index / 16): -+ lambda x: x[0] // 16): - line = [' \"'] - for _, bits in group: - line.append('\\x%02X' % bits) -@@ -386,7 +387,7 @@ - # Bitmap lookup. - # TODO(hidehiko): the bitmap has two huge 0-bits ranges. Reduce them. - category_map = [ -- (bits, category) for category, bits in CATEGORY_BITMAP.iteritems()] -+ (bits, category) for category, bits in CATEGORY_BITMAP.items()] - category_map.sort() - - lines.extend([ -@@ -451,7 +452,7 @@ - options.jisx0213file) - category_list = [ - categorizer.GetCategory(codepoint) -- for codepoint in xrange(categorizer.MaxCodePoint() + 1)] -+ for codepoint in range(categorizer.MaxCodePoint() + 1)] - generated_character_set_header = GenerateCharacterSetHeader(category_list) - - # Write the result. ---- /src/base/gen_config_file_stream_data.py -+++ /src/base/gen_config_file_stream_data.py -@@ -58,7 +58,7 @@ - result = [] - result.append(' { "%s", "' % os.path.basename(path)) - with open(path, 'rb') as stream: -- result.extend(r'\x%02X' % ord(byte) for byte in stream.read()) -+ result.extend(r'\x%02X' % byte for byte in stream.read()) - result.append('", %d }' % os.path.getsize(path)) - - return ''.join(result) -@@ -93,8 +93,8 @@ - def main(): - (options, args) = ParseOptions() - if not options.output: -- print >>sys.stderr, ( -- 'usage: gen_config_file_stream_data.py --output=filepath input ...') -+ print('usage: gen_config_file_stream_data.py --output=filepath input ...', -+ file=sys.stderr) - sys.exit(2) - - with open(options.output, 'w') as output: ---- /src/build_mozc.py -+++ /src/build_mozc.py -@@ -943,7 +943,7 @@ - logging.info('running %s...', binary) - try: - test_function(binary, gtest_report_dir, options) -- except RunOrDieError, e: -+ except RunOrDieError as e: - logging.error(e) - failed_tests.append(binary) - else: -@@ -1082,7 +1082,7 @@ - # and '-c' and 'Release' are build options. - targets = [] - build_options = [] -- for i in xrange(len(args)): -+ for i in range(len(args)): - if args[i].startswith('-'): - # starting with build options - build_options = args[i:] -@@ -1190,14 +1190,14 @@ - - def ShowHelpAndExit(): - """Shows the help message.""" -- print 'Usage: build_mozc.py COMMAND [ARGS]' -- print 'Commands: ' -- print ' gyp Generate project files.' -- print ' build Build the specified target.' -- print ' runtests Build all tests and run them.' -- print ' clean Clean all the build files and directories.' -- print '' -- print 'See also the comment in the script for typical usage.' -+ print('Usage: build_mozc.py COMMAND [ARGS]') -+ print('Commands: ') -+ print(' gyp Generate project files.') -+ print(' build Build the specified target.') -+ print(' runtests Build all tests and run them.') -+ print(' clean Clean all the build files and directories.') -+ print('') -+ print('See also the comment in the script for typical usage.') - sys.exit(1) - - ---- /src/build_tools/android_util.py -+++ /src/build_tools/android_util.py -@@ -548,7 +548,7 @@ - (devices_result, _) = process.communicate() - used_ports = set(int(port) for port - in re.findall(r'emulator-(\d+)', devices_result)) -- return [port for port in xrange(5554, 5586, 2) if port not in used_ports] -+ return [port for port in range(5554, 5586, 2) if port not in used_ports] - - - def SetUpTestingSdkHomeDirectory(dest_android_sdk_home, -@@ -575,7 +575,7 @@ - 'create', 'avd', - '--force', - '--sdcard', '512M',] -- for key, value in options.iteritems(): -+ for key, value in options.items(): - args.extend([key, value]) - env = {'ANDROID_SDK_HOME': os.path.abspath(dest_android_sdk_home)} - logging.info('Creating AVD: %s', args) -@@ -615,7 +615,7 @@ - def main(): - for arg in sys.argv[1:]: - for item in sorted(GetApkProperties(arg).items()): -- print '%s: %s' % item -+ print('%s: %s' % item) - - - if __name__ == '__main__': ---- /src/build_tools/binary_size_checker.py -+++ /src/build_tools/binary_size_checker.py -@@ -70,12 +70,12 @@ - actual_size = os.stat(filename).st_size - expected_size = EXPECTED_MAXIMUM_SIZES[basename] - if actual_size < expected_size * 1024 * 1024: -- print 'Pass: %s (size: %d) is smaller than expected (%d MB)' % ( -- filename, actual_size, expected_size) -+ print('Pass: %s (size: %d) is smaller than expected (%d MB)' % ( -+ filename, actual_size, expected_size)) - return True - else: -- print 'WARNING: %s (size: %d) is larger than expected (%d MB)' % ( -- filename, actual_size, expected_size) -+ print('WARNING: %s (size: %d) is larger than expected (%d MB)' % ( -+ filename, actual_size, expected_size)) - return False - - ---- /src/build_tools/build_and_sign_pkg_mac.py -+++ /src/build_tools/build_and_sign_pkg_mac.py -@@ -44,8 +44,8 @@ - import shutil - import sys - --from util import PrintErrorAndExit --from util import RunOrDie -+from .util import PrintErrorAndExit -+from .util import RunOrDie - - - def ParseOption(): ---- /src/build_tools/build_breakpad.py -+++ /src/build_tools/build_breakpad.py -@@ -54,9 +54,9 @@ - try: - subprocess.check_output(command) - except subprocess.CalledProcessError as e: -- print e.output -+ print(e.output) - sys.exit(e.returncode) -- print 'Done: %s' % ' '.join(command) -+ print('Done: %s' % ' '.join(command)) - - - def Xcodebuild(projdir, target, arch, sdk, outdir): ---- /src/build_tools/build_diskimage_mac.py -+++ /src/build_tools/build_diskimage_mac.py -@@ -90,7 +90,7 @@ - # setup volume directory - temp_dir = tempfile.mkdtemp() - CopyFile(path.join(build_dir, ".keystone_install"), temp_dir) -- os.chmod(path.join(temp_dir, ".keystone_install"), 0755) # rwxr-xr-x -+ os.chmod(path.join(temp_dir, ".keystone_install"), 0o755) # rwxr-xr-x - for a in args: - CopyFile(path.join(build_dir, a), temp_dir) - ---- /src/build_tools/change_reference_mac.py -+++ /src/build_tools/change_reference_mac.py -@@ -41,8 +41,8 @@ - import optparse - import os - --from util import PrintErrorAndExit --from util import RunOrDie -+from .util import PrintErrorAndExit -+from .util import RunOrDie - - - def ParseOption(): ---- /src/build_tools/code_generator_util.py -+++ /src/build_tools/code_generator_util.py -@@ -33,27 +33,26 @@ - __author__ = "hidehiko" - - import struct --import types - - - def ToCppStringLiteral(s): - """Returns C-style string literal, or NULL if given s is None.""" - if s is None: -- return 'NULL' -+ return b'NULL' - -- if all(0x20 <= ord(c) <= 0x7E for c in s): -+ if all(0x20 <= c <= 0x7E for c in s): - # All characters are in ascii code. -- return '"%s"' % s.replace('\\', r'\\').replace('"', r'\"') -+ return b'"%b"' % s.replace(b'\\', br'\\').replace(b'"', br'\"') - else: - # One or more characters are non-ascii. -- return '"%s"' % ''.join(r'\x%02X' % ord(c) for c in s) -+ return b'"%b"' % b''.join(br'\x%02X' % c for c in s) - - - def FormatWithCppEscape(format_text, *args): - """Returns a string filling format with args.""" - literal_list = [] - for arg in args: -- if isinstance(arg, (types.StringType, types.NoneType)): -+ if isinstance(arg, (bytes, type(None))): - arg = ToCppStringLiteral(arg) - literal_list.append(arg) - -@@ -95,7 +94,7 @@ - if target_compiler and target_compiler.startswith('msvs'): - stream.write('const uint64 k%s_data_wordtype[] = {\n' % variable_name) - -- for word_index in xrange(0, len(data), 8): -+ for word_index in range(0, len(data), 8): - word_chunk = data[word_index:word_index + 8].ljust(8, '\x00') - stream.write('0x%016X, ' % struct.unpack('<Q', word_chunk)) - if (word_index / 8) % 4 == 3: -@@ -111,7 +110,7 @@ - stream.write('const char k%s_data[] =\n' % variable_name) - # Output 16bytes per line. - chunk_size = 16 -- for index in xrange(0, len(data), chunk_size): -+ for index in range(0, len(data), chunk_size): - chunk = data[index:index + chunk_size] - stream.write('"') - stream.writelines(r'\x%02X' % ord(c) for c in chunk) -@@ -126,36 +125,50 @@ - if type(codepoint_list) is int: - codepoint_list = (codepoint_list,) - if codepoint_list is None or len(codepoint_list) == 0: -- return 'null' -- result = r'"' -+ return b'null' -+ result = b'"' - for codepoint in codepoint_list: -- utf16_string = unichr(codepoint).encode('utf-16be') -+ utf16_string = chr(codepoint).encode('utf-16be') - if len(utf16_string) == 2: - (u0, l0) = utf16_string -- result += r'\u%02X%02X' % (ord(u0), ord(l0)) -+ result += br'\u%02X%02X' % (u0, l0) - else: - (u0, l0, u1, l1) = utf16_string -- result += r'\u%02X%02X\u%02X%02X' % (ord(u0), ord(l0), ord(u1), ord(l1)) -- result += r'"' -+ result += br'\u%02X%02X\u%02X%02X' % (u0, l0, u1, l1) -+ result += b'"' - return result - - - def SkipLineComment(stream, comment_prefix='#'): - """Skips line comments from stream.""" - for line in stream: -+ if isinstance(line, bytes): -+ if isinstance(comment_prefix, str): -+ comment_prefix = comment_prefix.encode('utf-8') -+ line_ending = b'\n' -+ else: -+ line_ending = '\n' - stripped_line = line.strip() - if stripped_line and not stripped_line.startswith(comment_prefix): -- yield line.rstrip('\n') -+ yield line.rstrip(line_ending) - - - def ParseColumnStream(stream, num_column=None, delimiter=None): - """Returns parsed columns read from stream.""" - if num_column is None: - for line in stream: -- yield line.rstrip('\n').split(delimiter) -+ if isinstance(line, bytes): -+ line_ending = b'\n' -+ else: -+ line_ending = '\n' -+ yield line.rstrip(line_ending).split(delimiter) - else: - for line in stream: -- yield line.rstrip('\n').split(delimiter)[:num_column] -+ if isinstance(line, bytes): -+ line_ending = b'\n' -+ else: -+ line_ending = '\n' -+ yield line.rstrip(line_ending).split(delimiter)[:num_column] - - - def SelectColumn(stream, column_index): -@@ -172,5 +185,5 @@ - grouper extends the last chunk to make it an n-element chunk by adding - appropriate value, but this returns truncated chunk. - """ -- for index in xrange(0, len(iterable), n): -+ for index in range(0, len(iterable), n): - yield iterable[index:index + n] ---- /src/build_tools/codesign_mac.py -+++ /src/build_tools/codesign_mac.py -@@ -46,17 +46,17 @@ - - def RunOrDie(command): - """Run the command, or die if it failed.""" -- print "Running: " + command -+ print("Running: " + command) - try: - output = subprocess.check_output(command, shell=True) -- print >> sys.stderr, "==========" -- print >> sys.stderr, "COMMAND: " + command -- print >> sys.stderr, output -+ print("==========", file=sys.stderr) -+ print("COMMAND: " + command, file=sys.stderr) -+ print(output, file=sys.stderr) - except subprocess.CalledProcessError as e: -- print >> sys.stderr, "==========" -- print >> sys.stderr, "ERROR: " + command -- print >> sys.stderr, e.output -- print >> sys.stderr, "==========" -+ print("==========", file=sys.stderr) -+ print("ERROR: " + command, file=sys.stderr) -+ print(e.output, file=sys.stderr) -+ print("==========", file=sys.stderr) - sys.exit(1) - - -@@ -119,18 +119,18 @@ - (options, unused_args) = parser.parse_args() - - if not options.target: -- print "Error: --target should be specified." -- print parser.print_help() -+ print("Error: --target should be specified.") -+ print(parser.print_help()) - sys.exit(1) - - return options - - - def DumpEnviron(): -- print "=== os.environ ===" -+ print("=== os.environ ===") - for key in sorted(os.environ): -- print "%s = %s" % (key, os.getenv(key)) -- print "==================" -+ print("%s = %s" % (key, os.getenv(key))) -+ print("==================") - - - def main(): ---- /src/build_tools/copy_dll_and_symbol.py -+++ /src/build_tools/copy_dll_and_symbol.py -@@ -38,7 +38,7 @@ - import os - import shutil - --from util import PrintErrorAndExit -+from .util import PrintErrorAndExit - - def ParseOption(): - """Parse command line options.""" -@@ -98,7 +98,7 @@ - if _GetLastModifiedTime(src) <= target_file_mtime: - # Older file found. Ignore. - continue -- print 'Copying %s to %s' % (src, target_file_abspath) -+ print('Copying %s to %s' % (src, target_file_abspath)) - shutil.copy2(src, target_file_abspath) - break - ---- /src/build_tools/copy_file.py -+++ /src/build_tools/copy_file.py -@@ -52,7 +52,7 @@ - Args: - message: The error message to be printed to stderr. - """ -- print >>sys.stderr, message -+ print(message, file=sys.stderr) - sys.exit(1) - - ---- /src/build_tools/copy_qt_frameworks_mac.py -+++ /src/build_tools/copy_qt_frameworks_mac.py -@@ -41,9 +41,9 @@ - import optparse - import os - --from copy_file import CopyFiles --from util import PrintErrorAndExit --from util import RunOrDie -+from .copy_file import CopyFiles -+from .util import PrintErrorAndExit -+from .util import RunOrDie - - - def ParseOption(): ---- /src/build_tools/embed_file.py -+++ /src/build_tools/embed_file.py -@@ -46,10 +46,10 @@ - - def _FormatAsUint64LittleEndian(s): - """Formats a string as uint64 value in little endian order.""" -- for _ in xrange(len(s), 8): -- s += '\0' -+ for _ in range(len(s), 8): -+ s += b'\0' - s = s[::-1] # Reverse the string -- return '0x%s' % binascii.b2a_hex(s) -+ return b'0x%b' % binascii.b2a_hex(s) - - - def main(): -@@ -57,30 +57,30 @@ - with open(opts.input, 'rb') as infile: - with open(opts.output, 'wb') as outfile: - outfile.write( -- '#ifdef MOZC_EMBEDDED_FILE_%(name)s\n' -- '#error "%(name)s was already included or defined elsewhere"\n' -- '#else\n' -- '#define MOZC_EMBEDDED_FILE_%(name)s\n' -- 'const uint64 %(name)s_data[] = {\n' -- % {'name': opts.name}) -+ b'#ifdef MOZC_EMBEDDED_FILE_%(name)b\n' -+ b'#error "%(name)b was already included or defined elsewhere"\n' -+ b'#else\n' -+ b'#define MOZC_EMBEDDED_FILE_%(name)b\n' -+ b'const uint64 %(name)b_data[] = {\n' -+ % {b'name': opts.name.encode('utf-8')}) - - while True: - chunk = infile.read(8) - if not chunk: - break -- outfile.write(' ') -+ outfile.write(b' ') - outfile.write(_FormatAsUint64LittleEndian(chunk)) -- outfile.write(',\n') -+ outfile.write(b',\n') - - outfile.write( -- '};\n' -- 'const EmbeddedFile %(name)s = {\n' -- ' %(name)s_data,\n' -- ' %(size)d,\n' -- '};\n' -- '#endif // MOZC_EMBEDDED_FILE_%(name)s\n' -- % {'name': opts.name, -- 'size': os.stat(opts.input).st_size}) -+ b'};\n' -+ b'const EmbeddedFile %(name)b = {\n' -+ b' %(name)b_data,\n' -+ b' %(size)d,\n' -+ b'};\n' -+ b'#endif // MOZC_EMBEDDED_FILE_%(name)b\n' -+ % {b'name': opts.name.encode('utf-8'), -+ b'size': os.stat(opts.input).st_size}) - - - if __name__ == '__main__': ---- /src/build_tools/embed_pathname.py -+++ /src/build_tools/embed_pathname.py -@@ -28,7 +28,7 @@ - # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - --"""A script to embed the given (relative) path name to C/C++ characters array. -+r"""A script to embed the given (relative) path name to C/C++ characters array. - - Example: - ./embed_pathname.py --path_to_be_embedded=d:\data\mozc -@@ -53,7 +53,7 @@ - - (options, unused_args) = parser.parse_args() - if not all(vars(options).values()): -- print parser.print_help() -+ print(parser.print_help()) - sys.exit(1) - - return options -@@ -63,7 +63,7 @@ - opt = ParseOption() - path = os.path.abspath(opt.path_to_be_embedded) - # TODO(yukawa): Consider the case of non-ASCII characters. -- escaped_path = path.encode('string-escape') -+ escaped_path = path.replace('\\', '\\\\') - with open(opt.output, 'w') as output_file: - output_file.write( - 'const char %s[] = "%s";\n' % (opt.constant_name, escaped_path)) ---- /src/build_tools/ensure_gyp_module_path.py -+++ /src/build_tools/ensure_gyp_module_path.py -@@ -48,7 +48,7 @@ - - (options, _) = parser.parse_args() - if not options.expected: -- print parser.print_help() -+ print(parser.print_help()) - sys.exit(1) - - return options -@@ -59,20 +59,20 @@ - opt = ParseOption() - expected_path = os.path.abspath(opt.expected) - if not os.path.exists(expected_path): -- print '%s does not exist.' % expected_path -+ print('%s does not exist.' % expected_path) - sys.exit(1) - - try: - import gyp # NOLINT - except ImportError as e: -- print 'import gyp failed: %s' % e -+ print('import gyp failed: %s' % e) - sys.exit(1) - - actual_path = os.path.abspath(gyp.__path__[0]) - if expected_path != actual_path: -- print 'Unexpected gyp module is loaded on this environment.' -- print ' expected: %s' % expected_path -- print ' actual : %s' % actual_path -+ print('Unexpected gyp module is loaded on this environment.') -+ print(' expected: %s' % expected_path) -+ print(' actual : %s' % actual_path) - sys.exit(1) - - if __name__ == '__main__': ---- /src/build_tools/gen_win32_resource_header.py -+++ /src/build_tools/gen_win32_resource_header.py -@@ -39,7 +39,7 @@ - __author__ = "yukawa" - - import logging --import mozc_version -+from . import mozc_version - import optparse - import os - import sys ---- /src/build_tools/mozc_version.py -+++ /src/build_tools/mozc_version.py -@@ -94,7 +94,7 @@ - last_digit = TARGET_PLATFORM_TO_DIGIT.get(target_platform, None) - if last_digit is None: - logging.critical('target_platform %s is invalid. Accetable ones are %s', -- target_platform, TARGET_PLATFORM_TO_DIGIT.keys()) -+ target_platform, list(TARGET_PLATFORM_TO_DIGIT.keys())) - sys.exit(1) - - if not revision: -@@ -314,13 +314,14 @@ - self._properties = {} - if not os.path.isfile(path): - return -- for line in open(path): -- matchobj = re.match(r'(\w+)=(.*)', line.strip()) -- if matchobj: -- var = matchobj.group(1) -- val = matchobj.group(2) -- if var not in self._properties: -- self._properties[var] = val -+ with open(path) as file: -+ for line in file: -+ matchobj = re.match(r'(\w+)=(.*)', line.strip()) -+ if matchobj: -+ var = matchobj.group(1) -+ val = matchobj.group(2) -+ if var not in self._properties: -+ self._properties[var] = val - - # Check mandatory properties. - for key in VERSION_PROPERTIES: diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch deleted file mode 100644 index 456e8368049a..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch +++ /dev/null @@ -1,600 +0,0 @@ -https://github.com/google/mozc/issues/462 - ---- /src/build_tools/redirect.py -+++ /src/build_tools/redirect.py -@@ -58,14 +58,15 @@ - process = subprocess.Popen(sys.argv, stdout=subprocess.PIPE, - universal_newlines=True) - except: -- print '==========' -- print ' ERROR: %s' % ' '.join(sys.argv) -- print '==========' -+ print('==========') -+ print(' ERROR: %s' % ' '.join(sys.argv)) -+ print('==========') - raise - (stdout_content, _) = process.communicate() - # Write the stdout content to the output file. - output_file = open(output_file_name, 'w') - output_file.write(stdout_content) -+ output_file.close() - return process.wait() - - if __name__ == '__main__': ---- /src/build_tools/run_after_chdir.py -+++ /src/build_tools/run_after_chdir.py -@@ -57,7 +57,7 @@ - sys.argv.insert(0, sys.executable) # Inject the python interpreter path. - # We don't capture stdout and stderr from Popen. The output will just - # be emitted to a terminal or console. -- print sys.argv -+ print(sys.argv) - sys.exit(subprocess.call(sys.argv)) - - if __name__ == '__main__': ---- /src/build_tools/serialized_string_array_builder.py -+++ /src/build_tools/serialized_string_array_builder.py -@@ -58,11 +58,11 @@ - f.write(struct.pack('<I', array_size)) - - # Offset and length array of (4 + 4) * array_size bytes. -- for i in xrange(array_size): -+ for i in range(array_size): - f.write(struct.pack('<I', offsets[i])) - f.write(struct.pack('<I', lengths[i])) - - # Strings chunk. -- for i in xrange(array_size): -+ for i in range(array_size): - f.write(strings[i]) -- f.write('\0') -+ f.write(b'\0') ---- /src/build_tools/test_tools/gtest_report.py -+++ /src/build_tools/test_tools/gtest_report.py -@@ -36,9 +36,9 @@ - - __author__ = "nona" - --import cStringIO as StringIO -+import io - import logging --from xml.etree import cElementTree as ElementTree -+from xml.etree import ElementTree - - - class Failure(object): -@@ -87,13 +87,13 @@ - """Returns summarized error report text.""" - if self.fail_num == 0: - return '' -- output = StringIO.StringIO() -+ output = io.StringIO() - for testcase in self.testcases: - if not testcase.failures: - continue -- print >>output, '%s.%s:' % (self.name, testcase.name) -+ print('%s.%s:' % (self.name, testcase.name), file=output) - for failure in testcase.failures: -- print >>output, failure.contents.encode('utf-8') -+ print(failure.contents.encode('utf-8'), file=output) - return output.getvalue() - - @classmethod ---- /src/build_tools/test_tools/test_launcher.py -+++ /src/build_tools/test_tools/test_launcher.py -@@ -101,11 +101,11 @@ - time.sleep(1) - try: - shutil.rmtree(self._path) -- except OSError, e: -+ except OSError as e: - logging.error('Failed to remove %s. error: %s', self._path, e) - - --def _ExecuteTest((command, gtest_report_dir)): -+def _ExecuteTest(args): - """Executes tests with specified Test command. - - Args: -@@ -122,6 +122,7 @@ - module, which is used in multiprocessing module. - (http://docs.python.org/library/pickle.html) - """ -+ (command, gtest_report_dir) = args - binary = command[0] - binary_filename = os.path.basename(binary) - tmp_dir = tempfile.mkdtemp() ---- /src/build_tools/tweak_data.py -+++ /src/build_tools/tweak_data.py -@@ -55,7 +55,7 @@ - The value for the variable if the variable is defined in the - environment. Otherwise original string is returned. - """ -- if environment.has_key(matchobj.group(1)): -+ if matchobj.group(1) in environment: - return environment[matchobj.group(1)] - return matchobj.group(0) - ---- /src/build_tools/tweak_info_plist.py -+++ /src/build_tools/tweak_info_plist.py -@@ -42,8 +42,8 @@ - import logging - import optparse - import sys --import mozc_version --import tweak_data -+from . import mozc_version -+from . import tweak_data - - _COPYRIGHT_YEAR = datetime.date.today().year - -@@ -81,7 +81,7 @@ - - version = mozc_version.MozcVersion(options.version_file) - -- copyright_message = (u'© %d Google Inc.' % _COPYRIGHT_YEAR).encode('utf-8') -+ copyright_message = ('© %d Google Inc.' % _COPYRIGHT_YEAR).encode('utf-8') - long_version = version.GetVersionString() - short_version = version.GetVersionInFormat('@MAJOR@.@MINOR@.@BUILD@') - ---- /src/build_tools/tweak_info_plist_strings.py -+++ /src/build_tools/tweak_info_plist_strings.py -@@ -40,7 +40,7 @@ - import logging - import optparse - import sys --import tweak_data -+from . import tweak_data - - _COPYRIGHT_YEAR = datetime.date.today().year - -@@ -77,7 +77,7 @@ - if options.branding == 'GoogleJapaneseInput': - variables = { - 'CF_BUNDLE_NAME_EN': 'Google Japanese Input', -- 'CF_BUNDLE_NAME_JA': u'Google 日本語入力'.encode('utf-8'), -+ 'CF_BUNDLE_NAME_JA': 'Google 日本語入力'.encode('utf-8'), - 'NS_HUMAN_READABLE_COPYRIGHT': copyright_message, - 'INPUT_MODE_ANNOTATION': 'Google', - } ---- /src/build_tools/tweak_macinstaller_script.py -+++ /src/build_tools/tweak_macinstaller_script.py -@@ -39,7 +39,7 @@ - import logging - import optparse - --import mozc_version -+from . import mozc_version - - - def _ReplaceVariables(data, environment): ---- /src/build_tools/tweak_pkgproj.py -+++ /src/build_tools/tweak_pkgproj.py -@@ -45,7 +45,7 @@ - import os - import plistlib - import re --import mozc_version -+from . import mozc_version - - from os import path - -@@ -71,7 +71,7 @@ - The value for the variable if the variable is defined in the - environment. Otherwise original string is returned. - """ -- if environment.has_key(matchobj.group(1)): -+ if matchobj.group(1) in environment: - return environment[matchobj.group(1)] - return matchobj.group(0) - ---- /src/build_tools/util.py -+++ /src/build_tools/util.py -@@ -73,11 +73,11 @@ - return 1 - - --class RunOrDieError(StandardError): -+class RunOrDieError(Exception): - """The exception class for RunOrDie.""" - - def __init__(self, message): -- StandardError.__init__(self, message) -+ Exception.__init__(self, message) - - - def RunOrDie(argv): -@@ -105,7 +105,7 @@ - return # Do nothing if not exist. - if IsWindows(): - # Read-only files cannot be deleted on Windows. -- os.chmod(file_name, 0700) -+ os.chmod(file_name, 0o700) - logging.debug('Removing file: %s', file_name) - os.unlink(file_name) - ---- /src/build_tools/zlib_util.py -+++ /src/build_tools/zlib_util.py -@@ -58,7 +58,7 @@ - - def main(): - if len(sys.argv) != 4: -- print >>sys.stderr, 'Invalid arguments' -+ print('Invalid arguments', file=sys.stderr) - return - if sys.argv[1] == 'compress': - Compress(sys.argv[2], sys.argv[3]) -@@ -66,7 +66,7 @@ - if sys.argv[1] == 'decompress': - Decompress(sys.argv[2], sys.argv[3]) - return -- print >>sys.stderr, 'Unknown command:', sys.argv[1] -+ print('Unknown command:', sys.argv[1], file=sys.stderr) - - - if __name__ == '__main__': ---- /src/composer/internal/gen_typing_model.py -+++ /src/composer/internal/gen_typing_model.py -@@ -54,14 +54,13 @@ - __author__ = "noriyukit" - - import bisect --import codecs - import collections - import optparse - import struct - - UNDEFINED_COST = -1 --MAX_UINT16 = struct.unpack('H', '\xFF\xFF')[0] --MAX_UINT8 = struct.unpack('B', '\xFF')[0] -+MAX_UINT16 = struct.unpack('H', b'\xFF\xFF')[0] -+MAX_UINT8 = struct.unpack('B', b'\xFF')[0] - - - def ParseArgs(): -@@ -113,7 +112,7 @@ - sorted_values = list(sorted(set(values))) - mapping_table = sorted_values[0] - mapping_table_size_without_special_value = mapping_table_size - 1 -- span = len(sorted_values) / (mapping_table_size_without_special_value - 1) -+ span = len(sorted_values) // (mapping_table_size_without_special_value - 1) - mapping_table = [sorted_values[i * span] - for i - in range(0, mapping_table_size_without_special_value - 1)] -@@ -150,7 +149,7 @@ - - def GetValueTable(unique_characters, mapping_table, dictionary): - result = [] -- for key, value in dictionary.iteritems(): -+ for key, value in dictionary.items(): - index = GetIndexFromKey(unique_characters, key) - while len(result) <= index: - result.append(len(mapping_table) - 1) -@@ -167,13 +166,13 @@ - romaji_transition_cost) - with open(output_path, 'wb') as f: - f.write(struct.pack('<I', len(unique_characters))) -- f.write(''.join(unique_characters)) -+ f.write(''.join(unique_characters).encode('utf-8')) - offset = 4 + len(unique_characters) - - # Add padding to place value list size at 4-byte boundary. - if offset % 4: - padding_size = 4 - offset % 4 -- f.write('\x00' * padding_size) -+ f.write(b'\x00' * padding_size) - offset += padding_size - - f.write(struct.pack('<I', len(value_list))) -@@ -184,7 +183,7 @@ - # Add padding to place mapping_table at 4-byte boundary. - if offset % 4: - padding_size = 4 - offset % 4 -- f.write('\x00' * padding_size) -+ f.write(b'\x00' * padding_size) - offset += padding_size - - for v in mapping_table: -@@ -198,7 +197,8 @@ - # - trigram['vw']['x'] = -500 * log(P(x | 'vw')) - unigram = {} - trigram = collections.defaultdict(dict) -- for line in codecs.open(options.input_path, 'r', encoding='utf-8'): -+ input_file = open(options.input_path, 'r', encoding='utf-8') -+ for line in input_file: - line = line.rstrip() - ngram, cost = line.split('\t') - cost = int(cost) -@@ -206,6 +206,7 @@ - unigram[ngram] = cost - else: - trigram[ngram[:-1]][ngram[-1]] = cost -+ input_file.close() - - # Calculate ngram-related cost for each 'vw' and 'x': - # -500 * log( P('x' | 'vw') / P('x') ) ---- /src/converter/gen_boundary_data.py -+++ /src/converter/gen_boundary_data.py -@@ -70,7 +70,8 @@ - def LoadPatterns(file): - prefix = [] - suffix = [] -- for line in open(file, 'r'): -+ fh = open(file, 'r') -+ for line in fh: - if len(line) <= 1 or line[0] == '#': - continue - fields = line.split() -@@ -84,8 +85,9 @@ - elif label == 'SUFFIX': - suffix.append([re.compile(PatternToRegexp(feature)), cost]) - else: -- print 'format error %s' % (line) -+ print('format error %s' % (line)) - sys.exit(0) -+ fh.close() - return (prefix, suffix) - - -@@ -100,19 +102,23 @@ - - def LoadFeatures(filename): - features = [] -- for line in open(filename, 'r'): -+ fh = open(filename, 'r') -+ for line in fh: - fields = line.split() - features.append(fields[1]) -+ fh.close() - return features - - - def CountSpecialPos(filename): - count = 0 -- for line in open(filename, 'r'): -+ fh = open(filename, 'r') -+ for line in fh: - line = line.rstrip() - if not line or line[0] == '#': - continue - count += 1 -+ fh.close() - return count - - -@@ -141,7 +147,7 @@ - f.write(struct.pack('<H', GetCost(prefix, feature))) - f.write(struct.pack('<H', GetCost(suffix, feature))) - -- for _ in xrange(num_special_pos): -+ for _ in range(num_special_pos): - f.write(struct.pack('<H', 0)) - f.write(struct.pack('<H', 0)) - ---- /src/converter/gen_quality_regression_test_data.py -+++ /src/converter/gen_quality_regression_test_data.py -@@ -84,7 +84,7 @@ - else _ENABLED) - id = issue.attributes['id'].value - target = GetText(issue.getElementsByTagName('target')) -- for detail in issue.getElementsByTagName(u'detail'): -+ for detail in issue.getElementsByTagName('detail'): - fields = [] - fields.append('mozcsu_%s' % id) - for key in ('reading', 'output', 'actionStatus', 'rank', 'accuracy'): -@@ -104,19 +104,19 @@ - - def GenerateHeader(files): - try: -- print 'namespace mozc{' -- print 'struct TestCase {' -- print ' const bool enabled;' -- print ' const char *tsv;' -- print '} kTestData[] = {' -+ print('namespace mozc{') -+ print('struct TestCase {') -+ print(' const bool enabled;') -+ print(' const char *tsv;') -+ print('} kTestData[] = {') - for file in files: - for enabled, line in ParseFile(file): -- print ' {%s, "%s"},' % (enabled, EscapeString(line)) -- print ' {false, nullptr},' -- print '};' -- print '} // namespace mozc' -+ print(' {%s, "%s"},' % (enabled, EscapeString(line))) -+ print(' {false, nullptr},') -+ print('};') -+ print('} // namespace mozc') - except: -- print 'cannot open %s' % (file) -+ print('cannot open %s' % (file)) - sys.exit(1) - - ---- /src/converter/gen_segmenter_code.py -+++ /src/converter/gen_segmenter_code.py -@@ -54,18 +54,22 @@ - pos = {} - max_id = 0 - -- for line in open(id_file, "r"): -+ fh = open(id_file, "r") -+ for line in fh: - fields = line.split() - pos[fields[1]] = fields[0] - max_id = max(int(fields[0]), max_id) -+ fh.close() - - max_id = max_id + 1 -- for line in open(special_pos_file, "r"): -+ fh = open(special_pos_file, "r") -+ for line in fh: - if len(line) <= 1 or line[0] == '#': - continue - fields = line.split() - pos[fields[0]] = ("%d" % max_id) - max_id = max_id + 1 -+ fh.close() - - return pos - -@@ -79,8 +83,7 @@ - pat = re.compile(PatternToRegexp(pattern)) - min = -1; - max = -1; -- keys = pos.keys() -- keys.sort() -+ keys = sorted(pos.keys()) - - range = [] - -@@ -107,7 +110,7 @@ - tmp.append("(%s >= %s && %s <= %s)" % (name, r[0], name, r[1])) - - if len(tmp) == 0: -- print "FATAL: No rule fiind %s" % (pattern) -+ print("FATAL: No rule fiind %s" % (pattern)) - sys.exit(-1) - - return " || ".join(tmp) -@@ -115,19 +118,21 @@ - def main(): - pos = ReadPOSID(sys.argv[1], sys.argv[2]) - -- print HEADER % (len(pos.keys()), len(pos.keys())) -+ print(HEADER % (len(pos.keys()), len(pos.keys()))) - -- for line in open(sys.argv[3], "r"): -+ fh = open(sys.argv[3], "r") -+ for line in fh: - if len(line) <= 1 or line[0] == '#': - continue - (l, r, result) = line.split() - result = result.lower() - lcond = GetRange(pos, l, "rid") or "true"; - rcond = GetRange(pos, r, "lid") or "true"; -- print " // %s %s %s" % (l, r, result) -- print " if ((%s) && (%s)) { return %s; }" % (lcond, rcond, result) -+ print(" // %s %s %s" % (l, r, result)) -+ print(" if ((%s) && (%s)) { return %s; }" % (lcond, rcond, result)) -+ fh.close() - -- print FOOTER -+ print(FOOTER) - - if __name__ == "__main__": - main() ---- /src/data_manager/gen_connection_data.py -+++ /src/data_manager/gen_connection_data.py -@@ -32,8 +32,7 @@ - - __author__ = "hidehiko" - --import cStringIO as StringIO --import itertools -+import io - import logging - import optparse - import os -@@ -45,7 +44,7 @@ - INVALID_COST = 30000 - INVALID_1BYTE_COST = 255 - RESOLUTION_FOR_1BYTE = 64 --FILE_MAGIC = '\xAB\xCD' -+FILE_MAGIC = b'\xAB\xCD' - - FALSE_VALUES = ['f', 'false', '0'] - TRUE_VALUES = ['t', 'true', '1'] -@@ -79,28 +78,28 @@ - # The result is a square matrix. - mat_size = pos_size + special_pos_size - -- matrix = [[0] * mat_size for _ in xrange(mat_size)] -+ matrix = [[0] * mat_size for _ in range(mat_size)] - with open(text_connection_file) as stream: - stream = code_generator_util.SkipLineComment(stream) - # The first line contains the matrix column/row size. -- size = stream.next().rstrip() -+ size = next(stream).rstrip() - assert (int(size) == pos_size), '%s != %d' % (size, pos_size) - - for array_index, cost in enumerate(stream): - cost = int(cost.rstrip()) -- rid = array_index / pos_size -+ rid = array_index // pos_size - lid = array_index % pos_size - if rid == 0 and lid == 0: - cost = 0 - matrix[rid][lid] = cost - - # Fill INVALID_COST in matrix elements for special POS. -- for rid in xrange(pos_size, mat_size): -- for lid in xrange(1, mat_size): # Skip EOS -+ for rid in range(pos_size, mat_size): -+ for lid in range(1, mat_size): # Skip EOS - matrix[rid][lid] = INVALID_COST - -- for lid in xrange(pos_size, mat_size): -- for rid in xrange(1, mat_size): # Skip BOS -+ for lid in range(pos_size, mat_size): -+ for rid in range(1, mat_size): # Skip BOS - matrix[rid][lid] = INVALID_COST - - return matrix -@@ -116,7 +115,7 @@ - # Heuristically, we do not compress INVALID_COST. - continue - m[cost] = m.get(cost, 0) + 1 -- mode_value = max(m.iteritems(), key=lambda (_, count): count)[0] -+ mode_value = max(m.items(), key=lambda x: x[1])[0] - result.append(mode_value) - return result - -@@ -126,8 +125,8 @@ - # list, and fill None into the matrix if it equals to the corresponding - # mode value. - assert len(matrix) == len(mode_value_list) -- for row, mode_value in itertools.izip(matrix, mode_value_list): -- for index in xrange(len(row)): -+ for row, mode_value in zip(matrix, mode_value_list): -+ for index in range(len(row)): - if row[index] == mode_value: - row[index] = None - -@@ -179,7 +178,7 @@ - resolution = RESOLUTION_FOR_1BYTE - else: - resolution = 1 -- stream = StringIO.StringIO() -+ stream = io.BytesIO() - - # Output header. - stream.write(FILE_MAGIC) -@@ -194,7 +193,7 @@ - - # 4 bytes alignment. - if len(mode_value_list) % 2: -- stream.write('\x00\x00') -+ stream.write(b'\x00\x00') - - # Process each row: - for row in matrix: -@@ -218,7 +217,7 @@ - if cost == INVALID_COST: - cost = INVALID_1BYTE_COST - else: -- cost /= resolution -+ cost //= resolution - assert cost != INVALID_1BYTE_COST - values.append(cost) - -@@ -237,7 +236,7 @@ - values_size = len(values) * 2 - - # Output the bits for a row. -- stream.write(struct.pack('<HH', len(compact_bits) / 8, values_size)) -+ stream.write(struct.pack('<HH', len(compact_bits) // 8, values_size)) - OutputBitList(chunk_bits, stream) - OutputBitList(compact_bits, stream) - if use_1byte_cost: diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch deleted file mode 100644 index a5c5a2dc8038..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch +++ /dev/null @@ -1,583 +0,0 @@ -https://github.com/google/mozc/issues/462 - ---- /src/dictionary/gen_pos_map.py -+++ /src/dictionary/gen_pos_map.py -@@ -39,7 +39,7 @@ - from build_tools import code_generator_util - - --HEADER = """// Copyright 2009 Google Inc. All Rights Reserved. -+HEADER = b"""// Copyright 2009 Google Inc. All Rights Reserved. - // Author: keni - - #ifndef MOZC_DICTIONARY_POS_MAP_H_ -@@ -48,13 +48,13 @@ - // POS conversion rules - const POSMap kPOSMap[] = { - """ --FOOTER = """}; -+FOOTER = b"""}; - - #endif // MOZC_DICTIONARY_POS_MAP_H_ - """ - - def ParseUserPos(user_pos_file): -- with open(user_pos_file, 'r') as stream: -+ with open(user_pos_file, 'rb') as stream: - stream = code_generator_util.SkipLineComment(stream) - stream = code_generator_util.ParseColumnStream(stream, num_column=2) - return dict((key, enum_value) for key, enum_value in stream) -@@ -64,7 +64,7 @@ - user_pos_map = ParseUserPos(user_pos_file) - - result = {} -- with open(third_party_pos_map_file, 'r') as stream: -+ with open(third_party_pos_map_file, 'rb') as stream: - stream = code_generator_util.SkipLineComment(stream) - for columns in code_generator_util.ParseColumnStream(stream, num_column=2): - third_party_pos_name, mozc_pos = (columns + [None])[:2] -@@ -78,7 +78,7 @@ - result[third_party_pos_name] = mozc_pos - - # Create mozc_pos to mozc_pos map. -- for key, value in user_pos_map.iteritems(): -+ for key, value in user_pos_map.items(): - if key in result: - assert (result[key] == value) - continue -@@ -94,10 +94,10 @@ - if value is None: - # Invalid PosType. - value = ( -- 'static_cast< ::mozc::user_dictionary::UserDictionary::PosType>(-1)') -+ b'static_cast< ::mozc::user_dictionary::UserDictionary::PosType>(-1)') - else: -- value = '::mozc::user_dictionary::UserDictionary::' + value -- output.write(' { %s, %s },\n' % (key, value)) -+ value = b'::mozc::user_dictionary::UserDictionary::' + value -+ output.write(b' { %s, %s },\n' % (key, value)) - output.write(FOOTER) - - -@@ -121,7 +121,7 @@ - pos_map = GeneratePosMap(options.third_party_pos_map_file, - options.user_pos_file) - -- with open(options.output, 'w') as stream: -+ with open(options.output, 'wb') as stream: - OutputPosMap(pos_map, stream) - - ---- /src/dictionary/gen_pos_rewrite_rule.py -+++ /src/dictionary/gen_pos_rewrite_rule.py -@@ -46,29 +46,34 @@ - - - def LoadRewriteMapRule(filename): -- fh = open(filename) -+ fh = open(filename, 'rb') - rule = [] - for line in fh: -- line = line.rstrip('\n') -- if not line or line.startswith('#'): -+ line = line.rstrip(b'\n') -+ if not line or line.startswith(b'#'): - continue - fields = line.split() - rule.append([fields[0], fields[1]]) -+ fh.close() - return rule - - - def ReadPOSID(id_file, special_pos_file): - pos_list = [] - -- for line in open(id_file, 'r'): -+ fh = open(id_file, 'rb') -+ for line in fh: - fields = line.split() - pos_list.append(fields[1]) -+ fh.close() - -- for line in open(special_pos_file, 'r'): -- if len(line) <= 1 or line[0] == '#': -+ fh = open(special_pos_file, 'rb') -+ for line in fh: -+ if len(line) <= 1 or line[0:1] == b'#': - continue - fields = line.split() - pos_list.append(fields[0]) -+ fh.close() - - return pos_list - -@@ -112,7 +117,7 @@ - ids.append(id) - - with open(opts.output, 'wb') as f: -- f.write(''.join(chr(id) for id in ids)) -+ f.write(''.join(chr(id) for id in ids).encode('utf-8')) - - - if __name__ == '__main__': ---- /src/dictionary/gen_suffix_data.py -+++ /src/dictionary/gen_suffix_data.py -@@ -52,10 +52,10 @@ - opts = _ParseOptions() - - result = [] -- with open(opts.input, 'r') as stream: -+ with open(opts.input, 'rb') as stream: - for line in stream: -- line = line.rstrip('\r\n') -- fields = line.split('\t') -+ line = line.rstrip(b'\r\n') -+ fields = line.split(b'\t') - key = fields[0] - lid = int(fields[1]) - rid = int(fields[2]) -@@ -63,7 +63,7 @@ - value = fields[4] - - if key == value: -- value = '' -+ value = b'' - - result.append((key, value, lid, rid, cost)) - ---- /src/dictionary/gen_user_pos_data.py -+++ /src/dictionary/gen_user_pos_data.py -@@ -64,7 +64,7 @@ - f.write(struct.pack('<H', conjugation_id)) - - serialized_string_array_builder.SerializeToFile( -- sorted(string_index.iterkeys()), output_string_array) -+ sorted(x.encode('utf-8') for x in string_index.keys()), output_string_array) - - - def ParseOptions(): -@@ -100,7 +100,7 @@ - - if options.output_pos_list: - serialized_string_array_builder.SerializeToFile( -- [pos for (pos, _) in user_pos.data], options.output_pos_list) -+ [pos.encode('utf-8') for (pos, _) in user_pos.data], options.output_pos_list) - - - if __name__ == '__main__': ---- /src/dictionary/gen_zip_code_seed.py -+++ /src/dictionary/gen_zip_code_seed.py -@@ -83,7 +83,7 @@ - address = unicodedata.normalize('NFKC', self.address) - line = '\t'.join([zip_code, '0', '0', str(ZIP_CODE_COST), - address, ZIP_CODE_LABEL]) -- print line.encode('utf-8') -+ print(line.encode('utf-8')) - - - def ProcessZipCodeCSV(file_name): -@@ -105,26 +105,26 @@ - - def ReadZipCodeEntries(zip_code, level1, level2, level3): - """Read zip code entries.""" -- return [ZipEntry(zip_code, u''.join([level1, level2, town])) -+ return [ZipEntry(zip_code, ''.join([level1, level2, town])) - for town in ParseTownName(level3)] - - - def ReadJigyosyoEntry(zip_code, level1, level2, level3, name): - """Read jigyosyo entry.""" - return ZipEntry(zip_code, -- u''.join([level1, level2, level3, u' ', name])) -+ ''.join([level1, level2, level3, ' ', name])) - - - def ParseTownName(level3): - """Parse town name.""" -- if level3.find(u'以下に掲載がない場合') != -1: -+ if level3.find('以下に掲載がない場合') != -1: - return [''] - - assert CanParseAddress(level3), ('failed to be merged %s' - % level3.encode('utf-8')) - - # We ignore additional information here. -- level3 = re.sub(u'(.*)', u'', level3, re.U) -+ level3 = re.sub('(.*)', '', level3, re.U) - - # For 地割, we have these cases. - # XX1地割 -@@ -134,7 +134,7 @@ - # XX第1地割、XX第2地割、 - # XX第1地割〜XX第2地割、 - # We simply use XX for them. -- chiwari_match = re.match(u'(\D*?)第?\d+地割.*', level3, re.U) -+ chiwari_match = re.match('(\D*?)第?\d+地割.*', level3, re.U) - if chiwari_match: - town = chiwari_match.group(1) - return [town] -@@ -144,21 +144,21 @@ - # -> XX町YY and (XX町)ZZ - # YY、ZZ - # -> YY and ZZ -- chou_match = re.match(u'(.*町)?(.*)', level3, re.U) -+ chou_match = re.match('(.*町)?(.*)', level3, re.U) - if chou_match: -- chou = u'' -+ chou = '' - if chou_match.group(1): - chou = chou_match.group(1) - rests = chou_match.group(2) -- return [chou + rest for rest in rests.split(u'、')] -+ return [chou + rest for rest in rests.split('、')] - - return [level3] - - - def CanParseAddress(address): - """Return true for valid address.""" -- return (address.find(u'(') == -1 or -- address.find(u')') != -1) -+ return (address.find('(') == -1 or -+ address.find(')') != -1) - - - def ParseOptions(): ---- /src/dictionary/zip_code_util.py -+++ /src/dictionary/zip_code_util.py -@@ -86,11 +86,11 @@ - - - _SPECIAL_CASES = [ -- SpecialMergeZip(u'5900111', u'大阪府', u'堺市中区', [u'三原台']), -- SpecialMergeZip(u'8710046', u'大分県', u'中津市', -- [u'金谷', u'西堀端', u'東堀端', u'古金谷']), -- SpecialMergeZip(u'9218046', u'石川県', u'金沢市', -- [u'大桑町', u'三小牛町']), -+ SpecialMergeZip('5900111', '大阪府', '堺市中区', ['三原台']), -+ SpecialMergeZip('8710046', '大分県', '中津市', -+ ['金谷', '西堀端', '東堀端', '古金谷']), -+ SpecialMergeZip('9218046', '石川県', '金沢市', -+ ['大桑町', '三小牛町']), - ] - - ---- /src/gui/character_pad/data/gen_cp932_map.py -+++ /src/gui/character_pad/data/gen_cp932_map.py -@@ -32,7 +32,6 @@ - - import re - import sys --import string - - kUnicodePat = re.compile(r'0x[0-9A-Fa-f]{2,4}') - def IsValidUnicode(n): -@@ -42,28 +41,29 @@ - fh = open(sys.argv[1]) - result = {} - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue -- array = string.split(line) -+ array = line.split() - sjis = array[0] - ucs2 = array[1] - if eval(sjis) < 32 or not IsValidUnicode(ucs2): - continue - result.setdefault(ucs2, sjis) -+ fh.close() - - keys = sorted(result.keys()) - -- print "struct CP932MapData {" -- print " unsigned int ucs4;" -- print " unsigned short int sjis;" -- print "};" -- print "" -- print "static const size_t kCP932MapDataSize = %d;" % (len(keys)) -- print "static const CP932MapData kCP932MapData[] = {" -+ print("struct CP932MapData {") -+ print(" unsigned int ucs4;") -+ print(" unsigned short int sjis;") -+ print("};") -+ print("") -+ print("static const size_t kCP932MapDataSize = %d;" % (len(keys))) -+ print("static const CP932MapData kCP932MapData[] = {") - for n in keys: -- print " { %s, %s }," % (n ,result[n]) -- print " { 0, 0 }"; -- print "};" -+ print(" { %s, %s }," % (n ,result[n])) -+ print(" { 0, 0 }"); -+ print("};") - - if __name__ == "__main__": - main() ---- /src/gui/character_pad/data/gen_local_character_map.py -+++ /src/gui/character_pad/data/gen_local_character_map.py -@@ -30,7 +30,6 @@ - - __author__ = "taku" - --import string - import re - import sys - -@@ -43,9 +42,9 @@ - fh = open(filename) - result = [] - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue -- array = string.split(line) -+ array = line.split() - jis = array[0].replace('0x', '') - ucs2 = array[1].replace('0x', '') - if len(jis) == 2: -@@ -53,6 +52,7 @@ - - if IsValidUnicode(ucs2): - result.append([jis, ucs2]) -+ fh.close() - - return ["JISX0201", result] - -@@ -60,13 +60,14 @@ - fh = open(filename) - result = [] - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue - array = line.split() - jis = array[1].replace('0x', '') - ucs2 = array[2].replace('0x', '') - if IsValidUnicode(ucs2): - result.append([jis, ucs2]) -+ fh.close() - - return ["JISX0208", result] - -@@ -74,13 +75,14 @@ - fh = open(filename) - result = [] - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue - array = line.split() - jis = array[0].replace('0x', '') - ucs2 = array[1].replace('0x', '') - if IsValidUnicode(ucs2): - result.append([jis, ucs2]) -+ fh.close() - - return ["JISX0212", result] - -@@ -88,7 +90,7 @@ - fh = open(filename) - result = [] - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue - array = line.split() - sjis = array[0].replace('0x', '') -@@ -100,19 +102,20 @@ - - if IsValidUnicode(ucs2): - result.append([sjis, ucs2]) -+ fh.close() - - return ["CP932", result] - - def Output(arg): - name = arg[0] - result = arg[1] -- print "static const size_t k%sMapSize = %d;" % (name, len(result)) -- print "static const mozc::gui::CharacterPalette::LocalCharacterMap k%sMap[] = {" % (name) -+ print("static const size_t k%sMapSize = %d;" % (name, len(result))) -+ print("static const mozc::gui::CharacterPalette::LocalCharacterMap k%sMap[] = {" % (name)) - for n in result: -- print " { 0x%s, 0x%s }," % (n[0] ,n[1]) -- print " { 0, 0 }"; -- print "};" -- print "" -+ print(" { 0x%s, 0x%s }," % (n[0] ,n[1])) -+ print(" { 0, 0 }"); -+ print("};") -+ print("") - - if __name__ == "__main__": - Output(LoadJISX0201(sys.argv[1])) ---- /src/gui/character_pad/data/gen_unicode_blocks.py -+++ /src/gui/character_pad/data/gen_unicode_blocks.py -@@ -33,13 +33,13 @@ - import sys - import re - --re = re.compile('^(.....?)\.\.(.....?); (.+)') -+re = re.compile(r'^(.....?)\.\.(.....?); (.+)') - - def main(): -- print "static const mozc::gui::CharacterPalette::UnicodeBlock kUnicodeBlockTable[] = {" -+ print("static const mozc::gui::CharacterPalette::UnicodeBlock kUnicodeBlockTable[] = {") - fh = open(sys.argv[1]) - for line in fh.readlines(): -- if line[0] is '#': -+ if line[0] == '#': - continue - m = re.match(line) - if m is not None: -@@ -47,11 +47,12 @@ - end = int(m.group(2), 16) - name = m.group(3) - if start <= 0x2FFFF and end <= 0x2FFFF: -- print " { \"%s\", { %d, %d } }," % (name, start, end) -+ print(" { \"%s\", { %d, %d } }," % (name, start, end)) -+ fh.close() - -- print " { NULL, { 0, 0 } }" -- print "};" -- print "" -+ print(" { NULL, { 0, 0 } }") -+ print("};") -+ print("") - - if __name__ == "__main__": - main() ---- /src/gui/character_pad/data/gen_unicode_data.py -+++ /src/gui/character_pad/data/gen_unicode_data.py -@@ -46,18 +46,19 @@ - code = int(code, 16) - if code < 0x2FFFF: - results.append(" { %d, \"%s\" }," % (code, desc)) -+ fh.close() - -- print "struct UnicodeData {"; -- print " char32 ucs4;"; -- print " const char *description;"; -- print "};"; -- print "" -- print "static const size_t kUnicodeDataSize = %d;" % (len(results)) -- print "static const UnicodeData kUnicodeData[] = {"; -+ print("struct UnicodeData {"); -+ print(" char32 ucs4;"); -+ print(" const char *description;"); -+ print("};"); -+ print("") -+ print("static const size_t kUnicodeDataSize = %d;" % (len(results))) -+ print("static const UnicodeData kUnicodeData[] = {"); - for line in results: -- print line; -- print " { 0, NULL }"; -- print "};"; -+ print(line); -+ print(" { 0, NULL }"); -+ print("};"); - - if __name__ == "__main__": - main() ---- /src/gui/character_pad/data/gen_unihan_data.py -+++ /src/gui/character_pad/data/gen_unihan_data.py -@@ -31,35 +31,34 @@ - __author__ = "taku" - - import re --import string - import sys - rs = {} - - def Escape(n): -- if n is not "NULL": -+ if n != "NULL": - return "\"%s\"" % (n) - else: - return "NULL" - - def GetCode(n): -- if n is not "NULL": -- n = string.replace(n, '0-', 'JIS X 0208: 0x') -- n = string.replace(n, '1-', 'JIS X 0212: 0x') -- n = string.replace(n, '3-', 'JIS X 0213: 0x') -- n = string.replace(n, '4-', 'JIS X 0213: 0x') -- n = string.replace(n, 'A-', 'Vendors Ideographs: 0x') -- n = string.replace(n, '3A', 'JIS X 0213 2000: 0x') -+ if n != "NULL": -+ n = n.replace('0-', 'JIS X 0208: 0x') -+ n = n.replace('1-', 'JIS X 0212: 0x') -+ n = n.replace('3-', 'JIS X 0213: 0x') -+ n = n.replace('4-', 'JIS X 0213: 0x') -+ n = n.replace('A-', 'Vendors Ideographs: 0x') -+ n = n.replace('3A', 'JIS X 0213 2000: 0x') - return "\"%s\"" % n - else: - return "NULL" - - def GetRadical(n): - pat = re.compile(r'^(\d+)\.') -- if n is not "NULL": -+ if n != "NULL": - m = pat.match(n) - if m: - result = rs[m.group(1)] -- return "\"%s\"" % (result.encode('string_escape')) -+ return "\"%s\"" % result - else: - return "NULL" - else: -@@ -73,6 +72,7 @@ - id = array[1] - radical = array[2] - rs[id] = radical -+ fh.close() - - dic = {} - pat = re.compile(r'^U\+(\S+)\s+(kTotalStrokes|kJapaneseKun|kJapaneseOn|kRSUnicode|kIRG_JSource)\t(.+)') -@@ -86,23 +86,24 @@ - n = int(m.group(1), 16) - if n <= 65536: - dic.setdefault(key, {}).setdefault(field, value) -+ fh.close() - - keys = sorted(dic.keys()) - -- print "struct UnihanData {"; -- print " unsigned int ucs4;"; -+ print("struct UnihanData {"); -+ print(" unsigned int ucs4;"); - # Since the total strokes defined in Unihan data is Chinese-based - # number, we can't use it. - # print " unsigned char total_strokes;"; -- print " const char *japanese_kun;"; -- print " const char *japanese_on;"; -+ print(" const char *japanese_kun;"); -+ print(" const char *japanese_on;"); - # Since the radical information defined in Unihan data is Chinese-based - # number, we can't use it. - # print " const char *radical;"; -- print " const char *IRG_jsource;"; -- print "};" -- print "static const size_t kUnihanDataSize = %d;" % (len(keys)) -- print "static const UnihanData kUnihanData[] = {" -+ print(" const char *IRG_jsource;"); -+ print("};") -+ print("static const size_t kUnihanDataSize = %d;" % (len(keys))) -+ print("static const UnihanData kUnihanData[] = {") - - for key in keys: - total_strokes = dic[key].get("kTotalStrokes", "0") -@@ -111,9 +112,9 @@ - rad = GetRadical(dic[key].get("kRSUnicode", "NULL")) - code = GetCode(dic[key].get("kIRG_JSource", "NULL")) - # print " { 0x%s, %s, %s, %s, %s, %s }," % (key, total_strokes, kun, on, rad, code) -- print " { 0x%s, %s, %s, %s }," % (key, kun, on, code) -+ print(" { 0x%s, %s, %s, %s }," % (key, kun, on, code)) - -- print "};" -+ print("};") - - if __name__ == "__main__": - main() diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch deleted file mode 100644 index 41d2bf9eeb90..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch +++ /dev/null @@ -1,537 +0,0 @@ -https://github.com/google/mozc/issues/462 - ---- /src/prediction/gen_zero_query_data.py -+++ /src/prediction/gen_zero_query_data.py -@@ -59,20 +59,20 @@ - Returns: - A integer indicating parsed pua. - """ -- if not s or s[0] == '>': -+ if not s or s[0:1] == b'>': - return 0 - return int(s, 16) - - - def NormalizeString(string): - return unicodedata.normalize( -- 'NFKC', string.decode('utf-8')).encode('utf-8').replace('~', '〜') -+ 'NFKC', string.decode('utf-8')).replace('~', '〜').encode('utf-8') - - - def RemoveTrailingNumber(string): - if not string: -- return '' -- return re.sub(r'^([^0-9]+)[0-9]+$', r'\1', string) -+ return b'' -+ return re.sub(br'^([^0-9]+)[0-9]+$', r'\1', string) - - - def GetReadingsFromDescription(description): -@@ -84,19 +84,19 @@ - # - ビル・建物 - # \xE3\x83\xBB : "・" - return [RemoveTrailingNumber(token) for token -- in re.split(r'(?:\(|\)|/|\xE3\x83\xBB)+', normalized)] -+ in re.split(br'(?:\(|\)|/|\xE3\x83\xBB)+', normalized)] - - - def ReadEmojiTsv(stream): - """Reads emoji data from stream and returns zero query data.""" - zero_query_dict = defaultdict(list) - stream = code_generator_util.SkipLineComment(stream) -- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'): -+ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'): - if len(columns) != 13: -- logging.critical('format error: %s', '\t'.join(columns)) -+ logging.critical('format error: %s', b'\t'.join(columns)) - sys.exit(1) - -- code_points = columns[0].split(' ') -+ code_points = columns[0].split(b' ') - - # Emoji code point. - emoji = columns[1] -@@ -114,12 +114,12 @@ - # - Composite emoji which has multiple code point. - # NOTE: Some Unicode 6.0 emoji don't have PUA, and it is also omitted. - # TODO(hsumita): Check the availability of such emoji and enable it. -- logging.info('Skip %s', ' '.join(code_points)) -+ logging.info('Skip %s', b' '.join(code_points)) - continue - - reading_list = [] - # \xe3\x80\x80 is a full-width space -- for reading in re.split(r'(?: |\xe3\x80\x80)+', NormalizeString(readings)): -+ for reading in re.split(br'(?: |\xe3\x80\x80)+', NormalizeString(readings)): - if not reading: - continue - reading_list.append(reading) -@@ -158,15 +158,15 @@ - zero_query_dict = defaultdict(list) - - for line in input_stream: -- if line.startswith('#'): -+ if line.startswith(b'#'): - continue -- line = line.rstrip('\r\n') -+ line = line.rstrip(b'\r\n') - if not line: - continue - -- tokens = line.split('\t') -+ tokens = line.split(b'\t') - key = tokens[0] -- values = tokens[1].split(',') -+ values = tokens[1].split(b',') - - for value in values: - zero_query_dict[key].append( -@@ -179,16 +179,16 @@ - """Reads emoticon data from stream and returns zero query data.""" - zero_query_dict = defaultdict(list) - stream = code_generator_util.SkipLineComment(stream) -- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'): -+ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'): - if len(columns) != 3: -- logging.critical('format error: %s', '\t'.join(columns)) -+ logging.critical('format error: %s', b'\t'.join(columns)) - sys.exit(1) - - emoticon = columns[0] - readings = columns[2] - - # \xe3\x80\x80 is a full-width space -- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()): -+ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()): - if not reading: - continue - zero_query_dict[reading].append( -@@ -202,9 +202,9 @@ - """Reads emoji data from stream and returns zero query data.""" - zero_query_dict = defaultdict(list) - stream = code_generator_util.SkipLineComment(stream) -- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'): -+ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'): - if len(columns) < 3: -- logging.warning('format error: %s', '\t'.join(columns)) -+ logging.warning('format error: %s', b'\t'.join(columns)) - continue - - symbol = columns[1] -@@ -222,7 +222,7 @@ - continue - - # \xe3\x80\x80 is a full-width space -- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()): -+ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()): - if not reading: - continue - zero_query_dict[reading].append( -@@ -247,7 +247,7 @@ - - def IsValidKeyForZeroQuery(key): - """Returns if the key is valid for zero query trigger.""" -- is_ascii = all(ord(char) < 128 for char in key) -+ is_ascii = all(char < 128 for char in key) - return not is_ascii - - -@@ -301,13 +301,13 @@ - - def main(): - options = ParseOptions() -- with open(options.input_rule, 'r') as input_stream: -+ with open(options.input_rule, 'rb') as input_stream: - zero_query_rule_dict = ReadZeroQueryRuleData(input_stream) -- with open(options.input_symbol, 'r') as input_stream: -+ with open(options.input_symbol, 'rb') as input_stream: - zero_query_symbol_dict = ReadSymbolTsv(input_stream) -- with open(options.input_emoji, 'r') as input_stream: -+ with open(options.input_emoji, 'rb') as input_stream: - zero_query_emoji_dict = ReadEmojiTsv(input_stream) -- with open(options.input_emoticon, 'r') as input_stream: -+ with open(options.input_emoticon, 'rb') as input_stream: - zero_query_emoticon_dict = ReadEmoticonTsv(input_stream) - - merged_zero_query_dict = MergeZeroQueryData( ---- /src/prediction/gen_zero_query_number_data.py -+++ /src/prediction/gen_zero_query_number_data.py -@@ -41,15 +41,15 @@ - zero_query_dict = defaultdict(list) - - for line in input_stream: -- if line.startswith('#'): -+ if line.startswith(b'#'): - continue -- line = line.rstrip('\r\n') -+ line = line.rstrip(b'\r\n') - if not line: - continue - -- tokens = line.split('\t') -+ tokens = line.split(b'\t') - key = tokens[0] -- values = tokens[1].split(',') -+ values = tokens[1].split(b',') - - for value in values: - zero_query_dict[key].append( -@@ -71,7 +71,7 @@ - - def main(): - options = ParseOption() -- with open(options.input, 'r') as input_stream: -+ with open(options.input, 'rb') as input_stream: - zero_query_dict = ReadZeroQueryNumberData(input_stream) - util.WriteZeroQueryData(zero_query_dict, - options.output_token_array, ---- /src/prediction/gen_zero_query_util.py -+++ /src/prediction/gen_zero_query_util.py -@@ -69,7 +69,7 @@ - output_string_array): - # Collect all the strings and assing index in ascending order - string_index = {} -- for key, entry_list in zero_query_dict.iteritems(): -+ for key, entry_list in zero_query_dict.items(): - string_index[key] = 0 - for entry in entry_list: - string_index[entry.value] = 0 ---- /src/rewriter/gen_counter_suffix_array.py -+++ /src/rewriter/gen_counter_suffix_array.py -@@ -43,7 +43,7 @@ - with codecs.open(id_file, 'r', encoding='utf-8') as stream: - stream = code_generator_util.ParseColumnStream(stream, num_column=2) - for pos_id, pos_name in stream: -- if pos_name.startswith(u'名詞,接尾,助数詞'): -+ if pos_name.startswith('名詞,接尾,助数詞'): - pos_ids.add(pos_id) - return pos_ids - ---- /src/rewriter/gen_emoji_rewriter_data.py -+++ /src/rewriter/gen_emoji_rewriter_data.py -@@ -74,19 +74,19 @@ - the glyph (in other words, it has alternative (primary) code point, which - doesn't lead '>' and that's why we'll ignore it). - """ -- if not s or s[0] == '>': -+ if not s or s[0:1] == b'>': - return None - return int(s, 16) - - --_FULLWIDTH_RE = re.compile(ur'[!-~]') # U+FF01 - U+FF5E -+_FULLWIDTH_RE = re.compile(r'[!-~]') # U+FF01 - U+FF5E - - - def NormalizeString(string): - """Normalize full width ascii characters to half width characters.""" -- offset = ord(u'A') - ord(u'A') -- return _FULLWIDTH_RE.sub(lambda x: unichr(ord(x.group(0)) - offset), -- unicode(string, 'utf-8')).encode('utf-8') -+ offset = ord('A') - ord('A') -+ return _FULLWIDTH_RE.sub(lambda x: chr(ord(x.group(0)) - offset), -+ string.decode('utf-8')).encode('utf-8') - - - def ReadEmojiTsv(stream): -@@ -96,14 +96,14 @@ - token_dict = defaultdict(list) - - stream = code_generator_util.SkipLineComment(stream) -- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'): -+ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'): - if len(columns) != 13: -- logging.critical('format error: %s', '\t'.join(columns)) -+ logging.critical('format error: %s', b'\t'.join(columns)) - sys.exit(1) - -- code_points = columns[0].split(' ') -+ code_points = columns[0].split(b' ') - # Emoji code point. -- emoji = columns[1] if columns[1] else '' -+ emoji = columns[1] if columns[1] else b'' - android_pua = ParseCodePoint(columns[2]) - docomo_pua = ParseCodePoint(columns[3]) - softbank_pua = ParseCodePoint(columns[4]) -@@ -112,10 +112,10 @@ - readings = columns[6] - - # [7]: Name defined in Unicode. It is ignored in current implementation. -- utf8_description = columns[8] if columns[8] else '' -- docomo_description = columns[9] if columns[9] else '' -- softbank_description = columns[10] if columns[10] else '' -- kddi_description = columns[11] if columns[11] else '' -+ utf8_description = columns[8] if columns[8] else b'' -+ docomo_description = columns[9] if columns[9] else b'' -+ softbank_description = columns[10] if columns[10] else b'' -+ kddi_description = columns[11] if columns[11] else b'' - - if not android_pua or len(code_points) > 1: - # Skip some emoji, which is not supported on old devices. -@@ -123,7 +123,7 @@ - # - Composite emoji which has multiple code point. - # NOTE: Some Unicode 6.0 emoji don't have PUA, and it is also omitted. - # TODO(hsumita): Check the availability of such emoji and enable it. -- logging.info('Skip %s', ' '.join(code_points)) -+ logging.info('Skip %s', b' '.join(code_points)) - continue - - # Check consistency between carrier PUA codes and descriptions for Android -@@ -132,7 +132,7 @@ - (bool(softbank_pua) != bool(softbank_description)) or - (bool(kddi_pua) != bool(kddi_description))): - logging.warning('carrier PUA and description conflict: %s', -- '\t'.join(columns)) -+ b'\t'.join(columns)) - continue - - # Check if the character is usable on Android. -@@ -140,7 +140,7 @@ - android_pua = 0 # Replace None with 0. - - if not emoji and not android_pua: -- logging.info('Skip: %s', '\t'.join(columns)) -+ logging.info('Skip: %s', b'\t'.join(columns)) - continue - - index = len(emoji_data_list) -@@ -149,7 +149,7 @@ - kddi_description)) - - # \xe3\x80\x80 is a full-width space -- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()): -+ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()): - if reading: - token_dict[NormalizeString(reading)].append(index) - -@@ -159,7 +159,7 @@ - def OutputData(emoji_data_list, token_dict, - token_array_file, string_array_file): - """Output token and string arrays to files.""" -- sorted_token_dict = sorted(token_dict.iteritems()) -+ sorted_token_dict = sorted(token_dict.items()) - - strings = {} - for reading, _ in sorted_token_dict: -@@ -171,7 +171,7 @@ - strings[docomo_description] = 0 - strings[softbank_description] = 0 - strings[kddi_description] = 0 -- sorted_strings = sorted(strings.iterkeys()) -+ sorted_strings = sorted(strings.keys()) - for index, s in enumerate(sorted_strings): - strings[s] = index - -@@ -205,7 +205,7 @@ - - def main(): - options = ParseOptions() -- with open(options.input, 'r') as input_stream: -+ with open(options.input, 'rb') as input_stream: - (emoji_data_list, token_dict) = ReadEmojiTsv(input_stream) - - OutputData(emoji_data_list, token_dict, ---- /src/rewriter/gen_reading_correction_data.py -+++ /src/rewriter/gen_reading_correction_data.py -@@ -63,7 +63,7 @@ - def WriteData(input_path, output_value_array_path, output_error_array_path, - output_correction_array_path): - outputs = [] -- with open(input_path) as input_stream: -+ with open(input_path, 'rb') as input_stream: - input_stream = code_generator_util.SkipLineComment(input_stream) - input_stream = code_generator_util.ParseColumnStream(input_stream, - num_column=3) -@@ -73,7 +73,7 @@ - - # In order to lookup the entries via |error| with binary search, - # sort outputs here. -- outputs.sort(lambda x, y: cmp(x[1], y[1]) or cmp(x[0], y[0])) -+ outputs.sort(key=lambda x: (x[1], x[0])) - - serialized_string_array_builder.SerializeToFile( - [value for (value, _, _) in outputs], output_value_array_path) ---- /src/rewriter/gen_single_kanji_rewriter_data.py -+++ /src/rewriter/gen_single_kanji_rewriter_data.py -@@ -52,7 +52,7 @@ - stream = code_generator_util.ParseColumnStream(stream, num_column=2) - outputs = list(stream) - # For binary search by |key|, sort outputs here. -- outputs.sort(lambda x, y: cmp(x[0], y[0])) -+ outputs.sort(key=lambda x: x[0]) - - return outputs - -@@ -72,7 +72,7 @@ - variant_items.append([target, original, len(variant_types) - 1]) - - # For binary search by |target|, sort variant items here. -- variant_items.sort(lambda x, y: cmp(x[0], y[0])) -+ variant_items.sort(key=lambda x: x[0]) - - return (variant_types, variant_items) - -@@ -151,10 +151,10 @@ - def main(): - options = _ParseOptions() - -- with open(options.single_kanji_file, 'r') as single_kanji_stream: -+ with open(options.single_kanji_file, 'rb') as single_kanji_stream: - single_kanji = ReadSingleKanji(single_kanji_stream) - -- with open(options.variant_file, 'r') as variant_stream: -+ with open(options.variant_file, 'rb') as variant_stream: - variant_info = ReadVariant(variant_stream) - - WriteSingleKanji(single_kanji, ---- /src/session/gen_session_stress_test_data.py -+++ /src/session/gen_session_stress_test_data.py -@@ -50,24 +50,26 @@ - """ - result = '' - for c in s: -- hexstr = hex(ord(c)) -+ hexstr = hex(c) - # because hexstr contains '0x', remove the prefix and add our prefix - result += '\\x' + hexstr[2:] - return result - - def GenerateHeader(file): - try: -- print "const char *kTestSentences[] = {" -- for line in open(file, "r"): -- if line.startswith('#'): -+ print("const char *kTestSentences[] = {") -+ fh = open(file, "rb") -+ for line in fh: -+ if line.startswith(b'#'): - continue -- line = line.rstrip('\r\n') -+ line = line.rstrip(b'\r\n') - if not line: - continue -- print " \"%s\"," % escape_string(line) -- print "};" -+ print(" \"%s\"," % escape_string(line)) -+ fh.close() -+ print("};") - except: -- print "cannot open %s" % (file) -+ print("cannot open %s" % (file)) - sys.exit(1) - - def main(): ---- /src/unix/ibus/gen_mozc_xml.py -+++ /src/unix/ibus/gen_mozc_xml.py -@@ -74,7 +74,7 @@ - - - def OutputXmlElement(param_dict, element_name, value): -- print ' <%s>%s</%s>' % (element_name, (value % param_dict), element_name) -+ print(' <%s>%s</%s>' % (element_name, (value % param_dict), element_name)) - - - def OutputXml(param_dict, component, engine_common, engines, setup_arg): -@@ -90,26 +90,26 @@ - engines: A dictionary from a property name to a list of property values of - engines. For example, {'name': ['mozc-jp', 'mozc', 'mozc-dv']}. - """ -- print '<component>' -- for key in component: -+ print('<component>') -+ for key in sorted(component): - OutputXmlElement(param_dict, key, component[key]) -- print '<engines>' -+ print('<engines>') - for i in range(len(engines['name'])): -- print '<engine>' -- for key in engine_common: -+ print('<engine>') -+ for key in sorted(engine_common): - OutputXmlElement(param_dict, key, engine_common[key]) - if setup_arg: - OutputXmlElement(param_dict, 'setup', ' '.join(setup_arg)) -- for key in engines: -+ for key in sorted(engines): - OutputXmlElement(param_dict, key, engines[key][i]) -- print '</engine>' -- print '</engines>' -- print '</component>' -+ print('</engine>') -+ print('</engines>') -+ print('</component>') - - - def OutputCppVariable(param_dict, prefix, variable_name, value): -- print 'const char k%s%s[] = "%s";' % (prefix, variable_name.capitalize(), -- (value % param_dict)) -+ print('const char k%s%s[] = "%s";' % (prefix, variable_name.capitalize(), -+ (value % param_dict))) - - - def OutputCpp(param_dict, component, engine_common, engines): -@@ -122,18 +122,18 @@ - engines: ditto. - """ - guard_name = 'MOZC_UNIX_IBUS_MAIN_H_' -- print CPP_HEADER % (guard_name, guard_name) -- for key in component: -+ print(CPP_HEADER % (guard_name, guard_name)) -+ for key in sorted(component): - OutputCppVariable(param_dict, 'Component', key, component[key]) -- for key in engine_common: -+ for key in sorted(engine_common): - OutputCppVariable(param_dict, 'Engine', key, engine_common[key]) -- for key in engines: -- print 'const char* kEngine%sArray[] = {' % key.capitalize() -+ for key in sorted(engines): -+ print('const char* kEngine%sArray[] = {' % key.capitalize()) - for i in range(len(engines[key])): -- print '"%s",' % (engines[key][i] % param_dict) -- print '};' -- print 'const size_t kEngineArrayLen = %s;' % len(engines['name']) -- print CPP_FOOTER % guard_name -+ print('"%s",' % (engines[key][i] % param_dict)) -+ print('};') -+ print('const size_t kEngineArrayLen = %s;' % len(engines['name'])) -+ print(CPP_FOOTER % guard_name) - - - def CheckIBusVersion(options, minimum_version): ---- /src/usage_stats/gen_stats_list.py -+++ /src/usage_stats/gen_stats_list.py -@@ -37,23 +37,24 @@ - - def GetStatsNameList(filename): - stats = [] -- for line in open(filename, 'r'): -- stat = line.strip() -- if not stat or stat[0] == '#': -- continue -- stats.append(stat) -+ with open(filename, 'r') as file: -+ for line in file: -+ stat = line.strip() -+ if not stat or stat[0] == '#': -+ continue -+ stats.append(stat) - return stats - - - def main(): - stats_list = GetStatsNameList(sys.argv[1]) -- print '// This header file is generated by gen_stats_list.py' -+ print('// This header file is generated by gen_stats_list.py') - for stats in stats_list: -- print 'const char k%s[] = "%s";' % (stats, stats) -- print 'const char *kStatsList[] = {' -+ print('const char k%s[] = "%s";' % (stats, stats)) -+ print('const char *kStatsList[] = {') - for stats in stats_list: -- print ' k%s,' % (stats) -- print '};' -+ print(' k%s,' % (stats)) -+ print('};') - - - if __name__ == '__main__': diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-reiwa.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-reiwa.patch deleted file mode 100644 index 561a989a41d3..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-reiwa.patch +++ /dev/null @@ -1,35 +0,0 @@ -https://github.com/google/mozc/issues/461 - ---- /src/data/symbol/categorized.tsv -+++ /src/data/symbol/categorized.tsv -@@ -405,6 +405,7 @@ - ㍽ MATH たいしょう ねんごう - ㍼ MATH しょうわ ねんごう - ㍻ MATH へいせい ねんごう -+㋿ MATH れいわ ねんごう - - ㌣ MATH せんと たんい - ㌦ MATH どる たんい ---- /src/data/symbol/symbol.tsv -+++ /src/data/symbol/symbol.tsv -@@ -528,6 +528,7 @@ - 記号 ㊦ した げ 丸下 - 記号 ㊧ ひだり 丸左 - 記号 ㊨ みぎ 丸右 -+記号 ㋿ れいわ ねんごう 年号 OTHER - 記号 ㍻ へいせい ねんごう 年号 OTHER - 記号 ㍼ しょうわ ねんごう 年号 OTHER - 記号 ㍽ たいしょう ねんごう 年号 OTHER ---- /src/rewriter/date_rewriter.cc -+++ /src/rewriter/date_rewriter.cc -@@ -1239,6 +1239,10 @@ - 1989, - "平成", - "へいせい", -+ }, { -+ 2019, -+ "令和", -+ "れいわ", - } - }; - diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-server_path_check.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-server_path_check.patch deleted file mode 100644 index dd606e27fb56..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-server_path_check.patch +++ /dev/null @@ -1,95 +0,0 @@ -https://github.com/google/mozc/issues/471 - ---- /src/ipc/ipc_path_manager.cc -+++ /src/ipc/ipc_path_manager.cc -@@ -332,9 +332,21 @@ - return false; - } - -+ // Expand symbolic links in the expected server path to avoid false negatives -+ // during comparisons of the expected server path and the actual server path. -+ string real_server_path = server_path; -+#ifndef OS_WIN -+ char real_server_path_[PATH_MAX]; -+ if (realpath(server_path.c_str(), real_server_path_) == NULL) { -+ LOG(ERROR) << "realpath failed: " << strerror(errno); -+ return false; -+ } -+ real_server_path = real_server_path_; -+#endif -+ - // compare path name - if (pid == server_pid_) { -- return (server_path == server_path_); -+ return (real_server_path == server_path_); - } - - server_pid_ = 0; -@@ -344,17 +356,17 @@ - { - std::wstring expected_server_ntpath; - const std::map<string, std::wstring>::const_iterator it = -- expected_server_ntpath_cache_.find(server_path); -+ expected_server_ntpath_cache_.find(real_server_path); - if (it != expected_server_ntpath_cache_.end()) { - expected_server_ntpath = it->second; - } else { - std::wstring wide_server_path; -- Util::UTF8ToWide(server_path, &wide_server_path); -+ Util::UTF8ToWide(real_server_path, &wide_server_path); - if (WinUtil::GetNtPath(wide_server_path, &expected_server_ntpath)) { -- // Caches the relationship from |server_path| to -- // |expected_server_ntpath| in case |server_path| is renamed later. -+ // Caches the relationship from |real_server_path| to -+ // |expected_server_ntpath| in case |real_server_path| is renamed later. - // (This can happen during the updating). -- expected_server_ntpath_cache_[server_path] = expected_server_ntpath; -+ expected_server_ntpath_cache_[real_server_path] = expected_server_ntpath; - } - } - -@@ -371,9 +383,9 @@ - return false; - } - -- // Here we can safely assume that |server_path| (expected one) should be -+ // Here we can safely assume that |real_server_path| (expected one) should be - // the same to |server_path_| (actual one). -- server_path_ = server_path; -+ server_path_ = real_server_path; - server_pid_ = pid; - } - #endif // OS_WIN -@@ -399,7 +411,7 @@ - #ifdef OS_LINUX - // load from /proc/<pid>/exe - char proc[128]; -- char filename[512]; -+ char filename[PATH_MAX]; - snprintf(proc, sizeof(proc) - 1, "/proc/%u/exe", pid); - const ssize_t size = readlink(proc, filename, sizeof(filename) - 1); - if (size == -1) { -@@ -412,18 +424,18 @@ - server_pid_ = pid; - #endif // OS_LINUX - -- VLOG(1) << "server path: " << server_path << " " << server_path_; -- if (server_path == server_path_) { -+ VLOG(1) << "server path: " << real_server_path << " " << server_path_; -+ if (real_server_path == server_path_) { - return true; - } - - #ifdef OS_LINUX -- if ((server_path + " (deleted)") == server_path_) { -- LOG(WARNING) << server_path << " on disk is modified"; -+ if ((real_server_path + " (deleted)") == server_path_) { -+ LOG(WARNING) << real_server_path << " on disk is modified"; - // If a user updates the server binary on disk during the server is running, - // "readlink /proc/<pid>/exe" returns a path with the " (deleted)" suffix. - // We allow the special case. -- server_path_ = server_path; -+ server_path_ = real_server_path; - return true; - } - #endif // OS_LINUX diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-system_libraries.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-system_libraries.patch deleted file mode 100644 index 064b910c7026..000000000000 --- a/app-i18n/mozc/files/mozc-2.23.2815.102-system_libraries.patch +++ /dev/null @@ -1,274 +0,0 @@ ---- /src/gyp/defines.gypi -+++ /src/gyp/defines.gypi -@@ -71,6 +71,12 @@ - # use_libibus represents if ibus library is used or not. - # This option is only for Linux. - 'use_libibus%': '0', -+ -+ # use_libgtest represents if gtest library is used or not. -+ 'use_libgtest%': '0', -+ -+ # use_libjsoncpp represents if jsoncpp library is used or not. -+ 'use_libjsoncpp%': '0', - }, - 'target_defaults': { - 'defines': [ ---- /src/net/jsoncpp.gyp -+++ /src/net/jsoncpp.gyp -@@ -31,32 +31,57 @@ - 'targets': [ - { - 'target_name': 'jsoncpp', -- 'type': 'static_library', -- 'variables': { -- 'jsoncpp_root': '<(third_party_dir)/jsoncpp', -- 'jsoncpp_srcs': [ -- '<(jsoncpp_root)/src/lib_json/json_reader.cpp', -- '<(jsoncpp_root)/src/lib_json/json_value.cpp', -- '<(jsoncpp_root)/src/lib_json/json_writer.cpp', -- ], -- 'jsoncpp_include_dirs': ['<(jsoncpp_root)/include'], -- 'jsoncpp_additional_macros': ['JSON_USE_EXCEPTION=0'], -- }, -- 'defines': [ -- '<@(jsoncpp_additional_macros)', -+ 'conditions': [ -+ ['use_libjsoncpp==1', { -+ 'type': 'none', -+ 'variables': { -+ 'jsoncpp_additional_macros': ['JSON_USE_EXCEPTION=0'], -+ }, -+ 'all_dependent_settings': { -+ 'defines': [ -+ '<@(jsoncpp_additional_macros)', -+ ], -+ 'cflags': [ -+ '<!@(pkg-config --cflags jsoncpp)', -+ ], -+ 'link_settings': { -+ 'libraries': [ -+ '<!@(pkg-config --libs-only-l jsoncpp)', -+ ], -+ 'ldflags': [ -+ '<!@(pkg-config --libs-only-L jsoncpp)', -+ ], -+ } -+ }, -+ }, { -+ 'type': 'static_library', -+ 'variables': { -+ 'jsoncpp_root': '<(third_party_dir)/jsoncpp', -+ 'jsoncpp_srcs': [ -+ '<(jsoncpp_root)/src/lib_json/json_reader.cpp', -+ '<(jsoncpp_root)/src/lib_json/json_value.cpp', -+ '<(jsoncpp_root)/src/lib_json/json_writer.cpp', -+ ], -+ 'jsoncpp_include_dirs': ['<(jsoncpp_root)/include'], -+ 'jsoncpp_additional_macros': ['JSON_USE_EXCEPTION=0'], -+ }, -+ 'defines': [ -+ '<@(jsoncpp_additional_macros)', -+ ], -+ 'sources': [ -+ '<@(jsoncpp_srcs)', -+ 'jsoncpp.h', -+ ], -+ 'include_dirs': [ -+ '<@(jsoncpp_include_dirs)', -+ ], -+ 'all_dependent_settings': { -+ 'defines': [ -+ '<@(jsoncpp_additional_macros)', -+ ], -+ }, -+ }], - ], -- 'sources': [ -- '<@(jsoncpp_srcs)', -- 'jsoncpp.h', -- ], -- 'include_dirs': [ -- '<@(jsoncpp_include_dirs)', -- ], -- 'all_dependent_settings': { -- 'defines': [ -- '<@(jsoncpp_additional_macros)', -- ], -- }, - }, - ], - } ---- /src/net/jsoncpp.h -+++ /src/net/jsoncpp.h -@@ -35,7 +35,7 @@ - // Mozc basically disables C++ exception. - #define JSON_USE_EXCEPTION 0 - #endif // !JSON_USE_EXCEPTION --#include "third_party/jsoncpp/include/json/json.h" -+#include <json/json.h> - #define MOZC_JSONCPP_JSON_H_INCLUDED - #endif // !MOZC_JSONCPP_JSON_H_INCLUDED - ---- /src/testing/testing.gyp -+++ /src/testing/testing.gyp -@@ -53,66 +53,101 @@ - 'targets': [ - { - 'target_name': 'testing', -- 'type': 'static_library', -- 'variables': { -- 'gtest_defines': [ -- 'GTEST_LANG_CXX11=1', -- 'GTEST_HAS_TR1_TUPLE=0', # disable tr1 tuple in favor of C++11 tuple. -- ], -- 'gtest_dir': '<(third_party_dir)/gtest/googletest', -- 'gmock_dir': '<(third_party_dir)/gtest/googlemock', -- 'conditions': [ -- ['_toolset=="target" and target_platform=="Android"', { -- 'gtest_defines': [ -- 'GTEST_HAS_RTTI=0', # Android NDKr7 requires this. -- 'GTEST_HAS_CLONE=0', -- 'GTEST_HAS_GLOBAL_WSTRING=0', -- 'GTEST_HAS_POSIX_RE=0', -- 'GTEST_HAS_STD_WSTRING=0', -- 'GTEST_OS_LINUX=1', -- 'GTEST_OS_LINUX_ANDROID=1', -- ], -- }], -- ], -- }, -- 'sources': [ -- '<(gmock_dir)/src/gmock-cardinalities.cc', -- '<(gmock_dir)/src/gmock-internal-utils.cc', -- '<(gmock_dir)/src/gmock-matchers.cc', -- '<(gmock_dir)/src/gmock-spec-builders.cc', -- '<(gmock_dir)/src/gmock.cc', -- '<(gtest_dir)/src/gtest-death-test.cc', -- '<(gtest_dir)/src/gtest-filepath.cc', -- '<(gtest_dir)/src/gtest-port.cc', -- '<(gtest_dir)/src/gtest-printers.cc', -- '<(gtest_dir)/src/gtest-test-part.cc', -- '<(gtest_dir)/src/gtest-typed-test.cc', -- '<(gtest_dir)/src/gtest.cc', -- ], -- 'include_dirs': [ -- '<(gmock_dir)', -- '<(gmock_dir)/include', -- '<(gtest_dir)', -- '<(gtest_dir)/include', -- ], -- 'defines': [ -- '<@(gtest_defines)', -- ], -- 'all_dependent_settings': { -- 'defines': [ -- '<@(gtest_defines)', -- ], -- 'include_dirs': [ -- '<(gmock_dir)/include', -- '<(gtest_dir)/include', -- ], -- }, - 'conditions': [ -- ['(_toolset=="target" and compiler_target=="clang") or ' -- '(_toolset=="host" and compiler_host=="clang")', { -- 'cflags': [ -- '-Wno-missing-field-initializers', -- '-Wno-unused-private-field', -+ ['use_libgtest==1', { -+ 'type': 'none', -+ 'variables': { -+ 'gtest_defines': [ -+ 'GTEST_LANG_CXX11=1', -+ 'GTEST_HAS_TR1_TUPLE=0', # disable tr1 tuple in favor of C++11 tuple. -+ ], -+ 'conditions': [ -+ ['_toolset=="target" and target_platform=="Android"', { -+ 'gtest_defines': [ -+ 'GTEST_HAS_RTTI=0', # Android NDKr7 requires this. -+ 'GTEST_HAS_CLONE=0', -+ 'GTEST_HAS_GLOBAL_WSTRING=0', -+ 'GTEST_HAS_POSIX_RE=0', -+ 'GTEST_HAS_STD_WSTRING=0', -+ 'GTEST_OS_LINUX=1', -+ 'GTEST_OS_LINUX_ANDROID=1', -+ ], -+ }], -+ ], -+ }, -+ 'all_dependent_settings': { -+ 'defines': [ -+ '<@(gtest_defines)', -+ ], -+ 'link_settings': { -+ 'libraries': [ -+ '-lgmock -lgtest', -+ ], -+ }, -+ }, -+ }, { -+ 'type': 'static_library', -+ 'variables': { -+ 'gtest_defines': [ -+ 'GTEST_LANG_CXX11=1', -+ 'GTEST_HAS_TR1_TUPLE=0', # disable tr1 tuple in favor of C++11 tuple. -+ ], -+ 'gtest_dir': '<(third_party_dir)/gtest/googletest', -+ 'gmock_dir': '<(third_party_dir)/gtest/googlemock', -+ 'conditions': [ -+ ['_toolset=="target" and target_platform=="Android"', { -+ 'gtest_defines': [ -+ 'GTEST_HAS_RTTI=0', # Android NDKr7 requires this. -+ 'GTEST_HAS_CLONE=0', -+ 'GTEST_HAS_GLOBAL_WSTRING=0', -+ 'GTEST_HAS_POSIX_RE=0', -+ 'GTEST_HAS_STD_WSTRING=0', -+ 'GTEST_OS_LINUX=1', -+ 'GTEST_OS_LINUX_ANDROID=1', -+ ], -+ }], -+ ], -+ }, -+ 'sources': [ -+ '<(gmock_dir)/src/gmock-cardinalities.cc', -+ '<(gmock_dir)/src/gmock-internal-utils.cc', -+ '<(gmock_dir)/src/gmock-matchers.cc', -+ '<(gmock_dir)/src/gmock-spec-builders.cc', -+ '<(gmock_dir)/src/gmock.cc', -+ '<(gtest_dir)/src/gtest-death-test.cc', -+ '<(gtest_dir)/src/gtest-filepath.cc', -+ '<(gtest_dir)/src/gtest-port.cc', -+ '<(gtest_dir)/src/gtest-printers.cc', -+ '<(gtest_dir)/src/gtest-test-part.cc', -+ '<(gtest_dir)/src/gtest-typed-test.cc', -+ '<(gtest_dir)/src/gtest.cc', -+ ], -+ 'include_dirs': [ -+ '<(gmock_dir)', -+ '<(gmock_dir)/include', -+ '<(gtest_dir)', -+ '<(gtest_dir)/include', -+ ], -+ 'defines': [ -+ '<@(gtest_defines)', -+ ], -+ 'all_dependent_settings': { -+ 'defines': [ -+ '<@(gtest_defines)', -+ ], -+ 'include_dirs': [ -+ '<(gmock_dir)/include', -+ '<(gtest_dir)/include', -+ ], -+ }, -+ 'conditions': [ -+ ['(_toolset=="target" and compiler_target=="clang") or ' -+ '(_toolset=="host" and compiler_host=="clang")', { -+ 'cflags': [ -+ '-Wno-missing-field-initializers', -+ '-Wno-unused-private-field', -+ ], -+ }], - ], - }], - ], |