summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorArfrever Frehtes Taifersar Arahesis <Arfrever@Apache.Org>2020-01-17 17:39:29 +0000
committerMike Gilbert <floppym@gentoo.org>2020-01-17 14:08:54 -0500
commita1ff8a4ff637d2a7a8ed60297ab90fe8bdc7fb46 (patch)
treee3bc74e70c3b38e68021b3fcede2949e7f1cda1c /app-i18n/mozc
parentapp-i18n/tagainijisho: Delete old versions (<1.2.0_pre20191230120610_p2020010... (diff)
downloadgentoo-a1ff8a4ff637d2a7a8ed60297ab90fe8bdc7fb46.tar.gz
gentoo-a1ff8a4ff637d2a7a8ed60297ab90fe8bdc7fb46.tar.bz2
gentoo-a1ff8a4ff637d2a7a8ed60297ab90fe8bdc7fb46.zip
app-i18n/mozc: Use Python 3.
Closes: https://bugs.gentoo.org/695006 Signed-off-by: Arfrever Frehtes Taifersar Arahesis <Arfrever@Apache.Org> Signed-off-by: Mike Gilbert <floppym@gentoo.org>
Diffstat (limited to 'app-i18n/mozc')
-rw-r--r--app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch621
-rw-r--r--app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch600
-rw-r--r--app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch583
-rw-r--r--app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch537
-rw-r--r--app-i18n/mozc/mozc-2.23.2815.102.ebuild8
-rw-r--r--app-i18n/mozc/mozc-9999.ebuild8
6 files changed, 2353 insertions, 4 deletions
diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch
new file mode 100644
index 000000000000..2b9bbd720cde
--- /dev/null
+++ b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_1.patch
@@ -0,0 +1,621 @@
+https://github.com/google/mozc/issues/462
+
+--- /src/base/gen_character_set.py
++++ /src/base/gen_character_set.py
+@@ -33,7 +33,6 @@
+ import itertools
+ import optparse
+ import re
+-import string
+ import sys
+
+
+@@ -89,7 +88,8 @@
+ @staticmethod
+ def _LoadTable(filename, column_index, pattern, validater):
+ result = set()
+- for line in open(filename):
++ fh = open(filename)
++ for line in fh:
+ if line.startswith('#'):
+ # Skip a comment line.
+ continue
+@@ -100,6 +100,7 @@
+ ucs = int(match.group(1), 16)
+ if validater(ucs):
+ result.add(ucs)
++ fh.close()
+
+ return result
+
+@@ -250,7 +251,7 @@
+ # (at most) four code points.
+ bit_list = []
+ for _, group in itertools.groupby(enumerate(category_list),
+- lambda (codepoint, _): codepoint / 4):
++ lambda x: x[0] // 4):
+ # Fill bits from LSB to MSB for each group.
+ bits = 0
+ for index, (_, category) in enumerate(group):
+@@ -263,7 +264,7 @@
+
+ # Output the content. Each line would have (at most) 16 bytes.
+ for _, group in itertools.groupby(enumerate(bit_list),
+- lambda (index, _): index / 16):
++ lambda x: x[0] // 16):
+ line = [' \"']
+ for _, bits in group:
+ line.append('\\x%02X' % bits)
+@@ -386,7 +387,7 @@
+ # Bitmap lookup.
+ # TODO(hidehiko): the bitmap has two huge 0-bits ranges. Reduce them.
+ category_map = [
+- (bits, category) for category, bits in CATEGORY_BITMAP.iteritems()]
++ (bits, category) for category, bits in CATEGORY_BITMAP.items()]
+ category_map.sort()
+
+ lines.extend([
+@@ -451,7 +452,7 @@
+ options.jisx0213file)
+ category_list = [
+ categorizer.GetCategory(codepoint)
+- for codepoint in xrange(categorizer.MaxCodePoint() + 1)]
++ for codepoint in range(categorizer.MaxCodePoint() + 1)]
+ generated_character_set_header = GenerateCharacterSetHeader(category_list)
+
+ # Write the result.
+--- /src/base/gen_config_file_stream_data.py
++++ /src/base/gen_config_file_stream_data.py
+@@ -58,7 +58,7 @@
+ result = []
+ result.append(' { "%s", "' % os.path.basename(path))
+ with open(path, 'rb') as stream:
+- result.extend(r'\x%02X' % ord(byte) for byte in stream.read())
++ result.extend(r'\x%02X' % byte for byte in stream.read())
+ result.append('", %d }' % os.path.getsize(path))
+
+ return ''.join(result)
+@@ -93,8 +93,8 @@
+ def main():
+ (options, args) = ParseOptions()
+ if not options.output:
+- print >>sys.stderr, (
+- 'usage: gen_config_file_stream_data.py --output=filepath input ...')
++ print('usage: gen_config_file_stream_data.py --output=filepath input ...',
++ file=sys.stderr)
+ sys.exit(2)
+
+ with open(options.output, 'w') as output:
+--- /src/build_mozc.py
++++ /src/build_mozc.py
+@@ -943,7 +943,7 @@
+ logging.info('running %s...', binary)
+ try:
+ test_function(binary, gtest_report_dir, options)
+- except RunOrDieError, e:
++ except RunOrDieError as e:
+ logging.error(e)
+ failed_tests.append(binary)
+ else:
+@@ -1082,7 +1082,7 @@
+ # and '-c' and 'Release' are build options.
+ targets = []
+ build_options = []
+- for i in xrange(len(args)):
++ for i in range(len(args)):
+ if args[i].startswith('-'):
+ # starting with build options
+ build_options = args[i:]
+@@ -1190,14 +1190,14 @@
+
+ def ShowHelpAndExit():
+ """Shows the help message."""
+- print 'Usage: build_mozc.py COMMAND [ARGS]'
+- print 'Commands: '
+- print ' gyp Generate project files.'
+- print ' build Build the specified target.'
+- print ' runtests Build all tests and run them.'
+- print ' clean Clean all the build files and directories.'
+- print ''
+- print 'See also the comment in the script for typical usage.'
++ print('Usage: build_mozc.py COMMAND [ARGS]')
++ print('Commands: ')
++ print(' gyp Generate project files.')
++ print(' build Build the specified target.')
++ print(' runtests Build all tests and run them.')
++ print(' clean Clean all the build files and directories.')
++ print('')
++ print('See also the comment in the script for typical usage.')
+ sys.exit(1)
+
+
+--- /src/build_tools/android_util.py
++++ /src/build_tools/android_util.py
+@@ -548,7 +548,7 @@
+ (devices_result, _) = process.communicate()
+ used_ports = set(int(port) for port
+ in re.findall(r'emulator-(\d+)', devices_result))
+- return [port for port in xrange(5554, 5586, 2) if port not in used_ports]
++ return [port for port in range(5554, 5586, 2) if port not in used_ports]
+
+
+ def SetUpTestingSdkHomeDirectory(dest_android_sdk_home,
+@@ -575,7 +575,7 @@
+ 'create', 'avd',
+ '--force',
+ '--sdcard', '512M',]
+- for key, value in options.iteritems():
++ for key, value in options.items():
+ args.extend([key, value])
+ env = {'ANDROID_SDK_HOME': os.path.abspath(dest_android_sdk_home)}
+ logging.info('Creating AVD: %s', args)
+@@ -615,7 +615,7 @@
+ def main():
+ for arg in sys.argv[1:]:
+ for item in sorted(GetApkProperties(arg).items()):
+- print '%s: %s' % item
++ print('%s: %s' % item)
+
+
+ if __name__ == '__main__':
+--- /src/build_tools/binary_size_checker.py
++++ /src/build_tools/binary_size_checker.py
+@@ -70,12 +70,12 @@
+ actual_size = os.stat(filename).st_size
+ expected_size = EXPECTED_MAXIMUM_SIZES[basename]
+ if actual_size < expected_size * 1024 * 1024:
+- print 'Pass: %s (size: %d) is smaller than expected (%d MB)' % (
+- filename, actual_size, expected_size)
++ print('Pass: %s (size: %d) is smaller than expected (%d MB)' % (
++ filename, actual_size, expected_size))
+ return True
+ else:
+- print 'WARNING: %s (size: %d) is larger than expected (%d MB)' % (
+- filename, actual_size, expected_size)
++ print('WARNING: %s (size: %d) is larger than expected (%d MB)' % (
++ filename, actual_size, expected_size))
+ return False
+
+
+--- /src/build_tools/build_and_sign_pkg_mac.py
++++ /src/build_tools/build_and_sign_pkg_mac.py
+@@ -44,8 +44,8 @@
+ import shutil
+ import sys
+
+-from util import PrintErrorAndExit
+-from util import RunOrDie
++from .util import PrintErrorAndExit
++from .util import RunOrDie
+
+
+ def ParseOption():
+--- /src/build_tools/build_breakpad.py
++++ /src/build_tools/build_breakpad.py
+@@ -54,9 +54,9 @@
+ try:
+ subprocess.check_output(command)
+ except subprocess.CalledProcessError as e:
+- print e.output
++ print(e.output)
+ sys.exit(e.returncode)
+- print 'Done: %s' % ' '.join(command)
++ print('Done: %s' % ' '.join(command))
+
+
+ def Xcodebuild(projdir, target, arch, sdk, outdir):
+--- /src/build_tools/build_diskimage_mac.py
++++ /src/build_tools/build_diskimage_mac.py
+@@ -90,7 +90,7 @@
+ # setup volume directory
+ temp_dir = tempfile.mkdtemp()
+ CopyFile(path.join(build_dir, ".keystone_install"), temp_dir)
+- os.chmod(path.join(temp_dir, ".keystone_install"), 0755) # rwxr-xr-x
++ os.chmod(path.join(temp_dir, ".keystone_install"), 0o755) # rwxr-xr-x
+ for a in args:
+ CopyFile(path.join(build_dir, a), temp_dir)
+
+--- /src/build_tools/change_reference_mac.py
++++ /src/build_tools/change_reference_mac.py
+@@ -41,8 +41,8 @@
+ import optparse
+ import os
+
+-from util import PrintErrorAndExit
+-from util import RunOrDie
++from .util import PrintErrorAndExit
++from .util import RunOrDie
+
+
+ def ParseOption():
+--- /src/build_tools/code_generator_util.py
++++ /src/build_tools/code_generator_util.py
+@@ -33,27 +33,26 @@
+ __author__ = "hidehiko"
+
+ import struct
+-import types
+
+
+ def ToCppStringLiteral(s):
+ """Returns C-style string literal, or NULL if given s is None."""
+ if s is None:
+- return 'NULL'
++ return b'NULL'
+
+- if all(0x20 <= ord(c) <= 0x7E for c in s):
++ if all(0x20 <= c <= 0x7E for c in s):
+ # All characters are in ascii code.
+- return '"%s"' % s.replace('\\', r'\\').replace('"', r'\"')
++ return b'"%b"' % s.replace(b'\\', br'\\').replace(b'"', br'\"')
+ else:
+ # One or more characters are non-ascii.
+- return '"%s"' % ''.join(r'\x%02X' % ord(c) for c in s)
++ return b'"%b"' % b''.join(br'\x%02X' % c for c in s)
+
+
+ def FormatWithCppEscape(format_text, *args):
+ """Returns a string filling format with args."""
+ literal_list = []
+ for arg in args:
+- if isinstance(arg, (types.StringType, types.NoneType)):
++ if isinstance(arg, (bytes, type(None))):
+ arg = ToCppStringLiteral(arg)
+ literal_list.append(arg)
+
+@@ -95,7 +94,7 @@
+ if target_compiler and target_compiler.startswith('msvs'):
+ stream.write('const uint64 k%s_data_wordtype[] = {\n' % variable_name)
+
+- for word_index in xrange(0, len(data), 8):
++ for word_index in range(0, len(data), 8):
+ word_chunk = data[word_index:word_index + 8].ljust(8, '\x00')
+ stream.write('0x%016X, ' % struct.unpack('<Q', word_chunk))
+ if (word_index / 8) % 4 == 3:
+@@ -111,7 +110,7 @@
+ stream.write('const char k%s_data[] =\n' % variable_name)
+ # Output 16bytes per line.
+ chunk_size = 16
+- for index in xrange(0, len(data), chunk_size):
++ for index in range(0, len(data), chunk_size):
+ chunk = data[index:index + chunk_size]
+ stream.write('"')
+ stream.writelines(r'\x%02X' % ord(c) for c in chunk)
+@@ -126,36 +125,50 @@
+ if type(codepoint_list) is int:
+ codepoint_list = (codepoint_list,)
+ if codepoint_list is None or len(codepoint_list) == 0:
+- return 'null'
+- result = r'"'
++ return b'null'
++ result = b'"'
+ for codepoint in codepoint_list:
+- utf16_string = unichr(codepoint).encode('utf-16be')
++ utf16_string = chr(codepoint).encode('utf-16be')
+ if len(utf16_string) == 2:
+ (u0, l0) = utf16_string
+- result += r'\u%02X%02X' % (ord(u0), ord(l0))
++ result += br'\u%02X%02X' % (u0, l0)
+ else:
+ (u0, l0, u1, l1) = utf16_string
+- result += r'\u%02X%02X\u%02X%02X' % (ord(u0), ord(l0), ord(u1), ord(l1))
+- result += r'"'
++ result += br'\u%02X%02X\u%02X%02X' % (u0, l0, u1, l1)
++ result += b'"'
+ return result
+
+
+ def SkipLineComment(stream, comment_prefix='#'):
+ """Skips line comments from stream."""
+ for line in stream:
++ if isinstance(line, bytes):
++ if isinstance(comment_prefix, str):
++ comment_prefix = comment_prefix.encode('utf-8')
++ line_ending = b'\n'
++ else:
++ line_ending = '\n'
+ stripped_line = line.strip()
+ if stripped_line and not stripped_line.startswith(comment_prefix):
+- yield line.rstrip('\n')
++ yield line.rstrip(line_ending)
+
+
+ def ParseColumnStream(stream, num_column=None, delimiter=None):
+ """Returns parsed columns read from stream."""
+ if num_column is None:
+ for line in stream:
+- yield line.rstrip('\n').split(delimiter)
++ if isinstance(line, bytes):
++ line_ending = b'\n'
++ else:
++ line_ending = '\n'
++ yield line.rstrip(line_ending).split(delimiter)
+ else:
+ for line in stream:
+- yield line.rstrip('\n').split(delimiter)[:num_column]
++ if isinstance(line, bytes):
++ line_ending = b'\n'
++ else:
++ line_ending = '\n'
++ yield line.rstrip(line_ending).split(delimiter)[:num_column]
+
+
+ def SelectColumn(stream, column_index):
+@@ -172,5 +185,5 @@
+ grouper extends the last chunk to make it an n-element chunk by adding
+ appropriate value, but this returns truncated chunk.
+ """
+- for index in xrange(0, len(iterable), n):
++ for index in range(0, len(iterable), n):
+ yield iterable[index:index + n]
+--- /src/build_tools/codesign_mac.py
++++ /src/build_tools/codesign_mac.py
+@@ -46,17 +46,17 @@
+
+ def RunOrDie(command):
+ """Run the command, or die if it failed."""
+- print "Running: " + command
++ print("Running: " + command)
+ try:
+ output = subprocess.check_output(command, shell=True)
+- print >> sys.stderr, "=========="
+- print >> sys.stderr, "COMMAND: " + command
+- print >> sys.stderr, output
++ print("==========", file=sys.stderr)
++ print("COMMAND: " + command, file=sys.stderr)
++ print(output, file=sys.stderr)
+ except subprocess.CalledProcessError as e:
+- print >> sys.stderr, "=========="
+- print >> sys.stderr, "ERROR: " + command
+- print >> sys.stderr, e.output
+- print >> sys.stderr, "=========="
++ print("==========", file=sys.stderr)
++ print("ERROR: " + command, file=sys.stderr)
++ print(e.output, file=sys.stderr)
++ print("==========", file=sys.stderr)
+ sys.exit(1)
+
+
+@@ -119,18 +119,18 @@
+ (options, unused_args) = parser.parse_args()
+
+ if not options.target:
+- print "Error: --target should be specified."
+- print parser.print_help()
++ print("Error: --target should be specified.")
++ print(parser.print_help())
+ sys.exit(1)
+
+ return options
+
+
+ def DumpEnviron():
+- print "=== os.environ ==="
++ print("=== os.environ ===")
+ for key in sorted(os.environ):
+- print "%s = %s" % (key, os.getenv(key))
+- print "=================="
++ print("%s = %s" % (key, os.getenv(key)))
++ print("==================")
+
+
+ def main():
+--- /src/build_tools/copy_dll_and_symbol.py
++++ /src/build_tools/copy_dll_and_symbol.py
+@@ -38,7 +38,7 @@
+ import os
+ import shutil
+
+-from util import PrintErrorAndExit
++from .util import PrintErrorAndExit
+
+ def ParseOption():
+ """Parse command line options."""
+@@ -98,7 +98,7 @@
+ if _GetLastModifiedTime(src) <= target_file_mtime:
+ # Older file found. Ignore.
+ continue
+- print 'Copying %s to %s' % (src, target_file_abspath)
++ print('Copying %s to %s' % (src, target_file_abspath))
+ shutil.copy2(src, target_file_abspath)
+ break
+
+--- /src/build_tools/copy_file.py
++++ /src/build_tools/copy_file.py
+@@ -52,7 +52,7 @@
+ Args:
+ message: The error message to be printed to stderr.
+ """
+- print >>sys.stderr, message
++ print(message, file=sys.stderr)
+ sys.exit(1)
+
+
+--- /src/build_tools/copy_qt_frameworks_mac.py
++++ /src/build_tools/copy_qt_frameworks_mac.py
+@@ -41,9 +41,9 @@
+ import optparse
+ import os
+
+-from copy_file import CopyFiles
+-from util import PrintErrorAndExit
+-from util import RunOrDie
++from .copy_file import CopyFiles
++from .util import PrintErrorAndExit
++from .util import RunOrDie
+
+
+ def ParseOption():
+--- /src/build_tools/embed_file.py
++++ /src/build_tools/embed_file.py
+@@ -46,10 +46,10 @@
+
+ def _FormatAsUint64LittleEndian(s):
+ """Formats a string as uint64 value in little endian order."""
+- for _ in xrange(len(s), 8):
+- s += '\0'
++ for _ in range(len(s), 8):
++ s += b'\0'
+ s = s[::-1] # Reverse the string
+- return '0x%s' % binascii.b2a_hex(s)
++ return b'0x%b' % binascii.b2a_hex(s)
+
+
+ def main():
+@@ -57,30 +57,30 @@
+ with open(opts.input, 'rb') as infile:
+ with open(opts.output, 'wb') as outfile:
+ outfile.write(
+- '#ifdef MOZC_EMBEDDED_FILE_%(name)s\n'
+- '#error "%(name)s was already included or defined elsewhere"\n'
+- '#else\n'
+- '#define MOZC_EMBEDDED_FILE_%(name)s\n'
+- 'const uint64 %(name)s_data[] = {\n'
+- % {'name': opts.name})
++ b'#ifdef MOZC_EMBEDDED_FILE_%(name)b\n'
++ b'#error "%(name)b was already included or defined elsewhere"\n'
++ b'#else\n'
++ b'#define MOZC_EMBEDDED_FILE_%(name)b\n'
++ b'const uint64 %(name)b_data[] = {\n'
++ % {b'name': opts.name.encode('utf-8')})
+
+ while True:
+ chunk = infile.read(8)
+ if not chunk:
+ break
+- outfile.write(' ')
++ outfile.write(b' ')
+ outfile.write(_FormatAsUint64LittleEndian(chunk))
+- outfile.write(',\n')
++ outfile.write(b',\n')
+
+ outfile.write(
+- '};\n'
+- 'const EmbeddedFile %(name)s = {\n'
+- ' %(name)s_data,\n'
+- ' %(size)d,\n'
+- '};\n'
+- '#endif // MOZC_EMBEDDED_FILE_%(name)s\n'
+- % {'name': opts.name,
+- 'size': os.stat(opts.input).st_size})
++ b'};\n'
++ b'const EmbeddedFile %(name)b = {\n'
++ b' %(name)b_data,\n'
++ b' %(size)d,\n'
++ b'};\n'
++ b'#endif // MOZC_EMBEDDED_FILE_%(name)b\n'
++ % {b'name': opts.name.encode('utf-8'),
++ b'size': os.stat(opts.input).st_size})
+
+
+ if __name__ == '__main__':
+--- /src/build_tools/embed_pathname.py
++++ /src/build_tools/embed_pathname.py
+@@ -28,7 +28,7 @@
+ # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+
+-"""A script to embed the given (relative) path name to C/C++ characters array.
++r"""A script to embed the given (relative) path name to C/C++ characters array.
+
+ Example:
+ ./embed_pathname.py --path_to_be_embedded=d:\data\mozc
+@@ -53,7 +53,7 @@
+
+ (options, unused_args) = parser.parse_args()
+ if not all(vars(options).values()):
+- print parser.print_help()
++ print(parser.print_help())
+ sys.exit(1)
+
+ return options
+@@ -63,7 +63,7 @@
+ opt = ParseOption()
+ path = os.path.abspath(opt.path_to_be_embedded)
+ # TODO(yukawa): Consider the case of non-ASCII characters.
+- escaped_path = path.encode('string-escape')
++ escaped_path = path.replace('\\', '\\\\')
+ with open(opt.output, 'w') as output_file:
+ output_file.write(
+ 'const char %s[] = "%s";\n' % (opt.constant_name, escaped_path))
+--- /src/build_tools/ensure_gyp_module_path.py
++++ /src/build_tools/ensure_gyp_module_path.py
+@@ -48,7 +48,7 @@
+
+ (options, _) = parser.parse_args()
+ if not options.expected:
+- print parser.print_help()
++ print(parser.print_help())
+ sys.exit(1)
+
+ return options
+@@ -59,20 +59,20 @@
+ opt = ParseOption()
+ expected_path = os.path.abspath(opt.expected)
+ if not os.path.exists(expected_path):
+- print '%s does not exist.' % expected_path
++ print('%s does not exist.' % expected_path)
+ sys.exit(1)
+
+ try:
+ import gyp # NOLINT
+ except ImportError as e:
+- print 'import gyp failed: %s' % e
++ print('import gyp failed: %s' % e)
+ sys.exit(1)
+
+ actual_path = os.path.abspath(gyp.__path__[0])
+ if expected_path != actual_path:
+- print 'Unexpected gyp module is loaded on this environment.'
+- print ' expected: %s' % expected_path
+- print ' actual : %s' % actual_path
++ print('Unexpected gyp module is loaded on this environment.')
++ print(' expected: %s' % expected_path)
++ print(' actual : %s' % actual_path)
+ sys.exit(1)
+
+ if __name__ == '__main__':
+--- /src/build_tools/gen_win32_resource_header.py
++++ /src/build_tools/gen_win32_resource_header.py
+@@ -39,7 +39,7 @@
+ __author__ = "yukawa"
+
+ import logging
+-import mozc_version
++from . import mozc_version
+ import optparse
+ import os
+ import sys
+--- /src/build_tools/mozc_version.py
++++ /src/build_tools/mozc_version.py
+@@ -94,7 +94,7 @@
+ last_digit = TARGET_PLATFORM_TO_DIGIT.get(target_platform, None)
+ if last_digit is None:
+ logging.critical('target_platform %s is invalid. Accetable ones are %s',
+- target_platform, TARGET_PLATFORM_TO_DIGIT.keys())
++ target_platform, list(TARGET_PLATFORM_TO_DIGIT.keys()))
+ sys.exit(1)
+
+ if not revision:
+@@ -314,13 +314,14 @@
+ self._properties = {}
+ if not os.path.isfile(path):
+ return
+- for line in open(path):
+- matchobj = re.match(r'(\w+)=(.*)', line.strip())
+- if matchobj:
+- var = matchobj.group(1)
+- val = matchobj.group(2)
+- if var not in self._properties:
+- self._properties[var] = val
++ with open(path) as file:
++ for line in file:
++ matchobj = re.match(r'(\w+)=(.*)', line.strip())
++ if matchobj:
++ var = matchobj.group(1)
++ val = matchobj.group(2)
++ if var not in self._properties:
++ self._properties[var] = val
+
+ # Check mandatory properties.
+ for key in VERSION_PROPERTIES:
diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch
new file mode 100644
index 000000000000..456e8368049a
--- /dev/null
+++ b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_2.patch
@@ -0,0 +1,600 @@
+https://github.com/google/mozc/issues/462
+
+--- /src/build_tools/redirect.py
++++ /src/build_tools/redirect.py
+@@ -58,14 +58,15 @@
+ process = subprocess.Popen(sys.argv, stdout=subprocess.PIPE,
+ universal_newlines=True)
+ except:
+- print '=========='
+- print ' ERROR: %s' % ' '.join(sys.argv)
+- print '=========='
++ print('==========')
++ print(' ERROR: %s' % ' '.join(sys.argv))
++ print('==========')
+ raise
+ (stdout_content, _) = process.communicate()
+ # Write the stdout content to the output file.
+ output_file = open(output_file_name, 'w')
+ output_file.write(stdout_content)
++ output_file.close()
+ return process.wait()
+
+ if __name__ == '__main__':
+--- /src/build_tools/run_after_chdir.py
++++ /src/build_tools/run_after_chdir.py
+@@ -57,7 +57,7 @@
+ sys.argv.insert(0, sys.executable) # Inject the python interpreter path.
+ # We don't capture stdout and stderr from Popen. The output will just
+ # be emitted to a terminal or console.
+- print sys.argv
++ print(sys.argv)
+ sys.exit(subprocess.call(sys.argv))
+
+ if __name__ == '__main__':
+--- /src/build_tools/serialized_string_array_builder.py
++++ /src/build_tools/serialized_string_array_builder.py
+@@ -58,11 +58,11 @@
+ f.write(struct.pack('<I', array_size))
+
+ # Offset and length array of (4 + 4) * array_size bytes.
+- for i in xrange(array_size):
++ for i in range(array_size):
+ f.write(struct.pack('<I', offsets[i]))
+ f.write(struct.pack('<I', lengths[i]))
+
+ # Strings chunk.
+- for i in xrange(array_size):
++ for i in range(array_size):
+ f.write(strings[i])
+- f.write('\0')
++ f.write(b'\0')
+--- /src/build_tools/test_tools/gtest_report.py
++++ /src/build_tools/test_tools/gtest_report.py
+@@ -36,9 +36,9 @@
+
+ __author__ = "nona"
+
+-import cStringIO as StringIO
++import io
+ import logging
+-from xml.etree import cElementTree as ElementTree
++from xml.etree import ElementTree
+
+
+ class Failure(object):
+@@ -87,13 +87,13 @@
+ """Returns summarized error report text."""
+ if self.fail_num == 0:
+ return ''
+- output = StringIO.StringIO()
++ output = io.StringIO()
+ for testcase in self.testcases:
+ if not testcase.failures:
+ continue
+- print >>output, '%s.%s:' % (self.name, testcase.name)
++ print('%s.%s:' % (self.name, testcase.name), file=output)
+ for failure in testcase.failures:
+- print >>output, failure.contents.encode('utf-8')
++ print(failure.contents.encode('utf-8'), file=output)
+ return output.getvalue()
+
+ @classmethod
+--- /src/build_tools/test_tools/test_launcher.py
++++ /src/build_tools/test_tools/test_launcher.py
+@@ -101,11 +101,11 @@
+ time.sleep(1)
+ try:
+ shutil.rmtree(self._path)
+- except OSError, e:
++ except OSError as e:
+ logging.error('Failed to remove %s. error: %s', self._path, e)
+
+
+-def _ExecuteTest((command, gtest_report_dir)):
++def _ExecuteTest(args):
+ """Executes tests with specified Test command.
+
+ Args:
+@@ -122,6 +122,7 @@
+ module, which is used in multiprocessing module.
+ (http://docs.python.org/library/pickle.html)
+ """
++ (command, gtest_report_dir) = args
+ binary = command[0]
+ binary_filename = os.path.basename(binary)
+ tmp_dir = tempfile.mkdtemp()
+--- /src/build_tools/tweak_data.py
++++ /src/build_tools/tweak_data.py
+@@ -55,7 +55,7 @@
+ The value for the variable if the variable is defined in the
+ environment. Otherwise original string is returned.
+ """
+- if environment.has_key(matchobj.group(1)):
++ if matchobj.group(1) in environment:
+ return environment[matchobj.group(1)]
+ return matchobj.group(0)
+
+--- /src/build_tools/tweak_info_plist.py
++++ /src/build_tools/tweak_info_plist.py
+@@ -42,8 +42,8 @@
+ import logging
+ import optparse
+ import sys
+-import mozc_version
+-import tweak_data
++from . import mozc_version
++from . import tweak_data
+
+ _COPYRIGHT_YEAR = datetime.date.today().year
+
+@@ -81,7 +81,7 @@
+
+ version = mozc_version.MozcVersion(options.version_file)
+
+- copyright_message = (u'© %d Google Inc.' % _COPYRIGHT_YEAR).encode('utf-8')
++ copyright_message = ('© %d Google Inc.' % _COPYRIGHT_YEAR).encode('utf-8')
+ long_version = version.GetVersionString()
+ short_version = version.GetVersionInFormat('@MAJOR@.@MINOR@.@BUILD@')
+
+--- /src/build_tools/tweak_info_plist_strings.py
++++ /src/build_tools/tweak_info_plist_strings.py
+@@ -40,7 +40,7 @@
+ import logging
+ import optparse
+ import sys
+-import tweak_data
++from . import tweak_data
+
+ _COPYRIGHT_YEAR = datetime.date.today().year
+
+@@ -77,7 +77,7 @@
+ if options.branding == 'GoogleJapaneseInput':
+ variables = {
+ 'CF_BUNDLE_NAME_EN': 'Google Japanese Input',
+- 'CF_BUNDLE_NAME_JA': u'Google 日本語入力'.encode('utf-8'),
++ 'CF_BUNDLE_NAME_JA': 'Google 日本語入力'.encode('utf-8'),
+ 'NS_HUMAN_READABLE_COPYRIGHT': copyright_message,
+ 'INPUT_MODE_ANNOTATION': 'Google',
+ }
+--- /src/build_tools/tweak_macinstaller_script.py
++++ /src/build_tools/tweak_macinstaller_script.py
+@@ -39,7 +39,7 @@
+ import logging
+ import optparse
+
+-import mozc_version
++from . import mozc_version
+
+
+ def _ReplaceVariables(data, environment):
+--- /src/build_tools/tweak_pkgproj.py
++++ /src/build_tools/tweak_pkgproj.py
+@@ -45,7 +45,7 @@
+ import os
+ import plistlib
+ import re
+-import mozc_version
++from . import mozc_version
+
+ from os import path
+
+@@ -71,7 +71,7 @@
+ The value for the variable if the variable is defined in the
+ environment. Otherwise original string is returned.
+ """
+- if environment.has_key(matchobj.group(1)):
++ if matchobj.group(1) in environment:
+ return environment[matchobj.group(1)]
+ return matchobj.group(0)
+
+--- /src/build_tools/util.py
++++ /src/build_tools/util.py
+@@ -73,11 +73,11 @@
+ return 1
+
+
+-class RunOrDieError(StandardError):
++class RunOrDieError(Exception):
+ """The exception class for RunOrDie."""
+
+ def __init__(self, message):
+- StandardError.__init__(self, message)
++ Exception.__init__(self, message)
+
+
+ def RunOrDie(argv):
+@@ -105,7 +105,7 @@
+ return # Do nothing if not exist.
+ if IsWindows():
+ # Read-only files cannot be deleted on Windows.
+- os.chmod(file_name, 0700)
++ os.chmod(file_name, 0o700)
+ logging.debug('Removing file: %s', file_name)
+ os.unlink(file_name)
+
+--- /src/build_tools/zlib_util.py
++++ /src/build_tools/zlib_util.py
+@@ -58,7 +58,7 @@
+
+ def main():
+ if len(sys.argv) != 4:
+- print >>sys.stderr, 'Invalid arguments'
++ print('Invalid arguments', file=sys.stderr)
+ return
+ if sys.argv[1] == 'compress':
+ Compress(sys.argv[2], sys.argv[3])
+@@ -66,7 +66,7 @@
+ if sys.argv[1] == 'decompress':
+ Decompress(sys.argv[2], sys.argv[3])
+ return
+- print >>sys.stderr, 'Unknown command:', sys.argv[1]
++ print('Unknown command:', sys.argv[1], file=sys.stderr)
+
+
+ if __name__ == '__main__':
+--- /src/composer/internal/gen_typing_model.py
++++ /src/composer/internal/gen_typing_model.py
+@@ -54,14 +54,13 @@
+ __author__ = "noriyukit"
+
+ import bisect
+-import codecs
+ import collections
+ import optparse
+ import struct
+
+ UNDEFINED_COST = -1
+-MAX_UINT16 = struct.unpack('H', '\xFF\xFF')[0]
+-MAX_UINT8 = struct.unpack('B', '\xFF')[0]
++MAX_UINT16 = struct.unpack('H', b'\xFF\xFF')[0]
++MAX_UINT8 = struct.unpack('B', b'\xFF')[0]
+
+
+ def ParseArgs():
+@@ -113,7 +112,7 @@
+ sorted_values = list(sorted(set(values)))
+ mapping_table = sorted_values[0]
+ mapping_table_size_without_special_value = mapping_table_size - 1
+- span = len(sorted_values) / (mapping_table_size_without_special_value - 1)
++ span = len(sorted_values) // (mapping_table_size_without_special_value - 1)
+ mapping_table = [sorted_values[i * span]
+ for i
+ in range(0, mapping_table_size_without_special_value - 1)]
+@@ -150,7 +149,7 @@
+
+ def GetValueTable(unique_characters, mapping_table, dictionary):
+ result = []
+- for key, value in dictionary.iteritems():
++ for key, value in dictionary.items():
+ index = GetIndexFromKey(unique_characters, key)
+ while len(result) <= index:
+ result.append(len(mapping_table) - 1)
+@@ -167,13 +166,13 @@
+ romaji_transition_cost)
+ with open(output_path, 'wb') as f:
+ f.write(struct.pack('<I', len(unique_characters)))
+- f.write(''.join(unique_characters))
++ f.write(''.join(unique_characters).encode('utf-8'))
+ offset = 4 + len(unique_characters)
+
+ # Add padding to place value list size at 4-byte boundary.
+ if offset % 4:
+ padding_size = 4 - offset % 4
+- f.write('\x00' * padding_size)
++ f.write(b'\x00' * padding_size)
+ offset += padding_size
+
+ f.write(struct.pack('<I', len(value_list)))
+@@ -184,7 +183,7 @@
+ # Add padding to place mapping_table at 4-byte boundary.
+ if offset % 4:
+ padding_size = 4 - offset % 4
+- f.write('\x00' * padding_size)
++ f.write(b'\x00' * padding_size)
+ offset += padding_size
+
+ for v in mapping_table:
+@@ -198,7 +197,8 @@
+ # - trigram['vw']['x'] = -500 * log(P(x | 'vw'))
+ unigram = {}
+ trigram = collections.defaultdict(dict)
+- for line in codecs.open(options.input_path, 'r', encoding='utf-8'):
++ input_file = open(options.input_path, 'r', encoding='utf-8')
++ for line in input_file:
+ line = line.rstrip()
+ ngram, cost = line.split('\t')
+ cost = int(cost)
+@@ -206,6 +206,7 @@
+ unigram[ngram] = cost
+ else:
+ trigram[ngram[:-1]][ngram[-1]] = cost
++ input_file.close()
+
+ # Calculate ngram-related cost for each 'vw' and 'x':
+ # -500 * log( P('x' | 'vw') / P('x') )
+--- /src/converter/gen_boundary_data.py
++++ /src/converter/gen_boundary_data.py
+@@ -70,7 +70,8 @@
+ def LoadPatterns(file):
+ prefix = []
+ suffix = []
+- for line in open(file, 'r'):
++ fh = open(file, 'r')
++ for line in fh:
+ if len(line) <= 1 or line[0] == '#':
+ continue
+ fields = line.split()
+@@ -84,8 +85,9 @@
+ elif label == 'SUFFIX':
+ suffix.append([re.compile(PatternToRegexp(feature)), cost])
+ else:
+- print 'format error %s' % (line)
++ print('format error %s' % (line))
+ sys.exit(0)
++ fh.close()
+ return (prefix, suffix)
+
+
+@@ -100,19 +102,23 @@
+
+ def LoadFeatures(filename):
+ features = []
+- for line in open(filename, 'r'):
++ fh = open(filename, 'r')
++ for line in fh:
+ fields = line.split()
+ features.append(fields[1])
++ fh.close()
+ return features
+
+
+ def CountSpecialPos(filename):
+ count = 0
+- for line in open(filename, 'r'):
++ fh = open(filename, 'r')
++ for line in fh:
+ line = line.rstrip()
+ if not line or line[0] == '#':
+ continue
+ count += 1
++ fh.close()
+ return count
+
+
+@@ -141,7 +147,7 @@
+ f.write(struct.pack('<H', GetCost(prefix, feature)))
+ f.write(struct.pack('<H', GetCost(suffix, feature)))
+
+- for _ in xrange(num_special_pos):
++ for _ in range(num_special_pos):
+ f.write(struct.pack('<H', 0))
+ f.write(struct.pack('<H', 0))
+
+--- /src/converter/gen_quality_regression_test_data.py
++++ /src/converter/gen_quality_regression_test_data.py
+@@ -84,7 +84,7 @@
+ else _ENABLED)
+ id = issue.attributes['id'].value
+ target = GetText(issue.getElementsByTagName('target'))
+- for detail in issue.getElementsByTagName(u'detail'):
++ for detail in issue.getElementsByTagName('detail'):
+ fields = []
+ fields.append('mozcsu_%s' % id)
+ for key in ('reading', 'output', 'actionStatus', 'rank', 'accuracy'):
+@@ -104,19 +104,19 @@
+
+ def GenerateHeader(files):
+ try:
+- print 'namespace mozc{'
+- print 'struct TestCase {'
+- print ' const bool enabled;'
+- print ' const char *tsv;'
+- print '} kTestData[] = {'
++ print('namespace mozc{')
++ print('struct TestCase {')
++ print(' const bool enabled;')
++ print(' const char *tsv;')
++ print('} kTestData[] = {')
+ for file in files:
+ for enabled, line in ParseFile(file):
+- print ' {%s, "%s"},' % (enabled, EscapeString(line))
+- print ' {false, nullptr},'
+- print '};'
+- print '} // namespace mozc'
++ print(' {%s, "%s"},' % (enabled, EscapeString(line)))
++ print(' {false, nullptr},')
++ print('};')
++ print('} // namespace mozc')
+ except:
+- print 'cannot open %s' % (file)
++ print('cannot open %s' % (file))
+ sys.exit(1)
+
+
+--- /src/converter/gen_segmenter_code.py
++++ /src/converter/gen_segmenter_code.py
+@@ -54,18 +54,22 @@
+ pos = {}
+ max_id = 0
+
+- for line in open(id_file, "r"):
++ fh = open(id_file, "r")
++ for line in fh:
+ fields = line.split()
+ pos[fields[1]] = fields[0]
+ max_id = max(int(fields[0]), max_id)
++ fh.close()
+
+ max_id = max_id + 1
+- for line in open(special_pos_file, "r"):
++ fh = open(special_pos_file, "r")
++ for line in fh:
+ if len(line) <= 1 or line[0] == '#':
+ continue
+ fields = line.split()
+ pos[fields[0]] = ("%d" % max_id)
+ max_id = max_id + 1
++ fh.close()
+
+ return pos
+
+@@ -79,8 +83,7 @@
+ pat = re.compile(PatternToRegexp(pattern))
+ min = -1;
+ max = -1;
+- keys = pos.keys()
+- keys.sort()
++ keys = sorted(pos.keys())
+
+ range = []
+
+@@ -107,7 +110,7 @@
+ tmp.append("(%s >= %s && %s <= %s)" % (name, r[0], name, r[1]))
+
+ if len(tmp) == 0:
+- print "FATAL: No rule fiind %s" % (pattern)
++ print("FATAL: No rule fiind %s" % (pattern))
+ sys.exit(-1)
+
+ return " || ".join(tmp)
+@@ -115,19 +118,21 @@
+ def main():
+ pos = ReadPOSID(sys.argv[1], sys.argv[2])
+
+- print HEADER % (len(pos.keys()), len(pos.keys()))
++ print(HEADER % (len(pos.keys()), len(pos.keys())))
+
+- for line in open(sys.argv[3], "r"):
++ fh = open(sys.argv[3], "r")
++ for line in fh:
+ if len(line) <= 1 or line[0] == '#':
+ continue
+ (l, r, result) = line.split()
+ result = result.lower()
+ lcond = GetRange(pos, l, "rid") or "true";
+ rcond = GetRange(pos, r, "lid") or "true";
+- print " // %s %s %s" % (l, r, result)
+- print " if ((%s) && (%s)) { return %s; }" % (lcond, rcond, result)
++ print(" // %s %s %s" % (l, r, result))
++ print(" if ((%s) && (%s)) { return %s; }" % (lcond, rcond, result))
++ fh.close()
+
+- print FOOTER
++ print(FOOTER)
+
+ if __name__ == "__main__":
+ main()
+--- /src/data_manager/gen_connection_data.py
++++ /src/data_manager/gen_connection_data.py
+@@ -32,8 +32,7 @@
+
+ __author__ = "hidehiko"
+
+-import cStringIO as StringIO
+-import itertools
++import io
+ import logging
+ import optparse
+ import os
+@@ -45,7 +44,7 @@
+ INVALID_COST = 30000
+ INVALID_1BYTE_COST = 255
+ RESOLUTION_FOR_1BYTE = 64
+-FILE_MAGIC = '\xAB\xCD'
++FILE_MAGIC = b'\xAB\xCD'
+
+ FALSE_VALUES = ['f', 'false', '0']
+ TRUE_VALUES = ['t', 'true', '1']
+@@ -79,28 +78,28 @@
+ # The result is a square matrix.
+ mat_size = pos_size + special_pos_size
+
+- matrix = [[0] * mat_size for _ in xrange(mat_size)]
++ matrix = [[0] * mat_size for _ in range(mat_size)]
+ with open(text_connection_file) as stream:
+ stream = code_generator_util.SkipLineComment(stream)
+ # The first line contains the matrix column/row size.
+- size = stream.next().rstrip()
++ size = next(stream).rstrip()
+ assert (int(size) == pos_size), '%s != %d' % (size, pos_size)
+
+ for array_index, cost in enumerate(stream):
+ cost = int(cost.rstrip())
+- rid = array_index / pos_size
++ rid = array_index // pos_size
+ lid = array_index % pos_size
+ if rid == 0 and lid == 0:
+ cost = 0
+ matrix[rid][lid] = cost
+
+ # Fill INVALID_COST in matrix elements for special POS.
+- for rid in xrange(pos_size, mat_size):
+- for lid in xrange(1, mat_size): # Skip EOS
++ for rid in range(pos_size, mat_size):
++ for lid in range(1, mat_size): # Skip EOS
+ matrix[rid][lid] = INVALID_COST
+
+- for lid in xrange(pos_size, mat_size):
+- for rid in xrange(1, mat_size): # Skip BOS
++ for lid in range(pos_size, mat_size):
++ for rid in range(1, mat_size): # Skip BOS
+ matrix[rid][lid] = INVALID_COST
+
+ return matrix
+@@ -116,7 +115,7 @@
+ # Heuristically, we do not compress INVALID_COST.
+ continue
+ m[cost] = m.get(cost, 0) + 1
+- mode_value = max(m.iteritems(), key=lambda (_, count): count)[0]
++ mode_value = max(m.items(), key=lambda x: x[1])[0]
+ result.append(mode_value)
+ return result
+
+@@ -126,8 +125,8 @@
+ # list, and fill None into the matrix if it equals to the corresponding
+ # mode value.
+ assert len(matrix) == len(mode_value_list)
+- for row, mode_value in itertools.izip(matrix, mode_value_list):
+- for index in xrange(len(row)):
++ for row, mode_value in zip(matrix, mode_value_list):
++ for index in range(len(row)):
+ if row[index] == mode_value:
+ row[index] = None
+
+@@ -179,7 +178,7 @@
+ resolution = RESOLUTION_FOR_1BYTE
+ else:
+ resolution = 1
+- stream = StringIO.StringIO()
++ stream = io.BytesIO()
+
+ # Output header.
+ stream.write(FILE_MAGIC)
+@@ -194,7 +193,7 @@
+
+ # 4 bytes alignment.
+ if len(mode_value_list) % 2:
+- stream.write('\x00\x00')
++ stream.write(b'\x00\x00')
+
+ # Process each row:
+ for row in matrix:
+@@ -218,7 +217,7 @@
+ if cost == INVALID_COST:
+ cost = INVALID_1BYTE_COST
+ else:
+- cost /= resolution
++ cost //= resolution
+ assert cost != INVALID_1BYTE_COST
+ values.append(cost)
+
+@@ -237,7 +236,7 @@
+ values_size = len(values) * 2
+
+ # Output the bits for a row.
+- stream.write(struct.pack('<HH', len(compact_bits) / 8, values_size))
++ stream.write(struct.pack('<HH', len(compact_bits) // 8, values_size))
+ OutputBitList(chunk_bits, stream)
+ OutputBitList(compact_bits, stream)
+ if use_1byte_cost:
diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch
new file mode 100644
index 000000000000..a5c5a2dc8038
--- /dev/null
+++ b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_3.patch
@@ -0,0 +1,583 @@
+https://github.com/google/mozc/issues/462
+
+--- /src/dictionary/gen_pos_map.py
++++ /src/dictionary/gen_pos_map.py
+@@ -39,7 +39,7 @@
+ from build_tools import code_generator_util
+
+
+-HEADER = """// Copyright 2009 Google Inc. All Rights Reserved.
++HEADER = b"""// Copyright 2009 Google Inc. All Rights Reserved.
+ // Author: keni
+
+ #ifndef MOZC_DICTIONARY_POS_MAP_H_
+@@ -48,13 +48,13 @@
+ // POS conversion rules
+ const POSMap kPOSMap[] = {
+ """
+-FOOTER = """};
++FOOTER = b"""};
+
+ #endif // MOZC_DICTIONARY_POS_MAP_H_
+ """
+
+ def ParseUserPos(user_pos_file):
+- with open(user_pos_file, 'r') as stream:
++ with open(user_pos_file, 'rb') as stream:
+ stream = code_generator_util.SkipLineComment(stream)
+ stream = code_generator_util.ParseColumnStream(stream, num_column=2)
+ return dict((key, enum_value) for key, enum_value in stream)
+@@ -64,7 +64,7 @@
+ user_pos_map = ParseUserPos(user_pos_file)
+
+ result = {}
+- with open(third_party_pos_map_file, 'r') as stream:
++ with open(third_party_pos_map_file, 'rb') as stream:
+ stream = code_generator_util.SkipLineComment(stream)
+ for columns in code_generator_util.ParseColumnStream(stream, num_column=2):
+ third_party_pos_name, mozc_pos = (columns + [None])[:2]
+@@ -78,7 +78,7 @@
+ result[third_party_pos_name] = mozc_pos
+
+ # Create mozc_pos to mozc_pos map.
+- for key, value in user_pos_map.iteritems():
++ for key, value in user_pos_map.items():
+ if key in result:
+ assert (result[key] == value)
+ continue
+@@ -94,10 +94,10 @@
+ if value is None:
+ # Invalid PosType.
+ value = (
+- 'static_cast< ::mozc::user_dictionary::UserDictionary::PosType>(-1)')
++ b'static_cast< ::mozc::user_dictionary::UserDictionary::PosType>(-1)')
+ else:
+- value = '::mozc::user_dictionary::UserDictionary::' + value
+- output.write(' { %s, %s },\n' % (key, value))
++ value = b'::mozc::user_dictionary::UserDictionary::' + value
++ output.write(b' { %s, %s },\n' % (key, value))
+ output.write(FOOTER)
+
+
+@@ -121,7 +121,7 @@
+ pos_map = GeneratePosMap(options.third_party_pos_map_file,
+ options.user_pos_file)
+
+- with open(options.output, 'w') as stream:
++ with open(options.output, 'wb') as stream:
+ OutputPosMap(pos_map, stream)
+
+
+--- /src/dictionary/gen_pos_rewrite_rule.py
++++ /src/dictionary/gen_pos_rewrite_rule.py
+@@ -46,29 +46,34 @@
+
+
+ def LoadRewriteMapRule(filename):
+- fh = open(filename)
++ fh = open(filename, 'rb')
+ rule = []
+ for line in fh:
+- line = line.rstrip('\n')
+- if not line or line.startswith('#'):
++ line = line.rstrip(b'\n')
++ if not line or line.startswith(b'#'):
+ continue
+ fields = line.split()
+ rule.append([fields[0], fields[1]])
++ fh.close()
+ return rule
+
+
+ def ReadPOSID(id_file, special_pos_file):
+ pos_list = []
+
+- for line in open(id_file, 'r'):
++ fh = open(id_file, 'rb')
++ for line in fh:
+ fields = line.split()
+ pos_list.append(fields[1])
++ fh.close()
+
+- for line in open(special_pos_file, 'r'):
+- if len(line) <= 1 or line[0] == '#':
++ fh = open(special_pos_file, 'rb')
++ for line in fh:
++ if len(line) <= 1 or line[0:1] == b'#':
+ continue
+ fields = line.split()
+ pos_list.append(fields[0])
++ fh.close()
+
+ return pos_list
+
+@@ -112,7 +117,7 @@
+ ids.append(id)
+
+ with open(opts.output, 'wb') as f:
+- f.write(''.join(chr(id) for id in ids))
++ f.write(''.join(chr(id) for id in ids).encode('utf-8'))
+
+
+ if __name__ == '__main__':
+--- /src/dictionary/gen_suffix_data.py
++++ /src/dictionary/gen_suffix_data.py
+@@ -52,10 +52,10 @@
+ opts = _ParseOptions()
+
+ result = []
+- with open(opts.input, 'r') as stream:
++ with open(opts.input, 'rb') as stream:
+ for line in stream:
+- line = line.rstrip('\r\n')
+- fields = line.split('\t')
++ line = line.rstrip(b'\r\n')
++ fields = line.split(b'\t')
+ key = fields[0]
+ lid = int(fields[1])
+ rid = int(fields[2])
+@@ -63,7 +63,7 @@
+ value = fields[4]
+
+ if key == value:
+- value = ''
++ value = b''
+
+ result.append((key, value, lid, rid, cost))
+
+--- /src/dictionary/gen_user_pos_data.py
++++ /src/dictionary/gen_user_pos_data.py
+@@ -64,7 +64,7 @@
+ f.write(struct.pack('<H', conjugation_id))
+
+ serialized_string_array_builder.SerializeToFile(
+- sorted(string_index.iterkeys()), output_string_array)
++ sorted(x.encode('utf-8') for x in string_index.keys()), output_string_array)
+
+
+ def ParseOptions():
+@@ -100,7 +100,7 @@
+
+ if options.output_pos_list:
+ serialized_string_array_builder.SerializeToFile(
+- [pos for (pos, _) in user_pos.data], options.output_pos_list)
++ [pos.encode('utf-8') for (pos, _) in user_pos.data], options.output_pos_list)
+
+
+ if __name__ == '__main__':
+--- /src/dictionary/gen_zip_code_seed.py
++++ /src/dictionary/gen_zip_code_seed.py
+@@ -83,7 +83,7 @@
+ address = unicodedata.normalize('NFKC', self.address)
+ line = '\t'.join([zip_code, '0', '0', str(ZIP_CODE_COST),
+ address, ZIP_CODE_LABEL])
+- print line.encode('utf-8')
++ print(line.encode('utf-8'))
+
+
+ def ProcessZipCodeCSV(file_name):
+@@ -105,26 +105,26 @@
+
+ def ReadZipCodeEntries(zip_code, level1, level2, level3):
+ """Read zip code entries."""
+- return [ZipEntry(zip_code, u''.join([level1, level2, town]))
++ return [ZipEntry(zip_code, ''.join([level1, level2, town]))
+ for town in ParseTownName(level3)]
+
+
+ def ReadJigyosyoEntry(zip_code, level1, level2, level3, name):
+ """Read jigyosyo entry."""
+ return ZipEntry(zip_code,
+- u''.join([level1, level2, level3, u' ', name]))
++ ''.join([level1, level2, level3, ' ', name]))
+
+
+ def ParseTownName(level3):
+ """Parse town name."""
+- if level3.find(u'以下に掲載がない場合') != -1:
++ if level3.find('以下に掲載がない場合') != -1:
+ return ['']
+
+ assert CanParseAddress(level3), ('failed to be merged %s'
+ % level3.encode('utf-8'))
+
+ # We ignore additional information here.
+- level3 = re.sub(u'(.*)', u'', level3, re.U)
++ level3 = re.sub('(.*)', '', level3, re.U)
+
+ # For 地割, we have these cases.
+ # XX1地割
+@@ -134,7 +134,7 @@
+ # XX第1地割、XX第2地割、
+ # XX第1地割〜XX第2地割、
+ # We simply use XX for them.
+- chiwari_match = re.match(u'(\D*?)第?\d+地割.*', level3, re.U)
++ chiwari_match = re.match('(\D*?)第?\d+地割.*', level3, re.U)
+ if chiwari_match:
+ town = chiwari_match.group(1)
+ return [town]
+@@ -144,21 +144,21 @@
+ # -> XX町YY and (XX町)ZZ
+ # YY、ZZ
+ # -> YY and ZZ
+- chou_match = re.match(u'(.*町)?(.*)', level3, re.U)
++ chou_match = re.match('(.*町)?(.*)', level3, re.U)
+ if chou_match:
+- chou = u''
++ chou = ''
+ if chou_match.group(1):
+ chou = chou_match.group(1)
+ rests = chou_match.group(2)
+- return [chou + rest for rest in rests.split(u'、')]
++ return [chou + rest for rest in rests.split('、')]
+
+ return [level3]
+
+
+ def CanParseAddress(address):
+ """Return true for valid address."""
+- return (address.find(u'(') == -1 or
+- address.find(u')') != -1)
++ return (address.find('(') == -1 or
++ address.find(')') != -1)
+
+
+ def ParseOptions():
+--- /src/dictionary/zip_code_util.py
++++ /src/dictionary/zip_code_util.py
+@@ -86,11 +86,11 @@
+
+
+ _SPECIAL_CASES = [
+- SpecialMergeZip(u'5900111', u'大阪府', u'堺市中区', [u'三原台']),
+- SpecialMergeZip(u'8710046', u'大分県', u'中津市',
+- [u'金谷', u'西堀端', u'東堀端', u'古金谷']),
+- SpecialMergeZip(u'9218046', u'石川県', u'金沢市',
+- [u'大桑町', u'三小牛町']),
++ SpecialMergeZip('5900111', '大阪府', '堺市中区', ['三原台']),
++ SpecialMergeZip('8710046', '大分県', '中津市',
++ ['金谷', '西堀端', '東堀端', '古金谷']),
++ SpecialMergeZip('9218046', '石川県', '金沢市',
++ ['大桑町', '三小牛町']),
+ ]
+
+
+--- /src/gui/character_pad/data/gen_cp932_map.py
++++ /src/gui/character_pad/data/gen_cp932_map.py
+@@ -32,7 +32,6 @@
+
+ import re
+ import sys
+-import string
+
+ kUnicodePat = re.compile(r'0x[0-9A-Fa-f]{2,4}')
+ def IsValidUnicode(n):
+@@ -42,28 +41,29 @@
+ fh = open(sys.argv[1])
+ result = {}
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+- array = string.split(line)
++ array = line.split()
+ sjis = array[0]
+ ucs2 = array[1]
+ if eval(sjis) < 32 or not IsValidUnicode(ucs2):
+ continue
+ result.setdefault(ucs2, sjis)
++ fh.close()
+
+ keys = sorted(result.keys())
+
+- print "struct CP932MapData {"
+- print " unsigned int ucs4;"
+- print " unsigned short int sjis;"
+- print "};"
+- print ""
+- print "static const size_t kCP932MapDataSize = %d;" % (len(keys))
+- print "static const CP932MapData kCP932MapData[] = {"
++ print("struct CP932MapData {")
++ print(" unsigned int ucs4;")
++ print(" unsigned short int sjis;")
++ print("};")
++ print("")
++ print("static const size_t kCP932MapDataSize = %d;" % (len(keys)))
++ print("static const CP932MapData kCP932MapData[] = {")
+ for n in keys:
+- print " { %s, %s }," % (n ,result[n])
+- print " { 0, 0 }";
+- print "};"
++ print(" { %s, %s }," % (n ,result[n]))
++ print(" { 0, 0 }");
++ print("};")
+
+ if __name__ == "__main__":
+ main()
+--- /src/gui/character_pad/data/gen_local_character_map.py
++++ /src/gui/character_pad/data/gen_local_character_map.py
+@@ -30,7 +30,6 @@
+
+ __author__ = "taku"
+
+-import string
+ import re
+ import sys
+
+@@ -43,9 +42,9 @@
+ fh = open(filename)
+ result = []
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+- array = string.split(line)
++ array = line.split()
+ jis = array[0].replace('0x', '')
+ ucs2 = array[1].replace('0x', '')
+ if len(jis) == 2:
+@@ -53,6 +52,7 @@
+
+ if IsValidUnicode(ucs2):
+ result.append([jis, ucs2])
++ fh.close()
+
+ return ["JISX0201", result]
+
+@@ -60,13 +60,14 @@
+ fh = open(filename)
+ result = []
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+ array = line.split()
+ jis = array[1].replace('0x', '')
+ ucs2 = array[2].replace('0x', '')
+ if IsValidUnicode(ucs2):
+ result.append([jis, ucs2])
++ fh.close()
+
+ return ["JISX0208", result]
+
+@@ -74,13 +75,14 @@
+ fh = open(filename)
+ result = []
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+ array = line.split()
+ jis = array[0].replace('0x', '')
+ ucs2 = array[1].replace('0x', '')
+ if IsValidUnicode(ucs2):
+ result.append([jis, ucs2])
++ fh.close()
+
+ return ["JISX0212", result]
+
+@@ -88,7 +90,7 @@
+ fh = open(filename)
+ result = []
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+ array = line.split()
+ sjis = array[0].replace('0x', '')
+@@ -100,19 +102,20 @@
+
+ if IsValidUnicode(ucs2):
+ result.append([sjis, ucs2])
++ fh.close()
+
+ return ["CP932", result]
+
+ def Output(arg):
+ name = arg[0]
+ result = arg[1]
+- print "static const size_t k%sMapSize = %d;" % (name, len(result))
+- print "static const mozc::gui::CharacterPalette::LocalCharacterMap k%sMap[] = {" % (name)
++ print("static const size_t k%sMapSize = %d;" % (name, len(result)))
++ print("static const mozc::gui::CharacterPalette::LocalCharacterMap k%sMap[] = {" % (name))
+ for n in result:
+- print " { 0x%s, 0x%s }," % (n[0] ,n[1])
+- print " { 0, 0 }";
+- print "};"
+- print ""
++ print(" { 0x%s, 0x%s }," % (n[0] ,n[1]))
++ print(" { 0, 0 }");
++ print("};")
++ print("")
+
+ if __name__ == "__main__":
+ Output(LoadJISX0201(sys.argv[1]))
+--- /src/gui/character_pad/data/gen_unicode_blocks.py
++++ /src/gui/character_pad/data/gen_unicode_blocks.py
+@@ -33,13 +33,13 @@
+ import sys
+ import re
+
+-re = re.compile('^(.....?)\.\.(.....?); (.+)')
++re = re.compile(r'^(.....?)\.\.(.....?); (.+)')
+
+ def main():
+- print "static const mozc::gui::CharacterPalette::UnicodeBlock kUnicodeBlockTable[] = {"
++ print("static const mozc::gui::CharacterPalette::UnicodeBlock kUnicodeBlockTable[] = {")
+ fh = open(sys.argv[1])
+ for line in fh.readlines():
+- if line[0] is '#':
++ if line[0] == '#':
+ continue
+ m = re.match(line)
+ if m is not None:
+@@ -47,11 +47,12 @@
+ end = int(m.group(2), 16)
+ name = m.group(3)
+ if start <= 0x2FFFF and end <= 0x2FFFF:
+- print " { \"%s\", { %d, %d } }," % (name, start, end)
++ print(" { \"%s\", { %d, %d } }," % (name, start, end))
++ fh.close()
+
+- print " { NULL, { 0, 0 } }"
+- print "};"
+- print ""
++ print(" { NULL, { 0, 0 } }")
++ print("};")
++ print("")
+
+ if __name__ == "__main__":
+ main()
+--- /src/gui/character_pad/data/gen_unicode_data.py
++++ /src/gui/character_pad/data/gen_unicode_data.py
+@@ -46,18 +46,19 @@
+ code = int(code, 16)
+ if code < 0x2FFFF:
+ results.append(" { %d, \"%s\" }," % (code, desc))
++ fh.close()
+
+- print "struct UnicodeData {";
+- print " char32 ucs4;";
+- print " const char *description;";
+- print "};";
+- print ""
+- print "static const size_t kUnicodeDataSize = %d;" % (len(results))
+- print "static const UnicodeData kUnicodeData[] = {";
++ print("struct UnicodeData {");
++ print(" char32 ucs4;");
++ print(" const char *description;");
++ print("};");
++ print("")
++ print("static const size_t kUnicodeDataSize = %d;" % (len(results)))
++ print("static const UnicodeData kUnicodeData[] = {");
+ for line in results:
+- print line;
+- print " { 0, NULL }";
+- print "};";
++ print(line);
++ print(" { 0, NULL }");
++ print("};");
+
+ if __name__ == "__main__":
+ main()
+--- /src/gui/character_pad/data/gen_unihan_data.py
++++ /src/gui/character_pad/data/gen_unihan_data.py
+@@ -31,35 +31,34 @@
+ __author__ = "taku"
+
+ import re
+-import string
+ import sys
+ rs = {}
+
+ def Escape(n):
+- if n is not "NULL":
++ if n != "NULL":
+ return "\"%s\"" % (n)
+ else:
+ return "NULL"
+
+ def GetCode(n):
+- if n is not "NULL":
+- n = string.replace(n, '0-', 'JIS X 0208: 0x')
+- n = string.replace(n, '1-', 'JIS X 0212: 0x')
+- n = string.replace(n, '3-', 'JIS X 0213: 0x')
+- n = string.replace(n, '4-', 'JIS X 0213: 0x')
+- n = string.replace(n, 'A-', 'Vendors Ideographs: 0x')
+- n = string.replace(n, '3A', 'JIS X 0213 2000: 0x')
++ if n != "NULL":
++ n = n.replace('0-', 'JIS X 0208: 0x')
++ n = n.replace('1-', 'JIS X 0212: 0x')
++ n = n.replace('3-', 'JIS X 0213: 0x')
++ n = n.replace('4-', 'JIS X 0213: 0x')
++ n = n.replace('A-', 'Vendors Ideographs: 0x')
++ n = n.replace('3A', 'JIS X 0213 2000: 0x')
+ return "\"%s\"" % n
+ else:
+ return "NULL"
+
+ def GetRadical(n):
+ pat = re.compile(r'^(\d+)\.')
+- if n is not "NULL":
++ if n != "NULL":
+ m = pat.match(n)
+ if m:
+ result = rs[m.group(1)]
+- return "\"%s\"" % (result.encode('string_escape'))
++ return "\"%s\"" % result
+ else:
+ return "NULL"
+ else:
+@@ -73,6 +72,7 @@
+ id = array[1]
+ radical = array[2]
+ rs[id] = radical
++ fh.close()
+
+ dic = {}
+ pat = re.compile(r'^U\+(\S+)\s+(kTotalStrokes|kJapaneseKun|kJapaneseOn|kRSUnicode|kIRG_JSource)\t(.+)')
+@@ -86,23 +86,24 @@
+ n = int(m.group(1), 16)
+ if n <= 65536:
+ dic.setdefault(key, {}).setdefault(field, value)
++ fh.close()
+
+ keys = sorted(dic.keys())
+
+- print "struct UnihanData {";
+- print " unsigned int ucs4;";
++ print("struct UnihanData {");
++ print(" unsigned int ucs4;");
+ # Since the total strokes defined in Unihan data is Chinese-based
+ # number, we can't use it.
+ # print " unsigned char total_strokes;";
+- print " const char *japanese_kun;";
+- print " const char *japanese_on;";
++ print(" const char *japanese_kun;");
++ print(" const char *japanese_on;");
+ # Since the radical information defined in Unihan data is Chinese-based
+ # number, we can't use it.
+ # print " const char *radical;";
+- print " const char *IRG_jsource;";
+- print "};"
+- print "static const size_t kUnihanDataSize = %d;" % (len(keys))
+- print "static const UnihanData kUnihanData[] = {"
++ print(" const char *IRG_jsource;");
++ print("};")
++ print("static const size_t kUnihanDataSize = %d;" % (len(keys)))
++ print("static const UnihanData kUnihanData[] = {")
+
+ for key in keys:
+ total_strokes = dic[key].get("kTotalStrokes", "0")
+@@ -111,9 +112,9 @@
+ rad = GetRadical(dic[key].get("kRSUnicode", "NULL"))
+ code = GetCode(dic[key].get("kIRG_JSource", "NULL"))
+ # print " { 0x%s, %s, %s, %s, %s, %s }," % (key, total_strokes, kun, on, rad, code)
+- print " { 0x%s, %s, %s, %s }," % (key, kun, on, code)
++ print(" { 0x%s, %s, %s, %s }," % (key, kun, on, code))
+
+- print "};"
++ print("};")
+
+ if __name__ == "__main__":
+ main()
diff --git a/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch
new file mode 100644
index 000000000000..41d2bf9eeb90
--- /dev/null
+++ b/app-i18n/mozc/files/mozc-2.23.2815.102-python-3_4.patch
@@ -0,0 +1,537 @@
+https://github.com/google/mozc/issues/462
+
+--- /src/prediction/gen_zero_query_data.py
++++ /src/prediction/gen_zero_query_data.py
+@@ -59,20 +59,20 @@
+ Returns:
+ A integer indicating parsed pua.
+ """
+- if not s or s[0] == '>':
++ if not s or s[0:1] == b'>':
+ return 0
+ return int(s, 16)
+
+
+ def NormalizeString(string):
+ return unicodedata.normalize(
+- 'NFKC', string.decode('utf-8')).encode('utf-8').replace('~', '〜')
++ 'NFKC', string.decode('utf-8')).replace('~', '〜').encode('utf-8')
+
+
+ def RemoveTrailingNumber(string):
+ if not string:
+- return ''
+- return re.sub(r'^([^0-9]+)[0-9]+$', r'\1', string)
++ return b''
++ return re.sub(br'^([^0-9]+)[0-9]+$', r'\1', string)
+
+
+ def GetReadingsFromDescription(description):
+@@ -84,19 +84,19 @@
+ # - ビル・建物
+ # \xE3\x83\xBB : "・"
+ return [RemoveTrailingNumber(token) for token
+- in re.split(r'(?:\(|\)|/|\xE3\x83\xBB)+', normalized)]
++ in re.split(br'(?:\(|\)|/|\xE3\x83\xBB)+', normalized)]
+
+
+ def ReadEmojiTsv(stream):
+ """Reads emoji data from stream and returns zero query data."""
+ zero_query_dict = defaultdict(list)
+ stream = code_generator_util.SkipLineComment(stream)
+- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'):
++ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'):
+ if len(columns) != 13:
+- logging.critical('format error: %s', '\t'.join(columns))
++ logging.critical('format error: %s', b'\t'.join(columns))
+ sys.exit(1)
+
+- code_points = columns[0].split(' ')
++ code_points = columns[0].split(b' ')
+
+ # Emoji code point.
+ emoji = columns[1]
+@@ -114,12 +114,12 @@
+ # - Composite emoji which has multiple code point.
+ # NOTE: Some Unicode 6.0 emoji don't have PUA, and it is also omitted.
+ # TODO(hsumita): Check the availability of such emoji and enable it.
+- logging.info('Skip %s', ' '.join(code_points))
++ logging.info('Skip %s', b' '.join(code_points))
+ continue
+
+ reading_list = []
+ # \xe3\x80\x80 is a full-width space
+- for reading in re.split(r'(?: |\xe3\x80\x80)+', NormalizeString(readings)):
++ for reading in re.split(br'(?: |\xe3\x80\x80)+', NormalizeString(readings)):
+ if not reading:
+ continue
+ reading_list.append(reading)
+@@ -158,15 +158,15 @@
+ zero_query_dict = defaultdict(list)
+
+ for line in input_stream:
+- if line.startswith('#'):
++ if line.startswith(b'#'):
+ continue
+- line = line.rstrip('\r\n')
++ line = line.rstrip(b'\r\n')
+ if not line:
+ continue
+
+- tokens = line.split('\t')
++ tokens = line.split(b'\t')
+ key = tokens[0]
+- values = tokens[1].split(',')
++ values = tokens[1].split(b',')
+
+ for value in values:
+ zero_query_dict[key].append(
+@@ -179,16 +179,16 @@
+ """Reads emoticon data from stream and returns zero query data."""
+ zero_query_dict = defaultdict(list)
+ stream = code_generator_util.SkipLineComment(stream)
+- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'):
++ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'):
+ if len(columns) != 3:
+- logging.critical('format error: %s', '\t'.join(columns))
++ logging.critical('format error: %s', b'\t'.join(columns))
+ sys.exit(1)
+
+ emoticon = columns[0]
+ readings = columns[2]
+
+ # \xe3\x80\x80 is a full-width space
+- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()):
++ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()):
+ if not reading:
+ continue
+ zero_query_dict[reading].append(
+@@ -202,9 +202,9 @@
+ """Reads emoji data from stream and returns zero query data."""
+ zero_query_dict = defaultdict(list)
+ stream = code_generator_util.SkipLineComment(stream)
+- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'):
++ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'):
+ if len(columns) < 3:
+- logging.warning('format error: %s', '\t'.join(columns))
++ logging.warning('format error: %s', b'\t'.join(columns))
+ continue
+
+ symbol = columns[1]
+@@ -222,7 +222,7 @@
+ continue
+
+ # \xe3\x80\x80 is a full-width space
+- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()):
++ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()):
+ if not reading:
+ continue
+ zero_query_dict[reading].append(
+@@ -247,7 +247,7 @@
+
+ def IsValidKeyForZeroQuery(key):
+ """Returns if the key is valid for zero query trigger."""
+- is_ascii = all(ord(char) < 128 for char in key)
++ is_ascii = all(char < 128 for char in key)
+ return not is_ascii
+
+
+@@ -301,13 +301,13 @@
+
+ def main():
+ options = ParseOptions()
+- with open(options.input_rule, 'r') as input_stream:
++ with open(options.input_rule, 'rb') as input_stream:
+ zero_query_rule_dict = ReadZeroQueryRuleData(input_stream)
+- with open(options.input_symbol, 'r') as input_stream:
++ with open(options.input_symbol, 'rb') as input_stream:
+ zero_query_symbol_dict = ReadSymbolTsv(input_stream)
+- with open(options.input_emoji, 'r') as input_stream:
++ with open(options.input_emoji, 'rb') as input_stream:
+ zero_query_emoji_dict = ReadEmojiTsv(input_stream)
+- with open(options.input_emoticon, 'r') as input_stream:
++ with open(options.input_emoticon, 'rb') as input_stream:
+ zero_query_emoticon_dict = ReadEmoticonTsv(input_stream)
+
+ merged_zero_query_dict = MergeZeroQueryData(
+--- /src/prediction/gen_zero_query_number_data.py
++++ /src/prediction/gen_zero_query_number_data.py
+@@ -41,15 +41,15 @@
+ zero_query_dict = defaultdict(list)
+
+ for line in input_stream:
+- if line.startswith('#'):
++ if line.startswith(b'#'):
+ continue
+- line = line.rstrip('\r\n')
++ line = line.rstrip(b'\r\n')
+ if not line:
+ continue
+
+- tokens = line.split('\t')
++ tokens = line.split(b'\t')
+ key = tokens[0]
+- values = tokens[1].split(',')
++ values = tokens[1].split(b',')
+
+ for value in values:
+ zero_query_dict[key].append(
+@@ -71,7 +71,7 @@
+
+ def main():
+ options = ParseOption()
+- with open(options.input, 'r') as input_stream:
++ with open(options.input, 'rb') as input_stream:
+ zero_query_dict = ReadZeroQueryNumberData(input_stream)
+ util.WriteZeroQueryData(zero_query_dict,
+ options.output_token_array,
+--- /src/prediction/gen_zero_query_util.py
++++ /src/prediction/gen_zero_query_util.py
+@@ -69,7 +69,7 @@
+ output_string_array):
+ # Collect all the strings and assing index in ascending order
+ string_index = {}
+- for key, entry_list in zero_query_dict.iteritems():
++ for key, entry_list in zero_query_dict.items():
+ string_index[key] = 0
+ for entry in entry_list:
+ string_index[entry.value] = 0
+--- /src/rewriter/gen_counter_suffix_array.py
++++ /src/rewriter/gen_counter_suffix_array.py
+@@ -43,7 +43,7 @@
+ with codecs.open(id_file, 'r', encoding='utf-8') as stream:
+ stream = code_generator_util.ParseColumnStream(stream, num_column=2)
+ for pos_id, pos_name in stream:
+- if pos_name.startswith(u'名詞,接尾,助数詞'):
++ if pos_name.startswith('名詞,接尾,助数詞'):
+ pos_ids.add(pos_id)
+ return pos_ids
+
+--- /src/rewriter/gen_emoji_rewriter_data.py
++++ /src/rewriter/gen_emoji_rewriter_data.py
+@@ -74,19 +74,19 @@
+ the glyph (in other words, it has alternative (primary) code point, which
+ doesn't lead '>' and that's why we'll ignore it).
+ """
+- if not s or s[0] == '>':
++ if not s or s[0:1] == b'>':
+ return None
+ return int(s, 16)
+
+
+-_FULLWIDTH_RE = re.compile(ur'[!-~]') # U+FF01 - U+FF5E
++_FULLWIDTH_RE = re.compile(r'[!-~]') # U+FF01 - U+FF5E
+
+
+ def NormalizeString(string):
+ """Normalize full width ascii characters to half width characters."""
+- offset = ord(u'A') - ord(u'A')
+- return _FULLWIDTH_RE.sub(lambda x: unichr(ord(x.group(0)) - offset),
+- unicode(string, 'utf-8')).encode('utf-8')
++ offset = ord('A') - ord('A')
++ return _FULLWIDTH_RE.sub(lambda x: chr(ord(x.group(0)) - offset),
++ string.decode('utf-8')).encode('utf-8')
+
+
+ def ReadEmojiTsv(stream):
+@@ -96,14 +96,14 @@
+ token_dict = defaultdict(list)
+
+ stream = code_generator_util.SkipLineComment(stream)
+- for columns in code_generator_util.ParseColumnStream(stream, delimiter='\t'):
++ for columns in code_generator_util.ParseColumnStream(stream, delimiter=b'\t'):
+ if len(columns) != 13:
+- logging.critical('format error: %s', '\t'.join(columns))
++ logging.critical('format error: %s', b'\t'.join(columns))
+ sys.exit(1)
+
+- code_points = columns[0].split(' ')
++ code_points = columns[0].split(b' ')
+ # Emoji code point.
+- emoji = columns[1] if columns[1] else ''
++ emoji = columns[1] if columns[1] else b''
+ android_pua = ParseCodePoint(columns[2])
+ docomo_pua = ParseCodePoint(columns[3])
+ softbank_pua = ParseCodePoint(columns[4])
+@@ -112,10 +112,10 @@
+ readings = columns[6]
+
+ # [7]: Name defined in Unicode. It is ignored in current implementation.
+- utf8_description = columns[8] if columns[8] else ''
+- docomo_description = columns[9] if columns[9] else ''
+- softbank_description = columns[10] if columns[10] else ''
+- kddi_description = columns[11] if columns[11] else ''
++ utf8_description = columns[8] if columns[8] else b''
++ docomo_description = columns[9] if columns[9] else b''
++ softbank_description = columns[10] if columns[10] else b''
++ kddi_description = columns[11] if columns[11] else b''
+
+ if not android_pua or len(code_points) > 1:
+ # Skip some emoji, which is not supported on old devices.
+@@ -123,7 +123,7 @@
+ # - Composite emoji which has multiple code point.
+ # NOTE: Some Unicode 6.0 emoji don't have PUA, and it is also omitted.
+ # TODO(hsumita): Check the availability of such emoji and enable it.
+- logging.info('Skip %s', ' '.join(code_points))
++ logging.info('Skip %s', b' '.join(code_points))
+ continue
+
+ # Check consistency between carrier PUA codes and descriptions for Android
+@@ -132,7 +132,7 @@
+ (bool(softbank_pua) != bool(softbank_description)) or
+ (bool(kddi_pua) != bool(kddi_description))):
+ logging.warning('carrier PUA and description conflict: %s',
+- '\t'.join(columns))
++ b'\t'.join(columns))
+ continue
+
+ # Check if the character is usable on Android.
+@@ -140,7 +140,7 @@
+ android_pua = 0 # Replace None with 0.
+
+ if not emoji and not android_pua:
+- logging.info('Skip: %s', '\t'.join(columns))
++ logging.info('Skip: %s', b'\t'.join(columns))
+ continue
+
+ index = len(emoji_data_list)
+@@ -149,7 +149,7 @@
+ kddi_description))
+
+ # \xe3\x80\x80 is a full-width space
+- for reading in re.split(r'(?: |\xe3\x80\x80)+', readings.strip()):
++ for reading in re.split(br'(?: |\xe3\x80\x80)+', readings.strip()):
+ if reading:
+ token_dict[NormalizeString(reading)].append(index)
+
+@@ -159,7 +159,7 @@
+ def OutputData(emoji_data_list, token_dict,
+ token_array_file, string_array_file):
+ """Output token and string arrays to files."""
+- sorted_token_dict = sorted(token_dict.iteritems())
++ sorted_token_dict = sorted(token_dict.items())
+
+ strings = {}
+ for reading, _ in sorted_token_dict:
+@@ -171,7 +171,7 @@
+ strings[docomo_description] = 0
+ strings[softbank_description] = 0
+ strings[kddi_description] = 0
+- sorted_strings = sorted(strings.iterkeys())
++ sorted_strings = sorted(strings.keys())
+ for index, s in enumerate(sorted_strings):
+ strings[s] = index
+
+@@ -205,7 +205,7 @@
+
+ def main():
+ options = ParseOptions()
+- with open(options.input, 'r') as input_stream:
++ with open(options.input, 'rb') as input_stream:
+ (emoji_data_list, token_dict) = ReadEmojiTsv(input_stream)
+
+ OutputData(emoji_data_list, token_dict,
+--- /src/rewriter/gen_reading_correction_data.py
++++ /src/rewriter/gen_reading_correction_data.py
+@@ -63,7 +63,7 @@
+ def WriteData(input_path, output_value_array_path, output_error_array_path,
+ output_correction_array_path):
+ outputs = []
+- with open(input_path) as input_stream:
++ with open(input_path, 'rb') as input_stream:
+ input_stream = code_generator_util.SkipLineComment(input_stream)
+ input_stream = code_generator_util.ParseColumnStream(input_stream,
+ num_column=3)
+@@ -73,7 +73,7 @@
+
+ # In order to lookup the entries via |error| with binary search,
+ # sort outputs here.
+- outputs.sort(lambda x, y: cmp(x[1], y[1]) or cmp(x[0], y[0]))
++ outputs.sort(key=lambda x: (x[1], x[0]))
+
+ serialized_string_array_builder.SerializeToFile(
+ [value for (value, _, _) in outputs], output_value_array_path)
+--- /src/rewriter/gen_single_kanji_rewriter_data.py
++++ /src/rewriter/gen_single_kanji_rewriter_data.py
+@@ -52,7 +52,7 @@
+ stream = code_generator_util.ParseColumnStream(stream, num_column=2)
+ outputs = list(stream)
+ # For binary search by |key|, sort outputs here.
+- outputs.sort(lambda x, y: cmp(x[0], y[0]))
++ outputs.sort(key=lambda x: x[0])
+
+ return outputs
+
+@@ -72,7 +72,7 @@
+ variant_items.append([target, original, len(variant_types) - 1])
+
+ # For binary search by |target|, sort variant items here.
+- variant_items.sort(lambda x, y: cmp(x[0], y[0]))
++ variant_items.sort(key=lambda x: x[0])
+
+ return (variant_types, variant_items)
+
+@@ -151,10 +151,10 @@
+ def main():
+ options = _ParseOptions()
+
+- with open(options.single_kanji_file, 'r') as single_kanji_stream:
++ with open(options.single_kanji_file, 'rb') as single_kanji_stream:
+ single_kanji = ReadSingleKanji(single_kanji_stream)
+
+- with open(options.variant_file, 'r') as variant_stream:
++ with open(options.variant_file, 'rb') as variant_stream:
+ variant_info = ReadVariant(variant_stream)
+
+ WriteSingleKanji(single_kanji,
+--- /src/session/gen_session_stress_test_data.py
++++ /src/session/gen_session_stress_test_data.py
+@@ -50,24 +50,26 @@
+ """
+ result = ''
+ for c in s:
+- hexstr = hex(ord(c))
++ hexstr = hex(c)
+ # because hexstr contains '0x', remove the prefix and add our prefix
+ result += '\\x' + hexstr[2:]
+ return result
+
+ def GenerateHeader(file):
+ try:
+- print "const char *kTestSentences[] = {"
+- for line in open(file, "r"):
+- if line.startswith('#'):
++ print("const char *kTestSentences[] = {")
++ fh = open(file, "rb")
++ for line in fh:
++ if line.startswith(b'#'):
+ continue
+- line = line.rstrip('\r\n')
++ line = line.rstrip(b'\r\n')
+ if not line:
+ continue
+- print " \"%s\"," % escape_string(line)
+- print "};"
++ print(" \"%s\"," % escape_string(line))
++ fh.close()
++ print("};")
+ except:
+- print "cannot open %s" % (file)
++ print("cannot open %s" % (file))
+ sys.exit(1)
+
+ def main():
+--- /src/unix/ibus/gen_mozc_xml.py
++++ /src/unix/ibus/gen_mozc_xml.py
+@@ -74,7 +74,7 @@
+
+
+ def OutputXmlElement(param_dict, element_name, value):
+- print ' <%s>%s</%s>' % (element_name, (value % param_dict), element_name)
++ print(' <%s>%s</%s>' % (element_name, (value % param_dict), element_name))
+
+
+ def OutputXml(param_dict, component, engine_common, engines, setup_arg):
+@@ -90,26 +90,26 @@
+ engines: A dictionary from a property name to a list of property values of
+ engines. For example, {'name': ['mozc-jp', 'mozc', 'mozc-dv']}.
+ """
+- print '<component>'
+- for key in component:
++ print('<component>')
++ for key in sorted(component):
+ OutputXmlElement(param_dict, key, component[key])
+- print '<engines>'
++ print('<engines>')
+ for i in range(len(engines['name'])):
+- print '<engine>'
+- for key in engine_common:
++ print('<engine>')
++ for key in sorted(engine_common):
+ OutputXmlElement(param_dict, key, engine_common[key])
+ if setup_arg:
+ OutputXmlElement(param_dict, 'setup', ' '.join(setup_arg))
+- for key in engines:
++ for key in sorted(engines):
+ OutputXmlElement(param_dict, key, engines[key][i])
+- print '</engine>'
+- print '</engines>'
+- print '</component>'
++ print('</engine>')
++ print('</engines>')
++ print('</component>')
+
+
+ def OutputCppVariable(param_dict, prefix, variable_name, value):
+- print 'const char k%s%s[] = "%s";' % (prefix, variable_name.capitalize(),
+- (value % param_dict))
++ print('const char k%s%s[] = "%s";' % (prefix, variable_name.capitalize(),
++ (value % param_dict)))
+
+
+ def OutputCpp(param_dict, component, engine_common, engines):
+@@ -122,18 +122,18 @@
+ engines: ditto.
+ """
+ guard_name = 'MOZC_UNIX_IBUS_MAIN_H_'
+- print CPP_HEADER % (guard_name, guard_name)
+- for key in component:
++ print(CPP_HEADER % (guard_name, guard_name))
++ for key in sorted(component):
+ OutputCppVariable(param_dict, 'Component', key, component[key])
+- for key in engine_common:
++ for key in sorted(engine_common):
+ OutputCppVariable(param_dict, 'Engine', key, engine_common[key])
+- for key in engines:
+- print 'const char* kEngine%sArray[] = {' % key.capitalize()
++ for key in sorted(engines):
++ print('const char* kEngine%sArray[] = {' % key.capitalize())
+ for i in range(len(engines[key])):
+- print '"%s",' % (engines[key][i] % param_dict)
+- print '};'
+- print 'const size_t kEngineArrayLen = %s;' % len(engines['name'])
+- print CPP_FOOTER % guard_name
++ print('"%s",' % (engines[key][i] % param_dict))
++ print('};')
++ print('const size_t kEngineArrayLen = %s;' % len(engines['name']))
++ print(CPP_FOOTER % guard_name)
+
+
+ def CheckIBusVersion(options, minimum_version):
+--- /src/usage_stats/gen_stats_list.py
++++ /src/usage_stats/gen_stats_list.py
+@@ -37,23 +37,24 @@
+
+ def GetStatsNameList(filename):
+ stats = []
+- for line in open(filename, 'r'):
+- stat = line.strip()
+- if not stat or stat[0] == '#':
+- continue
+- stats.append(stat)
++ with open(filename, 'r') as file:
++ for line in file:
++ stat = line.strip()
++ if not stat or stat[0] == '#':
++ continue
++ stats.append(stat)
+ return stats
+
+
+ def main():
+ stats_list = GetStatsNameList(sys.argv[1])
+- print '// This header file is generated by gen_stats_list.py'
++ print('// This header file is generated by gen_stats_list.py')
+ for stats in stats_list:
+- print 'const char k%s[] = "%s";' % (stats, stats)
+- print 'const char *kStatsList[] = {'
++ print('const char k%s[] = "%s";' % (stats, stats))
++ print('const char *kStatsList[] = {')
+ for stats in stats_list:
+- print ' k%s,' % (stats)
+- print '};'
++ print(' k%s,' % (stats))
++ print('};')
+
+
+ if __name__ == '__main__':
diff --git a/app-i18n/mozc/mozc-2.23.2815.102.ebuild b/app-i18n/mozc/mozc-2.23.2815.102.ebuild
index 425b785cbfab..8a55204d5db9 100644
--- a/app-i18n/mozc/mozc-2.23.2815.102.ebuild
+++ b/app-i18n/mozc/mozc-2.23.2815.102.ebuild
@@ -1,8 +1,8 @@
-# Copyright 2010-2019 Gentoo Authors
+# Copyright 2010-2020 Gentoo Authors
# Distributed under the terms of the GNU General Public License v2
EAPI="7"
-PYTHON_COMPAT=(python2_7)
+PYTHON_COMPAT=(python{3_6,3_7,3_8})
inherit elisp-common multiprocessing python-any-r1 toolchain-funcs
@@ -105,6 +105,10 @@ src_unpack() {
}
src_prepare() {
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_1.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_2.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_3.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_4.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-system_libraries.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-gcc-8.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-protobuf_generated_classes_no_inheritance.patch"
diff --git a/app-i18n/mozc/mozc-9999.ebuild b/app-i18n/mozc/mozc-9999.ebuild
index d7c66b558d92..e82d56df3360 100644
--- a/app-i18n/mozc/mozc-9999.ebuild
+++ b/app-i18n/mozc/mozc-9999.ebuild
@@ -1,8 +1,8 @@
-# Copyright 2010-2019 Gentoo Authors
+# Copyright 2010-2020 Gentoo Authors
# Distributed under the terms of the GNU General Public License v2
EAPI="7"
-PYTHON_COMPAT=(python2_7)
+PYTHON_COMPAT=(python{3_6,3_7,3_8})
inherit elisp-common multiprocessing python-any-r1 toolchain-funcs
@@ -105,6 +105,10 @@ src_unpack() {
}
src_prepare() {
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_1.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_2.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_3.patch"
+ eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-python-3_4.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-system_libraries.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-gcc-8.patch"
eapply -p2 "${FILESDIR}/${PN}-2.23.2815.102-protobuf_generated_classes_no_inheritance.patch"