forked from openkylin/platform_build
Merge changes Idc7776be,I409fe30a
am: ec8e2b4e02
Change-Id: Ieea68e3ee56fc978761e9d2691667e5fb50bcf04
This commit is contained in:
commit
70b5b96565
|
@ -14,6 +14,7 @@
|
|||
|
||||
from __future__ import print_function
|
||||
|
||||
import base64
|
||||
import collections
|
||||
import copy
|
||||
import errno
|
||||
|
@ -30,7 +31,6 @@ import platform
|
|||
import re
|
||||
import shlex
|
||||
import shutil
|
||||
import string
|
||||
import subprocess
|
||||
import sys
|
||||
import tempfile
|
||||
|
@ -190,6 +190,8 @@ def Run(args, verbose=None, **kwargs):
|
|||
kwargs: Any additional args to be passed to subprocess.Popen(), such as env,
|
||||
stdin, etc. stdout and stderr will default to subprocess.PIPE and
|
||||
subprocess.STDOUT respectively unless caller specifies any of them.
|
||||
universal_newlines will default to True, as most of the users in
|
||||
releasetools expect string output.
|
||||
|
||||
Returns:
|
||||
A subprocess.Popen object.
|
||||
|
@ -197,6 +199,8 @@ def Run(args, verbose=None, **kwargs):
|
|||
if 'stdout' not in kwargs and 'stderr' not in kwargs:
|
||||
kwargs['stdout'] = subprocess.PIPE
|
||||
kwargs['stderr'] = subprocess.STDOUT
|
||||
if 'universal_newlines' not in kwargs:
|
||||
kwargs['universal_newlines'] = True
|
||||
# Don't log any if caller explicitly says so.
|
||||
if verbose != False:
|
||||
logger.info(" Running: \"%s\"", " ".join(args))
|
||||
|
@ -314,7 +318,7 @@ def LoadInfoDict(input_file, repacking=False):
|
|||
|
||||
def read_helper(fn):
|
||||
if isinstance(input_file, zipfile.ZipFile):
|
||||
return input_file.read(fn)
|
||||
return input_file.read(fn).decode()
|
||||
else:
|
||||
path = os.path.join(input_file, *fn.split("/"))
|
||||
try:
|
||||
|
@ -526,7 +530,7 @@ def LoadRecoveryFSTab(read_helper, fstab_version, recovery_fstab_path,
|
|||
# system. Other areas assume system is always at "/system" so point /system
|
||||
# at /.
|
||||
if system_root_image:
|
||||
assert not d.has_key("/system") and d.has_key("/")
|
||||
assert '/system' not in d and '/' in d
|
||||
d["/system"] = d["/"]
|
||||
return d
|
||||
|
||||
|
@ -953,7 +957,7 @@ def GetSparseImage(which, tmpdir, input_zip, allow_shared_blocks,
|
|||
# filename listed in system.map may contain an additional leading slash
|
||||
# (i.e. "//system/framework/am.jar"). Using lstrip to get consistent
|
||||
# results.
|
||||
arcname = string.replace(entry, which, which.upper(), 1).lstrip('/')
|
||||
arcname = entry.replace(which, which.upper(), 1).lstrip('/')
|
||||
|
||||
# Special handling another case, where files not under /system
|
||||
# (e.g. "/sbin/charger") are packed under ROOT/ in a target_files.zip.
|
||||
|
@ -1223,7 +1227,7 @@ def ReadApkCerts(tf_zip):
|
|||
if basename:
|
||||
installed_files.add(basename)
|
||||
|
||||
for line in tf_zip.read("META/apkcerts.txt").split("\n"):
|
||||
for line in tf_zip.read('META/apkcerts.txt').decode().split('\n'):
|
||||
line = line.strip()
|
||||
if not line:
|
||||
continue
|
||||
|
@ -1433,6 +1437,8 @@ class PasswordManager(object):
|
|||
|
||||
if not first:
|
||||
print("key file %s still missing some passwords." % (self.pwfile,))
|
||||
if sys.version_info[0] >= 3:
|
||||
raw_input = input # pylint: disable=redefined-builtin
|
||||
answer = raw_input("try to edit again? [y]> ").strip()
|
||||
if answer and answer[0] not in 'yY':
|
||||
raise RuntimeError("key passwords unavailable")
|
||||
|
@ -2185,7 +2191,7 @@ def ParseCertificate(data):
|
|||
This gives the same result as `openssl x509 -in <filename> -outform DER`.
|
||||
|
||||
Returns:
|
||||
The decoded certificate string.
|
||||
The decoded certificate bytes.
|
||||
"""
|
||||
cert_buffer = []
|
||||
save = False
|
||||
|
@ -2196,7 +2202,7 @@ def ParseCertificate(data):
|
|||
cert_buffer.append(line)
|
||||
if "--BEGIN CERTIFICATE--" in line:
|
||||
save = True
|
||||
cert = "".join(cert_buffer).decode('base64')
|
||||
cert = base64.b64decode("".join(cert_buffer))
|
||||
return cert
|
||||
|
||||
|
||||
|
@ -2338,7 +2344,7 @@ fi
|
|||
|
||||
logger.info("putting script in %s", sh_location)
|
||||
|
||||
output_sink(sh_location, sh)
|
||||
output_sink(sh_location, sh.encode())
|
||||
|
||||
|
||||
class DynamicPartitionUpdate(object):
|
||||
|
|
|
@ -579,7 +579,7 @@ class CommonApkUtilsTest(test_utils.ReleaseToolsTestCase):
|
|||
def test_ExtractPublicKey(self):
|
||||
cert = os.path.join(self.testdata_dir, 'testkey.x509.pem')
|
||||
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
||||
with open(pubkey, 'rb') as pubkey_fp:
|
||||
with open(pubkey) as pubkey_fp:
|
||||
self.assertEqual(pubkey_fp.read(), common.ExtractPublicKey(cert))
|
||||
|
||||
def test_ExtractPublicKey_invalidInput(self):
|
||||
|
@ -590,15 +590,16 @@ class CommonApkUtilsTest(test_utils.ReleaseToolsTestCase):
|
|||
def test_ExtractAvbPublicKey(self):
|
||||
privkey = os.path.join(self.testdata_dir, 'testkey.key')
|
||||
pubkey = os.path.join(self.testdata_dir, 'testkey.pubkey.pem')
|
||||
with open(common.ExtractAvbPublicKey(privkey)) as privkey_fp, \
|
||||
open(common.ExtractAvbPublicKey(pubkey)) as pubkey_fp:
|
||||
with open(common.ExtractAvbPublicKey(privkey), 'rb') as privkey_fp, \
|
||||
open(common.ExtractAvbPublicKey(pubkey), 'rb') as pubkey_fp:
|
||||
self.assertEqual(privkey_fp.read(), pubkey_fp.read())
|
||||
|
||||
def test_ParseCertificate(self):
|
||||
cert = os.path.join(self.testdata_dir, 'testkey.x509.pem')
|
||||
|
||||
cmd = ['openssl', 'x509', '-in', cert, '-outform', 'DER']
|
||||
proc = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
||||
proc = common.Run(cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE,
|
||||
universal_newlines=False)
|
||||
expected, _ = proc.communicate()
|
||||
self.assertEqual(0, proc.returncode)
|
||||
|
||||
|
@ -914,7 +915,7 @@ class CommonUtilsTest(test_utils.ReleaseToolsTestCase):
|
|||
target_files = common.MakeTempFile(prefix='target_files-', suffix='.zip')
|
||||
with zipfile.ZipFile(target_files, 'w') as target_files_zip:
|
||||
info_values = ''.join(
|
||||
['{}={}\n'.format(k, v) for k, v in sorted(info_dict.iteritems())])
|
||||
['{}={}\n'.format(k, v) for k, v in sorted(info_dict.items())])
|
||||
common.ZipWriteStr(target_files_zip, 'META/misc_info.txt', info_values)
|
||||
|
||||
FSTAB_TEMPLATE = "/dev/block/system {} ext4 ro,barrier=1 defaults"
|
||||
|
@ -1085,7 +1086,7 @@ class InstallRecoveryScriptFormatTest(test_utils.ReleaseToolsTestCase):
|
|||
loc = os.path.join(self._tempdir, prefix, name)
|
||||
if not os.path.exists(os.path.dirname(loc)):
|
||||
os.makedirs(os.path.dirname(loc))
|
||||
with open(loc, "w+") as f:
|
||||
with open(loc, "wb") as f:
|
||||
f.write(data)
|
||||
|
||||
def test_full_recovery(self):
|
||||
|
@ -1110,7 +1111,7 @@ class InstallRecoveryScriptFormatTest(test_utils.ReleaseToolsTestCase):
|
|||
validate_target_files.ValidateInstallRecoveryScript(self._tempdir,
|
||||
self._info)
|
||||
# Validate 'recovery-from-boot' with bonus argument.
|
||||
self._out_tmp_sink("etc/recovery-resource.dat", "bonus", "SYSTEM")
|
||||
self._out_tmp_sink("etc/recovery-resource.dat", b"bonus", "SYSTEM")
|
||||
common.MakeRecoveryPatch(self._tempdir, self._out_tmp_sink,
|
||||
recovery_image, boot_image, self._info)
|
||||
validate_target_files.ValidateInstallRecoveryScript(self._tempdir,
|
||||
|
@ -1118,25 +1119,30 @@ class InstallRecoveryScriptFormatTest(test_utils.ReleaseToolsTestCase):
|
|||
|
||||
|
||||
class MockScriptWriter(object):
|
||||
"""A class that mocks edify_generator.EdifyGenerator.
|
||||
"""
|
||||
"""A class that mocks edify_generator.EdifyGenerator."""
|
||||
|
||||
def __init__(self, enable_comments=False):
|
||||
self.lines = []
|
||||
self.enable_comments = enable_comments
|
||||
|
||||
def Comment(self, comment):
|
||||
if self.enable_comments:
|
||||
self.lines.append("# {}".format(comment))
|
||||
self.lines.append('# {}'.format(comment))
|
||||
|
||||
def AppendExtra(self, extra):
|
||||
self.lines.append(extra)
|
||||
|
||||
def __str__(self):
|
||||
return "\n".join(self.lines)
|
||||
return '\n'.join(self.lines)
|
||||
|
||||
|
||||
class MockBlockDifference(object):
|
||||
|
||||
def __init__(self, partition, tgt, src=None):
|
||||
self.partition = partition
|
||||
self.tgt = tgt
|
||||
self.src = src
|
||||
|
||||
def WriteScript(self, script, _, progress=None,
|
||||
write_verify_script=False):
|
||||
if progress:
|
||||
|
@ -1144,11 +1150,13 @@ class MockBlockDifference(object):
|
|||
script.AppendExtra("patch({});".format(self.partition))
|
||||
if write_verify_script:
|
||||
self.WritePostInstallVerifyScript(script)
|
||||
|
||||
def WritePostInstallVerifyScript(self, script):
|
||||
script.AppendExtra("verify({});".format(self.partition))
|
||||
|
||||
|
||||
class FakeSparseImage(object):
|
||||
|
||||
def __init__(self, size):
|
||||
self.blocksize = 4096
|
||||
self.total_blocks = size // 4096
|
||||
|
@ -1156,12 +1164,13 @@ class FakeSparseImage(object):
|
|||
|
||||
|
||||
class DynamicPartitionsDifferenceTest(test_utils.ReleaseToolsTestCase):
|
||||
|
||||
@staticmethod
|
||||
def get_op_list(output_path):
|
||||
with zipfile.ZipFile(output_path) as output_zip:
|
||||
with output_zip.open("dynamic_partitions_op_list") as op_list:
|
||||
return [line.strip() for line in op_list.readlines()
|
||||
if not line.startswith("#")]
|
||||
with output_zip.open('dynamic_partitions_op_list') as op_list:
|
||||
return [line.decode().strip() for line in op_list.readlines()
|
||||
if not line.startswith(b'#')]
|
||||
|
||||
def setUp(self):
|
||||
self.script = MockScriptWriter()
|
||||
|
|
|
@ -47,25 +47,22 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
}
|
||||
|
||||
self.hash_algorithm = "sha256"
|
||||
self.fixed_salt = \
|
||||
"aee087a5be3b982978c923f566a94613496b417f2af592639bc80d141e34dfe7"
|
||||
self.expected_root_hash = \
|
||||
"0b7c4565e87b1026e11fbab91c0bc29e185c847a5b44d40e6e86e461e8adf80d"
|
||||
self.fixed_salt = (
|
||||
"aee087a5be3b982978c923f566a94613496b417f2af592639bc80d141e34dfe7")
|
||||
self.expected_root_hash = (
|
||||
"0b7c4565e87b1026e11fbab91c0bc29e185c847a5b44d40e6e86e461e8adf80d")
|
||||
|
||||
def _create_simg(self, raw_data):
|
||||
def _CreateSimg(self, raw_data): # pylint: disable=no-self-use
|
||||
output_file = common.MakeTempFile()
|
||||
raw_image = common.MakeTempFile()
|
||||
with open(raw_image, 'wb') as f:
|
||||
f.write(raw_data)
|
||||
|
||||
cmd = ["img2simg", raw_image, output_file, '4096']
|
||||
p = common.Run(cmd)
|
||||
p.communicate()
|
||||
self.assertEqual(0, p.returncode)
|
||||
|
||||
common.RunAndCheckOutput(cmd)
|
||||
return output_file
|
||||
|
||||
def _generate_image(self):
|
||||
def _GenerateImage(self):
|
||||
partition_size = 1024 * 1024
|
||||
prop_dict = {
|
||||
'partition_size': str(partition_size),
|
||||
|
@ -79,11 +76,11 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
self.assertIsNotNone(verity_image_builder)
|
||||
adjusted_size = verity_image_builder.CalculateMaxImageSize()
|
||||
|
||||
raw_image = ""
|
||||
raw_image = bytearray(adjusted_size)
|
||||
for i in range(adjusted_size):
|
||||
raw_image += str(i % 10)
|
||||
raw_image[i] = ord('0') + i % 10
|
||||
|
||||
output_file = self._create_simg(raw_image)
|
||||
output_file = self._CreateSimg(raw_image)
|
||||
|
||||
# Append the verity metadata.
|
||||
verity_image_builder.Build(output_file)
|
||||
|
@ -92,7 +89,7 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
|
||||
@SkipIfExternalToolsUnavailable()
|
||||
def test_CreateHashtreeInfoGenerator(self):
|
||||
image_file = sparse_img.SparseImage(self._generate_image())
|
||||
image_file = sparse_img.SparseImage(self._GenerateImage())
|
||||
|
||||
generator = CreateHashtreeInfoGenerator(
|
||||
'system', image_file, self.prop_dict)
|
||||
|
@ -103,7 +100,7 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
|
||||
@SkipIfExternalToolsUnavailable()
|
||||
def test_DecomposeSparseImage(self):
|
||||
image_file = sparse_img.SparseImage(self._generate_image())
|
||||
image_file = sparse_img.SparseImage(self._GenerateImage())
|
||||
|
||||
generator = VerifiedBootVersion1HashtreeInfoGenerator(
|
||||
self.partition_size, 4096, True)
|
||||
|
@ -114,7 +111,7 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
|
||||
@SkipIfExternalToolsUnavailable()
|
||||
def test_ParseHashtreeMetadata(self):
|
||||
image_file = sparse_img.SparseImage(self._generate_image())
|
||||
image_file = sparse_img.SparseImage(self._GenerateImage())
|
||||
generator = VerifiedBootVersion1HashtreeInfoGenerator(
|
||||
self.partition_size, 4096, True)
|
||||
generator.DecomposeSparseImage(image_file)
|
||||
|
@ -131,12 +128,12 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
def test_ValidateHashtree_smoke(self):
|
||||
generator = VerifiedBootVersion1HashtreeInfoGenerator(
|
||||
self.partition_size, 4096, True)
|
||||
generator.image = sparse_img.SparseImage(self._generate_image())
|
||||
generator.image = sparse_img.SparseImage(self._GenerateImage())
|
||||
|
||||
generator.hashtree_info = info = HashtreeInfo()
|
||||
info.filesystem_range = RangeSet(data=[0, 991232 / 4096])
|
||||
info.filesystem_range = RangeSet(data=[0, 991232 // 4096])
|
||||
info.hashtree_range = RangeSet(
|
||||
data=[991232 / 4096, (991232 + 12288) / 4096])
|
||||
data=[991232 // 4096, (991232 + 12288) // 4096])
|
||||
info.hash_algorithm = self.hash_algorithm
|
||||
info.salt = self.fixed_salt
|
||||
info.root_hash = self.expected_root_hash
|
||||
|
@ -147,12 +144,12 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
def test_ValidateHashtree_failure(self):
|
||||
generator = VerifiedBootVersion1HashtreeInfoGenerator(
|
||||
self.partition_size, 4096, True)
|
||||
generator.image = sparse_img.SparseImage(self._generate_image())
|
||||
generator.image = sparse_img.SparseImage(self._GenerateImage())
|
||||
|
||||
generator.hashtree_info = info = HashtreeInfo()
|
||||
info.filesystem_range = RangeSet(data=[0, 991232 / 4096])
|
||||
info.filesystem_range = RangeSet(data=[0, 991232 // 4096])
|
||||
info.hashtree_range = RangeSet(
|
||||
data=[991232 / 4096, (991232 + 12288) / 4096])
|
||||
data=[991232 // 4096, (991232 + 12288) // 4096])
|
||||
info.hash_algorithm = self.hash_algorithm
|
||||
info.salt = self.fixed_salt
|
||||
info.root_hash = "a" + self.expected_root_hash[1:]
|
||||
|
@ -161,12 +158,12 @@ class VerifiedBootVersion1HashtreeInfoGeneratorTest(ReleaseToolsTestCase):
|
|||
|
||||
@SkipIfExternalToolsUnavailable()
|
||||
def test_Generate(self):
|
||||
image_file = sparse_img.SparseImage(self._generate_image())
|
||||
image_file = sparse_img.SparseImage(self._GenerateImage())
|
||||
generator = CreateHashtreeInfoGenerator('system', 4096, self.prop_dict)
|
||||
info = generator.Generate(image_file)
|
||||
|
||||
self.assertEqual(RangeSet(data=[0, 991232 / 4096]), info.filesystem_range)
|
||||
self.assertEqual(RangeSet(data=[991232 / 4096, (991232 + 12288) / 4096]),
|
||||
self.assertEqual(RangeSet(data=[0, 991232 // 4096]), info.filesystem_range)
|
||||
self.assertEqual(RangeSet(data=[991232 // 4096, (991232 + 12288) // 4096]),
|
||||
info.hashtree_range)
|
||||
self.assertEqual(self.hash_algorithm, info.hash_algorithm)
|
||||
self.assertEqual(self.fixed_salt, info.salt)
|
||||
|
|
|
@ -44,7 +44,7 @@ def _ReadFile(file_name, unpacked_name, round_up=False):
|
|||
"""Constructs and returns a File object. Rounds up its size if needed."""
|
||||
|
||||
assert os.path.exists(unpacked_name)
|
||||
with open(unpacked_name, 'r') as f:
|
||||
with open(unpacked_name, 'rb') as f:
|
||||
file_data = f.read()
|
||||
file_size = len(file_data)
|
||||
if round_up:
|
||||
|
|
|
@ -131,7 +131,8 @@ def Append(target, file_to_append, error_message):
|
|||
BuildVerityImageError: On error.
|
||||
"""
|
||||
try:
|
||||
with open(target, "a") as out_file, open(file_to_append, "r") as input_file:
|
||||
with open(target, 'ab') as out_file, \
|
||||
open(file_to_append, 'rb') as input_file:
|
||||
for line in input_file:
|
||||
out_file.write(line)
|
||||
except IOError:
|
||||
|
@ -178,6 +179,8 @@ def CreateVerityImageBuilder(prop_dict):
|
|||
# key_path and algorithm are only available when chain partition is used.
|
||||
key_path = prop_dict.get("avb_key_path")
|
||||
algorithm = prop_dict.get("avb_algorithm")
|
||||
|
||||
# Image uses hash footer.
|
||||
if prop_dict.get("avb_hash_enable") == "true":
|
||||
return VerifiedBootVersion2VerityImageBuilder(
|
||||
prop_dict["partition_name"],
|
||||
|
@ -188,16 +191,17 @@ def CreateVerityImageBuilder(prop_dict):
|
|||
algorithm,
|
||||
prop_dict.get("avb_salt"),
|
||||
prop_dict["avb_add_hash_footer_args"])
|
||||
else:
|
||||
return VerifiedBootVersion2VerityImageBuilder(
|
||||
prop_dict["partition_name"],
|
||||
partition_size,
|
||||
VerifiedBootVersion2VerityImageBuilder.AVB_HASHTREE_FOOTER,
|
||||
prop_dict["avb_avbtool"],
|
||||
key_path,
|
||||
algorithm,
|
||||
prop_dict.get("avb_salt"),
|
||||
prop_dict["avb_add_hashtree_footer_args"])
|
||||
|
||||
# Image uses hashtree footer.
|
||||
return VerifiedBootVersion2VerityImageBuilder(
|
||||
prop_dict["partition_name"],
|
||||
partition_size,
|
||||
VerifiedBootVersion2VerityImageBuilder.AVB_HASHTREE_FOOTER,
|
||||
prop_dict["avb_avbtool"],
|
||||
key_path,
|
||||
algorithm,
|
||||
prop_dict.get("avb_salt"),
|
||||
prop_dict["avb_add_hashtree_footer_args"])
|
||||
|
||||
return None
|
||||
|
||||
|
@ -605,19 +609,19 @@ class VerifiedBootVersion1HashtreeInfoGenerator(HashtreeInfoGenerator):
|
|||
self.metadata_size = metadata_size
|
||||
|
||||
self.hashtree_info.filesystem_range = RangeSet(
|
||||
data=[0, adjusted_size / self.block_size])
|
||||
data=[0, adjusted_size // self.block_size])
|
||||
self.hashtree_info.hashtree_range = RangeSet(
|
||||
data=[adjusted_size / self.block_size,
|
||||
(adjusted_size + verity_tree_size) / self.block_size])
|
||||
data=[adjusted_size // self.block_size,
|
||||
(adjusted_size + verity_tree_size) // self.block_size])
|
||||
|
||||
def _ParseHashtreeMetadata(self):
|
||||
"""Parses the hash_algorithm, root_hash, salt from the metadata block."""
|
||||
|
||||
metadata_start = self.filesystem_size + self.hashtree_size
|
||||
metadata_range = RangeSet(
|
||||
data=[metadata_start / self.block_size,
|
||||
(metadata_start + self.metadata_size) / self.block_size])
|
||||
meta_data = ''.join(self.image.ReadRangeSet(metadata_range))
|
||||
data=[metadata_start // self.block_size,
|
||||
(metadata_start + self.metadata_size) // self.block_size])
|
||||
meta_data = b''.join(self.image.ReadRangeSet(metadata_range))
|
||||
|
||||
# More info about the metadata structure available in:
|
||||
# system/extras/verity/build_verity_metadata.py
|
||||
|
@ -640,9 +644,9 @@ class VerifiedBootVersion1HashtreeInfoGenerator(HashtreeInfoGenerator):
|
|||
assert (int(table_entries[5]) * self.block_size == self.filesystem_size and
|
||||
int(table_entries[6]) * self.block_size == self.filesystem_size)
|
||||
|
||||
self.hashtree_info.hash_algorithm = table_entries[7]
|
||||
self.hashtree_info.root_hash = table_entries[8]
|
||||
self.hashtree_info.salt = table_entries[9]
|
||||
self.hashtree_info.hash_algorithm = table_entries[7].decode()
|
||||
self.hashtree_info.root_hash = table_entries[8].decode()
|
||||
self.hashtree_info.salt = table_entries[9].decode()
|
||||
|
||||
def ValidateHashtree(self):
|
||||
"""Checks that we can reconstruct the verity hash tree."""
|
||||
|
@ -669,8 +673,8 @@ class VerifiedBootVersion1HashtreeInfoGenerator(HashtreeInfoGenerator):
|
|||
|
||||
# Reads the generated hash tree and checks if it has the exact same bytes
|
||||
# as the one in the sparse image.
|
||||
with open(generated_verity_tree, "rb") as fd:
|
||||
return fd.read() == ''.join(self.image.ReadRangeSet(
|
||||
with open(generated_verity_tree, 'rb') as fd:
|
||||
return fd.read() == b''.join(self.image.ReadRangeSet(
|
||||
self.hashtree_info.hashtree_range))
|
||||
|
||||
def Generate(self, image):
|
||||
|
|
Loading…
Reference in New Issue