148 lines
5.8 KiB
Python
148 lines
5.8 KiB
Python
#!/usr/bin/env python
|
|
#
|
|
# Copyright (C) 2020 The Android Open Source Project
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
"""Unittests for parsing files in zip64 format"""
|
|
|
|
import os
|
|
import subprocess
|
|
import tempfile
|
|
import unittest
|
|
import zipfile
|
|
import time
|
|
|
|
class Zip64Test(unittest.TestCase):
|
|
@staticmethod
|
|
def _WriteFile(path, size_in_kib):
|
|
contents = os.path.basename(path)[0] * 1024
|
|
with open(path, 'w') as f:
|
|
for it in range(0, size_in_kib):
|
|
f.write(contents)
|
|
|
|
@staticmethod
|
|
def _AddEntriesToZip(output_zip, entries_dict=None):
|
|
for name, size in entries_dict.items():
|
|
file_path = tempfile.NamedTemporaryFile()
|
|
Zip64Test._WriteFile(file_path.name, size)
|
|
output_zip.write(file_path.name, arcname = name)
|
|
|
|
def _getEntryNames(self, zip_name):
|
|
cmd = ['ziptool', 'zipinfo', '-1', zip_name]
|
|
proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
|
output, _ = proc.communicate()
|
|
self.assertEquals(0, proc.returncode)
|
|
self.assertNotEqual(None, output)
|
|
return output.split()
|
|
|
|
def _ExtractEntries(self, zip_name):
|
|
temp_dir = tempfile.mkdtemp()
|
|
cmd = ['ziptool', 'unzip', '-d', temp_dir, zip_name]
|
|
proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
|
proc.communicate()
|
|
self.assertEquals(0, proc.returncode)
|
|
|
|
def test_entriesSmallerThan2G(self):
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
# Add a few entries with each of them smaller than 2GiB. But the entire zip file is larger
|
|
# than 4GiB in size.
|
|
with zipfile.ZipFile(zip_path, 'w', allowZip64=True) as output_zip:
|
|
entry_dict = {'a.txt': 1025 * 1024, 'b.txt': 1025 * 1024, 'c.txt': 1025 * 1024,
|
|
'd.txt': 1025 * 1024, 'e.txt': 1024}
|
|
self._AddEntriesToZip(output_zip, entry_dict)
|
|
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals(sorted(entry_dict.keys()), sorted(read_names))
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
def test_largeNumberOfEntries(self):
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
entry_dict = {}
|
|
# Add 100k entries (more than 65535|UINT16_MAX).
|
|
for num in range(0, 100 * 1024):
|
|
entry_dict[str(num)] = 50
|
|
|
|
with zipfile.ZipFile(zip_path, 'w', allowZip64=True) as output_zip:
|
|
self._AddEntriesToZip(output_zip, entry_dict)
|
|
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals(sorted(entry_dict.keys()), sorted(read_names))
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
def test_largeCompressedEntriesSmallerThan4G(self):
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
with zipfile.ZipFile(zip_path, 'w', compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True) as output_zip:
|
|
# Add entries close to 4GiB in size. Somehow the python library will put the (un)compressed
|
|
# sizes in the extra field. Test if our ziptool should be able to parse it.
|
|
entry_dict = {'e.txt': 4095 * 1024, 'f.txt': 4095 * 1024}
|
|
self._AddEntriesToZip(output_zip, entry_dict)
|
|
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals(sorted(entry_dict.keys()), sorted(read_names))
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
def test_forceDataDescriptor(self):
|
|
file_path = tempfile.NamedTemporaryFile(suffix='.txt')
|
|
self._WriteFile(file_path.name, 5000 * 1024)
|
|
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
with zipfile.ZipFile(zip_path, 'w', allowZip64=True) as output_zip:
|
|
pass
|
|
# The fd option force writes a data descriptor
|
|
cmd = ['zip', '-fd', zip_path.name, file_path.name]
|
|
proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)
|
|
proc.communicate()
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals([file_path.name[1:]], read_names)
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
def test_largeUncompressedEntriesLargerThan4G(self):
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
with zipfile.ZipFile(zip_path, 'w', compression=zipfile.ZIP_STORED,
|
|
allowZip64=True) as output_zip:
|
|
# Add entries close to 4GiB in size. Somehow the python library will put the (un)compressed
|
|
# sizes in the extra field. Test if our ziptool should be able to parse it.
|
|
entry_dict = {'g.txt': 5000 * 1024, 'h.txt': 6000 * 1024}
|
|
self._AddEntriesToZip(output_zip, entry_dict)
|
|
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals(sorted(entry_dict.keys()), sorted(read_names))
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
def test_largeCompressedEntriesLargerThan4G(self):
|
|
zip_path = tempfile.NamedTemporaryFile(suffix='.zip')
|
|
with zipfile.ZipFile(zip_path, 'w', compression=zipfile.ZIP_DEFLATED,
|
|
allowZip64=True) as output_zip:
|
|
# Add entries close to 4GiB in size. Somehow the python library will put the (un)compressed
|
|
# sizes in the extra field. Test if our ziptool should be able to parse it.
|
|
entry_dict = {'i.txt': 4096 * 1024, 'j.txt': 7000 * 1024}
|
|
self._AddEntriesToZip(output_zip, entry_dict)
|
|
|
|
read_names = self._getEntryNames(zip_path.name)
|
|
self.assertEquals(sorted(entry_dict.keys()), sorted(read_names))
|
|
self._ExtractEntries(zip_path.name)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
testsuite = unittest.TestLoader().discover(
|
|
os.path.dirname(os.path.realpath(__file__)))
|
|
unittest.TextTestRunner(verbosity=2).run(testsuite)
|