pull/1086/head
Piero Toffanin 2021-11-04 11:17:37 -04:00
rodzic 34736e63b0
commit 3378e67a2a
4 zmienionych plików z 360 dodań i 11 usunięć

Wyświetl plik

@ -340,6 +340,16 @@ def download_file_response(request, filePath, content_disposition, download_file
return response
def download_file_stream(request, stream, content_disposition, download_filename=None):
response = HttpResponse(FileWrapper(stream),
content_type=(mimetypes.guess_type(download_filename)[0] or "application/zip"))
response['Content-Type'] = mimetypes.guess_type(download_filename)[0] or "application/zip"
response['Content-Disposition'] = "{}; filename={}".format(content_disposition, download_filename)
return response
"""
Task downloads are simply aliases to download the task's assets
(but require a shorter path and look nicer the API user)
@ -353,14 +363,17 @@ class TaskDownloads(TaskNestedView):
# Check and download
try:
asset_path = task.get_asset_download_path(asset)
asset_fs, is_zipstream = task.get_asset_file_or_zipstream(asset)
except FileNotFoundError:
raise exceptions.NotFound(_("Asset does not exist"))
if not os.path.exists(asset_path):
if not is_zipstream and not os.path.isfile(asset_fs):
raise exceptions.NotFound(_("Asset does not exist"))
return download_file_response(request, asset_path, 'attachment', download_filename=request.GET.get('filename'))
if not is_zipstream:
return download_file_response(request, asset_fs, 'attachment', download_filename=request.GET.get('filename'))
else:
return download_file_stream(request, asset_fs, 'attachment', download_filename=request.GET.get('filename', asset))
"""
Raw access to the task's asset folder resources

Wyświetl plik

@ -29,6 +29,19 @@ def update_epsg_fields(apps, schema_editor):
t.epsg = epsg
t.save()
def remove_all_zip(apps, schema_editor):
Task = apps.get_model('app', 'Task')
for t in Task.objects.all():
asset = 'all.zip'
asset_path = os.path.join(settings.MEDIA_ROOT, "project", str(t.project.id), "task", str(t.id), "assets", asset)
if os.path.isfile(asset_path):
try:
os.remove(asset_path)
print("Cleaned up {}".format(asset_path))
except Exception as e:
print(e)
class Migration(migrations.Migration):
dependencies = [
@ -43,4 +56,6 @@ class Migration(migrations.Migration):
),
migrations.RunPython(update_epsg_fields),
migrations.RunPython(remove_all_zip),
]

Wyświetl plik

@ -3,6 +3,7 @@ import os
import shutil
import time
import uuid as uuid_module
from app.vendor import zipfly
import json
from shlex import quote
@ -168,7 +169,10 @@ def resize_image(image_path, resize_to, done=None):
class Task(models.Model):
ASSETS_MAP = {
'all.zip': 'all.zip',
'all.zip': {
'deferred_path': 'all.zip',
'deferred_compress_dir': '.'
},
'orthophoto.tif': os.path.join('odm_orthophoto', 'odm_orthophoto.tif'),
'orthophoto.png': os.path.join('odm_orthophoto', 'odm_orthophoto.png'),
'orthophoto.mbtiles': os.path.join('odm_orthophoto', 'odm_orthophoto.mbtiles'),
@ -436,14 +440,36 @@ class Task(models.Model):
logger.warning("Cannot duplicate task: {}".format(str(e)))
return False
def get_asset_file_or_zipstream(self, asset):
"""
Get a stream to an asset
:param asset: one of ASSETS_MAP keys
:return: (path|stream, is_zipstream:bool)
"""
if asset in self.ASSETS_MAP:
value = self.ASSETS_MAP[asset]
if isinstance(value, str):
return self.assets_path(value), False
elif isinstance(value, dict):
if 'deferred_path' in value and 'deferred_compress_dir' in value:
zip_dir = self.assets_path(value['deferred_compress_dir'])
paths = [{'n': os.path.relpath(os.path.join(dp, f), zip_dir), 'fs': os.path.join(dp, f)} for dp, dn, filenames in os.walk(zip_dir) for f in filenames]
return zipfly.ZipStream(paths), True
else:
raise FileNotFoundError("{} is not a valid asset (invalid dict values)".format(asset))
else:
raise FileNotFoundError("{} is not a valid asset (invalid map)".format(asset))
else:
raise FileNotFoundError("{} is not a valid asset".format(asset))
def get_asset_download_path(self, asset):
"""
Get the path to an asset download
:param asset: one of ASSETS_MAP keys
:return: path
"""
if asset in self.ASSETS_MAP:
value = self.ASSETS_MAP[asset]
if isinstance(value, str):
@ -451,10 +477,9 @@ class Task(models.Model):
elif isinstance(value, dict):
if 'deferred_path' in value and 'deferred_compress_dir' in value:
return self.generate_deferred_asset(value['deferred_path'], value['deferred_compress_dir'])
return value['deferred_path']
else:
raise FileNotFoundError("{} is not a valid asset (invalid dict values)".format(asset))
else:
raise FileNotFoundError("{} is not a valid asset (invalid map)".format(asset))
else:
@ -812,6 +837,9 @@ class Task(models.Model):
logger.info("Extracted all.zip for {}".format(self))
# Remove zip
os.remove(zip_path)
# Populate *_extent fields
extent_fields = [
(os.path.realpath(self.assets_path("odm_orthophoto", "odm_orthophoto.tif")),
@ -906,10 +934,11 @@ class Task(models.Model):
'public': self.public
}
def generate_deferred_asset(self, archive, directory):
def generate_deferred_asset(self, archive, directory, stream=False):
"""
:param archive: path of the destination .zip file (relative to /assets/ directory)
:param directory: path of the source directory to compress (relative to /assets/ directory)
:param stream: return a stream instead of a path to the file
:return: full path of the generated archive
"""
archive_path = self.assets_path(archive)

292
app/vendor/zipfly.py vendored 100644
Wyświetl plik

@ -0,0 +1,292 @@
# Modified from https://github.com/BuzonIO/zipfly
# This library was created by Buzon.io and is released under the MIT. Copyright 2021 Cardallot, Inc.
# -*- coding: utf-8 -*-
__version__ = '6.0.4'
# v
import io
import stat
import zipfile
ZIP64_LIMIT = (1 << 31) + 1
class LargePredictionSize(Exception):
"""
Raised when Buffer is larger than ZIP64
"""
class ZipflyStream(io.RawIOBase):
"""
The RawIOBase ABC extends IOBase. It deals with
the reading and writing of bytes to a stream. FileIO subclasses
RawIOBase to provide an interface to files in the machines file system.
"""
def __init__(self):
self._buffer = b''
self._size = 0
def writable(self):
return True
def write(self, b):
if self.closed:
raise RuntimeError("ZipFly stream was closed!")
self._buffer += b
return len(b)
def get(self):
chunk = self._buffer
self._buffer = b''
self._size += len(chunk)
return chunk
def size(self):
return self._size
class ZipFly:
def __init__(self,
mode = 'w',
paths = [],
chunksize = 0x8000,
compression = zipfile.ZIP_STORED,
allowZip64 = True,
compresslevel = None,
storesize = 0,
filesystem = 'fs',
arcname = 'n',
encode = 'utf-8',):
"""
@param store size : int : size of all files
in paths without compression
"""
if mode not in ('w',):
raise RuntimeError("ZipFly requires 'w' mode")
if compression not in ( zipfile.ZIP_STORED,):
raise RuntimeError("Not compression supported")
if compresslevel not in (None, ):
raise RuntimeError("Not compression level supported")
if isinstance(chunksize, str):
chunksize = int(chunksize, 16)
self.comment = f'Written using WebODM'
self.mode = mode
self.paths = paths
self.filesystem = filesystem
self.arcname = arcname
self.compression = compression
self.chunksize = chunksize
self.allowZip64 = allowZip64
self.compresslevel = compresslevel
self.storesize = storesize
self.encode = encode
self.ezs = int('0x8e', 16) # empty zip size in bytes
def set_comment(self, comment):
if not isinstance(comment, bytes):
comment = str.encode(comment)
if len(comment) >= zipfile.ZIP_MAX_COMMENT:
# trunk comment
comment = comment[:zipfile.ZIP_MAX_COMMENT]
self.comment = comment
def reader(self, entry):
def get_chunk():
return entry.read( self.chunksize )
return get_chunk()
def buffer_size(self):
'''
FOR UNIT TESTING (not used)
using to get the buffer size
this size is different from the size of each file added
'''
for i in self.generator(): pass
return self._buffer_size
def buffer_prediction_size(self):
if not self.allowZip64:
raise RuntimeError("ZIP64 extensions required")
# End of Central Directory Record
EOCD = int('0x16', 16)
FILE_OFFSET = int('0x5e', 16) * len(self.paths)
tmp_comment = self.comment
if isinstance(self.comment, bytes):
tmp_comment = ( self.comment ).decode()
size_comment = len(tmp_comment.encode( self.encode ))
# path-name
size_paths = 0
#for path in self.paths:
for idx in range(len(self.paths)):
'''
getting bytes from character in UTF-8 format
example:
'' has 3 bytes in utf-8 format ( b'\xe4\xbc\xa0' )
'''
#path = paths[idx]
name = self.arcname
if not self.arcname in self.paths[idx]:
name = self.filesystem
tmp_name = self.paths[idx][name]
if (tmp_name)[0] in ('/', ):
# is dir then trunk
tmp_name = (tmp_name)[ 1 : len( tmp_name ) ]
size_paths += (
len(
tmp_name.encode( self.encode )
) - int( '0x1', 16)
) * int('0x2', 16)
# zipsize
zs = sum([
EOCD,
FILE_OFFSET,
size_comment,
size_paths,
self.storesize,
])
if zs > ZIP64_LIMIT:
raise LargePredictionSize(
"Prediction size for zip file greater than 2 GB not supported"
)
return zs
def generator(self):
# stream
stream = ZipflyStream()
with zipfile.ZipFile(
stream,
mode = self.mode,
compression = self.compression,
allowZip64 = self.allowZip64,) as zf:
for path in self.paths:
if not self.filesystem in path:
raise RuntimeError(
f" '{self.filesystem}' key is required "
)
"""
filesystem should be the path to a file or directory on the filesystem.
arcname is the name which it will have within the archive (by default,
this will be the same as filename
"""
if not self.arcname in path:
# arcname will be default path
path[self.arcname] = path[self.filesystem]
z_info = zipfile.ZipInfo.from_file(
path[self.filesystem],
path[self.arcname]
)
with open( path[self.filesystem], 'rb' ) as e:
# Read from filesystem:
with zf.open( z_info, mode = self.mode ) as d:
"""
buffer = b''
while True:
chunk = e.read(self.chunksize)
if not chunk:
break
buffer += chunk
elements = buffer.split(b'\0')
for element in elements[:-1]:
d.write( element )
yield stream.get()
buffer = elements[-1]
if buffer:
# d.write( buffer )
yield stream.get()
"""
for chunk in iter( lambda: e.read( self.chunksize ), b'' ):
# (e.read( ... )) this get a small chunk of the file
# and return a callback to the next iterator
d.write( chunk )
yield stream.get()
self.set_comment(self.comment)
zf.comment = self.comment
# last chunk
yield stream.get()
# (TESTING)
# get the real size of the zipfile
self._buffer_size = stream.size()
# Flush and close this stream.
stream.close()
def get_size(self):
return self._buffer_size
class ZipStream:
def __init__(self, paths):
self.paths = paths
self.generator = None
def lazy_load(self, chunksize):
if self.generator is None:
zfly = ZipFly(paths=self.paths, mode='w', chunksize=chunksize)
self.generator = zfly.generator()
def read(self, count):
self.lazy_load(count)
return next(self.generator)