kopia lustrzana https://github.com/OpenDroneMap/WebODM
805 wiersze
32 KiB
Python
805 wiersze
32 KiB
Python
import os
|
|
import time
|
|
|
|
|
|
import logging
|
|
from datetime import timedelta
|
|
|
|
import json
|
|
import requests
|
|
from PIL import Image
|
|
from django.contrib.auth.models import User
|
|
from rest_framework import status
|
|
from rest_framework.test import APIClient
|
|
|
|
import worker
|
|
from django.utils import timezone
|
|
|
|
from app import pending_actions
|
|
from app.models import Project, Task, ImageUpload
|
|
from app.models.task import task_directory_path, full_task_directory_path, TaskInterruptedException
|
|
from app.plugins.signals import task_completed, task_removed, task_removing
|
|
from app.tests.classes import BootTransactionTestCase
|
|
from nodeodm import status_codes
|
|
from nodeodm.models import ProcessingNode, OFFLINE_MINUTES
|
|
from app.testwatch import testWatch
|
|
from .utils import start_processing_node, clear_test_media_root, catch_signal
|
|
|
|
# We need to test the task API in a TransactionTestCase because
|
|
# task processing happens on a separate thread, and normal TestCases
|
|
# do not commit changes to the DB, so spawning a new thread will show no
|
|
# data in it.
|
|
from webodm import settings
|
|
logger = logging.getLogger('app.logger')
|
|
|
|
DELAY = 2 # time to sleep for during process launch, background processing, etc.
|
|
|
|
class TestApiTask(BootTransactionTestCase):
|
|
def setUp(self):
|
|
super().setUp()
|
|
clear_test_media_root()
|
|
|
|
def test_task(self):
|
|
client = APIClient()
|
|
|
|
node_odm = start_processing_node()
|
|
|
|
user = User.objects.get(username="testuser")
|
|
self.assertFalse(user.is_superuser)
|
|
|
|
other_user = User.objects.get(username="testuser2")
|
|
|
|
project = Project.objects.create(
|
|
owner=user,
|
|
name="test project"
|
|
)
|
|
other_project = Project.objects.create(
|
|
owner=other_user,
|
|
name="another test project"
|
|
)
|
|
other_task = Task.objects.create(project=other_project)
|
|
|
|
# Start processing node
|
|
|
|
# Create processing node
|
|
pnode = ProcessingNode.objects.create(hostname="localhost", port=11223)
|
|
|
|
# Verify that it's working
|
|
self.assertTrue(pnode.api_version is not None)
|
|
|
|
# task creation via file upload
|
|
image1 = open("app/fixtures/tiny_drone_image.jpg", 'rb')
|
|
image2 = open("app/fixtures/tiny_drone_image_2.jpg", 'rb')
|
|
|
|
img1 = Image.open("app/fixtures/tiny_drone_image.jpg")
|
|
|
|
# Not authenticated?
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2]
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_403_FORBIDDEN);
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
client.login(username="testuser", password="test1234")
|
|
|
|
# Cannot create a task for a project that does not exist
|
|
res = client.post("/api/projects/0/tasks/", {
|
|
'images': [image1, image2]
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Cannot create a task for a project for which we have no access to
|
|
res = client.post("/api/projects/{}/tasks/".format(other_project.id), {
|
|
'images': [image1, image2]
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Cannot create a task without images
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': []
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_400_BAD_REQUEST)
|
|
|
|
# Cannot create a task with just 1 image
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': image1
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_400_BAD_REQUEST)
|
|
image1.seek(0)
|
|
|
|
# Normal case with images[], name and processing node parameter
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2],
|
|
'name': 'test_task',
|
|
'processing_node': pnode.id
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
multiple_param_task = Task.objects.latest('created_at')
|
|
self.assertTrue(multiple_param_task.name == 'test_task')
|
|
self.assertTrue(multiple_param_task.processing_node.id == pnode.id)
|
|
self.assertEqual(multiple_param_task.import_url, "")
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Uploaded images should be the same size as originals
|
|
with Image.open(multiple_param_task.task_path("tiny_drone_image.jpg")) as im:
|
|
self.assertTrue(im.size == img1.size)
|
|
|
|
# Normal case with images[], GCP, name and processing node parameter and resize_to option
|
|
testWatch.clear()
|
|
gcp = open("app/fixtures/gcp.txt", 'r')
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2, gcp],
|
|
'name': 'test_task',
|
|
'processing_node': pnode.id,
|
|
'resize_to': img1.size[0] / 2.0
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
resized_task = Task.objects.latest('created_at')
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
gcp.seek(0)
|
|
|
|
# Uploaded images should have been resized
|
|
with Image.open(resized_task.task_path("tiny_drone_image.jpg")) as im:
|
|
self.assertTrue(im.size[0] == img1.size[0] / 2.0)
|
|
|
|
# GCP should have been scaled
|
|
with open(resized_task.task_path("gcp.txt")) as f:
|
|
lines = list(map(lambda l: l.strip(), f.readlines()))
|
|
|
|
[x, y, z, px, py, imagename, *extras] = lines[1].split(' ')
|
|
self.assertTrue(imagename == "tiny_drone_image.JPG") # case insensitive
|
|
self.assertTrue(float(px) == 2.0) # scaled by half
|
|
self.assertTrue(float(py) == 3.0) # scaled by half
|
|
self.assertTrue(float(x) == 576529.22) # Didn't change
|
|
|
|
[x, y, z, px, py, imagename, *extras] = lines[5].split(' ')
|
|
self.assertTrue(imagename == "missing_image.jpg")
|
|
self.assertTrue(float(px) == 8.0) # Didn't change
|
|
self.assertTrue(float(py) == 8.0) # Didn't change
|
|
|
|
# Resize progress is 100%
|
|
resized_task.refresh_from_db()
|
|
self.assertEqual(resized_task.resize_progress, 1.0)
|
|
|
|
# Upload progress is 100%
|
|
self.assertEqual(resized_task.upload_progress, 1.0)
|
|
|
|
# Upload progress callback has been called
|
|
self.assertTrue(testWatch.get_calls_count("Task.process.callback") > 0)
|
|
|
|
# This is not a partial task
|
|
self.assertFalse(resized_task.partial)
|
|
|
|
# Case with malformed GCP file option
|
|
with open("app/fixtures/gcp_malformed.txt", 'r') as malformed_gcp:
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2, malformed_gcp],
|
|
'name': 'test_task',
|
|
'processing_node': pnode.id,
|
|
'resize_to': img1.size[0] / 2.0
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
malformed_gcp_task = Task.objects.latest('created_at')
|
|
|
|
# We just pass it along, it will get errored out during processing
|
|
# But we shouldn't fail.
|
|
with open(malformed_gcp_task.task_path("gcp_malformed.txt")) as f:
|
|
lines = list(map(lambda l: l.strip(), f.readlines()))
|
|
self.assertTrue(lines[1] == "<O_O>")
|
|
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Cannot create a task with images[], name, but invalid processing node parameter
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2],
|
|
'name': 'test_task',
|
|
'processing_node': 9999
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_400_BAD_REQUEST)
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Normal case with images[] parameter
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2],
|
|
'auto_processing_node': 'false'
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Should have returned the id of the newly created task
|
|
task = Task.objects.latest('created_at')
|
|
self.assertTrue('id' in res.data)
|
|
self.assertTrue(str(task.id) == res.data['id'])
|
|
|
|
# Progress is at 0%
|
|
self.assertEqual(task.running_progress, 0.0)
|
|
|
|
# Two images should have been uploaded
|
|
self.assertTrue(ImageUpload.objects.filter(task=task).count() == 2)
|
|
|
|
# Can_rerun_from should be an empty list
|
|
self.assertTrue(len(res.data['can_rerun_from']) == 0)
|
|
|
|
# processing_node_name should be null
|
|
self.assertTrue(res.data['processing_node_name'] is None)
|
|
|
|
# No processing node is set
|
|
self.assertTrue(task.processing_node is None)
|
|
|
|
# tiles.json should not be accessible at this point
|
|
tile_types = ['orthophoto', 'dsm', 'dtm']
|
|
for tile_type in tile_types:
|
|
res = client.get("/api/projects/{}/tasks/{}/{}/tiles.json".format(project.id, task.id, tile_type))
|
|
self.assertTrue(res.status_code == status.HTTP_400_BAD_REQUEST)
|
|
|
|
# Neither should an individual tile
|
|
# Z/X/Y coords are chosen based on node-odm test dataset for orthophoto_tiles/
|
|
res = client.get("/api/projects/{}/tasks/{}/orthophoto/tiles/16/16020/42443.png".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# Cannot access a tiles.json we have no access to
|
|
res = client.get("/api/projects/{}/tasks/{}/orthophoto/tiles.json".format(other_project.id, other_task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# Cannot access an individual tile we have no access to
|
|
res = client.get("/api/projects/{}/tasks/{}/orthophoto/tiles/16/16020/42443.png".format(other_project.id, other_task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# Cannot download assets (they don't exist yet)
|
|
for asset in list(task.ASSETS_MAP.keys()):
|
|
res = client.get("/api/projects/{}/tasks/{}/download/{}".format(project.id, task.id, asset))
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# Cannot access raw assets (they don't exist yet)
|
|
res = client.get("/api/projects/{}/tasks/{}/assets/odm_orthophoto/odm_orthophoto.tif".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# Cannot assign processing node to a task we have no access to
|
|
res = client.patch("/api/projects/{}/tasks/{}/".format(other_project.id, other_task.id), {
|
|
'processing_node': pnode.id
|
|
})
|
|
self.assertTrue(res.status_code == status.HTTP_404_NOT_FOUND)
|
|
|
|
# No UUID at this point
|
|
self.assertTrue(len(task.uuid) == 0)
|
|
|
|
# Assign processing node to task via API
|
|
res = client.patch("/api/projects/{}/tasks/{}/".format(project.id, task.id), {
|
|
'processing_node': pnode.id
|
|
})
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# On update worker.tasks.process_pending_tasks should have been called in the background
|
|
# (during tests this is sync)
|
|
|
|
# Processing should have started and a UUID is assigned
|
|
# Calling process pending tasks should finish the process
|
|
# and invoke the plugins completed signal
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.status in [status_codes.RUNNING, status_codes.COMPLETED]) # Sometimes this finishes before we get here
|
|
self.assertTrue(len(task.uuid) > 0)
|
|
|
|
with catch_signal(task_completed) as handler:
|
|
retry_count = 0
|
|
while task.status != status_codes.COMPLETED:
|
|
worker.tasks.process_pending_tasks()
|
|
time.sleep(DELAY)
|
|
task.refresh_from_db()
|
|
retry_count += 1
|
|
if retry_count > 10:
|
|
break
|
|
|
|
self.assertEqual(task.status, status_codes.COMPLETED)
|
|
|
|
# Progress is 100%
|
|
self.assertTrue(task.running_progress == 1.0)
|
|
|
|
handler.assert_any_call(
|
|
sender=Task,
|
|
task_id=task.id,
|
|
signal=task_completed,
|
|
)
|
|
|
|
# Processing node should have a "rerun_from" option
|
|
pnode_rerun_from_opts = list(filter(lambda d: 'name' in d and d['name'] == 'rerun-from', pnode.available_options))[0]
|
|
self.assertTrue(len(pnode_rerun_from_opts['domain']) > 0)
|
|
|
|
# The can_rerun_from field of a task should now be populated
|
|
# with the same values as the "rerun_from" domain values of
|
|
# the processing node
|
|
res = client.get("/api/projects/{}/tasks/{}/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
self.assertTrue(pnode_rerun_from_opts['domain'] == res.data['can_rerun_from'])
|
|
|
|
# processing_node_name should be the name of the pnode
|
|
self.assertEqual(res.data['processing_node_name'], str(pnode))
|
|
|
|
# Can download assets
|
|
for asset in list(task.ASSETS_MAP.keys()):
|
|
res = client.get("/api/projects/{}/tasks/{}/download/{}".format(project.id, task.id, asset))
|
|
self.assertEqual(res.status_code, status.HTTP_200_OK)
|
|
|
|
# We can stream downloads
|
|
res = client.get("/api/projects/{}/tasks/{}/download/{}?_force_stream=1".format(project.id, task.id, list(task.ASSETS_MAP.keys())[0]))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
self.assertTrue(res.has_header('_stream'))
|
|
|
|
# A textured mesh archive file should exist
|
|
self.assertTrue(os.path.exists(task.assets_path(task.ASSETS_MAP["textured_model.zip"]["deferred_path"])))
|
|
|
|
# Tiles archives should have been created
|
|
self.assertTrue(os.path.exists(task.assets_path(task.ASSETS_MAP["dsm_tiles.zip"]["deferred_path"])))
|
|
self.assertTrue(os.path.exists(task.assets_path(task.ASSETS_MAP["dtm_tiles.zip"]["deferred_path"])))
|
|
self.assertTrue(os.path.exists(task.assets_path(task.ASSETS_MAP["orthophoto_tiles.zip"]["deferred_path"])))
|
|
|
|
# Can download raw assets
|
|
res = client.get("/api/projects/{}/tasks/{}/assets/odm_orthophoto/odm_orthophoto.tif".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# Can access tiles.json
|
|
for tile_type in tile_types:
|
|
res = client.get("/api/projects/{}/tasks/{}/{}/tiles.json".format(project.id, task.id, tile_type))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# Bounds are what we expect them to be
|
|
# (4 coords in lat/lon)
|
|
tiles = json.loads(res.content.decode("utf-8"))
|
|
self.assertTrue(len(tiles['bounds']) == 4)
|
|
self.assertTrue(round(tiles['bounds'][0], 7) == -91.9945132)
|
|
|
|
# Can access individual tiles
|
|
for tile_type in tile_types:
|
|
res = client.get("/api/projects/{}/tasks/{}/{}/tiles/16/16020/42443.png".format(project.id, task.id, tile_type))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# Another user does not have access to the resources
|
|
other_client = APIClient()
|
|
other_client.login(username="testuser2", password="test1234")
|
|
|
|
def accessResources(expectedStatus):
|
|
for tile_type in tile_types:
|
|
res = other_client.get("/api/projects/{}/tasks/{}/{}/tiles.json".format(project.id, task.id, tile_type))
|
|
self.assertTrue(res.status_code == expectedStatus)
|
|
|
|
res = other_client.get("/api/projects/{}/tasks/{}/{}/tiles/16/16020/42443.png".format(project.id, task.id, tile_type))
|
|
self.assertTrue(res.status_code == expectedStatus)
|
|
|
|
res = other_client.get("/api/projects/{}/tasks/{}/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == expectedStatus)
|
|
|
|
accessResources(status.HTTP_404_NOT_FOUND)
|
|
|
|
# Original owner enables sharing
|
|
res = client.patch("/api/projects/{}/tasks/{}/".format(project.id, task.id), {
|
|
'public': True
|
|
})
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# Now other user can acccess resources
|
|
accessResources(status.HTTP_200_OK)
|
|
|
|
# He cannot change a task
|
|
res = other_client.patch("/api/projects/{}/tasks/{}/".format(project.id, task.id), {
|
|
'name': "Changed! Uh oh"
|
|
})
|
|
self.assertEqual(res.status_code, status.HTTP_404_NOT_FOUND)
|
|
|
|
# User logs out
|
|
other_client.logout()
|
|
|
|
# He can still access the resources as anonymous
|
|
accessResources(status.HTTP_200_OK)
|
|
|
|
# Restart a task
|
|
testWatch.clear()
|
|
res = client.post("/api/projects/{}/tasks/{}/restart/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
# process_task is called in the background
|
|
task.refresh_from_db()
|
|
|
|
self.assertTrue(task.status in [status_codes.RUNNING, status_codes.COMPLETED])
|
|
|
|
# Should return without issues
|
|
task.check_if_canceled()
|
|
|
|
# Cancel a task
|
|
res = client.post("/api/projects/{}/tasks/{}/cancel/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# task is processed right away
|
|
|
|
# Should have been canceled
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.status == status_codes.CANCELED)
|
|
self.assertTrue(task.pending_action is None)
|
|
|
|
# Manually set pending action
|
|
task.pending_action = pending_actions.CANCEL
|
|
task.save()
|
|
|
|
# Should raise TaskInterruptedException
|
|
self.assertRaises(TaskInterruptedException, task.check_if_canceled)
|
|
|
|
# Restore
|
|
task.pending_action = None
|
|
task.save()
|
|
|
|
# Remove a task and verify that it calls the proper plugins signals
|
|
with catch_signal(task_removing) as h1:
|
|
with catch_signal(task_removed) as h2:
|
|
res = client.post("/api/projects/{}/tasks/{}/remove/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
h1.assert_called_once_with(sender=Task, task_id=task.id, signal=task_removing)
|
|
h2.assert_called_once_with(sender=Task, task_id=task.id, signal=task_removed)
|
|
|
|
# task is processed right away
|
|
|
|
# Has been removed along with assets
|
|
self.assertFalse(Task.objects.filter(pk=task.id).exists())
|
|
self.assertFalse(ImageUpload.objects.filter(task=task).exists())
|
|
|
|
task_assets_path = os.path.join(settings.MEDIA_ROOT, task_directory_path(task.id, task.project.id))
|
|
self.assertFalse(os.path.exists(task_assets_path))
|
|
|
|
# Stop processing node
|
|
node_odm.terminate()
|
|
|
|
# Create a task
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2],
|
|
'name': 'test_task_offline',
|
|
'processing_node': pnode.id,
|
|
'auto_processing_node': 'false'
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
task = Task.objects.get(pk=res.data['id'])
|
|
image1.seek(0)
|
|
image2.seek(0)
|
|
|
|
# Processing should fail and set an error
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.last_error is not None)
|
|
self.assertTrue(task.status == status_codes.FAILED)
|
|
|
|
# Now bring it back online
|
|
node_odm = start_processing_node()
|
|
|
|
# Restart
|
|
res = client.post("/api/projects/{}/tasks/{}/restart/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
task.refresh_from_db()
|
|
|
|
# After processing, the task should have restarted, and have no UUID or status
|
|
self.assertTrue(task.status is None)
|
|
self.assertTrue(len(task.uuid) == 0)
|
|
|
|
# Another step and it should have acquired a UUID
|
|
worker.tasks.process_pending_tasks()
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.status in [status_codes.RUNNING, status_codes.COMPLETED])
|
|
self.assertTrue(len(task.uuid) > 0)
|
|
|
|
# Another step and it should be completed
|
|
time.sleep(DELAY)
|
|
worker.tasks.process_pending_tasks()
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.status == status_codes.COMPLETED)
|
|
|
|
|
|
# Test rerun-from clearing mechanism:
|
|
|
|
# 1 .Set some task options, including rerun_from
|
|
task.options = [{'name': 'mesh-size', 'value':1000},
|
|
{'name': 'rerun-from', 'value': 'odm_meshing'}]
|
|
task.save()
|
|
|
|
# 2. Remove the task directly from node-odm (simulate a task purge)
|
|
self.assertTrue(task.processing_node.remove_task(task.uuid))
|
|
|
|
# 3. Restart the task
|
|
res = client.post("/api/projects/{}/tasks/{}/restart/".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
|
|
# 4. Check that the rerun_from parameter has been cleared
|
|
# but the other parameters are still set
|
|
task.refresh_from_db()
|
|
self.assertTrue(len(task.uuid) == 0)
|
|
self.assertTrue(len(list(filter(lambda d: d['name'] == 'rerun-from', task.options))) == 0)
|
|
self.assertTrue(len(list(filter(lambda d: d['name'] == 'mesh-size', task.options))) == 1)
|
|
|
|
# Test connection, timeout errors
|
|
def connTimeout(*args, **kwargs):
|
|
raise requests.exceptions.ConnectTimeout("Simulated timeout")
|
|
|
|
testWatch.intercept("nodeodm.api_client.task_output", connTimeout)
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
# Timeout errors should be handled by retrying again at a later time
|
|
# and not fail
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.last_error is None)
|
|
|
|
|
|
# Reassigning the task to another project should move its assets
|
|
self.assertTrue(os.path.exists(full_task_directory_path(task.id, project.id)))
|
|
self.assertTrue(len(task.imageupload_set.all()) == 2)
|
|
for image in task.imageupload_set.all():
|
|
self.assertTrue('project/{}/'.format(project.id) in image.image.path)
|
|
|
|
task.project = other_project
|
|
task.save()
|
|
task.refresh_from_db()
|
|
self.assertFalse(os.path.exists(full_task_directory_path(task.id, project.id)))
|
|
self.assertTrue(os.path.exists(full_task_directory_path(task.id, other_project.id)))
|
|
|
|
for image in task.imageupload_set.all():
|
|
self.assertTrue('project/{}/'.format(other_project.id) in image.image.path)
|
|
|
|
node_odm.terminate()
|
|
|
|
# Restart node-odm as to not generate orthophotos
|
|
testWatch.clear()
|
|
node_odm = start_processing_node("--test_skip_orthophotos")
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'images': [image1, image2],
|
|
'name': 'test_task_no_orthophoto',
|
|
'processing_node': pnode.id,
|
|
'auto_processing_node': 'false'
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
time.sleep(DELAY)
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
task = Task.objects.get(pk=res.data['id'])
|
|
self.assertTrue(task.status == status_codes.COMPLETED)
|
|
|
|
# Orthophoto files/directories should be missing
|
|
self.assertFalse(os.path.exists(task.assets_path("odm_orthophoto", "odm_orthophoto.tif")))
|
|
self.assertFalse(os.path.exists(task.assets_path("orthophoto_tiles")))
|
|
|
|
# orthophoto_extent should be none
|
|
self.assertTrue(task.orthophoto_extent is None)
|
|
|
|
# but other extents should be populated
|
|
self.assertTrue(task.dsm_extent is not None)
|
|
self.assertTrue(task.dtm_extent is not None)
|
|
self.assertTrue(os.path.exists(task.assets_path("dsm_tiles")))
|
|
self.assertTrue(os.path.exists(task.assets_path("dtm_tiles")))
|
|
|
|
# Can access only tiles of available assets
|
|
res = client.get("/api/projects/{}/tasks/{}/dsm/tiles.json".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
res = client.get("/api/projects/{}/tasks/{}/dtm/tiles.json".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_200_OK)
|
|
res = client.get("/api/projects/{}/tasks/{}/orthophoto/tiles.json".format(project.id, task.id))
|
|
self.assertTrue(res.status_code == status.HTTP_400_BAD_REQUEST)
|
|
|
|
# Available assets should be missing orthophoto.tif type
|
|
# but others such as textured_model.zip should be available
|
|
res = client.get("/api/projects/{}/tasks/{}/".format(project.id, task.id))
|
|
self.assertFalse('orthophoto.tif' in res.data['available_assets'])
|
|
self.assertFalse('orthophoto_tiles.zip' in res.data['available_assets'])
|
|
self.assertTrue('textured_model.zip' in res.data['available_assets'])
|
|
|
|
image1.close()
|
|
image2.close()
|
|
gcp.close()
|
|
node_odm.terminate()
|
|
|
|
def test_task_auto_processing_node(self):
|
|
project = Project.objects.get(name="User Test Project")
|
|
task = Task.objects.create(project=project, name="Test")
|
|
pnode = ProcessingNode.objects.create(hostname="invalid-host", port=11223)
|
|
another_pnode = ProcessingNode.objects.create(hostname="invalid-host-2", port=11223)
|
|
|
|
# By default
|
|
self.assertTrue(task.auto_processing_node)
|
|
self.assertTrue(task.processing_node is None)
|
|
|
|
# Simulate an error
|
|
task.last_error = "Test error"
|
|
task.save()
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
# A processing node should not have been assigned
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node is None)
|
|
|
|
# Remove error
|
|
task.last_error = None
|
|
task.save()
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
# A processing node should not have been assigned because no processing nodes are online
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node is None)
|
|
|
|
# Bring a processing node online
|
|
pnode.last_refreshed = timezone.now()
|
|
pnode.save()
|
|
self.assertTrue(pnode.is_online())
|
|
|
|
# A processing node has been assigned
|
|
worker.tasks.process_pending_tasks()
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node.id == pnode.id)
|
|
|
|
# Task should have failed (no images provided, invalid host...)
|
|
self.assertTrue(task.last_error is not None)
|
|
|
|
# Bring another processing node online, and bring the old one offline
|
|
pnode.last_refreshed = timezone.now() - timedelta(minutes=OFFLINE_MINUTES)
|
|
pnode.save()
|
|
|
|
another_pnode.last_refreshed = timezone.now()
|
|
another_pnode.save()
|
|
|
|
# Remove error, set status to queued
|
|
task.last_error = None
|
|
task.status = status_codes.QUEUED
|
|
task.save()
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
# Processing node is now cleared and a new one will be assigned on the next tick
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node is None)
|
|
self.assertTrue(task.status is None)
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node.id == another_pnode.id)
|
|
|
|
# Set task to queued, bring node offline
|
|
task.last_error = None
|
|
task.status = status_codes.RUNNING
|
|
task.save()
|
|
another_pnode.last_refreshed = timezone.now() - timedelta(minutes=OFFLINE_MINUTES)
|
|
another_pnode.save()
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
task.refresh_from_db()
|
|
|
|
# Processing node is still there, but task should have failed
|
|
self.assertTrue(task.status == status_codes.FAILED)
|
|
self.assertTrue("Processing node went offline." in task.last_error)
|
|
|
|
|
|
def test_task_manual_processing_node(self):
|
|
user = User.objects.get(username="testuser")
|
|
project = Project.objects.create(name="User Test Project", owner=user)
|
|
task = Task.objects.create(project=project, name="Test", auto_processing_node=False)
|
|
|
|
# Bring a processing node online
|
|
pnode = ProcessingNode.objects.create(hostname="invalid-host", port=11223)
|
|
pnode.last_refreshed = timezone.now()
|
|
pnode.save()
|
|
self.assertTrue(pnode.is_online())
|
|
|
|
worker.tasks.process_pending_tasks()
|
|
|
|
# A processing node should not have been assigned because we asked
|
|
# not to via auto_processing_node = false
|
|
task.refresh_from_db()
|
|
self.assertTrue(task.processing_node is None)
|
|
|
|
def test_task_chunked_uploads(self):
|
|
node_odm = start_processing_node()
|
|
client = APIClient()
|
|
|
|
user = User.objects.get(username="testuser")
|
|
self.assertFalse(user.is_superuser)
|
|
|
|
project = Project.objects.create(
|
|
owner=user,
|
|
name="test project"
|
|
)
|
|
|
|
pnode = ProcessingNode.objects.create(hostname="localhost", port=11223)
|
|
|
|
# task creation via chunked upload
|
|
image1 = open("app/fixtures/tiny_drone_image.jpg", 'rb')
|
|
image2 = open("app/fixtures/tiny_drone_image_2.jpg", 'rb')
|
|
|
|
# Cannot create partial task without credentials
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'auto_processing_node': 'true',
|
|
'partial': 'true'
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_403_FORBIDDEN);
|
|
|
|
client.login(username="testuser", password="test1234")
|
|
|
|
# Can after login
|
|
res = client.post("/api/projects/{}/tasks/".format(project.id), {
|
|
'auto_processing_node': 'true',
|
|
'partial': 'true'
|
|
}, format="multipart")
|
|
self.assertTrue(res.status_code == status.HTTP_201_CREATED)
|
|
|
|
task = Task.objects.get(pk=res.data['id'])
|
|
|
|
# It's partial
|
|
self.assertTrue(task.partial)
|
|
|
|
# It should not get processed
|
|
worker.tasks.process_pending_tasks()
|
|
time.sleep(DELAY)
|
|
self.assertEqual(task.upload_progress, 0.0)
|
|
|
|
# Upload to inexisting task lead to 404
|
|
wrong_task_id = '11111111-1111-1111-1111-111111111111'
|
|
res = client.post("/api/projects/{}/tasks/{}/upload/".format(project.id, wrong_task_id), {
|
|
'images': [image1],
|
|
}, format="multipart")
|
|
self.assertEqual(res.status_code, status.HTTP_404_NOT_FOUND)
|
|
image1.seek(0)
|
|
|
|
# Upload works with one image
|
|
res = client.post("/api/projects/{}/tasks/{}/upload/".format(project.id, task.id), {
|
|
'images': [image1],
|
|
}, format="multipart")
|
|
self.assertEqual(res.status_code, status.HTTP_200_OK)
|
|
self.assertEqual(res.data['success'], True)
|
|
image1.seek(0)
|
|
|
|
# Cannot commit with a single image
|
|
res = client.post("/api/projects/{}/tasks/{}/commit/".format(project.id, task.id))
|
|
self.assertEqual(res.status_code, status.HTTP_400_BAD_REQUEST)
|
|
|
|
# And second image
|
|
res = client.post("/api/projects/{}/tasks/{}/upload/".format(project.id, task.id), {
|
|
'images': [image2],
|
|
}, format="multipart")
|
|
self.assertEqual(res.status_code, status.HTTP_200_OK)
|
|
self.assertEqual(res.data['success'], True)
|
|
image2.seek(0)
|
|
|
|
# Task hasn't started
|
|
self.assertEqual(task.upload_progress, 0.0)
|
|
|
|
# Can commit with two images
|
|
res = client.post("/api/projects/{}/tasks/{}/commit/".format(project.id, task.id))
|
|
self.assertEqual(res.status_code, status.HTTP_200_OK)
|
|
self.assertEqual(res.data['id'], str(task.id))
|
|
|
|
task.refresh_from_db()
|
|
|
|
# No longer partial
|
|
self.assertFalse(task.partial)
|
|
|
|
# Image count has been updated
|
|
self.assertEqual(task.images_count, 2)
|
|
|
|
# Make sure processing begins
|
|
worker.tasks.process_pending_tasks()
|
|
time.sleep(DELAY)
|
|
|
|
task.refresh_from_db()
|
|
self.assertEqual(task.upload_progress, 1.0)
|
|
|
|
image1.close()
|
|
image2.close()
|
|
|
|
node_odm.terminate()
|
|
|
|
|
|
|
|
|