- Adds two new crawl finished state, stopped_by_user and stopped_quota_reached - Tracking other possible 'stop reasons' in operator, though not making them distinct states for now. - Updated frontend with 'Stopped by User' and 'Stopped: Time Quota Reached', shown with same icon as current partial_complete - Added migration of partial_complete to either stopped_by_user or complete (no historical quota data available) - Addresses edge case in scaling: if crawl never scaled (no redis entry, no pod), automatically scale down - Edge case in status: if crawl is somehow 'canceled' but not deleted, immediately delete crawl object and begin finalizing. --------- Co-authored-by: Tessa Walsh <tessa@bitarchivist.net>
		
			
				
	
	
		
			179 lines
		
	
	
		
			4.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			179 lines
		
	
	
		
			4.7 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
| import requests
 | |
| import time
 | |
| import os
 | |
| import pytest
 | |
| 
 | |
| from .conftest import API_PREFIX
 | |
| 
 | |
| crawl_id = None
 | |
| 
 | |
| 
 | |
| def get_crawl(org_id, auth_headers, crawl_id):
 | |
|     r = requests.get(
 | |
|         f"{API_PREFIX}/orgs/{org_id}/crawls/{crawl_id}/replay.json",
 | |
|         headers=auth_headers,
 | |
|     )
 | |
|     assert r.status_code == 200
 | |
|     return r.json()
 | |
| 
 | |
| 
 | |
| def test_start_crawl_to_cancel(
 | |
|     default_org_id, crawler_config_id_only, crawler_auth_headers
 | |
| ):
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}/run",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     assert r.status_code == 200
 | |
|     data = r.json()
 | |
| 
 | |
|     assert data.get("started")
 | |
| 
 | |
|     global crawl_id
 | |
|     crawl_id = data["started"]
 | |
| 
 | |
| 
 | |
| def test_cancel_crawl(default_org_id, crawler_auth_headers):
 | |
|     data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
|     while data["state"] == "starting":
 | |
|         time.sleep(5)
 | |
|         data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
| 
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawls/{crawl_id}/cancel",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     data = r.json()
 | |
|     assert data["success"] == True
 | |
| 
 | |
|     data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
| 
 | |
|     while data["state"] in (
 | |
|         "starting",
 | |
|         "running",
 | |
|         "waiting_capacity",
 | |
|         "generate-wacz",
 | |
|         "uploading-wacz",
 | |
|         "pending-wait",
 | |
|     ):
 | |
|         time.sleep(5)
 | |
|         data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
| 
 | |
|     assert data["state"] == "canceled"
 | |
|     assert data["stopping"] == False
 | |
| 
 | |
|     assert len(data["resources"]) == 0
 | |
| 
 | |
| 
 | |
| def test_start_crawl_and_stop_immediately(
 | |
|     default_org_id, crawler_config_id_only, crawler_auth_headers
 | |
| ):
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}/run",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     assert r.status_code == 200
 | |
|     data = r.json()
 | |
| 
 | |
|     crawl_id = data["started"]
 | |
| 
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawls/{crawl_id}/stop",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     data = r.json()
 | |
|     assert data["success"] == True
 | |
| 
 | |
|     # test crawl
 | |
|     data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
|     assert data["stopping"] == True
 | |
| 
 | |
|     # test workflow
 | |
|     r = requests.get(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     assert r.json()["lastCrawlStopping"] == True
 | |
| 
 | |
|     while data["state"] in (
 | |
|         "starting",
 | |
|         "running",
 | |
|         "waiting_capacity",
 | |
|         "generate-wacz",
 | |
|         "uploading-wacz",
 | |
|         "pending-wait",
 | |
|     ):
 | |
|         time.sleep(5)
 | |
|         data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
| 
 | |
|     assert data["state"] in ("canceled", "stopped_by_user")
 | |
|     assert data["stopping"] == True
 | |
| 
 | |
| 
 | |
| def test_start_crawl_to_stop_partial(
 | |
|     default_org_id, crawler_config_id_only, crawler_auth_headers
 | |
| ):
 | |
|     while True:
 | |
|         time.sleep(2)
 | |
|         r = requests.get(
 | |
|             f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}",
 | |
|             headers=crawler_auth_headers,
 | |
|         )
 | |
|         if r.json().get("isCrawlRunning") is False:
 | |
|             break
 | |
| 
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}/run",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     assert r.status_code == 200
 | |
|     data = r.json()
 | |
| 
 | |
|     assert data.get("started")
 | |
| 
 | |
|     global crawl_id
 | |
|     crawl_id = data["started"]
 | |
| 
 | |
| 
 | |
| def test_stop_crawl_partial(
 | |
|     default_org_id, crawler_config_id_only, crawler_auth_headers
 | |
| ):
 | |
|     data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
|     done = False
 | |
|     while not done:
 | |
|         time.sleep(2)
 | |
|         data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
|         done = data.get("stats") and data.get("stats").get("done") > 0
 | |
| 
 | |
|     r = requests.post(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawls/{crawl_id}/stop",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     data = r.json()
 | |
|     assert data["success"] == True
 | |
| 
 | |
|     # test crawl
 | |
|     data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
|     assert data["stopping"] == True
 | |
| 
 | |
|     # test workflow
 | |
|     r = requests.get(
 | |
|         f"{API_PREFIX}/orgs/{default_org_id}/crawlconfigs/{crawler_config_id_only}",
 | |
|         headers=crawler_auth_headers,
 | |
|     )
 | |
|     assert r.json()["lastCrawlStopping"] == True
 | |
| 
 | |
|     while data["state"] in (
 | |
|         "running",
 | |
|         "generate-wacz",
 | |
|         "uploading-wacz",
 | |
|         "pending-wait",
 | |
|     ):
 | |
|         time.sleep(5)
 | |
|         data = get_crawl(default_org_id, crawler_auth_headers, crawl_id)
 | |
| 
 | |
|     assert data["state"] == "stopped_by_user"
 | |
|     assert data["stopping"] == True
 | |
| 
 | |
|     assert len(data["resources"]) == 1
 |