Backend: - add 'maxCrawlSize' to models and crawljob spec - add 'MAX_CRAWL_SIZE' to configmap - add maxCrawlSize to new crawlconfig + update APIs - operator: gracefully stop crawl if current size (from stats) exceeds maxCrawlSize - tests: add max crawl size tests Frontend: - Add Max Crawl Size text box Limits tab - Users enter max crawl size in GB, convert to bytes - Add BYTES_PER_GB as constant for converting to bytes - docs: Crawl Size Limit to user guide workflow setup section Operator Refactor: - use 'status.stopping' instead of 'crawl.stopping' to indicate crawl is being stopped, as changing later has no effect in operator - add is_crawl_stopping() to return if crawl is being stopped, based on crawl.stopping or size or time limit being reached - crawlerjob status: store byte size under 'size', human readable size under 'sizeHuman' for clarity - size stat always exists so remove unneeded conditional (defaults to 0) - store raw byte size in 'size', human readable size in 'sizeHuman' Charts: - subchart: update crawlerjob crd in btrix-crds to show status.stopping instead of spec.stopping - subchart: show 'sizeHuman' property instead of 'size' - bump subchart version to 0.1.1 --------- Co-authored-by: Ilya Kreymer <ikreymer@gmail.com> |
||
---|---|---|
.. | ||
.husky | ||
.vscode | ||
assets | ||
config | ||
scripts | ||
src | ||
tests | ||
xliff | ||
.dockerignore | ||
.editorconfig | ||
.gitignore | ||
.prettierignore | ||
00-browsertrix-nginx-init.sh | ||
Dockerfile | ||
frontend.conf.template | ||
index.d.ts | ||
lit-localize.json | ||
minio.conf | ||
package.json | ||
playwright.config.ts | ||
postcss.config.js | ||
README.md | ||
sample.env.local | ||
tailwind.config.js | ||
tsconfig.json | ||
web-test-runner.config.mjs | ||
webpack.config.js | ||
webpack.dev.js | ||
webpack.prod.js | ||
yarn.lock |