Merge pull request #526 from webrecorder/config-copy-updates
This commit is contained in:
commit
3c199419a2
@ -442,7 +442,7 @@ export class CrawlConfigEditor extends LiteElement {
|
|||||||
|
|
||||||
render() {
|
render() {
|
||||||
const tabLabels: Record<StepName, string> = {
|
const tabLabels: Record<StepName, string> = {
|
||||||
crawlSetup: msg("Crawl Setup"),
|
crawlSetup: msg("Crawl Scope"),
|
||||||
browserSettings: msg("Browser Settings"),
|
browserSettings: msg("Browser Settings"),
|
||||||
crawlScheduling: msg("Crawl Scheduling"),
|
crawlScheduling: msg("Crawl Scheduling"),
|
||||||
crawlInformation: msg("Crawl Information"),
|
crawlInformation: msg("Crawl Information"),
|
||||||
@ -944,7 +944,7 @@ https://example.com/path`}
|
|||||||
${this.renderFormCol(html`
|
${this.renderFormCol(html`
|
||||||
<sl-select
|
<sl-select
|
||||||
name="scopeType"
|
name="scopeType"
|
||||||
label=${msg("Crawl Scope")}
|
label=${msg("Start URL Scope")}
|
||||||
defaultValue=${this.formState.scopeType}
|
defaultValue=${this.formState.scopeType}
|
||||||
value=${this.formState.scopeType}
|
value=${this.formState.scopeType}
|
||||||
@sl-select=${(e: Event) =>
|
@sl-select=${(e: Event) =>
|
||||||
@ -987,7 +987,7 @@ https://example.net`}
|
|||||||
></sl-textarea>
|
></sl-textarea>
|
||||||
`)}
|
`)}
|
||||||
${this.renderHelpTextCol(
|
${this.renderHelpTextCol(
|
||||||
html`Crawl pages outside of Crawl Scope that begin with these URLs.`
|
html`If the crawler finds pages outside of the Start URL Scope they will only be saved if they begin with URLs listed here.`
|
||||||
)}
|
)}
|
||||||
${this.renderFormCol(html`
|
${this.renderFormCol(html`
|
||||||
<sl-checkbox
|
<sl-checkbox
|
||||||
@ -1096,8 +1096,8 @@ https://example.net`}
|
|||||||
</sl-radio-group>
|
</sl-radio-group>
|
||||||
`)}
|
`)}
|
||||||
${this.renderHelpTextCol(
|
${this.renderHelpTextCol(
|
||||||
html`Increasing parallel crawler instances will speed up crawls, but
|
html`Increasing parallel crawler instances can speed up crawls, but
|
||||||
take up more system resources.`
|
may increase the chances of getting rate limited.`
|
||||||
)}
|
)}
|
||||||
`;
|
`;
|
||||||
}
|
}
|
||||||
|
|||||||
@ -148,7 +148,7 @@ export class CrawlTemplatesNew extends LiteElement {
|
|||||||
<div class="text-lg font-medium my-3">${msg("URL List")}</div>
|
<div class="text-lg font-medium my-3">${msg("URL List")}</div>
|
||||||
<p class="text-sm text-neutral-500">
|
<p class="text-sm text-neutral-500">
|
||||||
${msg(
|
${msg(
|
||||||
"The crawler visits every URL you tell it to and optionally every URL linked on those pages."
|
"The crawler visits every URL specified in a list, and optionally every URL linked on those pages."
|
||||||
)}
|
)}
|
||||||
</p>
|
</p>
|
||||||
</figcaption>
|
</figcaption>
|
||||||
@ -169,7 +169,7 @@ export class CrawlTemplatesNew extends LiteElement {
|
|||||||
<div class="text-lg font-medium my-3">${msg("Seeded Crawl")}</div>
|
<div class="text-lg font-medium my-3">${msg("Seeded Crawl")}</div>
|
||||||
<p class="text-sm text-neutral-500">
|
<p class="text-sm text-neutral-500">
|
||||||
${msg(
|
${msg(
|
||||||
"The crawler automatically finds new pages and archives them."
|
"The crawler automatically discovers and archives pages starting from a single seed URL."
|
||||||
)}
|
)}
|
||||||
</p>
|
</p>
|
||||||
</figcaption>
|
</figcaption>
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user