Closes #1369 ### Changes - Adds improved getting started steps and intro contact information to the User Guide homepage - Adds a small section about the execution minutes graph for orgs with a quota set - Moves existing signup content to a dedicated signup page - Changes admonitions from using em dashes to using colons. - Em dashes are great and I love em.... But sometimes I love them a little _too_ much and they were a bad fit here. - Fixes user guide homepage link - Fixes `ReplayWeb.page` and `ArchiveWeb.page` names - Fixes broken links (would be good to have a CI system for this I think) --------- Co-authored-by: Emma Segal-Grossman <hi@emma.cafe> Co-authored-by: Tessa Walsh <tessa@bitarchivist.net> Co-authored-by: Ilya Kreymer <ikreymer@gmail.com>
30 lines
1.9 KiB
Markdown
30 lines
1.9 KiB
Markdown
# Browsertrix User Guide
|
|
|
|
Welcome to the Browsertrix User Guide. This page covers the basics of using Browsertrix, Webrecorder's high-fidelity web archiving system.
|
|
|
|
## Getting Started
|
|
|
|
To get started crawling with Browsertrix:
|
|
|
|
1. Create an account and join an Organization [as described here](signup).
|
|
2. After being redirected to the organization's [Overview page](overview), click the _Create New_ button in the top right and select _[Crawl Workflow](crawl-workflows)_ to begin configuring your first crawl!
|
|
3. For a simple crawl, choose the _Seeded Crawl_ option, and enter a page url in the _Crawl Start URL_ field. By default, the crawler will archive all pages under the starting path.
|
|
4. Next, click _Review & Save_, and ensure the _Run on Save_ option is selected. Then click _Save Workflow_.
|
|
5. Wait a moment for the crawler to start and watch as it archives the website!
|
|
|
|
---
|
|
|
|
After running your first crawl, check out the following to learn more about Browsertrix's features:
|
|
|
|
- A detailed list of [crawl workflow setup](workflow-setup) options.
|
|
- Adding [exclusions](workflow-setup/#exclusions) to limit your crawl's scope and evading crawler traps by [editing exclusion rules while crawling](crawl-workflows/#live-exclusion-editing).
|
|
- Best practices for crawling with [browser profiles](browser-profiles) to capture content only available when logged in to a website.
|
|
- Managing archived items, including [uploading previously archived content](archived-items/#uploading-web-archives).
|
|
- Organizing and combining archived items with [collections](collections) for sharing and export.
|
|
- If you're an admin: [Inviting collaborators to your org](org-settings/#members).
|
|
|
|
|
|
### Have more questions?
|
|
|
|
While our aim is to create intuitive interfaces, sometimes the complexities of web archiving require a little more explanation. If there's something that you found especially confusing or frustrating [please get in touch](mailto:docs-feedback@webrecorder.net)!
|