browsertrix/docs/user-guide/index.md
Henry Wilkinson ae8804d87f
Improves user documentation intro (#1376)
Closes #1369 

### Changes
- Adds improved getting started steps and intro contact information to
the User Guide homepage
- Adds a small section about the execution minutes graph for orgs with a
quota set
- Moves existing signup content to a dedicated signup page
- Changes admonitions from using em dashes to using colons.
- Em dashes are great and I love em.... But sometimes I love them a
little _too_ much and they were a bad fit here.
- Fixes user guide homepage link
- Fixes `ReplayWeb.page` and `ArchiveWeb.page` names
- Fixes broken links (would be good to have a CI system for this I
think)

---------
Co-authored-by: Emma Segal-Grossman <hi@emma.cafe>
Co-authored-by: Tessa Walsh <tessa@bitarchivist.net>
Co-authored-by: Ilya Kreymer <ikreymer@gmail.com>
2023-11-15 17:55:47 -08:00

30 lines
1.9 KiB
Markdown

# Browsertrix User Guide
Welcome to the Browsertrix User Guide. This page covers the basics of using Browsertrix, Webrecorder's high-fidelity web archiving system.
## Getting Started
To get started crawling with Browsertrix:
1. Create an account and join an Organization [as described here](signup).
2. After being redirected to the organization's [Overview page](overview), click the _Create New_ button in the top right and select _[Crawl Workflow](crawl-workflows)_ to begin configuring your first crawl!
3. For a simple crawl, choose the _Seeded Crawl_ option, and enter a page url in the _Crawl Start URL_ field. By default, the crawler will archive all pages under the starting path.
4. Next, click _Review & Save_, and ensure the _Run on Save_ option is selected. Then click _Save Workflow_.
5. Wait a moment for the crawler to start and watch as it archives the website!
---
After running your first crawl, check out the following to learn more about Browsertrix's features:
- A detailed list of [crawl workflow setup](workflow-setup) options.
- Adding [exclusions](workflow-setup/#exclusions) to limit your crawl's scope and evading crawler traps by [editing exclusion rules while crawling](crawl-workflows/#live-exclusion-editing).
- Best practices for crawling with [browser profiles](browser-profiles) to capture content only available when logged in to a website.
- Managing archived items, including [uploading previously archived content](archived-items/#uploading-web-archives).
- Organizing and combining archived items with [collections](collections) for sharing and export.
- If you're an admin: [Inviting collaborators to your org](org-settings/#members).
### Have more questions?
While our aim is to create intuitive interfaces, sometimes the complexities of web archiving require a little more explanation. If there's something that you found especially confusing or frustrating [please get in touch](mailto:docs-feedback@webrecorder.net)!