Closes #1369 ### Changes - Adds improved getting started steps and intro contact information to the User Guide homepage - Adds a small section about the execution minutes graph for orgs with a quota set - Moves existing signup content to a dedicated signup page - Changes admonitions from using em dashes to using colons. - Em dashes are great and I love em.... But sometimes I love them a little _too_ much and they were a bad fit here. - Fixes user guide homepage link - Fixes `ReplayWeb.page` and `ArchiveWeb.page` names - Fixes broken links (would be good to have a CI system for this I think) --------- Co-authored-by: Emma Segal-Grossman <hi@emma.cafe> Co-authored-by: Tessa Walsh <tessa@bitarchivist.net> Co-authored-by: Ilya Kreymer <ikreymer@gmail.com>
		
			
				
	
	
		
			30 lines
		
	
	
		
			1.9 KiB
		
	
	
	
		
			Markdown
		
	
	
	
	
	
			
		
		
	
	
			30 lines
		
	
	
		
			1.9 KiB
		
	
	
	
		
			Markdown
		
	
	
	
	
	
| # Browsertrix User Guide
 | |
| 
 | |
| Welcome to the Browsertrix User Guide. This page covers the basics of using Browsertrix, Webrecorder's high-fidelity web archiving system.
 | |
| 
 | |
| ## Getting Started
 | |
| 
 | |
| To get started crawling with Browsertrix:
 | |
| 
 | |
| 1. Create an account and join an Organization [as described here](signup).
 | |
| 2. After being redirected to the organization's [Overview page](overview), click the _Create New_ button in the top right and select _[Crawl Workflow](crawl-workflows)_ to begin configuring your first crawl!
 | |
| 3. For a simple crawl, choose the _Seeded Crawl_ option, and enter a page url in the _Crawl Start URL_ field. By default, the crawler will archive all pages under the starting path.
 | |
| 4. Next, click _Review & Save_, and ensure the _Run on Save_ option is selected. Then click _Save Workflow_.
 | |
| 5. Wait a moment for the crawler to start and watch as it archives the website!
 | |
| 
 | |
| ---
 | |
| 
 | |
| After running your first crawl, check out the following to learn more about Browsertrix's features:
 | |
| 
 | |
| - A detailed list of [crawl workflow setup](workflow-setup) options.
 | |
| - Adding [exclusions](workflow-setup/#exclusions) to limit your crawl's scope and evading crawler traps by [editing exclusion rules while crawling](crawl-workflows/#live-exclusion-editing).
 | |
| - Best practices for crawling with [browser profiles](browser-profiles) to capture content only available when logged in to a website.
 | |
| - Managing archived items, including [uploading previously archived content](archived-items/#uploading-web-archives).
 | |
| - Organizing and combining archived items with [collections](collections) for sharing and export.
 | |
| - If you're an admin: [Inviting collaborators to your org](org-settings/#members).
 | |
| 
 | |
| 
 | |
| ### Have more questions?
 | |
| 
 | |
| While our aim is to create intuitive interfaces, sometimes the complexities of web archiving require a little more explanation. If there's something that you found especially confusing or frustrating [please get in touch](mailto:docs-feedback@webrecorder.net)!
 |