@@ -134,64 +134,86 @@ firecrawl https://example.com --exclude-tags nav,aside,.ad
134134
135135---
136136
137- ### ` crawl ` - Crawl an entire website
138137
139- Crawl multiple pages from a website.
138+ ### ` search ` - Search the web
139+
140+ Search the web and optionally scrape content from search results.
140141
141142``` bash
142- # Start a crawl (returns job ID)
143- firecrawl crawl https://example.com
143+ # Basic search
144+ firecrawl search " firecrawl web scraping "
144145
145- # Wait for crawl to complete
146- firecrawl crawl https://example.com --wait
146+ # Limit results
147+ firecrawl search " AI news " --limit 10
147148
148- # With progress indicator
149- firecrawl crawl https://example.com --wait --progress
149+ # Search news sources
150+ firecrawl search " tech startups " --sources news
150151
151- # Check crawl status
152- firecrawl crawl < job-id >
152+ # Search images
153+ firecrawl search " landscape photography " --sources images
153154
154- # Limit pages
155- firecrawl crawl https://example.com --limit 100 --max-depth 3
155+ # Multiple sources
156+ firecrawl search " machine learning" --sources web,news,images
157+
158+ # Filter by category (GitHub, research papers, PDFs)
159+ firecrawl search " web scraping python" --categories github
160+ firecrawl search " transformer architecture" --categories research
161+ firecrawl search " machine learning" --categories github,research
162+
163+ # Time-based search
164+ firecrawl search " AI announcements" --tbs qdr:d # Past day
165+ firecrawl search " tech news" --tbs qdr:w # Past week
166+
167+ # Location-based search
168+ firecrawl search " restaurants" --location " San Francisco,California,United States"
169+ firecrawl search " local news" --country DE
170+
171+ # Search and scrape results
172+ firecrawl search " firecrawl tutorials" --scrape
173+ firecrawl search " API documentation" --scrape --scrape-formats markdown,links
174+
175+ # Output as pretty JSON
176+ firecrawl search " web scraping"
156177```
157178
158- #### Crawl Options
179+ #### Search Options
159180
160- | Option | Description |
161- | --------------------------- | ---------------------------------------- |
162- | ` --wait ` | Wait for crawl to complete |
163- | ` --progress ` | Show progress while waiting |
164- | ` --limit <n> ` | Maximum pages to crawl |
165- | ` --max-depth <n> ` | Maximum crawl depth |
166- | ` --include-paths <paths> ` | Only crawl matching paths |
167- | ` --exclude-paths <paths> ` | Skip matching paths |
168- | ` --sitemap <mode> ` | ` include ` , ` skip ` , or ` only ` |
169- | ` --allow-subdomains ` | Include subdomains |
170- | ` --allow-external-links ` | Follow external links |
171- | ` --crawl-entire-domain ` | Crawl entire domain |
172- | ` --ignore-query-parameters ` | Treat URLs with different params as same |
173- | ` --delay <ms> ` | Delay between requests |
174- | ` --max-concurrency <n> ` | Max concurrent requests |
175- | ` --timeout <seconds> ` | Timeout when waiting |
176- | ` --poll-interval <seconds> ` | Status check interval |
181+ | Option | Description |
182+ | ---------------------------- | ------------------------------------------------------------------------------------------- |
183+ | ` --limit <n> ` | Maximum results (default: 5, max: 100) |
184+ | ` --sources <sources> ` | Comma-separated: ` web ` , ` images ` , ` news ` (default: web) |
185+ | ` --categories <categories> ` | Comma-separated: ` github ` , ` research ` , ` pdf ` |
186+ | ` --tbs <value> ` | Time filter: ` qdr:h ` (hour), ` qdr:d ` (day), ` qdr:w ` (week), ` qdr:m ` (month), ` qdr:y ` (year) |
187+ | ` --location <location> ` | Geo-targeting (e.g., "Germany", "San Francisco,California,United States") |
188+ | ` --country <code> ` | ISO country code (default: US) |
189+ | ` --timeout <ms> ` | Timeout in milliseconds (default: 60000) |
190+ | ` --ignore-invalid-urls ` | Exclude URLs invalid for other Firecrawl endpoints |
191+ | ` --scrape ` | Enable scraping of search results |
192+ | ` --scrape-formats <formats> ` | Scrape formats when ` --scrape ` enabled (default: markdown) |
193+ | ` --only-main-content ` | Include only main content when scraping (default: true) |
194+ | ` -o, --output <path> ` | Save to file |
195+ | ` --json ` | Output as compact JSON (use ` -p ` for pretty JSON) |
177196
178197#### Examples
179198
180199``` bash
181- # Crawl blog section only
182- firecrawl crawl https://example.com --include-paths /blog,/posts
200+ # Research a topic with recent results
201+ firecrawl search " React Server Components " --tbs qdr:m --limit 10
183202
184- # Exclude admin pages
185- firecrawl crawl https://example.com --exclude-paths /admin,/login
203+ # Find GitHub repositories
204+ firecrawl search " web scraping library " --categories github --limit 20
186205
187- # Crawl with rate limiting
188- firecrawl crawl https://example.com --delay 1000 --max-concurrency 2
206+ # Search and get full content
207+ firecrawl search " firecrawl documentation " --scrape --scrape-formats markdown -p -o results.json
189208
190- # Deep crawl with high limit
191- firecrawl crawl https://example.com --limit 1000 --max-depth 10 --wait --progress
209+ # Find research papers
210+ firecrawl search " large language models " --categories research -p
192211
193- # Save results
194- firecrawl crawl https://example.com --wait -o crawl-results.json --pretty
212+ # Search with location targeting
213+ firecrawl search " best coffee shops" --location " Berlin,Germany" --country DE
214+
215+ # Get news from the past week
216+ firecrawl search " AI startups funding" --sources news --tbs qdr:w --limit 15
195217```
196218
197219---
@@ -245,89 +267,69 @@ firecrawl map https://example.com --include-subdomains --limit 1000
245267
246268---
247269
248- ### ` search ` - Search the web
270+ ### ` crawl ` - Crawl an entire website
249271
250- Search the web and optionally scrape content from search results .
272+ Crawl multiple pages from a website .
251273
252274``` bash
253- # Basic search
254- firecrawl search " firecrawl web scraping"
255-
256- # Limit results
257- firecrawl search " AI news" --limit 10
258-
259- # Search news sources
260- firecrawl search " tech startups" --sources news
261-
262- # Search images
263- firecrawl search " landscape photography" --sources images
264-
265- # Multiple sources
266- firecrawl search " machine learning" --sources web,news,images
267-
268- # Filter by category (GitHub, research papers, PDFs)
269- firecrawl search " web scraping python" --categories github
270- firecrawl search " transformer architecture" --categories research
271- firecrawl search " machine learning" --categories github,research
275+ # Start a crawl (returns job ID)
276+ firecrawl crawl https://example.com
272277
273- # Time-based search
274- firecrawl search " AI announcements" --tbs qdr:d # Past day
275- firecrawl search " tech news" --tbs qdr:w # Past week
278+ # Wait for crawl to complete
279+ firecrawl crawl https://example.com --wait
276280
277- # Location-based search
278- firecrawl search " restaurants" --location " San Francisco,California,United States"
279- firecrawl search " local news" --country DE
281+ # With progress indicator
282+ firecrawl crawl https://example.com --wait --progress
280283
281- # Search and scrape results
282- firecrawl search " firecrawl tutorials" --scrape
283- firecrawl search " API documentation" --scrape --scrape-formats markdown,links
284+ # Check crawl status
285+ firecrawl crawl < job-id>
284286
285- # Output as pretty JSON
286- firecrawl search " web scraping "
287+ # Limit pages
288+ firecrawl crawl https://example.com --limit 100 --max-depth 3
287289```
288290
289- #### Search Options
291+ #### Crawl Options
290292
291- | Option | Description |
292- | ---------------------------- | ------------------------------------------------------------------------------------------- |
293- | ` --limit <n> ` | Maximum results (default: 5, max: 100) |
294- | ` --sources <sources> ` | Comma-separated: ` web ` , ` images ` , ` news ` (default: web) |
295- | ` --categories <categories> ` | Comma-separated: ` github ` , ` research ` , ` pdf ` |
296- | ` --tbs <value> ` | Time filter: ` qdr:h ` (hour), ` qdr:d ` (day), ` qdr:w ` (week), ` qdr:m ` (month), ` qdr:y ` (year) |
297- | ` --location <location> ` | Geo-targeting (e.g., "Germany", "San Francisco,California,United States") |
298- | ` --country <code> ` | ISO country code (default: US) |
299- | ` --timeout <ms> ` | Timeout in milliseconds (default: 60000) |
300- | ` --ignore-invalid-urls ` | Exclude URLs invalid for other Firecrawl endpoints |
301- | ` --scrape ` | Enable scraping of search results |
302- | ` --scrape-formats <formats> ` | Scrape formats when ` --scrape ` enabled (default: markdown) |
303- | ` --only-main-content ` | Include only main content when scraping (default: true) |
304- | ` -o, --output <path> ` | Save to file |
305- | ` --json ` | Output as compact JSON (use ` -p ` for pretty JSON) |
293+ | Option | Description |
294+ | --------------------------- | ---------------------------------------- |
295+ | ` --wait ` | Wait for crawl to complete |
296+ | ` --progress ` | Show progress while waiting |
297+ | ` --limit <n> ` | Maximum pages to crawl |
298+ | ` --max-depth <n> ` | Maximum crawl depth |
299+ | ` --include-paths <paths> ` | Only crawl matching paths |
300+ | ` --exclude-paths <paths> ` | Skip matching paths |
301+ | ` --sitemap <mode> ` | ` include ` , ` skip ` , or ` only ` |
302+ | ` --allow-subdomains ` | Include subdomains |
303+ | ` --allow-external-links ` | Follow external links |
304+ | ` --crawl-entire-domain ` | Crawl entire domain |
305+ | ` --ignore-query-parameters ` | Treat URLs with different params as same |
306+ | ` --delay <ms> ` | Delay between requests |
307+ | ` --max-concurrency <n> ` | Max concurrent requests |
308+ | ` --timeout <seconds> ` | Timeout when waiting |
309+ | ` --poll-interval <seconds> ` | Status check interval |
306310
307311#### Examples
308312
309313``` bash
310- # Research a topic with recent results
311- firecrawl search " React Server Components" --tbs qdr:m --limit 10
312-
313- # Find GitHub repositories
314- firecrawl search " web scraping library" --categories github --limit 20
314+ # Crawl blog section only
315+ firecrawl crawl https://example.com --include-paths /blog,/posts
315316
316- # Search and get full content
317- firecrawl search " firecrawl documentation " --scrape --scrape-formats markdown -p -o results.json
317+ # Exclude admin pages
318+ firecrawl crawl https://example.com --exclude-paths /admin,/login
318319
319- # Find research papers
320- firecrawl search " large language models " --categories research -p
320+ # Crawl with rate limiting
321+ firecrawl crawl https://example.com --delay 1000 --max-concurrency 2
321322
322- # Search with location targeting
323- firecrawl search " best coffee shops " --location " Berlin,Germany " --country DE
323+ # Deep crawl with high limit
324+ firecrawl crawl https://example.com --limit 1000 --max-depth 10 --wait --progress
324325
325- # Get news from the past week
326- firecrawl search " AI startups funding " --sources news --tbs qdr:w --limit 15
326+ # Save results
327+ firecrawl crawl https://example.com --wait -o crawl-results.json --pretty
327328```
328329
329330---
330331
332+
331333### ` credit-usage ` - Check your credits
332334
333335``` bash
0 commit comments