Website Scraper Command Line Automation
Automate website scrape in A1 Website Scraper. Download websites using automation, e.g. during the night.
You can use a command line interface to automate all the major website scraper tools in the program.
This means that you can also use external applications and bat / batch / script files.
This enables you to run our website scraper software at regular intervals using e.g. Windows Task Scheduler.
- ":%project-path%" :
Where %project-path% should contain the active project file path. (Remember the colon before project path.)
- "@override_initfromproject=c:\example\projects\initfrom.ini@" :
Overwrite the initial project settings used.
- "@override_rootpath=http://example.com@" :
Overwrite website rootpath.
- "-autocreate" :
Automatically create project files and related if they do not exist already.
- "-exit" :
Exits when done.
- "-hide" :
Always invisible and exits when done.
- "-preferieoverwv2" :
If you prefer using Internet Explorer engine over WebView2 based on Chromium.
- "-scan" :
Runs website scanner.
- "-stop0000" :
Stops scan after a number of seconds, e.g. -stop600 stops scan after 10 minutes.
- "-stopurls0000" :
Stops scan after a number of URLs has been both found and handled, e.g. -stopurls500.
- "-save" :
- "@override_exportpathdir=c:\example\exports\@" :
Overwrite general directory path used for e.g. CSV export data files.
- "-exportexternalcsv" :
Exports all URLs data listed in "external" tree view into a file called "external.csv" located in project directory.
- "-exportinternalcsv" :
Exports all URLs data listed in "internal" tree view into a file called "internal.csv" located in project directory.
- "-exportsitemapcsv" :
Exports all URLs data listed in "internal" tree view into a file called "sitemap.csv" located in project directory.
- "-scrapesinglepage" :
Only scrape the single page URL defined Scraper options - useful if you only want data from a single page.
- Examples for usage on Windows:
- [ "c:\microsys\website\scraper.exe" -exit -scan -build -save ":c:\microsys\website\scraper\my-project.ini" ].
- [ "Scraper.exe" -exit -scrapesinglepage ":my-project.ini" ] - Here it is assumed my-project.ini is in the same directory as the executable.
- [ "Scraper.exe" -scan -build @override_rootpath=http://example.com@ ]
[ start "" "Scraper.exe" -scan -build @override_rootpath=http://example.com@ ] - Launches asynchronously. Do not use spaces in parameters.
[ timeout 2 ] - Idle time to avoid problems with multiple instances launching at the exact same time.
- Examples for usage on Mac OS:
- [ open -n A1WebsiteScraper.app --args -scan -build @override_rootpath=http://example.com@ ].
- [ open -n A1WebsiteScraper.app --args -exit -scan -save -autocreate ":/users/%name%/myprojects/shopexample.ini" @override_initfromproject=/users/%name%/myprojecs/mydefaults.ini@ @override_rootpath=https://shop.example.com@ ].
- To prevent a parameter value that contains spaces (e.g. if you are passing a directory path) from being broken up, enclose it inside a couple of "".
- In the above examples
%name% refers to your user name in the given operating system.