Website Scraper Command Line Automation

Automate website scrape in A1 Website Scraper. Download websites using automation, e.g. during the night.

Command Line Support in A1 Website Scraper

You can use a command line interface to automate all the major website scraper tools in the program.

This means that you can also use external applications and bat / batch / script files.

This enables you to run our website scraper software at regular intervals using e.g. Windows Task Scheduler.
  • Parameters:
    • ":%project-path%" : Where %project-path% should contain the active project file path. (Remember the colon before project path.)
    • "@override_initfromproject=c:\example\projects\initfrom.ini@" : Overwrite the initial project settings used.
    • "@override_rootpath=" : Overwrite website rootpath.
    • "-autocreate" : Automatically create project files and related if they do not exist already.
    • "-exit" : Exits when done.
    • "-hide" : Always invisible and exits when done.
    • "-preferieoverwv2" : If you prefer using Internet Explorer engine over WebView2 based on Chromium.
    • "-scan" : Runs website scanner.
    • "-stop0000" : Stops scan after a number of seconds, e.g. -stop600 stops scan after 10 minutes.
    • "-stopurls0000" : Stops scan after a number of URLs has been both found and handled, e.g. -stopurls500.
    • "-save" : Saves project.
    • "@override_exportpathdir=c:\example\exports\@" : Overwrite general directory path used for e.g. CSV export data files.
    • "-exportexternalcsv" : Exports all URLs data listed in "external" tree view into a file called "external.csv" located in project directory.
    • "-exportinternalcsv" : Exports all URLs data listed in "internal" tree view into a file called "internal.csv" located in project directory.
    • "-exportsitemapcsv" : Exports all URLs data listed in "internal" tree view into a file called "sitemap.csv" located in project directory.
    • "-scrapesinglepage" : Only scrape the single page URL defined Scraper options - useful if you only want data from a single page.

  • Examples for usage on Windows:
    • [ "c:\microsys\website\scraper.exe" -exit -scan -build -save ":c:\microsys\website\scraper\my-project.ini" ].
    • [ "Scraper.exe" -exit -scrapesinglepage ":my-project.ini" ] - Here it is assumed my-project.ini is in the same directory as the executable.
    • [ "Scraper.exe" -scan -build @override_rootpath= ]
    • [ start "" "Scraper.exe" -scan -build @override_rootpath= ] - Launches asynchronously. Do not use spaces in parameters.
      [ timeout 2 ] - Idle time to avoid problems with multiple instances launching at the exact same time.

  • Examples for usage on Mac OS:
    • [ open -n --args -scan -build @override_rootpath= ].
    • [ open -n --args -exit -scan -save -autocreate ":/users/%name%/myprojects/shopexample.ini" @override_initfromproject=/users/%name%/myprojecs/mydefaults.ini@ @override_rootpath= ].

  • Tips:
    • To prevent a parameter value that contains spaces (e.g. if you are passing a directory path) from being broken up, enclose it inside a couple of "".
    • In the above examples %name% refers to your user name in the given operating system.

Automate Website Scraper with Command Line and Batch Files

  • Create a batch file using any standard text editor:
    • Windows: batch-file.bat
    • Mac: batch-file.command
  • Example of what to write underlined:
    • Windows: [ "c:\microsys\website\scraper.exe" -exit -scan -build -save ":c:\microsys\website\scraper\my-project.ini" ].
    • Mac: [ open -n --args -scan -build @override_rootpath= ].
  • Save your batch file. You can now call it yourself or from other programs and scripts.

Schedule and automate Website Scraper with Windows Task Scheduler

schedule automate website scraper

  • Open Control Panel | Scheduled Tasks | Add Scheduled Task. Follow the guide.
  • Open the generated website scraper time scheduled item to edit details, e.g. command line parameters.
A1 Website Scraper
A1 Website Scraper | help | previous | next
Extract data from sites into CSV files. By scraping websites, you can grab data on websites and transform it into CSV files ready to be imported anywhere, e.g. SQL databases
This help page is maintained by
As one of the lead developers, his hands have touched most of the code in the software from Microsys. If you email any questions, chances are that he will be the one answering.
Share this page with friends   LinkedIn   Twitter   Facebook   Pinterest   YouTube  
 © Copyright 1997-2024 Microsys

 Usage of this website constitutes an accept of our legal, privacy policy and cookies information.