Website Scraper Command Line Automation
Automate website scrape in A1 Website Scraper. Download websites using automation, e.g. during the night.
      You can use a command line interface to automate all the major website scraper tools in the program.
      
      This means that you can also use external applications and bat / batch / script files.
      
      This enables you to run our website scraper software at regular intervals using e.g. Windows Task Scheduler.
      
      
        - Parameters:
          
            - ":%project-path%" :
            Where %project-path% should contain the active project file path. (Remember the colon before project path.)
            
- "@override_initfromproject=c:\example\projects\initfrom.ini@" :
            Overwrite the initial project settings used.
            
- "@override_rootpath=http://example.com@" :
            Overwrite website rootpath.
            
- "-autocreate" :
            Automatically create project files and related if they do not exist already.
            
- "-exit" :
            Exits when done.
            
- "-hide" :
            Always invisible and exits when done.
            
- "-preferieoverwv2" :
            If you prefer using Internet Explorer engine over WebView2 based on Chromium.
            
- "-scan" :
            Runs website scanner.
            
- "-stop0000" :
            Stops scan after a number of seconds, e.g. -stop600 stops scan after 10 minutes.
            
- "-stopurls0000" :
            Stops scan after a number of URLs has been both found and handled, e.g. -stopurls500.
            
- "-save" :
            Saves project.
            
- "@override_exportpathdir=c:\example\exports\@" :
            Overwrite general directory path used for e.g. CSV export data files.
            
- "-exportexternalcsv" :
            Exports all URLs data listed in "external" tree view into a file called "external.csv" located in project directory.
            
- "-exportinternalcsv" :
            Exports all URLs data listed in "internal" tree view into a file called "internal.csv" located in project directory.
            
- "-exportsitemapcsv" :
            Exports all URLs data listed in "internal" tree view into a file called "sitemap.csv" located in project directory.
            
- "-scrapesinglepage" :
            Only scrape the single page URL defined Scraper options - useful if you only want data from a single page.
            
 
 
- Examples for usage on Windows:
          
            - [ "c:\microsys\website\scraper.exe" -exit -scan -build -save ":c:\microsys\website\scraper\my-project.ini" ].
- [ "Scraper.exe" -exit -scrapesinglepage ":my-project.ini" ] - Here it is assumed my-project.ini is in the same directory as the executable.
- [ "Scraper.exe" -scan -build @override_rootpath=http://example.com@ ]
-               
                [ start "" "Scraper.exe" -scan -build @override_rootpath=http://example.com@ ] - Launches asynchronously. Do not use spaces in parameters.
                
 [ timeout 2 ] - Idle time to avoid problems with multiple instances launching at the exact same time.
 
 
- Examples for usage on Mac OS:
          
            - [ open -n A1WebsiteScraper.app --args -scan -build @override_rootpath=http://example.com@ ].
- [ open -n A1WebsiteScraper.app --args -exit -scan -save -autocreate ":/users/%name%/myprojects/shopexample.ini" @override_initfromproject=/users/%name%/myprojecs/mydefaults.ini@ @override_rootpath=https://shop.example.com@ ].
 
 
- Tips:
          
            - To prevent a parameter value that contains spaces (e.g. if you are passing a directory path) from being broken up, enclose it inside a couple of "".
- In the above examples %name%refers to your user name in the given operating system.