Chilkat HOME .NET Core C# Android™ AutoIt C C# C++ Chilkat2-Python CkPython Classic ASP DataFlex Delphi ActiveX Delphi DLL Go Java Lianja Mono C# Node.js Objective-C PHP ActiveX PHP Extension Perl PowerBuilder PowerShell PureBasic Ruby SQL Server Swift 2 Swift 3,4,5... Tcl Unicode C Unicode C++ VB.NET VBScript Visual Basic 6.0 Visual FoxPro Xojo Plugin
(PowerShell) Using the Disk CacheThe Chilkat Spider component has disk caching capabilities. To setup a disk cache, create a new directory anywhere on your local hard drive and set the CacheDir property to the path. For example, you might create "c:/spiderCache/". The UpdateCache property controls whether downloaded pages are saved to the cache. The FetchFromCache property controls whether the cache is first checked for pages. The LastFromCache property tells whether the last URL fetched came from cache or not.
Add-Type -Path "C:\chilkat\ChilkatDotNet47-9.5.0-x64\ChilkatDotNet47.dll" $spider = New-Object Chilkat.Spider # Set our cache directory and make sure saving-to-cache and fetching-from-cache # are both turned on: $spider.CacheDir = "c:/spiderCache/" $spider.FetchFromCache = $true $spider.UpdateCache = $true # If you run this code twice, you'll find that the 2nd run is extremely fast # because the pages will be retrieved from cache. # The spider object crawls a single web site at a time. As you'll see # in later examples, you can collect outbound links and use them to # crawl the web. For now, we'll simply spider 10 pages of chilkatsoft.com $spider.Initialize("www.chilkatsoft.com") # Add the 1st URL: $spider.AddUnspidered("http://www.chilkatsoft.com/") # Begin crawling the site by calling CrawlNext repeatedly. for ($i = 0; $i -le 9; $i++) { $success = $spider.CrawlNext() if ($success -eq $true) { # Show the URL of the page just spidered. $($spider.LastUrl) # The HTML is available in the LastHtml property } else { # Did we get an error or are there no more URLs to crawl? if ($spider.NumUnspidered -eq 0) { $("No more URLs to spider") } else { $($spider.LastErrorText) } } # Sleep 1 second before spidering the next URL. # The reason for waiting a short time before the next fetch is to prevent # undue stress on the web server. However, if the last page was retrieved # from cache, there is no need to pause. if ($spider.LastFromCache -ne $true) { $spider.SleepMs(1000) } } |
© 2000-2024 Chilkat Software, Inc. All Rights Reserved.