If a website returns a non-standard Content-Encoding value (or one currently not supported by WebCopy), no attempt will be made to decompress the file and it will be downloaded as-is. Fixed a crash that occurred if a 401 challenge was received and the www-authenticate header was a bare type Speculative fix for a crash that could occur when finishing the New Project Wizard A crash would occur when editing a file that didn't have a mime type Speed and estimated downtime time calculations were incorrect and could cause a crash when downloading large files A crash which could occur when loading localised text is no longer fatal Fixed a crash that could occur when switching between empty virtual list views during a crawl and items were then subsequently added Fixed an issue where command line arguments sometimes didn't correctly process ambiguous relative arguments that could be a file name or a unqualified URI A crash could occur if a content type header was malformed and was either utf or utf. The URI transformation service incorrectly attempted to add prefixes to email addresses, this in turn caused a crash if the mailto: reference was malformed Brotli decompression could fail with streams larger than 65535 bytes One-time project validation checks were ignoring the content encoding settings of the project (which by default is Gzip and Deflate) and were requesting content with Brotli compression A blank error message was displayed for Brotli decompression errors If a URL was previously skipped but then included in future scans, the original skip reason could be retained Custom headers were not applied when posting forms Posting a form did not set an appropriate content type 401 challenges no longer display credential dialogues unless the authentication type is either Basic or Digest as no other values have been tested due to lack of resource Test URL dialogue now uses load on demand for settings pages Added cookie, cookie-jar and discard-session-cookies command line parameters (User Manual) Test URL dialogue now allows configuring cookies Added the ability to read cookies from an external file View and customize a visual diagram of your website, which can also be exported to an image. The built-in editor allows you to easily test expressions. Several configuration options make use of regular expressions. There are many settings you can make to configure how your website will be crawled, in addition to rules and forms mentioned above, you can also configure domain aliases, user agent strings, default documents and more.Īfter scanning a website, you can view lists of pages, errors, missing pages, media resources, and more. Filtering allows you to easily view the different links found. HTTP 401 challenge authentication is also supported, so if your website contains protected areas, you can either pre-define user names and passwords or be automatically prompted for credentials while scanning.Īfter you have analyzed your website, the Link Map Viewer allows you to view all the links found in your website, both internal and external. Additional options are also available such as downloading a URL to include in the copy, but not crawling it.īefore analyzing a website, you can optionally post one or more forms, for example to login to an administration area. Rules control the scan behavior, for example excluding a section of the website. While it will do its best to create an offline copy of a website, advanced data-driven websites may not work as expected once they have been copied. It does not download the raw source code of a web site, it can only download what the HTTP server returns. If a website makes heavy use of JavaScript to operate, it is unlikely It will be able to make a true copy if it is unable to discover all of the websites due to JavaScript being used to dynamically generate links. It does not include a virtual DOM or any form of JavaScript parsing. In this manner, WebcCopy can "crawl" an entire website and download everything it sees in an effort to create a reasonable facsimile of the source website. It will download all of these resources, and continue to search for more. The Web Copy Tool will examine the HTML mark-up of a website and attempt to discover all linked resources such as other pages, images, videos, file downloads - anything and everything. This software may be used free of charge, but as with all free software, there are costs involved to develop and maintain. Using its extensive configuration you can define which parts of a website will be copied and how. Links to resources such as style-sheets, images, and other pages on the website will automatically be remapped to match the local path. It will scan the specified website and download its content onto your harddisk. Cyotek WebCopy is a free tool for copying full or partial websites locally onto your harddisk for offline viewing.
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |