[Update] HTTrack Website Copier v.3.40

  • Thread starter Thread starter CoMa
  • Start date Start date
C

CoMa

HTTrack Website Copier
http://www.httrack.com/

This program is an offline browser for Windows
and Unix/Linux. It allows you to transfer one or
several websites from the Internet to your hard
disk. The robot is fully configurable, with a LOT
of parameters: possibility of update/continue,
multiple connections, filters, proxy, wizard...
Windows interface + command-line release
Freeware/GPL



HTTrack Website Copier release history:
--------------------------------------
This file lists all changes and fixes that have been made for HTTrack.

3.40
+ New: mms:// streaming capture (thanks to Nicolas Benoit!)
+ New: proxyTrack project released
+ New: new experimental parser that no longer needs link testing ('testing
link type..')
+ New: Redirect handled transparently with delayed type check and broken
links made external when the "no error page" option is enabled
+ New: improved background download to handle large sites
+ New: '--assume foo/bar.cgi=text/html' is now possible
+ New: MIME type scan rules (such as -mime:video/* +mime:video/mpeg)
+ New: size scan rules now allows to rewrite uncaught links as external
links
+ Fixed: crash fixed when ftime()/localtime()==NULL
+ Fixed: iso-9660 option now using '_' for collision character
+ Fixed: collision problems with CaSe SeNsItIvItY
+ Fixed: a href='..' fixed!
+ Fixed: redirects are now handled by the new experimental parser
+ Fixed: "./" links generated with fixed outbound links (../../)
+ Fixed: 'base href' bogus in many cases
+ Fixed: enforce security limits to avoid bandwidth abuses
+ Fixed: bogus external (swf) parser, fixed remaining .delayed files
+ New: new check-mime and save-file2 callbacks
+ New: "always delayed type check" enabled
+ Fixed: totally bogus finalizer causing compressed files not to be
uncompressed, and many files to be truncated
+ Shell: new Finnish interface added!
+ Fixed: "..html" bogus type
+ Fixed: remaining bogus .delayed entries
+ Fixed: flush before user-defined command
+ Fixed: fixed user-defined command call and background cleaner
+ Fixed: fixed 'Crash adding error, unexpected error found.. [4250]' error
+ Fixed: fixed cache absolute file reference (the reference is now
relative) preventing the cache form being moved to another place
+ Fixed: webhttrack 'Browse Sites' path bug
+ Fixed: old httrack cache format fixes (import of older versions did not
work anymore)
+ Fixed: port fixes in htsnet.h
+ Fixed: -N option with advanced extraction (bogus "not found" member)
+ Fixed: javascript: location=URL was not recognized
+ Fixed: no more character escaping when not needed (such as UTF-8 codes)
+ Fixed: possibly temporary files left on disk with bogus servers giving
compressed content on HEAD reuests
+ Fixed: URL hack caused unexpected filename collisions (index.html vs
INDEX.HTML)
+ Fixed: "do not erase already downloaded file" option now correctly works
(it leaves files linked in the mirror)
+ Fixed: UCS2 encoded pages are now converted properly into UTF-8
+ New: "near" option now also catch embedded (images, css, ..) files
+ Fixed: bogus chunked multimedia link text files (such as x-ms-asf files)



/CoMa
 
HTTrack Website Copier
http://www.httrack.com/

This program is an offline browser for Windows
and Unix/Linux. It allows you to transfer one or
several websites from the Internet to your hard
disk. The robot is fully configurable, with a LOT
of parameters: possibility of update/continue,
multiple connections, filters, proxy, wizard...
Windows interface + command-line release
Freeware/GPL

Has anyone got some user experience of this?
Any good? Bugs? etc... TIA
 
hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it on two sites. It works well, I did not experience any
bugs. In using this, it's important to be a good 'netizen' (using w/in
rights, not pounding on a target server too hard, etc) so I strongly
recommend reading the documentation and forums beforehand.

This is a good product w/decent support.

-Craig
 
Craig said:
I've used it on two sites. It works well, I did not experience any
bugs. In using this, it's important to be a good 'netizen' (using w/in
rights, not pounding on a target server too hard, etc) so I strongly
recommend reading the documentation and forums beforehand.

This is a good product w/decent support.

-Craig
Ditto.

My experience is based on about 15 sites that I store and update with
HTTrack.

The good netizenship rules are here:
http://www.httrack.com/html/abuse.html

Greetings,
Demetris
 
hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it for a number of sites and it worked well. However, recently
when I only wanted all pdf files stored on a website it didn't work. I
don't know whether this was my bad manipulation, a deliberate blocking
of downloads by the website (I don't think this can be reason, since the
site hosts academic publications to facilitate distributing of research
results) of the program.

Frank
 
hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it for a number of sites and it worked well. However, recently
when I only wanted all pdf files stored on a website it didn't work. I
don't know whether this was my bad manipulation, a deliberate blocking
of downloads by the website (I don't think this can be reason, since the
site hosts academic publications to facilitate distributing of research
results) of the program.

I must add that I use SpiderZilla, the Mozilla/Firefox extension.

Frank
 
Back
Top