[Update] HTTrack Website Copier v.3.40

C

CoMa

HTTrack Website Copier
http://www.httrack.com/

This program is an offline browser for Windows
and Unix/Linux. It allows you to transfer one or
several websites from the Internet to your hard
disk. The robot is fully configurable, with a LOT
of parameters: possibility of update/continue,
multiple connections, filters, proxy, wizard...
Windows interface + command-line release
Freeware/GPL



HTTrack Website Copier release history:
--------------------------------------
This file lists all changes and fixes that have been made for HTTrack.

3.40
+ New: mms:// streaming capture (thanks to Nicolas Benoit!)
+ New: proxyTrack project released
+ New: new experimental parser that no longer needs link testing ('testing
link type..')
+ New: Redirect handled transparently with delayed type check and broken
links made external when the "no error page" option is enabled
+ New: improved background download to handle large sites
+ New: '--assume foo/bar.cgi=text/html' is now possible
+ New: MIME type scan rules (such as -mime:video/* +mime:video/mpeg)
+ New: size scan rules now allows to rewrite uncaught links as external
links
+ Fixed: crash fixed when ftime()/localtime()==NULL
+ Fixed: iso-9660 option now using '_' for collision character
+ Fixed: collision problems with CaSe SeNsItIvItY
+ Fixed: a href='..' fixed!
+ Fixed: redirects are now handled by the new experimental parser
+ Fixed: "./" links generated with fixed outbound links (../../)
+ Fixed: 'base href' bogus in many cases
+ Fixed: enforce security limits to avoid bandwidth abuses
+ Fixed: bogus external (swf) parser, fixed remaining .delayed files
+ New: new check-mime and save-file2 callbacks
+ New: "always delayed type check" enabled
+ Fixed: totally bogus finalizer causing compressed files not to be
uncompressed, and many files to be truncated
+ Shell: new Finnish interface added!
+ Fixed: "..html" bogus type
+ Fixed: remaining bogus .delayed entries
+ Fixed: flush before user-defined command
+ Fixed: fixed user-defined command call and background cleaner
+ Fixed: fixed 'Crash adding error, unexpected error found.. [4250]' error
+ Fixed: fixed cache absolute file reference (the reference is now
relative) preventing the cache form being moved to another place
+ Fixed: webhttrack 'Browse Sites' path bug
+ Fixed: old httrack cache format fixes (import of older versions did not
work anymore)
+ Fixed: port fixes in htsnet.h
+ Fixed: -N option with advanced extraction (bogus "not found" member)
+ Fixed: javascript: location=URL was not recognized
+ Fixed: no more character escaping when not needed (such as UTF-8 codes)
+ Fixed: possibly temporary files left on disk with bogus servers giving
compressed content on HEAD reuests
+ Fixed: URL hack caused unexpected filename collisions (index.html vs
INDEX.HTML)
+ Fixed: "do not erase already downloaded file" option now correctly works
(it leaves files linked in the mirror)
+ Fixed: UCS2 encoded pages are now converted properly into UTF-8
+ New: "near" option now also catch embedded (images, css, ..) files
+ Fixed: bogus chunked multimedia link text files (such as x-ms-asf files)



/CoMa
 
H

hummingbird

HTTrack Website Copier
http://www.httrack.com/

This program is an offline browser for Windows
and Unix/Linux. It allows you to transfer one or
several websites from the Internet to your hard
disk. The robot is fully configurable, with a LOT
of parameters: possibility of update/continue,
multiple connections, filters, proxy, wizard...
Windows interface + command-line release
Freeware/GPL

Has anyone got some user experience of this?
Any good? Bugs? etc... TIA
 
C

Craig

hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it on two sites. It works well, I did not experience any
bugs. In using this, it's important to be a good 'netizen' (using w/in
rights, not pounding on a target server too hard, etc) so I strongly
recommend reading the documentation and forums beforehand.

This is a good product w/decent support.

-Craig
 
D

Demetris

Craig said:
I've used it on two sites. It works well, I did not experience any
bugs. In using this, it's important to be a good 'netizen' (using w/in
rights, not pounding on a target server too hard, etc) so I strongly
recommend reading the documentation and forums beforehand.

This is a good product w/decent support.

-Craig
Ditto.

My experience is based on about 15 sites that I store and update with
HTTrack.

The good netizenship rules are here:
http://www.httrack.com/html/abuse.html

Greetings,
Demetris
 
F

FTR

hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it for a number of sites and it worked well. However, recently
when I only wanted all pdf files stored on a website it didn't work. I
don't know whether this was my bad manipulation, a deliberate blocking
of downloads by the website (I don't think this can be reason, since the
site hosts academic publications to facilitate distributing of research
results) of the program.

Frank
 
F

FTR

hummingbird said:
Has anyone got some user experience of this?
Any good? Bugs? etc... TIA

I've used it for a number of sites and it worked well. However, recently
when I only wanted all pdf files stored on a website it didn't work. I
don't know whether this was my bad manipulation, a deliberate blocking
of downloads by the website (I don't think this can be reason, since the
site hosts academic publications to facilitate distributing of research
results) of the program.

I must add that I use SpiderZilla, the Mozilla/Firefox extension.

Frank
 

Ask a Question

Want to reply to this thread or ask your own question?

You'll need to choose a username for the site, which only take a couple of moments. After that, you can post your question and our members will help you out.

Ask a Question

Top