Looking for Web sites harvesting tool
emiller at smtpgwy.isinet.com
emiller at smtpgwy.isinet.com
Tue Dec 9 15:26:18 EST 1997
I have a suggestion... The new version of ProCite (a bibliographic
management software package) for Windows does include a new feature
that may do what you want it to. You can automatically capture a URL
and title information from a web page and store the "reference" in a
ProCite record. You can also launch Netscape from within ProCite to
open the URL and...maybe most importantly, you can output ProCite
database records in HTML to publish bibliographies of Web pages on the
Web.
For more information, go to: http://www.risinc.com/pc/pcprod.html
Sincerely,
Elisa
Elisa Miller "Whatever you can do or
Institute for Scientific Information dream you can, begin it.
3501 Market Street Boldness has genius, power
Philadelphia, Pa 19104 and magic in it."
(215)386-0100 x 1395 Goethe
emiller at isinet.com
URL - http://www.isinet.com
______________________________ Reply Separator
_________________________________
Subject: Looking for Web sites harvesting tool
Author: Shu.Chen at rrcc.cccoes.edu at INTERNET
Date: 12/5/97 5:23 PM
I am on a mission to collect Web sites for different topics, and I am
looking for a tool that allows me to collect Web sites into my own html
files with some ease. Ideally the tool would take down the url from the
current active Web browser and gives me a window to put in clickable
text and description/abstract type of information, and also allows me to
save the collection in an html file.
I have spent two days searching for such a thing on the Web, but have
not found one yet. Does anyone know if there is such a tool out there?
Or do I have to do an awful lot of copy and paste to get the job done?
Thanks in advance for any hints and directions!
> ----------
>
More information about the Web4lib
mailing list