156.txt
来自「This complete matlab for neural network」· 文本 代码 · 共 35 行
TXT
35 行
发信人: fervvac (高远), 信区: DataMining
标 题: Re: How to get the website topology
发信站: 南京大学小百合站 (Sat Dec 29 14:49:39 2001), 站内信件
Yes, wget has been ported to Win32 platform. There are other similar command
line driven downloading tools out there as well. In particular, if you are
using Windows, you can try Teleport Pro and Offline Exploroer Pro. Both are
sharewares, but you can easily find cracks for them.
Using either tools, you can grab all the data (might not be all) to your local
storage devices and test your algorithms with a much faster speed. I know
telport Pro and Offline Explorer Pro can convert the links to local links, but
I don't know if wget can do the same thing or not.
As for the web site changing problem, I think there is no ideal solution out
there. Because there is no global snapshot available for a site. Just like the
fundamental issues in distributed system, no global timer, :-)
For the ynamic web site, I remember someone tried to guess the DB structure
and content from dynamically generated pages, but I am not sure of that.
【 在 iamgufeng (古风) 的大作中提到: 】
: in http://space.tin.it/computer/hherold/
: you can find sources and precompiled binaries for wget's windows version
: 【 在 helloboy (hello) 的大作中提到: 】
: : I agree with you.Though we have good algorithms,but we should not
: : write all from scratches.Some tools and products can help us facilitate
: : our research process.Such as WUM ,etc.
: : But I want to know something about "GNU wget"?
: : I use windows platform.Are their any useful web usage mining tools ?
--
※ 来源:.南京大学小百合站 bbs.nju.edu.cn.[FROM: 饮水思源BBS]
⌨️ 快捷键说明
复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?