32047739fe
The crawl utility starts a depth-first traversal of the web at the specified URLs. It stores all JPEG images that match the configured constraints. Crawl is fairly fast and allows for graceful termination. After terminating crawl, it is possible to restart it at exactly the same spot where it was terminated. Crawl keeps a persistent database that allows multiple crawls without revisiting sites. The main features of crawl are: * Saves encountered images or other media types * Media selection based on regular expressions and size contraints * Resume previous crawl after graceful termination * Persistent database of visited URLs * Very small and efficient code * Asynchronous DNS lookups * Supports robots.txt
14 lines
385 B
Text
14 lines
385 B
Text
$NetBSD: patch-ab,v 1.1.1.1 2005/01/18 17:46:31 peter Exp $
|
|
|
|
--- config.h.in.orig 2003-05-18 02:54:45.000000000 +0200
|
|
+++ config.h.in 2004-06-12 00:06:58.000000000 +0200
|
|
@@ -42,6 +42,9 @@
|
|
/* Define if your system has libdb */
|
|
#undef HAVE_DB_H
|
|
|
|
+/* Define if your system has libdb */
|
|
+#undef HAVE_DB1_H
|
|
+
|
|
/* Define to 1 if you have the `dirname' function. */
|
|
#undef HAVE_DIRNAME
|
|
|