static struct scanqueue *scanqueue = NULL;
static struct sharepoint *shares = NULL;
static struct hashcache *hashcache = NULL;
-static pid_t hashjob = 0;
+/* Set initially to -1, but changed to 0 the first time run() is
+ * called. This is to avoid forking a hash job before daemonizing,
+ * since that would make the daemon unable to wait() for the hash
+ * job. */
+static pid_t hashjob = -1;
struct sharecache *shareroot = NULL;
unsigned long long sharesize = 0;
GCBCHAIN(sharechangecb, unsigned long long);
addtobuf(buf, ':');
bufcat(buf, whbuf, wcslen(whbuf));
free(whbuf);
+ addtobuf(buf, 0);
return(buf);
}
+int hashcmp(struct hash *h1, struct hash *h2)
+{
+ if(wcscmp(h1->algo, h2->algo))
+ return(0);
+ if(h1->len != h2->len)
+ return(0);
+ if(memcmp(h1->buf, h2->buf, h1->len))
+ return(0);
+ return(1);
+}
+
static struct hashcache *newhashcache(void)
{
struct hashcache *new;
static int run(void)
{
+ if(hashjob == -1)
+ {
+ hashjob = 0;
+ checkhashes();
+ }
return(doscan(10));
}