home *** CD-ROM | disk | FTP | other *** search
- Path: sparky!uunet!iadpsa!iowegia!kjhoule
- From: kjhoule@iowegia.uucp (Kevin Houle)
- Newsgroups: comp.bbs.waffle
- Subject: Re: dupweed
- Message-ID: <1ZuPuB6w165w@iowegia.uucp>
- Date: Mon, 23 Nov 92 17:55:59 CST
- References: <10aLuB1w165w@eastwind.mcds.com>
- Distribution: world
- Organization: Iowegia Public Access Usenet/UUCP, Clive IA USA.
- Lines: 18
-
- chorn@eastwind.mcds.com (Christopher Horn) writes:
-
- > My solution is to kill the dupes, batch, and then go back and resequence.
- > Currently the dupe kill routine in my code assumes you get all your news
- > from a single upstream site. A painful restriction, but the result is some
- > VERY VERY fast code. And my directory recurse function uses very little
- > stack space, with no limit on subdirectories, etc.. It can handle 3000
- > files per directory, probably closer to 5000 if one wanted it too. I'm
- > currently trying to decide how to effectively clean dupes when getting
- > multiple feeds, as this does require every article file be opened since
- > identicle articles may have come via different paths. If anyone is
- > interested, let me know and I'll either post it or an announcement of
- > where/how to get it when I'm done.
-
- Alright! A better mouse trap :) I'm interested.
-
- --
- Kevin Houle kjhoule@iowegia.uucp
-