[Home] [By Thread] [By Date] [Recent Entries]

  • To: xml-dev@l...
  • Subject: hashing
  • From: Eric Hanson <eric@a...>
  • Date: Thu, 29 Apr 2004 19:58:17 +0000
  • User-agent: Mutt/1.2i

I have a large collection of XML documents, and want to find and
group any duplicates.  The obvious but slow way of doing this is
to just compare them all to each other.  Is there a better
approach?

Particularly, is there any APIs or standards for "hashing" a
document so that duplicates could be identified in a similar way
to what you'd do with a hash table?

Thanks,
Eric

Site Map | Privacy Policy | Terms of Use | Trademarks
Free Stylus Studio XML Training:
W3C Member