I'm developing a well formatted search engine type of site which will have over 10 million pages. I'm optimizing it to decrease rate of duplicated content within the site, but I need actual data to proceed.
I'm looking for a strategy/tool to detect duplicate content rate within the site. Any suggestions?
Note that only content is an issue, not meta tags, URLs or cannocials.
Thanks in advance
I'm looking for a strategy/tool to detect duplicate content rate within the site. Any suggestions?
Note that only content is an issue, not meta tags, URLs or cannocials.
Thanks in advance