Duplicate content is when you have an exact copy of a site. Page for page, file name for file name, image for image, code for code. And exact replica of a page or site. This was put in place for dynamic page spawners, duplicate websites, and doorway pages that were designed by blackhat seo?s and spammers that are trying to control the natural search results.
Each article directory, Press release site and even the pages that host RSS feeds have different code, images, file names, java scrips and a multitude of other differences that stops them from getting hit by duplicate content.
There is more on a page than your article when a directory publishes your article. Search engines read all of the code, not just your articles text on the page. You do not have to worry about duplicate content with articles or press releases.
Jim
Thank you for your explanation on duplicate content.
I respectfully differ on a number of points.
Right from the outset, may i state that a lot of faulty deductions have been made about "duplicate content" which actually was never stated by Google nor was Google's intention, hence the confusion.
Next. On what authority do i speak? I have done an in depth research on this particular topic as depicted by the post on it on my blog which i have not stated the link to here to avoid been labelled as self serving. On request however, i am willing to post the link for details on this topic.
Now, any content that is significantly the same (does not necessarily have to be 100%) whether on the same site or across sites is duplicate content. The issue however is that duplicate content is not necessarily penalized by the search engines.
About the only reason stated by Google for penalizing is " Duplicate content with malicious intent or deceptive in origin". Apart from this, the only other case which actually is not a penalty is where only one preferred version (as determined by Google) of a duplicated or replicated web page on the same site is indexed while the others are not (where the duplicate is not deceptive in origin nor did it arise with malicious intent)
Thus the illustration made by the OP arrives at the right conclusion that duplicate content arising from duplicate article submissions is not removed from Google or other search engines index, but in my view, for the wrong reasons. It is not because the codes, images, file name, java scripts etc on the different sites web pages differs it from that on other sites but rather simply because it is not Google's policy to penalize such duplicate content.
One must however be careful here in that presence in Google index does not mean ranking highly in the index. Again, apart from the primary index, there is also the supplementary index.
Even though not also directly stated by Google, it is unlikely that a content which is significantly the same with another will feature on the first page of Google. Distinct this however from "spinned" articles or the application of "article leverage" which may have made the articles significantly different, even if bearing the same title.
Regarding content on the same site, it is also not simply based on the same website codes, images, file name, java scripts etc that duplicate content is not indexed but rather purely on the policy pronunciation of Google on this. Even if the codes etc. on different pages of the same site differ as it happens on different websites, the duplicate content will still not be indexed as long as it is on the same domain.