Is there a way to prevent Google Alerts from picking up old press releases?
-
I have a client that wants a lot of old press releases (pdfs) added to their news page, but they don't want these to show up in Google Alerts. Is there a way for me to prevent this?
-
Thanks for the post Keri.
Yep, the OCR option would still make the image option for hiding "moo"
-
Harder, but certainly not impossible. I had Google Alerts come up on scanned PDF copies of newsletters from the 1980s and 1990s that were images.
The files recently moved and aren't showing up for the query, but I did see something else interesting. When I went to view one of the newsletters (https://docs.google.com/file/d/0B2S0WP3ixBdTVWg3RmFadF91ek0/edit?pli=1), it said "extracting text" for a few moments, then had a search box where I could search the document. On the fly, Google was doing OCR work and seemed decently accurate in the couple of tests I had done. There's a whole bunch of these newsletters at http://www.modelwarshipcombat.com/howto.shtml#hullbusters if you want to mess around with it at all.
-
Well that is how to exclude them from an alert that they setup, but I think they are talking about anyone who would setup an alert that might find the PDFs.
One other idea I had, that I think may help. If you setup the PDFs as images vs text then it would be harder for Google to "read" the PDFs and therefore not catalog them properly for the alert, but then this would have the same net effect of not having the PDFs in the index at all.
Danielle, my other question would be - why do they give a crap about Google Alerts specifically. There has been all kinds of issues with the service and if someone is really interested in finding out info on the company, there are other ways to monitor a website than Google Alerts. I used to use services that simply monitor a page (say the news release page) and lets me know when it is updated, this was often faster than Google Alerts and I would find stuff on a page before others who did only use Google Alerts. I think they are being kind of myopic about the whole approach and that blocking for Google Alerts may not help them as much as they think. Way more people simply search on Google vs using Alerts.
-
The easiest thing to do in this situation would be to add negative keywords or advanced operators to your google alert that prevent the new pages from triggering the alert. You can do this be adding advanced operators that exclude an exact match phrase, a file type, the clients domain or just a specific directory. If all the new pdf files will be in the same directory or share a common url structure you can exclude using the "inurl:-" operator.
-
That also presumes Google Alerts is anything near accurate. I've had it come up with things that have been on the web for years and for whatever reason, Google thinks they are new.
-
That was what I was thinking would have to be done... It's a little complicated on why they don't want them showing up in Alerts. They do want them showing up on the web, just not as an Alert. I'll let them know they can't have it both ways!
-
Robots.txt and exclude those files. Note that this takes them out of the web index in general so they will not show up in searches.
You need to ask your client why they are putting things on the web if they do not want them to be found. If they do not want them found, dont put them up on the web.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google index new data from my website page
Hi All, We have pages which are created few weeks before hand for Movie reviews in those pages we add value with adding the Movie cast and crew info and what ever info possible before the movie releases. The the movie releases we watch the movies and write reviews which is 500+ words. Now the issue is the pages are indexed a week before... How can i have these review pages scanned immediately when i have the complete review as the review content is not indexed for 3 to 5 days and the first day or 2 is when its important for the reviews to be seen in Google. Regards
On-Page Optimization | | AlexisWithers0 -
Webmaster sitemap how to tell Google to recheck and no of times to check a day?
Webmaster sitemap how to tell Google to recheck and no of times to check a day? But how often google check if the sitemap version changed and how to notify google about the change happened
On-Page Optimization | | bsharath0 -
Any idea how Google is doing this? Is it schematic? http://techcrunch.com/2014/02/28/google-adds-full-restaurant-menus-to-its-search-results-pages/
Google is now showing menus on select searches. Any idea how they are getting this information? I would like to make sure my clients get visibility this way.
On-Page Optimization | | Ron_McCabe0 -
DMCA Complaint to Google
I have several sites copying the content from my pages. I filed numerous complaints with Google and checked under the DMCA dashboard and it's showing my requests are approved; however, still several days later the infringing pages still remain. I'm a bit confused by this as I thought Google was supposed to remove the page. I guess I'm not understand why it was "approved" when I don't see changes. Am I supposed to contact the site's owner to get the pages down now? I don't want to show duplicate content to my site. Please help!
On-Page Optimization | | tutugirl0 -
Google Webmaster Tools notice of detected unnatural links
Hi Has anybody else ever recieved this message from Google in Google Webmaster Tools and what action did you take to overcome the problem and get your position back - Dear site owner or webmaster of............... We've detected that some of your site's pages may be using techniques that are outside Google's Webmaster Guidelines. Specifically, look for possibly artificial or unnatural links pointing to your site that could be intended to manipulate PageRank. Examples of unnatural linking could include buying links to pass PageRank or participating in link schemes. We encourage you to make changes to your site so that it meets our quality guidelines. Once you've made these changes, please submit your site for reconsideration in Google's search results. If you find unnatural links to your site that you are unable to control or remove, please provide the details in your reconsideration request. If you have any questions about how to resolve this issue, please see our Webmaster Help Forum for support. Sincerely, Google Search Quality Team
On-Page Optimization | | AMG1000 -
What is the best way to format an xml sitemap?
I am wondering if the urls should be in alphabetical order or if they should be set out in a way that reflects the sites hierarchy? Thanks.
On-Page Optimization | | Webat0 -
I think I`ve caught some kind of google filter on my site.
What if the PA and the DA on my domain and the entire site is 1. Most of the pages on the site were empty or not unique. Now I`m adding new pages with unique content. I have only one position in the top 10. The remaining 15 positions are above the top. What should I do to increase my PA & DA and to have top 10 positions by other keywords?
On-Page Optimization | | ATCnik0 -
Should I let Google index tags?
Should I let Google index tags? Positive? Negative Right now Google index every page, including tags... looks like I am risking to get duplicate content errors? If thats true should I just block /tag in robots.txt Also is it better to have as many pages indexed by google or it's should be as lees as possible and specific to the content as much as possible. Cheers
On-Page Optimization | | DiamondJewelryEmpire0