OSE Advanced for specific directory only?
-
Pls how do I set up Open Site Explorer Advanced to look for backlinks to files in ONLY a certain directory?
so if the domain is 1234.com, I don't care about links to 1234.com. OSE can find those very neatly!
What I do care about is ONLY links to 1234.com/profiles and I can't seem to figure out a way to do this.
the /profiles directory has thousands of profiles in it - and we think we have hundreds of thousands of backlinks - so ideally, I'd like to use regex or DOS like filtering to look at only those which start with 9 or 8 or whatever.
-
so yeah - this has worked.
thanks everybody!
you don't see the data in the normal interface, but there in the download from advanced - woot there it is!
well worth the wait!
-
thanks - already tried that but just in case I didn't do it right, I just tried again.
not happening - any one of the profiles has hundreds of backlinks from dozens of domains.
the Link Metrics (shown at the top of the page) result from /profile shows only 3 domains and 300 links. This is easily off by 100 to 1000.
of course, the advanced download may show more as it has been a few minutes and it has not yet completed the spidering process.
what was is not happening- at least in the OSE standard reports - is OSE is not looking at all files in that subdirectory somehow.
I'll let you know shortly if the download shows different results.
-
OSE does allow you to export backlinks to a particular subfolder, but it can be a little tricky.
1. On the OSE homepage, enter the exact directory path you want OSE to explore. This is the screen you want to enter it on, not the advanced report screen. Hit search.
2. The directory needs to end in a backslash "/". Sometimes OSE will try to redirect you to a directory without a slash. For example, it tries to redirect "seomoz.org/blog/" to "seomoz.org/blog" (no slash). But it gives you the option to see your original entry. Always choose the entry with the backslash.
3. After this, on the Advanced reports tab, choose the "Links that Link to;" option that says Any page in this subfolder. Choose any other options you want on this screen.
Viola. You are done!
This should return links only pointing to the subdirectory you specified in step 1. If you have any trouble, feel free to contact customer support.
-
tried it, (with and without the trailing slashs) - it doesn't seem to work - maybe you'll have better luck.
getting them all and filtering also doesn't work as we are well in excess of the size of the data set.
can we work with boolean or regex filters?
RE the size - well this is both good and bad news... we'll get into that in a bit.
-
Nope didn't work. Not sure what that feature actually does now. Once I download the CVS, it was possible to use excel to sort by the url that included the subdir. I know that is not what you are looking for, but it may help until someone else that knows chimes in.
-
I am not 100% sure, but it looks like on the advanced tab you can download a CSV that will return links to a sub folder. In section 2 of the tab there is an option for: "Any page in this subfolder".
I entered a www.seomoz.org/users to run the test. I will report back once the report is done.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Block Moz (or any other robot) from crawling pages with specific URLs
Hello! Moz reports that my site has around 380 duplicate page content. Most of them come from dynamic generated URLs that have some specific parameters. I have sorted this out for Google in webmaster tools (the new Google Search Console) by blocking the pages with these parameters. However, Moz is still reporting the same amount of duplicate content pages and, to stop it, I know I must use robots.txt. The trick is that, I don't want to block every page, but just the pages with specific parameters. I want to do this because among these 380 pages there are some other pages with no parameters (or different parameters) that I need to take care of. Basically, I need to clean this list to be able to use the feature properly in the future. I have read through Moz forums and found a few topics related to this, but there is no clear answer on how to block only pages with specific URLs. Therefore, I have done my research and come up with these lines for robots.txt: User-agent: dotbot
Moz Pro | | Blacktie
Disallow: /*numberOfStars=0 User-agent: rogerbot
Disallow: /*numberOfStars=0 My questions: 1. Are the above lines correct and would block Moz (dotbot and rogerbot) from crawling only pages that have numberOfStars=0 parameter in their URLs, leaving other pages intact? 2. Do I need to have an empty line between the two groups? (I mean between "Disallow: /*numberOfStars=0" and "User-agent: rogerbot")? (or does it even matter?) I think this would help many people as there is no clear answer on how to block crawling only pages with specific URLs. Moreover, this should be valid for any robot out there. Thank you for your help!0 -
Very Slow Advanced Reports
Hello All - I've been running some Advanced Reports again lately, and they seem much slower than I remember from last time I ran some. Currently I've got one (Inbound Links) report at 2,500 out of 10,000 links retrieved through LSAPI, and it's been at that point for about 6 hours. Did something get cloggered on the reports, or is it this just the expected performance?
Moz Pro | | icecarats0 -
Bulk OSE Open Site Explorer Tool?
I am trying to do some spring cleaning for a client and hoping to prune any unnecessary domains. Is there a tool that will check, in bulk, these domains through Open Site Explorer? I've looked through all the different Excel spread sheet apps and google doc apps but they are incredibly buggy if they work at all since SEOmoz changed their data limits. Maybe a new tool has been updated in the last few months that I am not aware of. Thanks!
Moz Pro | | kerplow0 -
Why has my OSE csv report been finalizing for 3 days?
Hi guys, My CSV report has been finalizing for 3 days. Is that right? I read some posts from last year when there was a backlog Is that happening again. Can someone have a look at that please? Many thanks.
Moz Pro | | Rosewood0 -
Scrolling in OSE
Hi Guys, Is anyone else having issues when scrolling in OSE remembering what the column headers are? Just throwing out a suggestion for the next OSE update, could the column headers remain at the top of the links table when scrolling? This would be very useful especially on the top pages table which now has many columns! Thanks guys! Danni
Moz Pro | | MB070 -
OSE Link Report Question.
Hi, when i export a link report as a csv file via the export tool on OSE, there is a column in the exported table called "Numder of Link". Please explain me, what does this column says exactly. Its located between "Domain Authorithy" and "Number of Domains Linking to Domain". Thanx, Seb.
Moz Pro | | TheHecksler0 -
What is the quickest way to get OSE data for many URLs all at once?
I have over 400 URLs in a spreadsheet and I would like to get Open Site Explorer data (domain/page authority/trust etc) for each URL. Would I use the Linkscape API to do this quickly (ie not manually entering every single site into OSE)? Or is there something in OSE or a tool I am overlooking? And whatever the best process is, can you give a brief overview? Thanks!! -Dan
Moz Pro | | evolvingSEO0 -
Can someone explain why I have been seeing an increase in the number of Linking Page URLs in OSE that link directly to downloads?
Ever since the last couple Linkscape updates when doing competitive back link analysis I have noticed a large increase in the number of URLs of Linking Pages in OSE that result in an immediate file download. The majority of the time these downloads are not common files ie PDF, DOC files. For example, these were all in a competitors back link profile: http://download.unesp.br/linux/debian/pool/main/i/isc-dhcp/isc-dhcp-relay-dbg_4.1.1-P1-17_ia64.deb http://snow.fmi.fi/data/20090210_eurasia_sd_025grid.mat http://www.rose-hulman.edu/class/me/HTML/ES204_0708_S/working model examples/Le25 mad hatter.wm?a=p&id=145880&g=5&p=sia&date=iso&o=ajgrep These are just a few I came across for a single competitor. Is this sketchy black hat SEO, some sort of error, actual links, or something else? Any information on this subject would be helpful. Thank you.
Moz Pro | | Gyi0