How do batched URL metrics work in terms of rows and rate limit?
-
I am using the free API plan to get URL metrics and batching my calls like this:
https://github.com/seomoz/SEOmozAPISamples/blob/master/php/batching_urls_sample.php
How does this work in terms of rows and limits? If I do a batch of 10 urls does it count as 1 row? or 10? Do I have to wait 10 seconds before calling the next batch?
-
Hi there - Kristina from Moz's Help Team here.
A batch of 10 URLs will count as 10 rows and you would need to wait 10 seconds before making the next query. If you make your next query too quickly you will be hit with a throttle which lasts 30 seconds. If you make another query during that period, you potentially could push that 30 second throttle even further out.
I hope this helps! In the future, feel free to email our team anytime at help@moz.com and someone from our team will be happy to help out!
Have a great day.
-Kristina
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Api limits on free plan
Good afternoon. I tried connecting free access to api https://moz.com/help/links-api/v1-archive/v1-free-api-access to retrieve DA data. After several requests to the api I got a message: "Your monthly row limit reached" Can you tell me what this error is related to? Or Moz doesn't have free api access anymore?
API | | Dima124dsfg1 -
API metrics for discovered and lost linking domains
Hello, Could you tell me, can I use the api to get such data as ‘Discovered in the last 60 days’ and ‘Lost Links’?
API | | OlegKireyenka
If yes, which values of ‘Cols’ in URL Metrics I can use? Thank!0 -
Do you fetch website titles from paid api https://moz.com/help/guides/moz-api/mozscape/api-reference/url-metrics?
We are using one of your API: https://moz.com/help/guides/moz-api/mozscape/api-reference/url-metrics on our website and it does not show up the title for each website.
API | | SOSCreatives
But when I see the title of the same website through your extension Moz Bar, then it does show the title from the extension of that website. Can you tell me what is missing here?0 -
Alternatives to Supermetrics to get Moz metrics into Google Sheets?
Hi, Interested in getting Moz API metrics into Google Sheets on an automated scheduled. Supermetrics can do this, but I am curious to know if there are any alternatives, free or paid. Thanks! 🙂
API | | GOODSIR1 -
Mozcheck.com not working with API, anyone else having this problem?
We have been using MozCheck.com with our API for 3 years, today it stopped working. Our account is in good standing, nothing has been changed.
API | | troytlb0 -
Moz Thumbs not working? @MozHelp
Hi all, I'm noticing an issue with Moz thumbs. Last night before bed, I am pretty positive I gave EGOL and Laura a thumbs up for their answers to this question. This morning I noticed there were no thumbs up for their responses. At first I thought maybe I didn't give them a thumbs up, and proceeded to thumb them again. Everything appeared to work, however after returning to the post I noticed my thumbs didn't appear to be recorded. Has anybody noticed this issue?
API | | donford0 -
Suggestion - How to improve OSE metrics for DA & PA
I am sure everyone is aware at Moz, that although the Moz link metrics ( primarily I am talking about DA & PA) are good, there is a lot of room for improvement, and that there are a lot of areas where the metric values given to some types of site are well out of whack with what their "real" values should be. Some examples
API | | James77
www.somuch.com (Link Directory) - DA 72
www.articlesbase.com (Article Directory) - DA 89
www.ezinearticles.com (Article Directory) - DA 91 I'm sure everyone would agree that links from these domains are not as powerful (if of any value at all), as their DA would suggest, and therefore by definition of how moz metrics work, the sites these have links from such sites are also inflated - thus they throw the whole link graph out of whack. I have 2 suggestions which could be used to singularly or in conjunction (and obviously with other factors that Moz use to calculate DA and PA) which could help move these values to what they should more realistically be. 1/. Incorporate rank values.
This is effectively using rank values to reverse engine what google (or other engines) as a "value" on a website. This could be achieved (if moz were not to build the data gathering system itself), by intergrating with a company that already provides this data - eg searchmetrics, semrush etc. As an example you would take a domian and pull in some rank values eg http://www.semrush.com/info/somuch.com?db=us - where you could use traffic, traffic price, traffic history as a metric as part of the overall Moz scoring alogrithm. As you can see from my example according to SEMRush the amount of traffic and traffic price is extreamly low for what you would expect of a website that has a DA of 72. Likewise you will find this for the other two sites and similarly to pretty much any other site you will test. This is essentially because your tapping into Googles own ranking factors, and thereby more inline with what real values (according to Google) are with respect to the quality of a website. Therefore if you were to incorporate these values, I believe you could improve the Moz metrics. 2/. Social Sharing Value
Another strong indicator of quality the amount of social sharing of a document or website as a whole, and again you will find as with my examples, that pages on these sites have low social metrics in comparison to what you would normally associate with sites of these DA values. Obviously to do this you would need to pull social metrics of all the pages in your link DB. Or if this we to tech intense to achieve, again work with a partner such as searchmetrics, which provide "Total Social Interations" on a domain level basis. Divide this value by the number of Moz crawled pages and you would have a crude value of the overall average social scorability of a webpage on a given site. Obviously both the above, do have their flaws if you looked at them in complete isolation, however in combination they could provide a robust metric to use in any alogrithm, and in combination with current moz values used in the alogrithm I believe you could make big strides into improving overall Moz metrics.1 -
Batch URL Error 413
Hello, i am using the free mozscape api to get the domain and page authority of urls. i am batching the url as shown in the sample code however i have over 500 URL i need to check and when running the request i get the following: stdClass Object ( [status] => 413 [error_message] => Too many urls in batch. Batches must be less than or equal to 200 urls. ) When i change it to 200 urls the request works fine, now is there a way to enable me to batch all 500 urls at once? i did read that the beta api is capable of batching urls in one request : http://moz.com/blog/400-higher-throughput-mozscape-api-now-in-beta-and-seeking-testers Has this been implemented yet into the current api? Thanks
API | | pauledwards0