SE::Google::TrustCheck - Site trust check
Scraper Overview
The scraper allows you to check the trust of a site in Google. All the capabilities of the SE::Google scraper are available. Thanks to the multi-threaded work of A-Parser, the processing speed of requests can reach 900 requests per minute, which on average allows to obtain up to 6200 results per minute.
The functionality of A-Parser allows you to save parsing settings for further use (presets), set parsing schedules, and much more.
Saving results is possible in the form and structure that you need, thanks to the built-in powerful templating engine Template Toolkit which allows you to apply additional logic to the results and output data in various formats, including JSON, SQL, and CSV.
Collected Data
- Google Trust Check
- Possible results -
0
,1
, and2
:0
- no additional link blocks1
or2
means that Google trusts the site, as it displays additional link blocks.1
- means the site has a horizontal block of links, and2
- means it has a large vertical block of links
Capabilities
- Collecting a database of trusted sites
- Supports selection of search country, domain, language of results, and other settings
Queries
As queries, you need to specify the URL of the desired site, for example:
http://uraldekor.ru/
http://a-parser.com/
http://www.yandex.ru/
http://google.com/
http://vk.com/
http://facebook.com/
http://youtube.com/
Query Substitutions
You can use built-in macros for automatic substitution of subqueries from files, for example, if we want to check sites/site based on a key database, we specify several main queries:
ria.ru
lenta.ru
rbc.ru
yandex.ru
In the query format, we specify a macro for substituting additional words from the Keywords.txt file, this method allows to check the site base based on the key base and get positions as a result:
$query {subs:Keywords}
This macro will create as many additional queries as there are in the file for each original search query, which in total will give [number of original queries (domains)] x [number of queries in the Keywords file] = [total number of queries]
as a result of the macro's operation.
You can also specify the protocol in the query format, so that only domains can be used as queries:
http://$query
This format will add http://
to each query.
Output Result Options
A-Parser supports flexible result formatting thanks to the built-in templating engine Template Toolkit, which allows it to output results in any form, as well as in a structured form, for example CSV or JSON
Exporting the trust check list
Result format:
$query: $trustrank\n
The result will display a list of links and their trust check.
Example result:
http://www.yandex.ru/: 2
http://a-parser.com/: 1
http://vk.com/: 2
http://uraldekor.ru/: 0
http://google.com/: 2
...
Links + anchors + snippets with position output
Output of links, anchors, and snippets in a CSV table
Saving related keywords
Keyword competition
Link indexing check
Saving in SQL format
Dumping results to JSON
Results processing
A-Parser allows processing results directly during scraping, in this section we have provided the most popular use cases for the SE::Google::TrustCheck scraper.
Saving domains with trust value "1"
Add a filter and select the trust variable $trustrank - Trust rank
from the drop-down list. Select the type: String equals
. Then you need to specify the trust value we need in the String 1
. With this filter, you will be able to remove all results with unnecessary trust.
Download example
How to import the example into A-Parser
eJx1VEtz2jAQ/iuMJod2hjhw6MU3woROOzSkCTkBB4HXRLWsNXrwGA//vSvZ2CYt
J3tf37f7raSSWW4y86LBgDUsXpSsCP8sZm9PcfwdcSshjufaGTv+gE3Wu+8FQ3OV
9VIhLWjWZwXXBrQHWNyoo6QEUu6kZf2S2VMBRIF70FokQEGRkJ2izrmlBkIa23Pp
fNrdzoE+xb07e2FeLhU738YpNB5PGqwWYDo4w8Fg0C1r2q8ZY9YQkLOKzqtU2Dku
u1D0j4UVqMgwoAw7r1YXIDMJc/jOi2FUC9oE3/ge5ljRQ+uekPXM8zBvwi34aFQJ
8uVrZI8egSeJ8JxcVgxe8Jb1XYldaE4h5XrNaPyJxtwPBgEgCHnpbsE+rC3ih4dK
X0ZILkD8rkpZnHJpoM8MdTzh1E/yOSJIIW5Rz4IU5C8ZqpGUU9iDbNMC/qMTMqFD
Mkqp6Edd+P+U2T8Y52bKLhWt/aCphwYlWI+zX21VglPckgDJmsaXIheWbDNGp/x+
BuTMAIpGumcvXY4aGpoauWanm1KA8mes3dyoaF1XY1xt59q5QZWK7aw+t5dMp+Z0
HWdqjHkhwc+lnJS0FgOv7SkZmXoN3mgb/Fw8DhR+9Mu9YxZRmp9vVauFFnQKv/kG
c1Kyy1pDbriU76/TboS1J4sMpznNlKGOtFsqfl89A9EG86U6HA7RiasEjiG4DS9C
Fdpn1TflG1gj1tYJnXXrkMI8t4Ut0qkkZc6r5oVpXqry1jsTl2da+x/zUhV4jXw6
+UhsEy7s8PwXfU7A2Q==
See also: Results filters
Link deduplication
Link deduplication by domain
Extracting domains
Removing tags from anchors and snippets
Filtering links by inclusion
Possible settings
Supports all settings of the SE::Google scraper, as well as additionally:
Parameter name | Default value | Description |
---|---|---|
Pages count | 1 | Number of search result pages to scrape (from 1 to 10) |
Links per page | 10 | Number of links in the search result on each page (from 10 to 100) |