Skip to main content

SE::Google::Compromised - Checking for the inscription This site may be hacked in Google

Overview of the scraper

The Google Compromised scraper allows you to check for the presence of the inscription This site may be hacked in Google's search results. Using the Google Compromised scraper, you can check your own domain databases for the inscription. You can find more details about this inscription in the Google Search Help Center.

A-Parser's functionality allows you to save parsing settings for future use (presets), set a parsing schedule, and much more.

Results can be saved in the form and structure you need, thanks to the built-in powerful template engine Template Toolkit which allows you to apply additional logic to the results and output data in various formats, including JSON, SQL and CSV.

Collected data

  • Checking for the presence of the inscription This site may be hacked in Google

what data the SE::Google::Compromised scraper collects

Capabilities

Use cases

  • Checking a list of domains for the presence of the inscription This site may be hacked in Google
  • Monitoring your own domains

Queries

As queries, you must specify the URL of the site being searched, for example:

http://a-parser.com/  
http://www.yandex.ru/
http://google.com/
http://russbehnke.com/
http://www.bmlaroca.cat/
http://vk.com/
http://facebook.com/
http://youtube.com/

Query substitutions

You can use built-in macros for automatic substitution of subqueries from files, for example, we want to check sites/a site against a keyword database, we will specify several main queries:

ria.ru
lenta.ru
rbc.ru
yandex.ru

In the query format, we specify a macro for substituting additional words from the file Keywords.txt, this method allows you to check a database of sites against a database of keywords and get positions as a result:

$query {subs:Keywords}

This macro will create as many additional queries as there are in the file for each initial search query, which in total will give [number of initial queries (domains)] x [number of queries in the Keywords file] = [total number of queries] as a result of the macro execution.

You can also specify the protocol in the query format so that only domains can be used as queries:

http://$query 

This format will append http:// to each query.

Output results examples

A-Parser supports flexible result formatting thanks to the built-in template engine Template Toolkit, which allows it to output results in arbitrary form, as well as in a structured form, such as CSV or JSON

Exporting the list of compromise checks

Result format:

$query: $compromised\n

An example of a result showing the URL and the presence of the inscription This site may be hacked in Google:

http://a-parser.com/: 0
http://www.bmlaroca.cat/: 1
http://russbehnke.com/: 0
http://www.yandex.ru/: 0
http://google.com/: 0

Same as in SE::Google.

Same as in SE::Google.

Same as in SE::Google.

Keyword competition

Same as in SE::Google.

Checking links indexation

Same as in SE::Google.

Saving in SQL format

Same as in SE::Google.

Dumping results to JSON

Same as in SE::Google.

Results processing

A-Parser allows you to process results directly during scraping; in this section, we have presented the most popular use cases for the SE::Google::Compromised scraper

Saving domains with a check value of "1"

Add a filter and select the check value variable $compromised - Is site compromised from the drop-down list. Select type: String equals. Next, in String you need to enter the value we need 1. With this filter, you can remove all results with an unnecessary value.

Example of filtering
Download example

How to import the example into A-Parser

eJx1VE1z2jAQ/SseTQ7tDDFw6MU3woROOzSkITkBB2GtiYosCUnmYzz8965kYxta
btZ+vH37dtclcdRu7asBC86SZFESHb5JQubPSfJdqY2AJBmrXBuVcwsseowmXDgw
ERxprgWQHtHUWDA+f3EvDaMYZLQQjvRK4k4asITagzGceQjO8J0pk1OHBEIY2VNR
+LCHXQHmlEQPaYu3XEpyvo+EYceTAWc42A7ScDAYdNOy0AkG1DUTkl5RrvzvVTDs
Ciq6YPittONK4sOCtOS8Wl2g7CT04tnrYVyL2jjndA/vqiIArRmFhReah54ZdeC9
cSXKl6+xO3oEyhj3NamoKnjV26ofku8COakw1uuGAkywITQ5CABBzAu7RS0uQYgi
5P6uckiSUWGhRyxSnVAkwm49HKWhTplZ0ADtJVFyJMQU9iDasID/VHDBcEVGGSb9
qBP/HzL7B+PctNcthRM/GOTQoITX0+xXm8XUVG2wc7bGvgXPucO3HatC+sEM0LgF
0I1mL16zXBloytTIdXU8Ew3Sr1c7spFuTVdtXI3l2pgqmfHNrF7ZS2Qh3/EWZ9If
jQDflyyEwLFYeGvXY2TrMfhHS/A2eRxK+NYvR0ecUsL+nFdUteG4ft88wRyV7Fat
IVMqxMfbtOsh7Urh49M5nfT79LG6/hgvpx9FS1nbD4dDfKKSwTE2RdexCT+H23BT
WLuGT7mtPFcw61xQo1Iap9T1iafmYKNwaVG486r5/TR/sfLuTygpz7gWf+xrleE1
9PFow2HYcMnD81/p/MfQ
tip

See also: Result Filters

Same as in SE::Google.

Same as in SE::Google.

Extracting domains

Same as in SE::Google.

Removing tags from anchors and snippets

Same as in SE::Google.

Same as in SE::Google.

Possible settings

Supports all settings of the SE::GoogleSE::Google, scraper, as well as additionally:

Parameter NameDefault ValueDescription
Pages count1Number of search result pages to scrape (from 1 to 10)