Skip to main content

Frequently Asked Questions

1.1. How to download results in Demo version?

In the Demo version, the results of the work are not available for download. We provide them upon your request. Send us your requests and tell us which scraper you are interested in, and we will send you the results (the number of results is limited in the demo version).

1.2. Do I need to pay for anything after purchasing A-Parser?
1.3. Where and how can I pay for proxies?

When purchasing a license, you are provided with bonus proxies.

Lite - 20 threads for 2 weeks, Pro and Enterprise - 50 threads per month.

You can buy more threads or extend them in the Members Area in the Shop tab, in the Proxies subsection.

1.4. Could you set up a task for me for money?

Technical support for issues related to the operation of A-Parser is provided free of charge. Regarding paid assistance in composing tasks, setting up and training to work with A-Parser, you can contact here: Paid services for composing tasks, assistance with setup and training to work with A-Parser.

1.5. Can I pay for the scraper through Privat24 bank? Through KIWI?

The list of payment systems we work with is indicated here: buy A-Parser.

1.6. If I need to parse only the number of indexed pages in Yandex, which scraper should I buy?

For such purposes, the Lite version is sufficient, but Pro is more practical and flexible in operation.

1.7. Where can I see information about my license?
1.8. Is it possible to use purchased proxies from multiple IPs?


2. Installation, launch and update questions

2.1. I click the Download button, but the archive is not downloaded. What to do?

Check if you have free space on your hard drive, disable antivirus. Follow the installation instructions. Also, familiarize yourself with How to get started.

2.2. I bought the Enterprise version, but PRO is still being installed. What to do?

Uninstall the previous version. In the Members Area, check if your IP address is correct. Click the Update button before downloading. Download a newer version. More details in the installation instructions.

2.3. I installed the program, but it does not start. What to do?

Check running applications, disable antivirus, check available free RAM. Also, in the Members Area, check if your IP address is correct. More details: installation instructions.

2.4. What to do if I have a dynamic IP address?

Nothing serious, A-Parser supports working with dynamic IP addresses. Just every time it changes, you need to write it in the Members Area. To avoid these manipulations, it is recommended to use a static IP address.

2.5. What are the optimal server and computer parameters for installing the scraper?

All system requirements can be viewed here: system requirements.

2.6. I started the task. The scraper crashed and no longer starts. What to do?

You need to stop the server, check if the process is not hanging in memory, and try to start it again. You can also try to start A-Parser with stopping all tasks. To do this, you need to start with the -stoptasks parameter. Details on starting with a parameter.

2.7. What password to enter when opening the address

If this is the first launch, the password is empty. If not, it is the one you set. If you forgot your password, use password reset.

2.8. I enter my IP in the Personal Account, but it does not change in the field "Your current IP". Why?

The field Your current IP displays the IP that is currently valid for you, and it should not change. You should enter it in the IP 1 field.

2.9. Can I run two copies simultaneously?

You can only run two copies on one machine if they have different ports specified in the configuration file.

You can only run two A-Parsers on different machines simultaneously if you have purchased an additional IP in the Personal Account.

2.10. Is the parser tied to hardware?

No. Your IP is used to control licenses.

2.11. Question about update - should I update only .exe? config/config.db and files/Rank-CMS/apps.json - what are these files for?

Unless otherwise specified, only update .exe. The first file is for storing the A-Parser configuration, and the second is the database for determining CMS and the actual operation of the parser Rank::CMSRank::CMS.

2.12. I have Win Server 2008 Web Edition - the parser does not start...

A-Parser will not work on this version of the OS. The only option is to change the OS.

2.13. I have a 4-core processor. Why does A-Parser only use one core?

A-Parser uses from 2 to 4 cores, additional cores are only used for filtering, Result Builder, Parse custom result.

2.14. I started to get a segmentation failed error. What to do?

Most likely your IP has changed. Check in the Personal Account.

2.15. I have Linux. A-Parser started, but it does not open in the browser. How to solve?

Check the firewall - most likely it is blocking access.

2.16. I have Windows 7. A-Parser started, but it does not open in the browser and there is no Node.js process in the task manager. How to solve?

You need to check for Windows updates and install the latest available ones. Namely, you need Windows 7 SP1 update.

2.17. A-Parser does not start and the aparser.log shows the error FATAL: padding_depad failed: Invalid argument provided. at ./Crypt/Mode/ line 20.

Most likely there is a problem with some task (folder /config/tasks/) due to a disk error (for example, if the PC was powered off without proper shutdown), you can find out more by running A-Parser with the -morelogs flag.

Solution: start A-Parser with the -stoptasks parameter. If this does not help, then clean up the entire /config/tasks/. If the problem persists even after this, then install the parser again in a new directory and add the config from the old one (if it is not damaged).

3. Questions about A-Parser settings and other settings

3.1. How to configure the proxy checker?

Detailed instructions can be found here: proxy settings.

3.2. There are no live proxies - why?

Check your internet connection and the correctness of the proxy checker settings. If everything is done correctly, then this means that at the moment your proxy list does not contain working servers. The solution to this problem is to either use other proxies or try again later. If you are using our proxies, check the IP address in the Personal Account in the Proxy section. It is also possible that your provider is blocking access to other DNS, try to do the steps described here:

3.3. How to connect anti-captcha?

Detailed instructions on setting up anti-captcha here.

3.4. I changed the parameters in the parser settings, but they did not apply. Why?

The default preset (default) cannot be changed. If any changes have been made, you need to click Save as new preset, and then use it in your task.

3.5. Can I change the settings of a running task?

Yes, but not all. In the running task, you can click on pause and select Edit in the drop-down menu.

3.6. How to import a preset?

Click the button next to the task selection field in the Task Editor. Details here.

3.7. How to configure the scraper to not use a proxy?

In the settings of the required scraper, uncheck Use proxy.

3.8. I don't have the Add Override option button!

This option can be added directly in the Task Editor. Scraper options.

3.9. How to overwrite the same file with the results?

When creating a task, select the Overwrite file option.

3.10. Where to change the password for the scraper?
3.11. I set 6 million keys for parsing, also indicated that all domains should be unique. How can I make it so that when I put in new 6 million keys, only unique domains that do not intersect with the previous parsing are recorded?

You need to use the Save uniqueness option when creating the first task, and specify the saved database in the second one. Details in Additional options of the task editor.

3.12. How to bypass the limit of 1000 results for Google?

Use the option Parse all results.

3.13. How to bypass the limit of 1024 threads on Linux?
3.14. What is the thread limit on Windows?

Up to 10000 threads.

3.15. How to make requests unique?

Use the Unique queries option in the Queries block in the Task Editor. Unique queries option

3.16. How to disable proxy checking?

In Settings - Proxy checker settings select the required proxy checker and check Do not check proxies. Save and select the saved preset.

3.17. What is Proxy ban time? Can I set it to 0?

The time of proxy ban in seconds. Yes, you can.

3.18. What is the difference between Exact Domain and Top Level Domain in the SE::Google::PositionSE::Google::Position scraper?

Exact Domain - is an exact match, i.e. if is in the search results, and we are looking for, there will be no match. Top Level Domain compares the entire top-level domain, i.e. there will be a match here.

3.19. If I run a test parsing, everything works, if I run a regular one, I get an error Some error. Why?

Most likely the problem is with DNS, try to follow this instruction for configuring DNS.

3.20. Where is the Result format set?
3.21. The Dutch language is missing in SE::GoogleSE::Google, although it is in the Google settings. Why?

The Dutch language is Dutch, it is in the list. Details in improvement for adding the Dutch language.

4. Questions about scraping and errors during scraping

4.1. What are threads?

All modern processors can perform tasks in multiple threads, which significantly increases their execution speed. For comparison, an ordinary bus that transports a certain number of people in one unit of time is a regular, single-threaded processing, and a double-decker bus that transports twice as many people in the same time is a multi-threaded processing. A-Parser can process up to 10000 threads simultaneously.

4.2. The task does not start - it writes Some Error - why?

Check the IP address in the Personal account.

4.3. All requests fail, what should I do?

Most likely, the task is incorrectly composed or the request format is incorrect. Also, check if there are live proxies. You can also try increasing the Request retries option (more details here: failed requests).

4.4. How many accounts do I need to register to parse 1,000,000 keywords with SE::Yandex::WordstatSE::Yandex::Wordstat?

It is impossible to say exactly how many accounts are needed, as an account may become invalid after an unknown number of requests. But you can always register new accounts using the SE::Yandex::RegisterSE::Yandex::Register parser or simply add existing accounts to the files/SE-Yandex/accounts.txt file.

4.5. The task does not start, it says Error: Lock 100 threads failed (20 of limit 100 used), what should I do?

You need to increase the maximum number of threads available in the parser settings or decrease it in the task settings. Details in Settings.

4.6. Can I run 2 tasks simultaneously?

Yes, A-Parser supports the execution of multiple tasks simultaneously. The number of simultaneously running tasks is regulated in Settings - General settings: Maximum active tasks.

4.7. Where is the results file located?

On the Task queue tab, after each task is completed, you can download the results of the work. Physically, they are located in the results folder.

4.8. Can I download the results file if the parsing is not finished?

No, you cannot download the results until the parsing is finished. But you can copy it from the aparser/results folder when the task is stopped or paused.

4.9. Can I use your parser to parse 1,000,000 links for one query?

Yes, using the Parse all results option.

4.10. Can I parse Rank::CMSRank::CMS, Net::WhoisNet::Whois without a proxy?

Rank::CMSRank::CMS - yes, and even necessary. Net::WhoisNet::Whois - not desirable.

4.11. How to parse links from Google?

You need to use the SE::GoogleSE::Google.

4.12. Can the parser follow links?

Yes, the HTML::LinkExtractorHTML::LinkExtractor parser can do this when using the Parse to level option.

4.13. Google parsing is very slow, what should I do?

First of all, you need to look at the task logs, perhaps all requests are unsuccessful. If this is the case, you need to find out why the requests are unsuccessful and fix them. When parsing with SE::GoogleSE::Google, unsuccessful attempts in the task logs are often associated with the fact that Google shows captchas, which is normal. You can connect Antigate to bypass captchas so that the parser does not try to guess them. Also, there is an article that describes the factors that affect the parsing speed and how they affect it: speed and principle of parser operation.

4.14. Can I use your parser to parse links with text only in Japanese?

Yes, to do this, you need to set the required language in the parser settings, as well as use Japanese keywords.

4.15. Can I use your parser to parse links only in the .de or .ru domain zone?

Yes. To do this, you need to use a filter.

4.16. How to get each result in a file on a new line?

When formatting the result, use \n. Example:


4.17. How to parse the top 10 sites from Google?

Here is the preset:


How to import a sample into A-Parser.

4.18. I add a task, go to the Task queue tab - and it's not there! Why?

Either there is an error in the task composition, or it has already been completed and moved to Completed.

4.19. It says the file is not in utf-8, but I didn't change it, it's already utf-8, what should I do?

Check again. Also try changing the encoding, for example using Notepad++.

4.20. In the results file, everything is in one line, although I put line breaks in the task - why?

In additional A-Parser settings, use the line break CRLF (Windows).

But if you have already parsed without this option, use a more advanced viewer, such as Notepad++.

4.21. How long does it take to check the frequency of queries on Yandex for 1,000 queries?

This indicator depends heavily on the task parameters, server characteristics, proxy quality, etc., so it is impossible to give a definite answer.

4.22. How do I set up the parser to get a request-link as a result?

Result format:

$p1.serp.format('$query: $link\n')

As a result, it will be:

запрос: ссылка 1
запрос: ссылка 2
запрос: ссылка 3

4.23. How do I re-parse failed requests and where are they stored?

To save failed requests, select the corresponding option in the Queries section in the Task Editor. Failed requests are stored in queries\failed. You need to create a new task and specify the file with failed requests as the query file.

4.24. How to remove HTML tags when parsing text?

Use the Remove HTML tags option in the Results Builder.

4.25. How to extract only domains when parsing?

Use the Extract Domain option in the Results Builder.

4.26. What is the maximum file size of queries that can be used in the parser?

The sizes of query and result files are not limited and can reach terabyte values.

4.27. Why does the parser give Queries length limited to 8192 characters when I enter text in the query field?

This happens because the length of the query is limited to 8192 characters. To use longer queries, use files as queries.

4.28. What does Waiting threads - 3 mean?

This means that there are not enough proxies. Reduce the number of threads or increase the number of proxies.

4.29. In test parsing, it writes 596 SOCKS proxy error: Hello read error(Connection reset by peer) (0 KB) and does not parse, why?

This indicates non-working proxies.

4.30. What is the difference between the language of the results and the search country in the Google parser?

The difference is as follows: the search country is a binding of the results to a specific country. For example, if you search for buy windows with a binding to a specific country, then sites offering to buy windows specifically in this country will be a priority. And the language of the results is the language in which the results should be displayed.

4.31. A certain site is not being parsed. What could be the problem?

Often the problem is that there is a block due to an old user agent on the server side. Solved with a new user agent or the following code in the User agent parameter:

[% %]

4.32. The parser hangs, crashes. The log contains the line syswrite: No space left on device

A-Parser does not have enough space on the hard drive. Free up more space.

4.33. The parser constantly outputs none in the results (or an obviously incorrect result)

Perhaps the output format in the used service has changed. Check in Issues or in Versions, there may already be a message (solution) to this problem.

4.34. A window with the inscription Failed fetch news constantly appears
4.35. How to output the first n results of a search query?
4.36. How to track the chain of redirects?
4.37. How to check the indexation of a link on the donor?

For these purposes, there is a separate scraper: Check::BackLinkCheck::BackLink. More details in the discussion.

4.38. The scraper crashes on Linux. The log contains the following record: EV: error in callback (ignoring): syswrite() on closed filehandle at AnyEvent/ line...

Most likely, you need to tune the number of threads, as written in the Documentation: Tuning Linux for more threads.

4.39. Where can I see all possible parameters for their use through the API?

Getting an API request in the interface.

Also, you can generate a full task config in JSON. To do this, take the task code and decode it from base64.

4.40. I download images using Net::HTTPNet::HTTP, but they are all broken for some reason. What to do?

1) Check the Max body size parameter - you may need to increase it. 2) Check the line break format in A-Parser settings: Additional settings - Line break.

To prevent the image from being broken, UNIX format must be used.

4.41. How to get the admin contact from WHOIS?

This task can be easily solved using the Parse custom result function and a regular expression. Details in the discussion.

4.42. Regular expression for parsing phones
4.43. Determining sites without a mobile version
4.44. How to find out the name of the ns-server?
4.45. How to parse links to Yandex cache?
4.46. How to parse links to all pages of the site?
4.47. How to parse the title from the page?
4.48. How to parse all sites in a given domain zone?
4.49. How to collect all urls with parameters?
4.50. How to filter results by multiple criteria and split them in the report?
4.51. How to simplify the filter construction?
4.52. How to sort by files depending on the result?
4.53. Create new result directory every X number of files (English)
4.54. First steps with WordStat
4.55. Collecting text blocks >1000 characters
4.56. Outputting a certain amount of text from a page

This is also solved using Template Toolkit. More details in the discussion.

4.57. Checking competition and inclusion in the Google title
4.58. Filtering by the number of occurrences of the query in the anchor and snippet
4.59. How to get the content of an article in one line?
4.60. How to compare two string dates?
4.61. How to parse highlighted words from a snippet?
4.62. Example of a task using multiple scrapers
4.63. How to shuffle strings in the result and how to output a random number of results?
4.64. How to sign the result using MD5?
4.65. How to convert a Unix timestamp to a string representation of a date?
4.66. Parse to level, how to parse with a limit?
4.67. The scraper crashes on Linux when starting the task. The log contains the following lines: Can't call method "if_list" on an undefined value at IO/Interface/

You need to execute the command in the console:

apt-get --reinstall --purge install netbase

4.68. Error Cannot init Parser: Error: Failed to launch the browser process! [0429/] Running as root without --no-sandbox is not supported...

You need to run A-Parser not as root. Specifically: from the root user, you need to create a new user without root privileges (if there is one, then just use it) and then allow this user to interact with the A-Parser directory, then you need to log in from the new user and already run it from it.

To create a user under the root user, you can use this guide.

To allow the created user to interact with the A-Parser directory, you need to give the user rights. To do this, log in as the root user and use the command to give rights:

chown -R user:user aparser

4.69. Error Cannot init Parser: Error: Failed to launch the browser process! [0429/] No usable sandbox! Update your kernel or see...

Execute the command under the root user:

sysctl -w kernel.unprivileged_userns_clone=1

A-Parser does not need to be restarted.

For CentOS 7, the solution is in this thread. Execute the command under the root user:

echo "user.max_user_namespaces=15000" >> /etc/sysctl.conf

Then restart sysctl with the command:

sysctl -p

4.70. Error JavaScript execution error(): Error: Failed to launch the browser process! /aparser/dist/nodejs/node_modules/puppeteer/.local-chromium/linux-884014/chrome-linux/chrome: error while loading shared libraries: cannot open shared object file: No such file or directory...

The error occurs due to the absence of libraries in the OS to work with Chrome.

The list of necessary libraries for Chrome to work can be found in Chrome headless doesn't launch on UNIX.

4.71. Why is the captcha not being solved? In the log, it can be seen that A-Parser received question marks instead of the captcha answer from Xevil

In the region settings, you need to change to Russian. You only need to change on the additional tab. It does not affect captcha solving, but there will be a problem with encoding in the Hrefer if you change both there and there. option Unique requests option Unique requests