Modify my web crawler to not have race condition

Completado Publicado hace 7 años Pagado a la entrega
Completado Pagado a la entrega

I would like my web crawler script be fixed so that when saving the file into chunks of specified number of lines to not have a race condition.

The race condition occurs when I run multiple instances of the crawler from the shell.

*For example, I need to have the output from the crawler, [login to view URL], [login to view URL], ... files have only 10 lines per file even when accessed simultaneously from other instances of the same script. It now has a problem of each outputted file having over 10 lines per file, and this is not desired.

The crawler needs to be able to make sure the outputted chunks of files are specified number of lines even when running multiple instances and accessing the same file. The actual specified number of the lines per file is 50,000 lines per file.

The variable in [login to view URL]

$lines_per_file determines how many lines per line

I believe the needed knowledge is: race condition, concurrency, forking etc.

Algoritmos Perl Arquitectura de software Extracción de datos web

Nº del proyecto: #10696959

Sobre el proyecto

2 propuestas Proyecto remoto Activo hace 7 años

Adjudicado a:

gangabass

I can update the code so you can run multiple instances and it will still have correct number of line in the output file. Thanks. Roman

$105 USD en 1 día
(602 comentarios)
7.6

2 freelancers están ofertando un promedio de $178 por este trabajo

mantislin

Hi sir, I am scraping expert, I have did too many similar projects, please check my feedback then you will know. Can you tell me more details? then I will provide demo data for you. Thanks, Kimi

$250 USD en 5 días
(308 comentarios)
7.7