I’ve implemented my first web-scraper with Crawly following the tutorial published on the official docs.
I’m at the point where I start a single crawler module from iex and it seems it’s successfully crawling but it’s not writing to disk.
I’ve checked it’s crawling the website and extracting some items using
I’ve limited the total number of items to make sure it completed gracefully a run.
Crawly.Pipelines.WriteToFile module is on the pipeline and it’s configured to write to the user-writable folder
/tmp/crawly, but at the end of the run it’s as empty as ever.
I’ve enabled logging pointing in the same folder but that’s not written either.
I’ve never debugged an Elixir/Erlang app and I still find the idea a bit daunting.
Anyone has a clue why it’s not writing to disk?