Hello everyone,
Would like some direction, as I want to start a project and I'm not even sure if I'm headed the right way.
I have a local news site, which I would like to "scrape" various of the news items off it.
I already talked with their webmaster, and he said it's good to go.
Ok, so I believe (please correct) that a good tool for the job would be PHP and cURL.
What about using PHP Simple HTML DOM Parser?
I ask because I'm just not sure of where to head.
I'm a n00b at this, so diving into this project is various hours... before I even realize if what I'm doing will work or not.
So, that's the general direction.
Should I use PHP and cURL? (a reference doc I found here)
Also, I don't know how this works, yet I would like to "scrape" the page 4-5 times per day (at pre-set times), and then save the info over in my server.
So when a user to my website visits, I server the scraped information from my site (as opposed to re-scraping from the original site?).
Any thoughts on this project?
Thank you very much everyone!