05-30-2016, 01:53 PM
Ive got a bunch of sites id like to scrape like this :
- scrape list of profile pages from a suppliers directory listing site
- for each profile page extract the external url (eg the suppliers site), name and phone number (using custom data grabber)
- for the url of each page check if it has the a particular word on the page (using the page scanner add on)
- for each url check the find any email addresses on page
Some of those steps have been shorted, but in principal these are stand alone tasks, but each of which would have its own results page, what id like to do is have all of this data stored in a excel or csv sheet, in rows something like :
Profile Page, Suppliers URL, Name, Phone Number, Does the Suppliers page contain a particular phrase (yes / no), email address
Is it possible to return the data from these different functions and keep adding them to a spreadsheet to give me a list of results like above ?
- scrape list of profile pages from a suppliers directory listing site
- for each profile page extract the external url (eg the suppliers site), name and phone number (using custom data grabber)
- for the url of each page check if it has the a particular word on the page (using the page scanner add on)
- for each url check the find any email addresses on page
Some of those steps have been shorted, but in principal these are stand alone tasks, but each of which would have its own results page, what id like to do is have all of this data stored in a excel or csv sheet, in rows something like :
Profile Page, Suppliers URL, Name, Phone Number, Does the Suppliers page contain a particular phrase (yes / no), email address
Is it possible to return the data from these different functions and keep adding them to a spreadsheet to give me a list of results like above ?