Reputation: 4003
I have a project am working on. I am using php and mysql as backends. The problem is that i will need to create an online catalog for more than 10k products...each with pdf files as attachment. Now there is no way I could go through them one by one...so am here to ask you the expert for advise. How can I get this data from the suppliers web pages with some having .jsp extensions? Is there a way to extract this huge amount of data without doing it one by one? Thank you and looking forward to your advises.
Upvotes: 0
Views: 83
Reputation: 5245
Ask the suppliers if they have a web service or an API that you could interact with.
Upvotes: 0
Reputation: 14856
This is very bad practice -it's slow, your suppliers won't like it I guess and your code gets rubbish whenever the suppliers html code on their site changes.
So the best approach is to use some API if they provided any. If you are really in need of parsing other sites html code you should give cURL a try.
Upvotes: 1
Reputation: 151720
Yes, contact your supplier for product info. They won't be amused if you scrape and re-publish their website.
Is there a way to extract this huge amount of data without doing it one by one?
No, there is no other way to do this than visiting every product page. You can't store data that you don't receive.
Upvotes: 1