EaglesNiko
EaglesNiko

Reputation: 31

Downloading all PDF files from a website

I need to make a windows desktop application in c# that downloads all the PDFs from a website. I have the link to the website but the problem i am facing is that the PDFs are not in a specific folder on the website but are scattered all over.

The thing i need is help at finding all those links so i can download them or any other advices that could help me with my problem.

Thanks to all help in advanced.

Upvotes: 3

Views: 5716

Answers (2)

Sandeep Kumar Narware
Sandeep Kumar Narware

Reputation: 250

What you are trying to do is known as Web scraping, there are some libraries which can make your task easy one of them is IronWebScraper but its paid one.

An extensive list of NuGet packages is available here which can be used for web scraping purpose.

Upvotes: 0

lukas.pukenis
lukas.pukenis

Reputation: 13587

  1. Scrape through all the pages
  2. Find all the "*.pdf" URLs
  3. Reconstruct them and simply download :)

Please be more specific are you trying to get all the PDFs from the html page or from the whole domain ?

Upvotes: 1

Related Questions