This post has been de-listed
It is no longer included in search results and normal feeds (front page, hot posts, subreddit posts, etc). It remains visible only via the author's post history.
Hey all. I'm looking for a program that can scan a webpage's source code for a specific footprint, then use a link on that page to jump to a new URL and repeat the process.
So for example, say we have a series of web-articles, it would go something like 1. scan Article 1 source code for term, 2. find the 'Next Article' link's URL, 3. scan Article 2 source code for term, etc. etc etc.
So basically I'm looking for something similar to ScrapeBox's page scanner but where you only need to load one URL, set a target-footprint and target-'next link', and it will keep jumping/returning results until it runs out of 'next links'. The kind of thing that would help me find all instances of the word "consumer interest" in a series of 10,000 sequential/linked URLs without having to load each URL manually.
Any ideas?
Subreddit
Post Details
- Posted
- 6 years ago
- Reddit URL
- View post on reddit.com
- External URL
- reddit.com/r/AskProgramm...