Check out Link Leecher, enter a URL and it will make a list of all the hyperlink URL’s on the webpage.
If I filter this for the term “api” I have limited it to only 2 items out of the initial 5
…you can also limit it by document format extension.
This is very similar to Tony Hirst’s bookmarklet, only this version outputs into an OPML URL, whereas Link Leecher outputs in a browser, text file, or CSV file for excel.
There is also a bookmarklet.
With these types of tools how do you stop it from collecting links other than story links eg. navigation links?
Also, what if you could say “generate a list for all the links on this page that have the hyperlink label “google”.
I’d also like it to temporarily mark up the actual webpage…each hyperlink in the body of the post could have a number next to it, this number corresponds to the number in the generated list, acting like a footnote. I’d like to do this in order to print it out, so when I’m reading a print version I know the URL of a given hyperlink (this can sometimes be helpful for some added context, as URL’s can sometimes contain the title and/or date of a blog post).
See my post, OPML of links on a given page, which points to a tool that can do this very thing…if Link Leecher incorporated this feature as well as OPML output, it would be a handy multi-purpose tool.