I have been always been entranced with how much data is publicly available on the web. Although, most of the time the data is not directly usable for other purposes than reading. With so many HTML parsing libraries out there, it is common practice to build an application for a particular site and a specific data set. Parsz tries to generalize the "parsing" and focus on the data structure you need. Instead of downloading, processing, and transforming the data each time you need to parse a web page, use parsz. Enough of the TV commercial…
I did NOT originally think of this. This guy did. I saw it, thought it was a great idea and decided such a simple idea could be implemented much easier by using Node.
The idea that caught me was to not
Since JSON has become very popular, I have not built in XML based output yet, although that may be just a npm module away.
There are many improvements and features to add, like regular expression support and custom functions. But overall, the tool/library (however the community decides to use it) is in a fairly usable state.
Here is the repo.
Have fun. Make sure to follow T&C.