node-readability

Turning any web page into a clean view.

Readability

Turn any web page into a clean view. This module is based on arc90's readability project.

  1. Optimized for more websites.
  2. Supporting HTML5 tags(article, section) and Microdata API.
  3. Focusing on both accuracy and performance. 4x times faster than arc90's version.
  4. Supporting encodings such as GBK and GB2312.
  5. Converting relative urls to absolute for images and links automatically(Thank Guillermo Baigorria & Tom Sutton).

Before -> After

npm install node-readability

read(html [, options], callback)

Where

  • html url or html code.
  • options is an optional options object
  • callback is the callback to run - callback(error, article, meta)

Example

var read = require('node-readability');
 
read('http://howtonode.org/really-simple-file-uploads', function(errarticlemeta) {
  // Main Article 
  console.log(article.content);
  // Title 
  console.log(article.title);
 
  // HTML Source Code 
  console.log(article.html);
  // DOM 
  console.log(article.document);
 
  // Response Object from Request Lib 
  console.log(meta);
 
  // Close article to clean up jsdom and prevent leaks 
  article.close();
});

NB If the page has been marked with charset other than utf-8, it will be converted automatically. Charsets such as GBK, GB2312 is also supported.

node-readability will pass the options to request directly. See request lib to view all available options.

node-readability has two additional options:

  • cleanRulers which allow set your own validation rule for tags.

If true rule is valid, otherwise no. options.cleanRulers = [callback(obj, tagName)]

read(url, {
  cleanRulers: [
    function(objtag) {
      if(tag === 'object') {
        if(obj.getAttribute('class') === 'BrightcoveExperience') {
          return true;
        }
      }
    }
  ]}, function(errarticleresponse) {
    //... 
  });
  • preprocess which should be a function to check or modify downloaded source before passing it to readability.

options.preprocess = callback(source, response, contentType, callback);

read(url, {
  preprocessfunction(sourceresponsecontentTypecallback) {
    if (source.length > maxBodySize) {
      return callback(new Error('too big'));
    }
    callback(null, source);
  }, function(errarticleresponse) {
    //... 
  });

article object

The article content of the web page. Return false if failed.

The article title of the web page. It's may not same to the text in the <title> tag.

The original html of the web page.

The document of the web page generated by jsdom. You can use it to access the DOM directly(for example, article.document.getElementById('main')).

response object from request lib. If you need to get current url after all redirect or get some headers it can be useful.

This lib is using jsdom to parser HTML instead of cheerio because some data such as image size and element visibility isn't able to acquire when using cheerio, which will significantly affect the result.

https://github.com/luin/node-readability/graphs/contributors

This code is under the Apache License 2.0. http://www.apache.org/licenses/LICENSE-2.0