Awesome Open Source
Awesome Open Source

jfiveparse: a java html5 parser

Maven Central Build Status

jfiveparse pass all the non scripted tests for the tokenizer and tree construction from the html5lib-tests suite.

It provide both fragment and full document parsing. It can parse directly from a String or by streaming through a Reader (note: the encoding must be known, currently the parser does not implement an autodetect feature).

Requires java 8.

[email protected].


As far as I know, there is no pure java html5 parser that currently pass the html5lib-tests suite (well, the more relevant tests :D).

Additionally, I wanted a library with a reduced footprint (and no dependencies). Currently the jar weight around ~150kb. The target is to keep it under 200kb.

Performance should be competitive with other java parsers.


jfiveparse is licensed under the Apache License Version 2.0.





compile 'ch.digitalfondue.jfiveparse:jfiveparse:0.9.0'


import ch.digitalfondue.jfiveparse.Document;
import ch.digitalfondue.jfiveparse.JFiveParse;
import ch.digitalfondue.jfiveparse.Node;

import java.util.List;

public class Example {

    public static void main(String[] args) {
        // directly from String
        Document doc = JFiveParse.parse("<html><body>Hello world!</body></html>");

        // from reader
        Document doc2 = JFiveParse.parse(new StringReader("<html><body>Hello world!</body></html>"));

        // parse fragment
        List<Node> fragment = JFiveParse.parseFragment("<p><span>Hello world</span></p>");

        // parse fragment from reader
        List<Node> fragment2 = JFiveParse.parseFragment(new StringReader("<p><span>Hello world</span></p>"));

It will print:

<html><head></head><body>Hello world!</body></html>
<html><head></head><body>Hello world!</body></html>
<p><span>Hello world</span></p>
<p><span>Hello world</span></p>


See directory:

Fetch all titles+links on the front page of HN

import ch.digitalfondue.jfiveparse.Element;
import ch.digitalfondue.jfiveparse.JFiveParse;
import ch.digitalfondue.jfiveparse.NodeMatcher;
import ch.digitalfondue.jfiveparse.Selector;

import java.nio.charset.StandardCharsets;

public class LoadHNTitle {

    public static void main(String[] args) throws IOException {
        try (Reader reader = new InputStreamReader(new URL("").openStream(), StandardCharsets.UTF_8)) {
            // select td.title > a.storylink
            NodeMatcher matcher ="td").hasClass("title").withChild().element("a").hasClass("storylink").toMatcher();
                    .forEach(a -> System.out.println(a.getTextContent() + " [" + a.getAttribute("href") + "]"));

Convert to the java DOM representation

If you need to generate a org.w3c.dom.Document from the ch.digitalfondue.jfiveparse.Document representation, there is a static method in the helper class: W3CDom.toW3CDocument.



template element handling

The template element is a "normal" element, so the child nodes are not placed inside a documentFragment. This will be fixed.


The &ntities; are by default (and by specification) parsed and interpreted. This behavior can be disabled by:

  • passing the enum "Option.DONT_TRANSFORM_ENTITIES" to the Parser
  • when calling Node.get{Inner,Outer}HTML(), pass the enum "Option.DONT_TRANSFORM_ENTITIES" for disabling the escaping. It's possible that something will be wrong in the generated document.

Preserving as much as possible the original document when serializing

By default, when parsing/serializing, the following transformations will be applied:

  • entities will be interpreted and converted
  • all the attribute values will be double quoted
  • tag and attribute names will be lower case
  • the "/" character used in self closing tag will be ignored
  • some whitespace will be ignored

Currently, jfiveparse can preserve the entities, the attribute quoting type and the case and the tag name case.

If you require to preserve as much as possible the document when serializing back in a string, pass the following parameters:

  • pass the enum "Option.DONT_TRANSFORM_ENTITIES" to the Parser
  • when calling Node.get{Inner,Outer}HTML(), pass the enums:

Uppercase handling in the tokenizer

Note: this is a deviation from the specification in term of implementation of the tokenizer, but globally, the end result is correct, as the attributes and tag names are then converted to lower case.

In the tokenizer, instead of applying the toLowerCase function on each character, the transformation is done in a single call in the TreeConstructor (see setTagName). This is used for saving the original case of the attributes and tag names.


  • additional doc
  • expand the typesafe matcher api
  • expand the Node api
  • add some methods from jquery too
  • cleanup code: but beware, no measurable slowdown is acceptable
  • keep track of lines, eventually chars too
  • profile
    • various optimizations...
    • TokenizerRCDataStates.handleRCDataState could be optimized
      • (textarea related)

mvn clean test jacoco:report

Get A Weekly Email With Trending Projects For These Topics
No Spam. Unsubscribe easily at any time.
Java (709,909
Html (432,729
Html5 (12,511
Html Parser (199
Related Projects