Package Exports
- hyntax
- hyntax/lib/constants/ast-nodes
This package does not declare an exports field, so the exports above have been automatically detected and optimized by JSPM instead. If any package subpath is missing, it is recommended to post an issue to the original package (hyntax) to support the "exports" field. If that is not possible, create a JSPM override to customize the exports field for this package.
Readme
Hyntax
Simple HTML tokenizer and parser .
Not base on regexps. It's a legit parser.
Supports streaming. Can tokenize and parse HTML in chunks, meaning it can handle memory-sensitive scenarios.
Zero dependency. Hyntax is written from scratch as a case-study. You can read about how it works in my blog post.
Both Node.js and browser.
Usage
const util = require('util')
const { tokenize, parse } = require('hyntax')
const inputHTML = `
<html>
<body>
<input type="text" placeholder="Don't type">
<button>Don't press</button>
</body>
</html>
`
const { tokens } = tokenize(inputHTML)
const { ast } = parse(tokens)
console.log(JSON.stringify(tokens, null, 2))
console.log(util.inspect(ast, { showHidden: false, depth: null }))Usage in Browser
You can bundle Hyntax into your front-end application without any problems with Webpack, Rollup or Browserify. Single Node.js specific piece of code is native Node's streams. All mentioned bundlers have a client-side replacement for stream.
All components of Hyntax are separate files, so you can bundle only parts you actually need.
import tokenize from 'hyntax/lib/tokenize'
import parse from 'hyntax/lib/parse'
import TokenizeStream from 'hyntax/lib/tokenize-stream'
import ParseStream from 'hyntax/lib/parse-stream'Streaming
Stream parsing can be handy in a couple of cases:
- You have a huge HTML and you don't want or can't store it whole in the memory
- You need to generate tokens and AST while HTML is still being downloaded
With Hyntax it looks like this
const http = require('http')
const util = require('util')
const { TokenizeStream, ParseStream } = require('hyntax')
http.get('http://info.cern.ch', (res) => {
const tokenizeStream = new TokenizeStream()
const parseStream = new ParseStream()
let resultTokens = []
let resultAst
res.pipe(tokenizeStream).pipe(parseStream)
tokenizeStream
.on('data', (tokens) => {
resultTokens = resultTokens.concat(tokens)
})
.on('end', () => {
console.log(JSON.stringify(resultTokens, null, 2))
})
parseStream
.on('data', (ast) => {
resultAst = ast
})
.on('end', () => {
console.log(util.inspect(resultAst, { showHidden: false, depth: null }))
})
}).on('error', (err) => {
throw err;
})Tokenizer
Hyntax has tokenizer as a separate module. You can use generated tokens on their own or pass them further to a parser to build an AST. You can use default Hyntax parser or write a custom one which builds AST for some specific need.
Interface
tokenize(input<String>, [existingState<Object>], [options<Object>])For basic usage input argument is sufficient.
All other arguments are needed only for stream parsing and being used internaly by TokenizeStream class. You should worry about those only if you're going to have a custom stream implementation.
Arguments
inputHTML string to process
existingStateWhen input is coming in chunks and multiple calls of
tokenize(chunk)are required,existingStateparameter is used to pass result of previous call.Default value —
undefined.options.isFinalChunk<Boolean>Signal that current input chunk is the last one. Used for creating of the last token which does not have explicit ending. For example when input is interrupted in the middle of a tag content without reaching closing tag.
Default value —
trueoptions.positionOffset<Number>Number of characters processed by previous
tokenize(chunk)calls. Needed for correct calculation of tokens position when input is coming in chunks.Default value —
0
Returns
tokenize(input) -> { state<Object>, tokens<Array> }stateIs the current state of tokenizer. It can be persist and passed to the next tokenizer call if input is coming in chunks.
tokens<Array>Array of resulting tokens.
Types of tokens
Here is a high level overview of all possible tokens.

Token object
Each token is an object with several properties
{
type: <String>,
content: <String>,
startPosition: <Number>,
endPosition: <Number>
}typeOne of the type constants from lib/constants/token-types.js.
contentPiece of original HTML input which was recognized as a token.
startPositionIndex of a character in the input HTML string where token starts.
endPositionIndex of a character in the input HTML string where token ends.