74 lines
		
	
	
		
			2.1 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			74 lines
		
	
	
		
			2.1 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
var tokenize = require('../tokenizer');
 | 
						|
var TokenStream = require('../common/TokenStream');
 | 
						|
var tokenStream = new TokenStream();
 | 
						|
var astToTokens = {
 | 
						|
    decorator: function(handlers) {
 | 
						|
        var curNode = null;
 | 
						|
        var prev = { len: 0, node: null };
 | 
						|
        var nodes = [prev];
 | 
						|
        var buffer = '';
 | 
						|
 | 
						|
        return {
 | 
						|
            children: handlers.children,
 | 
						|
            node: function(node) {
 | 
						|
                var tmp = curNode;
 | 
						|
                curNode = node;
 | 
						|
                handlers.node.call(this, node);
 | 
						|
                curNode = tmp;
 | 
						|
            },
 | 
						|
            chunk: function(chunk) {
 | 
						|
                buffer += chunk;
 | 
						|
                if (prev.node !== curNode) {
 | 
						|
                    nodes.push({
 | 
						|
                        len: chunk.length,
 | 
						|
                        node: curNode
 | 
						|
                    });
 | 
						|
                } else {
 | 
						|
                    prev.len += chunk.length;
 | 
						|
                }
 | 
						|
            },
 | 
						|
            result: function() {
 | 
						|
                return prepareTokens(buffer, nodes);
 | 
						|
            }
 | 
						|
        };
 | 
						|
    }
 | 
						|
};
 | 
						|
 | 
						|
function prepareTokens(str, nodes) {
 | 
						|
    var tokens = [];
 | 
						|
    var nodesOffset = 0;
 | 
						|
    var nodesIndex = 0;
 | 
						|
    var currentNode = nodes ? nodes[nodesIndex].node : null;
 | 
						|
 | 
						|
    tokenize(str, tokenStream);
 | 
						|
 | 
						|
    while (!tokenStream.eof) {
 | 
						|
        if (nodes) {
 | 
						|
            while (nodesIndex < nodes.length && nodesOffset + nodes[nodesIndex].len <= tokenStream.tokenStart) {
 | 
						|
                nodesOffset += nodes[nodesIndex++].len;
 | 
						|
                currentNode = nodes[nodesIndex].node;
 | 
						|
            }
 | 
						|
        }
 | 
						|
 | 
						|
        tokens.push({
 | 
						|
            type: tokenStream.tokenType,
 | 
						|
            value: tokenStream.getTokenValue(),
 | 
						|
            index: tokenStream.tokenIndex, // TODO: remove it, temporary solution
 | 
						|
            balance: tokenStream.balance[tokenStream.tokenIndex], // TODO: remove it, temporary solution
 | 
						|
            node: currentNode
 | 
						|
        });
 | 
						|
        tokenStream.next();
 | 
						|
        // console.log({ ...tokens[tokens.length - 1], node: undefined });
 | 
						|
    }
 | 
						|
 | 
						|
    return tokens;
 | 
						|
}
 | 
						|
 | 
						|
module.exports = function(value, syntax) {
 | 
						|
    if (typeof value === 'string') {
 | 
						|
        return prepareTokens(value, null);
 | 
						|
    }
 | 
						|
 | 
						|
    return syntax.generate(value, astToTokens);
 | 
						|
};
 |