1015 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
			
		
		
	
	
			1015 lines
		
	
	
		
			38 KiB
		
	
	
	
		
			JavaScript
		
	
	
	
	
	
| "use strict";
 | |
| 
 | |
| exports.__esModule = true;
 | |
| exports["default"] = void 0;
 | |
| var _root = _interopRequireDefault(require("./selectors/root"));
 | |
| var _selector = _interopRequireDefault(require("./selectors/selector"));
 | |
| var _className = _interopRequireDefault(require("./selectors/className"));
 | |
| var _comment = _interopRequireDefault(require("./selectors/comment"));
 | |
| var _id = _interopRequireDefault(require("./selectors/id"));
 | |
| var _tag = _interopRequireDefault(require("./selectors/tag"));
 | |
| var _string = _interopRequireDefault(require("./selectors/string"));
 | |
| var _pseudo = _interopRequireDefault(require("./selectors/pseudo"));
 | |
| var _attribute = _interopRequireWildcard(require("./selectors/attribute"));
 | |
| var _universal = _interopRequireDefault(require("./selectors/universal"));
 | |
| var _combinator = _interopRequireDefault(require("./selectors/combinator"));
 | |
| var _nesting = _interopRequireDefault(require("./selectors/nesting"));
 | |
| var _sortAscending = _interopRequireDefault(require("./sortAscending"));
 | |
| var _tokenize = _interopRequireWildcard(require("./tokenize"));
 | |
| var tokens = _interopRequireWildcard(require("./tokenTypes"));
 | |
| var types = _interopRequireWildcard(require("./selectors/types"));
 | |
| var _util = require("./util");
 | |
| var _WHITESPACE_TOKENS, _Object$assign;
 | |
| function _getRequireWildcardCache(nodeInterop) { if (typeof WeakMap !== "function") return null; var cacheBabelInterop = new WeakMap(); var cacheNodeInterop = new WeakMap(); return (_getRequireWildcardCache = function _getRequireWildcardCache(nodeInterop) { return nodeInterop ? cacheNodeInterop : cacheBabelInterop; })(nodeInterop); }
 | |
| function _interopRequireWildcard(obj, nodeInterop) { if (!nodeInterop && obj && obj.__esModule) { return obj; } if (obj === null || typeof obj !== "object" && typeof obj !== "function") { return { "default": obj }; } var cache = _getRequireWildcardCache(nodeInterop); if (cache && cache.has(obj)) { return cache.get(obj); } var newObj = {}; var hasPropertyDescriptor = Object.defineProperty && Object.getOwnPropertyDescriptor; for (var key in obj) { if (key !== "default" && Object.prototype.hasOwnProperty.call(obj, key)) { var desc = hasPropertyDescriptor ? Object.getOwnPropertyDescriptor(obj, key) : null; if (desc && (desc.get || desc.set)) { Object.defineProperty(newObj, key, desc); } else { newObj[key] = obj[key]; } } } newObj["default"] = obj; if (cache) { cache.set(obj, newObj); } return newObj; }
 | |
| function _interopRequireDefault(obj) { return obj && obj.__esModule ? obj : { "default": obj }; }
 | |
| function _defineProperties(target, props) { for (var i = 0; i < props.length; i++) { var descriptor = props[i]; descriptor.enumerable = descriptor.enumerable || false; descriptor.configurable = true; if ("value" in descriptor) descriptor.writable = true; Object.defineProperty(target, descriptor.key, descriptor); } }
 | |
| function _createClass(Constructor, protoProps, staticProps) { if (protoProps) _defineProperties(Constructor.prototype, protoProps); if (staticProps) _defineProperties(Constructor, staticProps); Object.defineProperty(Constructor, "prototype", { writable: false }); return Constructor; }
 | |
| var WHITESPACE_TOKENS = (_WHITESPACE_TOKENS = {}, _WHITESPACE_TOKENS[tokens.space] = true, _WHITESPACE_TOKENS[tokens.cr] = true, _WHITESPACE_TOKENS[tokens.feed] = true, _WHITESPACE_TOKENS[tokens.newline] = true, _WHITESPACE_TOKENS[tokens.tab] = true, _WHITESPACE_TOKENS);
 | |
| var WHITESPACE_EQUIV_TOKENS = Object.assign({}, WHITESPACE_TOKENS, (_Object$assign = {}, _Object$assign[tokens.comment] = true, _Object$assign));
 | |
| function tokenStart(token) {
 | |
|   return {
 | |
|     line: token[_tokenize.FIELDS.START_LINE],
 | |
|     column: token[_tokenize.FIELDS.START_COL]
 | |
|   };
 | |
| }
 | |
| function tokenEnd(token) {
 | |
|   return {
 | |
|     line: token[_tokenize.FIELDS.END_LINE],
 | |
|     column: token[_tokenize.FIELDS.END_COL]
 | |
|   };
 | |
| }
 | |
| function getSource(startLine, startColumn, endLine, endColumn) {
 | |
|   return {
 | |
|     start: {
 | |
|       line: startLine,
 | |
|       column: startColumn
 | |
|     },
 | |
|     end: {
 | |
|       line: endLine,
 | |
|       column: endColumn
 | |
|     }
 | |
|   };
 | |
| }
 | |
| function getTokenSource(token) {
 | |
|   return getSource(token[_tokenize.FIELDS.START_LINE], token[_tokenize.FIELDS.START_COL], token[_tokenize.FIELDS.END_LINE], token[_tokenize.FIELDS.END_COL]);
 | |
| }
 | |
| function getTokenSourceSpan(startToken, endToken) {
 | |
|   if (!startToken) {
 | |
|     return undefined;
 | |
|   }
 | |
|   return getSource(startToken[_tokenize.FIELDS.START_LINE], startToken[_tokenize.FIELDS.START_COL], endToken[_tokenize.FIELDS.END_LINE], endToken[_tokenize.FIELDS.END_COL]);
 | |
| }
 | |
| function unescapeProp(node, prop) {
 | |
|   var value = node[prop];
 | |
|   if (typeof value !== "string") {
 | |
|     return;
 | |
|   }
 | |
|   if (value.indexOf("\\") !== -1) {
 | |
|     (0, _util.ensureObject)(node, 'raws');
 | |
|     node[prop] = (0, _util.unesc)(value);
 | |
|     if (node.raws[prop] === undefined) {
 | |
|       node.raws[prop] = value;
 | |
|     }
 | |
|   }
 | |
|   return node;
 | |
| }
 | |
| function indexesOf(array, item) {
 | |
|   var i = -1;
 | |
|   var indexes = [];
 | |
|   while ((i = array.indexOf(item, i + 1)) !== -1) {
 | |
|     indexes.push(i);
 | |
|   }
 | |
|   return indexes;
 | |
| }
 | |
| function uniqs() {
 | |
|   var list = Array.prototype.concat.apply([], arguments);
 | |
|   return list.filter(function (item, i) {
 | |
|     return i === list.indexOf(item);
 | |
|   });
 | |
| }
 | |
| var Parser = /*#__PURE__*/function () {
 | |
|   function Parser(rule, options) {
 | |
|     if (options === void 0) {
 | |
|       options = {};
 | |
|     }
 | |
|     this.rule = rule;
 | |
|     this.options = Object.assign({
 | |
|       lossy: false,
 | |
|       safe: false
 | |
|     }, options);
 | |
|     this.position = 0;
 | |
|     this.css = typeof this.rule === 'string' ? this.rule : this.rule.selector;
 | |
|     this.tokens = (0, _tokenize["default"])({
 | |
|       css: this.css,
 | |
|       error: this._errorGenerator(),
 | |
|       safe: this.options.safe
 | |
|     });
 | |
|     var rootSource = getTokenSourceSpan(this.tokens[0], this.tokens[this.tokens.length - 1]);
 | |
|     this.root = new _root["default"]({
 | |
|       source: rootSource
 | |
|     });
 | |
|     this.root.errorGenerator = this._errorGenerator();
 | |
|     var selector = new _selector["default"]({
 | |
|       source: {
 | |
|         start: {
 | |
|           line: 1,
 | |
|           column: 1
 | |
|         }
 | |
|       },
 | |
|       sourceIndex: 0
 | |
|     });
 | |
|     this.root.append(selector);
 | |
|     this.current = selector;
 | |
|     this.loop();
 | |
|   }
 | |
|   var _proto = Parser.prototype;
 | |
|   _proto._errorGenerator = function _errorGenerator() {
 | |
|     var _this = this;
 | |
|     return function (message, errorOptions) {
 | |
|       if (typeof _this.rule === 'string') {
 | |
|         return new Error(message);
 | |
|       }
 | |
|       return _this.rule.error(message, errorOptions);
 | |
|     };
 | |
|   };
 | |
|   _proto.attribute = function attribute() {
 | |
|     var attr = [];
 | |
|     var startingToken = this.currToken;
 | |
|     this.position++;
 | |
|     while (this.position < this.tokens.length && this.currToken[_tokenize.FIELDS.TYPE] !== tokens.closeSquare) {
 | |
|       attr.push(this.currToken);
 | |
|       this.position++;
 | |
|     }
 | |
|     if (this.currToken[_tokenize.FIELDS.TYPE] !== tokens.closeSquare) {
 | |
|       return this.expected('closing square bracket', this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|     }
 | |
|     var len = attr.length;
 | |
|     var node = {
 | |
|       source: getSource(startingToken[1], startingToken[2], this.currToken[3], this.currToken[4]),
 | |
|       sourceIndex: startingToken[_tokenize.FIELDS.START_POS]
 | |
|     };
 | |
|     if (len === 1 && !~[tokens.word].indexOf(attr[0][_tokenize.FIELDS.TYPE])) {
 | |
|       return this.expected('attribute', attr[0][_tokenize.FIELDS.START_POS]);
 | |
|     }
 | |
|     var pos = 0;
 | |
|     var spaceBefore = '';
 | |
|     var commentBefore = '';
 | |
|     var lastAdded = null;
 | |
|     var spaceAfterMeaningfulToken = false;
 | |
|     while (pos < len) {
 | |
|       var token = attr[pos];
 | |
|       var content = this.content(token);
 | |
|       var next = attr[pos + 1];
 | |
|       switch (token[_tokenize.FIELDS.TYPE]) {
 | |
|         case tokens.space:
 | |
|           // if (
 | |
|           //     len === 1 ||
 | |
|           //     pos === 0 && this.content(next) === '|'
 | |
|           // ) {
 | |
|           //     return this.expected('attribute', token[TOKEN.START_POS], content);
 | |
|           // }
 | |
|           spaceAfterMeaningfulToken = true;
 | |
|           if (this.options.lossy) {
 | |
|             break;
 | |
|           }
 | |
|           if (lastAdded) {
 | |
|             (0, _util.ensureObject)(node, 'spaces', lastAdded);
 | |
|             var prevContent = node.spaces[lastAdded].after || '';
 | |
|             node.spaces[lastAdded].after = prevContent + content;
 | |
|             var existingComment = (0, _util.getProp)(node, 'raws', 'spaces', lastAdded, 'after') || null;
 | |
|             if (existingComment) {
 | |
|               node.raws.spaces[lastAdded].after = existingComment + content;
 | |
|             }
 | |
|           } else {
 | |
|             spaceBefore = spaceBefore + content;
 | |
|             commentBefore = commentBefore + content;
 | |
|           }
 | |
|           break;
 | |
|         case tokens.asterisk:
 | |
|           if (next[_tokenize.FIELDS.TYPE] === tokens.equals) {
 | |
|             node.operator = content;
 | |
|             lastAdded = 'operator';
 | |
|           } else if ((!node.namespace || lastAdded === "namespace" && !spaceAfterMeaningfulToken) && next) {
 | |
|             if (spaceBefore) {
 | |
|               (0, _util.ensureObject)(node, 'spaces', 'attribute');
 | |
|               node.spaces.attribute.before = spaceBefore;
 | |
|               spaceBefore = '';
 | |
|             }
 | |
|             if (commentBefore) {
 | |
|               (0, _util.ensureObject)(node, 'raws', 'spaces', 'attribute');
 | |
|               node.raws.spaces.attribute.before = spaceBefore;
 | |
|               commentBefore = '';
 | |
|             }
 | |
|             node.namespace = (node.namespace || "") + content;
 | |
|             var rawValue = (0, _util.getProp)(node, 'raws', 'namespace') || null;
 | |
|             if (rawValue) {
 | |
|               node.raws.namespace += content;
 | |
|             }
 | |
|             lastAdded = 'namespace';
 | |
|           }
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.dollar:
 | |
|           if (lastAdded === "value") {
 | |
|             var oldRawValue = (0, _util.getProp)(node, 'raws', 'value');
 | |
|             node.value += "$";
 | |
|             if (oldRawValue) {
 | |
|               node.raws.value = oldRawValue + "$";
 | |
|             }
 | |
|             break;
 | |
|           }
 | |
|         // Falls through
 | |
|         case tokens.caret:
 | |
|           if (next[_tokenize.FIELDS.TYPE] === tokens.equals) {
 | |
|             node.operator = content;
 | |
|             lastAdded = 'operator';
 | |
|           }
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.combinator:
 | |
|           if (content === '~' && next[_tokenize.FIELDS.TYPE] === tokens.equals) {
 | |
|             node.operator = content;
 | |
|             lastAdded = 'operator';
 | |
|           }
 | |
|           if (content !== '|') {
 | |
|             spaceAfterMeaningfulToken = false;
 | |
|             break;
 | |
|           }
 | |
|           if (next[_tokenize.FIELDS.TYPE] === tokens.equals) {
 | |
|             node.operator = content;
 | |
|             lastAdded = 'operator';
 | |
|           } else if (!node.namespace && !node.attribute) {
 | |
|             node.namespace = true;
 | |
|           }
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.word:
 | |
|           if (next && this.content(next) === '|' && attr[pos + 2] && attr[pos + 2][_tokenize.FIELDS.TYPE] !== tokens.equals &&
 | |
|           // this look-ahead probably fails with comment nodes involved.
 | |
|           !node.operator && !node.namespace) {
 | |
|             node.namespace = content;
 | |
|             lastAdded = 'namespace';
 | |
|           } else if (!node.attribute || lastAdded === "attribute" && !spaceAfterMeaningfulToken) {
 | |
|             if (spaceBefore) {
 | |
|               (0, _util.ensureObject)(node, 'spaces', 'attribute');
 | |
|               node.spaces.attribute.before = spaceBefore;
 | |
|               spaceBefore = '';
 | |
|             }
 | |
|             if (commentBefore) {
 | |
|               (0, _util.ensureObject)(node, 'raws', 'spaces', 'attribute');
 | |
|               node.raws.spaces.attribute.before = commentBefore;
 | |
|               commentBefore = '';
 | |
|             }
 | |
|             node.attribute = (node.attribute || "") + content;
 | |
|             var _rawValue = (0, _util.getProp)(node, 'raws', 'attribute') || null;
 | |
|             if (_rawValue) {
 | |
|               node.raws.attribute += content;
 | |
|             }
 | |
|             lastAdded = 'attribute';
 | |
|           } else if (!node.value && node.value !== "" || lastAdded === "value" && !(spaceAfterMeaningfulToken || node.quoteMark)) {
 | |
|             var _unescaped = (0, _util.unesc)(content);
 | |
|             var _oldRawValue = (0, _util.getProp)(node, 'raws', 'value') || '';
 | |
|             var oldValue = node.value || '';
 | |
|             node.value = oldValue + _unescaped;
 | |
|             node.quoteMark = null;
 | |
|             if (_unescaped !== content || _oldRawValue) {
 | |
|               (0, _util.ensureObject)(node, 'raws');
 | |
|               node.raws.value = (_oldRawValue || oldValue) + content;
 | |
|             }
 | |
|             lastAdded = 'value';
 | |
|           } else {
 | |
|             var insensitive = content === 'i' || content === "I";
 | |
|             if ((node.value || node.value === '') && (node.quoteMark || spaceAfterMeaningfulToken)) {
 | |
|               node.insensitive = insensitive;
 | |
|               if (!insensitive || content === "I") {
 | |
|                 (0, _util.ensureObject)(node, 'raws');
 | |
|                 node.raws.insensitiveFlag = content;
 | |
|               }
 | |
|               lastAdded = 'insensitive';
 | |
|               if (spaceBefore) {
 | |
|                 (0, _util.ensureObject)(node, 'spaces', 'insensitive');
 | |
|                 node.spaces.insensitive.before = spaceBefore;
 | |
|                 spaceBefore = '';
 | |
|               }
 | |
|               if (commentBefore) {
 | |
|                 (0, _util.ensureObject)(node, 'raws', 'spaces', 'insensitive');
 | |
|                 node.raws.spaces.insensitive.before = commentBefore;
 | |
|                 commentBefore = '';
 | |
|               }
 | |
|             } else if (node.value || node.value === '') {
 | |
|               lastAdded = 'value';
 | |
|               node.value += content;
 | |
|               if (node.raws.value) {
 | |
|                 node.raws.value += content;
 | |
|               }
 | |
|             }
 | |
|           }
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.str:
 | |
|           if (!node.attribute || !node.operator) {
 | |
|             return this.error("Expected an attribute followed by an operator preceding the string.", {
 | |
|               index: token[_tokenize.FIELDS.START_POS]
 | |
|             });
 | |
|           }
 | |
|           var _unescapeValue = (0, _attribute.unescapeValue)(content),
 | |
|             unescaped = _unescapeValue.unescaped,
 | |
|             quoteMark = _unescapeValue.quoteMark;
 | |
|           node.value = unescaped;
 | |
|           node.quoteMark = quoteMark;
 | |
|           lastAdded = 'value';
 | |
|           (0, _util.ensureObject)(node, 'raws');
 | |
|           node.raws.value = content;
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.equals:
 | |
|           if (!node.attribute) {
 | |
|             return this.expected('attribute', token[_tokenize.FIELDS.START_POS], content);
 | |
|           }
 | |
|           if (node.value) {
 | |
|             return this.error('Unexpected "=" found; an operator was already defined.', {
 | |
|               index: token[_tokenize.FIELDS.START_POS]
 | |
|             });
 | |
|           }
 | |
|           node.operator = node.operator ? node.operator + content : content;
 | |
|           lastAdded = 'operator';
 | |
|           spaceAfterMeaningfulToken = false;
 | |
|           break;
 | |
|         case tokens.comment:
 | |
|           if (lastAdded) {
 | |
|             if (spaceAfterMeaningfulToken || next && next[_tokenize.FIELDS.TYPE] === tokens.space || lastAdded === 'insensitive') {
 | |
|               var lastComment = (0, _util.getProp)(node, 'spaces', lastAdded, 'after') || '';
 | |
|               var rawLastComment = (0, _util.getProp)(node, 'raws', 'spaces', lastAdded, 'after') || lastComment;
 | |
|               (0, _util.ensureObject)(node, 'raws', 'spaces', lastAdded);
 | |
|               node.raws.spaces[lastAdded].after = rawLastComment + content;
 | |
|             } else {
 | |
|               var lastValue = node[lastAdded] || '';
 | |
|               var rawLastValue = (0, _util.getProp)(node, 'raws', lastAdded) || lastValue;
 | |
|               (0, _util.ensureObject)(node, 'raws');
 | |
|               node.raws[lastAdded] = rawLastValue + content;
 | |
|             }
 | |
|           } else {
 | |
|             commentBefore = commentBefore + content;
 | |
|           }
 | |
|           break;
 | |
|         default:
 | |
|           return this.error("Unexpected \"" + content + "\" found.", {
 | |
|             index: token[_tokenize.FIELDS.START_POS]
 | |
|           });
 | |
|       }
 | |
|       pos++;
 | |
|     }
 | |
|     unescapeProp(node, "attribute");
 | |
|     unescapeProp(node, "namespace");
 | |
|     this.newNode(new _attribute["default"](node));
 | |
|     this.position++;
 | |
|   }
 | |
| 
 | |
|   /**
 | |
|    * return a node containing meaningless garbage up to (but not including) the specified token position.
 | |
|    * if the token position is negative, all remaining tokens are consumed.
 | |
|    *
 | |
|    * This returns an array containing a single string node if all whitespace,
 | |
|    * otherwise an array of comment nodes with space before and after.
 | |
|    *
 | |
|    * These tokens are not added to the current selector, the caller can add them or use them to amend
 | |
|    * a previous node's space metadata.
 | |
|    *
 | |
|    * In lossy mode, this returns only comments.
 | |
|    */;
 | |
|   _proto.parseWhitespaceEquivalentTokens = function parseWhitespaceEquivalentTokens(stopPosition) {
 | |
|     if (stopPosition < 0) {
 | |
|       stopPosition = this.tokens.length;
 | |
|     }
 | |
|     var startPosition = this.position;
 | |
|     var nodes = [];
 | |
|     var space = "";
 | |
|     var lastComment = undefined;
 | |
|     do {
 | |
|       if (WHITESPACE_TOKENS[this.currToken[_tokenize.FIELDS.TYPE]]) {
 | |
|         if (!this.options.lossy) {
 | |
|           space += this.content();
 | |
|         }
 | |
|       } else if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.comment) {
 | |
|         var spaces = {};
 | |
|         if (space) {
 | |
|           spaces.before = space;
 | |
|           space = "";
 | |
|         }
 | |
|         lastComment = new _comment["default"]({
 | |
|           value: this.content(),
 | |
|           source: getTokenSource(this.currToken),
 | |
|           sourceIndex: this.currToken[_tokenize.FIELDS.START_POS],
 | |
|           spaces: spaces
 | |
|         });
 | |
|         nodes.push(lastComment);
 | |
|       }
 | |
|     } while (++this.position < stopPosition);
 | |
|     if (space) {
 | |
|       if (lastComment) {
 | |
|         lastComment.spaces.after = space;
 | |
|       } else if (!this.options.lossy) {
 | |
|         var firstToken = this.tokens[startPosition];
 | |
|         var lastToken = this.tokens[this.position - 1];
 | |
|         nodes.push(new _string["default"]({
 | |
|           value: '',
 | |
|           source: getSource(firstToken[_tokenize.FIELDS.START_LINE], firstToken[_tokenize.FIELDS.START_COL], lastToken[_tokenize.FIELDS.END_LINE], lastToken[_tokenize.FIELDS.END_COL]),
 | |
|           sourceIndex: firstToken[_tokenize.FIELDS.START_POS],
 | |
|           spaces: {
 | |
|             before: space,
 | |
|             after: ''
 | |
|           }
 | |
|         }));
 | |
|       }
 | |
|     }
 | |
|     return nodes;
 | |
|   }
 | |
| 
 | |
|   /**
 | |
|    *
 | |
|    * @param {*} nodes
 | |
|    */;
 | |
|   _proto.convertWhitespaceNodesToSpace = function convertWhitespaceNodesToSpace(nodes, requiredSpace) {
 | |
|     var _this2 = this;
 | |
|     if (requiredSpace === void 0) {
 | |
|       requiredSpace = false;
 | |
|     }
 | |
|     var space = "";
 | |
|     var rawSpace = "";
 | |
|     nodes.forEach(function (n) {
 | |
|       var spaceBefore = _this2.lossySpace(n.spaces.before, requiredSpace);
 | |
|       var rawSpaceBefore = _this2.lossySpace(n.rawSpaceBefore, requiredSpace);
 | |
|       space += spaceBefore + _this2.lossySpace(n.spaces.after, requiredSpace && spaceBefore.length === 0);
 | |
|       rawSpace += spaceBefore + n.value + _this2.lossySpace(n.rawSpaceAfter, requiredSpace && rawSpaceBefore.length === 0);
 | |
|     });
 | |
|     if (rawSpace === space) {
 | |
|       rawSpace = undefined;
 | |
|     }
 | |
|     var result = {
 | |
|       space: space,
 | |
|       rawSpace: rawSpace
 | |
|     };
 | |
|     return result;
 | |
|   };
 | |
|   _proto.isNamedCombinator = function isNamedCombinator(position) {
 | |
|     if (position === void 0) {
 | |
|       position = this.position;
 | |
|     }
 | |
|     return this.tokens[position + 0] && this.tokens[position + 0][_tokenize.FIELDS.TYPE] === tokens.slash && this.tokens[position + 1] && this.tokens[position + 1][_tokenize.FIELDS.TYPE] === tokens.word && this.tokens[position + 2] && this.tokens[position + 2][_tokenize.FIELDS.TYPE] === tokens.slash;
 | |
|   };
 | |
|   _proto.namedCombinator = function namedCombinator() {
 | |
|     if (this.isNamedCombinator()) {
 | |
|       var nameRaw = this.content(this.tokens[this.position + 1]);
 | |
|       var name = (0, _util.unesc)(nameRaw).toLowerCase();
 | |
|       var raws = {};
 | |
|       if (name !== nameRaw) {
 | |
|         raws.value = "/" + nameRaw + "/";
 | |
|       }
 | |
|       var node = new _combinator["default"]({
 | |
|         value: "/" + name + "/",
 | |
|         source: getSource(this.currToken[_tokenize.FIELDS.START_LINE], this.currToken[_tokenize.FIELDS.START_COL], this.tokens[this.position + 2][_tokenize.FIELDS.END_LINE], this.tokens[this.position + 2][_tokenize.FIELDS.END_COL]),
 | |
|         sourceIndex: this.currToken[_tokenize.FIELDS.START_POS],
 | |
|         raws: raws
 | |
|       });
 | |
|       this.position = this.position + 3;
 | |
|       return node;
 | |
|     } else {
 | |
|       this.unexpected();
 | |
|     }
 | |
|   };
 | |
|   _proto.combinator = function combinator() {
 | |
|     var _this3 = this;
 | |
|     if (this.content() === '|') {
 | |
|       return this.namespace();
 | |
|     }
 | |
|     // We need to decide between a space that's a descendant combinator and meaningless whitespace at the end of a selector.
 | |
|     var nextSigTokenPos = this.locateNextMeaningfulToken(this.position);
 | |
|     if (nextSigTokenPos < 0 || this.tokens[nextSigTokenPos][_tokenize.FIELDS.TYPE] === tokens.comma || this.tokens[nextSigTokenPos][_tokenize.FIELDS.TYPE] === tokens.closeParenthesis) {
 | |
|       var nodes = this.parseWhitespaceEquivalentTokens(nextSigTokenPos);
 | |
|       if (nodes.length > 0) {
 | |
|         var last = this.current.last;
 | |
|         if (last) {
 | |
|           var _this$convertWhitespa = this.convertWhitespaceNodesToSpace(nodes),
 | |
|             space = _this$convertWhitespa.space,
 | |
|             rawSpace = _this$convertWhitespa.rawSpace;
 | |
|           if (rawSpace !== undefined) {
 | |
|             last.rawSpaceAfter += rawSpace;
 | |
|           }
 | |
|           last.spaces.after += space;
 | |
|         } else {
 | |
|           nodes.forEach(function (n) {
 | |
|             return _this3.newNode(n);
 | |
|           });
 | |
|         }
 | |
|       }
 | |
|       return;
 | |
|     }
 | |
|     var firstToken = this.currToken;
 | |
|     var spaceOrDescendantSelectorNodes = undefined;
 | |
|     if (nextSigTokenPos > this.position) {
 | |
|       spaceOrDescendantSelectorNodes = this.parseWhitespaceEquivalentTokens(nextSigTokenPos);
 | |
|     }
 | |
|     var node;
 | |
|     if (this.isNamedCombinator()) {
 | |
|       node = this.namedCombinator();
 | |
|     } else if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.combinator) {
 | |
|       node = new _combinator["default"]({
 | |
|         value: this.content(),
 | |
|         source: getTokenSource(this.currToken),
 | |
|         sourceIndex: this.currToken[_tokenize.FIELDS.START_POS]
 | |
|       });
 | |
|       this.position++;
 | |
|     } else if (WHITESPACE_TOKENS[this.currToken[_tokenize.FIELDS.TYPE]]) {
 | |
|       // pass
 | |
|     } else if (!spaceOrDescendantSelectorNodes) {
 | |
|       this.unexpected();
 | |
|     }
 | |
|     if (node) {
 | |
|       if (spaceOrDescendantSelectorNodes) {
 | |
|         var _this$convertWhitespa2 = this.convertWhitespaceNodesToSpace(spaceOrDescendantSelectorNodes),
 | |
|           _space = _this$convertWhitespa2.space,
 | |
|           _rawSpace = _this$convertWhitespa2.rawSpace;
 | |
|         node.spaces.before = _space;
 | |
|         node.rawSpaceBefore = _rawSpace;
 | |
|       }
 | |
|     } else {
 | |
|       // descendant combinator
 | |
|       var _this$convertWhitespa3 = this.convertWhitespaceNodesToSpace(spaceOrDescendantSelectorNodes, true),
 | |
|         _space2 = _this$convertWhitespa3.space,
 | |
|         _rawSpace2 = _this$convertWhitespa3.rawSpace;
 | |
|       if (!_rawSpace2) {
 | |
|         _rawSpace2 = _space2;
 | |
|       }
 | |
|       var spaces = {};
 | |
|       var raws = {
 | |
|         spaces: {}
 | |
|       };
 | |
|       if (_space2.endsWith(' ') && _rawSpace2.endsWith(' ')) {
 | |
|         spaces.before = _space2.slice(0, _space2.length - 1);
 | |
|         raws.spaces.before = _rawSpace2.slice(0, _rawSpace2.length - 1);
 | |
|       } else if (_space2.startsWith(' ') && _rawSpace2.startsWith(' ')) {
 | |
|         spaces.after = _space2.slice(1);
 | |
|         raws.spaces.after = _rawSpace2.slice(1);
 | |
|       } else {
 | |
|         raws.value = _rawSpace2;
 | |
|       }
 | |
|       node = new _combinator["default"]({
 | |
|         value: ' ',
 | |
|         source: getTokenSourceSpan(firstToken, this.tokens[this.position - 1]),
 | |
|         sourceIndex: firstToken[_tokenize.FIELDS.START_POS],
 | |
|         spaces: spaces,
 | |
|         raws: raws
 | |
|       });
 | |
|     }
 | |
|     if (this.currToken && this.currToken[_tokenize.FIELDS.TYPE] === tokens.space) {
 | |
|       node.spaces.after = this.optionalSpace(this.content());
 | |
|       this.position++;
 | |
|     }
 | |
|     return this.newNode(node);
 | |
|   };
 | |
|   _proto.comma = function comma() {
 | |
|     if (this.position === this.tokens.length - 1) {
 | |
|       this.root.trailingComma = true;
 | |
|       this.position++;
 | |
|       return;
 | |
|     }
 | |
|     this.current._inferEndPosition();
 | |
|     var selector = new _selector["default"]({
 | |
|       source: {
 | |
|         start: tokenStart(this.tokens[this.position + 1])
 | |
|       },
 | |
|       sourceIndex: this.tokens[this.position + 1][_tokenize.FIELDS.START_POS]
 | |
|     });
 | |
|     this.current.parent.append(selector);
 | |
|     this.current = selector;
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.comment = function comment() {
 | |
|     var current = this.currToken;
 | |
|     this.newNode(new _comment["default"]({
 | |
|       value: this.content(),
 | |
|       source: getTokenSource(current),
 | |
|       sourceIndex: current[_tokenize.FIELDS.START_POS]
 | |
|     }));
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.error = function error(message, opts) {
 | |
|     throw this.root.error(message, opts);
 | |
|   };
 | |
|   _proto.missingBackslash = function missingBackslash() {
 | |
|     return this.error('Expected a backslash preceding the semicolon.', {
 | |
|       index: this.currToken[_tokenize.FIELDS.START_POS]
 | |
|     });
 | |
|   };
 | |
|   _proto.missingParenthesis = function missingParenthesis() {
 | |
|     return this.expected('opening parenthesis', this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|   };
 | |
|   _proto.missingSquareBracket = function missingSquareBracket() {
 | |
|     return this.expected('opening square bracket', this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|   };
 | |
|   _proto.unexpected = function unexpected() {
 | |
|     return this.error("Unexpected '" + this.content() + "'. Escaping special characters with \\ may help.", this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|   };
 | |
|   _proto.unexpectedPipe = function unexpectedPipe() {
 | |
|     return this.error("Unexpected '|'.", this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|   };
 | |
|   _proto.namespace = function namespace() {
 | |
|     var before = this.prevToken && this.content(this.prevToken) || true;
 | |
|     if (this.nextToken[_tokenize.FIELDS.TYPE] === tokens.word) {
 | |
|       this.position++;
 | |
|       return this.word(before);
 | |
|     } else if (this.nextToken[_tokenize.FIELDS.TYPE] === tokens.asterisk) {
 | |
|       this.position++;
 | |
|       return this.universal(before);
 | |
|     }
 | |
|     this.unexpectedPipe();
 | |
|   };
 | |
|   _proto.nesting = function nesting() {
 | |
|     if (this.nextToken) {
 | |
|       var nextContent = this.content(this.nextToken);
 | |
|       if (nextContent === "|") {
 | |
|         this.position++;
 | |
|         return;
 | |
|       }
 | |
|     }
 | |
|     var current = this.currToken;
 | |
|     this.newNode(new _nesting["default"]({
 | |
|       value: this.content(),
 | |
|       source: getTokenSource(current),
 | |
|       sourceIndex: current[_tokenize.FIELDS.START_POS]
 | |
|     }));
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.parentheses = function parentheses() {
 | |
|     var last = this.current.last;
 | |
|     var unbalanced = 1;
 | |
|     this.position++;
 | |
|     if (last && last.type === types.PSEUDO) {
 | |
|       var selector = new _selector["default"]({
 | |
|         source: {
 | |
|           start: tokenStart(this.tokens[this.position])
 | |
|         },
 | |
|         sourceIndex: this.tokens[this.position][_tokenize.FIELDS.START_POS]
 | |
|       });
 | |
|       var cache = this.current;
 | |
|       last.append(selector);
 | |
|       this.current = selector;
 | |
|       while (this.position < this.tokens.length && unbalanced) {
 | |
|         if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.openParenthesis) {
 | |
|           unbalanced++;
 | |
|         }
 | |
|         if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.closeParenthesis) {
 | |
|           unbalanced--;
 | |
|         }
 | |
|         if (unbalanced) {
 | |
|           this.parse();
 | |
|         } else {
 | |
|           this.current.source.end = tokenEnd(this.currToken);
 | |
|           this.current.parent.source.end = tokenEnd(this.currToken);
 | |
|           this.position++;
 | |
|         }
 | |
|       }
 | |
|       this.current = cache;
 | |
|     } else {
 | |
|       // I think this case should be an error. It's used to implement a basic parse of media queries
 | |
|       // but I don't think it's a good idea.
 | |
|       var parenStart = this.currToken;
 | |
|       var parenValue = "(";
 | |
|       var parenEnd;
 | |
|       while (this.position < this.tokens.length && unbalanced) {
 | |
|         if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.openParenthesis) {
 | |
|           unbalanced++;
 | |
|         }
 | |
|         if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.closeParenthesis) {
 | |
|           unbalanced--;
 | |
|         }
 | |
|         parenEnd = this.currToken;
 | |
|         parenValue += this.parseParenthesisToken(this.currToken);
 | |
|         this.position++;
 | |
|       }
 | |
|       if (last) {
 | |
|         last.appendToPropertyAndEscape("value", parenValue, parenValue);
 | |
|       } else {
 | |
|         this.newNode(new _string["default"]({
 | |
|           value: parenValue,
 | |
|           source: getSource(parenStart[_tokenize.FIELDS.START_LINE], parenStart[_tokenize.FIELDS.START_COL], parenEnd[_tokenize.FIELDS.END_LINE], parenEnd[_tokenize.FIELDS.END_COL]),
 | |
|           sourceIndex: parenStart[_tokenize.FIELDS.START_POS]
 | |
|         }));
 | |
|       }
 | |
|     }
 | |
|     if (unbalanced) {
 | |
|       return this.expected('closing parenthesis', this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|     }
 | |
|   };
 | |
|   _proto.pseudo = function pseudo() {
 | |
|     var _this4 = this;
 | |
|     var pseudoStr = '';
 | |
|     var startingToken = this.currToken;
 | |
|     while (this.currToken && this.currToken[_tokenize.FIELDS.TYPE] === tokens.colon) {
 | |
|       pseudoStr += this.content();
 | |
|       this.position++;
 | |
|     }
 | |
|     if (!this.currToken) {
 | |
|       return this.expected(['pseudo-class', 'pseudo-element'], this.position - 1);
 | |
|     }
 | |
|     if (this.currToken[_tokenize.FIELDS.TYPE] === tokens.word) {
 | |
|       this.splitWord(false, function (first, length) {
 | |
|         pseudoStr += first;
 | |
|         _this4.newNode(new _pseudo["default"]({
 | |
|           value: pseudoStr,
 | |
|           source: getTokenSourceSpan(startingToken, _this4.currToken),
 | |
|           sourceIndex: startingToken[_tokenize.FIELDS.START_POS]
 | |
|         }));
 | |
|         if (length > 1 && _this4.nextToken && _this4.nextToken[_tokenize.FIELDS.TYPE] === tokens.openParenthesis) {
 | |
|           _this4.error('Misplaced parenthesis.', {
 | |
|             index: _this4.nextToken[_tokenize.FIELDS.START_POS]
 | |
|           });
 | |
|         }
 | |
|       });
 | |
|     } else {
 | |
|       return this.expected(['pseudo-class', 'pseudo-element'], this.currToken[_tokenize.FIELDS.START_POS]);
 | |
|     }
 | |
|   };
 | |
|   _proto.space = function space() {
 | |
|     var content = this.content();
 | |
|     // Handle space before and after the selector
 | |
|     if (this.position === 0 || this.prevToken[_tokenize.FIELDS.TYPE] === tokens.comma || this.prevToken[_tokenize.FIELDS.TYPE] === tokens.openParenthesis || this.current.nodes.every(function (node) {
 | |
|       return node.type === 'comment';
 | |
|     })) {
 | |
|       this.spaces = this.optionalSpace(content);
 | |
|       this.position++;
 | |
|     } else if (this.position === this.tokens.length - 1 || this.nextToken[_tokenize.FIELDS.TYPE] === tokens.comma || this.nextToken[_tokenize.FIELDS.TYPE] === tokens.closeParenthesis) {
 | |
|       this.current.last.spaces.after = this.optionalSpace(content);
 | |
|       this.position++;
 | |
|     } else {
 | |
|       this.combinator();
 | |
|     }
 | |
|   };
 | |
|   _proto.string = function string() {
 | |
|     var current = this.currToken;
 | |
|     this.newNode(new _string["default"]({
 | |
|       value: this.content(),
 | |
|       source: getTokenSource(current),
 | |
|       sourceIndex: current[_tokenize.FIELDS.START_POS]
 | |
|     }));
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.universal = function universal(namespace) {
 | |
|     var nextToken = this.nextToken;
 | |
|     if (nextToken && this.content(nextToken) === '|') {
 | |
|       this.position++;
 | |
|       return this.namespace();
 | |
|     }
 | |
|     var current = this.currToken;
 | |
|     this.newNode(new _universal["default"]({
 | |
|       value: this.content(),
 | |
|       source: getTokenSource(current),
 | |
|       sourceIndex: current[_tokenize.FIELDS.START_POS]
 | |
|     }), namespace);
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.splitWord = function splitWord(namespace, firstCallback) {
 | |
|     var _this5 = this;
 | |
|     var nextToken = this.nextToken;
 | |
|     var word = this.content();
 | |
|     while (nextToken && ~[tokens.dollar, tokens.caret, tokens.equals, tokens.word].indexOf(nextToken[_tokenize.FIELDS.TYPE])) {
 | |
|       this.position++;
 | |
|       var current = this.content();
 | |
|       word += current;
 | |
|       if (current.lastIndexOf('\\') === current.length - 1) {
 | |
|         var next = this.nextToken;
 | |
|         if (next && next[_tokenize.FIELDS.TYPE] === tokens.space) {
 | |
|           word += this.requiredSpace(this.content(next));
 | |
|           this.position++;
 | |
|         }
 | |
|       }
 | |
|       nextToken = this.nextToken;
 | |
|     }
 | |
|     var hasClass = indexesOf(word, '.').filter(function (i) {
 | |
|       // Allow escaped dot within class name
 | |
|       var escapedDot = word[i - 1] === '\\';
 | |
|       // Allow decimal numbers percent in @keyframes
 | |
|       var isKeyframesPercent = /^\d+\.\d+%$/.test(word);
 | |
|       return !escapedDot && !isKeyframesPercent;
 | |
|     });
 | |
|     var hasId = indexesOf(word, '#').filter(function (i) {
 | |
|       return word[i - 1] !== '\\';
 | |
|     });
 | |
|     // Eliminate Sass interpolations from the list of id indexes
 | |
|     var interpolations = indexesOf(word, '#{');
 | |
|     if (interpolations.length) {
 | |
|       hasId = hasId.filter(function (hashIndex) {
 | |
|         return !~interpolations.indexOf(hashIndex);
 | |
|       });
 | |
|     }
 | |
|     var indices = (0, _sortAscending["default"])(uniqs([0].concat(hasClass, hasId)));
 | |
|     indices.forEach(function (ind, i) {
 | |
|       var index = indices[i + 1] || word.length;
 | |
|       var value = word.slice(ind, index);
 | |
|       if (i === 0 && firstCallback) {
 | |
|         return firstCallback.call(_this5, value, indices.length);
 | |
|       }
 | |
|       var node;
 | |
|       var current = _this5.currToken;
 | |
|       var sourceIndex = current[_tokenize.FIELDS.START_POS] + indices[i];
 | |
|       var source = getSource(current[1], current[2] + ind, current[3], current[2] + (index - 1));
 | |
|       if (~hasClass.indexOf(ind)) {
 | |
|         var classNameOpts = {
 | |
|           value: value.slice(1),
 | |
|           source: source,
 | |
|           sourceIndex: sourceIndex
 | |
|         };
 | |
|         node = new _className["default"](unescapeProp(classNameOpts, "value"));
 | |
|       } else if (~hasId.indexOf(ind)) {
 | |
|         var idOpts = {
 | |
|           value: value.slice(1),
 | |
|           source: source,
 | |
|           sourceIndex: sourceIndex
 | |
|         };
 | |
|         node = new _id["default"](unescapeProp(idOpts, "value"));
 | |
|       } else {
 | |
|         var tagOpts = {
 | |
|           value: value,
 | |
|           source: source,
 | |
|           sourceIndex: sourceIndex
 | |
|         };
 | |
|         unescapeProp(tagOpts, "value");
 | |
|         node = new _tag["default"](tagOpts);
 | |
|       }
 | |
|       _this5.newNode(node, namespace);
 | |
|       // Ensure that the namespace is used only once
 | |
|       namespace = null;
 | |
|     });
 | |
|     this.position++;
 | |
|   };
 | |
|   _proto.word = function word(namespace) {
 | |
|     var nextToken = this.nextToken;
 | |
|     if (nextToken && this.content(nextToken) === '|') {
 | |
|       this.position++;
 | |
|       return this.namespace();
 | |
|     }
 | |
|     return this.splitWord(namespace);
 | |
|   };
 | |
|   _proto.loop = function loop() {
 | |
|     while (this.position < this.tokens.length) {
 | |
|       this.parse(true);
 | |
|     }
 | |
|     this.current._inferEndPosition();
 | |
|     return this.root;
 | |
|   };
 | |
|   _proto.parse = function parse(throwOnParenthesis) {
 | |
|     switch (this.currToken[_tokenize.FIELDS.TYPE]) {
 | |
|       case tokens.space:
 | |
|         this.space();
 | |
|         break;
 | |
|       case tokens.comment:
 | |
|         this.comment();
 | |
|         break;
 | |
|       case tokens.openParenthesis:
 | |
|         this.parentheses();
 | |
|         break;
 | |
|       case tokens.closeParenthesis:
 | |
|         if (throwOnParenthesis) {
 | |
|           this.missingParenthesis();
 | |
|         }
 | |
|         break;
 | |
|       case tokens.openSquare:
 | |
|         this.attribute();
 | |
|         break;
 | |
|       case tokens.dollar:
 | |
|       case tokens.caret:
 | |
|       case tokens.equals:
 | |
|       case tokens.word:
 | |
|         this.word();
 | |
|         break;
 | |
|       case tokens.colon:
 | |
|         this.pseudo();
 | |
|         break;
 | |
|       case tokens.comma:
 | |
|         this.comma();
 | |
|         break;
 | |
|       case tokens.asterisk:
 | |
|         this.universal();
 | |
|         break;
 | |
|       case tokens.ampersand:
 | |
|         this.nesting();
 | |
|         break;
 | |
|       case tokens.slash:
 | |
|       case tokens.combinator:
 | |
|         this.combinator();
 | |
|         break;
 | |
|       case tokens.str:
 | |
|         this.string();
 | |
|         break;
 | |
|       // These cases throw; no break needed.
 | |
|       case tokens.closeSquare:
 | |
|         this.missingSquareBracket();
 | |
|       case tokens.semicolon:
 | |
|         this.missingBackslash();
 | |
|       default:
 | |
|         this.unexpected();
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   /**
 | |
|    * Helpers
 | |
|    */;
 | |
|   _proto.expected = function expected(description, index, found) {
 | |
|     if (Array.isArray(description)) {
 | |
|       var last = description.pop();
 | |
|       description = description.join(', ') + " or " + last;
 | |
|     }
 | |
|     var an = /^[aeiou]/.test(description[0]) ? 'an' : 'a';
 | |
|     if (!found) {
 | |
|       return this.error("Expected " + an + " " + description + ".", {
 | |
|         index: index
 | |
|       });
 | |
|     }
 | |
|     return this.error("Expected " + an + " " + description + ", found \"" + found + "\" instead.", {
 | |
|       index: index
 | |
|     });
 | |
|   };
 | |
|   _proto.requiredSpace = function requiredSpace(space) {
 | |
|     return this.options.lossy ? ' ' : space;
 | |
|   };
 | |
|   _proto.optionalSpace = function optionalSpace(space) {
 | |
|     return this.options.lossy ? '' : space;
 | |
|   };
 | |
|   _proto.lossySpace = function lossySpace(space, required) {
 | |
|     if (this.options.lossy) {
 | |
|       return required ? ' ' : '';
 | |
|     } else {
 | |
|       return space;
 | |
|     }
 | |
|   };
 | |
|   _proto.parseParenthesisToken = function parseParenthesisToken(token) {
 | |
|     var content = this.content(token);
 | |
|     if (token[_tokenize.FIELDS.TYPE] === tokens.space) {
 | |
|       return this.requiredSpace(content);
 | |
|     } else {
 | |
|       return content;
 | |
|     }
 | |
|   };
 | |
|   _proto.newNode = function newNode(node, namespace) {
 | |
|     if (namespace) {
 | |
|       if (/^ +$/.test(namespace)) {
 | |
|         if (!this.options.lossy) {
 | |
|           this.spaces = (this.spaces || '') + namespace;
 | |
|         }
 | |
|         namespace = true;
 | |
|       }
 | |
|       node.namespace = namespace;
 | |
|       unescapeProp(node, "namespace");
 | |
|     }
 | |
|     if (this.spaces) {
 | |
|       node.spaces.before = this.spaces;
 | |
|       this.spaces = '';
 | |
|     }
 | |
|     return this.current.append(node);
 | |
|   };
 | |
|   _proto.content = function content(token) {
 | |
|     if (token === void 0) {
 | |
|       token = this.currToken;
 | |
|     }
 | |
|     return this.css.slice(token[_tokenize.FIELDS.START_POS], token[_tokenize.FIELDS.END_POS]);
 | |
|   };
 | |
|   /**
 | |
|    * returns the index of the next non-whitespace, non-comment token.
 | |
|    * returns -1 if no meaningful token is found.
 | |
|    */
 | |
|   _proto.locateNextMeaningfulToken = function locateNextMeaningfulToken(startPosition) {
 | |
|     if (startPosition === void 0) {
 | |
|       startPosition = this.position + 1;
 | |
|     }
 | |
|     var searchPosition = startPosition;
 | |
|     while (searchPosition < this.tokens.length) {
 | |
|       if (WHITESPACE_EQUIV_TOKENS[this.tokens[searchPosition][_tokenize.FIELDS.TYPE]]) {
 | |
|         searchPosition++;
 | |
|         continue;
 | |
|       } else {
 | |
|         return searchPosition;
 | |
|       }
 | |
|     }
 | |
|     return -1;
 | |
|   };
 | |
|   _createClass(Parser, [{
 | |
|     key: "currToken",
 | |
|     get: function get() {
 | |
|       return this.tokens[this.position];
 | |
|     }
 | |
|   }, {
 | |
|     key: "nextToken",
 | |
|     get: function get() {
 | |
|       return this.tokens[this.position + 1];
 | |
|     }
 | |
|   }, {
 | |
|     key: "prevToken",
 | |
|     get: function get() {
 | |
|       return this.tokens[this.position - 1];
 | |
|     }
 | |
|   }]);
 | |
|   return Parser;
 | |
| }();
 | |
| exports["default"] = Parser;
 | |
| module.exports = exports.default; |