40 lines
1.3 KiB
JavaScript
40 lines
1.3 KiB
JavaScript
|
"use strict";
|
||
|
var __importDefault = (this && this.__importDefault) || function (mod) {
|
||
|
return (mod && mod.__esModule) ? mod : { "default": mod };
|
||
|
};
|
||
|
Object.defineProperty(exports, "__esModule", { value: true });
|
||
|
const tokenize_1 = __importDefault(require("postcss/lib/tokenize"));
|
||
|
function templateTokenize(...args) {
|
||
|
const tokenizer = (0, tokenize_1.default)(...args);
|
||
|
function nextToken(...args) {
|
||
|
const returned = [];
|
||
|
let token, lastPos;
|
||
|
let depth = 0;
|
||
|
while ((token = tokenizer.nextToken(...args))) {
|
||
|
if (token[0] !== 'word') {
|
||
|
if (token[0] === '{') {
|
||
|
++depth;
|
||
|
}
|
||
|
else if (token[0] === '}') {
|
||
|
--depth;
|
||
|
}
|
||
|
}
|
||
|
if (depth || returned.length) {
|
||
|
lastPos = token[3] || token[2] || lastPos;
|
||
|
returned.push(token);
|
||
|
}
|
||
|
if (!depth) {
|
||
|
break;
|
||
|
}
|
||
|
}
|
||
|
if (returned.length) {
|
||
|
token = ['word', returned.map((token) => token[1]).join(''), returned[0][2], lastPos];
|
||
|
}
|
||
|
return token;
|
||
|
}
|
||
|
return Object.assign({}, tokenizer, {
|
||
|
nextToken
|
||
|
});
|
||
|
}
|
||
|
exports.default = templateTokenize;
|