From: Tan Kian-ting Date: Tue, 10 Oct 2023 14:15:02 +0000 (+0800) Subject: basic parser built X-Git-Url: https://git.kianting.info/?a=commitdiff_plain;h=df1419e45c41c5681b9cdad914c360ff16052fc5;p=clo basic parser built --- diff --git a/README.md b/README.md index 69d3c81..a11749a 100644 --- a/README.md +++ b/README.md @@ -24,4 +24,5 @@ License: MIT - 20230928:basically fix `issue1`。其他ê物件猶著做。 - 20230929:add multi args parsing for `callee`. - 20230930:tîng khí parser, using `js-token`. - - 20231016: tîng siá parser, using `ts-parsec`. \ No newline at end of file + - 20231006: tîng siá parser, using `ts-parsec`. + - 20231010: 初步完成tsit ê階段ê Parser`。 \ No newline at end of file diff --git a/src/index.js b/src/index.js index 8fc8c39..6b49d56 100644 --- a/src/index.js +++ b/src/index.js @@ -62,43 +62,65 @@ var TokenKind; TokenKind[TokenKind["Number"] = 2] = "Number"; TokenKind[TokenKind["Op"] = 3] = "Op"; TokenKind[TokenKind["ExprMark"] = 4] = "ExprMark"; - TokenKind[TokenKind["Paren"] = 5] = "Paren"; - TokenKind[TokenKind["SpaceNL"] = 6] = "SpaceNL"; - TokenKind[TokenKind["Id"] = 7] = "Id"; - TokenKind[TokenKind["Str"] = 8] = "Str"; + TokenKind[TokenKind["ExcapeAt"] = 5] = "ExcapeAt"; + TokenKind[TokenKind["Paren"] = 6] = "Paren"; + TokenKind[TokenKind["SpaceNL"] = 7] = "SpaceNL"; + TokenKind[TokenKind["Id"] = 8] = "Id"; + TokenKind[TokenKind["Str"] = 9] = "Str"; + TokenKind[TokenKind["Comment"] = 10] = "Comment"; })(TokenKind || (TokenKind = {})); /** * Parsing */ const lexer = p.buildLexer([ [true, /^\d+(\.\d+)?/g, TokenKind.Number], + [true, /^\\\@/g, TokenKind.ExcapeAt], + [true, /^\/\*([^/]|\/[^*])*\*\//g, TokenKind.Comment], [true, /^\;/g, TokenKind.Semicolon], [true, /^[-][-][-]/g, TokenKind.Seperator], [true, /^[\+\-\*\/\&\|\!\^\<\>\~\=\?]+/g, TokenKind.Op], - [true, /^\@+/g, TokenKind.ExprMark], + [true, /^\@/g, TokenKind.ExprMark], [true, /^[()\[\]{}]/g, TokenKind.Paren], [true, /^["]([\"]|[\\].)*["]/g, TokenKind.Str], [true, /^[']([\']|[\\].)*[']/g, TokenKind.Str], [true, /^[()\[\]{}]/g, TokenKind.Paren], - [true, /^[^\s\n\t\r;]+/g, TokenKind.Id], - [false, /^(\s|\n|\r|\t)+/g, TokenKind.SpaceNL] + [true, /^[^\/\\\@\s\n\t\r;]+/g, TokenKind.Id], + [true, /^(\s|\n|\r|\t)+/g, TokenKind.SpaceNL], ]); /** * * # TEST */ -const inputTxt = `import ast; +const inputTxt = `import a as b; /*bacourt*/ +/* ba choir +ipsum lorem*/ + +import you as john; --- -122`; + +臺中市\\\@ + +政府 +@2+2==4; + +console.log("122");@ + +人民 +`; const PROG = p.rule(); -const UNIT = p.rule(); +const SEGMENT = p.rule(); +const IMPORT = p.rule(); const IMPORTS = p.rule(); const SEMICOLON = p.rule(); +const EXCAPE_AT = p.rule(); +const NOT_AT_TEXT = p.rule(); +const CONTENT = p.rule(); let doubleMinus = { type: 'Punctuator', value: '--' }; let doubleMinus2 = p.str('--'); const TERM = p.rule(); -function applyUnit(value) { - return value.text; +function applySegment(input) { + let unpackedInnerExprs = input[1].map((x) => { return x.text; }); + return ["%exprs", unpackedInnerExprs]; } function applySemiColon(value) { return value.text; @@ -106,14 +128,80 @@ function applySemiColon(value) { function applyParts(first, second) { return ["%clo", first, second[1]]; } -PROG.setPattern(p.lrec_sc(IMPORTS, p.seq(p.str('---'), UNIT), applyParts)); -function applyImports(input) { +function applyComment(value) { + return [value.text]; +} +function applyImport(input) { let importTail = input[1].map(x => x.text); return ["import"].concat(importTail); } ; -IMPORTS.setPattern(p.apply(p.seq(p.str('import'), p.rep_sc(p.tok(TokenKind.Id)), SEMICOLON), applyImports)); +/* +function applyImportComment(input: [Token,Token[], + tkTree, Token]) : tkTree{ + let importTail = input[1].map(x=>x.text); + let comment = [input[3].text]; + return ["import"].concat(importTail).concat(comment); +};*/ +function applyImports(input) { + let resultBody = [input[0]].concat(input[1]); + let resultWrapper = ["%import", resultBody]; + return resultWrapper; +} +; +function applyNotAtText(value) { + if (value.text == "\\\@") { + return '@'; + } + else { + return value.text; + } +} +; +function applyText(input) { + return ["%text", input]; +} +; +function applyContent(input) { + return ["%content", input]; +} +; +function applySpaceNL(value) { + return value.text; +} +/** + * IMPORTEE: Number, Op, Paren, Id, Str, Comment, + */ +let IMPORTEE = p.alt(p.tok(TokenKind.Number), p.tok(TokenKind.Op), p.tok(TokenKind.Paren), p.tok(TokenKind.Id), p.tok(TokenKind.Str), p.tok(TokenKind.SpaceNL), p.tok(TokenKind.Comment)); +let NOT_AT = p.alt(p.tok(TokenKind.Seperator), p.tok(TokenKind.Semicolon), p.tok(TokenKind.Number), p.tok(TokenKind.ExcapeAt), p.tok(TokenKind.Op), p.tok(TokenKind.Paren), p.tok(TokenKind.SpaceNL), p.tok(TokenKind.Id), p.tok(TokenKind.Str), p.tok(TokenKind.Comment)); +/** + * PROG : IMPORTS '---' CONTENT; + */ +PROG.setPattern(p.lrec_sc(IMPORTS, p.seq(p.str('---'), CONTENT), applyParts)); +/** + * NOT_AT_TEXT : NOT_AT + */ +NOT_AT_TEXT.setPattern(p.apply(NOT_AT, applyNotAtText)); +IMPORTS.setPattern(p.apply(p.seq(IMPORT, p.rep(IMPORT)), applyImports)); +/** + * IMPORT : + * 'import' IMPORTEE* SEMICOLON | + * COMMENT | + */ +IMPORT.setPattern(p.alt(p.apply(p.seq(p.str('import'), p.rep_sc(IMPORTEE), SEMICOLON), applyImport), p.apply(p.tok(TokenKind.Comment), applyComment), p.apply(p.tok(TokenKind.SpaceNL), applySpaceNL))); +/** + * SEMICOLON : ';'; + */ SEMICOLON.setPattern(p.apply(p.tok(TokenKind.Semicolon), applySemiColon)); -UNIT.setPattern(p.apply(p.tok(TokenKind.Number), applyUnit)); +/** + * SEGMENT : '@' NOT_AT* '@' | + * (NOT_AT_TEXT | EXCAPE_AT)* + */ +SEGMENT.setPattern(p.alt(p.apply(p.rep_sc(NOT_AT_TEXT), applyText), p.apply(p.seq(p.str('@'), p.rep(NOT_AT), p.str('@')), applySegment))); +/** + * CONTENT : SEGMENT* + */ +CONTENT.setPattern(p.apply(p.rep(SEGMENT), applyContent)); +console.log(repr(lexer.parse(inputTxt))); let tree = p.expectSingleResult(p.expectEOF(PROG.parse(lexer.parse(inputTxt)))); console.log("RESULT=" + tkTreeToSExp(tree)); diff --git a/src/index.ts b/src/index.ts index 157b160..f5f55af 100644 --- a/src/index.ts +++ b/src/index.ts @@ -3,6 +3,7 @@ import jsTokens from "js-tokens"; import * as util from 'util'; import * as p from 'typescript-parsec'; import { Token } from 'typescript-parsec'; +import { TokenType } from "./tokenize"; /** * * # REPRESENTATION @@ -40,15 +41,17 @@ let repr = (x : any)=>{return util.inspect(x, {depth: null})}; type tkTree = string | tkTree[]; enum TokenKind { - Seperator, - Semicolon, + Seperator, // --- + Semicolon, // ; Number, Op, - ExprMark, + ExprMark, // @ + ExcapeAt, // \@ Paren, - SpaceNL, + SpaceNL, // \s\t\n\r Id, Str, + Comment, // /* ooo */ } /** @@ -56,16 +59,19 @@ enum TokenKind { */ const lexer = p.buildLexer([ [true, /^\d+(\.\d+)?/g, TokenKind.Number], + [true, /^\\\@/g, TokenKind.ExcapeAt], + [true, /^\/\*([^/]|\/[^*])*\*\//g, TokenKind.Comment], [true, /^\;/g, TokenKind.Semicolon], [true, /^[-][-][-]/g, TokenKind.Seperator], [true, /^[\+\-\*\/\&\|\!\^\<\>\~\=\?]+/g, TokenKind.Op], - [true, /^\@+/g, TokenKind.ExprMark], + [true, /^\@/g, TokenKind.ExprMark], [true, /^[()\[\]{}]/g, TokenKind.Paren], [true, /^["]([\"]|[\\].)*["]/g, TokenKind.Str], [true, /^[']([\']|[\\].)*[']/g, TokenKind.Str], [true, /^[()\[\]{}]/g, TokenKind.Paren], - [true, /^[^\s\n\t\r;]+/g, TokenKind.Id], - [false, /^(\s|\n|\r|\t)+/g, TokenKind.SpaceNL] + [true, /^[^\/\\\@\s\n\t\r;]+/g, TokenKind.Id], + [true, /^(\s|\n|\r|\t)+/g, TokenKind.SpaceNL], + ]); /** @@ -73,23 +79,41 @@ const lexer = p.buildLexer([ * # TEST */ const inputTxt= -`import ast; +`import a as b; /*bacourt*/ +/* ba choir +ipsum lorem*/ + +import you as john; --- -122`; + +臺中市\\\@ + +政府 +@2+2==4; + +console.log("122");@ + +人民 +`; const PROG = p.rule(); -const UNIT = p.rule(); +const SEGMENT = p.rule(); +const IMPORT = p.rule(); const IMPORTS = p.rule(); const SEMICOLON = p.rule(); - +const EXCAPE_AT = p.rule(); +const NOT_AT_TEXT = p.rule(); +const CONTENT = p.rule(); let doubleMinus = { type: 'Punctuator', value: '--' }; let doubleMinus2 = p.str('--'); const TERM = p.rule(); -function applyUnit(value: Token): tkTree{ - return value.text; +function applySegment(input: [Token, Token[], + Token]): tkTree[]{ + let unpackedInnerExprs = input[1].map((x)=>{return x.text}); + return ["%exprs", unpackedInnerExprs]; } function applySemiColon(value: Token): tkTree{ @@ -102,26 +126,108 @@ function applyParts(first: tkTree, } +function applyComment(value: Token): tkTree[]{ + return [value.text]; +} -function applyImports(input: [Token,Token[], tkTree]):tkTree{ +function applyImport(input: [Token,Token[], tkTree]) : tkTree{ let importTail = input[1].map(x=>x.text); return ["import"].concat(importTail); }; + +/* +function applyImportComment(input: [Token,Token[], + tkTree, Token]) : tkTree{ + let importTail = input[1].map(x=>x.text); + let comment = [input[3].text]; + return ["import"].concat(importTail).concat(comment); +};*/ + +function applyImports(input : [tkTree, tkTree[]]): tkTree{ + let resultBody = [input[0]].concat(input[1]); + let resultWrapper = ["%import", resultBody]; + return resultWrapper; +}; + + + + +function applyNotAtText(value : Token): tkTree{ + if (value.text == "\\\@"){ + return '@'; + } + else{return value.text;} +}; + +function applyText (input : tkTree): tkTree[]{ + return ["%text", input]; +}; + +function applyContent(input : tkTree[]): tkTree[]{ + return ["%content", input]; +}; + +function applySpaceNL(value : Token): tkTree{ + return value.text; +} + /** - * PROG : IMPORTS '---' UNIT; + * IMPORTEE: Number, Op, Paren, Id, Str, Comment, + */ +let IMPORTEE = p.alt(p.tok(TokenKind.Number), + p.tok(TokenKind.Op), + p.tok(TokenKind.Paren), + p.tok(TokenKind.Id), + p.tok(TokenKind.Str), + p.tok(TokenKind.SpaceNL), + p.tok(TokenKind.Comment)); + +let NOT_AT = p.alt(p.tok(TokenKind.Seperator), + p.tok(TokenKind.Semicolon), + p.tok(TokenKind.Number), + p.tok(TokenKind.ExcapeAt), + p.tok(TokenKind.Op), + p.tok(TokenKind.Paren), + p.tok(TokenKind.SpaceNL), + p.tok(TokenKind.Id), + p.tok(TokenKind.Str), + p.tok(TokenKind.Comment), + ); + +/** + * PROG : IMPORTS '---' CONTENT; */ PROG.setPattern( - p.lrec_sc(IMPORTS, p.seq(p.str('---'), UNIT), applyParts) + p.lrec_sc(IMPORTS, p.seq(p.str('---'), CONTENT), applyParts) ) /** - * PROG : 'import' Id* SEMICOLON; + * NOT_AT_TEXT : NOT_AT */ +NOT_AT_TEXT.setPattern( + p.apply(NOT_AT, applyNotAtText) +); + IMPORTS.setPattern( - p.apply(p.seq(p.str('import'), p.rep_sc(p.tok(TokenKind.Id)), SEMICOLON) , applyImports) + p.apply( p.seq(IMPORT, p.rep(IMPORT)), applyImports) +); + +/** + * IMPORT : + * 'import' IMPORTEE* SEMICOLON | + * COMMENT | + */ +IMPORT.setPattern( + p.alt( + p.apply(p.seq(p.str('import'), p.rep_sc(IMPORTEE), SEMICOLON), + applyImport), + p.apply(p.tok(TokenKind.Comment), applyComment), + p.apply(p.tok(TokenKind.SpaceNL), applySpaceNL) + + ) ); /** @@ -131,13 +237,28 @@ SEMICOLON.setPattern( p.apply(p.tok(TokenKind.Semicolon), applySemiColon) ); + + /** - * UNIT : Number; + * SEGMENT : '@' NOT_AT* '@' | + * (NOT_AT_TEXT | EXCAPE_AT)* */ -UNIT.setPattern( - p.apply(p.tok(TokenKind.Number), applyUnit) +SEGMENT.setPattern( + p.alt( + p.apply(p.rep_sc(NOT_AT_TEXT), applyText), + p.apply(p.seq(p.str('@'), p.rep(NOT_AT), p.str('@')), applySegment), + ) ); +/** + * CONTENT : SEGMENT* + */ +CONTENT.setPattern( + p.apply(p.rep(SEGMENT), applyContent) +); + + +console.log(repr(lexer.parse(inputTxt))); let tree = p.expectSingleResult(p.expectEOF(PROG.parse(lexer.parse(inputTxt))));