| | |
| |
|
| | const fs = require('fs'); |
| | const ts = require('typescript'); |
| | const { tsquery } = require('@phenomnomnominal/tsquery'); |
| |
|
| | let data = ''; |
| |
|
| | process.stdin.on('data', function (chunk) { |
| | data += chunk; |
| | }); |
| |
|
| | process.stdin.on('end', function () { |
| | const sourceFile = ts.createSourceFile('temp.ts', data.toString(), ts.ScriptTarget.ES2015, true); |
| |
|
| | |
| | const semanticChunkNodeTypes = [ |
| | 'FunctionDeclaration', |
| | 'ArrowFunction', |
| | 'ClassDeclaration', |
| | 'InterfaceDeclaration', |
| | 'EnumDeclaration', |
| | 'TypeAliasDeclaration', |
| | 'MethodDeclaration', |
| | ]; |
| |
|
| | const semanticChunks = semanticChunkNodeTypes.flatMap(nodeType => |
| | tsquery(sourceFile, nodeType) |
| | ); |
| |
|
| | const jsonl = semanticChunks.map(chunk => { |
| | const comments = ts.getLeadingCommentRanges(sourceFile.getFullText(), chunk.getFullStart()) || []; |
| | const commentTexts = comments.map(comment => sourceFile.getFullText().slice(comment.pos, comment.end)).join('\n'); |
| |
|
| | |
| | const contentWithComments = commentTexts + '\n' + chunk.getText(sourceFile); |
| |
|
| | return JSON.stringify({ |
| | type: ts.SyntaxKind[chunk.kind], |
| | content: contentWithComments |
| | }); |
| | }).join('\n'); |
| |
|
| | fs.writeFileSync('semantic_chunks.jsonl', jsonl); |
| | }); |
| |
|