Skip to content

Commit c3363e7

Browse files
committed
format code
1 parent 25db599 commit c3363e7

File tree

3 files changed

+27
-27
lines changed

3 files changed

+27
-27
lines changed

src/dictionary/dictionary.ts

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,10 +7,10 @@ const rules: Rule[] = [
77
type: 'keyword',
88
default: 'import',
99
conflicts: [],
10-
description:'Determines which keyword should be used to import modules using defined in an imports statement.'
10+
description: 'Determines which keyword should be used to import modules using defined in an imports statement.'
1111
},
1212
{
13-
name:'imports-with-quotes',
13+
name: 'imports-with-quotes',
1414
type: 'boolean',
1515
default: true,
1616
conflicts: [],
@@ -47,7 +47,7 @@ const rules: Rule[] = [
4747
{
4848
name: 'export-required',
4949
type: 'boolean',
50-
default : false,
50+
default: false,
5151
conflicts: [],
5252
description: 'Determines whether is it required to export a definable in order to reach it from another file.'
5353
},
@@ -63,7 +63,7 @@ const rules: Rule[] = [
6363
type: 'boolean',
6464
default: true,
6565
conflicts: [],
66-
description : 'Determines whether is it possible to export a definable with a keyword.'
66+
description: 'Determines whether is it possible to export a definable with a keyword.'
6767
}
6868
];
6969

src/lexer.ts

Lines changed: 15 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -91,21 +91,21 @@ export function tokenizeSyx(source: string): Token[] {
9191
curPos++;
9292
}
9393
}
94-
if (src[0] === '(') tokens.push({ type: inString?20:TokenType.OpenParen, value: src.shift(), range: opr(curLine, curPos++) });
95-
else if (src[0] === ')') tokens.push({ type: inString?20:TokenType.CloseParen, value: src.shift(), range: opr(curLine, curPos++) });
96-
else if (src[0] === '{') tokens.push({ type: inString?20:TokenType.OpenBrace, value: src.shift(), range: opr(curLine, curPos++) });
97-
else if (src[0] === '}') tokens.push({ type: inString?20:TokenType.CloseBrace, value: src.shift(), range: opr(curLine, curPos++) });
98-
else if (src[0] === '[') tokens.push({ type: inString?20:TokenType.OpenSquare, value: src.shift(), range: opr(curLine, curPos++) });
99-
else if (src[0] === ']') tokens.push({ type: inString?20:TokenType.CloseSquare, value: src.shift(), range: opr(curLine, curPos++) });
100-
else if (src[0] === ',') tokens.push({ type: inString?20:TokenType.Comma, value: src.shift(), range: opr(curLine, curPos++) });
101-
else if (src[0] === ';') tokens.push({ type: inString?20:TokenType.Semicolon, value: src.shift(), range: opr(curLine, curPos++) });
102-
else if (src[0] === '<') tokens.push({ type: inString?20:TokenType.OpenDiamond, value: src.shift(), range: opr(curLine, curPos++) });
103-
else if (src[0] === '>') tokens.push({ type: inString?20:TokenType.CloseDiamond, value: src.shift(), range: opr(curLine, curPos++) });
94+
if (src[0] === '(') tokens.push({ type: inString ? 20 : TokenType.OpenParen, value: src.shift(), range: opr(curLine, curPos++) });
95+
else if (src[0] === ')') tokens.push({ type: inString ? 20 : TokenType.CloseParen, value: src.shift(), range: opr(curLine, curPos++) });
96+
else if (src[0] === '{') tokens.push({ type: inString ? 20 : TokenType.OpenBrace, value: src.shift(), range: opr(curLine, curPos++) });
97+
else if (src[0] === '}') tokens.push({ type: inString ? 20 : TokenType.CloseBrace, value: src.shift(), range: opr(curLine, curPos++) });
98+
else if (src[0] === '[') tokens.push({ type: inString ? 20 : TokenType.OpenSquare, value: src.shift(), range: opr(curLine, curPos++) });
99+
else if (src[0] === ']') tokens.push({ type: inString ? 20 : TokenType.CloseSquare, value: src.shift(), range: opr(curLine, curPos++) });
100+
else if (src[0] === ',') tokens.push({ type: inString ? 20 : TokenType.Comma, value: src.shift(), range: opr(curLine, curPos++) });
101+
else if (src[0] === ';') tokens.push({ type: inString ? 20 : TokenType.Semicolon, value: src.shift(), range: opr(curLine, curPos++) });
102+
else if (src[0] === '<') tokens.push({ type: inString ? 20 : TokenType.OpenDiamond, value: src.shift(), range: opr(curLine, curPos++) });
103+
else if (src[0] === '>') tokens.push({ type: inString ? 20 : TokenType.CloseDiamond, value: src.shift(), range: opr(curLine, curPos++) });
104104
else if (src[0] === '\'') { tokens.push({ type: TokenType.SingleQuote, value: src.shift(), range: opr(curLine, curPos++) }); t('\''); }
105105
else if (src[0] === '"') { tokens.push({ type: TokenType.DoubleQuote, value: src.shift(), range: opr(curLine, curPos++) }); t('"'); }
106-
else if (src[0] === '|') tokens.push({ type: inString?20:TokenType.VarSeperator, value: src.shift(), range: opr(curLine, curPos++) });
106+
else if (src[0] === '|') tokens.push({ type: inString ? 20 : TokenType.VarSeperator, value: src.shift(), range: opr(curLine, curPos++) });
107107
else if (src[0] === '+' && chars.includes(src[1])) {
108-
if (src[1] === 's') tokens.push({ type: inString?20:TokenType.WhitespaceIdentifier, value: '+s', range: tpr(pos(curLine, curPos), pos(curLine, curPos + 2)) });
108+
if (src[1] === 's') tokens.push({ type: inString ? 20 : TokenType.WhitespaceIdentifier, value: '+s', range: tpr(pos(curLine, curPos), pos(curLine, curPos + 2)) });
109109
curPos += 2;
110110
src.shift(); src.shift();
111111
} else if (isInt(src[0])) {
@@ -116,7 +116,7 @@ export function tokenizeSyx(source: string): Token[] {
116116
}
117117

118118
curPos += ident.length;
119-
tokens.push({ type: inString?20:TokenType.IntNumber, value: ident, range: tpr(pos(curLine, startPos), pos(curLine, curPos)) });
119+
tokens.push({ type: inString ? 20 : TokenType.IntNumber, value: ident, range: tpr(pos(curLine, startPos), pos(curLine, curPos)) });
120120
} else if (isAlphabetic(src[0])) {
121121
let ident = '';
122122
const startPos = curPos;
@@ -126,7 +126,7 @@ export function tokenizeSyx(source: string): Token[] {
126126
}
127127

128128
const reserved = keywords[ident];
129-
tokens.push({ type: inString?20:reserved ?? TokenType.Identifier, value: ident, range: tpr(pos(curLine, startPos), pos(curLine, curPos)) });
129+
tokens.push({ type: inString ? 20 : reserved ?? TokenType.Identifier, value: ident, range: tpr(pos(curLine, startPos), pos(curLine, curPos)) });
130130
} else if (isSkippable(src[0]) && !inString) {
131131
src.shift();
132132
curPos++;
@@ -161,7 +161,7 @@ export function tokenizeSys(source: string): Token[] {
161161
let curLine = 1;
162162

163163
while (src.length > 0 && `${src[0]}${src[1]}${src[2]}` !== ':::') {
164-
if (src[0] === ';') tokens.push({ type: inString?20:TokenType.Semicolon, value: src.shift(), range: opr(curLine, curPos++) });
164+
if (src[0] === ';') tokens.push({ type: inString ? 20 : TokenType.Semicolon, value: src.shift(), range: opr(curLine, curPos++) });
165165
else if (src[0] === '\'') { tokens.push({ type: TokenType.SingleQuote, value: src.shift(), range: opr(curLine, curPos++) }); t('\''); }
166166
else if (src[0] === '"') { tokens.push({ type: TokenType.DoubleQuote, value: src.shift(), range: opr(curLine, curPos++) }); t('"'); }
167167
else if (isAlphabetic(src[0])) {

src/test/compiler.test.ts

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -119,56 +119,56 @@ describe('Compiler module', () => {
119119
inst(
120120
_case(
121121
'operator <int>"+"<int> {',
122-
[TokenType.OperatorKeyword,TokenType.OpenDiamond,TokenType.Identifier,TokenType.CloseDiamond,TokenType.DoubleQuote,TokenType.Raw,TokenType.DoubleQuote,TokenType.OpenDiamond,TokenType.Identifier,TokenType.CloseDiamond,TokenType.OpenBrace,TokenType.EndOfFile]
122+
[TokenType.OperatorKeyword, TokenType.OpenDiamond, TokenType.Identifier, TokenType.CloseDiamond, TokenType.DoubleQuote, TokenType.Raw, TokenType.DoubleQuote, TokenType.OpenDiamond, TokenType.Identifier, TokenType.CloseDiamond, TokenType.OpenBrace, TokenType.EndOfFile]
123123
)
124124
);
125125

126126
inst(
127127
_case(
128128
'o-+?',
129-
[TokenType.Identifier,TokenType.Raw,TokenType.Raw,TokenType.Raw,TokenType.EndOfFile]
129+
[TokenType.Identifier, TokenType.Raw, TokenType.Raw, TokenType.Raw, TokenType.EndOfFile]
130130
)
131131
);
132132

133133
inst(
134134
_case(
135135
'rmh09345kg9',
136-
[TokenType.Identifier,TokenType.IntNumber,TokenType.Identifier,TokenType.IntNumber, TokenType.EndOfFile]
136+
[TokenType.Identifier, TokenType.IntNumber, TokenType.Identifier, TokenType.IntNumber, TokenType.EndOfFile]
137137
)
138138
);
139139

140140
inst(
141141
_case(
142142
'rule \'custom-random-rule?\';',
143-
[TokenType.RuleKeyword,TokenType.SingleQuote,20,20,20,20,20,20,TokenType.SingleQuote,TokenType.Semicolon, TokenType.EndOfFile]
143+
[TokenType.RuleKeyword, TokenType.SingleQuote, 20, 20, 20, 20, 20, 20, TokenType.SingleQuote, TokenType.Semicolon, TokenType.EndOfFile]
144144
)
145145
);
146146

147147
inst(
148148
_case(
149149
'keyword pray;rule\'imports-keyword\': pray;',
150-
[TokenType.KeywordKeyword,TokenType.Identifier,TokenType.Semicolon,TokenType.RuleKeyword,TokenType.SingleQuote,20,20,20,TokenType.SingleQuote,TokenType.Raw,TokenType.Identifier,TokenType.Semicolon, TokenType.EndOfFile]
150+
[TokenType.KeywordKeyword, TokenType.Identifier, TokenType.Semicolon, TokenType.RuleKeyword, TokenType.SingleQuote, 20, 20, 20, TokenType.SingleQuote, TokenType.Raw, TokenType.Identifier, TokenType.Semicolon, TokenType.EndOfFile]
151151
)
152152
);
153153

154154
inst(
155155
_case(
156156
'çş',
157-
[TokenType.Raw,TokenType.Raw, TokenType.EndOfFile]
157+
[TokenType.Raw, TokenType.Raw, TokenType.EndOfFile]
158158
)
159159
);
160160

161161
inst(
162162
_case(
163163
'keyword altınasıçĞ;',
164-
[TokenType.KeywordKeyword,TokenType.Identifier,20,TokenType.Identifier,20,20,20,TokenType.Semicolon, TokenType.EndOfFile]
164+
[TokenType.KeywordKeyword, TokenType.Identifier, 20, TokenType.Identifier, 20, 20, 20, TokenType.Semicolon, TokenType.EndOfFile]
165165
)
166166
);
167167

168168
inst(
169169
_case(
170170
'keyword imsodonewiththistest12casesisenough',
171-
[TokenType.KeywordKeyword,TokenType.Identifier,TokenType.IntNumber,TokenType.Identifier, TokenType.EndOfFile]
171+
[TokenType.KeywordKeyword, TokenType.Identifier, TokenType.IntNumber, TokenType.Identifier, TokenType.EndOfFile]
172172
)
173173
);
174174

0 commit comments

Comments
 (0)