Compare commits
98 Commits
abc6e44fb2
...
func-secti
| Author | SHA1 | Date | |
|---|---|---|---|
| 06a99ce232 | |||
|
|
1c2194b68a | ||
|
|
453a596587 | ||
| c3b6330b22 | |||
| 723b506005 | |||
| 6bbee2e13b | |||
|
|
9fd285920b | ||
|
|
e630ec6f04 | ||
|
|
0ac71fa1c3 | ||
|
|
9232432c35 | ||
|
|
b536b01eeb | ||
|
|
8175a9d4c5 | ||
|
|
3dd2ea83d3 | ||
|
|
b7631530bc | ||
|
|
fa1d8efe55 | ||
| e74aff3299 | |||
|
|
89a60e620e | ||
|
|
cd528552c8 | ||
|
|
067bf2f4df | ||
| 777c8df6a4 | |||
| c470997887 | |||
| 715766edb4 | |||
| 821fa0ecb3 | |||
| e316eb7791 | |||
| 731cc828ce | |||
| 05aa0e6177 | |||
| fb43f96acc | |||
|
|
b64fbd9fc4 | ||
|
|
0d366964ca | ||
| a5477717eb | |||
|
|
0b80a55f79 | ||
|
|
08935d69c0 | ||
|
|
39f8d7e4ac | ||
|
|
1f88b54eaa | ||
| b0d4ecc83f | |||
| 4eac5c67aa | |||
| 441b036a1c | |||
| 8817d72cb3 | |||
| 3ef1e706b3 | |||
| 944fc8514e | |||
|
|
cd55a0ad8d | ||
|
|
f95c7e0b1c | ||
| 15d1b602b3 | |||
|
|
c29efd97ba | ||
|
|
aa84d9a429 | ||
|
|
5dcf3b3d1a | ||
|
|
d8074fa5cb | ||
|
|
6a6fe8353e | ||
|
|
c4f763af5b | ||
|
|
6fbda34300 | ||
|
|
59126f60cc | ||
|
|
ca80a5968d | ||
| 61819311e9 | |||
|
|
f3b2d11f59 | ||
|
|
3900bbe7bf | ||
|
|
b878017b81 | ||
|
|
5271876196 | ||
|
|
617d76fc46 | ||
|
|
0ceaedbcd8 | ||
|
|
edb9c1a0b6 | ||
|
|
bd433fc65d | ||
|
|
c847d2187d | ||
|
|
cb2264977a | ||
|
|
790e7e632e | ||
|
|
fc1568aece | ||
|
|
222c47ced9 | ||
|
|
da6d587a48 | ||
| 018499310c | |||
|
|
78b8b9dacd | ||
|
|
2605d1fb09 | ||
|
|
9dce9b2f75 | ||
|
|
9b4279c052 | ||
|
|
2296765e81 | ||
|
|
19d0b3f455 | ||
|
|
e25e7bdf14 | ||
|
|
63419165dd | ||
|
|
69aaae8f14 | ||
|
|
717474a59e | ||
|
|
ef90115a1b | ||
|
|
cced825f74 | ||
|
|
9fd3fb1263 | ||
|
|
5c2a7aeb07 | ||
|
|
bc9beb0317 | ||
|
|
a548dcc585 | ||
|
|
15eb96e8ac | ||
|
|
120976a0f3 | ||
|
|
bde4bf8493 | ||
|
|
a013d4caad | ||
|
|
be9a3603d2 | ||
|
|
54de3d1270 | ||
|
|
a87973c141 | ||
|
|
85996b2554 | ||
|
|
4780d9cc28 | ||
|
|
bb89009742 | ||
|
|
9e66305001 | ||
|
|
39e4fbe844 | ||
|
|
d42d0c5b34 | ||
|
|
ca5f8202bb |
12
README.md
12
README.md
@@ -16,7 +16,7 @@ A directory of ARF files is called a module, and modules will compile to object
|
||||
files (one per module) using C as an intermediate language (maybe LLVM IR in the
|
||||
future).
|
||||
|
||||
## Design aspects
|
||||
## Design Aspects
|
||||
|
||||
These are some design goals that I have followed/am following:
|
||||
|
||||
@@ -32,7 +32,7 @@ These are some design goals that I have followed/am following:
|
||||
- One line at a time - the language's syntax should encourage writing code that
|
||||
flows vertically and not horizontally, with minimal nesting
|
||||
|
||||
## Planned features
|
||||
## Planned Features
|
||||
|
||||
- Type definition through inheritence
|
||||
- Struct member functions
|
||||
@@ -49,3 +49,11 @@ These are some design goals that I have followed/am following:
|
||||
- [ ] Semantic tree -> C -> object file
|
||||
- [ ] Figure out HOW to implement generics
|
||||
- [ ] Create a standard library
|
||||
|
||||
## Compiler Progress
|
||||
|
||||
<img src="assets/heatmap.png" alt="Progress heatmap" width="400">
|
||||
|
||||
- Yellow: needs to be completed for the MVP
|
||||
- Lime: ongoing progress in this area
|
||||
- Green: Already completed
|
||||
|
||||
BIN
assets/heatmap.png
Normal file
BIN
assets/heatmap.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 119 KiB |
@@ -1,8 +1,8 @@
|
||||
<svg xmlns="http://www.w3.org/2000/svg" viewBox="0 0 288 80" width="288" height="80">
|
||||
<path d="M48 0L112 0L112 32L96 32L96 16L56 16L40 32L16 32L48 0Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M96 64L136 64L136 80L104 80L96 72L96 64Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M120 0L120 32L136 32L136 16L184 16L184 32L176 40L8 40L0 48L0 56L184 56L200 40L200 32L200 0L120 0Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M191 61L204 48L236 80L210 80L191 61Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M256 40L208 40L224 56L256 56L256 40Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M208 0L288 0L288 16L224 16L224 32L208 32L208 0Z" fill="#bf616a" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M48 0L112 0L112 32L96 32L96 16L56 16L40 32L16 32L48 0Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M96 64L136 64L136 80L104 80L96 72L96 64Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M120 0L120 32L136 32L136 16L184 16L184 32L176 40L8 40L0 48L0 56L184 56L200 40L200 32L200 0L120 0Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M191 61L204 48L236 80L210 80L191 61Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M256 40L208 40L224 56L256 56L256 40Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
<path d="M208 0L288 0L288 16L224 16L224 32L208 32L208 0Z" fill="#b81414" fill-rule="evenodd" opacity="1" stroke="none"/>
|
||||
</svg>
|
||||
|
||||
|
Before Width: | Height: | Size: 847 B After Width: | Height: | Size: 847 B |
@@ -1,3 +1,9 @@
|
||||
:arf
|
||||
require "io"
|
||||
---
|
||||
|
||||
func rr main
|
||||
func ro main
|
||||
> arguments:{String}
|
||||
< status:Int 0
|
||||
---
|
||||
io.println "hello world"
|
||||
|
||||
@@ -5,37 +5,31 @@ require "io"
|
||||
---
|
||||
|
||||
# this is a global variable
|
||||
data wn helloText:String "Hello, world!"
|
||||
data pv helloText:String "Hello, world!"
|
||||
|
||||
# this is a struct definition
|
||||
type rr Greeter:Obj
|
||||
# "Hi." is a string constant. all Greeters will be initialized with a
|
||||
# pointer to it. I don't know really it depends on what I decide that
|
||||
# a String type even is.
|
||||
wr text:String "Hi."
|
||||
"sdfdsf" "ahh"
|
||||
"asdf"
|
||||
objt ro Greeter:Obj
|
||||
rw text:String "Hi."
|
||||
|
||||
# this is a function
|
||||
func rr main
|
||||
> argc:Int
|
||||
> argv:{String}
|
||||
< status:Int 0
|
||||
---
|
||||
let greeter:Greeter:mut
|
||||
greeter.setText helloText
|
||||
greeter.greet
|
||||
func ro main
|
||||
> arguments:{String}
|
||||
< status:Int 0
|
||||
---
|
||||
set greeter:Greeter:mut
|
||||
greeter.setText helloText
|
||||
greeter.greet
|
||||
|
||||
# this is a member function
|
||||
func rr greet
|
||||
@ greeter:{Greeter}
|
||||
---
|
||||
io.println greeter.text
|
||||
func ro greet
|
||||
@ greeter:{Greeter}
|
||||
---
|
||||
io.println greeter.text
|
||||
|
||||
# this is mutator member function
|
||||
func rr setText
|
||||
@ greeter:{Greeter}
|
||||
> text:String
|
||||
---
|
||||
greeter.text.set text
|
||||
func ro setText
|
||||
@ greeter:{Greeter}
|
||||
> text:String
|
||||
---
|
||||
greeter.text.set text
|
||||
|
||||
|
||||
@@ -1,11 +0,0 @@
|
||||
:arf
|
||||
require io
|
||||
---
|
||||
|
||||
func rr main
|
||||
> argc:Int
|
||||
> argv:{String}
|
||||
< status:Int
|
||||
---
|
||||
io.println [io.readln]
|
||||
= status 0
|
||||
@@ -1,13 +0,0 @@
|
||||
:arf
|
||||
---
|
||||
|
||||
data:{Int 6}
|
||||
-39480 398 29 0x3AFe3 0b10001010110 0o666
|
||||
|
||||
func rr literals
|
||||
---
|
||||
= stringLiteral:String "skadjlsakdj"
|
||||
= intArrayLiteral:{Int 3} 2398
|
||||
-2938 324
|
||||
= runeLiteral:Rune 'a'
|
||||
= floatArrayLiteral:{F64 5} 3248.23 0.324 -94.29
|
||||
21
face_test.go
Normal file
21
face_test.go
Normal file
@@ -0,0 +1,21 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestFace (test *testing.T) {
|
||||
checkTree ("../tests/parser/face",
|
||||
`:arf
|
||||
---
|
||||
face ro Destroyer:Face
|
||||
destroy
|
||||
face ro ReadWriter:Face
|
||||
read
|
||||
> into:{Byte ..}
|
||||
< read:Int
|
||||
< err:Error
|
||||
write
|
||||
> data:{Byte ..}
|
||||
< wrote:Int
|
||||
< err:Error
|
||||
`, test)
|
||||
}
|
||||
12
file/file.go
12
file/file.go
@@ -8,6 +8,8 @@ type File struct {
|
||||
path string
|
||||
file *os.File
|
||||
reader *bufio.Reader
|
||||
realLine int
|
||||
realColumn int
|
||||
currentLine int
|
||||
currentColumn int
|
||||
lines []string
|
||||
@@ -42,6 +44,9 @@ func (file *File) Read (bytes []byte) (amountRead int, err error) {
|
||||
|
||||
// store the character in the file
|
||||
for _, char := range bytes {
|
||||
file.realLine = file.currentLine
|
||||
file.realColumn = file.currentColumn
|
||||
|
||||
if char == '\n' {
|
||||
file.lines = append(file.lines, "")
|
||||
file.currentLine ++
|
||||
@@ -61,6 +66,9 @@ func (file *File) Read (bytes []byte) (amountRead int, err error) {
|
||||
func (file *File) ReadRune () (char rune, size int, err error) {
|
||||
char, size, err = file.reader.ReadRune()
|
||||
|
||||
file.realLine = file.currentLine
|
||||
file.realColumn = file.currentColumn
|
||||
|
||||
if char == '\n' {
|
||||
file.lines = append(file.lines, "")
|
||||
file.currentLine ++
|
||||
@@ -106,8 +114,8 @@ func (file *File) Close () {
|
||||
func (file *File) Location (width int) (location Location) {
|
||||
return Location {
|
||||
file: file,
|
||||
row: file.currentLine,
|
||||
column: file.currentColumn,
|
||||
row: file.realLine,
|
||||
column: file.realColumn,
|
||||
width: width,
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,5 +1,7 @@
|
||||
package file
|
||||
|
||||
import "fmt"
|
||||
|
||||
// Location represents a specific point in a file. It is used for error
|
||||
// reporting.
|
||||
type Location struct {
|
||||
@@ -31,3 +33,17 @@ func (location Location) Column () (column int) {
|
||||
func (location Location) Width () (width int) {
|
||||
return location.width
|
||||
}
|
||||
|
||||
// SetWidth sets the location's width
|
||||
func (location *Location) SetWidth (width int) {
|
||||
location.width = width
|
||||
}
|
||||
|
||||
// Describe generates a description of the location for debug purposes
|
||||
func (location Location) Describe () (description string) {
|
||||
return fmt.Sprint (
|
||||
"in ", location.file.Path(),
|
||||
" row ", location.row,
|
||||
" column ", location.column,
|
||||
" width ", location.width)
|
||||
}
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
package file
|
||||
package infoerr
|
||||
|
||||
import "os"
|
||||
import "fmt"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
|
||||
type ErrorKind int
|
||||
|
||||
@@ -11,14 +12,14 @@ const (
|
||||
)
|
||||
|
||||
type Error struct {
|
||||
Location
|
||||
file.Location
|
||||
message string
|
||||
kind ErrorKind
|
||||
}
|
||||
|
||||
// NewError creates a new error at the specified location.
|
||||
func NewError (
|
||||
location Location,
|
||||
location file.Location,
|
||||
message string,
|
||||
kind ErrorKind,
|
||||
) (
|
||||
@@ -41,24 +42,23 @@ func (err Error) Error () (formattedMessage string) {
|
||||
}
|
||||
|
||||
// print information about the location of the mistake
|
||||
if err.width > 0 {
|
||||
if err.Width() > 0 {
|
||||
formattedMessage += fmt.Sprint (
|
||||
" \033[34m", err.Location.row + 1,
|
||||
":", err.Location.column + 1)
|
||||
" \033[34m", err.Row() + 1,
|
||||
":", err.Column() + 1)
|
||||
}
|
||||
formattedMessage +=
|
||||
" \033[90min\033[0m " +
|
||||
err.Location.file.path + "\n"
|
||||
err.File().Path() + "\n"
|
||||
|
||||
if err.width > 0 {
|
||||
if err.Width() > 0 {
|
||||
// print erroneous line
|
||||
line := err.Location.file.lines[err.Location.row]
|
||||
formattedMessage +=
|
||||
err.Location.file.lines[err.Location.row] + "\n"
|
||||
line := err.File().GetLine(err.Row())
|
||||
formattedMessage += line + "\n"
|
||||
|
||||
// position error marker
|
||||
var index int
|
||||
for index = 0; index < err.Location.column; index ++ {
|
||||
for index = 0; index < err.Column(); index ++ {
|
||||
if line[index] == '\t' {
|
||||
formattedMessage += "\t"
|
||||
} else {
|
||||
@@ -67,7 +67,7 @@ func (err Error) Error () (formattedMessage string) {
|
||||
}
|
||||
|
||||
// print an arrow with a tail spanning the width of the mistake
|
||||
for err.width > 1 {
|
||||
for index < err.Column() + err.Width() - 1 {
|
||||
if line[index] == '\t' {
|
||||
formattedMessage += "--------"
|
||||
} else {
|
||||
@@ -3,6 +3,7 @@ package lexer
|
||||
import "io"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// LexingOperation holds information about an ongoing lexing operataion.
|
||||
type LexingOperation struct {
|
||||
@@ -34,10 +35,10 @@ func (lexer *LexingOperation) tokenize () (err error) {
|
||||
err = lexer.nextRune()
|
||||
|
||||
if err != nil || shebangCheck[index] != lexer.char {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"not an arf file",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
}
|
||||
@@ -92,14 +93,14 @@ func (lexer *LexingOperation) tokenizeAlphaBeginning () (err error) {
|
||||
}
|
||||
|
||||
token.value = got
|
||||
token.location.SetWidth(len(got))
|
||||
|
||||
if len(got) == 2 {
|
||||
firstValid := got[0] == 'n' || got[0] == 'r' || got[0] == 'w'
|
||||
secondValid := got[1] == 'n' || got[1] == 'r' || got[1] == 'w'
|
||||
|
||||
if firstValid && secondValid {
|
||||
permission, isPermission := types.PermissionFrom(got)
|
||||
|
||||
if isPermission {
|
||||
token.kind = TokenKindPermission
|
||||
token.value = types.PermissionFrom(got)
|
||||
token.value = permission
|
||||
}
|
||||
}
|
||||
|
||||
@@ -123,10 +124,10 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if !previousToken.Is(TokenKindNewline) {
|
||||
err = lexer.nextRune()
|
||||
|
||||
file.NewError (
|
||||
infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"tab not used as indent",
|
||||
file.ErrorKindWarn).Print()
|
||||
infoerr.ErrorKindWarn).Print()
|
||||
return
|
||||
}
|
||||
|
||||
@@ -142,6 +143,7 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
}
|
||||
|
||||
token.value = indentLevel
|
||||
token.location.SetWidth(indentLevel)
|
||||
lexer.addToken(token)
|
||||
case '\n':
|
||||
// line break
|
||||
@@ -182,6 +184,7 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '.' {
|
||||
token.kind = TokenKindElipsis
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case ',':
|
||||
@@ -217,6 +220,7 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '+' {
|
||||
token.kind = TokenKindIncrement
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '-':
|
||||
@@ -238,17 +242,40 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
err = lexer.nextRune()
|
||||
case '!':
|
||||
token := lexer.newToken()
|
||||
token.kind = TokenKindExclamation
|
||||
lexer.addToken(token)
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
token.kind = TokenKindExclamation
|
||||
if lexer.char == '=' {
|
||||
token.kind = TokenKindNotEqualTo
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '%':
|
||||
token := lexer.newToken()
|
||||
token.kind = TokenKindPercent
|
||||
lexer.addToken(token)
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
token.kind = TokenKindPercent
|
||||
if lexer.char == '=' {
|
||||
token.kind = TokenKindPercentAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '~':
|
||||
token := lexer.newToken()
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
token.kind = TokenKindTilde
|
||||
if lexer.char == '=' {
|
||||
token.kind = TokenKindTildeAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '=':
|
||||
token := lexer.newToken()
|
||||
token.kind = TokenKindEqualTo
|
||||
lexer.addToken(token)
|
||||
err = lexer.nextRune()
|
||||
case '<':
|
||||
@@ -259,6 +286,16 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '<' {
|
||||
token.kind = TokenKindLShift
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
if lexer.char == '=' {
|
||||
token.kind = TokenKindLShiftAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
} else if lexer.char == '=' {
|
||||
token.kind = TokenKindLessThanEqualTo
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '>':
|
||||
@@ -269,6 +306,16 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '>' {
|
||||
token.kind = TokenKindRShift
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
if lexer.char == '=' {
|
||||
token.kind = TokenKindRShiftAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
} else if lexer.char == '=' {
|
||||
token.kind = TokenKindGreaterThanEqualTo
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '|':
|
||||
@@ -279,6 +326,11 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '|' {
|
||||
token.kind = TokenKindLogicalOr
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
} else if lexer.char == '=' {
|
||||
token.kind = TokenKindBinaryOrAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
case '&':
|
||||
@@ -289,14 +341,19 @@ func (lexer *LexingOperation) tokenizeSymbolBeginning () (err error) {
|
||||
if lexer.char == '&' {
|
||||
token.kind = TokenKindLogicalAnd
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
} else if lexer.char == '=' {
|
||||
token.kind = TokenKindBinaryAndAssignment
|
||||
err = lexer.nextRune()
|
||||
token.location.SetWidth(2)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
default:
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"unexpected symbol character " +
|
||||
string(lexer.char),
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
@@ -310,6 +367,7 @@ func (lexer *LexingOperation) tokenizeDashBeginning () (err error) {
|
||||
if lexer.char == '-' {
|
||||
token := lexer.newToken()
|
||||
token.kind = TokenKindDecrement
|
||||
token.location.SetWidth(2)
|
||||
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
@@ -317,11 +375,13 @@ func (lexer *LexingOperation) tokenizeDashBeginning () (err error) {
|
||||
if lexer.char == '-' {
|
||||
token.kind = TokenKindSeparator
|
||||
lexer.nextRune()
|
||||
token.location.SetWidth(3)
|
||||
}
|
||||
lexer.addToken(token)
|
||||
} else if lexer.char == '>' {
|
||||
token := lexer.newToken()
|
||||
token.kind = TokenKindReturnDirection
|
||||
token.location.SetWidth(2)
|
||||
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
@@ -362,9 +422,9 @@ func (lexer *LexingOperation) skipSpaces () (err error) {
|
||||
func (lexer *LexingOperation) nextRune () (err error) {
|
||||
lexer.char, _, err = lexer.file.ReadRune()
|
||||
if err != nil && err != io.EOF {
|
||||
return file.NewError (
|
||||
return infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
err.Error(), file.ErrorKindError)
|
||||
err.Error(), infoerr.ErrorKindError)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
@@ -3,8 +3,17 @@ package lexer
|
||||
import "testing"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
func checkTokenSlice (filePath string, correct []Token, test *testing.T) {
|
||||
func quickToken (width int, kind TokenKind, value any) (token Token) {
|
||||
token.location.SetWidth(width)
|
||||
token.kind = kind
|
||||
token.value = value
|
||||
return
|
||||
}
|
||||
|
||||
func checkTokenSlice (filePath string, test *testing.T, correct ...Token) {
|
||||
test.Log("checking lexer results for", filePath)
|
||||
file, err := file.Open(filePath)
|
||||
if err != nil {
|
||||
test.Log(err)
|
||||
@@ -34,6 +43,15 @@ func checkTokenSlice (filePath string, correct []Token, test *testing.T) {
|
||||
test.Log("token slice length match", len(tokens), "=", len(correct))
|
||||
|
||||
for index, token := range tokens {
|
||||
if token.location.Width() != correct[index].location.Width() {
|
||||
test.Log("token", index, "has bad width")
|
||||
test.Log (
|
||||
"have", token.location.Width(),
|
||||
"want", correct[index].location.Width())
|
||||
test.Fail()
|
||||
return
|
||||
}
|
||||
|
||||
if !token.Equals(correct[index]) {
|
||||
test.Log("token", index, "not equal")
|
||||
test.Log (
|
||||
@@ -46,118 +64,209 @@ func checkTokenSlice (filePath string, correct []Token, test *testing.T) {
|
||||
test.Log("token slice content match")
|
||||
}
|
||||
|
||||
func compareErr (
|
||||
filePath string,
|
||||
correctKind infoerr.ErrorKind,
|
||||
correctMessage string,
|
||||
correctRow int,
|
||||
correctColumn int,
|
||||
correctWidth int,
|
||||
test *testing.T,
|
||||
) {
|
||||
test.Log("testing errors in", filePath)
|
||||
file, err := file.Open(filePath)
|
||||
if err != nil {
|
||||
test.Log(err)
|
||||
test.Fail()
|
||||
return
|
||||
}
|
||||
|
||||
_, err = Tokenize(file)
|
||||
check := err.(infoerr.Error)
|
||||
|
||||
test.Log("error that was recieved:")
|
||||
test.Log(check)
|
||||
|
||||
if check.Kind() != correctKind {
|
||||
test.Log("mismatched error kind")
|
||||
test.Log("- want:", correctKind)
|
||||
test.Log("- have:", check.Kind())
|
||||
test.Fail()
|
||||
}
|
||||
|
||||
if check.Message() != correctMessage {
|
||||
test.Log("mismatched error message")
|
||||
test.Log("- want:", correctMessage)
|
||||
test.Log("- have:", check.Message())
|
||||
test.Fail()
|
||||
}
|
||||
|
||||
if check.Row() != correctRow {
|
||||
test.Log("mismatched error row")
|
||||
test.Log("- want:", correctRow)
|
||||
test.Log("- have:", check.Row())
|
||||
test.Fail()
|
||||
}
|
||||
|
||||
if check.Column() != correctColumn {
|
||||
test.Log("mismatched error column")
|
||||
test.Log("- want:", correctColumn)
|
||||
test.Log("- have:", check.Column())
|
||||
test.Fail()
|
||||
}
|
||||
|
||||
if check.Width() != correctWidth {
|
||||
test.Log("mismatched error width")
|
||||
test.Log("- want:", check.Width())
|
||||
test.Log("- have:", correctWidth)
|
||||
test.Fail()
|
||||
}
|
||||
}
|
||||
|
||||
func TestTokenizeAll (test *testing.T) {
|
||||
checkTokenSlice("../tests/lexer/all.arf", []Token {
|
||||
Token { kind: TokenKindSeparator },
|
||||
Token { kind: TokenKindPermission, value: types.Permission {
|
||||
Internal: types.ModeRead,
|
||||
External: types.ModeWrite,
|
||||
}},
|
||||
Token { kind: TokenKindReturnDirection },
|
||||
Token { kind: TokenKindInt, value: int64(-349820394) },
|
||||
Token { kind: TokenKindUInt, value: uint64(932748397) },
|
||||
Token { kind: TokenKindFloat, value: 239485.37520 },
|
||||
Token { kind: TokenKindString, value: "hello world!\n" },
|
||||
Token { kind: TokenKindRune, value: 'E' },
|
||||
Token { kind: TokenKindName, value: "helloWorld" },
|
||||
Token { kind: TokenKindColon },
|
||||
Token { kind: TokenKindDot },
|
||||
Token { kind: TokenKindComma },
|
||||
Token { kind: TokenKindElipsis },
|
||||
Token { kind: TokenKindLBracket },
|
||||
Token { kind: TokenKindRBracket },
|
||||
Token { kind: TokenKindLBrace },
|
||||
Token { kind: TokenKindRBrace },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindPlus },
|
||||
Token { kind: TokenKindMinus },
|
||||
Token { kind: TokenKindIncrement },
|
||||
Token { kind: TokenKindDecrement },
|
||||
Token { kind: TokenKindAsterisk },
|
||||
Token { kind: TokenKindSlash },
|
||||
Token { kind: TokenKindAt },
|
||||
Token { kind: TokenKindExclamation },
|
||||
Token { kind: TokenKindPercent },
|
||||
Token { kind: TokenKindTilde },
|
||||
Token { kind: TokenKindLessThan },
|
||||
Token { kind: TokenKindLShift },
|
||||
Token { kind: TokenKindGreaterThan },
|
||||
Token { kind: TokenKindRShift },
|
||||
Token { kind: TokenKindBinaryOr },
|
||||
Token { kind: TokenKindLogicalOr },
|
||||
Token { kind: TokenKindBinaryAnd },
|
||||
Token { kind: TokenKindLogicalAnd },
|
||||
Token { kind: TokenKindNewline },
|
||||
}, test)
|
||||
checkTokenSlice("../tests/lexer/all.arf", test,
|
||||
quickToken(3, TokenKindSeparator, nil),
|
||||
quickToken(2, TokenKindPermission, types.PermissionReadWrite),
|
||||
quickToken(2, TokenKindReturnDirection, nil),
|
||||
quickToken(10, TokenKindInt, int64(-349820394)),
|
||||
quickToken(9, TokenKindUInt, uint64(932748397)),
|
||||
quickToken(12, TokenKindFloat, 239485.37520),
|
||||
quickToken(16, TokenKindString, "hello world!\n"),
|
||||
quickToken(3, TokenKindRune, 'E'),
|
||||
quickToken(10, TokenKindName, "helloWorld"),
|
||||
quickToken(1, TokenKindColon, nil),
|
||||
quickToken(1, TokenKindDot, nil),
|
||||
quickToken(1, TokenKindComma, nil),
|
||||
quickToken(2, TokenKindElipsis, nil),
|
||||
quickToken(1, TokenKindLBracket, nil),
|
||||
quickToken(1, TokenKindRBracket, nil),
|
||||
quickToken(1, TokenKindLBrace, nil),
|
||||
quickToken(1, TokenKindRBrace, nil),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(1, TokenKindPlus, nil),
|
||||
quickToken(1, TokenKindMinus, nil),
|
||||
quickToken(2, TokenKindIncrement, nil),
|
||||
quickToken(2, TokenKindDecrement, nil),
|
||||
quickToken(1, TokenKindAsterisk, nil),
|
||||
quickToken(1, TokenKindSlash, nil),
|
||||
quickToken(1, TokenKindAt, nil),
|
||||
quickToken(1, TokenKindExclamation, nil),
|
||||
quickToken(1, TokenKindPercent, nil),
|
||||
quickToken(2, TokenKindPercentAssignment, nil),
|
||||
quickToken(1, TokenKindTilde, nil),
|
||||
quickToken(2, TokenKindTildeAssignment, nil),
|
||||
quickToken(1, TokenKindEqualTo, nil),
|
||||
quickToken(2, TokenKindNotEqualTo, nil),
|
||||
quickToken(1, TokenKindLessThan, nil),
|
||||
quickToken(2, TokenKindLessThanEqualTo, nil),
|
||||
quickToken(2, TokenKindLShift, nil),
|
||||
quickToken(3, TokenKindLShiftAssignment, nil),
|
||||
quickToken(1, TokenKindGreaterThan, nil),
|
||||
quickToken(2, TokenKindGreaterThanEqualTo, nil),
|
||||
quickToken(2, TokenKindRShift, nil),
|
||||
quickToken(3, TokenKindRShiftAssignment, nil),
|
||||
quickToken(1, TokenKindBinaryOr, nil),
|
||||
quickToken(2, TokenKindBinaryOrAssignment, nil),
|
||||
quickToken(2, TokenKindLogicalOr, nil),
|
||||
quickToken(1, TokenKindBinaryAnd, nil),
|
||||
quickToken(2, TokenKindBinaryAndAssignment, nil),
|
||||
quickToken(2, TokenKindLogicalAnd, nil),
|
||||
quickToken(1, TokenKindBinaryXor, nil),
|
||||
quickToken(2, TokenKindBinaryXorAssignment, nil),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
)
|
||||
}
|
||||
|
||||
func TestTokenizeNumbers (test *testing.T) {
|
||||
checkTokenSlice("../tests/lexer/numbers.arf", []Token {
|
||||
Token { kind: TokenKindUInt, value: uint64(0) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindUInt, value: uint64(8) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindUInt, value: uint64(83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindUInt, value: uint64(83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindUInt, value: uint64(83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindUInt, value: uint64(83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
checkTokenSlice("../tests/lexer/numbers.arf", test,
|
||||
quickToken(1, TokenKindUInt, uint64(0)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(1, TokenKindUInt, uint64(8)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(8, TokenKindUInt, uint64(83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(29, TokenKindUInt, uint64(83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(9, TokenKindUInt, uint64(83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(10, TokenKindUInt, uint64(83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
|
||||
Token { kind: TokenKindInt, value: int64(-83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindInt, value: int64(-83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindInt, value: int64(-83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindInt, value: int64(-83628266) },
|
||||
Token { kind: TokenKindNewline },
|
||||
quickToken(9, TokenKindInt, int64(-83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(30, TokenKindInt, int64(-83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(10, TokenKindInt, int64(-83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(11, TokenKindInt, int64(-83628266)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
|
||||
Token { kind: TokenKindFloat, value: float64(0.123478) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindFloat, value: float64(234.3095) },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindFloat, value: float64(-2.312) },
|
||||
Token { kind: TokenKindNewline },
|
||||
}, test)
|
||||
quickToken(8, TokenKindFloat, float64(0.123478)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(8, TokenKindFloat, float64(234.3095)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(6, TokenKindFloat, float64(-2.312)),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
)
|
||||
}
|
||||
|
||||
func TestTokenizeText (test *testing.T) {
|
||||
checkTokenSlice("../tests/lexer/text.arf", []Token {
|
||||
Token { kind: TokenKindString, value: "hello world!\a\b\f\n\r\t\v'\"\\" },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindRune, value: '\a' },
|
||||
Token { kind: TokenKindRune, value: '\b' },
|
||||
Token { kind: TokenKindRune, value: '\f' },
|
||||
Token { kind: TokenKindRune, value: '\n' },
|
||||
Token { kind: TokenKindRune, value: '\r' },
|
||||
Token { kind: TokenKindRune, value: '\t' },
|
||||
Token { kind: TokenKindRune, value: '\v' },
|
||||
Token { kind: TokenKindRune, value: '\'' },
|
||||
Token { kind: TokenKindRune, value: '"' },
|
||||
Token { kind: TokenKindRune, value: '\\' },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindString, value: "hello world \x40\u0040\U00000040!" },
|
||||
Token { kind: TokenKindNewline },
|
||||
}, test)
|
||||
checkTokenSlice("../tests/lexer/text.arf", test,
|
||||
quickToken(34, TokenKindString, "hello world!\a\b\f\n\r\t\v'\"\\"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(4, TokenKindRune, '\a'),
|
||||
quickToken(4, TokenKindRune, '\b'),
|
||||
quickToken(4, TokenKindRune, '\f'),
|
||||
quickToken(4, TokenKindRune, '\n'),
|
||||
quickToken(4, TokenKindRune, '\r'),
|
||||
quickToken(4, TokenKindRune, '\t'),
|
||||
quickToken(4, TokenKindRune, '\v'),
|
||||
quickToken(4, TokenKindRune, '\''),
|
||||
quickToken(4, TokenKindRune, '"' ),
|
||||
quickToken(4, TokenKindRune, '\\'),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(35, TokenKindString, "hello world \x40\u0040\U00000040!"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
)
|
||||
}
|
||||
|
||||
func TestTokenizeIndent (test *testing.T) {
|
||||
checkTokenSlice("../tests/lexer/indent.arf", []Token {
|
||||
Token { kind: TokenKindName, value: "line1" },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindIndent, value: 1 },
|
||||
Token { kind: TokenKindName, value: "line2" },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindIndent, value: 4 },
|
||||
Token { kind: TokenKindName, value: "line3" },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindName, value: "line4" },
|
||||
Token { kind: TokenKindNewline },
|
||||
Token { kind: TokenKindIndent, value: 2 },
|
||||
Token { kind: TokenKindName, value: "line5" },
|
||||
Token { kind: TokenKindNewline },
|
||||
}, test)
|
||||
checkTokenSlice("../tests/lexer/indent.arf", test,
|
||||
quickToken(5, TokenKindName, "line1"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(1, TokenKindIndent, 1),
|
||||
quickToken(5, TokenKindName, "line2"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(4, TokenKindIndent, 4),
|
||||
quickToken(5, TokenKindName, "line3"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(5, TokenKindName, "line4"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
quickToken(2, TokenKindIndent, 2),
|
||||
quickToken(5, TokenKindName, "line5"),
|
||||
quickToken(1, TokenKindNewline, nil),
|
||||
)
|
||||
}
|
||||
|
||||
func TestTokenizeErr (test *testing.T) {
|
||||
compareErr (
|
||||
"../tests/lexer/error/unexpectedSymbol.arf",
|
||||
infoerr.ErrorKindError,
|
||||
"unexpected symbol character ;",
|
||||
1, 5, 1,
|
||||
test)
|
||||
|
||||
compareErr (
|
||||
"../tests/lexer/error/excessDataRune.arf",
|
||||
infoerr.ErrorKindError,
|
||||
"excess data in rune literal",
|
||||
1, 1, 7,
|
||||
test)
|
||||
|
||||
compareErr (
|
||||
"../tests/lexer/error/unknownEscape.arf",
|
||||
infoerr.ErrorKindError,
|
||||
"unknown escape character g",
|
||||
1, 2, 1,
|
||||
test)
|
||||
}
|
||||
|
||||
@@ -1,32 +1,55 @@
|
||||
package lexer
|
||||
|
||||
import "strconv"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// tokenizeSymbolBeginning lexes a token that starts with a number.
|
||||
func (lexer *LexingOperation) tokenizeNumberBeginning (negative bool) (err error) {
|
||||
var intNumber uint64
|
||||
var floatNumber float64
|
||||
var isFloat bool
|
||||
var amountRead int
|
||||
var totalRead int
|
||||
|
||||
token := lexer.newToken()
|
||||
|
||||
if lexer.char == '0' {
|
||||
lexer.nextRune()
|
||||
totalRead ++
|
||||
|
||||
if lexer.char == 'x' {
|
||||
lexer.nextRune()
|
||||
intNumber, floatNumber, isFloat, err = lexer.tokenizeNumber(16)
|
||||
totalRead ++
|
||||
intNumber, floatNumber,
|
||||
isFloat, amountRead,
|
||||
err = lexer.tokenizeNumber(16)
|
||||
|
||||
} else if lexer.char == 'b' {
|
||||
lexer.nextRune()
|
||||
intNumber, floatNumber, isFloat, err = lexer.tokenizeNumber(2)
|
||||
totalRead ++
|
||||
intNumber, floatNumber,
|
||||
isFloat, amountRead,
|
||||
err = lexer.tokenizeNumber(2)
|
||||
|
||||
} else if lexer.char == '.' {
|
||||
intNumber, floatNumber, isFloat, err = lexer.tokenizeNumber(10)
|
||||
intNumber, floatNumber,
|
||||
isFloat, amountRead,
|
||||
err = lexer.tokenizeNumber(10)
|
||||
|
||||
} else if lexer.char >= '0' && lexer.char <= '9' {
|
||||
intNumber, floatNumber, isFloat, err = lexer.tokenizeNumber(8)
|
||||
intNumber, floatNumber,
|
||||
isFloat, amountRead,
|
||||
err = lexer.tokenizeNumber(8)
|
||||
}
|
||||
} else {
|
||||
intNumber, floatNumber, isFloat, err = lexer.tokenizeNumber(10)
|
||||
intNumber, floatNumber,
|
||||
isFloat, amountRead,
|
||||
err = lexer.tokenizeNumber(10)
|
||||
}
|
||||
|
||||
totalRead += amountRead
|
||||
if negative {
|
||||
totalRead += 1
|
||||
}
|
||||
|
||||
if err != nil { return }
|
||||
@@ -47,7 +70,8 @@ func (lexer *LexingOperation) tokenizeNumberBeginning (negative bool) (err error
|
||||
token.value = uint64(intNumber)
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
token.location.SetWidth(totalRead)
|
||||
lexer.addToken(token)
|
||||
return
|
||||
}
|
||||
@@ -82,6 +106,7 @@ func (lexer *LexingOperation) tokenizeNumber (
|
||||
intNumber uint64,
|
||||
floatNumber float64,
|
||||
isFloat bool,
|
||||
amountRead int,
|
||||
err error,
|
||||
) {
|
||||
got := ""
|
||||
@@ -89,10 +114,10 @@ func (lexer *LexingOperation) tokenizeNumber (
|
||||
if !runeIsDigit(lexer.char, radix) { break }
|
||||
if lexer.char == '.' {
|
||||
if radix != 10 {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"floats must have radix of 10",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
isFloat = true
|
||||
@@ -103,6 +128,8 @@ func (lexer *LexingOperation) tokenizeNumber (
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
amountRead = len(got)
|
||||
|
||||
if isFloat {
|
||||
floatNumber, err = strconv.ParseFloat(got, 64)
|
||||
} else {
|
||||
@@ -110,10 +137,10 @@ func (lexer *LexingOperation) tokenizeNumber (
|
||||
}
|
||||
|
||||
if err != nil {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"could not parse number: " + err.Error(),
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
package lexer
|
||||
|
||||
import "strconv"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// tokenizeString tokenizes a string or rune literal.
|
||||
func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) {
|
||||
@@ -9,17 +9,20 @@ func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) {
|
||||
if err != nil { return }
|
||||
|
||||
token := lexer.newToken()
|
||||
got := ""
|
||||
tokenWidth := 2
|
||||
|
||||
got := ""
|
||||
|
||||
beginning := lexer.file.Location(1)
|
||||
for {
|
||||
// TODO: add hexadecimal escape codes
|
||||
if lexer.char == '\\' {
|
||||
err = lexer.nextRune()
|
||||
tokenWidth ++
|
||||
if err != nil { return }
|
||||
|
||||
var actual rune
|
||||
actual, err = lexer.getEscapeSequence()
|
||||
var actual rune
|
||||
var amountRead int
|
||||
actual, amountRead, err = lexer.getEscapeSequence()
|
||||
tokenWidth += amountRead
|
||||
if err != nil { return }
|
||||
|
||||
got += string(actual)
|
||||
@@ -27,6 +30,7 @@ func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) {
|
||||
got += string(lexer.char)
|
||||
|
||||
err = lexer.nextRune()
|
||||
tokenWidth ++
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
@@ -40,12 +44,13 @@ func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) {
|
||||
err = lexer.nextRune()
|
||||
if err != nil { return }
|
||||
|
||||
beginning.SetWidth(len(got))
|
||||
if isRuneLiteral {
|
||||
if len(got) > 1 {
|
||||
err = file.NewError (
|
||||
lexer.file.Location(1),
|
||||
err = infoerr.NewError (
|
||||
beginning,
|
||||
"excess data in rune literal",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
@@ -56,6 +61,7 @@ func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) {
|
||||
token.value = got
|
||||
}
|
||||
|
||||
token.location.SetWidth(tokenWidth)
|
||||
lexer.addToken(token)
|
||||
return
|
||||
}
|
||||
@@ -76,16 +82,22 @@ var escapeSequenceMap = map[rune] rune {
|
||||
}
|
||||
|
||||
// getEscapeSequence reads an escape sequence in a string or rune literal.
|
||||
func (lexer *LexingOperation) getEscapeSequence () (result rune, err error) {
|
||||
func (lexer *LexingOperation) getEscapeSequence () (
|
||||
result rune,
|
||||
amountRead int,
|
||||
err error,
|
||||
) {
|
||||
result, exists := escapeSequenceMap[lexer.char]
|
||||
if exists {
|
||||
err = lexer.nextRune()
|
||||
amountRead ++
|
||||
return
|
||||
} else if lexer.char >= '0' && lexer.char <= '7' {
|
||||
// octal escape sequence
|
||||
number := string(lexer.char)
|
||||
|
||||
err = lexer.nextRune()
|
||||
amountRead ++
|
||||
if err != nil { return }
|
||||
|
||||
for len(number) < 3 {
|
||||
@@ -94,14 +106,15 @@ func (lexer *LexingOperation) getEscapeSequence () (result rune, err error) {
|
||||
number += string(lexer.char)
|
||||
|
||||
err = lexer.nextRune()
|
||||
amountRead ++
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
if len(number) < 3 {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"octal escape sequence too short",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
@@ -117,6 +130,7 @@ func (lexer *LexingOperation) getEscapeSequence () (result rune, err error) {
|
||||
number := ""
|
||||
|
||||
err = lexer.nextRune()
|
||||
amountRead ++
|
||||
if err != nil { return }
|
||||
|
||||
for len(number) < want {
|
||||
@@ -128,24 +142,25 @@ func (lexer *LexingOperation) getEscapeSequence () (result rune, err error) {
|
||||
number += string(lexer.char)
|
||||
|
||||
err = lexer.nextRune()
|
||||
amountRead ++
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
if len(number) < want {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"hex escape sequence too short ",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
parsedNumber, _ := strconv.ParseInt(number, 16, want * 4)
|
||||
result = rune(parsedNumber)
|
||||
} else {
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
lexer.file.Location(1),
|
||||
"unknown escape character " +
|
||||
string(lexer.char), file.ErrorKindError)
|
||||
string(lexer.char), infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
@@ -2,6 +2,7 @@ package lexer
|
||||
|
||||
import "fmt"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// TokenKind is an enum represzenting what role a token has.
|
||||
type TokenKind int
|
||||
@@ -42,16 +43,28 @@ const (
|
||||
TokenKindAt
|
||||
TokenKindExclamation
|
||||
TokenKindPercent
|
||||
TokenKindPercentAssignment
|
||||
TokenKindTilde
|
||||
TokenKindTildeAssignment
|
||||
|
||||
TokenKindEqualTo
|
||||
TokenKindNotEqualTo
|
||||
TokenKindLessThanEqualTo
|
||||
TokenKindLessThan
|
||||
TokenKindLShift
|
||||
TokenKindLShiftAssignment
|
||||
TokenKindGreaterThan
|
||||
TokenKindGreaterThanEqualTo
|
||||
TokenKindRShift
|
||||
TokenKindRShiftAssignment
|
||||
TokenKindBinaryOr
|
||||
TokenKindBinaryOrAssignment
|
||||
TokenKindLogicalOr
|
||||
TokenKindBinaryAnd
|
||||
TokenKindBinaryAndAssignment
|
||||
TokenKindLogicalAnd
|
||||
TokenKindBinaryXor
|
||||
TokenKindBinaryXorAssignment
|
||||
)
|
||||
|
||||
// Token represents a single token. It holds its location in the file, as well
|
||||
@@ -89,8 +102,13 @@ func (token Token) Location () (location file.Location) {
|
||||
}
|
||||
|
||||
// NewError creates a new error at this token's location.
|
||||
func (token Token) NewError (message string, kind file.ErrorKind) (err file.Error) {
|
||||
return token.location.NewError(message, kind)
|
||||
func (token Token) NewError (
|
||||
message string,
|
||||
kind infoerr.ErrorKind,
|
||||
) (
|
||||
err infoerr.Error,
|
||||
) {
|
||||
return infoerr.NewError(token.location, message, kind)
|
||||
}
|
||||
|
||||
// Describe generates a textual description of the token to be used in debug
|
||||
@@ -165,24 +183,48 @@ func (tokenKind TokenKind) Describe () (description string) {
|
||||
description = "Exclamation"
|
||||
case TokenKindPercent:
|
||||
description = "Percent"
|
||||
case TokenKindPercentAssignment:
|
||||
description = "PercentAssignment"
|
||||
case TokenKindTilde:
|
||||
description = "Tilde"
|
||||
case TokenKindTildeAssignment:
|
||||
description = "TildeAssignment"
|
||||
case TokenKindEqualTo:
|
||||
description = "EqualTo"
|
||||
case TokenKindNotEqualTo:
|
||||
description = "NotEqualTo"
|
||||
case TokenKindLessThan:
|
||||
description = "LessThan"
|
||||
case TokenKindLessThanEqualTo:
|
||||
description = "LessThanEqualTo"
|
||||
case TokenKindLShift:
|
||||
description = "LShift"
|
||||
case TokenKindLShiftAssignment:
|
||||
description = "LShiftAssignment"
|
||||
case TokenKindGreaterThan:
|
||||
description = "GreaterThan"
|
||||
case TokenKindGreaterThanEqualTo:
|
||||
description = "GreaterThanEqualTo"
|
||||
case TokenKindRShift:
|
||||
description = "RShift"
|
||||
case TokenKindRShiftAssignment:
|
||||
description = "RShiftAssignment"
|
||||
case TokenKindBinaryOr:
|
||||
description = "BinaryOr"
|
||||
case TokenKindBinaryOrAssignment:
|
||||
description = "BinaryOrAssignment"
|
||||
case TokenKindLogicalOr:
|
||||
description = "LogicalOr"
|
||||
case TokenKindBinaryAnd:
|
||||
description = "BinaryAnd"
|
||||
case TokenKindBinaryAndAssignment:
|
||||
description = "BinaryAndAssignment"
|
||||
case TokenKindLogicalAnd:
|
||||
description = "LogicalAnd"
|
||||
case TokenKindBinaryXor:
|
||||
description = "BinaryXor"
|
||||
case TokenKindBinaryXorAssignment:
|
||||
description = "BinaryXorAssignment"
|
||||
}
|
||||
|
||||
return
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
var validArgumentStartTokens = []lexer.TokenKind {
|
||||
lexer.TokenKindName,
|
||||
@@ -37,7 +37,7 @@ func (parser *ParsingOperation) parseArgument () (argument Argument, err error)
|
||||
err = parser.token.NewError (
|
||||
"cannot use member selection in " +
|
||||
"a variable definition",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
@@ -1,7 +1,7 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parse body parses the body of an arf file, after the metadata header.
|
||||
func (parser *ParsingOperation) parseBody () (err error) {
|
||||
@@ -21,13 +21,54 @@ func (parser *ParsingOperation) parseBody () (err error) {
|
||||
parser.tree.dataSections[section.name] = section
|
||||
if err != nil { return }
|
||||
case "type":
|
||||
var section *TypeSection
|
||||
section, err = parser.parseTypeSection()
|
||||
if parser.tree.typeSections == nil {
|
||||
parser.tree.typeSections =
|
||||
make(map[string] *TypeSection)
|
||||
}
|
||||
parser.tree.typeSections[section.name] = section
|
||||
if err != nil { return }
|
||||
case "objt":
|
||||
var section *ObjtSection
|
||||
section, err = parser.parseObjtSection()
|
||||
if parser.tree.objtSections == nil {
|
||||
parser.tree.objtSections =
|
||||
make(map[string] *ObjtSection)
|
||||
}
|
||||
parser.tree.objtSections[section.name] = section
|
||||
if err != nil { return }
|
||||
case "face":
|
||||
var section *FaceSection
|
||||
section, err = parser.parseFaceSection()
|
||||
if parser.tree.faceSections == nil {
|
||||
parser.tree.faceSections =
|
||||
make(map[string] *FaceSection)
|
||||
}
|
||||
parser.tree.faceSections[section.name] = section
|
||||
if err != nil { return }
|
||||
case "enum":
|
||||
var section *EnumSection
|
||||
section, err = parser.parseEnumSection()
|
||||
if parser.tree.enumSections == nil {
|
||||
parser.tree.enumSections =
|
||||
make(map[string] *EnumSection)
|
||||
}
|
||||
parser.tree.enumSections[section.name] = section
|
||||
if err != nil { return }
|
||||
case "func":
|
||||
var section *FuncSection
|
||||
section, err = parser.parseFuncSection()
|
||||
if parser.tree.funcSections == nil {
|
||||
parser.tree.funcSections =
|
||||
make(map[string] *FuncSection)
|
||||
}
|
||||
parser.tree.funcSections[section.name] = section
|
||||
if err != nil { return }
|
||||
default:
|
||||
err = parser.token.NewError (
|
||||
"unknown section type \"" + sectionType + "\"",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseData parses a data section.
|
||||
func (parser *ParsingOperation) parseDataSection () (
|
||||
@@ -94,9 +94,6 @@ func (parser *ParsingOperation) parseObjectInitializationValues () (
|
||||
initializationValues ObjectInitializationValues,
|
||||
err error,
|
||||
) {
|
||||
println("BEGIN")
|
||||
defer println("END")
|
||||
|
||||
initializationValues.attributes = make(map[string] Argument)
|
||||
|
||||
baseIndent := 0
|
||||
@@ -116,8 +113,6 @@ func (parser *ParsingOperation) parseObjectInitializationValues () (
|
||||
// do not parse any further if the indent has changed
|
||||
if indent != baseIndent { break }
|
||||
|
||||
println("HIT")
|
||||
|
||||
// move on to the beginning of the line, which must contain
|
||||
// a member initialization value
|
||||
err = parser.nextToken(lexer.TokenKindDot)
|
||||
@@ -132,7 +127,7 @@ func (parser *ParsingOperation) parseObjectInitializationValues () (
|
||||
err = parser.token.NewError (
|
||||
"duplicate member \"" + name + "\" in object " +
|
||||
"member initialization",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
@@ -273,7 +268,7 @@ func (parser *ParsingOperation) parseType () (what Type, err error) {
|
||||
default:
|
||||
err = parser.token.NewError (
|
||||
"unknown type qualifier \"" + qualifier + "\"",
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
@@ -294,8 +289,6 @@ func (parser *ParsingOperation) parseIdentifier () (
|
||||
identifier.location = parser.token.Location()
|
||||
|
||||
for {
|
||||
// TODO: eat up newlines and tabs after the dot, but not before
|
||||
// it.
|
||||
if !parser.token.Is(lexer.TokenKindName) { break }
|
||||
|
||||
identifier.trail = append (
|
||||
@@ -306,6 +299,18 @@ func (parser *ParsingOperation) parseIdentifier () (
|
||||
if err != nil { return }
|
||||
|
||||
if !parser.token.Is(lexer.TokenKindDot) { break }
|
||||
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
// allow the identifier to continue on to the next line if there
|
||||
// is a line break right after the dot
|
||||
for parser.token.Is(lexer.TokenKindNewline) ||
|
||||
parser.token.Is(lexer.TokenKindIndent) {
|
||||
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
}
|
||||
}
|
||||
|
||||
return
|
||||
|
||||
38
parser/data_test.go
Normal file
38
parser/data_test.go
Normal file
@@ -0,0 +1,38 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestData (test *testing.T) {
|
||||
checkTree ("../tests/parser/data",
|
||||
`:arf
|
||||
---
|
||||
data ro integer:Int 3202
|
||||
data ro integerArray16:{Int 16}
|
||||
data ro integerArrayInitialized:{Int 16}
|
||||
3948
|
||||
293
|
||||
293049
|
||||
948
|
||||
912
|
||||
340
|
||||
0
|
||||
2304
|
||||
0
|
||||
4785
|
||||
92
|
||||
data ro integerArrayVariable:{Int ..}
|
||||
data ro integerPointer:{Int}
|
||||
data ro mutInteger:Int:mut 3202
|
||||
data ro mutIntegerPointer:{Int}:mut
|
||||
data ro nestedObject:Obj
|
||||
.that
|
||||
.bird2 123.8439
|
||||
.bird3 9328.21348239
|
||||
.this
|
||||
.bird0 324
|
||||
.bird1 "hello world"
|
||||
data ro object:thing.thing.thing.thing
|
||||
.that 2139
|
||||
.this 324
|
||||
`, test)
|
||||
}
|
||||
93
parser/enum.go
Normal file
93
parser/enum.go
Normal file
@@ -0,0 +1,93 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
func (parser *ParsingOperation) parseEnumSection () (
|
||||
section *EnumSection,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
|
||||
section = &EnumSection { location: parser.token.Location() }
|
||||
|
||||
// get permission
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
section.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.name = parser.token.Value().(string)
|
||||
|
||||
// parse inherited type
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
section.what, err = parser.parseType()
|
||||
if err != nil { return }
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
// parse members
|
||||
err = parser.parseEnumMembers(section)
|
||||
if err != nil { return }
|
||||
|
||||
if len(section.members) == 0 {
|
||||
infoerr.NewError (
|
||||
section.location,
|
||||
"defining an enum with no members",
|
||||
infoerr.ErrorKindWarn).Print()
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
// parseEnumMembers parses a list of members for an enum section. Indentation
|
||||
// level is assumed.
|
||||
func (parser *ParsingOperation) parseEnumMembers (
|
||||
into *EnumSection,
|
||||
) (
|
||||
err error,
|
||||
) {
|
||||
|
||||
for {
|
||||
// if we've left the block, stop parsing
|
||||
if !parser.token.Is(lexer.TokenKindIndent) { return }
|
||||
if parser.token.Value().(int) != 1 { return }
|
||||
|
||||
member := EnumMember { }
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
member.location = parser.token.Location()
|
||||
member.name = parser.token.Value().(string)
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
// parse default value
|
||||
if parser.token.Is(lexer.TokenKindNewline) {
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
member.value, err = parser.parseInitializationValues(1)
|
||||
into.members = append(into.members, member)
|
||||
if err != nil { return }
|
||||
} else {
|
||||
member.value, err = parser.parseArgument()
|
||||
into.members = append(into.members, member)
|
||||
if err != nil { return }
|
||||
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
}
|
||||
}
|
||||
}
|
||||
38
parser/enum_test.go
Normal file
38
parser/enum_test.go
Normal file
@@ -0,0 +1,38 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestEnum (test *testing.T) {
|
||||
checkTree ("../tests/parser/enum",
|
||||
`:arf
|
||||
---
|
||||
enum ro AffrontToGod:{Int 4}
|
||||
bird0
|
||||
28394
|
||||
9328
|
||||
398
|
||||
9
|
||||
bird1
|
||||
23
|
||||
932832
|
||||
398
|
||||
2349
|
||||
bird2
|
||||
1
|
||||
2
|
||||
3
|
||||
4
|
||||
enum ro NamedColor:U32
|
||||
red 16711680
|
||||
green 65280
|
||||
blue 255
|
||||
enum ro Weekday:Int
|
||||
sunday
|
||||
monday
|
||||
tuesday
|
||||
wednesday
|
||||
thursday
|
||||
friday
|
||||
saturday
|
||||
`, test)
|
||||
}
|
||||
132
parser/face.go
Normal file
132
parser/face.go
Normal file
@@ -0,0 +1,132 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseFaceSection parses an interface section.
|
||||
func (parser *ParsingOperation) parseFaceSection () (
|
||||
section *FaceSection,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
|
||||
section = &FaceSection {
|
||||
location: parser.token.Location(),
|
||||
behaviors: make(map[string] FaceBehavior),
|
||||
}
|
||||
|
||||
// get permission
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
section.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.name = parser.token.Value().(string)
|
||||
|
||||
// parse inherited interface
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.inherits, err = parser.parseIdentifier()
|
||||
if err != nil { return }
|
||||
err = parser.nextToken(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
// parse members
|
||||
for {
|
||||
// if we've left the block, stop parsing
|
||||
if !parser.token.Is(lexer.TokenKindIndent) { return }
|
||||
if parser.token.Value().(int) != 1 { return }
|
||||
|
||||
// parse behavior
|
||||
behaviorBeginning := parser.token.Location()
|
||||
var behavior FaceBehavior
|
||||
behavior, err = parser.parseFaceBehavior()
|
||||
|
||||
// add to section
|
||||
_, exists := section.behaviors[behavior.name]
|
||||
if exists {
|
||||
err = infoerr.NewError (
|
||||
behaviorBeginning,
|
||||
"multiple behaviors named " + behavior.name +
|
||||
" in this interface",
|
||||
infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
section.behaviors[behavior.name] = behavior
|
||||
|
||||
if err != nil { return }
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
// parseFaceBehavior parses a single interface behavior. Indentation level is
|
||||
// assumed.
|
||||
func (parser *ParsingOperation) parseFaceBehavior () (
|
||||
behavior FaceBehavior,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindIndent)
|
||||
if err != nil { return }
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
behavior.name = parser.token.Value().(string)
|
||||
|
||||
err = parser.nextToken(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
for {
|
||||
// if we've left the block, stop parsing
|
||||
if !parser.token.Is(lexer.TokenKindIndent) { return }
|
||||
if parser.token.Value().(int) != 2 { return }
|
||||
|
||||
// get preceding symbol
|
||||
err = parser.nextToken (
|
||||
lexer.TokenKindGreaterThan,
|
||||
lexer.TokenKindLessThan)
|
||||
if err != nil { return }
|
||||
kind := parser.token.Kind()
|
||||
|
||||
var declaration Declaration
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
declaration.name = parser.token.Value().(string)
|
||||
|
||||
// parse inherited type
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
declaration.what, err = parser.parseType()
|
||||
if err != nil { return }
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
if kind == lexer.TokenKindGreaterThan {
|
||||
behavior.inputs = append (
|
||||
behavior.inputs,
|
||||
declaration)
|
||||
} else {
|
||||
behavior.outputs = append (
|
||||
behavior.outputs,
|
||||
declaration)
|
||||
}
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
30
parser/func.go
Normal file
30
parser/func.go
Normal file
@@ -0,0 +1,30 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
// import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseFunc parses a function section.
|
||||
func (parser *ParsingOperation) parseFuncSection () (
|
||||
section *FuncSection,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
|
||||
section = &FuncSection { location: parser.token.Location() }
|
||||
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
section.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.name = parser.token.Value().(string)
|
||||
|
||||
err = parser.nextToken(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
111
parser/func_test.go
Normal file
111
parser/func_test.go
Normal file
@@ -0,0 +1,111 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestFunc (test *testing.T) {
|
||||
checkTree ("../tests/parser/func",
|
||||
`:arf
|
||||
---
|
||||
func ro aBasicExternal
|
||||
> someInput:Int:mut
|
||||
< someOutput:Int 4
|
||||
---
|
||||
external
|
||||
func ro bMethod
|
||||
@ bird:{Bird}
|
||||
> someInput:Int:mut
|
||||
< someOutput:Int 4
|
||||
---
|
||||
external
|
||||
func ro cBasicPhrases
|
||||
---
|
||||
[fn 329 983 09]
|
||||
[fn 329 983 09]
|
||||
[fn 329 983 091]
|
||||
[fn [gn 329 983 091] 123]
|
||||
func ro dArgumentTypes
|
||||
---
|
||||
[bird tree butterfly.wing "hello world" grass:{Int:mut 8}]
|
||||
func ro eMath
|
||||
[> x:Int]
|
||||
[> y:Int]
|
||||
[< z:Int]
|
||||
[---]
|
||||
[++ x]
|
||||
[-- y]
|
||||
[set z [+ [* 0392 00] 98 x [/ 9832 y] 930]]
|
||||
[! true]
|
||||
[~ 0b01]
|
||||
[% 873 32]
|
||||
[= 5 5]
|
||||
[!= 4 4]
|
||||
[<= 4 98]
|
||||
[< 4 98]
|
||||
[<< 0x0F 4]
|
||||
[>= 98 4]
|
||||
[> 98 4]
|
||||
[>> 0xF0 4]
|
||||
[| 0b01 0b10]
|
||||
[& 0b110 0b011]
|
||||
[&& true true]
|
||||
[|| true false]
|
||||
func ro fReturnDirection
|
||||
< err:Error
|
||||
---
|
||||
[someFunc 498 2980 90] -> thing:Int err
|
||||
[otherFunc] -> thing err:Error
|
||||
[fn 329 983 091] -> thing:Int err
|
||||
func ro gControlFlow
|
||||
---
|
||||
[if condition]
|
||||
[something]
|
||||
[if condition]
|
||||
[something]
|
||||
[elseif]
|
||||
[otherThing]
|
||||
[else]
|
||||
[finalThing]
|
||||
[while [< x 432]]
|
||||
[something]
|
||||
[switch value]
|
||||
[: 324]
|
||||
[something]
|
||||
[: 93284]
|
||||
otherThing
|
||||
[: 9128 34738 7328]
|
||||
multipleCases
|
||||
[:]
|
||||
[defaultThing]
|
||||
[for index:Size element:Int someArray]
|
||||
[something]
|
||||
[someNextThing]
|
||||
[justMakingSureBlockParsingWorks]
|
||||
[if condition]
|
||||
[if condition]
|
||||
[nestedThing]
|
||||
[else]
|
||||
[otherThing]
|
||||
[else]
|
||||
[if condition]
|
||||
[nestedThing]
|
||||
[else]
|
||||
[otherThing]
|
||||
func hSetPhrase
|
||||
---
|
||||
[set x:Int 3]
|
||||
[set y:{Int} [. x]]
|
||||
[set z:{Int 8}]
|
||||
398
|
||||
9
|
||||
2309
|
||||
983
|
||||
-2387
|
||||
478
|
||||
555
|
||||
123
|
||||
[set bird:Bird]
|
||||
.that
|
||||
.whenYou 99999
|
||||
.this 324
|
||||
`, test)
|
||||
}
|
||||
@@ -1,7 +1,7 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseMeta parsese the metadata header at the top of an arf file.
|
||||
func (parser *ParsingOperation) parseMeta () (err error) {
|
||||
@@ -35,7 +35,7 @@ func (parser *ParsingOperation) parseMeta () (err error) {
|
||||
default:
|
||||
parser.token.NewError (
|
||||
"unrecognized metadata field: " + field,
|
||||
file.ErrorKindError)
|
||||
infoerr.ErrorKindError)
|
||||
}
|
||||
|
||||
err = parser.nextToken(lexer.TokenKindNewline)
|
||||
|
||||
14
parser/meta_test.go
Normal file
14
parser/meta_test.go
Normal file
@@ -0,0 +1,14 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestMeta (test *testing.T) {
|
||||
checkTree ("../tests/parser/meta",
|
||||
`:arf
|
||||
author "Sasha Koshka"
|
||||
license "GPLv3"
|
||||
require "someModule"
|
||||
require "otherModule"
|
||||
---
|
||||
`, test)
|
||||
}
|
||||
127
parser/objt.go
Normal file
127
parser/objt.go
Normal file
@@ -0,0 +1,127 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseObjtSection parses an object type definition. This allows for structured
|
||||
// types to be defined, and for member variables to be added and overridden.
|
||||
func (parser *ParsingOperation) parseObjtSection () (
|
||||
section *ObjtSection,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
|
||||
section = &ObjtSection { location: parser.token.Location() }
|
||||
|
||||
// get permission
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
section.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.name = parser.token.Value().(string)
|
||||
|
||||
// parse inherited type
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
section.inherits, err = parser.parseIdentifier()
|
||||
if err != nil { return }
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
// parse members
|
||||
err = parser.parseObjtMembers(section)
|
||||
if err != nil { return }
|
||||
|
||||
if len(section.members) == 0 {
|
||||
infoerr.NewError (
|
||||
section.location,
|
||||
"defining an object with no members",
|
||||
infoerr.ErrorKindWarn).Print()
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
// parseObjtMembers parses a list of members for an object section. Indentation
|
||||
// level is assumed.
|
||||
func (parser *ParsingOperation) parseObjtMembers (
|
||||
into *ObjtSection,
|
||||
) (
|
||||
err error,
|
||||
) {
|
||||
for {
|
||||
// if we've left the block, stop parsing
|
||||
if !parser.token.Is(lexer.TokenKindIndent) { return }
|
||||
if parser.token.Value().(int) != 1 { return }
|
||||
|
||||
// add member to object section
|
||||
var member ObjtMember
|
||||
member, err = parser.parseObjtMember()
|
||||
into.members = append(into.members, member)
|
||||
if err != nil { return }
|
||||
}
|
||||
}
|
||||
|
||||
// parseObjtMember parses a single member of an object section. Indentation
|
||||
// level is assumed.
|
||||
func (parser *ParsingOperation) parseObjtMember () (
|
||||
member ObjtMember,
|
||||
err error,
|
||||
) {
|
||||
// get permission
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
member.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
member.name = parser.token.Value().(string)
|
||||
|
||||
// get type
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
member.what, err = parser.parseType()
|
||||
if err != nil { return }
|
||||
|
||||
println(parser.token.Describe())
|
||||
|
||||
// if there is a bit width, get it
|
||||
if parser.token.Is(lexer.TokenKindBinaryAnd) {
|
||||
err = parser.nextToken(lexer.TokenKindUInt)
|
||||
if err != nil { return }
|
||||
member.bitWidth = parser.token.Value().(uint64)
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
// parse default value
|
||||
if parser.token.Is(lexer.TokenKindNewline) {
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
member.defaultValue,
|
||||
err = parser.parseInitializationValues(1)
|
||||
if err != nil { return }
|
||||
} else {
|
||||
member.defaultValue, err = parser.parseArgument()
|
||||
if err != nil { return }
|
||||
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
31
parser/objt_test.go
Normal file
31
parser/objt_test.go
Normal file
@@ -0,0 +1,31 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestObjt (test *testing.T) {
|
||||
checkTree ("../tests/parser/objt",
|
||||
`:arf
|
||||
---
|
||||
objt ro Basic:Obj
|
||||
ro that:Basic
|
||||
ro this:Basic
|
||||
objt ro BitFields:Obj
|
||||
ro that:Int & 1
|
||||
ro this:Int & 24 298
|
||||
objt ro ComplexInit:Obj
|
||||
ro whatever:{Int 3}
|
||||
230984
|
||||
849
|
||||
394580
|
||||
ro complex0:Bird
|
||||
.that 98
|
||||
.this 2
|
||||
ro complex1:Bird
|
||||
.that 98902
|
||||
.this 235
|
||||
ro basic:Int 87
|
||||
objt ro Init:Obj
|
||||
ro that:String "hello world"
|
||||
ro this:Int 23
|
||||
`, test)
|
||||
}
|
||||
@@ -5,6 +5,7 @@ import "os"
|
||||
import "path/filepath"
|
||||
import "git.tebibyte.media/sashakoshka/arf/file"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// ParsingOperation holds information about an ongoing parsing operation.
|
||||
type ParsingOperation struct {
|
||||
@@ -96,9 +97,9 @@ func (parser *ParsingOperation) expect (allowed ...lexer.TokenKind) (err error)
|
||||
message += allowedItem.Describe()
|
||||
}
|
||||
|
||||
err = file.NewError (
|
||||
err = infoerr.NewError (
|
||||
parser.token.Location(),
|
||||
message, file.ErrorKindError)
|
||||
message, infoerr.ErrorKindError)
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
@@ -1,6 +1,7 @@
|
||||
package parser
|
||||
|
||||
import "io"
|
||||
import "strings"
|
||||
import "testing"
|
||||
// import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
|
||||
@@ -10,9 +11,9 @@ func checkTree (modulePath string, correct string, test *testing.T) {
|
||||
treeRunes := []rune(treeString)
|
||||
|
||||
test.Log("CORRECT TREE:")
|
||||
test.Log(correct)
|
||||
logWithLineNumbers(correct, test)
|
||||
test.Log("WHAT WAS PARSED:")
|
||||
test.Log(treeString)
|
||||
logWithLineNumbers(treeString, test)
|
||||
|
||||
if err != io.EOF && err != nil {
|
||||
test.Log("returned error:")
|
||||
@@ -63,49 +64,11 @@ func checkTree (modulePath string, correct string, test *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
func TestMeta (test *testing.T) {
|
||||
checkTree ("../tests/parser/meta",
|
||||
`:arf
|
||||
author "Sasha Koshka"
|
||||
license "GPLv3"
|
||||
require "someModule"
|
||||
require "otherModule"
|
||||
---
|
||||
`, test)
|
||||
}
|
||||
func logWithLineNumbers (bigString string, test *testing.T) {
|
||||
lines := strings.Split (
|
||||
strings.Replace(bigString, "\t", " ", -1), "\n")
|
||||
|
||||
func TestData (test *testing.T) {
|
||||
checkTree ("../tests/parser/data",
|
||||
`:arf
|
||||
---
|
||||
data wr integer:Int 3202
|
||||
data wr integerArray16:{Int 16}
|
||||
data wr integerArrayInitialized:{Int 16}
|
||||
3948
|
||||
293
|
||||
293049
|
||||
948
|
||||
912
|
||||
340
|
||||
0
|
||||
2304
|
||||
0
|
||||
4785
|
||||
92
|
||||
data wr integerArrayVariable:{Int ..}
|
||||
data wr integerPointer:{Int}
|
||||
data wr mutInteger:Int:mut 3202
|
||||
data wr mutIntegerPointer:{Int}:mut
|
||||
data wr nestedObject:Obj
|
||||
.that
|
||||
.bird2 123.8439
|
||||
.bird3 9328.21348239
|
||||
.this
|
||||
.bird0 324
|
||||
.bird1 "hello world"
|
||||
data wr object:Obj
|
||||
.that 2139
|
||||
.this 324
|
||||
`, test)
|
||||
for index, line := range lines {
|
||||
test.Logf("%3d | %s", index + 1, line)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -46,6 +46,26 @@ func (tree *SyntaxTree) ToString (indent int) (output string) {
|
||||
|
||||
output += doIndent(indent, "---\n")
|
||||
|
||||
typeSectionKeys := sortMapKeysAlphabetically(tree.typeSections)
|
||||
for _, name := range typeSectionKeys {
|
||||
output += tree.typeSections[name].ToString(indent)
|
||||
}
|
||||
|
||||
objtSectionKeys := sortMapKeysAlphabetically(tree.objtSections)
|
||||
for _, name := range objtSectionKeys {
|
||||
output += tree.objtSections[name].ToString(indent)
|
||||
}
|
||||
|
||||
enumSectionKeys := sortMapKeysAlphabetically(tree.enumSections)
|
||||
for _, name := range enumSectionKeys {
|
||||
output += tree.enumSections[name].ToString(indent)
|
||||
}
|
||||
|
||||
faceSectionKeys := sortMapKeysAlphabetically(tree.faceSections)
|
||||
for _, name := range faceSectionKeys {
|
||||
output += tree.faceSections[name].ToString(indent)
|
||||
}
|
||||
|
||||
dataSectionKeys := sortMapKeysAlphabetically(tree.dataSections)
|
||||
for _, name := range dataSectionKeys {
|
||||
output += tree.dataSections[name].ToString(indent)
|
||||
@@ -246,3 +266,126 @@ func (section *DataSection) ToString (indent int) (output string) {
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (section *TypeSection) ToString (indent int) (output string) {
|
||||
output += doIndent (
|
||||
indent,
|
||||
"type ",
|
||||
section.permission.ToString(), " ",
|
||||
section.name, ":",
|
||||
section.inherits.ToString())
|
||||
|
||||
isComplexInitialization :=
|
||||
section.defaultValue.kind == ArgumentKindObjectInitializationValues ||
|
||||
section.defaultValue.kind == ArgumentKindArrayInitializationValues
|
||||
|
||||
if section.defaultValue.value == nil {
|
||||
output += "\n"
|
||||
} else if isComplexInitialization {
|
||||
output += "\n"
|
||||
output += section.defaultValue.ToString(indent + 1, true)
|
||||
} else {
|
||||
output += " " + section.defaultValue.ToString(0, false)
|
||||
output += "\n"
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (member ObjtMember) ToString (indent int) (output string) {
|
||||
output += doIndent(indent)
|
||||
|
||||
output += member.permission.ToString() + " "
|
||||
output += member.name + ":"
|
||||
output += member.what.ToString()
|
||||
|
||||
if member.bitWidth > 0 {
|
||||
output += fmt.Sprint(" & ", member.bitWidth)
|
||||
}
|
||||
|
||||
isComplexInitialization :=
|
||||
member.defaultValue.kind == ArgumentKindObjectInitializationValues ||
|
||||
member.defaultValue.kind == ArgumentKindArrayInitializationValues
|
||||
|
||||
if member.defaultValue.value == nil {
|
||||
output += "\n"
|
||||
} else if isComplexInitialization {
|
||||
output += "\n"
|
||||
output += member.defaultValue.ToString(indent + 1, true)
|
||||
} else {
|
||||
output += " " + member.defaultValue.ToString(0, false)
|
||||
output += "\n"
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func (section *ObjtSection) ToString (indent int) (output string) {
|
||||
output += doIndent (
|
||||
indent,
|
||||
"objt ",
|
||||
section.permission.ToString(), " ",
|
||||
section.name, ":",
|
||||
section.inherits.ToString(), "\n")
|
||||
|
||||
for _, member := range section.members {
|
||||
output += member.ToString(indent + 1)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (section *EnumSection) ToString (indent int) (output string) {
|
||||
output += doIndent (
|
||||
indent,
|
||||
"enum ",
|
||||
section.permission.ToString(), " ",
|
||||
section.name, ":",
|
||||
section.what.ToString(), "\n")
|
||||
|
||||
for _, member := range section.members {
|
||||
output += doIndent(indent + 1, member.name)
|
||||
|
||||
isComplexInitialization :=
|
||||
member.value.kind == ArgumentKindObjectInitializationValues ||
|
||||
member.value.kind == ArgumentKindArrayInitializationValues
|
||||
|
||||
if member.value.value == nil {
|
||||
output += "\n"
|
||||
} else if isComplexInitialization {
|
||||
output += "\n"
|
||||
output += member.value.ToString(indent + 2, true)
|
||||
} else {
|
||||
output += " " + member.value.ToString(0, false)
|
||||
output += "\n"
|
||||
}
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (section *FaceSection) ToString (indent int) (output string) {
|
||||
output += doIndent (
|
||||
indent,
|
||||
"face ",
|
||||
section.permission.ToString(), " ",
|
||||
section.name, ":",
|
||||
section.inherits.ToString(), "\n")
|
||||
|
||||
for _, name := range sortMapKeysAlphabetically(section.behaviors) {
|
||||
behavior := section.behaviors[name]
|
||||
output += behavior.ToString(indent + 1)
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
func (behavior *FaceBehavior) ToString (indent int) (output string) {
|
||||
output += doIndent(indent, behavior.name, "\n")
|
||||
|
||||
for _, inputItem := range behavior.inputs {
|
||||
output += doIndent(indent + 1, "> ", inputItem.ToString(), "\n")
|
||||
}
|
||||
|
||||
for _, outputItem := range behavior.outputs {
|
||||
output += doIndent(indent + 1, "< ", outputItem.ToString(), "\n")
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
@@ -11,7 +11,12 @@ type SyntaxTree struct {
|
||||
author string
|
||||
|
||||
requires []string
|
||||
typeSections map[string] *TypeSection
|
||||
objtSections map[string] *ObjtSection
|
||||
enumSections map[string] *EnumSection
|
||||
faceSections map[string] *FaceSection
|
||||
dataSections map[string] *DataSection
|
||||
funcSections map[string] *FuncSection
|
||||
}
|
||||
|
||||
// Identifier represents a chain of arguments separated by a dot.
|
||||
@@ -156,6 +161,89 @@ type DataSection struct {
|
||||
name string
|
||||
|
||||
what Type
|
||||
value Argument
|
||||
permission types.Permission
|
||||
value Argument
|
||||
}
|
||||
|
||||
// TypeSection represents a blind type definition.
|
||||
type TypeSection struct {
|
||||
location file.Location
|
||||
name string
|
||||
|
||||
inherits Type
|
||||
permission types.Permission
|
||||
defaultValue Argument
|
||||
}
|
||||
|
||||
// ObjtMember represents a part of an object type definition.
|
||||
type ObjtMember struct {
|
||||
location file.Location
|
||||
name string
|
||||
|
||||
what Type
|
||||
bitWidth uint64
|
||||
permission types.Permission
|
||||
defaultValue Argument
|
||||
}
|
||||
|
||||
// ObjtSection represents an object type definition
|
||||
type ObjtSection struct {
|
||||
location file.Location
|
||||
name string
|
||||
|
||||
inherits Identifier
|
||||
permission types.Permission
|
||||
members []ObjtMember
|
||||
}
|
||||
|
||||
type EnumMember struct {
|
||||
location file.Location
|
||||
name string
|
||||
value Argument
|
||||
}
|
||||
|
||||
// EnumSection represents an enumerated type section.
|
||||
type EnumSection struct {
|
||||
location file.Location
|
||||
name string
|
||||
|
||||
what Type
|
||||
permission types.Permission
|
||||
members []EnumMember
|
||||
}
|
||||
|
||||
// FaceBehavior represents a behavior of an interface section.
|
||||
type FaceBehavior struct {
|
||||
location file.Location
|
||||
name string
|
||||
|
||||
inputs []Declaration
|
||||
outputs []Declaration
|
||||
}
|
||||
|
||||
// FaceSection represents an interface type section.
|
||||
type FaceSection struct {
|
||||
location file.Location
|
||||
name string
|
||||
inherits Identifier
|
||||
|
||||
permission types.Permission
|
||||
behaviors map[string] FaceBehavior
|
||||
}
|
||||
|
||||
// Block represents a scoped/indented block of code.
|
||||
// TODO: blocks will not directly nest. nested blocks will be stored as a part
|
||||
// of certain control flow statements.
|
||||
type Block []Phrase
|
||||
|
||||
// FuncSection represents a function section.
|
||||
type FuncSection struct {
|
||||
location file.Location
|
||||
name string
|
||||
permission types.Permission
|
||||
|
||||
receiver *Declaration
|
||||
inputs []Declaration
|
||||
outputs []Declaration
|
||||
root *Block
|
||||
}
|
||||
|
||||
53
parser/type.go
Normal file
53
parser/type.go
Normal file
@@ -0,0 +1,53 @@
|
||||
package parser
|
||||
|
||||
import "git.tebibyte.media/sashakoshka/arf/types"
|
||||
import "git.tebibyte.media/sashakoshka/arf/lexer"
|
||||
// import "git.tebibyte.media/sashakoshka/arf/infoerr"
|
||||
|
||||
// parseTypeSection parses a blind type definition, meaning it can inherit from
|
||||
// anything including primitives, but cannot define structure.
|
||||
func (parser *ParsingOperation) parseTypeSection () (
|
||||
section *TypeSection,
|
||||
err error,
|
||||
) {
|
||||
err = parser.expect(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
|
||||
section = &TypeSection { location: parser.token.Location() }
|
||||
|
||||
// get permission
|
||||
err = parser.nextToken(lexer.TokenKindPermission)
|
||||
if err != nil { return }
|
||||
section.permission = parser.token.Value().(types.Permission)
|
||||
|
||||
// get name
|
||||
err = parser.nextToken(lexer.TokenKindName)
|
||||
if err != nil { return }
|
||||
section.name = parser.token.Value().(string)
|
||||
|
||||
// parse inherited type
|
||||
err = parser.nextToken(lexer.TokenKindColon)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
section.inherits, err = parser.parseType()
|
||||
if err != nil { return }
|
||||
|
||||
// parse default values
|
||||
if parser.token.Is(lexer.TokenKindNewline) {
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
|
||||
section.defaultValue, err = parser.parseInitializationValues(0)
|
||||
if err != nil { return }
|
||||
} else {
|
||||
section.defaultValue, err = parser.parseArgument()
|
||||
if err != nil { return }
|
||||
|
||||
err = parser.expect(lexer.TokenKindNewline)
|
||||
if err != nil { return }
|
||||
err = parser.nextToken()
|
||||
if err != nil { return }
|
||||
}
|
||||
return
|
||||
}
|
||||
17
parser/type_test.go
Normal file
17
parser/type_test.go
Normal file
@@ -0,0 +1,17 @@
|
||||
package parser
|
||||
|
||||
import "testing"
|
||||
|
||||
func TestType (test *testing.T) {
|
||||
checkTree ("../tests/parser/type",
|
||||
`:arf
|
||||
---
|
||||
type ro Basic:Int
|
||||
type ro BasicInit:Int 6
|
||||
type ro IntArray:{Int ..}
|
||||
type ro IntArrayInit:{Int 3}
|
||||
3298
|
||||
923
|
||||
92
|
||||
`, test)
|
||||
}
|
||||
@@ -1,3 +1,3 @@
|
||||
:arf
|
||||
--- rw -> -349820394 932748397 239485.37520 "hello world!\n" 'E' helloWorld:.,..[]{}
|
||||
+ - ++ -- * / @ ! % ~ < << > >> | || & &&
|
||||
+ - ++ -- * / @ ! % %= ~ ~= = != < <= << <<= > >= >> >>= | |= || & &= && ^ ^=
|
||||
|
||||
2
tests/lexer/error/excessDataRune.arf
Normal file
2
tests/lexer/error/excessDataRune.arf
Normal file
@@ -0,0 +1,2 @@
|
||||
:arf
|
||||
'aaaaaaa'
|
||||
2
tests/lexer/error/unexpectedSymbol.arf
Normal file
2
tests/lexer/error/unexpectedSymbol.arf
Normal file
@@ -0,0 +1,2 @@
|
||||
:arf
|
||||
hello;
|
||||
2
tests/lexer/error/unknownEscape.arf
Normal file
2
tests/lexer/error/unknownEscape.arf
Normal file
@@ -0,0 +1,2 @@
|
||||
:arf
|
||||
"\g"
|
||||
@@ -1,19 +1,19 @@
|
||||
:arf
|
||||
---
|
||||
|
||||
data wr integer:Int 3202
|
||||
data ro integer:Int 3202
|
||||
|
||||
data wr mutInteger:Int:mut 3202
|
||||
data ro mutInteger:Int:mut 3202
|
||||
|
||||
data wr integerPointer:{Int}
|
||||
data ro integerPointer:{Int}
|
||||
|
||||
data wr mutIntegerPointer:{Int}:mut
|
||||
data ro mutIntegerPointer:{Int}:mut
|
||||
|
||||
data wr integerArray16:{Int 16}
|
||||
data ro integerArray16:{Int 16}
|
||||
|
||||
data wr integerArrayVariable:{Int ..}
|
||||
data ro integerArrayVariable:{Int ..}
|
||||
|
||||
data wr integerArrayInitialized:{Int 16}
|
||||
data ro integerArrayInitialized:{Int 16}
|
||||
3948 293 293049 948 912
|
||||
340 0 2304 0 4785 92
|
||||
|
||||
@@ -22,11 +22,13 @@ data wr integerArrayInitialized:{Int 16}
|
||||
|
||||
# data wr mutIntegerPointerInit:{Int}:mut [& integer]
|
||||
|
||||
data wr object:Obj
|
||||
# TODO: maybe test identifiers somewhere else?
|
||||
data ro object:thing.thing.
|
||||
thing.thing
|
||||
.this 324
|
||||
.that 2139
|
||||
|
||||
data wr nestedObject:Obj
|
||||
data ro nestedObject:Obj
|
||||
.this
|
||||
.bird0 324
|
||||
.bird1 "hello world"
|
||||
@@ -35,7 +37,7 @@ data wr nestedObject:Obj
|
||||
.bird3 9328.21348239
|
||||
|
||||
|
||||
# func rr main
|
||||
# func ro main
|
||||
# ---
|
||||
# # TODO: set should be a special case, checking under itself for object
|
||||
# member initialization args. it should also check for args in general
|
||||
|
||||
30
tests/parser/enum/main.arf
Normal file
30
tests/parser/enum/main.arf
Normal file
@@ -0,0 +1,30 @@
|
||||
:arf
|
||||
---
|
||||
|
||||
enum ro Weekday:Int
|
||||
sunday
|
||||
monday
|
||||
tuesday
|
||||
wednesday
|
||||
thursday
|
||||
friday
|
||||
saturday
|
||||
|
||||
enum ro NamedColor:U32
|
||||
red 0xFF0000
|
||||
green 0x00FF00
|
||||
blue 0x0000FF
|
||||
|
||||
enum ro AffrontToGod:{Int 4}
|
||||
bird0
|
||||
28394 9328
|
||||
398 9
|
||||
bird1
|
||||
23 932832
|
||||
398
|
||||
2349
|
||||
bird2
|
||||
1
|
||||
2
|
||||
3
|
||||
4
|
||||
15
tests/parser/face/main.arf
Normal file
15
tests/parser/face/main.arf
Normal file
@@ -0,0 +1,15 @@
|
||||
:arf
|
||||
---
|
||||
|
||||
face ro ReadWriter:Face
|
||||
write
|
||||
> data:{Byte ..}
|
||||
< wrote:Int
|
||||
< err:Error
|
||||
read
|
||||
> into:{Byte ..}
|
||||
< read:Int
|
||||
< err:Error
|
||||
|
||||
face ro Destroyer:Face
|
||||
destroy
|
||||
134
tests/parser/func/main.arf
Normal file
134
tests/parser/func/main.arf
Normal file
@@ -0,0 +1,134 @@
|
||||
:arf
|
||||
---
|
||||
func ro aBasicExternal
|
||||
> someInput:Int:mut
|
||||
< someOutput:Int 4
|
||||
---
|
||||
external
|
||||
|
||||
func ro bMethod
|
||||
@ bird:{Bird}
|
||||
> someInput:Int:mut
|
||||
< someOutput:Int 4
|
||||
---
|
||||
external
|
||||
|
||||
func ro cBasicPhrases
|
||||
---
|
||||
fn 329 983 09
|
||||
[fn 329 983 09]
|
||||
[fn
|
||||
329
|
||||
983
|
||||
091]
|
||||
fn [gn
|
||||
329 983
|
||||
091] 123
|
||||
|
||||
func ro dArgumentTypes
|
||||
---
|
||||
[bird tree butterfly.wing "hello world"
|
||||
grass:{Int:mut 8}]
|
||||
|
||||
func ro eMath
|
||||
> x:Int
|
||||
> y:Int
|
||||
< z:Int
|
||||
---
|
||||
++ x
|
||||
-- y
|
||||
set z [+ [* 0392 00] 98 x [/ 9832 y] 930]
|
||||
|
||||
# TODO: need tokens ~=
|
||||
|
||||
! true
|
||||
~ 0b01
|
||||
# ~= x
|
||||
% 873 32
|
||||
|
||||
= 5 5
|
||||
!= 4 4
|
||||
|
||||
<= 4 98
|
||||
< 4 98
|
||||
<< 0x0F 4
|
||||
# <<= x 4
|
||||
|
||||
>= 98 4
|
||||
> 98 4
|
||||
>> 0xF0 4
|
||||
# >>= x 4
|
||||
|
||||
| 0b01 0b10
|
||||
# |= x 0b10
|
||||
& 0b110 0b011
|
||||
# &= x 0b011
|
||||
|
||||
&& true true
|
||||
|| true false
|
||||
|
||||
func ro fReturnDirection
|
||||
< err:Error
|
||||
---
|
||||
someFunc 498 2980 90 -> thing:Int err
|
||||
otherFunc -> thing err:Error
|
||||
|
||||
[fn
|
||||
329
|
||||
983
|
||||
091] -> thing:Int err
|
||||
|
||||
func ro gControlFlow
|
||||
---
|
||||
if condition
|
||||
something
|
||||
|
||||
if condition
|
||||
something
|
||||
elseif
|
||||
[otherThing]
|
||||
else
|
||||
finalThing
|
||||
|
||||
while [< x 432]
|
||||
something
|
||||
|
||||
switch value
|
||||
: 324
|
||||
something
|
||||
[: 93284]
|
||||
otherThing
|
||||
: 9128 34738 7328
|
||||
multipleCases
|
||||
:
|
||||
[defaultThing]
|
||||
|
||||
for index:Size element:Int someArray
|
||||
something
|
||||
someNextThing
|
||||
justMakingSureBlockParsingWorks
|
||||
|
||||
[if condition]
|
||||
if condition
|
||||
nestedThing
|
||||
else
|
||||
otherThing
|
||||
else
|
||||
if condition
|
||||
nestedThing
|
||||
else
|
||||
otherThing
|
||||
|
||||
func hSetPhrase
|
||||
---
|
||||
set x:Int 3
|
||||
# TODO: this should be the "location of" phrase. update other things to
|
||||
# match.
|
||||
set y:{Int} [. x]
|
||||
set z:{Int 8}
|
||||
398 9 2309 983 -2387
|
||||
478 555 123
|
||||
set bird:Bird
|
||||
.that
|
||||
.whenYou 99999
|
||||
.this 324
|
||||
25
tests/parser/objt/main.arf
Normal file
25
tests/parser/objt/main.arf
Normal file
@@ -0,0 +1,25 @@
|
||||
:arf
|
||||
---
|
||||
objt ro Basic:Obj
|
||||
ro that:Basic
|
||||
ro this:Basic
|
||||
|
||||
objt ro BitFields:Obj
|
||||
ro that:Int & 1
|
||||
ro this:Int & 24 298
|
||||
|
||||
objt ro Init:Obj
|
||||
ro that:String "hello world"
|
||||
ro this:Int 23
|
||||
|
||||
objt ro ComplexInit:Obj
|
||||
ro whatever:{Int 3}
|
||||
230984
|
||||
849 394580
|
||||
ro complex0:Bird
|
||||
.that 98
|
||||
.this 2
|
||||
ro complex1:Bird
|
||||
.that 98902
|
||||
.this 235
|
||||
ro basic:Int 87
|
||||
10
tests/parser/type/main.arf
Normal file
10
tests/parser/type/main.arf
Normal file
@@ -0,0 +1,10 @@
|
||||
:arf
|
||||
---
|
||||
type ro Basic:Int
|
||||
|
||||
type ro BasicInit:Int 6
|
||||
|
||||
type ro IntArray:{Int ..}
|
||||
|
||||
type ro IntArrayInit:{Int 3}
|
||||
3298 923 92
|
||||
@@ -1,48 +1,63 @@
|
||||
package types
|
||||
|
||||
type Mode int
|
||||
type Permission int
|
||||
|
||||
const (
|
||||
ModeNone = iota
|
||||
ModeRead
|
||||
ModeWrite
|
||||
// Displays as: pv
|
||||
//
|
||||
// Other modules cannot access the section or member.
|
||||
PermissionPrivate Permission = iota
|
||||
|
||||
// Displays as: ro
|
||||
//
|
||||
// Other modules can access the section or member, but can only read its
|
||||
// value. It is effectively immutable.
|
||||
//
|
||||
// Data sections, member variables, etc: The value can be read by other
|
||||
// modules but not altered by them
|
||||
//
|
||||
// Functions: The function can be called by other modules.
|
||||
//
|
||||
// Methods: The method can be called by other modules, but cannot be
|
||||
// overriden by a type defined in another module inheriting from this
|
||||
// method's reciever.
|
||||
PermissionReadOnly
|
||||
|
||||
// Displays as: rw
|
||||
//
|
||||
// Other modules cannot only access the section or member's value, but
|
||||
// can alter it. It is effectively mutable.
|
||||
//
|
||||
// Data sections, member variables, etc: The value can be read and
|
||||
// altered at will by other modules.
|
||||
//
|
||||
// Functions: This permission cannot be applied to non-method functions.
|
||||
//
|
||||
// Methods: The method can be called by other modules, and overridden by
|
||||
// types defined in other modules inheriting from the method's reciever.
|
||||
PermissionReadWrite
|
||||
)
|
||||
|
||||
type Permission struct {
|
||||
Internal Mode
|
||||
External Mode
|
||||
}
|
||||
|
||||
func ModeFrom (char rune) (mode Mode) {
|
||||
switch (char) {
|
||||
case 'n': mode = ModeNone
|
||||
case 'r': mode = ModeRead
|
||||
case 'w': mode = ModeWrite
|
||||
// PermissionFrom creates a new permission value from the specified text. If the
|
||||
// input text was not valid, the function returns false for worked. Otherwise,
|
||||
// it returns true.
|
||||
func PermissionFrom (data string) (permission Permission, worked bool) {
|
||||
worked = true
|
||||
switch data {
|
||||
case "pv": permission = PermissionPrivate
|
||||
case "ro": permission = PermissionReadOnly
|
||||
case "rw": permission = PermissionReadWrite
|
||||
default: worked = false
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
func PermissionFrom (data string) (permission Permission) {
|
||||
if len(data) != 2 { return }
|
||||
|
||||
permission.Internal = ModeFrom(rune(data[0]))
|
||||
permission.External = ModeFrom(rune(data[1]))
|
||||
return
|
||||
}
|
||||
|
||||
func (mode Mode) ToString () (output string) {
|
||||
switch mode {
|
||||
case ModeNone: output = "n"
|
||||
case ModeRead: output = "r"
|
||||
case ModeWrite: output = "w"
|
||||
}
|
||||
|
||||
return
|
||||
}
|
||||
|
||||
// ToString converts the permission value into a string.
|
||||
func (permission Permission) ToString () (output string) {
|
||||
output += permission.Internal.ToString()
|
||||
output += permission.External.ToString()
|
||||
switch permission {
|
||||
case PermissionPrivate: output = "pv"
|
||||
case PermissionReadOnly: output = "ro"
|
||||
case PermissionReadWrite: output = "rw"
|
||||
}
|
||||
return
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user