diff --git a/arfc/main.go b/arfc/main.go index 99de746..42627a4 100644 --- a/arfc/main.go +++ b/arfc/main.go @@ -2,7 +2,7 @@ package arfc import "os" import "fmt" -import "github.com/sashakoshka/arf" +import "git.tebibyte.media/sashakoshka/arf" func main () { if len(os.Args) != 2 { diff --git a/go.mod b/go.mod index 2da2f4b..743d639 100644 --- a/go.mod +++ b/go.mod @@ -1,3 +1,3 @@ -module github.com/sashakoshka/arf +module git.tebibyte.media/sashakoshka/arf go 1.18 diff --git a/lexer/lexer.go b/lexer/lexer.go index 9703759..30445b6 100644 --- a/lexer/lexer.go +++ b/lexer/lexer.go @@ -1,8 +1,8 @@ package lexer import "io" -import "github.com/sashakoshka/arf/file" -import "github.com/sashakoshka/arf/types" +import "git.tebibyte.media/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/types" // LexingOperation holds information about an ongoing lexing operataion. type LexingOperation struct { diff --git a/lexer/lexer_test.go b/lexer/lexer_test.go index 742ea5b..352959c 100644 --- a/lexer/lexer_test.go +++ b/lexer/lexer_test.go @@ -1,8 +1,8 @@ package lexer import "testing" -import "github.com/sashakoshka/arf/file" -import "github.com/sashakoshka/arf/types" +import "git.tebibyte.media/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/types" func checkTokenSlice (filePath string, correct []Token, test *testing.T) { file, err := file.Open(filePath) diff --git a/lexer/numbers.go b/lexer/numbers.go index 5518f58..c96f1c8 100644 --- a/lexer/numbers.go +++ b/lexer/numbers.go @@ -1,6 +1,6 @@ package lexer -import "github.com/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/file" // tokenizeSymbolBeginning lexes a token that starts with a number. func (lexer *LexingOperation) tokenizeNumberBeginning (negative bool) (err error) { diff --git a/lexer/text.go b/lexer/text.go index 31f3e3a..9a2adb0 100644 --- a/lexer/text.go +++ b/lexer/text.go @@ -1,7 +1,7 @@ package lexer import "strconv" -import "github.com/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/file" // tokenizeString tokenizes a string or rune literal. func (lexer *LexingOperation) tokenizeString (isRuneLiteral bool) (err error) { diff --git a/lexer/token.go b/lexer/token.go index 0ac8298..6247031 100644 --- a/lexer/token.go +++ b/lexer/token.go @@ -1,7 +1,7 @@ package lexer import "fmt" -import "github.com/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/file" // TokenKind is an enum represzenting what role a token has. type TokenKind int diff --git a/main.go b/main.go index 1703906..3f4c7e1 100644 --- a/main.go +++ b/main.go @@ -1,7 +1,7 @@ package arf import "io" -import "github.com/sashakoshka/arf/parser" +import "git.tebibyte.media/sashakoshka/arf/parser" func CompileModule (modulePath string, output io.Writer) (err error) { _, err = parser.Parse(modulePath) diff --git a/parser/parser.go b/parser/parser.go index 3c877a0..e20254d 100644 --- a/parser/parser.go +++ b/parser/parser.go @@ -2,8 +2,8 @@ package parser import "os" import "path/filepath" -import "github.com/sashakoshka/arf/file" -import "github.com/sashakoshka/arf/lexer" +import "git.tebibyte.media/sashakoshka/arf/file" +import "git.tebibyte.media/sashakoshka/arf/lexer" // ParsingOperation holds information about an ongoing parsing operation. type ParsingOperation struct {