firefly-compiler 0.5.39 → 0.5.41
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.hintrc +4 -4
- package/.vscode/settings.json +4 -4
- package/bin/Release.ff +158 -157
- package/bin/firefly.mjs +1 -1
- package/compiler/Builder.ff +275 -275
- package/compiler/Compiler.ff +234 -234
- package/compiler/Dependencies.ff +186 -186
- package/compiler/DependencyLock.ff +17 -17
- package/compiler/JsEmitter.ff +1437 -1437
- package/compiler/LspHook.ff +202 -202
- package/compiler/ModuleCache.ff +178 -178
- package/compiler/Workspace.ff +88 -88
- package/core/.firefly/include/package.json +5 -5
- package/core/.firefly/package.ff +2 -2
- package/core/Any.ff +25 -25
- package/core/Array.ff +298 -298
- package/core/Atomic.ff +63 -63
- package/core/Box.ff +7 -7
- package/core/BrowserSystem.ff +40 -40
- package/core/BuildSystem.ff +156 -156
- package/core/Crypto.ff +94 -94
- package/core/Equal.ff +41 -41
- package/core/Error.ff +25 -25
- package/core/HttpClient.ff +142 -142
- package/core/Instant.ff +24 -24
- package/core/Js.ff +305 -305
- package/core/JsSystem.ff +135 -135
- package/core/Json.ff +423 -423
- package/core/List.ff +482 -482
- package/core/Lock.ff +108 -108
- package/core/NodeSystem.ff +198 -198
- package/core/Ordering.ff +160 -160
- package/core/Path.ff +377 -378
- package/core/Queue.ff +90 -90
- package/core/Random.ff +140 -140
- package/core/RbMap.ff +216 -216
- package/core/Show.ff +44 -44
- package/core/SourceLocation.ff +68 -68
- package/core/Task.ff +165 -165
- package/experimental/benchmarks/ListGrab.ff +23 -23
- package/experimental/benchmarks/ListGrab.java +55 -55
- package/experimental/benchmarks/Pyrotek45.ff +30 -30
- package/experimental/benchmarks/Pyrotek45.java +64 -64
- package/experimental/bidirectional/Bidi.ff +88 -88
- package/experimental/lines/Main.ff +40 -40
- package/experimental/random/Index.ff +53 -53
- package/experimental/random/Process.ff +120 -120
- package/experimental/random/RunLength.ff +65 -65
- package/experimental/random/Scrape.ff +51 -51
- package/experimental/random/Symbols.ff +73 -73
- package/experimental/random/Tensor.ff +52 -52
- package/experimental/random/Units.ff +36 -36
- package/experimental/s3/S3TestAuthorizationHeader.ff +39 -39
- package/experimental/s3/S3TestPut.ff +16 -16
- package/experimental/tests/TestJson.ff +26 -26
- package/firefly.sh +0 -0
- package/fireflysite/.firefly/package.ff +4 -4
- package/fireflysite/CommunityOverview.ff +20 -20
- package/fireflysite/CountingButtonDemo.ff +58 -58
- package/fireflysite/DocumentParser.ff +325 -325
- package/fireflysite/ExamplesOverview.ff +40 -40
- package/fireflysite/FrontPage.ff +344 -344
- package/fireflysite/GettingStarted.ff +45 -45
- package/fireflysite/Guide.ff +456 -456
- package/fireflysite/Main.ff +163 -163
- package/fireflysite/MatchingPasswordsDemo.ff +82 -82
- package/fireflysite/PackagesOverview.ff +49 -49
- package/fireflysite/PostgresqlDemo.ff +34 -34
- package/fireflysite/ReferenceAll.ff +18 -18
- package/fireflysite/ReferenceIntroduction.ff +11 -11
- package/fireflysite/Styles.ff +567 -567
- package/fireflysite/Test.ff +121 -121
- package/fireflysite/assets/markdown/reference/BaseTypes.md +209 -209
- package/fireflysite/assets/markdown/reference/EmittedJavascript.md +65 -65
- package/fireflysite/assets/markdown/reference/Exceptions.md +101 -101
- package/fireflysite/assets/markdown/reference/FunctionsAndMethods.md +364 -364
- package/fireflysite/assets/markdown/reference/JavascriptInterop.md +235 -235
- package/fireflysite/assets/markdown/reference/ModulesAndPackages.md +162 -162
- package/fireflysite/assets/markdown/reference/OldStructuredConcurrency.md +48 -48
- package/fireflysite/assets/markdown/reference/PatternMatching.md +224 -224
- package/fireflysite/assets/markdown/reference/StatementsAndExpressions.md +86 -86
- package/fireflysite/assets/markdown/reference/StructuredConcurrency.md +99 -99
- package/fireflysite/assets/markdown/reference/TraitsAndInstances.md +100 -100
- package/fireflysite/assets/markdown/reference/UserDefinedTypes.md +184 -184
- package/fireflysite/assets/markdown/scratch/ControlFlow.md +136 -136
- package/fireflysite/assets/markdown/scratch/Toc.md +40 -40
- package/lsp/.firefly/package.ff +1 -1
- package/lsp/CompletionHandler.ff +827 -827
- package/lsp/Handler.ff +714 -714
- package/lsp/HoverHandler.ff +79 -79
- package/lsp/LanguageServer.ff +272 -272
- package/lsp/SignatureHelpHandler.ff +55 -55
- package/lsp/SymbolHandler.ff +181 -181
- package/lsp/TestReferences.ff +17 -17
- package/lsp/TestReferencesCase.ff +7 -7
- package/lsp/stderr.txt +1 -1
- package/lsp/stdout.txt +34 -34
- package/lux/.firefly/package.ff +1 -1
- package/lux/Css.ff +648 -648
- package/lux/CssTest.ff +48 -48
- package/lux/Lux.ff +608 -608
- package/lux/LuxEvent.ff +79 -79
- package/lux/Main.ff +123 -123
- package/lux/Main2.ff +143 -143
- package/lux/TestDry.ff +28 -28
- package/output/js/ff/compiler/Builder.mjs +36 -36
- package/output/js/ff/core/Path.mjs +0 -2
- package/package.json +1 -1
- package/rpc/.firefly/package.ff +1 -1
- package/rpc/Rpc.ff +70 -70
- package/s3/.firefly/package.ff +1 -1
- package/s3/S3.ff +92 -92
- package/vscode/LICENSE.txt +21 -21
- package/vscode/Prepublish.ff +15 -15
- package/vscode/README.md +16 -16
- package/vscode/client/package-lock.json +544 -544
- package/vscode/client/package.json +22 -22
- package/vscode/client/src/extension.ts +104 -104
- package/vscode/icons/firefly-icon.svg +10 -10
- package/vscode/language-configuration.json +61 -61
- package/vscode/package-lock.json +3623 -3623
- package/vscode/package.json +1 -1
- package/vscode/snippets.json +241 -241
- package/vscode/syntaxes/firefly-markdown-injection.json +45 -45
- package/webserver/.firefly/include/package.json +5 -5
- package/webserver/.firefly/package.ff +2 -2
- package/webserver/WebServer.ff +647 -647
- package/websocket/.firefly/package.ff +1 -1
- package/websocket/WebSocket.ff +100 -100
|
@@ -1,65 +1,65 @@
|
|
|
1
|
-
main(system: NodeSystem) {
|
|
2
|
-
let inBuffers = 1.to(100).map {_ => Buffer.fromByteList([5, 5, 3, 1, 2].toArray())}
|
|
3
|
-
let inStream = inBuffers.toStream()
|
|
4
|
-
let outStream = toRunLength(inStream, 3)
|
|
5
|
-
printRunLength(outStream)
|
|
6
|
-
}
|
|
7
|
-
|
|
8
|
-
printRunLength(outStream: Stream[Buffer]) {
|
|
9
|
-
let buffer = Buffer.fromBufferList(outStream.toArray())
|
|
10
|
-
Log.debug(buffer.toHex())
|
|
11
|
-
let stack = Stack.make()
|
|
12
|
-
mutable i = 0
|
|
13
|
-
while {i < buffer.size()} {
|
|
14
|
-
1.to(buffer.grabUint8(i)).each {_ => stack.push(buffer.grabUint8(i + 1))}
|
|
15
|
-
i += 2
|
|
16
|
-
}
|
|
17
|
-
Log.debug(Buffer.fromByteList(stack.drain()).toHex())
|
|
18
|
-
}
|
|
19
|
-
|
|
20
|
-
toRunLength(stream: Stream[Buffer], bufferSize: Int = 65536): Stream[Buffer] {
|
|
21
|
-
mutable outBuffer = Buffer.new(bufferSize)
|
|
22
|
-
mutable outOffset = 0
|
|
23
|
-
let outBuffers = Stack.make()
|
|
24
|
-
function writeByte(byte: Int): Unit {
|
|
25
|
-
if(outOffset >= outBuffer.size()) {
|
|
26
|
-
outBuffers.push(outBuffer)
|
|
27
|
-
outBuffer = Buffer.new(bufferSize)
|
|
28
|
-
outOffset = 0
|
|
29
|
-
}
|
|
30
|
-
outBuffer.setUint8(outOffset, byte)
|
|
31
|
-
outOffset += 1
|
|
32
|
-
}
|
|
33
|
-
mutable extraCount = 0
|
|
34
|
-
mutable value = 0
|
|
35
|
-
let result = stream.flatMap {inBuffer =>
|
|
36
|
-
mutable i = 0
|
|
37
|
-
while {i < inBuffer.size()} {
|
|
38
|
-
value = if(extraCount > 0) {value} else {inBuffer.grabUint8(i)}
|
|
39
|
-
mutable j = if(extraCount > 0) {-1} else {0}
|
|
40
|
-
doWhile {
|
|
41
|
-
j += 1
|
|
42
|
-
j + extraCount < 256 && i + j < inBuffer.size() && inBuffer.grabUint8(i + j) == value
|
|
43
|
-
}
|
|
44
|
-
i += j
|
|
45
|
-
if(i < inBuffer.size() || j + extraCount == 256) {
|
|
46
|
-
writeByte(j + extraCount)
|
|
47
|
-
writeByte(value)
|
|
48
|
-
extraCount = 0
|
|
49
|
-
} else {
|
|
50
|
-
extraCount += j
|
|
51
|
-
}
|
|
52
|
-
}
|
|
53
|
-
outBuffers.drain().toStream()
|
|
54
|
-
}
|
|
55
|
-
result.addAll(Stream.do {
|
|
56
|
-
if(extraCount > 0) {
|
|
57
|
-
writeByte(extraCount)
|
|
58
|
-
writeByte(value)
|
|
59
|
-
}
|
|
60
|
-
if(outOffset > 0) {
|
|
61
|
-
outBuffers.push(outBuffer.view(0, outOffset))
|
|
62
|
-
}
|
|
63
|
-
outBuffers.drain().toStream()
|
|
64
|
-
})
|
|
65
|
-
}
|
|
1
|
+
main(system: NodeSystem) {
|
|
2
|
+
let inBuffers = 1.to(100).map {_ => Buffer.fromByteList([5, 5, 3, 1, 2].toArray())}
|
|
3
|
+
let inStream = inBuffers.toStream()
|
|
4
|
+
let outStream = toRunLength(inStream, 3)
|
|
5
|
+
printRunLength(outStream)
|
|
6
|
+
}
|
|
7
|
+
|
|
8
|
+
printRunLength(outStream: Stream[Buffer]) {
|
|
9
|
+
let buffer = Buffer.fromBufferList(outStream.toArray())
|
|
10
|
+
Log.debug(buffer.toHex())
|
|
11
|
+
let stack = Stack.make()
|
|
12
|
+
mutable i = 0
|
|
13
|
+
while {i < buffer.size()} {
|
|
14
|
+
1.to(buffer.grabUint8(i)).each {_ => stack.push(buffer.grabUint8(i + 1))}
|
|
15
|
+
i += 2
|
|
16
|
+
}
|
|
17
|
+
Log.debug(Buffer.fromByteList(stack.drain()).toHex())
|
|
18
|
+
}
|
|
19
|
+
|
|
20
|
+
toRunLength(stream: Stream[Buffer], bufferSize: Int = 65536): Stream[Buffer] {
|
|
21
|
+
mutable outBuffer = Buffer.new(bufferSize)
|
|
22
|
+
mutable outOffset = 0
|
|
23
|
+
let outBuffers = Stack.make()
|
|
24
|
+
function writeByte(byte: Int): Unit {
|
|
25
|
+
if(outOffset >= outBuffer.size()) {
|
|
26
|
+
outBuffers.push(outBuffer)
|
|
27
|
+
outBuffer = Buffer.new(bufferSize)
|
|
28
|
+
outOffset = 0
|
|
29
|
+
}
|
|
30
|
+
outBuffer.setUint8(outOffset, byte)
|
|
31
|
+
outOffset += 1
|
|
32
|
+
}
|
|
33
|
+
mutable extraCount = 0
|
|
34
|
+
mutable value = 0
|
|
35
|
+
let result = stream.flatMap {inBuffer =>
|
|
36
|
+
mutable i = 0
|
|
37
|
+
while {i < inBuffer.size()} {
|
|
38
|
+
value = if(extraCount > 0) {value} else {inBuffer.grabUint8(i)}
|
|
39
|
+
mutable j = if(extraCount > 0) {-1} else {0}
|
|
40
|
+
doWhile {
|
|
41
|
+
j += 1
|
|
42
|
+
j + extraCount < 256 && i + j < inBuffer.size() && inBuffer.grabUint8(i + j) == value
|
|
43
|
+
}
|
|
44
|
+
i += j
|
|
45
|
+
if(i < inBuffer.size() || j + extraCount == 256) {
|
|
46
|
+
writeByte(j + extraCount)
|
|
47
|
+
writeByte(value)
|
|
48
|
+
extraCount = 0
|
|
49
|
+
} else {
|
|
50
|
+
extraCount += j
|
|
51
|
+
}
|
|
52
|
+
}
|
|
53
|
+
outBuffers.drain().toStream()
|
|
54
|
+
}
|
|
55
|
+
result.addAll(Stream.do {
|
|
56
|
+
if(extraCount > 0) {
|
|
57
|
+
writeByte(extraCount)
|
|
58
|
+
writeByte(value)
|
|
59
|
+
}
|
|
60
|
+
if(outOffset > 0) {
|
|
61
|
+
outBuffers.push(outBuffer.view(0, outOffset))
|
|
62
|
+
}
|
|
63
|
+
outBuffers.drain().toStream()
|
|
64
|
+
})
|
|
65
|
+
}
|
|
@@ -1,51 +1,51 @@
|
|
|
1
|
-
|
|
2
|
-
nodeMain(system: NodeSystem) {
|
|
3
|
-
|
|
4
|
-
let urlFile = system.arguments().grab(0)
|
|
5
|
-
let urls = system.path(urlFile).readText().lines()
|
|
6
|
-
|
|
7
|
-
let urlChannel = system.mainTask().channel(urls.size())
|
|
8
|
-
let htmlChannel = system.mainTask().channel(urls.size())
|
|
9
|
-
let resultChannel = system.mainTask().channel(urls.size())
|
|
10
|
-
|
|
11
|
-
1.to(1000).each {_ =>
|
|
12
|
-
system.mainTask().spawn {task =>
|
|
13
|
-
let url = urlChannel.read()
|
|
14
|
-
let html = system.httpClient().get(url, []) {_.readText()}
|
|
15
|
-
htmlChannel.write(html)
|
|
16
|
-
}
|
|
17
|
-
}
|
|
18
|
-
|
|
19
|
-
1.to(10).each {_ =>
|
|
20
|
-
system.mainTask().spawn {task =>
|
|
21
|
-
let html = htmlChannel.read()
|
|
22
|
-
let result = scrape(html)
|
|
23
|
-
resultChannel.write(result)
|
|
24
|
-
}
|
|
25
|
-
}
|
|
26
|
-
|
|
27
|
-
urls.each {url => urlChannel.write(url)}
|
|
28
|
-
let results = urls.map {_ => resultChannel.read()}
|
|
29
|
-
system.writeText(results.join("\n"))
|
|
30
|
-
|
|
31
|
-
}
|
|
32
|
-
|
|
33
|
-
scrape(html: String): String {
|
|
34
|
-
html // Uhm...
|
|
35
|
-
}
|
|
36
|
-
|
|
37
|
-
foo() {
|
|
38
|
-
|
|
39
|
-
let result = urls.toStream()
|
|
40
|
-
.parallelMap(threads = 1000) {url => download(url)}
|
|
41
|
-
.parallelMap(threads = 10) {result => result.map(scrape)}
|
|
42
|
-
|
|
43
|
-
let result = urls.toStream()
|
|
44
|
-
.parallelMap(threads = 1000) {task, url => task.timeout(Duration(5)) {download(url)}}
|
|
45
|
-
.parallelMap(threads = 10) {_, result => result.map(scrape)}
|
|
46
|
-
|
|
47
|
-
let result = urls.toStream()
|
|
48
|
-
.parallelMap(1000) {task, url => task.deadline(task.now().add(Duration(5))) {download(url)}}
|
|
49
|
-
.parallelMap(10) {_, result => result.map(scrape)}
|
|
50
|
-
|
|
51
|
-
}
|
|
1
|
+
|
|
2
|
+
nodeMain(system: NodeSystem) {
|
|
3
|
+
|
|
4
|
+
let urlFile = system.arguments().grab(0)
|
|
5
|
+
let urls = system.path(urlFile).readText().lines()
|
|
6
|
+
|
|
7
|
+
let urlChannel = system.mainTask().channel(urls.size())
|
|
8
|
+
let htmlChannel = system.mainTask().channel(urls.size())
|
|
9
|
+
let resultChannel = system.mainTask().channel(urls.size())
|
|
10
|
+
|
|
11
|
+
1.to(1000).each {_ =>
|
|
12
|
+
system.mainTask().spawn {task =>
|
|
13
|
+
let url = urlChannel.read()
|
|
14
|
+
let html = system.httpClient().get(url, []) {_.readText()}
|
|
15
|
+
htmlChannel.write(html)
|
|
16
|
+
}
|
|
17
|
+
}
|
|
18
|
+
|
|
19
|
+
1.to(10).each {_ =>
|
|
20
|
+
system.mainTask().spawn {task =>
|
|
21
|
+
let html = htmlChannel.read()
|
|
22
|
+
let result = scrape(html)
|
|
23
|
+
resultChannel.write(result)
|
|
24
|
+
}
|
|
25
|
+
}
|
|
26
|
+
|
|
27
|
+
urls.each {url => urlChannel.write(url)}
|
|
28
|
+
let results = urls.map {_ => resultChannel.read()}
|
|
29
|
+
system.writeText(results.join("\n"))
|
|
30
|
+
|
|
31
|
+
}
|
|
32
|
+
|
|
33
|
+
scrape(html: String): String {
|
|
34
|
+
html // Uhm...
|
|
35
|
+
}
|
|
36
|
+
|
|
37
|
+
foo() {
|
|
38
|
+
|
|
39
|
+
let result = urls.toStream()
|
|
40
|
+
.parallelMap(threads = 1000) {url => download(url)}
|
|
41
|
+
.parallelMap(threads = 10) {result => result.map(scrape)}
|
|
42
|
+
|
|
43
|
+
let result = urls.toStream()
|
|
44
|
+
.parallelMap(threads = 1000) {task, url => task.timeout(Duration(5)) {download(url)}}
|
|
45
|
+
.parallelMap(threads = 10) {_, result => result.map(scrape)}
|
|
46
|
+
|
|
47
|
+
let result = urls.toStream()
|
|
48
|
+
.parallelMap(1000) {task, url => task.deadline(task.now().add(Duration(5))) {download(url)}}
|
|
49
|
+
.parallelMap(10) {_, result => result.map(scrape)}
|
|
50
|
+
|
|
51
|
+
}
|
|
@@ -1,74 +1,74 @@
|
|
|
1
|
-
nodeMain(system: NodeSystem) {
|
|
2
|
-
/*
|
|
3
|
-
Log.show(splitCamelCased(""))
|
|
4
|
-
Log.show(splitCamelCased("ABC"))
|
|
5
|
-
Log.show(splitCamelCased("abc"))
|
|
6
|
-
Log.show(splitCamelCased("AbC"))
|
|
7
|
-
Log.show(splitCamelCased("AbCd"))
|
|
8
|
-
Log.show(splitCamelCased("aaBccDee"))
|
|
9
|
-
Log.show(splitCamelCased("AaBccDee"))
|
|
10
|
-
|
|
11
|
-
Log.show(prefixes(""))
|
|
12
|
-
Log.show(prefixes("abcde"))
|
|
13
|
-
|
|
14
|
-
test(["add", "addAll", "addToList"], "al")
|
|
15
|
-
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "abcdefghijklmnopqrstuvwxyz")
|
|
16
|
-
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "abcdefghijklmnopqrstuvwxyzA")
|
|
17
|
-
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "aabcdefghijklmnopqrstuvwxyz")
|
|
18
|
-
*/
|
|
19
|
-
printTime(system.time(), "matchSymbol") {
|
|
20
|
-
Log.show(matchSymbol(splitCamelCased("aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ", True), "abcdefghijklmnopqrstuvwxyzA", True))
|
|
21
|
-
}
|
|
22
|
-
|
|
23
|
-
}
|
|
24
|
-
|
|
25
|
-
test(symbols: List[String], query: String) {
|
|
26
|
-
Log.debug("")
|
|
27
|
-
Log.debug("#" + query)
|
|
28
|
-
symbols.each {s =>
|
|
29
|
-
let symbolWords = splitCamelCased(s, True)
|
|
30
|
-
let match = matchSymbol(symbolWords, query, !query.any {_.isAsciiUpper()})
|
|
31
|
-
Log.debug(s + " (" + Show.show(symbolWords) + " ): " + Show.show(match))
|
|
32
|
-
}
|
|
33
|
-
}
|
|
34
|
-
|
|
35
|
-
matchSymbol(symbolWords: List[String], query: String, lower: Bool): Bool {
|
|
36
|
-
let memo: IntMap[IntMap[Bool]] = IntMap.new()
|
|
37
|
-
function go(wordIndex: Int, queryOffset: Int): Bool {
|
|
38
|
-
memo.getOrSet(wordIndex) {IntMap.new()}.getOrSet(queryOffset):
|
|
39
|
-
if(queryOffset >= query.size()) {True} else:
|
|
40
|
-
if(wordIndex >= symbolWords.size()) {False} else:
|
|
41
|
-
prefixes(query.dropFirst(queryOffset)).any {prefix =>
|
|
42
|
-
symbolWords.grab(wordIndex).startsWith(prefix) &&
|
|
43
|
-
go(wordIndex + 1, queryOffset + prefix.size())
|
|
44
|
-
}
|
|
45
|
-
}
|
|
46
|
-
go(0, 0)
|
|
47
|
-
}
|
|
48
|
-
|
|
49
|
-
prefixes(string: String): List[String] {
|
|
50
|
-
if(string == "") {
|
|
51
|
-
[""]
|
|
52
|
-
} else {
|
|
53
|
-
[string, ...prefixes(string.dropLast())]
|
|
54
|
-
}
|
|
55
|
-
}
|
|
56
|
-
|
|
57
|
-
splitCamelCased(name: String, lower: Bool): List[String] {
|
|
58
|
-
name.first().{
|
|
59
|
-
| None => []
|
|
60
|
-
| Some(c) =>
|
|
61
|
-
let word = c.toString() + name.dropFirst().takeWhile {_.isAsciiLower()}
|
|
62
|
-
let rest = name.dropFirst(word.size())
|
|
63
|
-
[if(lower) {word.lower()} else {word}, ...splitCamelCased(rest, lower)]
|
|
64
|
-
}
|
|
65
|
-
}
|
|
66
|
-
|
|
67
|
-
printTime[R](ts: TimeSystem, label: String, body: () => R): R {
|
|
68
|
-
let start = ts.elapsed()
|
|
69
|
-
let result = body()
|
|
70
|
-
let stop = ts.elapsed()
|
|
71
|
-
let duration = ("" + (stop.seconds - start.seconds)).slice(0, 5)
|
|
72
|
-
Log.trace(label + ": " + duration + "s")
|
|
73
|
-
result
|
|
1
|
+
nodeMain(system: NodeSystem) {
|
|
2
|
+
/*
|
|
3
|
+
Log.show(splitCamelCased(""))
|
|
4
|
+
Log.show(splitCamelCased("ABC"))
|
|
5
|
+
Log.show(splitCamelCased("abc"))
|
|
6
|
+
Log.show(splitCamelCased("AbC"))
|
|
7
|
+
Log.show(splitCamelCased("AbCd"))
|
|
8
|
+
Log.show(splitCamelCased("aaBccDee"))
|
|
9
|
+
Log.show(splitCamelCased("AaBccDee"))
|
|
10
|
+
|
|
11
|
+
Log.show(prefixes(""))
|
|
12
|
+
Log.show(prefixes("abcde"))
|
|
13
|
+
|
|
14
|
+
test(["add", "addAll", "addToList"], "al")
|
|
15
|
+
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "abcdefghijklmnopqrstuvwxyz")
|
|
16
|
+
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "abcdefghijklmnopqrstuvwxyzA")
|
|
17
|
+
test(["aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ"], "aabcdefghijklmnopqrstuvwxyz")
|
|
18
|
+
*/
|
|
19
|
+
printTime(system.time(), "matchSymbol") {
|
|
20
|
+
Log.show(matchSymbol(splitCamelCased("aAbBcCdDeEfFgGhHiIjJkKlLmMnNoOpPqQrRsStTuUvVwWxXyYzZ", True), "abcdefghijklmnopqrstuvwxyzA", True))
|
|
21
|
+
}
|
|
22
|
+
|
|
23
|
+
}
|
|
24
|
+
|
|
25
|
+
test(symbols: List[String], query: String) {
|
|
26
|
+
Log.debug("")
|
|
27
|
+
Log.debug("#" + query)
|
|
28
|
+
symbols.each {s =>
|
|
29
|
+
let symbolWords = splitCamelCased(s, True)
|
|
30
|
+
let match = matchSymbol(symbolWords, query, !query.any {_.isAsciiUpper()})
|
|
31
|
+
Log.debug(s + " (" + Show.show(symbolWords) + " ): " + Show.show(match))
|
|
32
|
+
}
|
|
33
|
+
}
|
|
34
|
+
|
|
35
|
+
matchSymbol(symbolWords: List[String], query: String, lower: Bool): Bool {
|
|
36
|
+
let memo: IntMap[IntMap[Bool]] = IntMap.new()
|
|
37
|
+
function go(wordIndex: Int, queryOffset: Int): Bool {
|
|
38
|
+
memo.getOrSet(wordIndex) {IntMap.new()}.getOrSet(queryOffset):
|
|
39
|
+
if(queryOffset >= query.size()) {True} else:
|
|
40
|
+
if(wordIndex >= symbolWords.size()) {False} else:
|
|
41
|
+
prefixes(query.dropFirst(queryOffset)).any {prefix =>
|
|
42
|
+
symbolWords.grab(wordIndex).startsWith(prefix) &&
|
|
43
|
+
go(wordIndex + 1, queryOffset + prefix.size())
|
|
44
|
+
}
|
|
45
|
+
}
|
|
46
|
+
go(0, 0)
|
|
47
|
+
}
|
|
48
|
+
|
|
49
|
+
prefixes(string: String): List[String] {
|
|
50
|
+
if(string == "") {
|
|
51
|
+
[""]
|
|
52
|
+
} else {
|
|
53
|
+
[string, ...prefixes(string.dropLast())]
|
|
54
|
+
}
|
|
55
|
+
}
|
|
56
|
+
|
|
57
|
+
splitCamelCased(name: String, lower: Bool): List[String] {
|
|
58
|
+
name.first().{
|
|
59
|
+
| None => []
|
|
60
|
+
| Some(c) =>
|
|
61
|
+
let word = c.toString() + name.dropFirst().takeWhile {_.isAsciiLower()}
|
|
62
|
+
let rest = name.dropFirst(word.size())
|
|
63
|
+
[if(lower) {word.lower()} else {word}, ...splitCamelCased(rest, lower)]
|
|
64
|
+
}
|
|
65
|
+
}
|
|
66
|
+
|
|
67
|
+
printTime[R](ts: TimeSystem, label: String, body: () => R): R {
|
|
68
|
+
let start = ts.elapsed()
|
|
69
|
+
let result = body()
|
|
70
|
+
let stop = ts.elapsed()
|
|
71
|
+
let duration = ("" + (stop.seconds - start.seconds)).slice(0, 5)
|
|
72
|
+
Log.trace(label + ": " + duration + "s")
|
|
73
|
+
result
|
|
74
74
|
}
|
|
@@ -1,52 +1,52 @@
|
|
|
1
|
-
// Brainstorming representation for Tensors where the scalar types and operators are left open ended.
|
|
2
|
-
// The purpose of this type is to build up a syntax tree of tensor operations that can then be
|
|
3
|
-
// optimized and executed on CPU/GPU/TPU, enabling machine learning and other numerical applications.
|
|
4
|
-
|
|
5
|
-
data Tensor(scalar: String, shape: List[Int]) {
|
|
6
|
-
TensorVariable(name: String)
|
|
7
|
-
TensorInts(data: Array[Int])
|
|
8
|
-
TensorFloats(data: Array[Float])
|
|
9
|
-
TensorBuffer(data: Buffer)
|
|
10
|
-
TensorOperator(operator: String, operands: List[Tensor])
|
|
11
|
-
}
|
|
12
|
-
|
|
13
|
-
data TensorException(message: String)
|
|
14
|
-
|
|
15
|
-
capability TensorProcessor(
|
|
16
|
-
toFloats: Tensor => Array[Float]
|
|
17
|
-
toInts: Tensor => Array[Int]
|
|
18
|
-
toBuffer: Tensor => Buffer
|
|
19
|
-
)
|
|
20
|
-
|
|
21
|
-
defaultProcessor(): TensorProcessor {
|
|
22
|
-
throw(TensorException("Default tensor processor not yet implemented"))
|
|
23
|
-
}
|
|
24
|
-
|
|
25
|
-
f32(shape: List[Int], data: List[Float]): Tensor {
|
|
26
|
-
TensorFloats("ff_f32", shape, data.toArray())
|
|
27
|
-
}
|
|
28
|
-
|
|
29
|
-
f32a(shape: List[Int], data: Array[Float]): Tensor {
|
|
30
|
-
TensorFloats("ff_f64", shape, data)
|
|
31
|
-
}
|
|
32
|
-
|
|
33
|
-
extend this: Tensor {
|
|
34
|
-
|
|
35
|
-
multiply(that: Tensor): Tensor {
|
|
36
|
-
if(this.scalar != that.scalar) {
|
|
37
|
-
throw(TensorException("Scalar type mismatch: " + this.scalar + " vs. " + that.scalar))
|
|
38
|
-
}
|
|
39
|
-
let scalar = this.scalar
|
|
40
|
-
let shape = [] // TODO: Check and compute
|
|
41
|
-
TensorOperator(scalar, shape, "ff_multiply", [this, that])
|
|
42
|
-
}
|
|
43
|
-
|
|
44
|
-
toFloats(processor: TensorProcessor = defaultProcessor()): Array[Float] {
|
|
45
|
-
processor.toFloats(this)
|
|
46
|
-
}
|
|
47
|
-
|
|
48
|
-
}
|
|
49
|
-
|
|
50
|
-
main(system: NodeSystem) {
|
|
51
|
-
let tensor = Tensor.f32([], [1.0])
|
|
52
|
-
}
|
|
1
|
+
// Brainstorming representation for Tensors where the scalar types and operators are left open ended.
|
|
2
|
+
// The purpose of this type is to build up a syntax tree of tensor operations that can then be
|
|
3
|
+
// optimized and executed on CPU/GPU/TPU, enabling machine learning and other numerical applications.
|
|
4
|
+
|
|
5
|
+
data Tensor(scalar: String, shape: List[Int]) {
|
|
6
|
+
TensorVariable(name: String)
|
|
7
|
+
TensorInts(data: Array[Int])
|
|
8
|
+
TensorFloats(data: Array[Float])
|
|
9
|
+
TensorBuffer(data: Buffer)
|
|
10
|
+
TensorOperator(operator: String, operands: List[Tensor])
|
|
11
|
+
}
|
|
12
|
+
|
|
13
|
+
data TensorException(message: String)
|
|
14
|
+
|
|
15
|
+
capability TensorProcessor(
|
|
16
|
+
toFloats: Tensor => Array[Float]
|
|
17
|
+
toInts: Tensor => Array[Int]
|
|
18
|
+
toBuffer: Tensor => Buffer
|
|
19
|
+
)
|
|
20
|
+
|
|
21
|
+
defaultProcessor(): TensorProcessor {
|
|
22
|
+
throw(TensorException("Default tensor processor not yet implemented"))
|
|
23
|
+
}
|
|
24
|
+
|
|
25
|
+
f32(shape: List[Int], data: List[Float]): Tensor {
|
|
26
|
+
TensorFloats("ff_f32", shape, data.toArray())
|
|
27
|
+
}
|
|
28
|
+
|
|
29
|
+
f32a(shape: List[Int], data: Array[Float]): Tensor {
|
|
30
|
+
TensorFloats("ff_f64", shape, data)
|
|
31
|
+
}
|
|
32
|
+
|
|
33
|
+
extend this: Tensor {
|
|
34
|
+
|
|
35
|
+
multiply(that: Tensor): Tensor {
|
|
36
|
+
if(this.scalar != that.scalar) {
|
|
37
|
+
throw(TensorException("Scalar type mismatch: " + this.scalar + " vs. " + that.scalar))
|
|
38
|
+
}
|
|
39
|
+
let scalar = this.scalar
|
|
40
|
+
let shape = [] // TODO: Check and compute
|
|
41
|
+
TensorOperator(scalar, shape, "ff_multiply", [this, that])
|
|
42
|
+
}
|
|
43
|
+
|
|
44
|
+
toFloats(processor: TensorProcessor = defaultProcessor()): Array[Float] {
|
|
45
|
+
processor.toFloats(this)
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
}
|
|
49
|
+
|
|
50
|
+
main(system: NodeSystem) {
|
|
51
|
+
let tensor = Tensor.f32([], [1.0])
|
|
52
|
+
}
|
|
@@ -1,36 +1,36 @@
|
|
|
1
|
-
|
|
2
|
-
nodeMain(system: NodeSystem) {
|
|
3
|
-
if(True) {
|
|
4
|
-
1 + 1
|
|
5
|
-
} else {
|
|
6
|
-
"foo"
|
|
7
|
-
}
|
|
8
|
-
if(True) {
|
|
9
|
-
1
|
|
10
|
-
} elseIf {False} {
|
|
11
|
-
'x'
|
|
12
|
-
} else {
|
|
13
|
-
"foo"
|
|
14
|
-
}
|
|
15
|
-
if(True) {
|
|
16
|
-
if(True) {
|
|
17
|
-
1
|
|
18
|
-
} else {
|
|
19
|
-
"foo"
|
|
20
|
-
}
|
|
21
|
-
} else {
|
|
22
|
-
"foo"
|
|
23
|
-
}
|
|
24
|
-
try {
|
|
25
|
-
1
|
|
26
|
-
} catchAny {_ =>
|
|
27
|
-
"foo"
|
|
28
|
-
} grab()
|
|
29
|
-
[1, 2, 3].each {_ =>
|
|
30
|
-
try {
|
|
31
|
-
if(False) {1} else {2}
|
|
32
|
-
} catchAny {_ =>
|
|
33
|
-
"foo"
|
|
34
|
-
} grab()
|
|
35
|
-
}
|
|
36
|
-
}
|
|
1
|
+
|
|
2
|
+
nodeMain(system: NodeSystem) {
|
|
3
|
+
if(True) {
|
|
4
|
+
1 + 1
|
|
5
|
+
} else {
|
|
6
|
+
"foo"
|
|
7
|
+
}
|
|
8
|
+
if(True) {
|
|
9
|
+
1
|
|
10
|
+
} elseIf {False} {
|
|
11
|
+
'x'
|
|
12
|
+
} else {
|
|
13
|
+
"foo"
|
|
14
|
+
}
|
|
15
|
+
if(True) {
|
|
16
|
+
if(True) {
|
|
17
|
+
1
|
|
18
|
+
} else {
|
|
19
|
+
"foo"
|
|
20
|
+
}
|
|
21
|
+
} else {
|
|
22
|
+
"foo"
|
|
23
|
+
}
|
|
24
|
+
try {
|
|
25
|
+
1
|
|
26
|
+
} catchAny {_ =>
|
|
27
|
+
"foo"
|
|
28
|
+
} grab()
|
|
29
|
+
[1, 2, 3].each {_ =>
|
|
30
|
+
try {
|
|
31
|
+
if(False) {1} else {2}
|
|
32
|
+
} catchAny {_ =>
|
|
33
|
+
"foo"
|
|
34
|
+
} grab()
|
|
35
|
+
}
|
|
36
|
+
}
|
|
@@ -1,40 +1,40 @@
|
|
|
1
|
-
dependency ff:s3:0.0.0
|
|
2
|
-
import S3 from ff:s3
|
|
3
|
-
|
|
4
|
-
nodeMain(system: NodeSystem) {
|
|
5
|
-
// Trying to reproduce the result from an example here
|
|
6
|
-
// https://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-header-based-auth.html
|
|
7
|
-
let expectedHeader = "AWS4-HMAC-SHA256 Credential=AKIAIOSFODNN7EXAMPLE/20130524/us-east-1/s3/aws4_request,SignedHeaders=date;host;x-amz-content-sha256;x-amz-date;x-amz-storage-class,Signature=98ad721746da40c64f1a55b78f14c238d841ea1380cd77a1b5971af0ece108bd"
|
|
8
|
-
let amzDate = "20130524T000000Z" // TODO
|
|
9
|
-
let headerDate = "Fri, 24 May 2013 00:00:00 GMT" // TODO
|
|
10
|
-
let bucket = "examplebucket"
|
|
11
|
-
let body = "Welcome to Amazon S3.".toBuffer()
|
|
12
|
-
let actualHeader = S3.makeS3AuthorizationHeader(
|
|
13
|
-
system = system
|
|
14
|
-
accessKeyId = "AKIAIOSFODNN7EXAMPLE"
|
|
15
|
-
secretAccessKey = "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
|
|
16
|
-
region = "us-east-1"
|
|
17
|
-
bucket = bucket
|
|
18
|
-
encodedKey = S3.encode("test$file.text")
|
|
19
|
-
body = body
|
|
20
|
-
canonicalHeaders = [
|
|
21
|
-
Pair("date", headerDate)
|
|
22
|
-
Pair("host", bucket + ".s3.amazonaws.com")
|
|
23
|
-
Pair("x-amz-date", amzDate)
|
|
24
|
-
Pair("x-amz-content-sha256", system.crypto().sha256(body).toHex())
|
|
25
|
-
Pair("x-amz-storage-class", "REDUCED_REDUNDANCY")
|
|
26
|
-
]
|
|
27
|
-
amzDate = amzDate
|
|
28
|
-
)
|
|
29
|
-
system.writeLine("")
|
|
30
|
-
system.writeLine("actual : " + actualHeader)
|
|
31
|
-
system.writeLine("")
|
|
32
|
-
system.writeLine("expected: " + expectedHeader)
|
|
33
|
-
|
|
34
|
-
system.writeLine("")
|
|
35
|
-
if(actualHeader == expectedHeader) {
|
|
36
|
-
system.writeLine("Test passed")
|
|
37
|
-
} else {
|
|
38
|
-
system.writeLine("Test failed")
|
|
39
|
-
}
|
|
1
|
+
dependency ff:s3:0.0.0
|
|
2
|
+
import S3 from ff:s3
|
|
3
|
+
|
|
4
|
+
nodeMain(system: NodeSystem) {
|
|
5
|
+
// Trying to reproduce the result from an example here
|
|
6
|
+
// https://docs.aws.amazon.com/AmazonS3/latest/API/sig-v4-header-based-auth.html
|
|
7
|
+
let expectedHeader = "AWS4-HMAC-SHA256 Credential=AKIAIOSFODNN7EXAMPLE/20130524/us-east-1/s3/aws4_request,SignedHeaders=date;host;x-amz-content-sha256;x-amz-date;x-amz-storage-class,Signature=98ad721746da40c64f1a55b78f14c238d841ea1380cd77a1b5971af0ece108bd"
|
|
8
|
+
let amzDate = "20130524T000000Z" // TODO
|
|
9
|
+
let headerDate = "Fri, 24 May 2013 00:00:00 GMT" // TODO
|
|
10
|
+
let bucket = "examplebucket"
|
|
11
|
+
let body = "Welcome to Amazon S3.".toBuffer()
|
|
12
|
+
let actualHeader = S3.makeS3AuthorizationHeader(
|
|
13
|
+
system = system
|
|
14
|
+
accessKeyId = "AKIAIOSFODNN7EXAMPLE"
|
|
15
|
+
secretAccessKey = "wJalrXUtnFEMI/K7MDENG/bPxRfiCYEXAMPLEKEY"
|
|
16
|
+
region = "us-east-1"
|
|
17
|
+
bucket = bucket
|
|
18
|
+
encodedKey = S3.encode("test$file.text")
|
|
19
|
+
body = body
|
|
20
|
+
canonicalHeaders = [
|
|
21
|
+
Pair("date", headerDate)
|
|
22
|
+
Pair("host", bucket + ".s3.amazonaws.com")
|
|
23
|
+
Pair("x-amz-date", amzDate)
|
|
24
|
+
Pair("x-amz-content-sha256", system.crypto().sha256(body).toHex())
|
|
25
|
+
Pair("x-amz-storage-class", "REDUCED_REDUNDANCY")
|
|
26
|
+
]
|
|
27
|
+
amzDate = amzDate
|
|
28
|
+
)
|
|
29
|
+
system.writeLine("")
|
|
30
|
+
system.writeLine("actual : " + actualHeader)
|
|
31
|
+
system.writeLine("")
|
|
32
|
+
system.writeLine("expected: " + expectedHeader)
|
|
33
|
+
|
|
34
|
+
system.writeLine("")
|
|
35
|
+
if(actualHeader == expectedHeader) {
|
|
36
|
+
system.writeLine("Test passed")
|
|
37
|
+
} else {
|
|
38
|
+
system.writeLine("Test failed")
|
|
39
|
+
}
|
|
40
40
|
}
|