first commit
This commit is contained in:
commit
46e1a1e2ce
8 changed files with 590 additions and 0 deletions
2
.gitignore
vendored
Normal file
2
.gitignore
vendored
Normal file
|
|
@ -0,0 +1,2 @@
|
||||||
|
zig-out/
|
||||||
|
.zig-cache/
|
||||||
156
build.zig
Normal file
156
build.zig
Normal file
|
|
@ -0,0 +1,156 @@
|
||||||
|
const std = @import("std");
|
||||||
|
|
||||||
|
// Although this function looks imperative, it does not perform the build
|
||||||
|
// directly and instead it mutates the build graph (`b`) that will be then
|
||||||
|
// executed by an external runner. The functions in `std.Build` implement a DSL
|
||||||
|
// for defining build steps and express dependencies between them, allowing the
|
||||||
|
// build runner to parallelize the build automatically (and the cache system to
|
||||||
|
// know when a step doesn't need to be re-run).
|
||||||
|
pub fn build(b: *std.Build) void {
|
||||||
|
// Standard target options allow the person running `zig build` to choose
|
||||||
|
// what target to build for. Here we do not override the defaults, which
|
||||||
|
// means any target is allowed, and the default is native. Other options
|
||||||
|
// for restricting supported target set are available.
|
||||||
|
const target = b.standardTargetOptions(.{});
|
||||||
|
// Standard optimization options allow the person running `zig build` to select
|
||||||
|
// between Debug, ReleaseSafe, ReleaseFast, and ReleaseSmall. Here we do not
|
||||||
|
// set a preferred release mode, allowing the user to decide how to optimize.
|
||||||
|
const optimize = b.standardOptimizeOption(.{});
|
||||||
|
// It's also possible to define more custom flags to toggle optional features
|
||||||
|
// of this build script using `b.option()`. All defined flags (including
|
||||||
|
// target and optimize options) will be listed when running `zig build --help`
|
||||||
|
// in this directory.
|
||||||
|
|
||||||
|
// This creates a module, which represents a collection of source files alongside
|
||||||
|
// some compilation options, such as optimization mode and linked system libraries.
|
||||||
|
// Zig modules are the preferred way of making Zig code available to consumers.
|
||||||
|
// addModule defines a module that we intend to make available for importing
|
||||||
|
// to our consumers. We must give it a name because a Zig package can expose
|
||||||
|
// multiple modules and consumers will need to be able to specify which
|
||||||
|
// module they want to access.
|
||||||
|
const mod = b.addModule("al", .{
|
||||||
|
// The root source file is the "entry point" of this module. Users of
|
||||||
|
// this module will only be able to access public declarations contained
|
||||||
|
// in this file, which means that if you have declarations that you
|
||||||
|
// intend to expose to consumers that were defined in other files part
|
||||||
|
// of this module, you will have to make sure to re-export them from
|
||||||
|
// the root file.
|
||||||
|
.root_source_file = b.path("src/root.zig"),
|
||||||
|
// Later on we'll use this module as the root module of a test executable
|
||||||
|
// which requires us to specify a target.
|
||||||
|
.target = target,
|
||||||
|
});
|
||||||
|
|
||||||
|
// Here we define an executable. An executable needs to have a root module
|
||||||
|
// which needs to expose a `main` function. While we could add a main function
|
||||||
|
// to the module defined above, it's sometimes preferable to split business
|
||||||
|
// logic and the CLI into two separate modules.
|
||||||
|
//
|
||||||
|
// If your goal is to create a Zig library for others to use, consider if
|
||||||
|
// it might benefit from also exposing a CLI tool. A parser library for a
|
||||||
|
// data serialization format could also bundle a CLI syntax checker, for example.
|
||||||
|
//
|
||||||
|
// If instead your goal is to create an executable, consider if users might
|
||||||
|
// be interested in also being able to embed the core functionality of your
|
||||||
|
// program in their own executable in order to avoid the overhead involved in
|
||||||
|
// subprocessing your CLI tool.
|
||||||
|
//
|
||||||
|
// If neither case applies to you, feel free to delete the declaration you
|
||||||
|
// don't need and to put everything under a single module.
|
||||||
|
const exe = b.addExecutable(.{
|
||||||
|
.name = "al",
|
||||||
|
.root_module = b.createModule(.{
|
||||||
|
// b.createModule defines a new module just like b.addModule but,
|
||||||
|
// unlike b.addModule, it does not expose the module to consumers of
|
||||||
|
// this package, which is why in this case we don't have to give it a name.
|
||||||
|
.root_source_file = b.path("src/main.zig"),
|
||||||
|
// Target and optimization levels must be explicitly wired in when
|
||||||
|
// defining an executable or library (in the root module), and you
|
||||||
|
// can also hardcode a specific target for an executable or library
|
||||||
|
// definition if desireable (e.g. firmware for embedded devices).
|
||||||
|
.target = target,
|
||||||
|
.optimize = optimize,
|
||||||
|
// List of modules available for import in source files part of the
|
||||||
|
// root module.
|
||||||
|
.imports = &.{
|
||||||
|
// Here "al" is the name you will use in your source code to
|
||||||
|
// import this module (e.g. `@import("al")`). The name is
|
||||||
|
// repeated because you are allowed to rename your imports, which
|
||||||
|
// can be extremely useful in case of collisions (which can happen
|
||||||
|
// importing modules from different packages).
|
||||||
|
.{ .name = "al", .module = mod },
|
||||||
|
},
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
|
||||||
|
// This declares intent for the executable to be installed into the
|
||||||
|
// install prefix when running `zig build` (i.e. when executing the default
|
||||||
|
// step). By default the install prefix is `zig-out/` but can be overridden
|
||||||
|
// by passing `--prefix` or `-p`.
|
||||||
|
b.installArtifact(exe);
|
||||||
|
|
||||||
|
// This creates a top level step. Top level steps have a name and can be
|
||||||
|
// invoked by name when running `zig build` (e.g. `zig build run`).
|
||||||
|
// This will evaluate the `run` step rather than the default step.
|
||||||
|
// For a top level step to actually do something, it must depend on other
|
||||||
|
// steps (e.g. a Run step, as we will see in a moment).
|
||||||
|
const run_step = b.step("run", "Run the app");
|
||||||
|
|
||||||
|
// This creates a RunArtifact step in the build graph. A RunArtifact step
|
||||||
|
// invokes an executable compiled by Zig. Steps will only be executed by the
|
||||||
|
// runner if invoked directly by the user (in the case of top level steps)
|
||||||
|
// or if another step depends on it, so it's up to you to define when and
|
||||||
|
// how this Run step will be executed. In our case we want to run it when
|
||||||
|
// the user runs `zig build run`, so we create a dependency link.
|
||||||
|
const run_cmd = b.addRunArtifact(exe);
|
||||||
|
run_step.dependOn(&run_cmd.step);
|
||||||
|
|
||||||
|
// By making the run step depend on the default step, it will be run from the
|
||||||
|
// installation directory rather than directly from within the cache directory.
|
||||||
|
run_cmd.step.dependOn(b.getInstallStep());
|
||||||
|
|
||||||
|
// This allows the user to pass arguments to the application in the build
|
||||||
|
// command itself, like this: `zig build run -- arg1 arg2 etc`
|
||||||
|
if (b.args) |args| {
|
||||||
|
run_cmd.addArgs(args);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Creates an executable that will run `test` blocks from the provided module.
|
||||||
|
// Here `mod` needs to define a target, which is why earlier we made sure to
|
||||||
|
// set the releative field.
|
||||||
|
const mod_tests = b.addTest(.{
|
||||||
|
.root_module = mod,
|
||||||
|
});
|
||||||
|
|
||||||
|
// A run step that will run the test executable.
|
||||||
|
const run_mod_tests = b.addRunArtifact(mod_tests);
|
||||||
|
|
||||||
|
// Creates an executable that will run `test` blocks from the executable's
|
||||||
|
// root module. Note that test executables only test one module at a time,
|
||||||
|
// hence why we have to create two separate ones.
|
||||||
|
const exe_tests = b.addTest(.{
|
||||||
|
.root_module = exe.root_module,
|
||||||
|
});
|
||||||
|
|
||||||
|
// A run step that will run the second test executable.
|
||||||
|
const run_exe_tests = b.addRunArtifact(exe_tests);
|
||||||
|
|
||||||
|
// A top level step for running all tests. dependOn can be called multiple
|
||||||
|
// times and since the two run steps do not depend on one another, this will
|
||||||
|
// make the two of them run in parallel.
|
||||||
|
const test_step = b.step("test", "Run tests");
|
||||||
|
test_step.dependOn(&run_mod_tests.step);
|
||||||
|
test_step.dependOn(&run_exe_tests.step);
|
||||||
|
|
||||||
|
// Just like flags, top level steps are also listed in the `--help` menu.
|
||||||
|
//
|
||||||
|
// The Zig build system is entirely implemented in userland, which means
|
||||||
|
// that it cannot hook into private compiler APIs. All compilation work
|
||||||
|
// orchestrated by the build system will result in other Zig compiler
|
||||||
|
// subcommands being invoked with the right flags defined. You can observe
|
||||||
|
// these invocations when one fails (or you pass a flag to increase
|
||||||
|
// verbosity) to validate assumptions and diagnose problems.
|
||||||
|
//
|
||||||
|
// Lastly, the Zig build system is relatively simple and self-contained,
|
||||||
|
// and reading its source code will allow you to master it.
|
||||||
|
}
|
||||||
81
build.zig.zon
Normal file
81
build.zig.zon
Normal file
|
|
@ -0,0 +1,81 @@
|
||||||
|
.{
|
||||||
|
// This is the default name used by packages depending on this one. For
|
||||||
|
// example, when a user runs `zig fetch --save <url>`, this field is used
|
||||||
|
// as the key in the `dependencies` table. Although the user can choose a
|
||||||
|
// different name, most users will stick with this provided value.
|
||||||
|
//
|
||||||
|
// It is redundant to include "zig" in this name because it is already
|
||||||
|
// within the Zig package namespace.
|
||||||
|
.name = .al,
|
||||||
|
// This is a [Semantic Version](https://semver.org/).
|
||||||
|
// In a future version of Zig it will be used for package deduplication.
|
||||||
|
.version = "0.0.0",
|
||||||
|
// Together with name, this represents a globally unique package
|
||||||
|
// identifier. This field is generated by the Zig toolchain when the
|
||||||
|
// package is first created, and then *never changes*. This allows
|
||||||
|
// unambiguous detection of one package being an updated version of
|
||||||
|
// another.
|
||||||
|
//
|
||||||
|
// When forking a Zig project, this id should be regenerated (delete the
|
||||||
|
// field and run `zig build`) if the upstream project is still maintained.
|
||||||
|
// Otherwise, the fork is *hostile*, attempting to take control over the
|
||||||
|
// original project's identity. Thus it is recommended to leave the comment
|
||||||
|
// on the following line intact, so that it shows up in code reviews that
|
||||||
|
// modify the field.
|
||||||
|
.fingerprint = 0x793b656ad110237c, // Changing this has security and trust implications.
|
||||||
|
// Tracks the earliest Zig version that the package considers to be a
|
||||||
|
// supported use case.
|
||||||
|
.minimum_zig_version = "0.15.2",
|
||||||
|
// This field is optional.
|
||||||
|
// Each dependency must either provide a `url` and `hash`, or a `path`.
|
||||||
|
// `zig build --fetch` can be used to fetch all dependencies of a package, recursively.
|
||||||
|
// Once all dependencies are fetched, `zig build` no longer requires
|
||||||
|
// internet connectivity.
|
||||||
|
.dependencies = .{
|
||||||
|
// See `zig fetch --save <url>` for a command-line interface for adding dependencies.
|
||||||
|
//.example = .{
|
||||||
|
// // When updating this field to a new URL, be sure to delete the corresponding
|
||||||
|
// // `hash`, otherwise you are communicating that you expect to find the old hash at
|
||||||
|
// // the new URL. If the contents of a URL change this will result in a hash mismatch
|
||||||
|
// // which will prevent zig from using it.
|
||||||
|
// .url = "https://example.com/foo.tar.gz",
|
||||||
|
//
|
||||||
|
// // This is computed from the file contents of the directory of files that is
|
||||||
|
// // obtained after fetching `url` and applying the inclusion rules given by
|
||||||
|
// // `paths`.
|
||||||
|
// //
|
||||||
|
// // This field is the source of truth; packages do not come from a `url`; they
|
||||||
|
// // come from a `hash`. `url` is just one of many possible mirrors for how to
|
||||||
|
// // obtain a package matching this `hash`.
|
||||||
|
// //
|
||||||
|
// // Uses the [multihash](https://multiformats.io/multihash/) format.
|
||||||
|
// .hash = "...",
|
||||||
|
//
|
||||||
|
// // When this is provided, the package is found in a directory relative to the
|
||||||
|
// // build root. In this case the package's hash is irrelevant and therefore not
|
||||||
|
// // computed. This field and `url` are mutually exclusive.
|
||||||
|
// .path = "foo",
|
||||||
|
//
|
||||||
|
// // When this is set to `true`, a package is declared to be lazily
|
||||||
|
// // fetched. This makes the dependency only get fetched if it is
|
||||||
|
// // actually used.
|
||||||
|
// .lazy = false,
|
||||||
|
//},
|
||||||
|
},
|
||||||
|
// Specifies the set of files and directories that are included in this package.
|
||||||
|
// Only files and directories listed here are included in the `hash` that
|
||||||
|
// is computed for this package. Only files listed here will remain on disk
|
||||||
|
// when using the zig package manager. As a rule of thumb, one should list
|
||||||
|
// files required for compilation plus any license(s).
|
||||||
|
// Paths are relative to the build root. Use the empty string (`""`) to refer to
|
||||||
|
// the build root itself.
|
||||||
|
// A directory listed here means that all files within, recursively, are included.
|
||||||
|
.paths = .{
|
||||||
|
"build.zig",
|
||||||
|
"build.zig.zon",
|
||||||
|
"src",
|
||||||
|
// For example...
|
||||||
|
//"LICENSE",
|
||||||
|
//"README.md",
|
||||||
|
},
|
||||||
|
}
|
||||||
5
graph
Executable file
5
graph
Executable file
|
|
@ -0,0 +1,5 @@
|
||||||
|
#!/bin/bash
|
||||||
|
|
||||||
|
zig build run 2>&1 >/dev/null | dot -Tpdf -o graph.pdf
|
||||||
|
zathura graph.pdf
|
||||||
|
rm graph.pdf
|
||||||
132
src/Lexer.zig
Normal file
132
src/Lexer.zig
Normal file
|
|
@ -0,0 +1,132 @@
|
||||||
|
const std = @import("std");
|
||||||
|
const Lexer = @This();
|
||||||
|
|
||||||
|
index: usize,
|
||||||
|
source: []u8,
|
||||||
|
start: usize,
|
||||||
|
|
||||||
|
pub const TokenType = enum {
|
||||||
|
plus,
|
||||||
|
minus,
|
||||||
|
star,
|
||||||
|
slash,
|
||||||
|
integer,
|
||||||
|
float,
|
||||||
|
identifier,
|
||||||
|
|
||||||
|
eof,
|
||||||
|
illegal,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const Token = struct {
|
||||||
|
@"type": TokenType,
|
||||||
|
lexeme: []u8,
|
||||||
|
};
|
||||||
|
|
||||||
|
pub fn parseSpecial(lexer: *Lexer) Token {
|
||||||
|
var token_type: TokenType = .eof;
|
||||||
|
switch (lexer.source[lexer.index]) {
|
||||||
|
'+' => token_type = .plus,
|
||||||
|
'-' => token_type = .minus,
|
||||||
|
'*' => token_type = .star,
|
||||||
|
'/' => token_type = .slash,
|
||||||
|
else => token_type = .eof,
|
||||||
|
}
|
||||||
|
|
||||||
|
lexer.index += 1;
|
||||||
|
|
||||||
|
return .{
|
||||||
|
.@"type" = token_type,
|
||||||
|
.lexeme = undefined,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn next(lexer: *Lexer) Token {
|
||||||
|
lexer.skipWhitespaceAndComments();
|
||||||
|
|
||||||
|
if (lexer.index >= lexer.source.len) {
|
||||||
|
return lexer.makeToken(.eof);
|
||||||
|
}
|
||||||
|
|
||||||
|
const c = lexer.source[lexer.index];
|
||||||
|
|
||||||
|
// Numbers
|
||||||
|
if (std.ascii.isDigit(c)) {
|
||||||
|
return lexer.number();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Identifiers
|
||||||
|
if (std.ascii.isAlphabetic(c)) {
|
||||||
|
}
|
||||||
|
|
||||||
|
// Single Character Tokens
|
||||||
|
lexer.index += 1;
|
||||||
|
switch (c) {
|
||||||
|
'+' => return lexer.makeToken(.plus),
|
||||||
|
'-' => return lexer.makeToken(.minus),
|
||||||
|
'*' => return lexer.makeToken(.star),
|
||||||
|
'/' => return lexer.makeToken(.slash),
|
||||||
|
else => return lexer.makeToken(.illegal),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn makeToken(lexer: *Lexer, @"type": TokenType) Token {
|
||||||
|
return .{
|
||||||
|
.@"type" = @"type",
|
||||||
|
// Safely slice the source
|
||||||
|
.lexeme = if (lexer.index <= lexer.source.len)
|
||||||
|
lexer.source[lexer.start..lexer.index]
|
||||||
|
else
|
||||||
|
"",
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
fn skipWhitespaceAndComments(lexer: *Lexer) void {
|
||||||
|
while (lexer.index < lexer.source.len) {
|
||||||
|
const c = lexer.source[lexer.index];
|
||||||
|
switch (c) {
|
||||||
|
// Whitespace
|
||||||
|
' ', '\t', '\r', '\n' => {
|
||||||
|
lexer.index += 1;
|
||||||
|
},
|
||||||
|
// Comments
|
||||||
|
'#' => {
|
||||||
|
while (lexer.index < lexer.source.len and lexer.source[lexer.index] != '\n') {
|
||||||
|
lexer.index += 1;
|
||||||
|
}
|
||||||
|
},
|
||||||
|
else => {
|
||||||
|
lexer.start = lexer.index;
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn number(lexer: *Lexer) Token {
|
||||||
|
while (lexer.index < lexer.source.len and std.ascii.isDigit(lexer.source[lexer.index])) {
|
||||||
|
lexer.index += 1;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (lexer.index < lexer.source.len and lexer.source[lexer.index] == '.') {
|
||||||
|
if (lexer.index + 1 < lexer.source.len and std.ascii.isDigit(lexer.source[lexer.index + 1])) {
|
||||||
|
lexer.index += 1; // consume dot
|
||||||
|
while (lexer.index < lexer.source.len and std.ascii.isDigit(lexer.source[lexer.index])) {
|
||||||
|
lexer.index += 1;
|
||||||
|
}
|
||||||
|
return lexer.makeToken(.float);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return lexer.makeToken(.integer);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// If `source` was allocated on the heap,
|
||||||
|
/// the caller must free it.
|
||||||
|
pub fn init(source: []u8) Lexer {
|
||||||
|
return .{
|
||||||
|
.index = 0,
|
||||||
|
.source = source,
|
||||||
|
.start = 0,
|
||||||
|
};
|
||||||
|
}
|
||||||
170
src/Parser.zig
Normal file
170
src/Parser.zig
Normal file
|
|
@ -0,0 +1,170 @@
|
||||||
|
const std = @import("std");
|
||||||
|
const Lexer = @import("Lexer.zig");
|
||||||
|
const Parser = @This();
|
||||||
|
|
||||||
|
lexer: *Lexer,
|
||||||
|
allocator: std.mem.Allocator,
|
||||||
|
node_table: std.AutoHashMap(u64, *Node),
|
||||||
|
previous: Lexer.Token,
|
||||||
|
current: Lexer.Token,
|
||||||
|
|
||||||
|
pub const NodeType = enum {
|
||||||
|
add,
|
||||||
|
sub,
|
||||||
|
mul,
|
||||||
|
div,
|
||||||
|
|
||||||
|
integer,
|
||||||
|
float,
|
||||||
|
|
||||||
|
start,
|
||||||
|
@"return",
|
||||||
|
};
|
||||||
|
|
||||||
|
pub const Node = struct {
|
||||||
|
@"type": NodeType,
|
||||||
|
id: u64,
|
||||||
|
inputs: std.ArrayList(*Node),
|
||||||
|
outputs: std.ArrayList(*Node),
|
||||||
|
data: extern union {
|
||||||
|
integer: u64,
|
||||||
|
float: f64,
|
||||||
|
},
|
||||||
|
|
||||||
|
pub fn init(parser: *Parser, @"type": NodeType) !*Node {
|
||||||
|
var node = try parser.allocator.create(Node);
|
||||||
|
node.@"type" = @"type";
|
||||||
|
node.inputs = .{};
|
||||||
|
node.outputs = .{};
|
||||||
|
node.data = undefined;
|
||||||
|
return node;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn globalNumbering(node: *Node, parser: *Parser) !*Node {
|
||||||
|
const node_hash = node.hash();
|
||||||
|
node.id = node_hash;
|
||||||
|
if (parser.node_table.get(node_hash)) |n| {
|
||||||
|
parser.allocator.destroy(node);
|
||||||
|
return n;
|
||||||
|
}
|
||||||
|
|
||||||
|
try parser.node_table.put(node_hash, node);
|
||||||
|
|
||||||
|
return node;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn hash(node: *Node) u64 {
|
||||||
|
var hasher = std.hash.Wyhash.init(0);
|
||||||
|
std.hash.autoHash(&hasher, node.@"type");
|
||||||
|
|
||||||
|
switch (node.@"type") {
|
||||||
|
.integer => std.hash.autoHash(&hasher, node.data.integer),
|
||||||
|
.float => std.hash.autoHash(&hasher, @as(u64, @bitCast(node.data.float))),
|
||||||
|
else => {},
|
||||||
|
}
|
||||||
|
|
||||||
|
for (node.inputs.items) |n| {
|
||||||
|
std.hash.autoHash(&hasher, @intFromPtr(n));
|
||||||
|
}
|
||||||
|
|
||||||
|
return hasher.final();
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn deinit(node: *Node, parser: *Parser) void {
|
||||||
|
parser.allocator.destroy(node);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
pub fn match(parser: *Parser, expected: Lexer.TokenType) bool {
|
||||||
|
if (parser.current.@"type" == expected) {
|
||||||
|
parser.advance();
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn advance(parser: *Parser) void {
|
||||||
|
parser.previous = parser.current;
|
||||||
|
parser.current = parser.lexer.next();
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn buildFactor(parser: *Parser) !?*Node {
|
||||||
|
const token = parser.current;
|
||||||
|
switch (token.@"type") {
|
||||||
|
.integer => {
|
||||||
|
parser.advance();
|
||||||
|
const node = try Node.init(parser, .integer);
|
||||||
|
node.data.integer = std.fmt.parseInt(u64, token.lexeme, 10) catch 0;
|
||||||
|
return node.globalNumbering(parser);
|
||||||
|
},
|
||||||
|
.float => {
|
||||||
|
parser.advance();
|
||||||
|
const node = try Node.init(parser, .float);
|
||||||
|
node.data.float = std.fmt.parseFloat(f64, token.lexeme) catch 0;
|
||||||
|
return node.globalNumbering(parser);
|
||||||
|
},
|
||||||
|
else => {}
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn buildTerm(parser: *Parser) !?*Node {
|
||||||
|
var lhs = try parser.buildFactor();
|
||||||
|
|
||||||
|
while (parser.match(.star) or parser.match(.slash)) {
|
||||||
|
const node_type: NodeType = switch (parser.previous.@"type") {
|
||||||
|
.star => .mul,
|
||||||
|
.slash => .div,
|
||||||
|
else => unreachable,
|
||||||
|
};
|
||||||
|
var node = try Node.init(parser, node_type);
|
||||||
|
try node.inputs.append(parser.allocator, (try parser.buildFactor()).?);
|
||||||
|
try node.inputs.append(parser.allocator, lhs.?);
|
||||||
|
node = try node.globalNumbering(parser);
|
||||||
|
lhs = node;
|
||||||
|
}
|
||||||
|
|
||||||
|
return lhs;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn buildExpression(parser: *Parser) !?*Node {
|
||||||
|
var lhs = try parser.buildTerm();
|
||||||
|
|
||||||
|
while (parser.match(.plus) or parser.match(.minus)) {
|
||||||
|
const node_type: NodeType = switch (parser.previous.@"type") {
|
||||||
|
.plus => .add,
|
||||||
|
.minus => .sub,
|
||||||
|
else => unreachable,
|
||||||
|
};
|
||||||
|
var node = try Node.init(parser, node_type);
|
||||||
|
try node.inputs.append(parser.allocator, (try parser.buildTerm()).?);
|
||||||
|
try node.inputs.append(parser.allocator, lhs.?);
|
||||||
|
node = try node.globalNumbering(parser);
|
||||||
|
lhs = node;
|
||||||
|
}
|
||||||
|
|
||||||
|
return lhs;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn buildGraph(parser: *Parser) !?*Node {
|
||||||
|
return try buildExpression(parser);
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn init(allocator: std.mem.Allocator, lexer: *Lexer) Parser {
|
||||||
|
var parser: Parser = .{
|
||||||
|
.lexer = lexer,
|
||||||
|
.allocator = allocator,
|
||||||
|
.node_table = std.AutoHashMap(u64, *Node).init(allocator),
|
||||||
|
.previous = undefined,
|
||||||
|
.current = undefined,
|
||||||
|
};
|
||||||
|
|
||||||
|
parser.advance();
|
||||||
|
|
||||||
|
return parser;
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn deinit(parser: *Parser) void {
|
||||||
|
parser.node_table.deinit();
|
||||||
|
}
|
||||||
42
src/main.zig
Normal file
42
src/main.zig
Normal file
|
|
@ -0,0 +1,42 @@
|
||||||
|
const std = @import("std");
|
||||||
|
const al = @import("al");
|
||||||
|
|
||||||
|
pub fn nodeName(node: *al.Parser.Node) void {
|
||||||
|
switch (node.@"type") {
|
||||||
|
.start => std.debug.print("{d} [label=\"start\",fillcolor=yellow, color=black, shape=box]", .{node.id}),
|
||||||
|
.add => std.debug.print("{d} [label=\"+\"]", .{node.id}),
|
||||||
|
.sub => std.debug.print("{d} [label=\"-\"]", .{node.id}),
|
||||||
|
.mul => std.debug.print("{d} [label=\"*\"]", .{node.id}),
|
||||||
|
.div => std.debug.print("{d} [label=\"/\"]", .{node.id}),
|
||||||
|
.integer => std.debug.print("{d} [label=\"{d}\"]", .{node.id, node.data.integer}),
|
||||||
|
.float => std.debug.print("{d} [label=\"{d}\"]", .{node.id, node.data.float}),
|
||||||
|
else => {},
|
||||||
|
}
|
||||||
|
std.debug.print("\n", .{});
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn printGraph(node: *al.Parser.Node) void {
|
||||||
|
for (node.inputs.items) |n| {
|
||||||
|
nodeName(n);
|
||||||
|
std.debug.print("{d}->{d}\n", .{node.id, n.id});
|
||||||
|
printGraph(n);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn main() !void {
|
||||||
|
var gpa = std.heap.DebugAllocator(.{}).init;
|
||||||
|
defer {
|
||||||
|
//_ = gpa.detectLeaks();
|
||||||
|
}
|
||||||
|
const allocator = gpa.allocator();
|
||||||
|
|
||||||
|
var lexer = al.Lexer.init(@constCast("3*2+2.2"));
|
||||||
|
var parser = al.Parser.init(allocator, &lexer);
|
||||||
|
defer parser.deinit();
|
||||||
|
const graph = try parser.buildGraph();
|
||||||
|
defer graph.?.deinit(&parser);
|
||||||
|
std.debug.print("digraph G {{\n", .{});
|
||||||
|
nodeName(graph.?);
|
||||||
|
printGraph(graph.?);
|
||||||
|
std.debug.print("}}\n", .{});
|
||||||
|
}
|
||||||
2
src/root.zig
Normal file
2
src/root.zig
Normal file
|
|
@ -0,0 +1,2 @@
|
||||||
|
pub const Lexer = @import("Lexer.zig");
|
||||||
|
pub const Parser = @import("Parser.zig");
|
||||||
Loading…
Add table
Add a link
Reference in a new issue