forked from mirror/vulkan-zig
This also includes a workaround for the fact that @"type" refers to the builtin and not to a variable called "type". See ziglang/zig#2897.
618 lines
18 KiB
Zig
618 lines
18 KiB
Zig
const std = @import("std");
|
|
const registry = @import("registry.zig");
|
|
const xml = @import("../xml.zig");
|
|
const mem = std.mem;
|
|
const Allocator = mem.Allocator;
|
|
const testing = std.testing;
|
|
const ArraySize = registry.Array.ArraySize;
|
|
const TypeInfo = registry.TypeInfo;
|
|
|
|
pub const Token = struct {
|
|
kind: Kind,
|
|
text: []const u8,
|
|
|
|
const Kind = enum {
|
|
id, // Any id thats not a keyword
|
|
name, // Vulkan <name>...</name>
|
|
type_name, // Vulkan <type>...</type>
|
|
enum_name, // Vulkan <enum>...</enum>
|
|
int,
|
|
star,
|
|
comma,
|
|
semicolon,
|
|
colon,
|
|
minus,
|
|
tilde,
|
|
dot,
|
|
hash,
|
|
lparen,
|
|
rparen,
|
|
lbracket,
|
|
rbracket,
|
|
kw_typedef,
|
|
kw_const,
|
|
kw_vkapi_ptr,
|
|
kw_struct,
|
|
};
|
|
};
|
|
|
|
pub const CTokenizer = struct {
|
|
source: []const u8,
|
|
offset: usize = 0,
|
|
in_comment: bool = false,
|
|
|
|
fn peek(self: CTokenizer) ?u8 {
|
|
return if (self.offset < self.source.len) self.source[self.offset] else null;
|
|
}
|
|
|
|
fn consumeNoEof(self: *CTokenizer) u8 {
|
|
const c = self.peek().?;
|
|
self.offset += 1;
|
|
return c;
|
|
}
|
|
|
|
fn consume(self: *CTokenizer) !u8 {
|
|
return if (self.offset < self.source.len)
|
|
return self.consumeNoEof()
|
|
else
|
|
return null;
|
|
}
|
|
|
|
fn keyword(self: *CTokenizer) Token {
|
|
const start = self.offset;
|
|
_ = self.consumeNoEof();
|
|
|
|
while (true) {
|
|
const c = self.peek() orelse break;
|
|
switch (c) {
|
|
'A'...'Z', 'a'...'z', '_', '0'...'9' => _ = self.consumeNoEof(),
|
|
else => break,
|
|
}
|
|
}
|
|
|
|
const token_text = self.source[start..self.offset];
|
|
|
|
const kind = if (mem.eql(u8, token_text, "typedef"))
|
|
Token.Kind.kw_typedef
|
|
else if (mem.eql(u8, token_text, "const"))
|
|
Token.Kind.kw_const
|
|
else if (mem.eql(u8, token_text, "VKAPI_PTR"))
|
|
Token.Kind.kw_vkapi_ptr
|
|
else if (mem.eql(u8, token_text, "struct"))
|
|
Token.Kind.kw_struct
|
|
else
|
|
Token.Kind.id;
|
|
|
|
return .{ .kind = kind, .text = token_text };
|
|
}
|
|
|
|
fn int(self: *CTokenizer) Token {
|
|
const start = self.offset;
|
|
_ = self.consumeNoEof();
|
|
|
|
while (true) {
|
|
const c = self.peek() orelse break;
|
|
switch (c) {
|
|
'0'...'9' => _ = self.consumeNoEof(),
|
|
else => break,
|
|
}
|
|
}
|
|
|
|
return .{
|
|
.kind = .int,
|
|
.text = self.source[start..self.offset],
|
|
};
|
|
}
|
|
|
|
fn skipws(self: *CTokenizer) void {
|
|
while (true) {
|
|
switch (self.peek() orelse break) {
|
|
' ', '\t', '\n', '\r' => _ = self.consumeNoEof(),
|
|
else => break,
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn next(self: *CTokenizer) !?Token {
|
|
self.skipws();
|
|
if (mem.startsWith(u8, self.source[self.offset..], "//") or self.in_comment) {
|
|
const end = mem.indexOfScalarPos(u8, self.source, self.offset, '\n') orelse {
|
|
self.offset = self.source.len;
|
|
self.in_comment = true;
|
|
return null;
|
|
};
|
|
self.in_comment = false;
|
|
self.offset = end + 1;
|
|
}
|
|
self.skipws();
|
|
|
|
const c = self.peek() orelse return null;
|
|
var kind: Token.Kind = undefined;
|
|
switch (c) {
|
|
'A'...'Z', 'a'...'z', '_' => return self.keyword(),
|
|
'0'...'9' => return self.int(),
|
|
'*' => kind = .star,
|
|
',' => kind = .comma,
|
|
';' => kind = .semicolon,
|
|
':' => kind = .colon,
|
|
'-' => kind = .minus,
|
|
'~' => kind = .tilde,
|
|
'.' => kind = .dot,
|
|
'#' => kind = .hash,
|
|
'[' => kind = .lbracket,
|
|
']' => kind = .rbracket,
|
|
'(' => kind = .lparen,
|
|
')' => kind = .rparen,
|
|
else => return error.UnexpectedCharacter,
|
|
}
|
|
|
|
const start = self.offset;
|
|
_ = self.consumeNoEof();
|
|
return Token{ .kind = kind, .text = self.source[start..self.offset] };
|
|
}
|
|
};
|
|
|
|
pub const XmlCTokenizer = struct {
|
|
it: xml.Element.ChildIterator,
|
|
ctok: ?CTokenizer = null,
|
|
current: ?Token = null,
|
|
|
|
pub fn init(elem: *xml.Element) XmlCTokenizer {
|
|
return .{
|
|
.it = elem.iterator(),
|
|
};
|
|
}
|
|
|
|
fn elemToToken(elem: *xml.Element) !?Token {
|
|
if (elem.children.items.len != 1 or elem.children.items[0] != .CharData) {
|
|
return error.InvalidXml;
|
|
}
|
|
|
|
const text = elem.children.items[0].CharData;
|
|
if (mem.eql(u8, elem.tag, "type")) {
|
|
return Token{ .kind = .type_name, .text = text };
|
|
} else if (mem.eql(u8, elem.tag, "enum")) {
|
|
return Token{ .kind = .enum_name, .text = text };
|
|
} else if (mem.eql(u8, elem.tag, "name")) {
|
|
return Token{ .kind = .name, .text = text };
|
|
} else if (mem.eql(u8, elem.tag, "comment")) {
|
|
return null;
|
|
} else {
|
|
return error.InvalidTag;
|
|
}
|
|
}
|
|
|
|
fn next(self: *XmlCTokenizer) !?Token {
|
|
if (self.current) |current| {
|
|
const token = current;
|
|
self.current = null;
|
|
return token;
|
|
}
|
|
|
|
var in_comment: bool = false;
|
|
|
|
while (true) {
|
|
if (self.ctok) |*ctok| {
|
|
if (try ctok.next()) |tok| {
|
|
return tok;
|
|
}
|
|
in_comment = ctok.in_comment;
|
|
}
|
|
|
|
self.ctok = null;
|
|
|
|
if (self.it.next()) |child| {
|
|
switch (child.*) {
|
|
.CharData => |cdata| self.ctok = CTokenizer{ .source = cdata, .in_comment = in_comment },
|
|
.Comment => {}, // xml comment
|
|
.Element => |elem| if (!in_comment) if (try elemToToken(elem)) |tok| return tok,
|
|
}
|
|
} else {
|
|
return null;
|
|
}
|
|
}
|
|
}
|
|
|
|
fn nextNoEof(self: *XmlCTokenizer) !Token {
|
|
return (try self.next()) orelse return error.UnexpectedEof;
|
|
}
|
|
|
|
fn peek(self: *XmlCTokenizer) !?Token {
|
|
if (self.current) |current| {
|
|
return current;
|
|
}
|
|
|
|
self.current = try self.next();
|
|
return self.current;
|
|
}
|
|
|
|
fn peekNoEof(self: *XmlCTokenizer) !Token {
|
|
return (try self.peek()) orelse return error.UnexpectedEof;
|
|
}
|
|
|
|
fn expect(self: *XmlCTokenizer, kind: Token.Kind) !Token {
|
|
const tok = (try self.next()) orelse return error.UnexpectedEof;
|
|
if (tok.kind != kind) {
|
|
return error.UnexpectedToken;
|
|
}
|
|
|
|
return tok;
|
|
}
|
|
};
|
|
|
|
// TYPEDEF = kw_typedef DECLARATION ';'
|
|
pub fn parseTypedef(allocator: *Allocator, xctok: *XmlCTokenizer) !registry.Declaration {
|
|
_ = try xctok.expect(.kw_typedef);
|
|
const decl = try parseDeclaration(allocator, xctok);
|
|
_ = try xctok.expect(.semicolon);
|
|
if (try xctok.peek()) |_| {
|
|
return error.InvalidSyntax;
|
|
}
|
|
|
|
return registry.Declaration{
|
|
.name = decl.name orelse return error.MissingTypeIdentifier,
|
|
.decl_type = .{ .typedef = decl.decl_type },
|
|
};
|
|
}
|
|
|
|
// MEMBER = DECLARATION (':' int)?
|
|
pub fn parseMember(allocator: *Allocator, xctok: *XmlCTokenizer) !registry.Container.Field {
|
|
const decl = try parseDeclaration(allocator, xctok);
|
|
var field = registry.Container.Field{
|
|
.name = decl.name orelse return error.MissingTypeIdentifier,
|
|
.field_type = decl.decl_type,
|
|
.bits = null,
|
|
.is_buffer_len = false,
|
|
};
|
|
|
|
if (try xctok.peek()) |tok| {
|
|
if (tok.kind != .colon) {
|
|
return error.InvalidSyntax;
|
|
}
|
|
|
|
_ = try xctok.nextNoEof();
|
|
const bits = try xctok.expect(.int);
|
|
field.bits = try std.fmt.parseInt(usize, bits.text, 10);
|
|
|
|
// Assume for now that there won't be any invalid C types like `char char* x : 4`.
|
|
|
|
if (try xctok.peek()) |_| {
|
|
return error.InvalidSyntax;
|
|
}
|
|
}
|
|
|
|
return field;
|
|
}
|
|
|
|
pub fn parseParamOrProto(allocator: *Allocator, xctok: *XmlCTokenizer) !registry.Declaration {
|
|
const decl = try parseDeclaration(allocator, xctok);
|
|
if (try xctok.peek()) |_| {
|
|
return error.InvalidSyntax;
|
|
}
|
|
return registry.Declaration{
|
|
.name = decl.name orelse return error.MissingTypeIdentifier,
|
|
.decl_type = .{ .typedef = decl.decl_type },
|
|
};
|
|
}
|
|
|
|
pub const Declaration = struct {
|
|
name: ?[]const u8, // Parameter names may be optional, especially in case of func(void)
|
|
decl_type: TypeInfo,
|
|
};
|
|
|
|
pub const ParseError = error{
|
|
OutOfMemory,
|
|
InvalidSyntax,
|
|
InvalidTag,
|
|
InvalidXml,
|
|
Overflow,
|
|
UnexpectedEof,
|
|
UnexpectedCharacter,
|
|
UnexpectedToken,
|
|
MissingTypeIdentifier,
|
|
};
|
|
|
|
// DECLARATION = kw_const? type_name DECLARATOR
|
|
// DECLARATOR = POINTERS (id | name)? ('[' ARRAY_DECLARATOR ']')*
|
|
// | POINTERS '(' FNPTRSUFFIX
|
|
fn parseDeclaration(allocator: *Allocator, xctok: *XmlCTokenizer) ParseError!Declaration {
|
|
// Parse declaration constness
|
|
var tok = try xctok.nextNoEof();
|
|
const inner_is_const = tok.kind == .kw_const;
|
|
if (inner_is_const) {
|
|
tok = try xctok.nextNoEof();
|
|
}
|
|
|
|
if (tok.kind == .kw_struct) {
|
|
tok = try xctok.nextNoEof();
|
|
}
|
|
// Parse type name
|
|
if (tok.kind != .type_name and tok.kind != .id) return error.InvalidSyntax;
|
|
const type_name = tok.text;
|
|
|
|
var type_info = TypeInfo{ .name = type_name };
|
|
|
|
// Parse pointers
|
|
type_info = try parsePointers(allocator, xctok, inner_is_const, type_info);
|
|
|
|
// Parse name / fn ptr
|
|
|
|
if (try parseFnPtrSuffix(allocator, xctok, type_info)) |decl| {
|
|
return decl;
|
|
}
|
|
|
|
const name = blk: {
|
|
const name_tok = (try xctok.peek()) orelse break :blk null;
|
|
if (name_tok.kind == .id or name_tok.kind == .name) {
|
|
_ = try xctok.nextNoEof();
|
|
break :blk name_tok.text;
|
|
} else {
|
|
break :blk null;
|
|
}
|
|
};
|
|
|
|
var inner_type = &type_info;
|
|
while (try parseArrayDeclarator(xctok)) |array_size| {
|
|
// Move the current inner type to a new node on the heap
|
|
const child = try allocator.create(TypeInfo);
|
|
child.* = inner_type.*;
|
|
|
|
// Re-assign the previous inner type for the array type info node
|
|
inner_type.* = .{
|
|
.array = .{
|
|
.size = array_size,
|
|
.child = child,
|
|
},
|
|
};
|
|
|
|
// update the inner_type pointer so it points to the proper
|
|
// inner type again
|
|
inner_type = child;
|
|
}
|
|
|
|
return Declaration{
|
|
.name = name,
|
|
.decl_type = type_info,
|
|
};
|
|
}
|
|
|
|
// FNPTRSUFFIX = kw_vkapi_ptr '*' name' ')' '(' ('void' | (DECLARATION (',' DECLARATION)*)?) ')'
|
|
fn parseFnPtrSuffix(allocator: *Allocator, xctok: *XmlCTokenizer, return_type: TypeInfo) !?Declaration {
|
|
const lparen = try xctok.peek();
|
|
if (lparen == null or lparen.?.kind != .lparen) {
|
|
return null;
|
|
}
|
|
_ = try xctok.nextNoEof();
|
|
_ = try xctok.expect(.kw_vkapi_ptr);
|
|
_ = try xctok.expect(.star);
|
|
const name = try xctok.expect(.name);
|
|
_ = try xctok.expect(.rparen);
|
|
_ = try xctok.expect(.lparen);
|
|
|
|
const return_type_heap = try allocator.create(TypeInfo);
|
|
return_type_heap.* = return_type;
|
|
|
|
var command_ptr = Declaration{
|
|
.name = name.text,
|
|
.decl_type = .{
|
|
.command_ptr = .{
|
|
.params = &[_]registry.Command.Param{},
|
|
.return_type = return_type_heap,
|
|
.success_codes = &[_][]const u8{},
|
|
.error_codes = &[_][]const u8{},
|
|
},
|
|
},
|
|
};
|
|
|
|
const first_param = try parseDeclaration(allocator, xctok);
|
|
if (first_param.name == null) {
|
|
if (first_param.decl_type != .name or !mem.eql(u8, first_param.decl_type.name, "void")) {
|
|
return error.InvalidSyntax;
|
|
}
|
|
|
|
_ = try xctok.expect(.rparen);
|
|
return command_ptr;
|
|
}
|
|
|
|
// There is no good way to estimate the number of parameters beforehand.
|
|
// Fortunately, there are usually a relatively low number of parameters to a function pointer,
|
|
// so an ArrayList backed by an arena allocator is good enough.
|
|
var params = std.ArrayList(registry.Command.Param).init(allocator);
|
|
try params.append(.{
|
|
.name = first_param.name.?,
|
|
.param_type = first_param.decl_type,
|
|
.is_buffer_len = false,
|
|
});
|
|
|
|
while (true) {
|
|
switch ((try xctok.peekNoEof()).kind) {
|
|
.rparen => break,
|
|
.comma => _ = try xctok.nextNoEof(),
|
|
else => return error.InvalidSyntax,
|
|
}
|
|
|
|
const decl = try parseDeclaration(allocator, xctok);
|
|
try params.append(.{
|
|
.name = decl.name orelse return error.MissingTypeIdentifier,
|
|
.param_type = decl.decl_type,
|
|
.is_buffer_len = false,
|
|
});
|
|
}
|
|
|
|
_ = try xctok.nextNoEof();
|
|
command_ptr.decl_type.command_ptr.params = params.toOwnedSlice();
|
|
return command_ptr;
|
|
}
|
|
|
|
// POINTERS = (kw_const? '*')*
|
|
fn parsePointers(allocator: *Allocator, xctok: *XmlCTokenizer, inner_const: bool, inner: TypeInfo) !TypeInfo {
|
|
var type_info = inner;
|
|
var first_const = inner_const;
|
|
|
|
while (true) {
|
|
var tok = (try xctok.peek()) orelse return type_info;
|
|
var is_const = first_const;
|
|
first_const = false;
|
|
|
|
if (tok.kind == .kw_const) {
|
|
is_const = true;
|
|
_ = try xctok.nextNoEof();
|
|
tok = (try xctok.peek()) orelse return type_info;
|
|
}
|
|
|
|
if (tok.kind != .star) {
|
|
// if `is_const` is true at this point, there was a trailing const,
|
|
// and the declaration itself is const.
|
|
return type_info;
|
|
}
|
|
|
|
_ = try xctok.nextNoEof();
|
|
|
|
const child = try allocator.create(TypeInfo);
|
|
child.* = type_info;
|
|
|
|
type_info = .{
|
|
.pointer = .{
|
|
.is_const = is_const or first_const,
|
|
.is_optional = false, // set elsewhere
|
|
.size = .one, // set elsewhere
|
|
.child = child,
|
|
},
|
|
};
|
|
}
|
|
}
|
|
|
|
// ARRAY_DECLARATOR = '[' (int | enum_name) ']'
|
|
fn parseArrayDeclarator(xctok: *XmlCTokenizer) !?ArraySize {
|
|
const lbracket = try xctok.peek();
|
|
if (lbracket == null or lbracket.?.kind != .lbracket) {
|
|
return null;
|
|
}
|
|
|
|
_ = try xctok.nextNoEof();
|
|
|
|
const size_tok = try xctok.nextNoEof();
|
|
const size: ArraySize = switch (size_tok.kind) {
|
|
.int => .{
|
|
.int = std.fmt.parseInt(usize, size_tok.text, 10) catch |err| switch (err) {
|
|
error.Overflow => return error.Overflow,
|
|
error.InvalidCharacter => unreachable,
|
|
},
|
|
},
|
|
.enum_name => .{ .alias = size_tok.text },
|
|
else => return error.InvalidSyntax,
|
|
};
|
|
|
|
_ = try xctok.expect(.rbracket);
|
|
return size;
|
|
}
|
|
|
|
pub fn parseVersion(xctok: *XmlCTokenizer) ![4][]const u8 {
|
|
_ = try xctok.expect(.hash);
|
|
const define = try xctok.expect(.id);
|
|
if (!mem.eql(u8, define.text, "define")) {
|
|
return error.InvalidVersion;
|
|
}
|
|
|
|
_ = try xctok.expect(.name);
|
|
const vk_make_version = try xctok.expect(.type_name);
|
|
if (!mem.eql(u8, vk_make_version.text, "VK_MAKE_API_VERSION")) {
|
|
return error.NotVersion;
|
|
}
|
|
|
|
_ = try xctok.expect(.lparen);
|
|
var version: [4][]const u8 = undefined;
|
|
for (version) |*part, i| {
|
|
if (i != 0) {
|
|
_ = try xctok.expect(.comma);
|
|
}
|
|
|
|
const tok = try xctok.nextNoEof();
|
|
switch (tok.kind) {
|
|
.id, .int => part.* = tok.text,
|
|
else => return error.UnexpectedToken,
|
|
}
|
|
}
|
|
_ = try xctok.expect(.rparen);
|
|
return version;
|
|
}
|
|
|
|
fn testTokenizer(tokenizer: anytype, expected_tokens: []const Token) !void {
|
|
for (expected_tokens) |expected| {
|
|
const tok = (tokenizer.next() catch unreachable).?;
|
|
try testing.expectEqual(expected.kind, tok.kind);
|
|
try testing.expectEqualSlices(u8, expected.text, tok.text);
|
|
}
|
|
|
|
if (tokenizer.next() catch unreachable) |_| unreachable;
|
|
}
|
|
|
|
test "CTokenizer" {
|
|
var ctok = CTokenizer{ .source = "typedef ([const)]** VKAPI_PTR 123,;aaaa" };
|
|
|
|
try testTokenizer(&ctok, &[_]Token{
|
|
.{ .kind = .kw_typedef, .text = "typedef" },
|
|
.{ .kind = .lparen, .text = "(" },
|
|
.{ .kind = .lbracket, .text = "[" },
|
|
.{ .kind = .kw_const, .text = "const" },
|
|
.{ .kind = .rparen, .text = ")" },
|
|
.{ .kind = .rbracket, .text = "]" },
|
|
.{ .kind = .star, .text = "*" },
|
|
.{ .kind = .star, .text = "*" },
|
|
.{ .kind = .kw_vkapi_ptr, .text = "VKAPI_PTR" },
|
|
.{ .kind = .int, .text = "123" },
|
|
.{ .kind = .comma, .text = "," },
|
|
.{ .kind = .semicolon, .text = ";" },
|
|
.{ .kind = .id, .text = "aaaa" },
|
|
});
|
|
}
|
|
|
|
test "XmlCTokenizer" {
|
|
const document = try xml.parse(testing.allocator,
|
|
\\<root>// comment <name>commented name</name> <type>commented type</type> trailing
|
|
\\ typedef void (VKAPI_PTR *<name>PFN_vkVoidFunction</name>)(void);
|
|
\\</root>
|
|
);
|
|
defer document.deinit();
|
|
|
|
var xctok = XmlCTokenizer.init(document.root);
|
|
|
|
try testTokenizer(&xctok, &[_]Token{
|
|
.{ .kind = .kw_typedef, .text = "typedef" },
|
|
.{ .kind = .id, .text = "void" },
|
|
.{ .kind = .lparen, .text = "(" },
|
|
.{ .kind = .kw_vkapi_ptr, .text = "VKAPI_PTR" },
|
|
.{ .kind = .star, .text = "*" },
|
|
.{ .kind = .name, .text = "PFN_vkVoidFunction" },
|
|
.{ .kind = .rparen, .text = ")" },
|
|
.{ .kind = .lparen, .text = "(" },
|
|
.{ .kind = .id, .text = "void" },
|
|
.{ .kind = .rparen, .text = ")" },
|
|
.{ .kind = .semicolon, .text = ";" },
|
|
});
|
|
}
|
|
|
|
test "parseTypedef" {
|
|
const document = try xml.parse(testing.allocator,
|
|
\\<root> // comment <name>commented name</name> trailing
|
|
\\ typedef const struct <type>Python</type>* pythons[4];
|
|
\\ // more comments
|
|
\\</root>
|
|
\\
|
|
);
|
|
defer document.deinit();
|
|
|
|
var arena = std.heap.ArenaAllocator.init(testing.allocator);
|
|
defer arena.deinit();
|
|
|
|
var xctok = XmlCTokenizer.init(document.root);
|
|
const decl = try parseTypedef(&arena.allocator, &xctok);
|
|
|
|
try testing.expectEqualSlices(u8, "pythons", decl.name);
|
|
const array = decl.decl_type.typedef.array;
|
|
try testing.expectEqual(ArraySize{ .int = 4 }, array.size);
|
|
const ptr = array.child.pointer;
|
|
try testing.expectEqual(true, ptr.is_const);
|
|
try testing.expectEqualSlices(u8, "Python", ptr.child.name);
|
|
}
|