diff --git a/src/codegen/spirv.zig b/src/codegen/spirv.zig index e7e6ebdef3..0b2fc2d037 100644 --- a/src/codegen/spirv.zig +++ b/src/codegen/spirv.zig @@ -209,6 +209,10 @@ const DeclGen = struct { /// See Object.type_map type_map: *TypeMap, + /// Child types of pointers that are currently in progress of being resolved. If a pointer + /// is already in this map, its recursive. + wip_pointers: std.AutoHashMapUnmanaged(struct { InternPool.Index, StorageClass }, CacheRef) = .{}, + /// We need to keep track of result ids for block labels, as well as the 'incoming' /// blocks for a block. blocks: BlockMap = .{}, @@ -295,6 +299,7 @@ const DeclGen = struct { pub fn deinit(self: *DeclGen) void { self.args.deinit(self.gpa); self.inst_results.deinit(self.gpa); + self.wip_pointers.deinit(self.gpa); self.blocks.deinit(self.gpa); self.func.deinit(self.gpa); self.base_line_stack.deinit(self.gpa); @@ -358,8 +363,7 @@ const DeclGen = struct { const mod = self.module; const ty = mod.intern_pool.typeOf(val).toType(); - const ty_ref = try self.resolveType(ty, .indirect); - const ptr_ty_ref = try self.spv.ptrType(ty_ref, storage_class); + const ptr_ty_ref = try self.ptrType(ty, storage_class); const var_id = self.spv.declPtr(spv_decl_index).result_id; @@ -582,66 +586,41 @@ const DeclGen = struct { } /// Construct a struct at runtime. - /// result_ty_ref must be a struct type. + /// ty must be a struct type. /// Constituents should be in `indirect` representation (as the elements of a struct should be). /// Result is in `direct` representation. - fn constructStruct(self: *DeclGen, result_ty_ref: CacheRef, constituents: []const IdRef) !IdRef { + fn constructStruct(self: *DeclGen, ty: Type, types: []const Type, constituents: []const IdRef) !IdRef { + assert(types.len == constituents.len); // The Khronos LLVM-SPIRV translator crashes because it cannot construct structs which' // operands are not constant. // See https://github.com/KhronosGroup/SPIRV-LLVM-Translator/issues/1349 // For now, just initialize the struct by setting the fields manually... // TODO: Make this OpCompositeConstruct when we can - const ptr_ty_ref = try self.spv.ptrType(result_ty_ref, .Function); - const ptr_composite_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(ptr_ty_ref), - .id_result = ptr_composite_id, - .storage_class = .Function, - }); - - const spv_composite_ty = self.spv.cache.lookup(result_ty_ref).struct_type; - const member_types = spv_composite_ty.member_types; - - for (constituents, member_types, 0..) |constitent_id, member_ty_ref, index| { - const ptr_member_ty_ref = try self.spv.ptrType(member_ty_ref, .Function); + const ptr_composite_id = try self.alloc(ty, .{ .storage_class = .Function }); + for (constituents, types, 0..) |constitent_id, member_ty, index| { + const ptr_member_ty_ref = try self.ptrType(member_ty, .Function); const ptr_id = try self.accessChain(ptr_member_ty_ref, ptr_composite_id, &.{@as(u32, @intCast(index))}); try self.func.body.emit(self.spv.gpa, .OpStore, .{ .pointer = ptr_id, .object = constitent_id, }); } - const result_id = self.spv.allocId(); - try self.func.body.emit(self.spv.gpa, .OpLoad, .{ - .id_result_type = self.typeId(result_ty_ref), - .id_result = result_id, - .pointer = ptr_composite_id, - }); - return result_id; + return try self.load(ty, ptr_composite_id, .{}); } /// Construct an array at runtime. - /// result_ty_ref must be an array type. + /// ty must be an array type. /// Constituents should be in `indirect` representation (as the elements of an array should be). /// Result is in `direct` representation. - fn constructArray(self: *DeclGen, result_ty_ref: CacheRef, constituents: []const IdRef) !IdRef { + fn constructArray(self: *DeclGen, ty: Type, constituents: []const IdRef) !IdRef { // The Khronos LLVM-SPIRV translator crashes because it cannot construct structs which' // operands are not constant. // See https://github.com/KhronosGroup/SPIRV-LLVM-Translator/issues/1349 // For now, just initialize the struct by setting the fields manually... // TODO: Make this OpCompositeConstruct when we can - // TODO: Make this Function storage type - const ptr_ty_ref = try self.spv.ptrType(result_ty_ref, .Function); - const ptr_composite_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(ptr_ty_ref), - .id_result = ptr_composite_id, - .storage_class = .Function, - }); - - const spv_composite_ty = self.spv.cache.lookup(result_ty_ref).array_type; - const elem_ty_ref = spv_composite_ty.element_type; - const ptr_elem_ty_ref = try self.spv.ptrType(elem_ty_ref, .Function); - + const mod = self.module; + const ptr_composite_id = try self.alloc(ty, .{ .storage_class = .Function }); + const ptr_elem_ty_ref = try self.ptrType(ty.elemType2(mod), .Function); for (constituents, 0..) |constitent_id, index| { const ptr_id = try self.accessChain(ptr_elem_ty_ref, ptr_composite_id, &.{@as(u32, @intCast(index))}); try self.func.body.emit(self.spv.gpa, .OpStore, .{ @@ -649,13 +628,8 @@ const DeclGen = struct { .object = constitent_id, }); } - const result_id = self.spv.allocId(); - try self.func.body.emit(self.spv.gpa, .OpLoad, .{ - .id_result_type = self.typeId(result_ty_ref), - .id_result = result_id, - .pointer = ptr_composite_id, - }); - return result_id; + + return try self.load(ty, ptr_composite_id, .{}); } /// This function generates a load for a constant in direct (ie, non-memory) representation. @@ -766,15 +740,18 @@ const DeclGen = struct { }.toValue(); var constituents: [2]IdRef = undefined; + var types: [2]Type = undefined; if (eu_layout.error_first) { constituents[0] = try self.constant(err_ty, err_val, .indirect); constituents[1] = try self.constant(payload_ty, payload_val, .indirect); + types = .{ err_ty, payload_ty }; } else { constituents[0] = try self.constant(payload_ty, payload_val, .indirect); constituents[1] = try self.constant(err_ty, err_val, .indirect); + types = .{ payload_ty, err_ty }; } - return try self.constructStruct(result_ty_ref, &constituents); + return try self.constructStruct(ty, &types, &constituents); }, .enum_tag => { const int_val = try val.intFromEnum(ty, mod); @@ -792,7 +769,11 @@ const DeclGen = struct { } const len_id = try self.constant(Type.usize, ptr.len.toValue(), .indirect); - return try self.constructStruct(result_ty_ref, &.{ ptr_id, len_id }); + return try self.constructStruct( + ty, + &.{ ptr_ty, Type.usize }, + &.{ ptr_id, len_id }, + ); }, .opt => { const payload_ty = ty.optionalChild(mod); @@ -819,7 +800,11 @@ const DeclGen = struct { else try self.spv.constUndef(try self.resolveType(payload_ty, .indirect)); - return try self.constructStruct(result_ty_ref, &.{ payload_id, has_pl_id }); + return try self.constructStruct( + ty, + &.{ payload_ty, Type.bool }, + &.{ payload_id, has_pl_id }, + ); }, .aggregate => |aggregate| switch (ip.indexToKey(ty.ip_index)) { inline .array_type, .vector_type => |array_type, tag| { @@ -857,7 +842,7 @@ const DeclGen = struct { else => {}, } - return try self.constructArray(result_ty_ref, constituents); + return try self.constructArray(ty, constituents); }, .struct_type => { const struct_type = mod.typeToStruct(ty).?; @@ -865,6 +850,9 @@ const DeclGen = struct { return self.todo("packed struct constants", .{}); } + var types = std.ArrayList(Type).init(self.gpa); + defer types.deinit(); + var constituents = std.ArrayList(IdRef).init(self.gpa); defer constituents.deinit(); @@ -880,22 +868,23 @@ const DeclGen = struct { const field_val = try val.fieldValue(mod, field_index); const field_id = try self.constant(field_ty, field_val, .indirect); + try types.append(field_ty); try constituents.append(field_id); } - return try self.constructStruct(result_ty_ref, constituents.items); + return try self.constructStruct(ty, types.items, constituents.items); }, .anon_struct_type => unreachable, // TODO else => unreachable, }, .un => |un| { const active_field = ty.unionTagFieldIndex(un.tag.toValue(), mod).?; - const layout = self.unionLayout(ty, active_field); - const payload = if (layout.active_field_size != 0) - try self.constant(layout.active_field_ty, un.val.toValue(), .indirect) + const union_obj = mod.typeToUnion(ty).?; + const field_ty = union_obj.field_types.get(ip)[active_field].toType(); + const payload = if (field_ty.hasRuntimeBitsIgnoreComptime(mod)) + try self.constant(field_ty, un.val.toValue(), .direct) else null; - return try self.unionInit(ty, active_field, payload); }, .memoized_call => unreachable, @@ -934,8 +923,7 @@ const DeclGen = struct { // TODO: Can we consolidate this in ptrElemPtr? const elem_ty = parent_ptr_ty.elemType2(mod); // use elemType() so that we get T for *[N]T. - const elem_ty_ref = try self.resolveType(elem_ty, .direct); - const elem_ptr_ty_ref = try self.spv.ptrType(elem_ty_ref, spvStorageClass(parent_ptr_ty.ptrAddressSpace(mod))); + const elem_ptr_ty_ref = try self.ptrType(elem_ty, spvStorageClass(parent_ptr_ty.ptrAddressSpace(mod))); if (elem_ptr_ty_ref == result_ty_ref) { return elem_ptr_id; @@ -997,8 +985,7 @@ const DeclGen = struct { }; const decl_id = try self.resolveAnonDecl(decl_val, actual_storage_class); - const decl_ty_ref = try self.resolveType(decl_ty, .indirect); - const decl_ptr_ty_ref = try self.spv.ptrType(decl_ty_ref, final_storage_class); + const decl_ptr_ty_ref = try self.ptrType(decl_ty, final_storage_class); const ptr_id = switch (final_storage_class) { .Generic => blk: { @@ -1054,8 +1041,7 @@ const DeclGen = struct { const final_storage_class = spvStorageClass(decl.@"addrspace"); - const decl_ty_ref = try self.resolveType(decl.ty, .indirect); - const decl_ptr_ty_ref = try self.spv.ptrType(decl_ty_ref, final_storage_class); + const decl_ptr_ty_ref = try self.ptrType(decl.ty, final_storage_class); const ptr_id = switch (final_storage_class) { .Generic => blk: { @@ -1123,29 +1109,52 @@ const DeclGen = struct { return try self.intType(.unsigned, self.getTarget().ptrBitWidth()); } - /// Generate a union type, optionally with a known field. If the tag alignment is greater - /// than that of the payload, a regular union (non-packed, with both tag and payload), will - /// be generated as follows: - /// If the active field is known: + fn ptrType(self: *DeclGen, child_ty: Type, storage_class: StorageClass) !CacheRef { + const key = .{ child_ty.toIntern(), storage_class }; + const entry = try self.wip_pointers.getOrPut(self.gpa, key); + if (entry.found_existing) { + const fwd_ref = entry.value_ptr.*; + try self.spv.cache.recursive_ptrs.put(self.spv.gpa, fwd_ref, {}); + return fwd_ref; + } + + const fwd_ref = try self.spv.resolve(.{ .fwd_ptr_type = .{ + .zig_child_type = child_ty.toIntern(), + .storage_class = storage_class, + } }); + entry.value_ptr.* = fwd_ref; + + const child_ty_ref = try self.resolveType(child_ty, .indirect); + _ = try self.spv.resolve(.{ .ptr_type = .{ + .storage_class = storage_class, + .child_type = child_ty_ref, + .fwd = fwd_ref, + } }); + + assert(self.wip_pointers.remove(key)); + + return fwd_ref; + } + + /// Generate a union type. Union types are always generated with the + /// most aligned field active. If the tag alignment is greater + /// than that of the payload, a regular union (non-packed, with both tag and + /// payload), will be generated as follows: /// struct { /// tag: TagType, - /// payload: ActivePayloadType, - /// payload_padding: [payload_size - @sizeOf(ActivePayloadType)]u8, + /// payload: MostAlignedFieldType, + /// payload_padding: [payload_size - @sizeOf(MostAlignedFieldType)]u8, /// padding: [padding_size]u8, /// } /// If the payload alignment is greater than that of the tag: /// struct { - /// payload: ActivePayloadType, - /// payload_padding: [payload_size - @sizeOf(ActivePayloadType)]u8, + /// payload: MostAlignedFieldType, + /// payload_padding: [payload_size - @sizeOf(MostAlignedFieldType)]u8, /// tag: TagType, /// padding: [padding_size]u8, /// } - /// If the active payload is unknown, it will default back to the most aligned field. This is - /// to make sure that the overal struct has the correct alignment in spir-v. /// If any of the fields' size is 0, it will be omitted. - /// NOTE: When the active field is set to something other than the most aligned field, the - /// resulting struct will be *underaligned*. - fn resolveUnionType(self: *DeclGen, ty: Type, maybe_active_field: ?usize) !CacheRef { + fn resolveUnionType(self: *DeclGen, ty: Type) !CacheRef { const mod = self.module; const ip = &mod.intern_pool; const union_obj = mod.typeToUnion(ty).?; @@ -1154,17 +1163,13 @@ const DeclGen = struct { return self.todo("packed union types", .{}); } - const layout = self.unionLayout(ty, maybe_active_field); - - if (layout.payload_size == 0) { + const layout = self.unionLayout(ty); + if (!layout.has_payload) { // No payload, so represent this as just the tag type. return try self.resolveType(union_obj.enum_tag_ty.toType(), .indirect); } - // TODO: We need to add the active field to the key, somehow. - if (maybe_active_field == null) { - if (self.type_map.get(ty.toIntern())) |info| return info.ty_ref; - } + if (self.type_map.get(ty.toIntern())) |info| return info.ty_ref; var member_types: [4]CacheRef = undefined; var member_names: [4]CacheString = undefined; @@ -1177,10 +1182,10 @@ const DeclGen = struct { member_names[layout.tag_index] = try self.spv.resolveString("(tag)"); } - if (layout.active_field_size != 0) { - const active_payload_ty_ref = try self.resolveType(layout.active_field_ty, .indirect); - member_types[layout.active_field_index] = active_payload_ty_ref; - member_names[layout.active_field_index] = try self.spv.resolveString("(payload)"); + if (layout.payload_size != 0) { + const payload_ty_ref = try self.resolveType(layout.payload_ty, .indirect); + member_types[layout.payload_index] = payload_ty_ref; + member_names[layout.payload_index] = try self.spv.resolveString("(payload)"); } if (layout.payload_padding_size != 0) { @@ -1201,9 +1206,7 @@ const DeclGen = struct { .member_names = member_names[0..layout.total_fields], } }); - if (maybe_active_field == null) { - try self.type_map.put(self.gpa, ty.toIntern(), .{ .ty_ref = ty_ref }); - } + try self.type_map.put(self.gpa, ty.toIntern(), .{ .ty_ref = ty_ref }); return ty_ref; } @@ -1351,12 +1354,12 @@ const DeclGen = struct { .Pointer => { const ptr_info = ty.ptrInfo(mod); + // Note: Don't cache this pointer type, it would mess up the recursive pointer functionality + // in ptrType()! + const storage_class = spvStorageClass(ptr_info.flags.address_space); - const child_ty_ref = try self.resolveType(ptr_info.child.toType(), .indirect); - const ptr_ty_ref = try self.spv.resolve(.{ .ptr_type = .{ - .storage_class = storage_class, - .child_type = child_ty_ref, - } }); + const ptr_ty_ref = try self.ptrType(ptr_info.child.toType(), storage_class); + if (ptr_info.flags.size != .Slice) { return ptr_ty_ref; } @@ -1471,7 +1474,7 @@ const DeclGen = struct { try self.type_map.put(self.gpa, ty.toIntern(), .{ .ty_ref = ty_ref }); return ty_ref; }, - .Union => return try self.resolveUnionType(ty, null), + .Union => return try self.resolveUnionType(ty), .ErrorSet => return try self.intType(.unsigned, 16), .ErrorUnion => { const payload_ty = ty.errorUnionPayload(mod); @@ -1585,14 +1588,16 @@ const DeclGen = struct { } const UnionLayout = struct { - active_field: u32, - active_field_ty: Type, - payload_size: u32, - + /// If false, this union is represented + /// by only an integer of the tag type. + has_payload: bool, tag_size: u32, tag_index: u32, - active_field_size: u32, - active_field_index: u32, + /// Note: This is the size of the payload type itself, NOT the size of the ENTIRE payload. + /// Use `has_payload` instead!! + payload_ty: Type, + payload_size: u32, + payload_index: u32, payload_padding_size: u32, payload_padding_index: u32, padding_size: u32, @@ -1600,23 +1605,19 @@ const DeclGen = struct { total_fields: u32, }; - fn unionLayout(self: *DeclGen, ty: Type, maybe_active_field: ?usize) UnionLayout { + fn unionLayout(self: *DeclGen, ty: Type) UnionLayout { const mod = self.module; const ip = &mod.intern_pool; const layout = ty.unionGetLayout(self.module); const union_obj = mod.typeToUnion(ty).?; - const active_field = maybe_active_field orelse layout.most_aligned_field; - const active_field_ty = union_obj.field_types.get(ip)[active_field].toType(); - var union_layout = UnionLayout{ - .active_field = @intCast(active_field), - .active_field_ty = active_field_ty, - .payload_size = @intCast(layout.payload_size), + .has_payload = layout.payload_size != 0, .tag_size = @intCast(layout.tag_size), .tag_index = undefined, - .active_field_size = undefined, - .active_field_index = undefined, + .payload_ty = undefined, + .payload_size = undefined, + .payload_index = undefined, .payload_padding_size = undefined, .payload_padding_index = undefined, .padding_size = @intCast(layout.padding), @@ -1624,11 +1625,16 @@ const DeclGen = struct { .total_fields = undefined, }; - union_layout.active_field_size = if (active_field_ty.hasRuntimeBitsIgnoreComptime(mod)) - @intCast(active_field_ty.abiSize(mod)) - else - 0; - union_layout.payload_padding_size = @intCast(layout.payload_size - union_layout.active_field_size); + if (union_layout.has_payload) { + const most_aligned_field = layout.most_aligned_field; + const most_aligned_field_ty = union_obj.field_types.get(ip)[most_aligned_field].toType(); + union_layout.payload_ty = most_aligned_field_ty; + union_layout.payload_size = @intCast(most_aligned_field_ty.abiSize(mod)); + } else { + union_layout.payload_size = 0; + } + + union_layout.payload_padding_size = @intCast(layout.payload_size - union_layout.payload_size); const tag_first = layout.tag_align.compare(.gte, layout.payload_align); var field_index: u32 = 0; @@ -1638,8 +1644,8 @@ const DeclGen = struct { field_index += 1; } - if (union_layout.active_field_size != 0) { - union_layout.active_field_index = field_index; + if (union_layout.payload_size != 0) { + union_layout.payload_index = field_index; field_index += 1; } @@ -1683,7 +1689,7 @@ const DeclGen = struct { /// the name of an error in the text executor. fn generateTestEntryPoint(self: *DeclGen, name: []const u8, spv_test_decl_index: SpvModule.Decl.Index) !void { const anyerror_ty_ref = try self.resolveType(Type.anyerror, .direct); - const ptr_anyerror_ty_ref = try self.spv.ptrType(anyerror_ty_ref, .CrossWorkgroup); + const ptr_anyerror_ty_ref = try self.ptrType(Type.anyerror, .CrossWorkgroup); const void_ty_ref = try self.resolveType(Type.void, .direct); const kernel_proto_ty_ref = try self.spv.resolve(.{ .function_type = .{ @@ -1718,6 +1724,7 @@ const DeclGen = struct { .id_result = error_id, .function = test_id, }); + // Note: Convert to direct not required. try section.emit(self.spv.gpa, .OpStore, .{ .pointer = p_error_id, .object = error_id, @@ -1822,8 +1829,7 @@ const DeclGen = struct { else => final_storage_class, }; - const ty_ref = try self.resolveType(decl.ty, .indirect); - const ptr_ty_ref = try self.spv.ptrType(ty_ref, actual_storage_class); + const ptr_ty_ref = try self.ptrType(decl.ty, actual_storage_class); const begin = self.spv.beginGlobal(); try self.spv.globals.section.emit(self.spv.gpa, .OpVariable, .{ @@ -1928,11 +1934,15 @@ const DeclGen = struct { return try self.convertToDirect(result_ty, result_id); } - fn load(self: *DeclGen, value_ty: Type, ptr_id: IdRef, is_volatile: bool) !IdRef { + const MemoryOptions = struct { + is_volatile: bool = false, + }; + + fn load(self: *DeclGen, value_ty: Type, ptr_id: IdRef, options: MemoryOptions) !IdRef { const indirect_value_ty_ref = try self.resolveType(value_ty, .indirect); const result_id = self.spv.allocId(); const access = spec.MemoryAccess.Extended{ - .Volatile = is_volatile, + .Volatile = options.is_volatile, }; try self.func.body.emit(self.spv.gpa, .OpLoad, .{ .id_result_type = self.typeId(indirect_value_ty_ref), @@ -1943,10 +1953,10 @@ const DeclGen = struct { return try self.convertToDirect(value_ty, result_id); } - fn store(self: *DeclGen, value_ty: Type, ptr_id: IdRef, value_id: IdRef, is_volatile: bool) !void { + fn store(self: *DeclGen, value_ty: Type, ptr_id: IdRef, value_id: IdRef, options: MemoryOptions) !void { const indirect_value_id = try self.convertToIndirect(value_ty, value_id); const access = spec.MemoryAccess.Extended{ - .Volatile = is_volatile, + .Volatile = options.is_volatile, }; try self.func.body.emit(self.spv.gpa, .OpStore, .{ .pointer = ptr_id, @@ -2118,9 +2128,7 @@ const DeclGen = struct { constituent.* = try self.convertToIndirect(child_ty, result_id); } - const result_ty = try self.resolveType(child_ty, .indirect); - const result_ty_ref = try self.spv.arrayType(vector_len, result_ty); - return try self.constructArray(result_ty_ref, constituents); + return try self.constructArray(ty, constituents); } const result_id = self.spv.allocId(); @@ -2181,7 +2189,7 @@ const DeclGen = struct { const info = try self.arithmeticTypeInfo(result_ty); // TODO: Use fmin for OpenCL - const cmp_id = try self.cmp(op, result_ty, lhs_id, rhs_id); + const cmp_id = try self.cmp(op, Type.bool, result_ty, lhs_id, rhs_id); const selection_id = switch (info.class) { .float => blk: { // cmp uses OpFOrd. When we have 0 [<>] nan this returns false, @@ -2316,7 +2324,7 @@ const DeclGen = struct { constituent.* = try self.arithOp(child_ty, lhs_index_id, rhs_index_id, fop, sop, uop, modular); } - return self.constructArray(result_ty_ref, constituents); + return self.constructArray(ty, constituents); } // Binary operations are generally applicable to both scalar and vector operations @@ -2472,11 +2480,11 @@ const DeclGen = struct { // Construct the struct that Zig wants as result. // The value should already be the correct type. const ov_id = try self.intFromBool(ov_ty_ref, overflowed_id); - const result_ty_ref = try self.resolveType(result_ty, .direct); - return try self.constructStruct(result_ty_ref, &.{ - value_id, - ov_id, - }); + return try self.constructStruct( + result_ty, + &.{ operand_ty, ov_ty }, + &.{ value_id, ov_id }, + ); } fn airShuffle(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -2634,6 +2642,7 @@ const DeclGen = struct { fn cmp( self: *DeclGen, op: std.math.CompareOperator, + result_ty: Type, ty: Type, lhs_id: IdRef, rhs_id: IdRef, @@ -2674,7 +2683,7 @@ const DeclGen = struct { if (ty.optionalReprIsPayload(mod)) { assert(payload_ty.hasRuntimeBitsIgnoreComptime(mod)); assert(!payload_ty.isSlice(mod)); - return self.cmp(op, payload_ty, lhs_id, rhs_id); + return self.cmp(op, Type.bool, payload_ty, lhs_id, rhs_id); } const lhs_valid_id = if (payload_ty.hasRuntimeBitsIgnoreComptime(mod)) @@ -2687,7 +2696,7 @@ const DeclGen = struct { else try self.convertToDirect(Type.bool, rhs_id); - const valid_cmp_id = try self.cmp(op, Type.bool, lhs_valid_id, rhs_valid_id); + const valid_cmp_id = try self.cmp(op, Type.bool, Type.bool, lhs_valid_id, rhs_valid_id); if (!payload_ty.hasRuntimeBitsIgnoreComptime(mod)) { return valid_cmp_id; } @@ -2698,7 +2707,7 @@ const DeclGen = struct { const lhs_pl_id = try self.extractField(payload_ty, lhs_id, 0); const rhs_pl_id = try self.extractField(payload_ty, rhs_id, 0); - const pl_cmp_id = try self.cmp(op, payload_ty, lhs_pl_id, rhs_pl_id); + const pl_cmp_id = try self.cmp(op, Type.bool, payload_ty, lhs_pl_id, rhs_pl_id); // op == .eq => lhs_valid == rhs_valid && lhs_pl == rhs_pl // op == .neq => lhs_valid != rhs_valid || lhs_pl != rhs_pl @@ -2720,7 +2729,6 @@ const DeclGen = struct { .Vector => { const child_ty = ty.childType(mod); const vector_len = ty.vectorLen(mod); - const bool_ty_ref_indirect = try self.resolveType(Type.bool, .indirect); var constituents = try self.gpa.alloc(IdRef, vector_len); defer self.gpa.free(constituents); @@ -2728,12 +2736,11 @@ const DeclGen = struct { for (constituents, 0..) |*constituent, i| { const lhs_index_id = try self.extractField(child_ty, cmp_lhs_id, @intCast(i)); const rhs_index_id = try self.extractField(child_ty, cmp_rhs_id, @intCast(i)); - const result_id = try self.cmp(op, child_ty, lhs_index_id, rhs_index_id); + const result_id = try self.cmp(op, Type.bool, child_ty, lhs_index_id, rhs_index_id); constituent.* = try self.convertToIndirect(Type.bool, result_id); } - const result_ty_ref = try self.spv.arrayType(vector_len, bool_ty_ref_indirect); - return try self.constructArray(result_ty_ref, constituents); + return try self.constructArray(result_ty, constituents); }, else => unreachable, }; @@ -2806,8 +2813,9 @@ const DeclGen = struct { const lhs_id = try self.resolve(bin_op.lhs); const rhs_id = try self.resolve(bin_op.rhs); const ty = self.typeOf(bin_op.lhs); + const result_ty = self.typeOfIndex(inst); - return try self.cmp(op, ty, lhs_id, rhs_id); + return try self.cmp(op, result_ty, ty, lhs_id, rhs_id); } fn airVectorCmp(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -2819,8 +2827,9 @@ const DeclGen = struct { const rhs_id = try self.resolve(vec_cmp.rhs); const op = vec_cmp.compareOperator(); const ty = self.typeOf(vec_cmp.lhs); + const result_ty = self.typeOfIndex(inst); - return try self.cmp(op, ty, lhs_id, rhs_id); + return try self.cmp(op, result_ty, ty, lhs_id, rhs_id); } fn bitCast( @@ -2865,23 +2874,17 @@ const DeclGen = struct { return result_id; } - const src_ptr_ty_ref = try self.spv.ptrType(src_ty_ref, .Function); - const dst_ptr_ty_ref = try self.spv.ptrType(dst_ty_ref, .Function); + const dst_ptr_ty_ref = try self.ptrType(dst_ty, .Function); - const tmp_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(src_ptr_ty_ref), - .id_result = tmp_id, - .storage_class = .Function, - }); - try self.store(src_ty, tmp_id, src_id, false); + const tmp_id = try self.alloc(src_ty, .{ .storage_class = .Function }); + try self.store(src_ty, tmp_id, src_id, .{}); const casted_ptr_id = self.spv.allocId(); try self.func.body.emit(self.spv.gpa, .OpBitcast, .{ .id_result_type = self.typeId(dst_ptr_ty_ref), .id_result = casted_ptr_id, .operand = tmp_id, }); - return try self.load(dst_ty, casted_ptr_id, false); + return try self.load(dst_ty, casted_ptr_id, .{}); } fn airBitCast(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -3060,7 +3063,6 @@ const DeclGen = struct { const elem_ptr_ty = slice_ty.slicePtrFieldType(mod); const elem_ptr_ty_ref = try self.resolveType(elem_ptr_ty, .direct); - const slice_ty_ref = try self.resolveType(slice_ty, .direct); const size_ty_ref = try self.sizeType(); const array_ptr_id = try self.resolve(ty_op.operand); @@ -3073,7 +3075,11 @@ const DeclGen = struct { // Convert the pointer-to-array to a pointer to the first element. try self.accessChain(elem_ptr_ty_ref, array_ptr_id, &.{0}); - return try self.constructStruct(slice_ty_ref, &.{ elem_ptr_id, len_id }); + return try self.constructStruct( + slice_ty, + &.{ elem_ptr_ty, Type.usize }, + &.{ elem_ptr_id, len_id }, + ); } fn airSlice(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -3083,13 +3089,16 @@ const DeclGen = struct { const bin_op = self.air.extraData(Air.Bin, ty_pl.payload).data; const ptr_id = try self.resolve(bin_op.lhs); const len_id = try self.resolve(bin_op.rhs); + const ptr_ty = self.typeOf(bin_op.lhs); const slice_ty = self.typeOfIndex(inst); - const slice_ty_ref = try self.resolveType(slice_ty, .direct); - return try self.constructStruct(slice_ty_ref, &.{ - ptr_id, // Note: Type should not need to be converted to direct. - len_id, // Note: Type should not need to be converted to direct. - }); + // Note: Types should not need to be converted to direct, these types + // dont need to be converted. + return try self.constructStruct( + slice_ty, + &.{ ptr_ty, Type.usize }, + &.{ ptr_id, len_id }, + ); } fn airAggregateInit(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -3099,7 +3108,6 @@ const DeclGen = struct { const ip = &mod.intern_pool; const ty_pl = self.air.instructions.items(.data)[inst].ty_pl; const result_ty = self.typeOfIndex(inst); - const result_ty_ref = try self.resolveType(result_ty, .direct); const len: usize = @intCast(result_ty.arrayLen(mod)); const elements: []const Air.Inst.Ref = @ptrCast(self.air.extra[ty_pl.payload..][0..len]); @@ -3111,6 +3119,8 @@ const DeclGen = struct { unreachable; // TODO } + const types = try self.gpa.alloc(Type, elements.len); + defer self.gpa.free(types); const constituents = try self.gpa.alloc(IdRef, elements.len); defer self.gpa.free(constituents); var index: usize = 0; @@ -3122,6 +3132,7 @@ const DeclGen = struct { assert(field_ty.toType().hasRuntimeBits(mod)); const id = try self.resolve(element); + types[index] = field_ty.toType(); constituents[index] = try self.convertToIndirect(field_ty.toType(), id); index += 1; } @@ -3135,6 +3146,7 @@ const DeclGen = struct { assert(field_ty.hasRuntimeBitsIgnoreComptime(mod)); const id = try self.resolve(element); + types[index] = field_ty; constituents[index] = try self.convertToIndirect(field_ty, id); index += 1; } @@ -3142,7 +3154,11 @@ const DeclGen = struct { else => unreachable, } - return try self.constructStruct(result_ty_ref, constituents[0..index]); + return try self.constructStruct( + result_ty, + types[0..index], + constituents[0..index], + ); }, .Array => { const array_info = result_ty.arrayInfo(mod); @@ -3159,7 +3175,7 @@ const DeclGen = struct { elem_ids[n_elems - 1] = try self.constant(array_info.elem_type, sentinel_val, .indirect); } - return try self.constructArray(result_ty_ref, elem_ids); + return try self.constructArray(result_ty, elem_ids); }, else => unreachable, } @@ -3244,15 +3260,14 @@ const DeclGen = struct { const slice_ptr = try self.extractField(ptr_ty, slice_id, 0); const elem_ptr = try self.ptrAccessChain(ptr_ty_ref, slice_ptr, index_id, &.{}); - return try self.load(slice_ty.childType(mod), elem_ptr, slice_ty.isVolatilePtr(mod)); + return try self.load(slice_ty.childType(mod), elem_ptr, .{ .is_volatile = slice_ty.isVolatilePtr(mod) }); } fn ptrElemPtr(self: *DeclGen, ptr_ty: Type, ptr_id: IdRef, index_id: IdRef) !IdRef { const mod = self.module; // Construct new pointer type for the resulting pointer const elem_ty = ptr_ty.elemType2(mod); // use elemType() so that we get T for *[N]T. - const elem_ty_ref = try self.resolveType(elem_ty, .direct); - const elem_ptr_ty_ref = try self.spv.ptrType(elem_ty_ref, spvStorageClass(ptr_ty.ptrAddressSpace(mod))); + const elem_ptr_ty_ref = try self.ptrType(elem_ty, spvStorageClass(ptr_ty.ptrAddressSpace(mod))); if (ptr_ty.isSinglePointer(mod)) { // Pointer-to-array. In this case, the resulting pointer is not of the same type // as the ptr_ty (we want a *T, not a *[N]T), and hence we need to use accessChain. @@ -3288,9 +3303,7 @@ const DeclGen = struct { const mod = self.module; const bin_op = self.air.instructions.items(.data)[inst].bin_op; const array_ty = self.typeOf(bin_op.lhs); - const array_ty_ref = try self.resolveType(array_ty, .direct); const elem_ty = array_ty.childType(mod); - const elem_ty_ref = try self.resolveType(elem_ty, .indirect); const array_id = try self.resolve(bin_op.lhs); const index_id = try self.resolve(bin_op.rhs); @@ -3298,22 +3311,12 @@ const DeclGen = struct { // For now, just generate a temporary and use that. // TODO: This backend probably also should use isByRef from llvm... - const array_ptr_ty_ref = try self.spv.ptrType(array_ty_ref, .Function); - const elem_ptr_ty_ref = try self.spv.ptrType(elem_ty_ref, .Function); - - const tmp_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(array_ptr_ty_ref), - .id_result = tmp_id, - .storage_class = .Function, - }); - try self.func.body.emit(self.spv.gpa, .OpStore, .{ - .pointer = tmp_id, - .object = array_id, - }); + const elem_ptr_ty_ref = try self.ptrType(elem_ty, .Function); + const tmp_id = try self.alloc(array_ty, .{ .storage_class = .Function }); + try self.store(array_ty, tmp_id, array_id, .{}); const elem_ptr_id = try self.accessChainId(elem_ptr_ty_ref, tmp_id, &.{index_id}); - return try self.load(elem_ty, elem_ptr_id, false); + return try self.load(elem_ty, elem_ptr_id, .{}); } fn airPtrElemVal(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -3326,7 +3329,7 @@ const DeclGen = struct { const ptr_id = try self.resolve(bin_op.lhs); const index_id = try self.resolve(bin_op.rhs); const elem_ptr_id = try self.ptrElemPtr(ptr_ty, ptr_id, index_id); - return try self.load(elem_ty, elem_ptr_id, ptr_ty.isVolatilePtr(mod)); + return try self.load(elem_ty, elem_ptr_id, .{ .is_volatile = ptr_ty.isVolatilePtr(mod) }); } fn airSetUnionTag(self: *DeclGen, inst: Air.Inst.Index) !void { @@ -3334,22 +3337,21 @@ const DeclGen = struct { const bin_op = self.air.instructions.items(.data)[inst].bin_op; const un_ptr_ty = self.typeOf(bin_op.lhs); const un_ty = un_ptr_ty.childType(mod); - const layout = self.unionLayout(un_ty, null); + const layout = self.unionLayout(un_ty); if (layout.tag_size == 0) return; const tag_ty = un_ty.unionTagTypeSafety(mod).?; - const tag_ty_ref = try self.resolveType(tag_ty, .indirect); - const tag_ptr_ty_ref = try self.spv.ptrType(tag_ty_ref, spvStorageClass(un_ptr_ty.ptrAddressSpace(mod))); + const tag_ptr_ty_ref = try self.ptrType(tag_ty, spvStorageClass(un_ptr_ty.ptrAddressSpace(mod))); const union_ptr_id = try self.resolve(bin_op.lhs); const new_tag_id = try self.resolve(bin_op.rhs); - if (layout.payload_size == 0) { - try self.store(tag_ty, union_ptr_id, new_tag_id, un_ptr_ty.isVolatilePtr(mod)); + if (!layout.has_payload) { + try self.store(tag_ty, union_ptr_id, new_tag_id, .{ .is_volatile = un_ptr_ty.isVolatilePtr(mod) }); } else { const ptr_id = try self.accessChain(tag_ptr_ty_ref, union_ptr_id, &.{layout.tag_index}); - try self.store(tag_ty, ptr_id, new_tag_id, un_ptr_ty.isVolatilePtr(mod)); + try self.store(tag_ty, ptr_id, new_tag_id, .{ .is_volatile = un_ptr_ty.isVolatilePtr(mod) }); } } @@ -3360,11 +3362,11 @@ const DeclGen = struct { const un_ty = self.typeOf(ty_op.operand); const mod = self.module; - const layout = self.unionLayout(un_ty, null); + const layout = self.unionLayout(un_ty); if (layout.tag_size == 0) return null; const union_handle = try self.resolve(ty_op.operand); - if (layout.payload_size == 0) return union_handle; + if (!layout.has_payload) return union_handle; const tag_ty = un_ty.unionTagTypeSafety(mod).?; return try self.extractField(tag_ty, union_handle, layout.tag_index); @@ -3377,8 +3379,8 @@ const DeclGen = struct { payload: ?IdRef, ) !IdRef { // To initialize a union, generate a temporary variable with the - // type that has the right field active, then pointer-cast and store - // the active field, and finally load and return the entire union. + // union type, then get the field pointer and pointer-cast it to the + // right type to store it. Finally load the entire union. const mod = self.module; const ip = &mod.intern_pool; @@ -3389,7 +3391,7 @@ const DeclGen = struct { } const maybe_tag_ty = ty.unionTagTypeSafety(mod); - const layout = self.unionLayout(ty, active_field); + const layout = self.unionLayout(ty); const tag_int = if (layout.tag_size != 0) blk: { const tag_ty = maybe_tag_ty.?; @@ -3400,42 +3402,34 @@ const DeclGen = struct { break :blk tag_int_val.toUnsignedInt(mod); } else 0; - if (layout.payload_size == 0) { + if (!layout.has_payload) { const tag_ty_ref = try self.resolveType(maybe_tag_ty.?, .direct); return try self.constInt(tag_ty_ref, tag_int); } - const un_active_ty_ref = try self.resolveUnionType(ty, active_field); - const un_active_ptr_ty_ref = try self.spv.ptrType(un_active_ty_ref, .Function); - const un_general_ty_ref = try self.resolveType(ty, .direct); - const un_general_ptr_ty_ref = try self.spv.ptrType(un_general_ty_ref, .Function); - - const tmp_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(un_active_ptr_ty_ref), - .id_result = tmp_id, - .storage_class = .Function, - }); + const tmp_id = try self.alloc(ty, .{ .storage_class = .Function }); if (layout.tag_size != 0) { const tag_ty_ref = try self.resolveType(maybe_tag_ty.?, .direct); - const tag_ptr_ty_ref = try self.spv.ptrType(tag_ty_ref, .Function); + const tag_ptr_ty_ref = try self.ptrType(maybe_tag_ty.?, .Function); const ptr_id = try self.accessChain(tag_ptr_ty_ref, tmp_id, &.{@as(u32, @intCast(layout.tag_index))}); const tag_id = try self.constInt(tag_ty_ref, tag_int); - try self.func.body.emit(self.spv.gpa, .OpStore, .{ - .pointer = ptr_id, - .object = tag_id, - }); + try self.store(maybe_tag_ty.?, ptr_id, tag_id, .{}); } - if (layout.active_field_size != 0) { - const active_field_ty_ref = try self.resolveType(layout.active_field_ty, .indirect); - const active_field_ptr_ty_ref = try self.spv.ptrType(active_field_ty_ref, .Function); - const ptr_id = try self.accessChain(active_field_ptr_ty_ref, tmp_id, &.{@as(u32, @intCast(layout.active_field_index))}); - try self.func.body.emit(self.spv.gpa, .OpStore, .{ - .pointer = ptr_id, - .object = payload.?, + const payload_ty = union_ty.field_types.get(ip)[active_field].toType(); + if (payload_ty.hasRuntimeBitsIgnoreComptime(mod)) { + const pl_ptr_ty_ref = try self.ptrType(layout.payload_ty, .Function); + const pl_ptr_id = try self.accessChain(pl_ptr_ty_ref, tmp_id, &.{layout.payload_index}); + const active_pl_ptr_ty_ref = try self.ptrType(payload_ty, .Function); + const active_pl_ptr_id = self.spv.allocId(); + try self.func.body.emit(self.spv.gpa, .OpBitcast, .{ + .id_result_type = self.typeId(active_pl_ptr_ty_ref), + .id_result = active_pl_ptr_id, + .operand = pl_ptr_id, }); + + try self.store(payload_ty, active_pl_ptr_id, payload.?, .{}); } else { assert(payload == null); } @@ -3443,34 +3437,21 @@ const DeclGen = struct { // Just leave the padding fields uninitialized... // TODO: Or should we initialize them with undef explicitly? - // Now cast the pointer and load it as the 'generic' union type. - - const casted_var_id = self.spv.allocId(); - try self.func.body.emit(self.spv.gpa, .OpBitcast, .{ - .id_result_type = self.typeId(un_general_ptr_ty_ref), - .id_result = casted_var_id, - .operand = tmp_id, - }); - - const result_id = self.spv.allocId(); - try self.func.body.emit(self.spv.gpa, .OpLoad, .{ - .id_result_type = self.typeId(un_general_ty_ref), - .id_result = result_id, - .pointer = casted_var_id, - }); - - return result_id; + return try self.load(ty, tmp_id, .{}); } fn airUnionInit(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { if (self.liveness.isUnused(inst)) return null; + const mod = self.module; + const ip = &mod.intern_pool; const ty_pl = self.air.instructions.items(.data)[inst].ty_pl; const extra = self.air.extraData(Air.UnionInit, ty_pl.payload).data; const ty = self.typeOfIndex(inst); - const layout = self.unionLayout(ty, extra.field_index); - const payload = if (layout.active_field_size != 0) + const union_obj = mod.typeToUnion(ty).?; + const field_ty = union_obj.field_types.get(ip)[extra.field_index].toType(); + const payload = if (field_ty.hasRuntimeBitsIgnoreComptime(mod)) try self.resolve(extra.init) else null; @@ -3499,30 +3480,24 @@ const DeclGen = struct { .Union => switch (object_ty.containerLayout(mod)) { .Packed => unreachable, // TODO else => { - // Store, pointer-cast, load - const un_general_ty_ref = try self.resolveType(object_ty, .indirect); - const un_general_ptr_ty_ref = try self.spv.ptrType(un_general_ty_ref, .Function); - const un_active_ty_ref = try self.resolveUnionType(object_ty, field_index); - const un_active_ptr_ty_ref = try self.spv.ptrType(un_active_ty_ref, .Function); - const field_ty_ref = try self.resolveType(field_ty, .indirect); - const field_ptr_ty_ref = try self.spv.ptrType(field_ty_ref, .Function); + // Store, ptr-elem-ptr, pointer-cast, load + const layout = self.unionLayout(object_ty); + assert(layout.has_payload); - const tmp_id = self.spv.allocId(); - try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(un_general_ptr_ty_ref), - .id_result = tmp_id, - .storage_class = .Function, - }); - try self.store(object_ty, tmp_id, object_id, false); - const casted_tmp_id = self.spv.allocId(); + const tmp_id = try self.alloc(object_ty, .{ .storage_class = .Function }); + try self.store(object_ty, tmp_id, object_id, .{}); + + const pl_ptr_ty_ref = try self.ptrType(layout.payload_ty, .Function); + const pl_ptr_id = try self.accessChain(pl_ptr_ty_ref, tmp_id, &.{layout.payload_index}); + + const active_pl_ptr_ty_ref = try self.ptrType(field_ty, .Function); + const active_pl_ptr_id = self.spv.allocId(); try self.func.body.emit(self.spv.gpa, .OpBitcast, .{ - .id_result_type = self.typeId(un_active_ptr_ty_ref), - .id_result = casted_tmp_id, - .operand = tmp_id, + .id_result_type = self.typeId(active_pl_ptr_ty_ref), + .id_result = active_pl_ptr_id, + .operand = pl_ptr_id, }); - const layout = self.unionLayout(object_ty, field_index); - const field_ptr_id = try self.accessChain(field_ptr_ty_ref, casted_tmp_id, &.{layout.active_field_index}); - return try self.load(field_ty, field_ptr_id, false); + return try self.load(field_ty, active_pl_ptr_id, .{}); }, }, else => unreachable, @@ -3581,18 +3556,24 @@ const DeclGen = struct { .Union => switch (object_ty.containerLayout(mod)) { .Packed => unreachable, // TODO else => { - const storage_class = spvStorageClass(object_ptr_ty.ptrAddressSpace(mod)); - const un_active_ty_ref = try self.resolveUnionType(object_ty, field_index); - const un_active_ptr_ty_ref = try self.spv.ptrType(un_active_ty_ref, storage_class); + const layout = self.unionLayout(object_ty); + if (!layout.has_payload) { + // Asked to get a pointer to a zero-sized field. Just lower this + // to undefined, there is no reason to make it be a valid pointer. + return try self.spv.constUndef(result_ty_ref); + } - const casted_id = self.spv.allocId(); + const storage_class = spvStorageClass(object_ptr_ty.ptrAddressSpace(mod)); + const pl_ptr_ty_ref = try self.ptrType(layout.payload_ty, storage_class); + const pl_ptr_id = try self.accessChain(pl_ptr_ty_ref, object_ptr, &.{layout.payload_index}); + + const active_pl_ptr_id = self.spv.allocId(); try self.func.body.emit(self.spv.gpa, .OpBitcast, .{ - .id_result_type = self.typeId(un_active_ptr_ty_ref), - .id_result = casted_id, - .operand = object_ptr, + .id_result_type = self.typeId(result_ty_ref), + .id_result = active_pl_ptr_id, + .operand = pl_ptr_id, }); - const layout = self.unionLayout(object_ty, field_index); - return try self.accessChain(result_ty_ref, casted_id, &.{layout.active_field_index}); + return active_pl_ptr_id; }, }, else => unreachable, @@ -3608,23 +3589,13 @@ const DeclGen = struct { return try self.structFieldPtr(result_ptr_ty, struct_ptr_ty, struct_ptr, field_index); } - /// We cannot use an OpVariable directly in an OpSpecConstantOp, but we can - /// after we insert a dummy AccessChain... - /// TODO: Get rid of this - fn makePointerConstant( - self: *DeclGen, - section: *SpvSection, - ptr_ty_ref: CacheRef, - ptr_id: IdRef, - ) !IdRef { - const result_id = self.spv.allocId(); - try section.emitSpecConstantOp(self.spv.gpa, .OpInBoundsAccessChain, .{ - .id_result_type = self.typeId(ptr_ty_ref), - .id_result = result_id, - .base = ptr_id, - }); - return result_id; - } + const AllocOptions = struct { + initializer: ?IdRef = null, + /// The final storage class of the pointer. This may be either `.Generic` or `.Function`. + /// In either case, the local is allocated in the `.Function` storage class, and optionally + /// cast back to `.Generic`. + storage_class: StorageClass = .Generic, + }; // Allocate a function-local variable, with possible initializer. // This function returns a pointer to a variable of type `ty_ref`, @@ -3632,30 +3603,36 @@ const DeclGen = struct { // placed in the Function address space. fn alloc( self: *DeclGen, - ty_ref: CacheRef, - initializer: ?IdRef, + ty: Type, + options: AllocOptions, ) !IdRef { - const fn_ptr_ty_ref = try self.spv.ptrType(ty_ref, .Function); - const general_ptr_ty_ref = try self.spv.ptrType(ty_ref, .Generic); + const ptr_fn_ty_ref = try self.ptrType(ty, .Function); // SPIR-V requires that OpVariable declarations for locals go into the first block, so we are just going to // directly generate them into func.prologue instead of the body. const var_id = self.spv.allocId(); try self.func.prologue.emit(self.spv.gpa, .OpVariable, .{ - .id_result_type = self.typeId(fn_ptr_ty_ref), + .id_result_type = self.typeId(ptr_fn_ty_ref), .id_result = var_id, .storage_class = .Function, - .initializer = initializer, + .initializer = options.initializer, }); - // Convert to a generic pointer - const result_id = self.spv.allocId(); - try self.func.body.emit(self.spv.gpa, .OpPtrCastToGeneric, .{ - .id_result_type = self.typeId(general_ptr_ty_ref), - .id_result = result_id, - .pointer = var_id, - }); - return result_id; + switch (options.storage_class) { + .Generic => { + const ptr_gn_ty_ref = try self.ptrType(ty, .Generic); + // Convert to a generic pointer + const result_id = self.spv.allocId(); + try self.func.body.emit(self.spv.gpa, .OpPtrCastToGeneric, .{ + .id_result_type = self.typeId(ptr_gn_ty_ref), + .id_result = result_id, + .pointer = var_id, + }); + return result_id; + }, + .Function => return var_id, + else => unreachable, + } } fn airAlloc(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -3664,8 +3641,7 @@ const DeclGen = struct { const ptr_ty = self.typeOfIndex(inst); assert(ptr_ty.ptrAddressSpace(mod) == .generic); const child_ty = ptr_ty.childType(mod); - const child_ty_ref = try self.resolveType(child_ty, .indirect); - return try self.alloc(child_ty_ref, null); + return try self.alloc(child_ty, .{}); } fn airArg(self: *DeclGen) IdRef { @@ -3780,7 +3756,7 @@ const DeclGen = struct { const operand = try self.resolve(ty_op.operand); if (!ptr_ty.isVolatilePtr(mod) and self.liveness.isUnused(inst)) return null; - return try self.load(elem_ty, operand, ptr_ty.isVolatilePtr(mod)); + return try self.load(elem_ty, operand, .{ .is_volatile = ptr_ty.isVolatilePtr(mod) }); } fn airStore(self: *DeclGen, inst: Air.Inst.Index) !void { @@ -3790,7 +3766,7 @@ const DeclGen = struct { const ptr = try self.resolve(bin_op.lhs); const value = try self.resolve(bin_op.rhs); - try self.store(elem_ty, ptr, value, ptr_ty.isVolatilePtr(self.module)); + try self.store(elem_ty, ptr, value, .{ .is_volatile = ptr_ty.isVolatilePtr(self.module) }); } fn airLoop(self: *DeclGen, inst: Air.Inst.Index) !void { @@ -3854,7 +3830,7 @@ const DeclGen = struct { } const ptr = try self.resolve(un_op); - const value = try self.load(ret_ty, ptr, ptr_ty.isVolatilePtr(mod)); + const value = try self.load(ret_ty, ptr, .{ .is_volatile = ptr_ty.isVolatilePtr(mod) }); try self.func.body.emit(self.spv.gpa, .OpReturnValue, .{ .value = value, }); @@ -3980,8 +3956,11 @@ const DeclGen = struct { members[eu_layout.errorFieldIndex()] = operand_id; members[eu_layout.payloadFieldIndex()] = try self.spv.constUndef(payload_ty_ref); - const err_union_ty_ref = try self.resolveType(err_union_ty, .direct); - return try self.constructStruct(err_union_ty_ref, &members); + var types: [2]Type = undefined; + types[eu_layout.errorFieldIndex()] = Type.anyerror; + types[eu_layout.payloadFieldIndex()] = payload_ty; + + return try self.constructStruct(err_union_ty, &types, &members); } fn airWrapErrUnionPayload(self: *DeclGen, inst: Air.Inst.Index) !?IdRef { @@ -4002,8 +3981,11 @@ const DeclGen = struct { members[eu_layout.errorFieldIndex()] = try self.constInt(err_ty_ref, 0); members[eu_layout.payloadFieldIndex()] = try self.convertToIndirect(payload_ty, operand_id); - const err_union_ty_ref = try self.resolveType(err_union_ty, .direct); - return try self.constructStruct(err_union_ty_ref, &members); + var types: [2]Type = undefined; + types[eu_layout.errorFieldIndex()] = Type.anyerror; + types[eu_layout.payloadFieldIndex()] = payload_ty; + + return try self.constructStruct(err_union_ty, &types, &members); } fn airIsNull(self: *DeclGen, inst: Air.Inst.Index, pred: enum { is_null, is_non_null }) !?IdRef { @@ -4037,7 +4019,7 @@ const DeclGen = struct { .is_null => .eq, .is_non_null => .neq, }; - return try self.cmp(op, ptr_ty, ptr_id, null_id); + return try self.cmp(op, Type.bool, ptr_ty, ptr_id, null_id); } const is_non_null_id = if (payload_ty.hasRuntimeBitsIgnoreComptime(mod)) @@ -4135,10 +4117,10 @@ const DeclGen = struct { return operand_id; } - const optional_ty_ref = try self.resolveType(optional_ty, .direct); const payload_id = try self.convertToIndirect(payload_ty, operand_id); const members = [_]IdRef{ payload_id, try self.constBool(true, .indirect) }; - return try self.constructStruct(optional_ty_ref, &members); + const types = [_]Type{ payload_ty, Type.bool }; + return try self.constructStruct(optional_ty, &types, &members); } fn airSwitchBr(self: *DeclGen, inst: Air.Inst.Index) !void { @@ -4420,6 +4402,7 @@ const DeclGen = struct { } // TODO: Multiple results + // TODO: Check that the output type from assembly is the same as the type actually expected by Zig. } return null; diff --git a/src/codegen/spirv/Assembler.zig b/src/codegen/spirv/Assembler.zig index 8f466668ea..87e18d4bd8 100644 --- a/src/codegen/spirv/Assembler.zig +++ b/src/codegen/spirv/Assembler.zig @@ -304,10 +304,16 @@ fn processTypeInstruction(self: *Assembler) !AsmValue { // and so some consideration must be taken when entering this in the type system. return self.todo("process OpTypeArray", .{}); }, - .OpTypePointer => try self.spv.ptrType( - try self.resolveTypeRef(operands[2].ref_id), - @as(spec.StorageClass, @enumFromInt(operands[1].value)), - ), + .OpTypePointer => blk: { + break :blk try self.spv.resolve(.{ + .ptr_type = .{ + .storage_class = @enumFromInt(operands[1].value), + .child_type = try self.resolveTypeRef(operands[2].ref_id), + // TODO: This should be a proper reference resolved via OpTypeForwardPointer + .fwd = @enumFromInt(std.math.maxInt(u32)), + }, + }); + }, .OpTypeFunction => blk: { const param_operands = operands[2..]; const param_types = try self.spv.gpa.alloc(CacheRef, param_operands.len); diff --git a/src/codegen/spirv/Cache.zig b/src/codegen/spirv/Cache.zig index ad2d3442e5..93921204db 100644 --- a/src/codegen/spirv/Cache.zig +++ b/src/codegen/spirv/Cache.zig @@ -22,6 +22,8 @@ const Opcode = spec.Opcode; const IdResult = spec.IdResult; const StorageClass = spec.StorageClass; +const InternPool = @import("../../InternPool.zig"); + const Self = @This(); map: std.AutoArrayHashMapUnmanaged(void, void) = .{}, @@ -31,6 +33,8 @@ extra: std.ArrayListUnmanaged(u32) = .{}, string_bytes: std.ArrayListUnmanaged(u8) = .{}, strings: std.AutoArrayHashMapUnmanaged(void, u32) = .{}, +recursive_ptrs: std.AutoHashMapUnmanaged(Ref, void) = .{}, + const Item = struct { tag: Tag, /// The result-id that this item uses. @@ -62,18 +66,21 @@ const Tag = enum { /// Function (proto)type /// data is payload to FunctionType type_function, - /// Pointer type in the CrossWorkgroup storage class - /// data is child type - type_ptr_generic, - /// Pointer type in the CrossWorkgroup storage class - /// data is child type - type_ptr_crosswgp, - /// Pointer type in the Function storage class - /// data is child type - type_ptr_function, + // /// Pointer type in the CrossWorkgroup storage class + // /// data is child type + // type_ptr_generic, + // /// Pointer type in the CrossWorkgroup storage class + // /// data is child type + // type_ptr_crosswgp, + // /// Pointer type in the Function storage class + // /// data is child type + // type_ptr_function, /// Simple pointer type that does not have any decorations. /// data is payload to SimplePointerType type_ptr_simple, + /// A forward declaration for a pointer. + /// data is ForwardPointerType + type_fwd_ptr, /// Simple structure type that does not have any decorations. /// data is payload to SimpleStructType type_struct_simple, @@ -142,6 +149,12 @@ const Tag = enum { const SimplePointerType = struct { storage_class: StorageClass, child_type: Ref, + fwd: Ref, + }; + + const ForwardPointerType = struct { + storage_class: StorageClass, + zig_child_type: InternPool.Index, }; /// Trailing: @@ -163,14 +176,14 @@ const Tag = enum { fn encode(value: f64) Float64 { const bits = @as(u64, @bitCast(value)); return .{ - .low = @as(u32, @truncate(bits)), - .high = @as(u32, @truncate(bits >> 32)), + .low = @truncate(bits), + .high = @truncate(bits >> 32), }; } fn decode(self: Float64) f64 { const bits = @as(u64, self.low) | (@as(u64, self.high) << 32); - return @as(f64, @bitCast(bits)); + return @bitCast(bits); } }; @@ -192,8 +205,8 @@ const Tag = enum { fn encode(ty: Ref, value: u64) Int64 { return .{ .ty = ty, - .low = @as(u32, @truncate(value)), - .high = @as(u32, @truncate(value >> 32)), + .low = @truncate(value), + .high = @truncate(value >> 32), }; } @@ -210,8 +223,8 @@ const Tag = enum { fn encode(ty: Ref, value: i64) Int64 { return .{ .ty = ty, - .low = @as(u32, @truncate(@as(u64, @bitCast(value)))), - .high = @as(u32, @truncate(@as(u64, @bitCast(value)) >> 32)), + .low = @truncate(@as(u64, @bitCast(value))), + .high = @truncate(@as(u64, @bitCast(value)) >> 32), }; } @@ -237,6 +250,7 @@ pub const Key = union(enum) { array_type: ArrayType, function_type: FunctionType, ptr_type: PointerType, + fwd_ptr_type: ForwardPointerType, struct_type: StructType, opaque_type: OpaqueType, @@ -273,12 +287,18 @@ pub const Key = union(enum) { pub const PointerType = struct { storage_class: StorageClass, child_type: Ref, + fwd: Ref, // TODO: Decorations: // - Alignment // - ArrayStride, // - MaxByteOffset, }; + pub const ForwardPointerType = struct { + zig_child_type: InternPool.Index, + storage_class: StorageClass, + }; + pub const StructType = struct { // TODO: Decorations. /// The name of the structure. Can be `.none`. @@ -313,21 +333,21 @@ pub const Key = union(enum) { /// Turns this value into the corresponding 32-bit literal, 2s complement signed. fn toBits32(self: Int) u32 { return switch (self.value) { - .uint64 => |val| @as(u32, @intCast(val)), - .int64 => |val| if (val < 0) @as(u32, @bitCast(@as(i32, @intCast(val)))) else @as(u32, @intCast(val)), + .uint64 => |val| @intCast(val), + .int64 => |val| if (val < 0) @bitCast(@as(i32, @intCast(val))) else @intCast(val), }; } fn toBits64(self: Int) u64 { return switch (self.value) { .uint64 => |val| val, - .int64 => |val| @as(u64, @bitCast(val)), + .int64 => |val| @bitCast(val), }; } fn to(self: Int, comptime T: type) T { return switch (self.value) { - inline else => |val| @as(T, @intCast(val)), + inline else => |val| @intCast(val), }; } }; @@ -387,7 +407,7 @@ pub const Key = union(enum) { }, inline else => |key| std.hash.autoHash(&hasher, key), } - return @as(u32, @truncate(hasher.final())); + return @truncate(hasher.final()); } fn eql(a: Key, b: Key) bool { @@ -419,7 +439,7 @@ pub const Key = union(enum) { pub fn eql(ctx: @This(), a: Key, b_void: void, b_index: usize) bool { _ = b_void; - return ctx.self.lookup(@as(Ref, @enumFromInt(b_index))).eql(a); + return ctx.self.lookup(@enumFromInt(b_index)).eql(a); } pub fn hash(ctx: @This(), a: Key) u32 { @@ -450,6 +470,7 @@ pub fn deinit(self: *Self, spv: *const Module) void { self.extra.deinit(spv.gpa); self.string_bytes.deinit(spv.gpa); self.strings.deinit(spv.gpa); + self.recursive_ptrs.deinit(spv.gpa); } /// Actually materialize the database into spir-v instructions. @@ -460,7 +481,7 @@ pub fn materialize(self: *const Self, spv: *Module) !Section { var section = Section{}; errdefer section.deinit(spv.gpa); for (self.items.items(.result_id), 0..) |result_id, index| { - try self.emit(spv, result_id, @as(Ref, @enumFromInt(index)), §ion); + try self.emit(spv, result_id, @enumFromInt(index), §ion); } return section; } @@ -538,6 +559,15 @@ fn emit( }); // TODO: Decorations? }, + .fwd_ptr_type => |fwd| { + // Only emit the OpTypeForwardPointer if its actually required. + if (self.recursive_ptrs.contains(ref)) { + try section.emit(spv.gpa, .OpTypeForwardPointer, .{ + .pointer_type = result_id, + .storage_class = fwd.storage_class, + }); + } + }, .struct_type => |struct_type| { try section.emitRaw(spv.gpa, .OpTypeStruct, 1 + struct_type.member_types.len); section.writeOperand(IdResult, result_id); @@ -549,7 +579,7 @@ fn emit( } for (struct_type.memberNames(), 0..) |member_name, i| { if (self.getString(member_name)) |name| { - try spv.memberDebugName(result_id, @as(u32, @intCast(i)), name); + try spv.memberDebugName(result_id, @intCast(i), name); } } // TODO: Decorations? @@ -625,13 +655,12 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { const adapter: Key.Adapter = .{ .self = self }; const entry = try self.map.getOrPutAdapted(spv.gpa, key, adapter); if (entry.found_existing) { - return @as(Ref, @enumFromInt(entry.index)); + return @enumFromInt(entry.index); } - const result_id = spv.allocId(); const item: Item = switch (key) { inline .void_type, .bool_type => .{ .tag = .type_simple, - .result_id = result_id, + .result_id = spv.allocId(), .data = @intFromEnum(key.toSimpleType()), }, .int_type => |int| blk: { @@ -641,87 +670,104 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { }; break :blk .{ .tag = t, - .result_id = result_id, + .result_id = spv.allocId(), .data = int.bits, }; }, .float_type => |float| .{ .tag = .type_float, - .result_id = result_id, + .result_id = spv.allocId(), .data = float.bits, }, .vector_type => |vector| .{ .tag = .type_vector, - .result_id = result_id, + .result_id = spv.allocId(), .data = try self.addExtra(spv, vector), }, .array_type => |array| .{ .tag = .type_array, - .result_id = result_id, + .result_id = spv.allocId(), .data = try self.addExtra(spv, array), }, .function_type => |function| blk: { const extra = try self.addExtra(spv, Tag.FunctionType{ - .param_len = @as(u32, @intCast(function.parameters.len)), + .param_len = @intCast(function.parameters.len), .return_type = function.return_type, }); - try self.extra.appendSlice(spv.gpa, @as([]const u32, @ptrCast(function.parameters))); + try self.extra.appendSlice(spv.gpa, @ptrCast(function.parameters)); break :blk .{ .tag = .type_function, - .result_id = result_id, + .result_id = spv.allocId(), .data = extra, }; }, - .ptr_type => |ptr| switch (ptr.storage_class) { - .Generic => Item{ - .tag = .type_ptr_generic, - .result_id = result_id, - .data = @intFromEnum(ptr.child_type), - }, - .CrossWorkgroup => Item{ - .tag = .type_ptr_crosswgp, - .result_id = result_id, - .data = @intFromEnum(ptr.child_type), - }, - .Function => Item{ - .tag = .type_ptr_function, - .result_id = result_id, - .data = @intFromEnum(ptr.child_type), - }, - else => |storage_class| Item{ - .tag = .type_ptr_simple, - .result_id = result_id, - .data = try self.addExtra(spv, Tag.SimplePointerType{ - .storage_class = storage_class, - .child_type = ptr.child_type, - }), - }, + // .ptr_type => |ptr| switch (ptr.storage_class) { + // .Generic => Item{ + // .tag = .type_ptr_generic, + // .result_id = spv.allocId(), + // .data = @intFromEnum(ptr.child_type), + // }, + // .CrossWorkgroup => Item{ + // .tag = .type_ptr_crosswgp, + // .result_id = spv.allocId(), + // .data = @intFromEnum(ptr.child_type), + // }, + // .Function => Item{ + // .tag = .type_ptr_function, + // .result_id = spv.allocId(), + // .data = @intFromEnum(ptr.child_type), + // }, + // else => |storage_class| Item{ + // .tag = .type_ptr_simple, + // .result_id = spv.allocId(), + // .data = try self.addExtra(spv, Tag.SimplePointerType{ + // .storage_class = storage_class, + // .child_type = ptr.child_type, + // }), + // }, + // }, + .ptr_type => |ptr| Item{ + .tag = .type_ptr_simple, + .result_id = self.resultId(ptr.fwd), + .data = try self.addExtra(spv, Tag.SimplePointerType{ + .storage_class = ptr.storage_class, + .child_type = ptr.child_type, + .fwd = ptr.fwd, + }), + }, + .fwd_ptr_type => |fwd| Item{ + .tag = .type_fwd_ptr, + .result_id = spv.allocId(), + .data = try self.addExtra(spv, Tag.ForwardPointerType{ + .zig_child_type = fwd.zig_child_type, + .storage_class = fwd.storage_class, + }), }, .struct_type => |struct_type| blk: { const extra = try self.addExtra(spv, Tag.SimpleStructType{ .name = struct_type.name, - .members_len = @as(u32, @intCast(struct_type.member_types.len)), + .members_len = @intCast(struct_type.member_types.len), }); - try self.extra.appendSlice(spv.gpa, @as([]const u32, @ptrCast(struct_type.member_types))); + try self.extra.appendSlice(spv.gpa, @ptrCast(struct_type.member_types)); if (struct_type.member_names) |member_names| { - try self.extra.appendSlice(spv.gpa, @as([]const u32, @ptrCast(member_names))); + try self.extra.appendSlice(spv.gpa, @ptrCast(member_names)); break :blk Item{ .tag = .type_struct_simple_with_member_names, - .result_id = result_id, + .result_id = spv.allocId(), .data = extra, }; } else { break :blk Item{ .tag = .type_struct_simple, - .result_id = result_id, + .result_id = spv.allocId(), .data = extra, }; } }, .opaque_type => |opaque_type| Item{ .tag = .type_opaque, - .result_id = result_id, + .result_id = spv.allocId(), .data = @intFromEnum(opaque_type.name), }, .int => |int| blk: { @@ -729,13 +775,13 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { if (int_type.signedness == .unsigned and int_type.bits == 8) { break :blk .{ .tag = .uint8, - .result_id = result_id, + .result_id = spv.allocId(), .data = int.to(u8), }; } else if (int_type.signedness == .unsigned and int_type.bits == 32) { break :blk .{ .tag = .uint32, - .result_id = result_id, + .result_id = spv.allocId(), .data = int.to(u32), }; } @@ -745,32 +791,32 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { if (val >= 0 and val <= std.math.maxInt(u32)) { break :blk .{ .tag = .uint_small, - .result_id = result_id, + .result_id = spv.allocId(), .data = try self.addExtra(spv, Tag.UInt32{ .ty = int.ty, - .value = @as(u32, @intCast(val)), + .value = @intCast(val), }), }; } else if (val >= std.math.minInt(i32) and val <= std.math.maxInt(i32)) { break :blk .{ .tag = .int_small, - .result_id = result_id, + .result_id = spv.allocId(), .data = try self.addExtra(spv, Tag.Int32{ .ty = int.ty, - .value = @as(i32, @intCast(val)), + .value = @intCast(val), }), }; } else if (val < 0) { break :blk .{ .tag = .int_large, - .result_id = result_id, - .data = try self.addExtra(spv, Tag.Int64.encode(int.ty, @as(i64, @intCast(val)))), + .result_id = spv.allocId(), + .data = try self.addExtra(spv, Tag.Int64.encode(int.ty, @intCast(val))), }; } else { break :blk .{ .tag = .uint_large, - .result_id = result_id, - .data = try self.addExtra(spv, Tag.UInt64.encode(int.ty, @as(u64, @intCast(val)))), + .result_id = spv.allocId(), + .data = try self.addExtra(spv, Tag.UInt64.encode(int.ty, @intCast(val))), }; } }, @@ -779,29 +825,29 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { .float => |float| switch (self.lookup(float.ty).float_type.bits) { 16 => .{ .tag = .float16, - .result_id = result_id, + .result_id = spv.allocId(), .data = @as(u16, @bitCast(float.value.float16)), }, 32 => .{ .tag = .float32, - .result_id = result_id, + .result_id = spv.allocId(), .data = @as(u32, @bitCast(float.value.float32)), }, 64 => .{ .tag = .float64, - .result_id = result_id, + .result_id = spv.allocId(), .data = try self.addExtra(spv, Tag.Float64.encode(float.value.float64)), }, else => unreachable, }, .undef => |undef| .{ .tag = .undef, - .result_id = result_id, + .result_id = spv.allocId(), .data = @intFromEnum(undef.ty), }, .null => |null_info| .{ .tag = .null, - .result_id = result_id, + .result_id = spv.allocId(), .data = @intFromEnum(null_info.ty), }, .bool => |bool_info| .{ @@ -809,13 +855,13 @@ pub fn resolve(self: *Self, spv: *Module, key: Key) !Ref { true => Tag.bool_true, false => Tag.bool_false, }, - .result_id = result_id, + .result_id = spv.allocId(), .data = @intFromEnum(bool_info.ty), }, }; try self.items.append(spv.gpa, item); - return @as(Ref, @enumFromInt(entry.index)); + return @enumFromInt(entry.index); } /// Turn a Ref back into a Key. @@ -830,14 +876,14 @@ pub fn lookup(self: *const Self, ref: Ref) Key { }, .type_int_signed => .{ .int_type = .{ .signedness = .signed, - .bits = @as(u16, @intCast(data)), + .bits = @intCast(data), } }, .type_int_unsigned => .{ .int_type = .{ .signedness = .unsigned, - .bits = @as(u16, @intCast(data)), + .bits = @intCast(data), } }, .type_float => .{ .float_type = .{ - .bits = @as(u16, @intCast(data)), + .bits = @intCast(data), } }, .type_vector => .{ .vector_type = self.extraData(Tag.VectorType, data) }, .type_array => .{ .array_type = self.extraData(Tag.ArrayType, data) }, @@ -846,40 +892,50 @@ pub fn lookup(self: *const Self, ref: Ref) Key { return .{ .function_type = .{ .return_type = payload.data.return_type, - .parameters = @as([]const Ref, @ptrCast(self.extra.items[payload.trail..][0..payload.data.param_len])), + .parameters = @ptrCast(self.extra.items[payload.trail..][0..payload.data.param_len]), }, }; }, - .type_ptr_generic => .{ - .ptr_type = .{ - .storage_class = .Generic, - .child_type = @as(Ref, @enumFromInt(data)), - }, - }, - .type_ptr_crosswgp => .{ - .ptr_type = .{ - .storage_class = .CrossWorkgroup, - .child_type = @as(Ref, @enumFromInt(data)), - }, - }, - .type_ptr_function => .{ - .ptr_type = .{ - .storage_class = .Function, - .child_type = @as(Ref, @enumFromInt(data)), - }, - }, + // .type_ptr_generic => .{ + // .ptr_type = .{ + // .storage_class = .Generic, + // .child_type = @enumFromInt(data), + // }, + // }, + // .type_ptr_crosswgp => .{ + // .ptr_type = .{ + // .storage_class = .CrossWorkgroup, + // .child_type = @enumFromInt(data), + // }, + // }, + // .type_ptr_function => .{ + // .ptr_type = .{ + // .storage_class = .Function, + // .child_type = @enumFromInt(data), + // }, + // }, .type_ptr_simple => { const payload = self.extraData(Tag.SimplePointerType, data); return .{ .ptr_type = .{ .storage_class = payload.storage_class, .child_type = payload.child_type, + .fwd = payload.fwd, + }, + }; + }, + .type_fwd_ptr => { + const payload = self.extraData(Tag.ForwardPointerType, data); + return .{ + .fwd_ptr_type = .{ + .zig_child_type = payload.zig_child_type, + .storage_class = payload.storage_class, }, }; }, .type_struct_simple => { const payload = self.extraDataTrail(Tag.SimpleStructType, data); - const member_types = @as([]const Ref, @ptrCast(self.extra.items[payload.trail..][0..payload.data.members_len])); + const member_types: []const Ref = @ptrCast(self.extra.items[payload.trail..][0..payload.data.members_len]); return .{ .struct_type = .{ .name = payload.data.name, @@ -891,8 +947,8 @@ pub fn lookup(self: *const Self, ref: Ref) Key { .type_struct_simple_with_member_names => { const payload = self.extraDataTrail(Tag.SimpleStructType, data); const trailing = self.extra.items[payload.trail..]; - const member_types = @as([]const Ref, @ptrCast(trailing[0..payload.data.members_len])); - const member_names = @as([]const String, @ptrCast(trailing[payload.data.members_len..][0..payload.data.members_len])); + const member_types: []const Ref = @ptrCast(trailing[0..payload.data.members_len]); + const member_names: []const String = @ptrCast(trailing[payload.data.members_len..][0..payload.data.members_len]); return .{ .struct_type = .{ .name = payload.data.name, @@ -903,16 +959,16 @@ pub fn lookup(self: *const Self, ref: Ref) Key { }, .type_opaque => .{ .opaque_type = .{ - .name = @as(String, @enumFromInt(data)), + .name = @enumFromInt(data), }, }, .float16 => .{ .float = .{ .ty = self.get(.{ .float_type = .{ .bits = 16 } }), - .value = .{ .float16 = @as(f16, @bitCast(@as(u16, @intCast(data)))) }, + .value = .{ .float16 = @bitCast(@as(u16, @intCast(data))) }, } }, .float32 => .{ .float = .{ .ty = self.get(.{ .float_type = .{ .bits = 32 } }), - .value = .{ .float32 = @as(f32, @bitCast(data)) }, + .value = .{ .float32 = @bitCast(data) }, } }, .float64 => .{ .float = .{ .ty = self.get(.{ .float_type = .{ .bits = 64 } }), @@ -955,17 +1011,17 @@ pub fn lookup(self: *const Self, ref: Ref) Key { } }; }, .undef => .{ .undef = .{ - .ty = @as(Ref, @enumFromInt(data)), + .ty = @enumFromInt(data), } }, .null => .{ .null = .{ - .ty = @as(Ref, @enumFromInt(data)), + .ty = @enumFromInt(data), } }, .bool_true => .{ .bool = .{ - .ty = @as(Ref, @enumFromInt(data)), + .ty = @enumFromInt(data), .value = true, } }, .bool_false => .{ .bool = .{ - .ty = @as(Ref, @enumFromInt(data)), + .ty = @enumFromInt(data), .value = false, } }, }; @@ -981,7 +1037,7 @@ pub fn resultId(self: Self, ref: Ref) IdResult { fn get(self: *const Self, key: Key) Ref { const adapter: Key.Adapter = .{ .self = self }; const index = self.map.getIndexAdapted(key, adapter).?; - return @as(Ref, @enumFromInt(index)); + return @enumFromInt(index); } fn addExtra(self: *Self, spv: *Module, extra: anytype) !u32 { @@ -991,15 +1047,16 @@ fn addExtra(self: *Self, spv: *Module, extra: anytype) !u32 { } fn addExtraAssumeCapacity(self: *Self, extra: anytype) !u32 { - const payload_offset = @as(u32, @intCast(self.extra.items.len)); + const payload_offset: u32 = @intCast(self.extra.items.len); inline for (@typeInfo(@TypeOf(extra)).Struct.fields) |field| { const field_val = @field(extra, field.name); - const word = switch (field.type) { + const word: u32 = switch (field.type) { u32 => field_val, - i32 => @as(u32, @bitCast(field_val)), + i32 => @bitCast(field_val), Ref => @intFromEnum(field_val), StorageClass => @intFromEnum(field_val), String => @intFromEnum(field_val), + InternPool.Index => @intFromEnum(field_val), else => @compileError("Invalid type: " ++ @typeName(field.type)), }; self.extra.appendAssumeCapacity(word); @@ -1018,10 +1075,11 @@ fn extraDataTrail(self: Self, comptime T: type, offset: u32) struct { data: T, t const word = self.extra.items[offset + i]; @field(result, field.name) = switch (field.type) { u32 => word, - i32 => @as(i32, @bitCast(word)), - Ref => @as(Ref, @enumFromInt(word)), - StorageClass => @as(StorageClass, @enumFromInt(word)), - String => @as(String, @enumFromInt(word)), + i32 => @bitCast(word), + Ref => @enumFromInt(word), + StorageClass => @enumFromInt(word), + String => @enumFromInt(word), + InternPool.Index => @enumFromInt(word), else => @compileError("Invalid type: " ++ @typeName(field.type)), }; } @@ -1049,7 +1107,7 @@ pub const String = enum(u32) { _ = ctx; var hasher = std.hash.Wyhash.init(0); hasher.update(a); - return @as(u32, @truncate(hasher.final())); + return @truncate(hasher.final()); } }; }; @@ -1064,10 +1122,10 @@ pub fn addString(self: *Self, spv: *Module, str: []const u8) !String { try self.string_bytes.ensureUnusedCapacity(spv.gpa, 1 + str.len); self.string_bytes.appendSliceAssumeCapacity(str); self.string_bytes.appendAssumeCapacity(0); - entry.value_ptr.* = @as(u32, @intCast(offset)); + entry.value_ptr.* = @intCast(offset); } - return @as(String, @enumFromInt(entry.index)); + return @enumFromInt(entry.index); } pub fn getString(self: *const Self, ref: String) ?[]const u8 { diff --git a/src/codegen/spirv/Module.zig b/src/codegen/spirv/Module.zig index 1936b78826..4ecbc8d7a0 100644 --- a/src/codegen/spirv/Module.zig +++ b/src/codegen/spirv/Module.zig @@ -507,17 +507,6 @@ pub fn arrayType(self: *Module, len: u32, elem_ty_ref: CacheRef) !CacheRef { } }); } -pub fn ptrType( - self: *Module, - child: CacheRef, - storage_class: spec.StorageClass, -) !CacheRef { - return try self.resolve(.{ .ptr_type = .{ - .storage_class = storage_class, - .child_type = child, - } }); -} - pub fn constInt(self: *Module, ty_ref: CacheRef, value: anytype) !IdRef { const ty = self.cache.lookup(ty_ref).int_type; const Value = Cache.Key.Int.Value; diff --git a/test/behavior/bugs/12000.zig b/test/behavior/bugs/12000.zig index f15c691fd2..d7b856d2f5 100644 --- a/test/behavior/bugs/12000.zig +++ b/test/behavior/bugs/12000.zig @@ -9,7 +9,6 @@ test { if (builtin.zig_backend == .stage2_x86) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var t: T = .{ .next = null }; try std.testing.expect(t.next == null); diff --git a/test/behavior/bugs/1735.zig b/test/behavior/bugs/1735.zig index db98e94e48..68f851e578 100644 --- a/test/behavior/bugs/1735.zig +++ b/test/behavior/bugs/1735.zig @@ -44,7 +44,6 @@ const a = struct { test "initialization" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var t = a.init(); try std.testing.expect(t.foo.len == 0); diff --git a/test/behavior/bugs/1914.zig b/test/behavior/bugs/1914.zig index 14f4b0be27..bd3f830e60 100644 --- a/test/behavior/bugs/1914.zig +++ b/test/behavior/bugs/1914.zig @@ -12,8 +12,6 @@ const b_list: []B = &[_]B{}; const a = A{ .b_list_pointer = &b_list }; test "segfault bug" { - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; - const assert = std.debug.assert; const obj = B{ .a_pointer = &a }; assert(obj.a_pointer == &a); // this makes zig crash @@ -30,7 +28,5 @@ pub const B2 = struct { var b_value = B2{ .pointer_array = &[_]*A2{} }; test "basic stuff" { - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; - std.debug.assert(&b_value == &b_value); } diff --git a/test/behavior/bugs/2006.zig b/test/behavior/bugs/2006.zig index 40f03d5b0b..3a4e61b11d 100644 --- a/test/behavior/bugs/2006.zig +++ b/test/behavior/bugs/2006.zig @@ -7,7 +7,6 @@ const S = struct { }; test "bug 2006" { if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var a: S = undefined; a = S{ .p = undefined }; diff --git a/test/behavior/bugs/3007.zig b/test/behavior/bugs/3007.zig index 7adec60a3b..3ae1562e9b 100644 --- a/test/behavior/bugs/3007.zig +++ b/test/behavior/bugs/3007.zig @@ -22,7 +22,6 @@ test "fixed" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; default_foo = get_foo() catch null; // This Line try std.testing.expect(!default_foo.?.free); diff --git a/test/behavior/bugs/6947.zig b/test/behavior/bugs/6947.zig index dc24dac871..25b35bdb89 100644 --- a/test/behavior/bugs/6947.zig +++ b/test/behavior/bugs/6947.zig @@ -8,7 +8,6 @@ test { if (builtin.zig_backend == .stage2_x86) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var slice: []void = undefined; destroy(&slice[0]); diff --git a/test/behavior/bugs/7325.zig b/test/behavior/bugs/7325.zig index 78536f2b40..3652553b51 100644 --- a/test/behavior/bugs/7325.zig +++ b/test/behavior/bugs/7325.zig @@ -81,7 +81,6 @@ test { if (builtin.zig_backend == .stage2_x86) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var param: ParamType = .{ .one_of = .{ .name = "name" }, diff --git a/test/behavior/error.zig b/test/behavior/error.zig index 72003d1ea8..dfd511af2e 100644 --- a/test/behavior/error.zig +++ b/test/behavior/error.zig @@ -943,7 +943,6 @@ test "returning an error union containing a type with no runtime bits" { test "try used in recursive function with inferred error set" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const Value = union(enum) { values: []const @This(), diff --git a/test/behavior/eval.zig b/test/behavior/eval.zig index a57d537c30..39e9e32484 100644 --- a/test/behavior/eval.zig +++ b/test/behavior/eval.zig @@ -391,7 +391,6 @@ test "return 0 from function that has u0 return type" { test "statically initialized struct" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; st_init_str_foo.x += 1; try expect(st_init_str_foo.x == 14); @@ -498,7 +497,6 @@ test "comptime shlWithOverflow" { test "const ptr to variable data changes at runtime" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; try expect(foo_ref.name[0] == 'a'); foo_ref.name = "b"; @@ -1551,8 +1549,6 @@ test "comptime function turns function value to function pointer" { } test "container level const and var have unique addresses" { - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; - const S = struct { x: i32, y: i32, diff --git a/test/behavior/generics.zig b/test/behavior/generics.zig index 2981c73f22..d52ad20677 100644 --- a/test/behavior/generics.zig +++ b/test/behavior/generics.zig @@ -205,7 +205,6 @@ fn foo2(arg: anytype) bool { test "generic struct" { if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var a1 = GenNode(i32){ .value = 13, diff --git a/test/behavior/null.zig b/test/behavior/null.zig index ec57819580..3fe6e663e1 100644 --- a/test/behavior/null.zig +++ b/test/behavior/null.zig @@ -185,7 +185,6 @@ test "unwrap optional which is field of global var" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; struct_with_optional.field = null; if (struct_with_optional.field) |payload| { diff --git a/test/behavior/optional.zig b/test/behavior/optional.zig index 3528b3f571..648b698842 100644 --- a/test/behavior/optional.zig +++ b/test/behavior/optional.zig @@ -193,7 +193,6 @@ test "nested orelse" { test "self-referential struct through a slice of optional" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { const Node = struct { diff --git a/test/behavior/ptrcast.zig b/test/behavior/ptrcast.zig index 7f6323084f..08e199ae60 100644 --- a/test/behavior/ptrcast.zig +++ b/test/behavior/ptrcast.zig @@ -130,7 +130,6 @@ test "lower reinterpreted comptime field ptr (with under-aligned fields)" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; // Test lowering a field ptr comptime var bytes align(2) = [_]u8{ 1, 2, 3, 4, 5, 6 }; @@ -153,7 +152,6 @@ test "lower reinterpreted comptime field ptr" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; // Test lowering a field ptr comptime var bytes align(4) = [_]u8{ 1, 2, 3, 4, 5, 6, 7, 8 }; diff --git a/test/behavior/struct.zig b/test/behavior/struct.zig index 8e7aa59844..dbd33c2bb5 100644 --- a/test/behavior/struct.zig +++ b/test/behavior/struct.zig @@ -292,7 +292,6 @@ const Val = struct { test "struct point to self" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var root: Node = undefined; root.val.x = 1; @@ -347,7 +346,6 @@ test "self-referencing struct via array member" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const T = struct { children: [1]*@This(), @@ -370,7 +368,6 @@ const EmptyStruct = struct { test "align 1 field before self referential align 8 field as slice return type" { if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const result = alloc(Expr); try expect(result.len == 0); @@ -736,7 +733,6 @@ test "packed struct with u0 field access" { test "access to global struct fields" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; g_foo.bar.value = 42; try expect(g_foo.bar.value == 42); @@ -1423,7 +1419,6 @@ test "fieldParentPtr of a zero-bit field" { test "struct field has a pointer to an aligned version of itself" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const E = struct { next: *align(1) @This(), @@ -1519,7 +1514,6 @@ test "function pointer in struct returns the struct" { test "no dependency loop on optional field wrapped in generic function" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { fn Atomic(comptime T: type) type { diff --git a/test/behavior/struct_contains_null_ptr_itself.zig b/test/behavior/struct_contains_null_ptr_itself.zig index ded29039c8..7f0182af22 100644 --- a/test/behavior/struct_contains_null_ptr_itself.zig +++ b/test/behavior/struct_contains_null_ptr_itself.zig @@ -5,7 +5,6 @@ const builtin = @import("builtin"); test "struct contains null pointer which contains original struct" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var x: ?*NodeLineComment = null; try expect(x == null); diff --git a/test/behavior/struct_contains_slice_of_itself.zig b/test/behavior/struct_contains_slice_of_itself.zig index 6f63bfb8de..adb1c31047 100644 --- a/test/behavior/struct_contains_slice_of_itself.zig +++ b/test/behavior/struct_contains_slice_of_itself.zig @@ -13,7 +13,6 @@ const NodeAligned = struct { test "struct contains slice of itself" { if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var other_nodes = [_]Node{ Node{ @@ -54,7 +53,6 @@ test "struct contains slice of itself" { test "struct contains aligned slice of itself" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; var other_nodes = [_]NodeAligned{ NodeAligned{ diff --git a/test/behavior/union.zig b/test/behavior/union.zig index 6b87ab96ce..ebfde8899e 100644 --- a/test/behavior/union.zig +++ b/test/behavior/union.zig @@ -399,7 +399,6 @@ test "tagged union with no payloads" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const a = UnionEnumNoPayloads{ .B = {} }; switch (a) { @@ -474,7 +473,6 @@ test "update the tag value for zero-sized unions" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = union(enum) { U0: void, @@ -515,7 +513,6 @@ test "method call on an empty union" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { const MyUnion = union(MyUnionTag) { @@ -593,7 +590,6 @@ test "tagged union with all void fields but a meaningful tag" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { const B = union(enum) { @@ -795,7 +791,6 @@ test "@unionInit stored to a const" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { const U = union(enum) { @@ -867,7 +862,6 @@ test "union no tag with struct member" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const Struct = struct {}; const Union = union { @@ -1079,7 +1073,6 @@ test "@unionInit on union with tag but no fields" { if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const S = struct { const Type = enum(u8) { no_op = 105 }; @@ -1128,7 +1121,6 @@ test "global variable struct contains union initialized to non-most-aligned fiel if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const T = struct { const U = union(enum) { @@ -1348,7 +1340,6 @@ test "union field ptr - zero sized payload" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const U = union { foo: void, @@ -1363,7 +1354,6 @@ test "union field ptr - zero sized field" { if (builtin.zig_backend == .stage2_aarch64) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_arm) return error.SkipZigTest; // TODO if (builtin.zig_backend == .stage2_sparc64) return error.SkipZigTest; // TODO - if (builtin.zig_backend == .stage2_spirv64) return error.SkipZigTest; const U = union { foo: void,