mirror of
https://github.com/roc-lang/roc.git
synced 2025-08-03 19:58:18 +00:00
checkpoint
This commit is contained in:
parent
3c8dbce72e
commit
5e4f43e1d8
12 changed files with 296 additions and 109 deletions
|
@ -3,9 +3,9 @@ const builtin = @import("builtin");
|
|||
const always_inline = std.builtin.CallOptions.Modifier.always_inline;
|
||||
const Monotonic = std.builtin.AtomicOrder.Monotonic;
|
||||
|
||||
const DEBUG_INCDEC = false;
|
||||
const DEBUG_INCDEC = true;
|
||||
const DEBUG_TESTING_ALLOC = false;
|
||||
const DEBUG_ALLOC = false;
|
||||
const DEBUG_ALLOC = true;
|
||||
|
||||
pub fn WithOverflow(comptime T: type) type {
|
||||
return extern struct { value: T, has_overflowed: bool };
|
||||
|
@ -360,7 +360,7 @@ pub fn isUnique(
|
|||
const refcount = (isizes - 1)[0];
|
||||
|
||||
if (DEBUG_INCDEC and builtin.target.cpu.arch != .wasm32) {
|
||||
std.debug.print("| is unique {*}\n", .{isizes});
|
||||
std.debug.print("| is unique {*}\n", .{isizes - 1});
|
||||
}
|
||||
|
||||
return refcount == REFCOUNT_ONE_ISIZE;
|
||||
|
@ -436,7 +436,7 @@ pub fn allocateWithRefcount(
|
|||
var new_bytes: [*]u8 = alloc(length, alignment) orelse unreachable;
|
||||
|
||||
if (DEBUG_ALLOC and builtin.target.cpu.arch != .wasm32) {
|
||||
std.debug.print("+ allocated {*} ({} bytes with alignment {})\n", .{ new_bytes, data_bytes, element_alignment });
|
||||
std.debug.print("+ allocated {*} ({} bytes with alignment {})\n", .{ new_bytes, data_bytes, alignment });
|
||||
}
|
||||
|
||||
const data_ptr = new_bytes + alignment;
|
||||
|
|
|
@ -128,7 +128,13 @@ impl IntWidth {
|
|||
U128 | I128 => {
|
||||
// the C ABI defines 128-bit integers to always be 16B aligned,
|
||||
// according to https://reviews.llvm.org/D28990#655487
|
||||
16
|
||||
//
|
||||
// however, rust does not always think that this is true
|
||||
match target_info.architecture {
|
||||
Architecture::X86_64 => 8,
|
||||
Architecture::Aarch64 | Architecture::Aarch32 | Architecture::Wasm32 => 16,
|
||||
Architecture::X86_32 => 8,
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -987,7 +987,11 @@ impl<
|
|||
}
|
||||
// Note that on windows there is only 1 general return register so we can't use this optimisation
|
||||
LayoutRepr::I128 | LayoutRepr::U128 if CC::GENERAL_RETURN_REGS.len() > 1 => {
|
||||
let offset = self.storage_manager.claim_stack_area(dst, 16);
|
||||
let offset = self.storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*dst,
|
||||
Layout::U128,
|
||||
);
|
||||
|
||||
ASM::mov_base32_reg64(&mut self.buf, offset, CC::GENERAL_RETURN_REGS[0]);
|
||||
ASM::mov_base32_reg64(&mut self.buf, offset + 8, CC::GENERAL_RETURN_REGS[1]);
|
||||
|
@ -1264,9 +1268,11 @@ impl<
|
|||
|
||||
let buf = &mut self.buf;
|
||||
|
||||
let struct_size = self.layout_interner.stack_size(*return_layout);
|
||||
|
||||
let base_offset = self.storage_manager.claim_stack_area(dst, struct_size);
|
||||
let base_offset = self.storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*dst,
|
||||
*return_layout,
|
||||
);
|
||||
|
||||
match self.layout_interner.get_repr(*num_layout) {
|
||||
LayoutRepr::Builtin(Int(
|
||||
|
@ -1453,14 +1459,16 @@ impl<
|
|||
|
||||
let buf = &mut self.buf;
|
||||
|
||||
let struct_size = self.layout_interner.stack_size(*return_layout);
|
||||
|
||||
let base_offset = self.storage_manager.claim_stack_area(dst, struct_size);
|
||||
|
||||
match self.layout_interner.get_repr(*num_layout) {
|
||||
LayoutRepr::Builtin(Int(
|
||||
IntWidth::I64 | IntWidth::I32 | IntWidth::I16 | IntWidth::I8,
|
||||
)) => {
|
||||
let base_offset = self.storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*dst,
|
||||
*return_layout,
|
||||
);
|
||||
|
||||
let dst_reg = self
|
||||
.storage_manager
|
||||
.claim_general_reg(buf, &Symbol::DEV_TMP);
|
||||
|
@ -1481,14 +1489,27 @@ impl<
|
|||
self.free_symbol(&Symbol::DEV_TMP);
|
||||
self.free_symbol(&Symbol::DEV_TMP2);
|
||||
}
|
||||
LayoutRepr::Builtin(Int(
|
||||
IntWidth::U64 | IntWidth::U32 | IntWidth::U16 | IntWidth::U8,
|
||||
)) => {
|
||||
todo!("mulChecked for unsigned integers")
|
||||
LayoutRepr::Builtin(Builtin::Int(int_width)) => {
|
||||
self.build_fn_call(
|
||||
dst,
|
||||
bitcode::NUM_MUL_CHECKED_INT[int_width].to_string(),
|
||||
&[*src1, *src2],
|
||||
&[*num_layout, *num_layout],
|
||||
return_layout,
|
||||
);
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Float(_width)) => {
|
||||
todo!("mulChecked for floats")
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Decimal) => {
|
||||
self.build_fn_call(
|
||||
dst,
|
||||
bitcode::DEC_MUL_WITH_OVERFLOW.to_string(),
|
||||
&[*src1, *src2],
|
||||
&[Layout::DEC, Layout::DEC],
|
||||
return_layout,
|
||||
);
|
||||
}
|
||||
x => todo!("mulChecked: layout, {:?}", x),
|
||||
}
|
||||
}
|
||||
|
@ -2614,9 +2635,9 @@ impl<
|
|||
self.load_layout_stack_size(elem_layout, Symbol::DEV_TMP2);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let lowlevel_args = [
|
||||
capacity,
|
||||
|
@ -2677,9 +2698,9 @@ impl<
|
|||
);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let lowlevel_args = bumpalo::vec![
|
||||
in self.env.arena;
|
||||
|
@ -2751,9 +2772,9 @@ impl<
|
|||
self.load_layout_stack_size(elem_layout, Symbol::DEV_TMP2);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let lowlevel_args = [
|
||||
list,
|
||||
|
@ -2863,9 +2884,9 @@ impl<
|
|||
self.load_layout_stack_size(elem_layout, Symbol::DEV_TMP3);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let ret_fields =
|
||||
if let LayoutRepr::Struct(field_layouts) = self.layout_interner.get_repr(*ret_layout) {
|
||||
|
@ -2960,9 +2981,9 @@ impl<
|
|||
self.load_layout_stack_size(elem_layout, Symbol::DEV_TMP2);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let lowlevel_args = bumpalo::vec![
|
||||
in self.env.arena;
|
||||
|
@ -3028,9 +3049,9 @@ impl<
|
|||
self.load_layout_stack_size(elem_layout, Symbol::DEV_TMP3);
|
||||
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(dst, self.layout_interner.stack_size(*ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, *dst, *ret_layout);
|
||||
|
||||
let lowlevel_args = [
|
||||
list,
|
||||
|
@ -3074,7 +3095,10 @@ impl<
|
|||
}
|
||||
|
||||
fn create_empty_array(&mut self, sym: &Symbol) {
|
||||
let base_offset = self.storage_manager.claim_stack_area(sym, 24);
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area_with_alignment(*sym, 24, 8);
|
||||
|
||||
self.storage_manager
|
||||
.with_tmp_general_reg(&mut self.buf, |_storage_manager, buf, reg| {
|
||||
ASM::mov_reg64_imm64(buf, reg, 0);
|
||||
|
@ -3161,7 +3185,7 @@ impl<
|
|||
self.storage_manager.with_tmp_general_reg(
|
||||
&mut self.buf,
|
||||
|storage_manager, buf, tmp_reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(sym, 24);
|
||||
let base_offset = storage_manager.claim_stack_area_with_alignment(*sym, 24, 8);
|
||||
ASM::mov_base32_reg64(buf, base_offset, ptr_reg);
|
||||
|
||||
ASM::mov_reg64_imm64(buf, tmp_reg, elements.len() as i64);
|
||||
|
@ -3510,9 +3534,11 @@ impl<
|
|||
return;
|
||||
}
|
||||
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(&allocation, element_width);
|
||||
let base_offset = self.storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
allocation,
|
||||
element_layout,
|
||||
);
|
||||
|
||||
let ptr_reg = self.storage_manager.claim_general_reg(&mut self.buf, &ptr);
|
||||
|
||||
|
@ -3636,9 +3662,11 @@ impl<
|
|||
.chain(nullable_id + 1..number_of_tags);
|
||||
|
||||
let table = self.debug_symbol("tag_id_table");
|
||||
let table_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(&table, (number_of_tags * 2) as _);
|
||||
let table_offset = self.storage_manager.claim_stack_area_with_alignment(
|
||||
table,
|
||||
(number_of_tags * 2) as _,
|
||||
8,
|
||||
);
|
||||
|
||||
let mut offset = table_offset;
|
||||
for i in it {
|
||||
|
@ -3744,7 +3772,7 @@ impl<
|
|||
UnionLayout::NonRecursive(field_layouts) => {
|
||||
let id_offset = data_size - data_alignment;
|
||||
let base_offset = self.storage_manager.claim_stack_area_with_alignment(
|
||||
sym,
|
||||
*sym,
|
||||
data_size,
|
||||
data_alignment,
|
||||
);
|
||||
|
@ -3941,7 +3969,7 @@ impl<
|
|||
let (data_size, data_alignment) =
|
||||
union_layout.data_size_and_alignment(self.layout_interner);
|
||||
let to_offset = self.storage_manager.claim_stack_area_with_alignment(
|
||||
&scratch_space,
|
||||
scratch_space,
|
||||
data_size,
|
||||
data_alignment,
|
||||
);
|
||||
|
@ -3993,11 +4021,14 @@ impl<
|
|||
fields,
|
||||
);
|
||||
|
||||
let (data_size, _) = union_layout.data_size_and_alignment(self.layout_interner);
|
||||
let (data_size, data_alignment) =
|
||||
union_layout.data_size_and_alignment(self.layout_interner);
|
||||
let scratch_space = self.debug_symbol("scratch_space");
|
||||
let to_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(&scratch_space, data_size);
|
||||
let to_offset = self.storage_manager.claim_stack_area_with_alignment(
|
||||
scratch_space,
|
||||
data_size,
|
||||
data_alignment,
|
||||
);
|
||||
|
||||
// this is a cheaty copy, because the destination may be wider than the source
|
||||
let (from_offset, _) =
|
||||
|
@ -4070,7 +4101,11 @@ impl<
|
|||
self.storage_manager.with_tmp_general_reg(
|
||||
&mut self.buf,
|
||||
|storage_manager, buf, reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(sym, 16);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*sym,
|
||||
Layout::U128,
|
||||
);
|
||||
|
||||
let mut num_bytes = [0; 8];
|
||||
num_bytes.copy_from_slice(&bytes[..8]);
|
||||
|
@ -4108,7 +4143,11 @@ impl<
|
|||
self.storage_manager.with_tmp_general_reg(
|
||||
&mut self.buf,
|
||||
|storage_manager, buf, reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(sym, 16);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*sym,
|
||||
Layout::DEC,
|
||||
);
|
||||
|
||||
let mut num_bytes = [0; 8];
|
||||
num_bytes.copy_from_slice(&bytes[..8]);
|
||||
|
@ -4129,7 +4168,12 @@ impl<
|
|||
self.storage_manager.with_tmp_general_reg(
|
||||
&mut self.buf,
|
||||
|storage_manager, buf, reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(sym, 24);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*sym,
|
||||
Layout::STR,
|
||||
);
|
||||
|
||||
let mut bytes = [0; 24];
|
||||
bytes[..x.len()].copy_from_slice(x.as_bytes());
|
||||
bytes[23] = (x.len() as u8) | 0b1000_0000;
|
||||
|
@ -4455,7 +4499,11 @@ impl<
|
|||
(U64, U128) => {
|
||||
let src_reg = self.storage_manager.load_to_general_reg(buf, src);
|
||||
|
||||
let base_offset = self.storage_manager.claim_stack_area(dst, 16);
|
||||
let base_offset = self.storage_manager.claim_stack_area_layout(
|
||||
self.layout_interner,
|
||||
*dst,
|
||||
Layout::U128,
|
||||
);
|
||||
|
||||
let tmp = Symbol::DEV_TMP;
|
||||
let tmp_reg = self.storage_manager.claim_general_reg(buf, &tmp);
|
||||
|
@ -4553,9 +4601,9 @@ impl<
|
|||
dst: Symbol,
|
||||
) {
|
||||
// Setup the return location.
|
||||
let base_offset = self
|
||||
.storage_manager
|
||||
.claim_stack_area(&dst, self.layout_interner.stack_size(ret_layout));
|
||||
let base_offset =
|
||||
self.storage_manager
|
||||
.claim_stack_area_layout(self.layout_interner, dst, ret_layout);
|
||||
|
||||
let tmp = self.debug_symbol("call_with_stack_return_result");
|
||||
|
||||
|
@ -4811,7 +4859,7 @@ impl<
|
|||
tmp_reg: GeneralReg,
|
||||
offset: i32,
|
||||
) {
|
||||
let base_offset = storage_manager.claim_stack_area(&dst, 24);
|
||||
let base_offset = storage_manager.claim_stack_area_with_alignment(dst, 24, 8);
|
||||
|
||||
ASM::mov_reg64_mem64_offset32(buf, tmp_reg, ptr_reg, offset);
|
||||
ASM::mov_base32_reg64(buf, base_offset, tmp_reg);
|
||||
|
@ -4840,7 +4888,8 @@ impl<
|
|||
return;
|
||||
}
|
||||
|
||||
let base_offset = storage_manager.claim_stack_area(&dst, stack_size);
|
||||
// 16 is a guess here. In practice this is the highest alignment we encounter in roc datastructures
|
||||
let base_offset = storage_manager.claim_stack_area_with_alignment(dst, stack_size, 16);
|
||||
|
||||
if size - copied >= 8 {
|
||||
for _ in (0..(size - copied)).step_by(8) {
|
||||
|
@ -4896,7 +4945,11 @@ impl<
|
|||
storage_manager.with_tmp_general_reg(
|
||||
buf,
|
||||
|storage_manager, buf, tmp_reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(&dst, 16);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
layout_interner,
|
||||
dst,
|
||||
Layout::U128,
|
||||
);
|
||||
|
||||
ASM::mov_reg64_mem64_offset32(buf, tmp_reg, ptr_reg, offset);
|
||||
ASM::mov_base32_reg64(buf, base_offset, tmp_reg);
|
||||
|
@ -4943,7 +4996,11 @@ impl<
|
|||
Builtin::Decimal => {
|
||||
// same as 128-bit integer
|
||||
storage_manager.with_tmp_general_reg(buf, |storage_manager, buf, tmp_reg| {
|
||||
let base_offset = storage_manager.claim_stack_area(&dst, 16);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
layout_interner,
|
||||
dst,
|
||||
Layout::DEC,
|
||||
);
|
||||
|
||||
ASM::mov_reg64_mem64_offset32(buf, tmp_reg, ptr_reg, offset);
|
||||
ASM::mov_base32_reg64(buf, base_offset, tmp_reg);
|
||||
|
@ -5079,7 +5136,9 @@ impl<
|
|||
}
|
||||
|
||||
let (from_offset, stack_size) = storage_manager.stack_offset_and_size(&value);
|
||||
debug_assert!(from_offset % 8 == 0);
|
||||
|
||||
// on x86_64 this is too strict.
|
||||
// debug_assert_eq!(from_offset % 8, 0);
|
||||
|
||||
storage_manager.with_tmp_general_reg(buf, |_storage_manager, buf, tmp_reg| {
|
||||
let mut copied = 0;
|
||||
|
|
|
@ -679,7 +679,7 @@ impl<
|
|||
self.symbol_storage_map.insert(*sym, NoData);
|
||||
return;
|
||||
}
|
||||
let base_offset = self.claim_stack_area(sym, struct_size);
|
||||
let base_offset = self.claim_stack_area_layout(layout_interner, *sym, *layout);
|
||||
|
||||
let mut in_layout = *layout;
|
||||
let layout = loop {
|
||||
|
@ -955,7 +955,7 @@ impl<
|
|||
pub fn ensure_symbol_on_stack(&mut self, buf: &mut Vec<'a, u8>, sym: &Symbol) {
|
||||
match self.remove_storage_for_sym(sym) {
|
||||
Reg(reg_storage) => {
|
||||
let base_offset = self.claim_stack_size(8);
|
||||
let base_offset = self.claim_stack_size_with_alignment(8, 8);
|
||||
match reg_storage {
|
||||
General(reg) => ASM::mov_base32_reg64(buf, base_offset, reg),
|
||||
Float(reg) => ASM::mov_base32_freg64(buf, base_offset, reg),
|
||||
|
@ -1015,7 +1015,7 @@ impl<
|
|||
match self.remove_storage_for_sym(sym) {
|
||||
Reg(reg_storage) => {
|
||||
debug_assert_eq!(reg_storage, wanted_reg);
|
||||
let base_offset = self.claim_stack_size(8);
|
||||
let base_offset = self.claim_stack_size_with_alignment(8, 8);
|
||||
match reg_storage {
|
||||
General(reg) => ASM::mov_base32_reg64(buf, base_offset, reg),
|
||||
Float(reg) => ASM::mov_base32_freg64(buf, base_offset, reg),
|
||||
|
@ -1135,7 +1135,7 @@ impl<
|
|||
) {
|
||||
match layout_interner.get_repr(layout) {
|
||||
single_register_layouts!() | pointer_layouts!() => {
|
||||
let base_offset = self.claim_stack_size(8);
|
||||
let base_offset = self.claim_stack_size_with_alignment(8, 8);
|
||||
self.symbol_storage_map.insert(
|
||||
symbol,
|
||||
Stack(Primitive {
|
||||
|
@ -1156,7 +1156,7 @@ impl<
|
|||
if stack_size == 0 {
|
||||
self.no_data(&symbol);
|
||||
} else {
|
||||
self.claim_stack_area(&symbol, stack_size);
|
||||
self.claim_stack_area_layout(layout_interner, symbol, layout);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -1288,32 +1288,47 @@ impl<
|
|||
self.join_param_map.insert(*id, param_storage);
|
||||
}
|
||||
|
||||
/// claim_stack_area is the public wrapper around claim_stack_size.
|
||||
/// It also deals with updating symbol storage.
|
||||
/// It returns the base offset of the stack area.
|
||||
/// It should only be used for complex data and not primitives.
|
||||
pub fn claim_stack_area(&mut self, sym: &Symbol, size: u32) -> i32 {
|
||||
self.claim_stack_area_with_alignment(sym, size, 8)
|
||||
/// Claim space on the stack for a certain layout. Size and alignment are handled
|
||||
///
|
||||
/// This function:
|
||||
///
|
||||
/// - deals with updating symbol storage.
|
||||
/// - should only be used for complex data and not primitives.
|
||||
/// - returns the base offset of the stack area.
|
||||
pub(crate) fn claim_stack_area_layout(
|
||||
&mut self,
|
||||
layout_interner: &STLayoutInterner<'_>,
|
||||
sym: Symbol,
|
||||
layout: InLayout<'_>,
|
||||
) -> i32 {
|
||||
let (size, alignment) = layout_interner.stack_size_and_alignment(layout);
|
||||
self.claim_stack_area_with_alignment(sym, size, Ord::min(alignment, 8))
|
||||
}
|
||||
|
||||
/// Claim space on the stack of a certain size and alignment.
|
||||
///
|
||||
/// This function:
|
||||
///
|
||||
/// - deals with updating symbol storage.
|
||||
/// - should only be used for complex data and not primitives.
|
||||
/// - returns the base offset of the stack area.
|
||||
pub fn claim_stack_area_with_alignment(
|
||||
&mut self,
|
||||
sym: &Symbol,
|
||||
sym: Symbol,
|
||||
size: u32,
|
||||
alignment: u32,
|
||||
) -> i32 {
|
||||
let base_offset = self.claim_stack_size_with_alignment(size, alignment);
|
||||
let base_offset = self.claim_stack_size_with_alignment(size, Ord::min(alignment, 8));
|
||||
self.symbol_storage_map
|
||||
.insert(*sym, Stack(Complex { base_offset, size }));
|
||||
.insert(sym, Stack(Complex { base_offset, size }));
|
||||
self.allocation_map
|
||||
.insert(*sym, Rc::new((base_offset, size)));
|
||||
.insert(sym, Rc::new((base_offset, size)));
|
||||
base_offset
|
||||
}
|
||||
|
||||
pub fn claim_pointer_stack_area(&mut self, sym: Symbol) -> i32 {
|
||||
let size = 8;
|
||||
|
||||
let base_offset = self.claim_stack_size(size);
|
||||
// pointers are 8 bytes wide with an alignment of 8
|
||||
let base_offset = self.claim_stack_size_with_alignment(8, 8);
|
||||
|
||||
self.symbol_storage_map.insert(
|
||||
sym,
|
||||
|
@ -1326,13 +1341,9 @@ impl<
|
|||
base_offset
|
||||
}
|
||||
|
||||
/// claim_stack_size claims `amount` bytes from the stack alignind to 8.
|
||||
/// claim_stack_size claims `amount` bytes from the stack
|
||||
/// This may be free space in the stack or result in increasing the stack size.
|
||||
/// It returns base pointer relative offset of the new data.
|
||||
fn claim_stack_size(&mut self, amount: u32) -> i32 {
|
||||
self.claim_stack_size_with_alignment(amount, 8)
|
||||
}
|
||||
|
||||
fn claim_stack_size_with_alignment(&mut self, amount: u32, alignment: u32) -> i32 {
|
||||
debug_assert_ne!(amount, 0);
|
||||
|
||||
|
|
|
@ -253,7 +253,7 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Syste
|
|||
|
||||
#[inline(always)]
|
||||
fn load_args<'a>(
|
||||
_buf: &mut Vec<'a, u8>,
|
||||
buf: &mut Vec<'a, u8>,
|
||||
storage_manager: &mut StorageManager<
|
||||
'a,
|
||||
'_,
|
||||
|
@ -281,7 +281,7 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Syste
|
|||
}
|
||||
|
||||
for (in_layout, sym) in args.iter() {
|
||||
state.load_arg(storage_manager, layout_interner, *sym, *in_layout);
|
||||
state.load_arg(buf, storage_manager, layout_interner, *sym, *in_layout);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -307,7 +307,7 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Syste
|
|||
if Self::returns_via_arg_pointer(layout_interner, ret_layout) {
|
||||
// Save space on the stack for the result we will be return.
|
||||
let base_offset =
|
||||
storage_manager.claim_stack_area(dst, layout_interner.stack_size(*ret_layout));
|
||||
storage_manager.claim_stack_area_layout(layout_interner, *dst, *ret_layout);
|
||||
// Set the first reg to the address base + offset.
|
||||
let ret_reg = Self::GENERAL_PARAM_REGS[general_i];
|
||||
general_i += 1;
|
||||
|
@ -413,7 +413,8 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Syste
|
|||
}
|
||||
_ if !Self::returns_via_arg_pointer(layout_interner, layout) => {
|
||||
let size = layout_interner.stack_size(*layout);
|
||||
let offset = storage_manager.claim_stack_area(sym, size);
|
||||
let offset =
|
||||
storage_manager.claim_stack_area_layout(layout_interner, *sym, *layout);
|
||||
if size <= 8 {
|
||||
X86_64Assembler::mov_base32_reg64(buf, offset, Self::GENERAL_RETURN_REGS[0]);
|
||||
} else if size <= 16 {
|
||||
|
@ -971,6 +972,7 @@ struct X64_64SystemVLoadArgs {
|
|||
impl X64_64SystemVLoadArgs {
|
||||
fn load_arg<'a>(
|
||||
&mut self,
|
||||
buf: &mut Vec<u8>,
|
||||
storage_manager: &mut X86_64StorageManager<'a, '_, X86_64SystemV>,
|
||||
layout_interner: &mut STLayoutInterner<'a>,
|
||||
sym: Symbol,
|
||||
|
@ -990,6 +992,7 @@ impl X64_64SystemVLoadArgs {
|
|||
self.argument_offset += stack_size as i32;
|
||||
}
|
||||
LayoutRepr::LambdaSet(lambda_set) => self.load_arg(
|
||||
buf,
|
||||
storage_manager,
|
||||
layout_interner,
|
||||
sym,
|
||||
|
@ -1001,12 +1004,10 @@ impl X64_64SystemVLoadArgs {
|
|||
self.argument_offset += stack_size as i32;
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Int(IntWidth::U128 | IntWidth::I128)) => {
|
||||
storage_manager.complex_stack_arg(&sym, self.argument_offset, stack_size);
|
||||
self.argument_offset += stack_size as i32;
|
||||
self.load_arg_general_128bit(buf, storage_manager, sym);
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Decimal) => {
|
||||
storage_manager.complex_stack_arg(&sym, self.argument_offset, stack_size);
|
||||
self.argument_offset += stack_size as i32;
|
||||
self.load_arg_general_128bit(buf, storage_manager, sym);
|
||||
}
|
||||
LayoutRepr::Union(UnionLayout::NonRecursive(_)) => {
|
||||
// for now, just also store this on the stack
|
||||
|
@ -1036,6 +1037,33 @@ impl X64_64SystemVLoadArgs {
|
|||
}
|
||||
}
|
||||
|
||||
fn load_arg_general_128bit(
|
||||
&mut self,
|
||||
buf: &mut Vec<u8>,
|
||||
storage_manager: &mut X86_64StorageManager<'_, '_, X86_64SystemV>,
|
||||
sym: Symbol,
|
||||
) {
|
||||
type ASM = X86_64Assembler;
|
||||
|
||||
let reg1 = X86_64SystemV::GENERAL_PARAM_REGS.get(self.general_i);
|
||||
let reg2 = X86_64SystemV::GENERAL_PARAM_REGS.get(self.general_i + 1);
|
||||
|
||||
match (reg1, reg2) {
|
||||
(Some(reg1), Some(reg2)) => {
|
||||
let offset = storage_manager.claim_stack_area_with_alignment(sym, 16, 16);
|
||||
|
||||
ASM::mov_base32_reg64(buf, offset, *reg1);
|
||||
ASM::mov_base32_reg64(buf, offset + 8, *reg2);
|
||||
|
||||
self.general_i += 2;
|
||||
}
|
||||
_ => {
|
||||
storage_manager.complex_stack_arg(&sym, self.argument_offset, 16);
|
||||
self.argument_offset += 16;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn load_arg_float(
|
||||
&mut self,
|
||||
storage_manager: &mut X86_64StorageManager<'_, '_, X86_64SystemV>,
|
||||
|
@ -1081,7 +1109,11 @@ impl X64_64WindowsFastCallLoadArgs {
|
|||
match X86_64WindowsFastcall::GENERAL_PARAM_REGS.get(self.general_i) {
|
||||
Some(ptr_reg) => {
|
||||
// if there is a general purpose register available, use it to store a pointer to the value
|
||||
let base_offset = storage_manager.claim_stack_area(&sym, stack_size);
|
||||
let base_offset = storage_manager.claim_stack_area_layout(
|
||||
layout_interner,
|
||||
sym,
|
||||
in_layout,
|
||||
);
|
||||
let tmp_reg = X86_64WindowsFastcall::GENERAL_RETURN_REGS[0];
|
||||
|
||||
copy_to_base_offset::<_, _, ASM>(
|
||||
|
@ -1115,8 +1147,10 @@ impl X64_64WindowsFastCallLoadArgs {
|
|||
self.argument_offset += stack_size as i32;
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Int(IntWidth::U128 | IntWidth::I128)) => {
|
||||
storage_manager.complex_stack_arg(&sym, self.argument_offset, stack_size);
|
||||
self.argument_offset += stack_size as i32;
|
||||
self.load_arg_general_128bit(buf, storage_manager, sym);
|
||||
}
|
||||
LayoutRepr::Builtin(Builtin::Decimal) => {
|
||||
self.load_arg_general_128bit(buf, storage_manager, sym);
|
||||
}
|
||||
LayoutRepr::Union(UnionLayout::NonRecursive(_)) => {
|
||||
// for now, just also store this on the stack
|
||||
|
@ -1146,6 +1180,33 @@ impl X64_64WindowsFastCallLoadArgs {
|
|||
}
|
||||
}
|
||||
|
||||
fn load_arg_general_128bit(
|
||||
&mut self,
|
||||
buf: &mut Vec<u8>,
|
||||
storage_manager: &mut X86_64StorageManager<'_, '_, X86_64WindowsFastcall>,
|
||||
sym: Symbol,
|
||||
) {
|
||||
type ASM = X86_64Assembler;
|
||||
|
||||
let reg1 = X86_64WindowsFastcall::GENERAL_PARAM_REGS.get(self.general_i);
|
||||
let reg2 = X86_64WindowsFastcall::GENERAL_PARAM_REGS.get(self.general_i + 1);
|
||||
|
||||
match (reg1, reg2) {
|
||||
(Some(reg1), Some(reg2)) => {
|
||||
let offset = storage_manager.claim_stack_area_with_alignment(sym, 16, 16);
|
||||
|
||||
ASM::mov_base32_reg64(buf, offset, *reg1);
|
||||
ASM::mov_base32_reg64(buf, offset + 8, *reg2);
|
||||
|
||||
self.general_i += 2;
|
||||
}
|
||||
_ => {
|
||||
storage_manager.complex_stack_arg(&sym, self.argument_offset, 16);
|
||||
self.argument_offset += 16;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn load_arg_float(
|
||||
&mut self,
|
||||
storage_manager: &mut X86_64StorageManager<'_, '_, X86_64WindowsFastcall>,
|
||||
|
@ -1360,7 +1421,7 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Windo
|
|||
if Self::returns_via_arg_pointer(layout_interner, ret_layout) {
|
||||
// Save space on the stack for the result we will be return.
|
||||
let base_offset =
|
||||
storage_manager.claim_stack_area(dst, layout_interner.stack_size(*ret_layout));
|
||||
storage_manager.claim_stack_area_layout(layout_interner, *dst, *ret_layout);
|
||||
|
||||
// Set the first reg to the address base + offset.
|
||||
let ret_reg = Self::GENERAL_PARAM_REGS[general_i];
|
||||
|
@ -1464,8 +1525,8 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Windo
|
|||
// For windows (and zig 0.9 changes in zig 0.10) we need to match what zig does,
|
||||
// in this case uses RAX & RDX to return the value
|
||||
LayoutRepr::I128 | LayoutRepr::U128 => {
|
||||
let size = layout_interner.stack_size(*layout);
|
||||
let offset = storage_manager.claim_stack_area(sym, size);
|
||||
let offset =
|
||||
storage_manager.claim_stack_area_layout(layout_interner, *sym, *layout);
|
||||
X86_64Assembler::mov_base32_reg64(buf, offset, X86_64GeneralReg::RAX);
|
||||
X86_64Assembler::mov_base32_reg64(buf, offset + 0x08, X86_64GeneralReg::RDX);
|
||||
}
|
||||
|
@ -1474,7 +1535,8 @@ impl CallConv<X86_64GeneralReg, X86_64FloatReg, X86_64Assembler> for X86_64Windo
|
|||
}
|
||||
_ if !Self::returns_via_arg_pointer(layout_interner, layout) => {
|
||||
let size = layout_interner.stack_size(*layout);
|
||||
let offset = storage_manager.claim_stack_area(sym, size);
|
||||
let offset =
|
||||
storage_manager.claim_stack_area_layout(layout_interner, *sym, *layout);
|
||||
if size <= 8 {
|
||||
X86_64Assembler::mov_base32_reg64(buf, offset, Self::GENERAL_RETURN_REGS[0]);
|
||||
} else {
|
||||
|
|
|
@ -455,7 +455,6 @@ fn build_object<'a, B: Backend<'a>>(
|
|||
);
|
||||
}
|
||||
AssemblyBackendMode::Binary => { /* do nothing */ }
|
||||
>>>>>>> 12686f23b (repl helper codegen)
|
||||
}
|
||||
|
||||
build_proc_symbol(
|
||||
|
|
|
@ -3970,11 +3970,28 @@ fn mul_checked_dec() {
|
|||
assert_evals_to!(
|
||||
indoc!(
|
||||
r#"
|
||||
Num.mulChecked 5.0dec 2.0dec == Ok 10.0dec
|
||||
Num.mulChecked 5.0dec 2.0dec
|
||||
"#
|
||||
),
|
||||
true,
|
||||
bool
|
||||
RocResult::ok(RocDec::from_str("10.0").unwrap()),
|
||||
RocResult<RocDec, ()>
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[cfg(any(feature = "gen-llvm", feature = "gen-wasm", feature = "gen-dev"))]
|
||||
fn mul_checked_u128() {
|
||||
assert_evals_to!(
|
||||
indoc!(
|
||||
r#"
|
||||
x : Result U128 [ Overflow ]
|
||||
x = Num.mulChecked 5u128 2u128
|
||||
|
||||
x
|
||||
"#
|
||||
),
|
||||
RocResult::ok(5u128 * 2u128),
|
||||
RocResult<u128, ()>
|
||||
);
|
||||
}
|
||||
|
||||
|
|
|
@ -2061,7 +2061,7 @@ fn non_unary_union_with_lambda_set_with_imported_toplevels_issue_4733() {
|
|||
_ -> (\a -> a)
|
||||
|
||||
|
||||
main = ((fn "*") 3) * ((fn "+") 5)
|
||||
main = ((fn "*") 3i64) * ((fn "+") 5)
|
||||
"#
|
||||
),
|
||||
90,
|
||||
|
|
|
@ -216,7 +216,7 @@ pub fn helper(
|
|||
roc_bitcode::host_tempfile().expect("failed to write host builtins object to tempfile");
|
||||
|
||||
// TODO make this an environment variable
|
||||
if false {
|
||||
if true {
|
||||
let file_path = std::env::temp_dir().join("app.o");
|
||||
println!("gen-test object file written to {}", file_path.display());
|
||||
std::fs::copy(&app_o_file, file_path).unwrap();
|
||||
|
|
|
@ -425,9 +425,19 @@ fn list_contains() {
|
|||
|
||||
#[cfg(not(feature = "wasm"))]
|
||||
#[test]
|
||||
fn list_sum() {
|
||||
fn list_sum_empty() {
|
||||
expect_success("List.sum []", "0 : Num a");
|
||||
}
|
||||
|
||||
#[cfg(not(feature = "wasm"))]
|
||||
#[test]
|
||||
fn list_sum_num() {
|
||||
expect_success("List.sum [1, 2, 3]", "6 : Num *");
|
||||
}
|
||||
|
||||
#[cfg(not(feature = "wasm"))]
|
||||
#[test]
|
||||
fn list_sum_frac() {
|
||||
expect_success("List.sum [1.1, 2.2, 3.3]", "6.6 : Frac *");
|
||||
}
|
||||
|
||||
|
@ -677,16 +687,28 @@ fn type_problem() {
|
|||
}
|
||||
|
||||
#[test]
|
||||
fn issue_2149() {
|
||||
fn issue_2149_i8_ok() {
|
||||
expect_success(r#"Str.toI8 "127""#, "Ok 127 : Result I8 [InvalidNumStr]");
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn issue_2149_i8_err() {
|
||||
expect_success(
|
||||
r#"Str.toI8 "128""#,
|
||||
"Err InvalidNumStr : Result I8 [InvalidNumStr]",
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn issue_2149_i16_ok() {
|
||||
expect_success(
|
||||
r#"Str.toI16 "32767""#,
|
||||
"Ok 32767 : Result I16 [InvalidNumStr]",
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn issue_2149_i16_err() {
|
||||
expect_success(
|
||||
r#"Str.toI16 "32768""#,
|
||||
"Err InvalidNumStr : Result I16 [InvalidNumStr]",
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
//! Provides Rust representations of Roc data structures.
|
||||
#![cfg_attr(not(feature = "std"), no_std)]
|
||||
// #![cfg_attr(not(feature = "std"), no_std)]
|
||||
#![crate_type = "lib"]
|
||||
|
||||
use arrayvec::ArrayString;
|
||||
|
@ -227,10 +227,19 @@ impl<T, E> Drop for RocResult<T, E> {
|
|||
}
|
||||
}
|
||||
|
||||
#[derive(Copy, Clone, PartialEq, Eq, Hash, Debug)]
|
||||
#[derive(Copy, Clone, PartialEq, Eq, Hash)]
|
||||
#[repr(C)]
|
||||
pub struct RocDec([u8; 16]);
|
||||
|
||||
impl Debug for RocDec {
|
||||
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
||||
f.debug_tuple("RocDec")
|
||||
.field(&self.0)
|
||||
.field(&self.to_str())
|
||||
.finish()
|
||||
}
|
||||
}
|
||||
|
||||
impl RocDec {
|
||||
pub const MIN: Self = Self(i128::MIN.to_ne_bytes());
|
||||
pub const MAX: Self = Self(i128::MAX.to_ne_bytes());
|
||||
|
|
|
@ -547,6 +547,8 @@ impl<T> Drop for RocList<T> {
|
|||
ManuallyDrop::drop(&mut *elements.as_ptr().add(index));
|
||||
}
|
||||
|
||||
dbg!(self.ptr_to_allocation());
|
||||
|
||||
// Release the memory.
|
||||
roc_dealloc(self.ptr_to_allocation(), Self::alloc_alignment());
|
||||
}
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue