Dynamically sized but stored contiguously makes the systems performance engineer in me weep. If the lists get big, the kernel is going to do so much churn.
matlab likes to pick the smallest available spot in memory to store a list, so for loops that increase the size of a matrix it’s recommended to preallocate the space using a matrix full of zeros!
Dynamically sized but stored contiguously makes the systems performance engineer in me weep. If the lists get big, the kernel is going to do so much churn.
Contiguous storage is very fast in terms of iteration though often offsetting the cost of allocation
Modern CPUs are also extremely efficient at dealing with contiguous data structures. Branch prediction and caching get to shine on them.
Avoiding memory access or helping CPU access it all upfront switches physical domain of computation.
matlab likes to pick the smallest available spot in memory to store a list, so for loops that increase the size of a matrix it’s recommended to preallocate the space using a matrix full of zeros!
Is that churn or chum? (RN or M)
Churm