Open matanlurey opened 5 months ago
Labels: area-core-library
, type-enhancement
Summary: The Uint8List.fromList()
constructor is significantly slower than creating a Uint8List
with a specified length and then setting each element individually. This performance difference is likely due to the way fromList()
handles copying data.
It's unsurprising that .fromList
is slower. If nothing else, it allocates the normal list and initializes that first, which likely uses ~8x the memory of the Uint8List
on a 64-bit runtime. Then that list is passed to a constructor which spends some extra time checking if its input is a typed-data list, so it can do memcpy.
10x seems excessive, but that may just be beause Uint8List(10)..[0] = 0..[1] = 1..[2] = 2....
is incredibly fast, basically as fast as initializing the memory, and [0, ..., 9]
is just as efficient, but initializes 8x the memory, and then copies it once to the Uint8List
.
(The obvious optimization would be to optimize away the intermediate list so that Uint8List.fromList([e1, ..., en])
is compiled as Uint8List(n)..[0] = e1... ..[n-1] = en
.)
Doesn't seem to be list allocation that costs. Changing the list to a const [1, ... ,10]
only improves performance by ~5%.
A quick look at the code shows that the definition of .fromList
just calls setRange
.
That's probably what it does. If using a (reusable) Uint8List
as the source, performance improves ~x3. Still another x3 needed until the speed of just writing the bytes directly.
@mkustermann Typed data, there's always more to optimize! :)
If using a (reusable)
Uint8List
as the source, performance improves ~x3. Still another x3 needed until the speed of just writing the bytes directly.@mkustermann Typed data, there's always more to optimize! :)
For a 10-element list we are already far down the slippery slope of having too many tests to select special cases to close the final x3. I think the language needs something to help us dispatch more quickly to the appropriate specialized kernel.
I would like the full glory of control flow collections to be available to typed data lists, with both modifiable, unmodifiable, and perhaps const
variants. Let the compilers and runtime do things that we would be scared to make available to the general developer.
Indeed, the only thing here that would match the direct storing approach would be recognizing the pattern .fromList(literal)
and converting it to direct stores. Possible, but doesn't generalize.
A typed-data literal, fx Uint8List{1, 2, 3, 4 ...}
, could more easily be recognized and optimized, and/or allow general collection elements.
(I'd like the feature to be available to user types too, maybe requiring the type to have a special constructor or interface, we can still special-case platform types.)
It would be nice to have an intrinsic for a list-literal or perhaps document fromList as inefficient.
Uint8List.fromList()
isn't inefficient by itself, it depends on the argument being passed and what we know about that object.
_list = Uint8List.fromList([0, 1, 2, 3, 4, 5, 6, 7, 8, 9]);
Right now this compiles to this:
0: B0[graph]:0 {
v0 <- Constant(#null) T{Null?}
v3 <- Constant(#TypeArguments: (H16e735c6) [Type: int]) T{TypeArguments}
v4 <- Constant(#10) [10, 10] T{_Smi}
v6 <- Constant(#0) [0, 0] T{_Smi}
v7 <- Constant(#1) [1, 1] T{_Smi}
v8 <- Constant(#2) [2, 2] T{_Smi}
v9 <- Constant(#3) [3, 3] T{_Smi}
v10 <- Constant(#4) [4, 4] T{_Smi}
v11 <- Constant(#5) [5, 5] T{_Smi}
v12 <- Constant(#6) [6, 6] T{_Smi}
v13 <- Constant(#7) [7, 7] T{_Smi}
v14 <- Constant(#8) [8, 8] T{_Smi}
v15 <- Constant(#9) [9, 9] T{_Smi}
}
2: B49[function entry]:2 {
v2 <- Parameter(0 @rdi) T{_Uint8ListFromList}
}
3: ParallelMove fp[-1] <- rdi
4: CheckStackOverflow:8(stack=0, loop=0)
5: ParallelMove rbx <- C, r10 <- C
6: v5 <- CreateArray:12(v3, v4) T{_List}
8: ParallelMove fp[-2] <- rax
8: StoreIndexed([_List] v5, v6, v6, NoStoreBarrier)
10: StoreIndexed([_List] v5, v7, v7, NoStoreBarrier)
12: StoreIndexed([_List] v5, v8, v8, NoStoreBarrier)
14: StoreIndexed([_List] v5, v9, v9, NoStoreBarrier)
16: StoreIndexed([_List] v5, v10, v10, NoStoreBarrier)
18: StoreIndexed([_List] v5, v11, v11, NoStoreBarrier)
20: StoreIndexed([_List] v5, v12, v12, NoStoreBarrier)
22: StoreIndexed([_List] v5, v13, v13, NoStoreBarrier)
24: StoreIndexed([_List] v5, v14, v14, NoStoreBarrier)
26: StoreIndexed([_List] v5, v15, v15, NoStoreBarrier)
27: ParallelMove rdx <- C
28: v47 <- AllocateObject:10(cls=_GrowableList, v3 T{TypeArguments}) T{_GrowableList}
29: ParallelMove rcx <- rax, rax <- fp[-2]
30: ParallelMove fp[-3] <- rcx
30: StoreField(v47 . GrowableObjectArray.data = v5 T{_List}, NoStoreBarrier)
32: StoreField(v47 T{_GrowableList} . GrowableObjectArray.length = v4 T{_Smi}, NoStoreBarrier)
33: ParallelMove rax <- C
34: v64 <- AllocateTypedData:10(v4 T{_Smi}) T{_Uint8List}
35: ParallelMove rdi <- rax, rsi <- C, rdx <- C, rbx <- fp[-3], r8 <- C, rax <- rax
36: ParallelMove fp[-2] <- rax
36: StaticCall:166( _slowSetRange@7027147<0> v64 T{_Uint8List}, v169 T{_Smi}, v170 T{_Smi}, v47 T{_GrowableList}, v169 T{_Smi}, using unchecked entrypoint, result_type = T{Null?})
37: ParallelMove rax <- fp[-2], rcx <- fp[-1]
38: StoreField(v2 T{_Uint8ListFromList} . _list@15459445 = v64 T{_Uint8List})
39: ParallelMove rax <- C
40: DartReturn:22(v0)
*** END CFG
Without special compiler magic for the particular use (i.e. relying on normal compiler optimizations), the compiler would need to
_slowSetRange
here (which it doesn't because it's way too big)This is too much to ask from the compiler.
We could recognize this particular pattern higher up in the stack (e.g. kernel level) and rewrite it there. That would probably be ok, but not ideal as we introduce hand-crafted optimizations for library features where we depend on knowing the library implementation details.
Ideally we'd want to allow developers expressing the concept of bytes in the language - via const and non-const typed data literals. That would mostly solve this case.
If using a (reusable) Uint8List as the source, performance improves ~x3. Still another x3 needed until the speed of just writing the bytes directly.
@mkustermann Typed data, there's always more to optimize! :)
Just using a top-level Uint8List.fromList([...])
as source in this benchmark results in suboptimal code because we have to lazily initialize global, don't track length property in global type flow analysis, ... (also some other issues, e.g. filed https://github.com/dart-lang/sdk/issues/56096)
@matanlurey is your benchmark inspired by some pattern that is used frequently in Flutter framework code ?
Nothing specific - it was just surprising to me that what seemed like the easiest way to create a list of bytes was the least performant.
Possibly related: https://github.com/dart-lang/sdk/issues/32080.
It would be nice to have an intrinsic for a list-literal or perhaps document
fromList
as inefficient.Here is a micro-benchmark I wrote: