Discussion about this post

User's avatar
Maynard Handley's avatar

Here's what I think is happening with the FP16 rate:

https://forums.macrumors.com/threads/apple-silicon-deep-learning-performance.2319673/page-14?post=34127092#post-34127092

I *think* this also covers the "neural cores" claim but we'll see.

As for "unified image compression" I believe this refers to the compression of INTERMEDIATE buffers (g-buffer, bump-map, etc) used to construct a final image.

Obviously imagery that is brought into the GPU as resources (basically "textures" in all their variety) is compressed by everyone. And you can compress material at the very end of an operation, once you know you're done with it.

But compression of intermediate buffers is a tougher project because it's essentially a coherency problem -- as soon as you compress parts of a buffer you now have that "buffer state" is not just spread over multiple caches but also at two different address ranges (the compressed and the uncompressed range). It's tracking all this so that compression happens as the buffer is assembled, rather than as a final additional (and synchronous rather than asynchronous stage) that's the tech part of "Unified image compression".

Expand full comment
Farfolomew's avatar

I've always loved your writing style, Ryan: simple, thorough, and with impeccable grammar (well ok, there was one minor spelling mistake!). I'm glad you've found a new home here.

I also enjoy the re-use of AnandTech's Table colors and styling :-)

Expand full comment

No posts