Finite Field Assembly: A Language for Emulating GPUs on CPU
28 points
3 days ago
| 10 comments
| leetarxiv.substack.com
| HN
pwdisswordfishz
1 minute ago
[-]
> Field order (the number of elements your field can hold). i.e you can store 8 * 9 * 11 elements in the field

I thought a finite field's order has to be a prime power.

reply
muragekibicho
3 days ago
[-]
Introduction : Finite Field Assembly is a programming language that lets you emulate GPUs on CPUs

It's a CUDA alternative that uses finite field theory to convert GPU kernels to prime number fields.

Finite Field is the primary data structure : FF-asm is a CUDA alternative designed for computations over finite fields.

Recursive computing support : not cache-aware vectorization, not parallelization, but performing a calculation inside a calculation inside another calculation.

Extension of C89 - runs everywhere gcc is available. Context : I'm getting my math PhD and I built this language around my area of expertise, Number Theory and Finite Fields.

reply
zeroq
1 hour ago
[-]
I've read this and I've seen the site, and I still have no idea what it is, what's the application and why should I be interested.

Additionally I've tried earlier chapters and they are behind a paywall.

You need a better introduction.

reply
pizza
51 minutes ago
[-]
This is phrased in a kind of demanding way to an author who has been kind enough to share their novel work with us. Are you sure you spent enough time trying to understand?
reply
Conscat
35 minutes ago
[-]
It seems that pretty much everybody here is confused by this article. One user even accused it of LLM plagiarism, which is pretty telling in my opinion.

I for one have no clue what anything I read in there is supposed to mean. Emulating a GPU's semantics on a CPU is a topic which I thought I had a decent grasp on, but everything from the stated goals at the top of this article to the example code makes no sense to me.

reply
pizza
14 minutes ago
[-]
It just seems like residue numbering systems computation, which I'm already working with.
reply
almostgotcaught
35 minutes ago
[-]
> I'm getting my math PhD and I built this language around my area of expertise, Number Theory and Finite Fields.

Your LinkedIn says you're an undergrad that took a gap year 10 months ago (before completing your senior year) to do sales for a real estate company.

reply
pizza
8 minutes ago
[-]
Why bother doing a witch hunt and leaving out that they did Stats at Yale..
reply
almostgotcaught
2 minutes ago
[-]
Because why does it matter? Are you suggesting undergrad stats at Yale is comparable to a PhD in number theory?
reply
saghm
14 minutes ago
[-]
Depending on what properties they sold, they certainly could have gotten valuable real-world expertise with finite fields. It's certainly easier to sell them than infinite ones!
reply
saagarjha
32 minutes ago
[-]
Are you sure that’s their LinkedIn?
reply
almostgotcaught
27 minutes ago
[-]
Why wouldn't it be? All of the pics, names and details line up between GitHub, here, Reddit, and substack.
reply
adamvenis
45 minutes ago
[-]
I think I get it. You're using the Ring isomorphism from the Chinese Remainder Theorem to do "parallel computation". This is the same principle as how boolean algebra on binary strings computes the pairwise results of each bit in parallel. Unfortunately, there's no free lunch - if you want to perform K operations on N-bit integers in parallel, you still need to work with (K * N)-bit-wide vectors, which is essentially what SIMD does anyway.
reply
almostgotcaught
41 minutes ago
[-]
Yup that's exactly what this is and thus, notably, it is not actually about finite fields.
reply
vimarsh6739
32 minutes ago
[-]
One of the more subtle aspects of retargeting GPU code to run on the CPU is the presence of fine grained(read - block level and warp level) explicit synchronization mechanisms being available in the GPU. However, this is not the same in CPU land, so additional care has to be taken to handle this. One example of work which tries this is https://arxiv.org/pdf/2207.00257 .

Interestingly, in the same work, contrary to what you’d expect, transpiling GPU code to run on CPU gives ~76% speedups in HPC workloads compared to a hand optimized multi-core CPU implementation on Fugaku(a CPU only supercomputer), after accounting for these differences in synchronization.

reply
foota
50 minutes ago
[-]
It's a bit hard for me to tell the intention here. Is the idea that finite fields can take better advantage of CPU architecture than something like SIMD for parallel computation? Or is this just for experimentation?

Edit: this tickles my brain about some similar seeming sort of programming language experiment, where they were also trying to express concurrency (not inherently the same as parallelism) using some fancy math. I can't remember what it was though?

reply
hashxyz
1 hour ago
[-]
Pretty sure this is just vectorization. You can pack some 8bit ints into a machine-length 32bit int and add them together, that is vectorization.
reply
Conscat
41 minutes ago
[-]
I don't think that's true when the add overflows. You wouldn't want a lane's overflow to carry into an adjacent lane.
reply
catapart
1 hour ago
[-]
If matrix multiplication does get added to this, I imagine that there is some utility for game development. At that point, I'd be curious what the comparison would be from CPU to GPU. Like, given a clock speed of x, what would a comparable GPU (or set of GPU features) look like?

I know that's pretty abstract, but without that kind of "apples to apples" comparison, I have trouble contextualizing what kind of output is bring targeted with this kind of work.

reply
tooltechgeek
1 hour ago
[-]
What are some problems where this approach has advantages?
reply
imbusy111
47 minutes ago
[-]
I suspect this is just AI slop.
reply
almostgotcaught
42 minutes ago
[-]
It's hilarious how gullible hn is. All you gotta do is put GPU and math buzzwords in your README and you'll automatically be upvoted.

This was discussed on Reddit - this is not actually finite field arithmetic.

Also you can go to this dudes GitHub and see exactly how serious this project is.

https://github.com/LeetArxiv/Finite-Field-Assembly

Lol

reply