The reason sensor manufacturers use "seemingly ancient" (i.e. huge feature sizes) processes in their sensors is you really don't need a more advanced process like in processors.
When you manufacture something which computes, power consumption and internal noise improvement is more drastic with improved manufacturing processes. When you are measuring something, you don't need or want too small pixels or features to begin with.
So having a small gigapixel sensor just because your process allows creates more disadvantage over having a sensor same size with a lower resolution, from light capturing angle. So, low-light sensitivity and resolution is a trade-off.
Back-illuminated sensors used by all contemporary cameras created this leap rather than reducing feature size via improved processes. You already pack the sensor as dense as possible (you don't want gaps or "smaller" pixels w/o increasing resolution either), and moving data/power plane away from pixels is the biggest contributor to noise in the sensor.
See the link [0]. Top left image is full frame, top right is APS-C, bottom left is M4/3, and bottom right is full frame / high-res (60+MP) sensors.
When you look at the images, smaller the sensor, worse the noise performance. When you compare full-size images of top left to bottom right, top left image is better in terms of noise. I selected RAW to surface "what sensor sees" The selected spot is the darkest point in that scene.
You can select JPEG to see what in camera image processing does to these images. Shutter speed is around 1/40s and ISO is fixed at 12800 since it's the de-facto standard for night photography.
> Also you bypassed the possibility of timing multiple sensors separately to achieve 2000fps.
Working on an image which doesn't reflect real world is a bit dangerous, isn't it?
[0]: https://www.dpreview.com/reviews/image-comparison?attr18=low...