• 0 Posts
  • 6 Comments
Joined 1 year ago
cake
Cake day: February 14th, 2025

help-circle




  • This is one team that disagrees out of many that agree.

    To explain what you are seeing. The above image is the inverse Fourier transform (FT) of different frequencies of sinus waves that compose an image.

    The very large baseline interferometer (VLBI) applied in the event horizon telescope (EHT) is using different telescopes all over the world, in a technique called interferometry, to achieve high enough resolutions to observe different frequencies in Fourier space that make up an image. If you observe all, you can recreate the full image perfectly. They did not, they observed for a long time and thus got a hefty amount of these “spatial” frequencies. Then they use techniques that limit the image to physical reality (e.g. no negative intensities/fluxes) and clean it from artefacts. Then transform it to image space (via the inverse FT)

    Thereby, they get an actual image that approximates reality. There is no AI used at all. The researchers from Japan argued for different approach to the data, getting a slightly different inclination in that image. This may well be as the data is still too few to 100 % determine the shape, but looks more to me like they chose very different assumptions (which many other researchers do not agree with).

    Edit: They did use ML for simulations to compare their sampling of the Fourier space to.