Natanael

joined 1 year ago
[–] Natanael@slrpnk.net 1 points 7 months ago (4 children)

Statistical associations is not equivalent to a world model, especially because they're neither deterministic nor even tries to prevent giving up conflicting answers. It models only use of language

[–] Natanael@slrpnk.net 1 points 7 months ago* (last edited 7 months ago)

It varies, there's definitely generative pieces involved but they try to not make it blatant

If we're talking evidence in court then it's practically speaking more important if the photographer themselves can testify about how accurate they think it is and how well it corresponds to what they saw. Any significantly AI edited photo effectively becomes as strong evidence as a diary entry written by a person on the scene, it backs up their testimony to a certain degree by checking for the witness' consistency over time instead of trusting it directly. The photo can lie just as much as the diary entry can, so it's a test for credibility instead.

If you use face swap then those photos are likely nearly unusable. Editing for colors and contrast, etc, still usable. Upscaling depends entirely on what the testimony is about. Identifying a person that's just a pixelated blob? Nope, won't do. Same with verifying what a scene looked like, such as identifying very pixelated objects, not OK. But upscaling a clear photo which you just wanted to be larger, where the photographer can attest to who the subject is? Still usable.

[–] Natanael@slrpnk.net 1 points 7 months ago* (last edited 7 months ago)

Depends on implementation, if done properly and if they don't try to upscale and deblur too much then that kind of interpolation between multiple frames can be useful to extract more detail. If it's a moving subject then this type of zoom can create false results because the algorithm can't tell the difference and will think it's an optical artifact. For stationary subjects and photographers it can be useful

[–] Natanael@slrpnk.net 1 points 7 months ago

There's different types of computational photography, the ones which ensures to capture enough sensor data to then interpolate in a way which accurately simulates a different camera/lighting setup are in a way "more realistic" than the ones which heavily really on complex algorithms to do stuff like deblurring. My point is essentially that the calculations done has to be founded in physics rather than in just trying to produce something artistic.

[–] Natanael@slrpnk.net 1 points 7 months ago (2 children)

When it generates additional data instead of just interpolating captured data

[–] Natanael@slrpnk.net 1 points 7 months ago

I think there's a possibility for long format video of stable scenes to use ML for higher compression ratios by deriving a video specific model of the objects in the frame and then describing their movements (essentially reducing the actual frames to wire frame models instead of image frames, then painting them in from the model).

But that's a very specific thing that probably only work well for certain types of video content (think animated stuff)

[–] Natanael@slrpnk.net 8 points 7 months ago* (last edited 7 months ago)

License plates is an interesting case because with a known set of visual symbols (known fonts used by approved plate issuers) you can often accurately deblur even very very blurry text (but not with AI algorithms, but rather by modeling the blur of the cameras and the unique blur gradients this results in for each letter). It does require a certain minimum pixel resolution of the letters to guarantee unambiguity though.

[–] Natanael@slrpnk.net 9 points 7 months ago (2 children)

But if you don't do that then the ML engine doesn't have the introspective capability to realize it failed to recreate an image

[–] Natanael@slrpnk.net 4 points 7 months ago* (last edited 7 months ago)

There's a specific type of digital zoom which captures multiple frames and takes advantage of motion between frames (plus inertial sensor movement data) to interpolate to get higher detail. This is rather limited because you need a lot of sharp successive frames just to get a solid 2-3x resolution with minimal extra noise.

[–] Natanael@slrpnk.net 6 points 7 months ago

This is just smarter post processing, like better noise cancelation, error correction, interpolation, etc.

But ML tools extrapolate rather than interpolate which adds things that weren't there

[–] Natanael@slrpnk.net 1 points 7 months ago (1 children)

Entropy and information theory is very real, it's embedded in quantum physics

[–] Natanael@slrpnk.net 6 points 7 months ago* (last edited 7 months ago) (7 children)

There's a lot of other layers in brains that's missing in machine learning. These models don't form world models and ~~some~~don't have an understanding of facts and have no means of ensuring consistency, to start with.

view more: ‹ prev next ›