this post was submitted on 14 Jan 2024
1432 points (97.3% liked)
Greentext
4437 readers
1401 users here now
This is a place to share greentexts and witness the confounding life of Anon. If you're new to the Greentext community, think of it as a sort of zoo with Anon as the main attraction.
Be warned:
- Anon is often crazy.
- Anon is often depressed.
- Anon frequently shares thoughts that are immature, offensive, or incomprehensible.
If you find yourself getting angry (or god forbid, agreeing) with something Anon has said, you might be doing it wrong.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
as an analogue to your point about the camera being objectively better. My point is that you can't boil everything down to objective facts, even if it is true, there are a number of other variables.
With lossy compression, that quite literally gets thrown out the window the second it's used. If that's the standard then all lossy compression is bad. The question is at what point, does N amount of deviation from the original image, make it noticeably different from the original image, to the point that it negatively affects the image more than the space it saves. That's the hard part to quantify. And yet we use lossy compression everywhere. Literally nobody can agree what standard of compression is acceptable. I for one never touch HW accelerated encoding because it's not efficient, and introduces artifacts. Yet other people are perfectly content using it. I would much rather store the original source file, even if it's insanely big, over HW encoding it down to something more manageable, and potentially forever altering that file.
It depends on what standardized photo testing you use. If i can take a photo roughly 2-10 feet in front of me, and it looks decent. I do not care about anything else. If it's outside of that range my eyesight is bad enough it doesn't matter anyway. A phone with a built in zoom lense might be able to take better far shot photos. But i never take those, so it's useless to me.
i mean, if we include photo processing, that's just not true, unless major phone manufacturers have started open sourcing their software since i last checked.
i didnt look at it, but im not going to discount it either, frankly i just don't care. I just don't think more than like 30-40% of why people buy a new phone is to take better photos, maybe thats how they justify spending that money to themselves, i could see that. But JUST for better photos? idk. Maybe i'm just a bad capitalist who doesnt spend enough money.
as you already said "the best camera is the one you have with you, you continually brought up editing, and real world use cases where having a better camera would make sense. Which is where i would use that actual camera, i just don't really care about the quality of the pictures i take that aren't supposed to be actual media. It's fine enough as is. Being any better isn't going to appreciably change that.
i didn't say that, i just stated that at a certain point, an end user is going to stop caring about a "feature" when it's feature set is severely convoluted. Maybe i actually just care about what i spend my money on, and other people don't. But i like knowing what im buying, before i spend my money on it.
no my claim is that anything that is 80% efficacy is going to be more than fine, your claim is that 99% efficacy is worse than 100% efficacy, which is true, but not perceptible.
i claim my iphone 5 as having a better camera than the leapfrog leappad. My point there, even though you have butchered it incredibly, is quite literally the difference between a pixel 8 and an iphone 15. You can't go back to before multi cameras, because a modern single camera phone will still have improved since then.
i know, but for the same reason that i don't care about a 4090 ti being faster than a 1070 due to its price being fascinatingly high. I don't care about phones with more than one camera having better camera quality. I just dont want that feature.
Literally this entire thread started with "still don't understand the appeal of multi camera phones" or something like that, it's paraphrased. I know there are technical reasons one would do that, but i just can't justify it for what it provides. Unless i see an actual proper realistic breakdown, of which exactly NONE exist. So i couldn't line them up even if i wanted to. I'm just left to my own devices to see what else could be done. And so i just dont care. Same reason i dont care about phones having high refresh rates, it just wastes processing power, it feels smoother sure. I don't really care though. I use it like 10 minutes out of my day maybe. Swiping sideways at 90hz doesn't matter if i dont use it anyway. (dont bother explaining the difference to me, because i own multiple 144hz displays, i already know.)
And my point is that you can measure blurry image from a clear one. It's not subjective. Is a glass back better? Maybe. Is one photo better than another? That can be measured. It is only when two cameras take photos are almost indistinguishable that it becomes subjective.
There is a difference between blurring a little and blurring a lot. I already discussed the trade of of size vs quality that allows you to compare lossy compression. That's why JPEG has a quality option and chroma subsampling option. You can control how much detail is lost. If it literally got thrown out the window, a jpeg could not be recognizable as the original.
I already said it is a trade off of size vs appearance. You are setting up a strawman of "what if we had two cameras that made almost identical photos"
Again, no. This is about taking a photo. The process of taking a photo is the same in my current 3 camera phone as your single camera phone. Your claim that it adds complication to the user is false.
No I didn't. Please quote. You said you did video editing on a PC which was a complete non sequitur to my comment about your phone having 8 cores.
What are you talking about? Where did I mention percentages before? The Pixel 7pro camera is obviously better than my Pixel 3a in the same way my Nikon DSLR with 300mm lens was obviously better than my Canon S100 digital Elph compact camera. The Pixel 7pro needs 3 cameras to be better than the 3a. 1 camera is wide which the 3a couldn't do. One camera is about the same as the 3a (probably better if I zoomed carefully). One camera is optical zoom which is so much better than digital zoom it is clearly better. But although I happen to know the details because its my hobby, it requires nothing in the UI to use the 3 cameras. It is completely automatic with the same UI as my older Pixel 3a.
??? There is no significant difference between a Pixel 8 and an iPhone 15. They both have 2 camera modules on the back. The 8 pro and 15 pro both have 3 camera modules on the back.
You said you didn't see the need for more camera lenses. A Pixel 8 pro can take noticeably better photos than your current iPhone 5 because of the additional lenses.
It only seems convoluted to you because you haven't used it. You think it must be complicated in the same way someone would think having multiple cores in their phone cpu would be convoluted and hard to use.
The iPhone 5 is so old there aren't going to be direct comparisons. (or was that a typo and you meant you have an iphone 15) Can I argue that a 10 year old Intel PC is no faster than a current high end AMD PC because there are no websites that directly compare 10 year old PCs to current PCs? There are dozens of reputable websites that reviewed the iPhone 5. There are also reviews of the Pixel 7 pro. GSMarena.com even goes far enough back to directly compare an iPhone 7 to a Pixel 8pro. And that's not even factoring in optical zoom.
You don't care but other do. It doesn't have anything to do with modern phone culture. People upgraded their cameras to take better photos even before digital cameras.
if it was literally thrown out the window, we would be talking semantics and philosophy. Moot point anyway. Your proposal was that the pixel be identical to the pixel on the previous photo. You can make an image using a bunch of pennies in varying states of oxidation that resemble an image. You cant spit an image into an ASCII converter, and it will resemble that image. Those are quite literally not the same picture. Your proposal defined a provable system, to demonstrate that the images were identical. I said they were objectively not because they are not. Image compression also isn't explicitly blurring. But again moot point.
My point there was demonstrating that your approach to defining quality was bad.
it's more complicated than size vs appearance, but generally. Size vs quality, is how lossy compression is considered. HW accel av1 and software av1 are going to look and output vastly different media, at different sizes. They work differently, even though they use the same underlying codec. My point is that nothing is an objective binary state.
Yeah, if we obfuscate it down to ignore everything else, it's fully transparent. Much like cigarettes are good for you because they make you feel normal. Dont worry about the lung cancer stuff it's normal and happens to everybody.
do you not understand the concept of an analogy? Or even the concept of drastically simplifying concepts in a way that can be easily explained and translated between individuals without having complete and total understanding between those two individuals? it'd be weird since that's explicitly what you're doing. That statement is explaining my point of view, and explaining your arguments in turn. It should be fairly obvious why i just don't care.
the pixel 7 pro was released sometime in october in 2022. The 3a was released in august of 2020. That's 2 years between the models. Not to mention the obvious model disparity between the 3a and the 7 pro, the 7 pro being $900 and the 3a appearing to be $400, they are objectively not in the same class of phone, nor are they even in the same time period. The majority of difference in camera quality is going to be down to the sensors themselves improving, rather than having more cameras.
yeah, because any two phones today, one produced with one camera, and one produced with three, are going to take pretty similar photos. It has nothing to do with the amount of cameras, it has everything to do with the amount of capability between the two. If you think back, you'll remember my point about percentages, which seemed to have confused you. This is a literal product based interpretation of that statement. Which also seems to have gone over your head.
this is actually just wrong. The iphone 5, released in 2012. the pixel 8 pro seems to have released sometime in 2023. More than a decade apart. It takes better photos because the camera is just better. The sensor is significantly bigger, the bump well, exists now. All qualities that lead to a better camera quality. The pixel 8 pro with one camera, or three cameras, is going to take the same kind of photo regardless of what i do with it. It does have an ultrawide, of which i genuinely have no idea what i would use for. And a telephoto, which has some obvious uses. but nothing that i care about. The instances in which i would use those cameras, are still going to look bad.
if we're bringing the concept of multi core cpus in again, then i can actually hit you with some knowledge truths on this one. Multi core cpus are significantly more convoluted to handle from an OS perspective. This is why they didn't exist for the longest time. Instead of everything running sequentially on one core, there are multiple, some of them are even "phantom" cores, that only exist when the main core is busy. Which means you need to figure out a way to divide cpu time, across cores, presumably evenly because that would be the most effective manner of doing it. While also not incurring significant overhead costs, such as latency, and even cpu cycles. Because if your queue handling is bad, you might as well just have a single core. Especially if you can block up the queue, and crawl the system to a halt.
Yet another little fun fact btw. This is true for all processors, but especially so on multi core processors, to my knowledge single core processors just tended to function a little bit differently (prior to modern multi core architectures) to ensure this wasn't a problem (hardware interrupts) But on multi core cpus you need a way to ensure that a piece of software accessing the cpu can't hog the entire cpu, blocking out the OS that's managing it, and it's scheduler as well, this would cause a dead lock. Which means you wouldn't be able to do anything. Even if the end user experience is transparent, it's still more convoluted, shitting on the carpet and sweeping it under the rug doesn't remove the fact that you shit on the carpet. Unless we're arguing that convolution doesn't exist, im not sure how else it would work. Convolution incurs costs, and costs incur many things.
i could very well argue that you must think cpus are simple because you don't understand them. An objective truth of design and philosophy, is anything that adds more complexity, makes it more convoluted. There is no way around this.
even then it wouldn't be accurate, the iphone 5 is 12 years old now. The iphone 6 would be a better argument, though still not a good one. Ironically enough, dankpods recently did a video where he compared an iphone 15 and an iphone 5c. The 15 was better, but both were still bad.
there are websites for that, so you couldn't argue that anyway. Though again moot point.
because i guess taking photos was never a part of phone culture. That was just something people did randomly for no reason, and by accident. I dont integrate with the modern culture, which is why i detest it so much. A lot of people, you included don't mind it, and have opinions on it. A lot of people, me included, can't stand it because i don't want to spend exorbitant prices on what is more than required.
You keep trying to claim that quality difference is subjective. If that is your point then the JPEG quality option does nothing. Quality 1 is the same as Quality 99 because the pixels change. It is a false claim. You can objectively measure the difference from the original. Again this isn't discussing two images that are so similar that the differences are subjective such as two different encoders both at Quality 99.
I am going to repeat this again because you keep ignoring it. I AM NOT DISCUSSING TWO IMAGES WHERE THE IMAGE IS SO SIMLAR THAT QUALITY BECOMES SUBJECTIVE.
You said this, "your claim is that 99% efficacy is worse than 100% efficacy," after I had already said:
"I AM NOT DISCUSSING TWO IMAGES WHERE THE IMAGE IS SO SIMLAR THAT QUALITY BECOMES SUBJECTIVE."
There is no difference in the UI. It isn't obfuscation but presenting a complex task as simple to the user. Just like all modern technology. You don't have to know how the OS multitasks to post your replies.
Stop with the chatGPT crap. I was writing a VxD driver for Windows 3.1 32 years ago.
When typing your reply, what did you have to do to control your multiple CPU's so that you could type and post your reply. Did you have to set the affinity of the browser process to a particular core? Did you manually schedule the threads? No? I thought so.
Show me a website that compares a computer from when the iphone 5 came out directly to a computer today. Yes there are benchmarks archived of PC's from 2012 which can be compared to benchmarks in 2023. But no one has reviewed them side by side today. If you just want static numbers for an iPhone 5 there's DXOMark and GSMArena.com.
So that's your claim? An iPhone 5's camera is as good as any camera unless someone has done a review with a direct comparison? It must be as good as a Canon EOS R6 Mark II because no one has tested them side by side? Really? When you do video editing work, was the source all recorded nothing better than iPhone 5 because its all the same?
The first mass marketed camera was a produced by Alphonse Giroux in 1839. Although everyone didn't have a camera until the release of the Kodak brownie in 1900. It has nothing to do with the phone. Adding the camera to the phone made it possible to not have to buy and carry a separate device to do what everyone had been doing for over 100 years.
You use a desktop computer and do video editing! It's the phone camera users who are traditional.
i'm claiming that it's both objective, and subjective and that in this case, for me, it's primarily subjective, rather than objective, based on my usage.
you asked me where the percentages came from, i explained it.
or what you could quite literally argue is, the definition of obfuscation, though in this case i probably meant abstraction. They're basically the same tbh.
it's not chatgpt, also good to know you wrote insert device drivers for windows.
no, but my system did, having a cursory understanding of this stuff and how it works allows you to better utilize your hardware.
cpubenchmark and basically every other synthetic benchmark tool out there. cpus are a bundle of transistors, it's not hard to bench them (though it's not that simple either) phone cameras are a little more involved. You can't really just go "bigger number more better"
i literally have like two paragraphs explaining my thoughts. Not sure why you even put that in there.
amusingly, not my claim at all, the first phone that released, which really did numbers, had a camera on it. Every phone since then has had a camera. Although i should probably mention contextually here, that i am SPECIFICALLY referring to touch screen phones, the modern ones, which really invented the whole philosophy of modern phone culture, the thing that i was specifically referring to in my previous comment. Since we already established that lack of context is bad and what not.
cool, still not a normie though. Unfortunately.