Maybe eventually, it has to do with market share and if the service is a "core platform". Signal doesn't have enough market share to warrant it yet, even iMessage wasn't forced to since it's not that popular in EU. The law was mainly targeted at WhatsApp as that's THE messenger in the EU.
Not_mikey
I think you misunderstand what apples value proposition is, at least nowadays. The app store and not being able to use other app stores is not a reason people get iPhones. Maybe back when app stores were first created and the threat of malware was greater people might have considered it but nowadays no one cares. Even the idea of a unified ecosystem isn't as much a selling point any more because Google and Samsung offer similar seamless integrations with their accessories. You can see this in their marketing, they aren't focused on how all the apple products work together easily any more. In their marketing you can see what they think their value proposition is, and what was their big Superbowl ad this year, longer battery life ...
Apple at this point knows it doesn't have much of a value proposition for switching from android. So the only way they're gonna sell new phones is to get the kids who don't have a phone and convince the people who do have an iPhone to get a new one.
They convince the kids through their tried and true aesthetics and lifestyle marketing, this is about half there marketing these days. This along with iMessage in the U.S. and the general fear of being in the out group and obsession with brands that younger people have moved them towards iPhones.
They convince the current users with incremental upgrades, eg. Better battery life, better camera; and maintaining the walled garden and keeping exit costs high so they don't turn to androids for those incremental updates.
All this is to say that apple having a single app store isn't a sign of consumer sentiment, but a sign of apples desire to milk as much profits out of their current users as they can. Other app stores can only benefit the consumers, either they do get them for lower fees or don't because they put some value on the "ecosystem". From a company's perspective yes your right that they want to be able to do anything to their product they want, but the goal of regulation is to step in when the companies desires are at odds with the people or the consumers desire, this is one of those cases.
If anything this was worse under the old system. Making art previously costed a lot of money, you had to pay the artists for their time and money, and better artists cost more. So in the past that oil company could commission 100 top quality artists to make corporate propaganda while a person who cares for the environment but has no money could only make a drawing limited by their own personal technical artistic ability, which could be just stick figures.
This is why "high quality" consumerist and capitalist "art" and branding in the form of advertising is so abundant meanwhile anti-consumerist, anti-capitalist art is rarer, no one's paying to get it made.
Now any cause, regardless of money, can create at least mid art to get there message across. Those causes can also have way more people behind them then an oil company can reasonably hire
It's sort of like how the gun changed how power worked. Previously a king could use there resources to pay for a smaller army of well equipped highly trained knights to subjugate a group of people. Then when the gun came training and equipment didn't matter nearly as much and it became more of a numbers game, and to get those numbers rulers needed to give more power to the masses in order to be able to marshall them for their cause. Those rulers who didn't got overthrown in revolutions.
Why would real meaning and messages be harder to find, does AI generated art inherently have less meaning?
Let's say I wanted to convey the message that oil companies are destroying the environment so , throwing subtlety out the window, come up with an idea of "a vampiric oil baron draining mother nature of oil", does the picture that is generated from me putting that prompt into an AI generator have any less meaning then if I actually drew it myself?
For all the advances in AI it still lacks intentionality, and always will under these current models, that has to be supplied by the person in the form of a prompt. I'd say that intention is the source of messages and meaning in art. AI just allows people without technical abilities in art to express those intentions, feelings and messages.
degeneration might be the ONLY way to prevent Faceboom
So we should just start posting furry porn to try and scare them off?
That study is like giving a written test to an illiterate adult, seeing them do worse than a child and saying they aren't intelligent or innovative. Like I said earlier intelligence is multi-faceted, and chatgpt excels at rhetorical, conversational and other types of written intelligence. It does not, as that study shows, do well in spatial manipulation, that doesn't mean it's not intelligent. If you gave that same test to a paralyzed blind person with little to no concept of spatial reality they'd probably do just as bad. If you asked them to compose a short story or an essay they might be good at it because that's where they're capabilities lye. That short story could still be innovative in its composition and characters, and could be way better than anything a child wrote.
You have to measure different types of intelligence with different tests. If you asked chatgpt and a set of adults and children to write a short story about a wholey new subject chatgpt would beat most of the children and probably some of the adults.
And if that short story is about a new subject matter completey out of its training set what/who is it plagiarizing from? You could say it's taking common tropes, themes and story elements from other stories, but that's fundamentally what a lot of writing and culture is. If that's plagiarism then you should be more worried about the marvel franchise as it's a plagiarism machine that has way more cultural impact.
It can answer questions as well as any person. Just because you may need to check with another source doesn't mean it didn't answer the question it just means you can't fully trust it. If I ask someone who's the fourth u.s. president and they say Jefferson they still answered the question, they just answered it wrong. You also don't have to check with another source in the same way you do with asking a person a question, if it sounds right. If that person answered Madison and I faintly recall it and think it sounds right I will probably not check their answer and take it as fact.
For example I asked chatgpt for a chocolate chip cookie recipe once. I make cookies pretty often so would know if the recipe seemed off but the one it provided seemed good, I followed it and made some pretty good cookies. It answered the question correctly as shown by the cookies. You could argue it plagiarized but while the ingredients and steps were pretty close to some I found later none were a perfect match which is about as good as you can get with recipes which tend to converge in the same thing. The only real difference between most of them is the dumb story they give at the beginning which thankfully chatgpt doesn't do.
The 7th grader and plagiarism comment make me think you haven't played with them much or really tested them. I have had it write contracts, one of which I had reviewed by a lawyer who only had some small comments, as well as other letters and documents I needed for my mortgage and buying a home. All of these were looked over by proffesionals and none of them realized it was a bot. None of them were plagiarized too because the parameters I gave it and the output it created were way too unique to be in its training set.
I'm not saying it doesn't encode some of its training data, I'm saying it's not just encoding its training data. It probably does "memorize" a bunch of trivial facts from its training data and regurgitate them when asked. I'm saying that's not all they are and that's not what makes the intelligent, their ability to also answer questions outside their training data is.
Auto complete is not a lossy encoding of a database either, it's a product of a dataset, just like you are a product of your experiences, but it is not wholly representative of that dataset.
A wind tunnel is not intelligent because it doesn't answer questions or process knowledge/data it just creates data. A wind tunnel will not answer the question "is this aerodynamic" but you can observe a wind tunnel and use your intelligence to process that and answer the question.
Temperature and randomness don't explain hallucinations, they are a product of inference. If you turned the temperature down to 0 and asked it the question " what happened in the great Christmas fire of 1934" it will give it's best guess of what happened then even though that question is not in it's dataset and it can't look up the answer. The temperature would just mean that between runs it would consistently give the same story, the one that is most statistically probable, as opposed to another one that may be less probable but was pushed up due to randomness. Hallucinations are a product of inference, of taking something at face value then trying to explain it. People will do this too, if you tell someone a lie confidently then ask them about it they will use there intelligence to rationalize a story about what happened.
This is just like Canada banning foreign investment in real estate. It admits there's a problem, data harvesting , homes as investments, but just solves a small part of the problem pertaining to "foreign bad guys" while ignoring the larger domestic issue.
All it does is make the government look like they did something without actually confronting the powerful interests that are causing the problem.