Heh. Asstronomers.
Monument
A base plate that’s got a spring under it, except for a little nub that pokes the power button.
Terrible if you live in earthquake-prone areas.
Wait. Are we describing a bump stock for your computer?
(I have another response.)
Lean away, give her a long look, and gaze deep into her eyes. Tilt your head slightly to the side.
“…. dad?”
In the stunned silence, because she never expected you to be right, tell her that you didn’t think you would see her again after she disappeared on her way to get cigarettes all those years ago.
And then put your hand on her thigh and say “I’ve missed you daddy” with wide eyes.
Gets ‘em every time.
“Oooh, I don’t know. Maaaaybeee your lips, because I can’t stop looking at them. But if you did, then your doctor is really good.”
Deliver it light and kind of flirty and she’s not thinking about whatever trap that question is, she’s thinking about how you just told her you want to kiss her.
If it’s for any other reason - making sure an insecurity is unknown, wanting to springboard into a conversation, wanting to see how you react if she baits you into saying something dumb, or even having her own flirty line to deliver about it - she’ll probably get to it. Assuming you haven’t wooed her into a voracious make out session.
You say “Not even close.” in response to the suggestion that Apple’s research can be used to improve benchmarks for AI performance, but then later say the article talks about how we might need different approaches to achieve reasoning.
Now, mind you - achieving reasoning can only happen if the model is accurate and works well. And to have a good model, you must have good benchmarks.
Not to belabor the point, but here’s what the article and study says:
The article talks at length about the reliance on a standardized set of questions - GSM8K, and how the questions themselves may have made their way into the training data. It notes that modifying the questions dynamically leads to decreases in performance of the tested models, even if the complexity of the problem to be solved has not gone up.
The third sentence of the paper (Abstract section) says this “While the performance of LLMs on GSM8K has significantly improved in recent years, it remains unclear whether their mathematical reasoning capabilities have genuinely advanced, raising questions about the reliability of the reported metrics.” The rest of the abstract goes on to discuss (paraphrased in layman’s terms) that LLM’s are ‘studying for the test’ and not generally achieving real reasoning capabilities.
By presenting their methodology - dynamically changing the evaluation criteria to reduce data pollution and require models be capable of eliminating red herrings - the Apple researchers are offering a possible way benchmarking can be improved.
Which is what the person you replied to stated.
The commenter is fairly close, it seems.
they would have nowhere to retreat if their food supply ran out.
Um. Hello? There are scientists there.
Which means scientific papers, then tourists, then garbage and a symbiotic relationship, then the eventual domestication of polar bears.
Not, you know, the international scientific community treating scientists like cats.
That’s very fair, indeed.
Perhaps awareness of one will spark awareness of the other. I suppose my concern is that plasticisers are sort of a ‘hidden’ risk, for the most part. They’re used in nearly every food packaging (and prep, such as hoses) that isn’t contained in glass, or served up in its own peel.
Microplastics are terrifying and all that, but I’m sort of more worried about plasticisers like BPA, BPF, BPS and the rest of the alphabet of BP-whatever’s that was created and brought into use after the dangers of BPA were realized.
Just a heads up - if something plastic says it’s BPA-free, it probably uses a different bisphenol compound that is less studied than BPA. And is likely as toxic (or even more toxic)!
But nobody ever talks about those, because science words.
“In this here melting pot, we burn away all our differences until we’re left with only the pure white flame of Christian nationalism.”
I had to put the statement in quotes because while being hyperbole, it’s not too far from how some people think, and I don’t want to be confused with those folks.
The problem is that now the first page of results is all AI garbage and wrong, so you’re not 100% sure at what point you’ve reached the sane internet.
You’re one of the founding members of the greater Seattle area polycule, aren’t you?