FatCrab

joined 1 year ago
[–] FatCrab@lemmy.one 1 points 1 month ago (1 children)

AI in health and medtech has been around and in the field for ages. However, two persistent challenges make roll out slow-- and they're not going anywhere because of the stakes at hand.

The first is just straight regulatory. Regulators don't have a very good or very consistent working framework to apply to to these technologies, but that's in part due to how vast the field is in terms of application. The second is somewhat related to the first but really is also very market driven, and that is the issue of explainability of outputs. Regulators generally want it of course, but also customers (i.e., doctors) don't just want predictions/detections, but want and need to understand why a model "thinks" what it does. Doing that in a way that does not itself require significant training in the data and computer science underlying the particular model and architecture is often pretty damned hard.

I think it's an enormous oversimplification to say modern AI is just "fancy signal processing" unless all inference, including that done by humans, is also just signal processing. Modern AI applies rules it is given, explicitly or by virtue of complex pattern identification, to inputs to produce outputs according to those "given" rules. Now, what no current AI can really do is synthesize new rules uncoupled from the act of pattern matching. Effectively, a priori reasoning is still out of scope for the most part, but the reality is that that simply is not necessary for an enormous portion of the value proposition of "AI" to be realized.

[–] FatCrab@lemmy.one 9 points 1 month ago

Summary judgement is not a thing separate from a lawsuit. It's literally a standard filling made in nearly every lawsuit (even if just as a hail mary). You referenced "beyond a reasonable doubt" earlier. This is also not the standard used in (US) civil cases--it's typically a standard consisting of the preponderance of the evidence.

I'm also not sure what you mean by "court approved documentation." Different jurisdictions approach contract law differently, but courts don't "approve" most contracts--parties allege there was a binding and contractual agreement, present their evidence to the court, and a mix of judge and jury determines whether under the jurisdictions laws and enforceable agreement occurred and how it can be enforced (i.e., are the obligations severable, what damages, etc.).

[–] FatCrab@lemmy.one 1 points 1 month ago

There's plenty you could do if no label was produced with a sufficiently high confidence. These are continuous systems, so the idea of "rerunning" the model isn't that crazy, but you could pair that with an automatic decrease in speed to generate more frames, stop the whole vehicle (safely of course), divert path, and I'm sure plenty more an actual domain and subject matter expert might come up with--or a whole team of them. But while we're on the topic, it's not really right to even label these confidence intervals as such--they're just output weighting associated with respective levels. We've sort of decided they vaguely match up to something kind of sort approximate to confidence values but they aren't based on a ground truth like I'm understanding your comment to imply--they entirely derive out of the trained model weights and their confluence. Don't really have anywhere to go with that thought beyond the observation itself.

[–] FatCrab@lemmy.one 2 points 1 month ago (2 children)

Are you under the impression that I think Teslas approach to AI and computer vision is anything but fucking dumb? The person said a stupid and patently incorrect thing. I corrected them. Confidence values being literally baked into how most ML architectures work is unrelated to intentionally depriving your system of one of the most robust ccomputer vision signals we can come up with right now.

[–] FatCrab@lemmy.one 3 points 1 month ago (4 children)

All probabilistic models output a confidence value, and it's very common and basic practice to gate downstream processes around that value. This person just doesn't know what they're talking about. Though, that puts them on about the same footing as Elono when it comes to AI/ML.

[–] FatCrab@lemmy.one 3 points 1 month ago

I've worked on processing submissions for this project. Honestly, it probably ends up just costing them more to do this program, which is mostly just a paid PR activity. The overwhelming majority of submissions, and I mean like 99%, are either not prior art in the sense of patent law or were already retrieved by the law firm on the case.

[–] FatCrab@lemmy.one 2 points 1 month ago (1 children)

I agree. I think the effective entry into the public domain of AI generated material, in combination with a lot of reporting/marking laws coming online is an effective incentive to keep a lot of material human made for large corporate actors who don't like releasing stuff from their own control.

What I'd like to see in addition to this is a requirement that content-producing models all be open source as well. Note, I don't think we need weird new IP rights that are effectively a "right to learn from" or the like.

[–] FatCrab@lemmy.one 7 points 1 month ago (3 children)

This has been the copyright office's stance for quite a while now. Actually, most of the world's respective IP registrars and authorities do not grant IP rights to AI generated material.

[–] FatCrab@lemmy.one 6 points 1 month ago

My z flip is hands down my favorite phone I've ever owned and I didn't get it expecting to like it much. I just needed a new phone and with Samsung's recycling program, my old near-tablet sized phone made the switch like barely 100 bucks.

There are a lot of small advantages it provides that quickly add up to it being an overall superior experience. Now if only Bixby wasn't the worst fucking thing ever.

[–] FatCrab@lemmy.one 11 points 1 month ago (1 children)

Their non-profit status had nothing to do with the legality of their training data acquisition methods. Some of it was still legal and some of it was still illegal (torrenting a bunch of books off a piracy site).

[–] FatCrab@lemmy.one 2 points 2 months ago

Oh wow, this suit is shaping up to be silly. I didn't realize it was filed in Japan, too. That makes the patent aspect even shakier. Japan has no discovery process like in the US, which is generally very necessary for many software-related patents as, assuming they have a strong likelihood of surviving challenge, they are typically drawn to processes that are completely obfuscated from the user and outside observes.

[–] FatCrab@lemmy.one 2 points 2 months ago

There is an era of patents from the late 90s through the early-mid-00s that were insanely vague and rarely stand up to scrutiny, but most are expiring at this point, if they haven't already. Generally, though, patents are not granted on "concepts" but on implementations. That's a sometimes ambiguous line, but that's a fundamental principle of modern patents.

view more: next ›