this post was submitted on 09 Mar 2024
368 points (94.0% liked)

Technology

59569 readers
3431 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] eleitl@lemmy.ml 15 points 8 months ago* (last edited 8 months ago) (1 children)

Tracking a moving object in realtime with video is a standard task for a machine learning engineer. You can do it on an embedded platform with ML hardware support. I don't know what hardware newer Lancets use but they can already do it according from developer reports from Telegram channels like e.g Разработчик БПЛА.

[–] Warl0k3@lemmy.world 5 points 8 months ago* (last edited 8 months ago) (2 children)

Honestly, I was just objecting to the use of "AI". We've had both fire and forget and loitering munitions for decades now, neither of which use ML. Will it happen? Sure. For now, ML/AI is too unreliable to be trusted in a deployed direct attack platform, and we dont have computing hardware powerful enough to run ML models that we can jam in a missile.

(Though yeah we run tons of models against drone data feeds, none of those are done onboard...)

[–] barsoap@lemm.ee 3 points 8 months ago* (last edited 8 months ago)

For now, ML/AI is too unreliable to be trusted in a deployed direct attack platform

And probably can't ever be trusted. That "hallucinations can't ever be ruled out" result is for language models but should probably apply to vision, too. In any case researchers made cars see things and AFAIU they didn't even have to attack the model they simply confused the radar. Militaries are probably way better at that than anything that's out in the open, they've been doing ECM for ages and of course never tell anyone how any of it works.

That doesn't mean that ML can't be used, though, you can have additional non-ML mission parameters such as the drone only acquiring targets over enemy territory. Or that the AI is merely the gunner, there's still a human commander.

[–] eleitl@lemmy.ml 2 points 8 months ago

The point of modern deep learning approaches is that they're extremely easy on the developer skill. Decades ago realtime machine vision needed a machine vision expert, these days you throw the hardware at the problem at learning stage, and embedded devices to run the results are stupidly powerful (doesn't even take a Jetson board), if you compare to what has been available even a decade ago.