this post was submitted on 05 Nov 2025
-41 points (15.3% liked)

Technology

76711 readers
4521 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

... in the United States, public investment in science seems to be redirected and concentrated on AI at the expense of other disciplines. And Big Tech companies are consolidating their control over the AI ecosystem. In these ways and others, AI seems to be making everything worse.

This is not the whole story. We should not resign ourselves to AI being harmful to humanity. None of us should accept this as inevitable, especially those in a position to influence science, government, and society. Scientists and engineers can push AI towards a beneficial path. Here’s how.

The essential point is that, like with the climate crisis, a vision of what positive future outcomes look like is necessary to actually get things done. Things with the technology that would make life better. They give a handful of examples and provide broad categories if activities that can help steer what is done.

you are viewing a single comment's thread
view the rest of the comments
[–] Artisian@lemmy.world 0 points 4 days ago

Success would lead to AI use that properly accounted for its environmental impact and had to justify it's costs. That likely means much AI use stopping, and broader reuse of models that we've already invested in (less competition in the space please).

The main suggestion in the article is regulation, so I don't feel particularly understood atm. The practical problem is that, like oil, LLM use can be done locally at a variety of scales. It also provides something that some people want a lot:

  • Additional (poorly done) labor. Sometimes that's all you need for a project
  • Emulation of proof of work to existing infrastructure (eg, job apps)
  • Translation and communication customization

It's thus extremely difficult to regulate into non-existence globally (and would probably be bad if we did). So effective regulation must include persuasion and support for the folks who would most benefit from using it (or you need a huge enforcement effort, which I think has its own downsides).

The problem is that even if everyone else leaves the hole, there will still be these users. Just like drug use, piracy, or gambling, it's easier to regulate when we make a central easy to access service and do harm reduction. To do this you need a product that meets the needs and mitigates the harms.

Persuading me I'm directionally wrong would require such evidence as:

  • Everyone does want to leave the hole (hard, I know people who don't. And anti-AI messaging thus far has been more about signaling than persuasion)
  • That LLMs really can't/can be made difficult to be done locally (hard, the Internet gives too much data, and making computing time expensive has a lot of downsides)
  • Proposed regulation that would actually be enforceable at reasonable cost (haven't thought hard about it, maybe this is easy?)