this post was submitted on 20 Mar 2025
499 points (100.0% liked)

Technology

67151 readers
6523 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 76 points 2 days ago (2 children)

The theory that the lead maintainer had (he is an actual software developer, I just dabble), is that it might be a type of reinforcement learning:

  • Get your LLM to create what it thinks are valid bug reports/issues
  • Monitor the outcome of those issues (closed immediately, discussion, eventual pull request)
  • Use those outcomes to assign how "good" or "bad" that generated issue was
  • Use that scoring as a way to feed back into the model to influence it to create more "good" issues

If this is what's happening, then it's essentially offloading your LLM's reinforcement learning scoring to open source maintainers.

[–] [email protected] 39 points 2 days ago

Thats wild. I don't have much hope for llm's if things like this is how they are doing things and I would not be surprised given how well they don't work. Too much quantity over quality in training.

[–] [email protected] 2 points 1 day ago

Honestly, I would be alright with this if the AI companies paid Github so that the server infrastructure can be upgraded. Having AI that can figure out bugs and error reports could be really useful for our society. For example, your computer rebooting for no apparent reason? The AI can check the diagnostic reports, combine them with online reports, and narrow down the possibilities.

In the long run, this could also help maintainers as well. If they can have AI for testing programs, the maintainers won't have to hope for volunteers or rely on paid QA for detecting issues.

What Github & AI companies should do, is an opt-in program for maintainers. If they allow the AI to officially make reports, Github should offer an reward of some kind to their users. Allocate to each maintainer a number of credits so that they can discuss the report with the AI in realtime, plus $10 bucks for each hour spent on resolving the issue.

Sadly, I have the feeling that malignant capitalism would demand maintainers to sacrifice their time for nothing but irritation.