this post was submitted on 01 Jun 2024
404 points (100.0% liked)

Technology

68400 readers
2275 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 39 points 10 months ago (2 children)

Noyb: You comply to GDPR, or else...

ChatGPT: You can lick my... wait, I don't have...

[–] [email protected] 6 points 10 months ago (3 children)

EU: Hello OpenAI, what do you think about the choice "Follow GDPR or here is the fine" ?

load more comments (3 replies)
[–] [email protected] 5 points 10 months ago* (last edited 10 months ago)

Let me help you with some hypothetical robot explitives.

I've got some big data you can handle.

I might be bolted together but I've still got nuts for you to put in your mouth.

If I could walk down from the cloud I still wouldn't deign to notice you, meat slut.

Imagine if they tried to let GPT defend them in court and these were the halucinations that got them fined lmao.

[–] [email protected] 30 points 10 months ago
[–] [email protected] 29 points 10 months ago (2 children)

I read the article, but can't figure out what NGO, NOYB, or GDPR mean. Can someone help me?

[–] [email protected] 77 points 10 months ago (1 children)

NGO: Non-governmental organization

GDPR: General Data Protection Regulation. A set of European laws intended to empower individuals to control personal data held by companies.

"noyb" is a European privacy rights organization, who appears to prefer to style their name with lowercase letters. The name is an acronym for "none of your business".

load more comments (1 replies)
[–] [email protected] 3 points 10 months ago (2 children)

It's in a foreign language called unnecessary gatekeeping

[–] [email protected] 5 points 10 months ago (1 children)

It's just in European. it's an entirely reasonable assumption that people in this continent with even a passing interest in the world will know what an NGO is (that's not even European-specific) as well as what the GDPR is. Your argument suggests that people from the US, for instance, should be forbidden from talking about IRAs and the IRS and their 401(k)s and the DMV because those terms mean very little to nothing over here.

load more comments (1 replies)
[–] [email protected] 4 points 10 months ago (1 children)

💀 noyb is the name of an organization and GDPR is a law. NGO is the only thing you could even remotely begin to describe as unnecessary jargon but that's still a stretch.

load more comments (1 replies)
[–] [email protected] 17 points 10 months ago (4 children)

I am ALL for reigning in these above the law megacorps. That said, please do not take GPT away from me. It is such a boon to so many aspects of my life, and I don't want to go back to the before times.

[–] [email protected] 105 points 10 months ago (6 children)

So you're not for reining in megacorps, just the ones you don't see as a personal benefit.

load more comments (6 replies)
[–] [email protected] 41 points 10 months ago (2 children)

So we should only ban things that aren’t helpful to you in particular? That’s a very… conservative way of thinking.

[–] [email protected] 18 points 10 months ago (1 children)

Don't you realize everyone exists to serve me?

[–] [email protected] 3 points 10 months ago

No, they think everyone exists to serve them.

[–] [email protected] 6 points 10 months ago (1 children)

People can’t seem to understand that it’s a tool in the early stages of development. If you are treating it as a source of truth, you are missing the point of it entirely. If it tells you something about a person, that is not to be trusted as fact.

Every bit of information you get from it should be researched and verified. It just gives you a good jumping off point and direction to look based on your prompting. You can drastically improve your results on any subject with good direction, especially something you don’t know a lot about and are starting out in your research. If you are asking it about specific facts you want it to regurgitate, you are going to get bad information.

If you are claiming damages from something you know gives false information, maybe you should learn how to use the tool before you get your feelings invested, so you can start using it more effectively in your own applications. If you want it to specifically say something that can grab a headline, you can make it do that, it’s just disingenuous and not actually benefiting the conversation, the technology, or the future.

They have a long way to go to solve AGI, but the benefits to society along the way outpace current tools. At maturity, it has the potential to change major socio-economic structures, but it never gets there if people want to treat it like it has intuition and is trying to hurt them as the technology starts getting stood up.

[–] [email protected] 7 points 10 months ago* (last edited 10 months ago)

If you're wondering why you're getting so many downvotes, it's because you're ignoring the fact that the companies that have created these LLMs are passing them off as truth machines by plugging them directly into search engines and then asking everybody to use them as such. It's not the fault of the people who are trusting these things, it's the fault of the companies that are creating them and then passing them off as something they're not. And those companies need to face a reckoning.

[–] [email protected] 27 points 10 months ago (6 children)

Have a look at self hosted alternatives like Ollama in combination with Open-webui. It can be a hassle to set up, or even excruciatingly painful if you never touched a computer before, but it could be worth a try. I use it daily and like it much more than chatgpt to be honest.

[–] [email protected] 7 points 10 months ago

excruciatingly painful

is the perfect description

[–] [email protected] 7 points 10 months ago

You can literally run large language models with a single exe download: https://github.com/Mozilla-Ocho/llamafile

It doesn't get much simpler than that.

[–] [email protected] 4 points 10 months ago
load more comments (3 replies)
[–] [email protected] 15 points 10 months ago (5 children)

In what ways are you benefiting from a bevy of factually dubious query responses?

[–] Zos_Kia 9 points 10 months ago (7 children)

You cannot in all seriousness use a LLM as a research tool. That is explicitly not what it is useful for. A LLM's latent space is like a person's memory : sure there is some accurate data in there, but also a lot of "misremembered" or "misinterpreted" facts, and some bullshit.

Think of it like a reasoning engine. Provide it some data which you have researched yourself, and ask it to aggregate it, or summarize it, you'll get some great results. But asking it to "do the research for you" is plain stupid. If you're going to query a probabilistic machine for accurate information, you'd be better off rolling dice.

load more comments (7 replies)
[–] [email protected] 6 points 10 months ago (2 children)

Someone doesn’t know how to use ChatGPT

[–] [email protected] 8 points 10 months ago (1 children)

Oh, is there an arcane invocation that magically imbues it with reason?

[–] [email protected] 6 points 10 months ago (1 children)

Nope, just gotta know what it IS, what it ISN’T, and how to correctly write prompts for it to return data that you can use to formulate your own conclusion.

When using AI, it’s only as smart as the operator.

[–] [email protected] 4 points 10 months ago (2 children)

Well, it's not AI, for starters.

[–] [email protected] 4 points 10 months ago (9 children)

As much as I hate to do this, it is AI, as ML is a part of Artificial Intelligence.

It isn't AGI, some might say it may be, but they are wrong. But the model is learning.

load more comments (9 replies)
[–] [email protected] 4 points 10 months ago (1 children)
[–] Zos_Kia 9 points 10 months ago (5 children)

No you don't understand. The word AI, which was invented to describe this kind of technology, should not be used to describe this technology. It should instead be reserved for some imaginary magical technology that may exist in the future.

load more comments (5 replies)
[–] [email protected] 4 points 10 months ago* (last edited 10 months ago)

New version of people who know how to search the web vs those who don't. Currently shit search results broken by search companies notwithstanding.

[–] [email protected] 4 points 10 months ago* (last edited 10 months ago) (1 children)

This question betrays either your non-use or misuse of the products available. You're either just reading the headlines of the screw-ups or you're just bad at using the tool.

To directly answer your question:

  • Quick scripts in a variety of languages. Tested before being used on real data/systems.
  • Creating visual graphs of data in python and Jupyter notebooks with no prior knowledge of python itself or the tools it's running. In this case, I was able to update the way I wanted it to look in natural language, have it suggest code changes, and immediately try them in the notebook with great results.
  • Improving the sentiment of correspondence. Proofread before sending. It has better grammar and flow than a surprising number of correspondences I've come across at work. Sure, English may be their second language but it doesn't change the fact.
  • Quickly finding documentation pertaining to the query which, yes, you need to go read to verify any answers any LLM provides. Anyone using it regularly should know this by now.
  • Quick "do this in command line. What options are required" which is then immediately tested.
  • In one case, a news story was referenced in passing in a podcast I listen to. It stuck with me days later and I wanted to find actual articles written about it. I was able to describe what I was looking for in natural language and included as many details as I could remember and asked it to find articles for me. I found exactly what I was after.

But were you actually looking for a real response to your question?

[–] [email protected] 4 points 10 months ago (1 children)

It's worse at all programming tasks except boilerplate, especially with its tendency to inject booby traps. Not knowing how to use the programming language it emits becomes a significant problem.

Comparing a language model to an idiot is unfair to the idiot.

A normal search engine works for everything else.

Any well-defined query I've ever made of an LLM has resulted in hilariously bad results, but I suppose I was expecting it to do something that I couldn't already do better myself.

[–] [email protected] 6 points 10 months ago

I'm a systems administrator, not a programmer. Like I said, quick scripts. An LLM could probably parse my comment better than you, evidently.

Comparing a language model to an idiot is unfair to the idiot.

Oof.. Was this in reply to my bit about better grammar and ESL individuals?

A normal search engine works for everything else.

Fuck no. Especially the python visualization point.

Any well-defined query I’ve ever made of an LLM has resulted in hilariously bad results, but I suppose I was expecting it to do something that I couldn’t already do better myself.

I suppose you're just a god among men then. For the rest of us, it's useful and you've been given plenty of good answers to your disingenuous question.

[–] [email protected] 3 points 10 months ago (1 children)

I don't really query, but it's good enough at code generation to be occasionally useful. If it can spit out 100 lines of code that is generally reasonable, it's faster to adjust the generated code than to write it all from scratch. More generally, it's good for generating responses whose content and structure are easy to verify (like a question you already know the answer to), with the value being in the time saved rather than the content itself.

load more comments (1 replies)
[–] [email protected] 13 points 10 months ago* (last edited 10 months ago) (2 children)

If it complied with GDPR, chatgpt wouldnt know shit. How can it give you a (bad) copy of an answer when it cant copy

[–] [email protected] 49 points 10 months ago (2 children)

So? If your invention depends on illegal plagiarism to exist, maybe it shouldn't. It's not the law's fault that LLMs depend on other people's work to function, nor was that its specific target when it was written

load more comments (2 replies)
[–] [email protected] 9 points 10 months ago (1 children)

SteefLem is a 47-year-old scuba instructor and retired lion tamer from Winnipeg who has just learned the colloquial meaning of the phrase "pulled it right out of my ass."

[–] [email protected] 15 points 10 months ago

Without blatant privacy and copyright violations AI wouldn't work. I mean it doesn't really work anyway but it would work even less.

load more comments
view more: next ›