Programming, LLMs 

The one avenue I've found LLMs useful in, and peculiar in, and significantly irreplaceable, is manipulating language. You know, the thing it was designed for. Probability machine go brrr

Programming, LLMs 

The way every scammer is trying to use it to bypass the law and not pay human workers is its own phenomenon, but technically, the reason why it even copies people so much is because it's just trying its best to model its training data.

If you get a somehow usable training dataset for the one domain you want to deploy the probabilistic machine in, you could use the machine and find it not doing all of that stuff. It just behaves as a fuzzy transformer for the specific thing you trained it in

And I think current LLMs could've gotten to their current grasp of English using public domain text. Even speaking with a Victorian program is good enough.

Tech, LLMs 

It's just, I think the rest of the world is having a better opinion of this stuff not only because they're less politically literate, but because of the trait of LLMs where they can switch languages arbitrarily (because in the training data, words from every language around the same concept are placed close together). So people are getting all the jank of American LLMs, *but in their language*, it's "foreign" "more advanced" jank.

So like, I think the tech is like, actually something when it comes to language, cause that's where all the research had been going into in decades prior. This thing is basically an offshoot of Google Translate

Tech, LLMs 

Everything else though, is what you'd expect probabilistic machines to output for a domain. Believable (= "looks like the real thing" = "I'm comparing it with real-world data" = "I'm comparing it with what I trained it against) but amorphously corrupting and fundamentally untrue stuff. It's a machine starting from the output, not the input, it's not an intelligence machine.

Tech, LLMs 

If for some godforsaken reason some open source dev wants to make an open source version of this tech, they can probably win by just writing the system by hand and then training it with public domain data. And then selling it as a language transforming tool, that can run on a CPU.

Thing is, everyone in the free software world is "armed to the teeth" in software, ready to implement any permanent system by themselves. It's the complete opposite of the general public, who basically don't want to lift a finger to get anything done, beyond expressing what exactly they want

Follow

Tech, LLMs 

@snowyfox open translation models trained on publicly available data exist afaik? like the one built into Firefox for offline translation works pretty well. I was trying to figure out what training data they use exactly and couldn't (but I gave up pretty quickly) but there's definitely plenty around

Programming, LLMs 

@noiob Mm, but it's not translation I'm talking about, I mean the use case where people feed user messages into an LLM to get some tags/judgements/guesses, or feed JSON into an LLM to have it normalise and massaged slightly

Where English language is involved in the field values and such, some ability to deal with English is needed, and a local LLM trained on Victorian-era books could pass the test

Sign in to participate in the conversation
Awoo Space

Awoo.space is a Mastodon instance where members can rely on a team of moderators to help resolve conflict, and limits federation with other instances using a specific access list to minimize abuse.

While mature content is allowed here, we strongly believe in being able to choose to engage with content on your own terms, so please make sure to put mature and potentially sensitive content behind the CW feature with enough description that people know what it's about.

Before signing up, please read our community guidelines. While it's a very broad swath of topics it covers, please do your best! We believe that as long as you're putting forth genuine effort to limit harm you might cause – even if you haven't read the document – you'll be okay!