• 0 Posts
  • 13 Comments
Joined 1 year ago
cake
Cake day: July 22nd, 2023

help-circle


  • Try lots of therapists until you find the right one, don’t waste your time on ones that don’t (e: help) - you might as well talk to a wall.

    Reality: therapy is expensive, wait lists are long. And half the reason you’re feeling like this in the first place is because of how much harder it’s getting to keep ourselves fed, clothed and sheltered.




  • I’m going to take the time to illustrate here, how I can see LLMs affecting human speech through existing applications and technologies that are (or could) be made both available and popular enough to achieve this. We’re far enough down the comment chain I can reply to myself now right?

    So, we can all agree that people are increasingly using LLMs in the form of chatgpt and the like, to acquire knowledge/information. The same way as they would use a search engine to follow a link to that knowledge.

    Speech-to-text has been a thing for at least 3 decades (yeah it was pretty hopeless once, but not so much now). So let’s not argue about speech vs text. People already talk to Google and siri and whoever else to this end, llms. Pale have their responses read out via tts.

    I remember being blown away watching a blind sysadmin interacting with a Linux shell via tts at rates I couldn’t even understand the words in 1998. How far we’ve come. I digress, so.

    We’ve all experienced trouble getting the information we’re looking for even with all these tools. Because there’s so much information, and it can be very difficult to find the needle in the haystack. So we constantly have to refine our queries either to be more specific, or exclude relationships to other information.

    This in turn, causes us to think about the words we were using to get the results we want, more frequently because otherwise we spend too much time on recursion.

    In turn, the more we do this, and are trained to do this, the more it will bleed into human communication.

    Now look, there is absolutely a lot of hopium smoking going on here, but damn, this could have everlasting impact on verbal communication. If technology can train people - through inaccurate/incorrect results to think about the communication going out when they speak, we could drastically reduce the amount of miscommunication between people by that alone.

    Imagine:

    get me a chair

    wheels out an office chair from the study

    no I meant a chair for at the kitchen table

    Vs

    get me a chair for at the kitchen table

    You can apply the same thing to human prompted image generation and video generation.

    Now… We don’t need llms to do this, or know this. But we are never going to achieve this without a third party - the “llm”, and whatever it’s plugged into - because the human recipient will usually be more capable of translating these variances, or employ other contexts not as accessible via a single output as speech or text.

    But if machines train us to communicate out better (more accurately, precisely and/or concisely), that is an effect I can’t welcome enough.

    Realistically, the machines will learn to deal with us being dumb, before we adapt.

    e: formatting.


  • This is interesting and thought provoking discussion, ty.

    You’re absolutely right, I was looking for the dead end - plugging LLM into a solution.

    I’m more thinking LLMs used in conjunction with other tech will have these effects on our communicating. LLMs, or whatever replaces them to do that interpretation, are necessary to facilitate that.

    When we come up with something better, to do the same job better, then of course, LLMs will be redundant. If that happens, great.

    We are already seeing a boom in popularity of LLMs outside of professional use. Global ubiquity for anything is never going to happen, unless we can fix communication, which we probably can’t. We certainly can’t alone. It’s very much a chicken an egg problem, that we can only gain from by progressing towards.

    Imagining vocallising using programming languages gave me a chuckle. I have been known to do things like use s/x/y/ to correct in written chats though.

    Programming languages allow us to talk to and listen to machines. LLMs will hopefully allow machines to listen and talk to/between us.




  • Do you actually believe this?

    LLMs are the opposite of a dead end. More like the opening of a pipe. It’s not that they will burn out, it’s just that they’ll reach a point that they’re just one function of a more complete AI perhaps.

    At the very least they tackle a very difficult problem, of communication between human and machine. Their purpose is that. We have to tell machines what to do, when to do it, and how to do it. With such precision that there is no room for error. LLMs are not tools to prove truth, or anything.

    If you ask an LLM a question, and it gives you a response that indicates it has understood your question correctly, and you are able to understand its response that far, then the LLM has done it’s job, regardless of if the answer is correct.

    Validating the facts of the response is another function again, which would employ LLMs as a translation tool.

    It’s not a long leap from there to a language translation tool between humans, where an AI is an interpreter. deepl on roids.