good point. im trying to get it to write an original joke with a set up, assumption and expectation-subversion punchline. its kind of done it but its not funnyMaybe if that wasn't an old joke
To be clear, when you say “explanation”, it isn’t an attempt to explain a mysterious phenomenon. Humans built these chatbots and this is how they work. They create each word/sentence (that bit is a bit fuzzy to me) based on the current state of the system. There is no forward planning (“purpose”) and there is no memory of past states of the system (“experience”)what I dont get with the "its just predicting the next word, one by one" explanation …
Too philosophical for me, I've got nothing to say to that.To be clear, when you say “explanation”, it isn’t an attempt to explain a mysterious phenomenon. Humans built these chatbots and this is how they work. They create each word/sentence (that bit is a bit fuzzy to me) based on the current state of the system. There is no forward planning (“purpose”) and there is no memory of past states of the system (“experience”)
The way it works* is that the entire context of the chat is submitted to GPT. One of the reasons ChatGPT has taken off so much compared to GPT-3 and before, is the ease of use for non-technical people, and it's this chat context which makes the magic happen.yes..... in that moment between answering once and not getting the answer fully correct and then questioning its answer and improving on it its not using brute force to reread every bit of data it has been given... a different process is occurring which i have no idea about and which seems to be surprising experts
By leaving 3 empty, you were giving it clues on how to continue and it would give you more three more in addition to the two you provided.5 Reasons to start a succulent garden
1. Succulents are low-mainenance
2. Succulents are drough-tolerant
3.
The context limit includes both input and output, so if you write a lot of instructions to ChatGPT it might not have enough context to give you a proper answer. I think context sizes are going to be the thing that differentiates the next batch of LLM improvements. It's already 'Good Enough' to do lots of tasks, but only if the tasks are bit sized. You couldn't say, "write a book", for instance.What are tokens?
Tokens can be thought of as pieces of words. Before the API processes the prompts, the input is broken down into tokens. These tokens are not cut up exactly where the words start or end - tokens can include trailing spaces and even sub-words. Here are some helpful rules of thumb for understanding tokens in terms of lengths:
Or
- 1 token ~= 4 chars in English
- 1 token ~= ¾ words
- 100 tokens ~= 75 words
- 1-2 sentence ~= 30 tokens
- 1 paragraph ~= 100 tokens
- 1,500 words ~= 2048 tokens
It doesn’t “remember” what you’ve said earlier, it is just creating a response based on the current state of the chat, which includes the full script. If you turn it off and then go back later with a different session, it has no memory of your earlier “conversation” with it. There is no persistence of memory. If it were a person, it would be one with massively damaging retrograde amnesia.Too philosophical for me, I've got nothing to say to that.
If you have a session with gpt4 it remembers what you've said earlier and those earlier interactions colour later answers, so there is some memory of past states. What's more gpt4 is improving since launch directly because of interactions. It's is learning and improving. Wether that is "experience" or not isn't a big deal to me
complicated to talk about thisThe way it works* is that the entire context of the chat is submitted to GPT. One of the reasons ChatGPT has taken off so much compared to GPT-3 and before, is the ease of use for non-technical people, and it's this chat context which makes the magic happen.
Using GPT-3 before the release of ChatGPT, you had to submit a prompt which was something like this:
The bits in green are the response, and the first line is the prompt. It was literally like an auto-completer.
If you got it wrong, you had to re-write the entire prompt, and you got a complete new set of responses, with no knowledge/history of the previous attempts. You could include example prompts to guide GPT-3 how to answer, like this:
By leaving 3 empty, you were giving it clues on how to continue and it would give you more three more in addition to the two you provided.
You could include negative prompts to guide it away from things you don't want to see.
What ChatGPT did was simplify this by including the previous context in each request. Instead of prompting and giving lots of examples, you can just ask and then if it gets it wrong, you can say, "no, not that" and what's happening in the background is something like this:
You: Give me 5 reals to start a succulent garden
GPT does some analysis of your prompt and gives you 5 answers including things about space and cost, etc.
You: No, I meant things like the mental health benefits
GPT turns your initial prompt and your follow up into a new prompt. Something like: "5 reasons to start a succulent garden with a focus on lifestyle and wellbeing instead of practical concerns"
Then it gives you a new answer. If you ask it more, it re-prompts based on the context so far, plus the new info, and gives a new answer.
So it's still one answer one response, but the prompt is getting more complex in the background. So, the prompts will grow in size and complexity as you drill down further by chatting to it. There is a context limit (of I think 8k tokens) on GPT-4. (A token is not a word.) Here's OpenAI's descriptions of a token:
The context limit includes both input and output, so if you write a lot of instructions to ChatGPT it might not have enough context to give you a proper answer. I think context sizes are going to be the thing that differentiates the next batch of LLM improvements. It's already 'Good Enough' to do lots of tasks, but only if the tasks are bit sized. You couldn't say, "write a book", for instance.
But with more context, I think GPT4 could already do it.
*I think
That’s funny, but also terrifying from my perspective because it completely blows apart the controls we currently rely on to manage risk in the organisation.I'm quite interested in what happens when Microsoft's Co-Pilot comes online.
It raises the possibility of a lot of office workers becoming totally cut-off from the actual content of their job. E.g., someone uses an LLM to write a report for them, based on a bunch of data they don't understand. I, receiving their report via email, ask Co-Pilot to summarise it for me and put that summary in an email to my colleagues, and set an agenda for a meeting about it. Co-Pilot suggests talking points for the meeting, at which we all sit around pretending to understand the report. Co-Pilot then produces notes and action points from that meeting and we all finish the day feeling very productive.
im not that interested about a sense of AI personhood, at this point at least, im interested in the mechanics of its 'thinking'. a different emphasis.It doesn’t “remember” what you’ve said earlier, it is just creating a response based on the current state of the chat, which includes the full script. If you turn it off and then go back later with a different session, it has no memory of your earlier “conversation” with it. There is no persistence of memory. If it were a person, it would be one with massively damaging retrograde amnesia.
It could theoretically be given memory persistence — and persistence across all its interactions, to boot — but this ability to build up “experience” is being intentionally withheld because of concerns about the consequences. What all this means is that any sense that the machine is having an actual meaningful conversation with you is something generated by you — by your projection of humanity onto it — not by the chat bot itself. The sense of personhood comes from your representation of it in your own mind, not any consciousness in the machine. A “machine in the mind” rather than a “mind in the machine”
is it not simply recognising the question as an instruction to redo the task (which sometimes took a few attempts to recognise)? and in one of the examples it redid it and made a whole new mistake.complicated to talk about this
in the video about GPT4 teaching itself the example demonstrated was
DO TASK X -
It did it with a few mistakes
then it got asked DID YOU DO IT CORRECTLY?
it answered No, let me try again, and then did it correctly
-so its not got any new context or more complexity to deal with, its somehow reevaluated its answer against the question and improved on it. Does that contradict what you are saying?
Totally fair enough not to think this is the most interesting question atm. But I don't think conversations about consciousness, or sentience if you prefer (I prefer consciousness), are impossible. Plenty is now written about animal consciousness that has a lot of value and has a scientific basis - you just need to listen to the right people.im not that interested about a sense of AI personhood, at this point at least, im interested in the mechanics of its 'thinking'. a different emphasis.
ive heard career philosophers on tv arguing (wrongly, surely) that dogs are automatons - shows how impossible these kind of Is It Sentient conversations are
is it not simply recognising the question as an instruction to redo the task (which sometimes took a few attempts to recognise)?
That's not an audience, it's a workforce.right, so the big change is ease of use and bringing it to a mass market audience. great PR for AI and OpenAI.
The fact that you instinctively understand them to be wrong demonstrates that actually these conversations are very very possible.im not that interested about a sense of AI personhood, at this point at least, im interested in the mechanics of its 'thinking'. a different emphasis.
ive heard career philosophers on tv arguing (wrongly, surely) that dogs are automatons - shows how impossible these kind of Is It Sentient conversations are
I am also interested in the copyright angle.
Italy’s privacy watchdog bans ChatGPT over data breach concerns
Measure is in place ‘until ChatGPT respects privacy’, says Italian Data Protection Authoritywww.theguardian.com
yes agree on that especially - its 'sentience' may be unlike anything we have previously encounteredIt will work out its own rules.
what I dont get with the "its just predicting the next word, one by one" explanation
see aboveAs mentioned earlier it's not something thought up to explain the result, it's what the program does.
Regarding its potential for understanding, I don't think we can quite judge that yet. The history of AI has been full of confident predictions that something like general intelligence is just around the corner, ever since the 1950s.
At the moment, they aren’t linking chatbots to persistence in memory, as I understand it. Every line it gives is just text prediction based on the prompt at that time. There’s no experience involved.
As mentioned earlier it's not something thought up to explain the result, it's what the program does.
this is the most interesting video ive yet seen. most videos are about new developments, this is more on history and analysis
this man Geoffrey Hinton - Wikipedia has been working on AI for 40+ years and its his + colleagues neural network model, dismissed for years, that has resulted in the current breakthrough
interesting video for lots of reasons but highlight for me
-shows how these machines are modelled directly on the human brain and how they are able to learn and store the lessons that they learn, by making new "neural pathways"
-explains how they recognise images, again based on how animals (including us) do it
-on the question of sentience thinks people are being too quick to dismiss it outright, and we need to consider different definitions of sentience. Describes it as an idiot savant
-says that the Its Predicting The Next Word simplification is fundamentally true,but misses out that the bot understands the sentences and discerns meaning before responding - he gives an example of that in action. there is some degree of comprehension/context going on of the words being inputed
-also talks about some worries, particularly the military implications
- big issue about who decides what the truth is and adds that bias to the bot. We;ve already had one Left Wing Bias headline in the UK press
see above
33mins12 in the video
I saw an example where it referred to a previous conversation that the user hadn't had with it, then when he questioned it, it claimed that conversation had been with another user.There is a memory of a sort going on. GPT3.5 is definitely referencing things which I mentioned to it earlier in conversations.
I think the latest versions are beginning to output results that are increasingly hard to explain solely by "word by word prediction".
Early experiments with GPT-4 [...] we discuss the challenges ahead for advancing towards deeper and more comprehensive versions of AGI, including the possible need for pursuing a new paradigm that moves beyond next-word prediction.
[...]
Loosely speaking, we can therefore see the drawbacks of the model as a combination of “naive” attention mistakes with more fundamental limitations due to its “linear thinking” as a next-token prediction machine
[...]
the model also has several substantial flaws, some of which seem to be inherent to the next-word prediction paradigm that underlies its architecture.
Quite. I had a conversation last week with a PhD student who is studying exactly this for his PhD. He’s excited for but very cautious about what can be done if you do start allowing the machine to gain experience, but that’s definitely not what’s happening at the moment. What we’re seeing at the moment is the interesting way humans project their own perspective onto non-human things. It’s kind of like seeing faces in clouds.I saw an example where it referred to a previous conversation that the user hadn't had with it, then when he questioned it, it claimed that conversation had been with another user.
Again, that's what the people who wrote it say.