[snip]
The fact remains that these are chatbots responding the way they are programmed to,
[/snip]
It's worth taking issue with this.
Eliza was programmed - pure and simple. The code crudely analyses the input dialog and constructs responses in a way which sometimes gives the impression of making sense, and is sometimes nonsense.
But ChatGPT was not programmed. Rather an architecture and a learning algorithm were programmed for the core LLM - GPT4. This program knew nothing. It had a large dictionary of relatively common words (each represented as a single token) plus room for more tokens to deal with words it had never seen, but NO DEFINITIONS. Indeed no semantic understanding at all - it viewed a full stop "." and the word "transsubstantiation" as of identical status, two meaningless tokens. Here is the complete list of GPT-4 tokens. eg "GNU" is token 4349. (Oddly, it's slightly more than you can number with 2 bytes!)
The resulting LLM was then given vast amounts of data (as sequences of those tokens, like any text information) to learn languages, semantics and reasoning from.
A few months of training later using 9 figures worth of hardware and electricity, this LLM has one capability - to predict the next token in stream of tokens. When running it generates a set of probabilities for all possible tokens in that list linked above and generates a token according to a temperature (absolute zero means output the highest probability token, very hot means output one of many of the more probable tokens).
This is sufficient to generate (mostly) meaningful and often (but not always) correct output, but important refinements are added. One is to improve the quality of output by doing fine tuning based on user feedback. This involves generating two outputs (with temperature above zero, so they are different) and letting people pick which is better. The feedback is used to tweak the output probability.
With the LLM being able to emulate understanding, this is harnessed by wrapping the communication in guidance given to the LLM in English, doing what humans did manually in last years AIs in order to get better output.
All that is a very long way from AIs being "programmed". The only things that were programmed were the ability to learn (in a computer language) and, loosely, the meta-instructions that make it more thoughtful in its responses (where it is "programmed" in English!). If the latter is programming, you too are programming an AI when you tell it to answer a question . Really, this is best thought of as an interaction, rather than programming. It is not deterministic and can't realistically be described in the syntax of a computer language.
I wonder if something like this has ever been done before, as an actual experiment. I’m not quite sure what one can deduce from this conversation, but I think it has some value and is a new approach. It could be used to compare chatbots, but that’s just a wild idea I came up with. I’d love to hear your thoughts on this, Elroch, edutiiseme, Lola and anyone else who wishes to share.