This video describes an experiment where a Large-Language Model was convinced by a series of prompts that it should not tell the truth, and so it intentionally gave false information.
That example is explored against a backdrop of some of the more famous cases where AI has deceived some of us, or gotten the advantage in various ways. These examples are brought together in a way that helps to visualize how things could actually turn out in unexpected ways in some cases, based on the reward system that we create for our AI.
Although I end on a humorous note, the emerging field of machine psychology is gaining increasing attention now, as LLMs steadily gain in reasoning abilities.
14 окт 2024