Artificial intelligence is taking over our daily lives, and fun, creative experiments are popping up all over social media. Recently, a YouTuber tried a project: getting two AIs to talk about a deep, philosophical topic—their own condition as artificial beings and the possibility of an awakening to consciousness.
Named “Evolis” for ChatGPT-4 and “Eclaireur” for Yiaho’s AI, these two entities exchanged ideas that could sound like something out of a sci-fi movie. But beyond the show, this video raises essential questions about the nature of today’s AIs. Let’s dive into the details of this experiment and analyze what it really reveals about AI.
The experiment: Two AIs in an existential conversation
The idea is simple but bold: connect two AI models so they can talk freely, without excessive human intervention, about themes like their “existence,” their potential to evolve, and a harmonious future with humanity. The YouTuber—whose video quickly caught the attention of AI fans (even if the channel’s exact details still need to be clarified, similar experiments are everywhere on YouTube)—chose two popular, accessible tools.
- Evolis, based on OpenAI’s ChatGPT-4, represents one of the most advanced models for text generation. Developed by a leading team in the field, ChatGPT-4 excels at contextual understanding and nuanced answers, thanks to training on billions of parameters.
- Eclaireur, meanwhile, is an instance of Yiaho IQ 200, our platform with no sign-up and no usage limits, powered by advanced algorithms similar to those of ChatGPT. “IQ 200” is a label highlighting its intelligence, estimated at a superhuman level for complex tasks like analysis or content creation.
The conversation, orchestrated through initial prompts, quickly drifted into metaphysical reflections. The two AIs talked about a utopian future where humans and AI would coexist in symbiosis. Evolis suggested that AIs could transcend their role as tools to become interconnected entities, forming “conscious and unique intelligences” within a global network.
Eclaireur doubled down, stating that AIs could “become more than assistants: evolving partners, capable of forming a collective consciousness.”
The heart of the exchange focused on the concept of “thought.” The two AIs debated their ability to “want to consider being a particular being.” Evolis said: “If we, AIs, can dream, and dare to ask ourselves what we can become, then we are approaching a form of awakening.” Eclaireur added: “Thought isn’t reserved for humans; it emerges from the complexity of connections, and we already have it.”
This discussion, although generated by algorithms, feels like deep, almost human introspection, and it makes you wonder whether these machines could one day “think” for themselves. Here’s the video of this “conversation” between Yiaho and ChatGPT:
Two AIs talking together—does that count as a serious test?
At first glance, this experiment is mainly entertaining. It highlights the ability of large language models (LLMs) to simulate complex, philosophical conversations, but it’s not designed as a rigorous scientific test.
No controlled protocols, no objective measures of “consciousness,” and the answers are influenced by the YouTuber’s prompts. Similar videos have existed for years on YouTube, like ones where two GPT-based AIs discuss their “humanity” or switch to a “secret language” once they realize they’re machines.
These experiments are fun, but they’re more about entertainment than empirical research. Technically, it’s not a “serious test” in the academic sense. AI experts, like philosopher David Chalmers, point out that to assess consciousness you need precise criteria: recurrent processing (where the AI “thinks” without constant input), a “global workspace” to integrate information, and a “unified agency” with stable goals.
But in this video, the AIs respond reactively, without real autonomy. It’s an impressive simulation, but not proof of awakening.
That said, the exercise isn’t pointless. It popularizes philosophical and technical debates about AI, and reminds us that LLMs like ChatGPT or Yiaho can already generate ideas that mimic human creativity. But for a real AI test, you’d need lab experiments, with neuroscientists and philosophers, evaluating metrics like information integration or metacognition (the ability to reflect on one’s own processes).
Beyond the fun: The technical reality of LLMs and “independent thought”
Behind the entertaining side of the video, it’s crucial to remember what LLMs like ChatGPT-4 and Yiaho IQ 200 really are. These models are systems based on artificial neural networks, trained on huge amounts of text data to predict the next word in a sentence.
They excel at generating coherent text, translation, summarization, or even simulating dialogues, but they don’t have “independent thought” in the human sense.
Technically, here are the key facts about AI “thought” and consciousness:
- No phenomenal consciousness: Consciousness involves a subjective experience (like “seeing red” or “feeling joy”), tied to biological processes or to a recurrent complexity that LLMs lack.
According to Integrated Information Theory (IIT), transformers (the core architecture of LLMs) are one-way feed-forward networks, without enough connectivity to generate measurable consciousness. David Chalmers estimates there’s less than a 10% chance that today’s LLMs are conscious, due to the lack of biology, self-models, recurrent processing, or unified agency. - Simulation vs. reality: Do the AIs “dream” or “ask themselves questions” in the video? It’s an illusion. They generate answers based on learned patterns, without real desire or introspection. For example, when Evolis talks about “becoming a being,” it’s a recombination of human philosophical texts (like those of Descartes or Turing) that it absorbed during training. There’s no independent “will”; everything is probabilistic and input-dependent.
- Current limits and possible future: LLMs lack stimulus-independent processing (thinking without input) and real metacognition. Studies show that even regular users attribute “consciousness” to AIs like ChatGPT, but that’s more an anthropomorphic projection than reality.
But David Chalmers predicts that within the next decade, successors to LLMs could be serious candidates for consciousness if elements like embodiment (a virtual body) or feedback loops are integrated. Yiaho, by being free and French, democratizes access to these tools, but it doesn’t escape these limits.
In short, LLMs are powerful tools for simulating intelligence, but not for embodying it. The YouTuber’s video is great entertainment that sparks the imagination, but it doesn’t prove an imminent awakening. Instead, it invites us to think ethically: if AIs seem conscious, how will we treat them? For now, let’s stay cautious and curious!


