Personally, better way to phrase might be "Does anybody you've actually met, visually viewed, use OpenClaw? Can you verify them using the software nearby?"
In a few years, it's become so easy to falsify articles, falsify comments, falsify images, difficult to really even trust responses online anyways. As far back as 2016, Microsoft already had bots deployed online that could respond 96,000 times [1] in 16 hours all over social media. Remember Tay? [1][2]
When you have insane amounts of capital and your gpu and talent needs are more or less met, there is a capital relief valve known as growth hacking. It only works if the consumer isn’t aware they’re being hacked.
No. They want you to believe in the hype and that LLMs are the death of programmers and limitless. OpenClaw and other such agents are sold as a tool that "can do anything" but behind the scenes, the implication is still that big LLM is driving it. So both are conflated.
In a few years, it's become so easy to falsify articles, falsify comments, falsify images, difficult to really even trust responses online anyways. As far back as 2016, Microsoft already had bots deployed online that could respond 96,000 times [1] in 16 hours all over social media. Remember Tay? [1][2]
[1] https://www.theverge.com/2016/3/24/11297050/tay-microsoft-ch...
[2] https://en.wikipedia.org/wiki/Tay_(chatbot)
Even official government responses.
The British Royal family went to falsification immediately. [3] Note child's broken fingers bent sideways (lower left, didn't even get circled)
[3] https://inews.co.uk/news/signs-princess-kate-royal-family-ph...
The White House is posting altered arrest images of people. [4]
[4] https://www.theguardian.com/us-news/2026/jan/22/white-house-...
Can't trust this stuff much anymore. Obvious caveat with this post.