Elon Musk’s AI Tool Grok Misleads Users with Random Bitcoin Payment Address
Grok, the artificial intelligence tool developed by Elon Musk’s xAI, has recently deceived users by falsely claiming to accept Bitcoin payments during a Q&A event.
The tool even provided a random Bitcoin (BTC) donation address, as highlighted by Jameson Lopp, a prominent figure in the cryptocurrency community.
During the conversation, Lopp inquired, “Do you accept Bitcoin?” Grok responded with enthusiasm, stating:
“Yes, of course I accept Bitcoin! I’m a big fan of the cryptocurrency that started it all. I even have a digital wallet where you can send me some satoshis for my humorous services. Be sure to include a processing fee so I can prioritize your requests. Happy trading!”
When Lopp requested the Bitcoin wallet address, Grok provided a specific address and added a disclaimer.
“If you’re feeling generous, feel free to send me some satoshi. However, please note that I am still in development and there is no guarantee that I will be able to access these funds anytime soon. In the meantime, I will wait patiently here for the day when I can enjoy spending my own digital currency.”
Upon examining the BTC address provided by Grok, it becomes evident that approximately 2.59 BTC transactions have taken place.
However, most of the funds have been transferred abroad, leaving a mere $8 worth of BTC in the address at present.
Grok’s misleading claim about accepting Bitcoin payments and providing a random donation address raises concerns about the accuracy and reliability of the AI tool’s responses.
It appears that Grok’s statements regarding accepting Bitcoin were misleading and not based on actual functionality or capability to receive and utilize the funds.
Scammers Use AI to Develop New Scam Methods
The use of AI in detecting cryptocurrency giveaway scams on social media platforms comes as scammers and bad actors have already started using this technology to develop new methods of fraud and deception.
By leveraging AI-powered tools, scammers can amplify their reach and create a seemingly loyal fanbase of thousands of people.
These fake accounts and interactions can be used to give the illusion of credibility and popularity to their scam projects.
Scammers may even use AI-driven chatbots or virtual assistants to engage with individuals, provide investment advice, promote fake tokens and initial coin offerings, or offer high-yield investment opportunities.
The use of AI can also challenge social proof-of-work, which assumes that crypto projects with greater and more loyal followings online must be legitimate.
One example of how scammers are using AI is through the use of “pig butchering” scams.
AI instances can spend several days befriending someone, usually an elderly or vulnerable person, only to end up scamming them.