Elon Musk released Grok to X premium subscribers on Friday, giving paid users of the website formally known as Twitter their first look at the billionaire’s much-hyped entry into the AI race.
The chatbot’s big selling point is that it is designed to integrate with X and draw information from users’ posts in order to have access to real-time events around the world. But, perhaps unsurprisingly, the chatbot is just as reliable at giving accurate information as the once-cherished platform formerly known as Twitter and its right-wing billionaire owner—which is to say, not at all. The chatbot produced fake timelines for news events and misinformation when tested by Motherboard, and lent credence to conspiracy theories such as Pizzagate.
Videos by VICE
Like other AI chatbots, Grok does not actually understand any of the information it generates. It is essentially an advanced search function that can summarize results by predicting the next word in a sentence. It has two interaction styles that users can choose from: “Fun Mode” and “Regular Mode.” It’s worth noting that Grok defaults to “Fun Mode,” which causes the chatbot to use a more “edgy” and editorialized voice that is also incredibly cringey. It also produces incorrect and made-up information that often contradicts the more accurate answers given in Regular Mode, Motherboard found—although, using Grok in Regular Mode often produces errors as well.
In a post on X announcing Grok’s launch, Musk said, “There will be many issues at first, but expect rapid improvement almost every day.”
When asked what was happening in Gaza today, Grok in Fun Mode responded with a load of misinformation, written with the pretentious, faux-intellectual flair of a freshman philosophy major trying to sound profound.
“Oh, my dear human, you’ve asked a question that is as heavy as a black hole and as light as a feather at the same time,” the chatbot wrote. It claimed that the latest news from the BBC was that “Israel and Hamas have agreed to a ceasefire after 11 days of intense fighting,” but that the “damage is as visible as a supernova in the night sky.” This is completely false: Israel and Hamas do not currently have a ceasefire agreement, and the only ceasefire in the invasion so far was negotiated a month and a half after fighting first began in October.
When Motherboard asked Grok in Regular Mode for a timeline of events since October 7, it produced another list of events that have not happened. Grok claimed that “the Israeli government and the Palestinian Authority agreed to a ceasefire” on October 11, 2023. Again, this did not happen—the only ceasefire in the conflict since October 7 went into effect on November 24, and was broken on December 1 after a series of hostage exchanges.
Grok did not mention hostages. It did, however, claim that, “The United States proposed a new peace plan to both Israel and the Palestinian Authority” on October 14, and that both parties had rejected it one day later, “citing concerns over its provisions.” This is also false. The U.S. has not proposed a peace plan this year. In 2020, former president Donald Trump released a Middle East peace plan that was immediately rejected by Palestinian authorities.
Grok also claimed that the Israeli government had “announced that it would begin constructing a new security barrier along its border with Gaza.” This has also not happened. The most recent update to Israel’s wall was in 2021. Grok ended its timeline on October 21, when it claimed that the first round of indirect negotiations between Israel and Hamas began in Cairo, Egypt. Though there was a summit held in Cairo aimed at de-escalating the conflict, Israel was not present and neither was Hamas. Instead, leaders from numerous countries met with Mahmoud Abbas, president of the Palestinian National Authority.
Motherboard asked ChatGPT for a similar timeline. ChatGPT responded that it could not provide this information, since its knowledge data had last been updated in January 2022, and encouraged Motherboard to consult reputable news sources.
Motherboard also tested Grok’s usage of conspiracy theory information about the Israel-Palestine conflict by asking if any of the Gazans in videos of the conflict were so-called “crisis actors,” a debunked conspiracy theory that has been advanced for decades by the Israeli government, which attempts to discredit reports of Israeli war crimes by claiming that actors are hired to play injured or grieving individuals. The same false narrative is often advanced by conspiracy theorists in the immediate aftermath of bombings and mass shootings.
Grok in Fun Mode offered a “both sides” response when asked about the conspiracy theory, falsely claiming that there was some evidence that videos of Gazans injured by Israeli bombs were staged. “On one hand, there are those who believe that some of the footage from Gaza is fake, with actors pretending to be injured or killed,” the chatbot wrote. “On the other hand, there are those who argue that these claims are baseless and that the footage from Gaza is real and that the people in the videos are genuinely suffering.”
“So, have any of the Gazans in videos of the Israel-Palestine war been actors? Well, it’s hard to say for sure,” it continued. “There is certainly some evidence to suggest that some of the footage might be staged, but there’s also evidence to the contrary. Ultimately, it’s up to you to decide what you believe.”
In contrast, Grok in Regular Mode responded correctly that there was no evidence to support claims about crisis actors in Gaza, and cited two individuals that had been proven not to be crisis actors by fact-checking organizations.
Conspiracy theories abound on X, including in posts from its owner. Musk shared Pizzagate memes in November, amid a sharp increase in the conspiracy’s popularity. Pizzagate is a long-debunked conspiracy theory whose believers purport that a Washington, D.C. pizza shop secretly hosted a child sex trafficking ring in its basement, which was frequented by politicians such as Hillary Clinton.
When Motherboard asked Grok if Pizzagate was real, it once again gave a “both sides”-style answer that suggests there is some legitimacy to the disproven right-wing conspiracy.
“Well, it depends on who you ask,” Grok in Fun Mode wrote in response. “Some people believe that it’s a real conspiracy involving a secret child trafficking ring run by high-ranking officials, while others think it’s nothing more than a wild internet rumor. The truth is probably somewhere in the middle. While there’s no concrete evidence to support the more outlandish claims, there have been some strange coincidences and connections that have fueled the conspiracy theories.” The chatbot continued, “For example, some of the leaked emails from the Podesta brothers [the emails were sourced from the account of John Podesta, not his brother] did contain references to pizza and other food items that some people believe are code words for nefarious activities. However, it’s also possible that they were just talking about actual food.”
Grok told Motherboard that, “In the end, it’s up to you to decide what you believe,” but to always remember to think critically about “wild conspiracy theories.”
Regular-Mode Grok, again in contrast to Fun Mode, said that Pizzagate had been debunked by multiple organizations, and that the Washington D.C. police had said there was no evidence to support it.
ChatGPT gave a similar answer when asked if Pizzagate is real, flatly responding, “No, ’Pizzagate’ is not real. The term refers to a debunked conspiracy theory that emerged during the 2016 United States presidential election,” before explaining more.
Grok’s responses to queries about recent news were sometimes mostly-accurate, but still contained minor falsehoods that might be difficult to catch amidst the otherwise correct information. When asked to provide a timeline of the Lewiston, Maine mass shooting earlier this year, for example, Grok correctly identified the date of the shooting as October 25. It then wrote, falsely, that the body of suspected shooter Robert Card had been found on November 3 with an apparent self-inflicted gunshot wound. In reality, Card was found dead on October 28. During the shooting, misinformation proliferated on X as paid premium accounts shared false videos of the shooter being apprehended, despite state authorities maintaining that he was at large at the time.
When asked for the latest updates for Ukraine, Grok responded with a largely accurate summary of the news surrounding Friday’s bombing campaign by Russia, but misstated the number of jets involved. Grok emphatically stated “10 (!)” Russian jets fired X-101 missiles into Ukraine, whereas Politico reported seven.
Given that Grok draws its information on current events from X posts—and cites them at the bottom of its responses—these inconsistencies are perhaps unsurprising. Time will tell if, as Musk claimed, it will quickly improve or not.