Artificial intelligence is being programmed to spot disinformation – by the military.

Defense One reports on a new U.S. Air Force and Special Operations Command (SOCOM) program, using artificial intelligence to counter “information space” pollution – or, you know, software to fight hooey online:

“If you don’t compete in the information space, regardless of how good your operations are, your activities are, you will probably eat a shit sandwich of disinformation or false reporting later on,” Raymond “Tony” Thomas, a former SOCOM chief, said in an interview. “We certainly experienced that at the tactical level. That was the epiphany where we would have good raids, good strikes, etc. and the bad guys would spin it so fast that we would be eating collateral damage claims, etc. So the information space in that very tactical space is key.”

Primer’s neural network technology can scan large amounts of text and extract themes and other information based on the frequency and prominence of words and phrases. It’s the sort of thing that can be very useful if you have a lot of text you want to very quickly summarize in an accurate headline, a capability they demonstrate here. To train their headline-writing neural net, they used a corpus “of millions of publicly available document-title pairs: news articles and headlines” according to their paper on the subject.

It will be a slow process. There’s a big difference between teaching a neural net to summarize a news article or paper and write a headline and teaching it to separate fact from fiction. How do you train software to distinguish trustworthy information from untrustworthy claims? Primer plans to do it the same way one might teach a child: by teaching it to recognize sources of credible information versus less credible information, which takes time, practice, and consistent scrutiny. That will take time and a fair amount of important data input from users and operators, said John Bohannon, the company’s director of science.

Bohannon showed us an example of where the technology is today, in the context of the emerging conflict between Armenia and Azerbaijan. The network can find news, sources and social media posts about the conflict and segment that information into groups, based on who is saying what about a particular event or incident, such as a military strike. This immediately gives the user a sense of what different groups and different governments are claiming. You can also see how those reporting entities have changed the way they’ve discussed the situation in question overtime. Essentially, at present, the network gives you much of the same information that you might get from a newspaper story covering an incident or event.

Eventually, the platform will be able to award a particular claim or news item a sort of accuracy score based on those factors, whether the source is credible, what other perhaps more credible sources are saying, etc. But if you’re not sure of how the network reached its conclusion, you can see the process —and the news sources—it used to make that determination. That’s the ambition, anyway.

It’s a first step to tackling a larger problem, one that will only grow. Sean Gourley, Primer’s founder and CEO, described how the cost of using disinformation is relatively low for authoritarian regimes that aren’t accountable to voters or democratic allies when they lie. Couple that with how quick and easy it is to produce and spread disinformation online and you run into a big problem for which the U.S. and other militaries from democratic countries have little defense. “Information attacks are cheaper to carry out than identity. There’s an asymmetry here. It’s much like an IED explosion. You can put it down very cheaply but the cost of defending against it is very high.” he said. “This is not something that’s going to be solved by humans… You’re bringing a knife to a gunfight if you’re going to bring humans to this problem.”