Feds Want To Use AI In Form Of ChatGPT 

(Republicaninformer.com)- A new generation of artificially intelligent systems can generate text, video, picture, and audio material.  

As part of its ongoing effort to incorporate artificial intelligence into its operations and systems, the CIA is investigating the usage of ChatGPT, a software bot that can write like a person.  

Government agencies, besides the CIA, have also recognized the possible threat of generative AI to national security. During a January presentation, DISA’s chief technology officer, Stephen Wallace, discussed the agency’s investigation into the long-term effects that generative AI may have on its mission and the Department of Defense. TrueNorth is a family of “neuromorphic” computer chips developed by the Air Force Research Laboratory (AFRL) and IBM to run software designed to simulate the neurons in the human brain. 

The new brain-inspired chips used in defense systems could differentiate between military and civilian vehicles in radar-generated pictures while using just twice as much power as traditional computer systems. ChatGPT may be a potential next-generation smart AI producing more actionable intelligence in less time. 

Military experts may see chatbots searching through satellite data, looking for signs that an enemy country is secretly constructing chemical weapons or preparing to launch ballistic missiles. Concerns regarding generative AI extend beyond its potential uses by ourselves to include those of our potential adversaries. A regime like Russia’s, which feeds on state-sponsored propaganda, would find chatbots that can quickly produce compelling written prose disturbingly valuable. The spread of convincing false news posted by bots is unavoidable. 

U.S. intelligence agencies are investigating what their own generative AI technologies can tell us about how bots produce news, distribute it, and can be identified in time to be labeled as false. ChatGPT does not verify its sources, and its final products may include incorrect information. It also fails to provide source information, making it more difficult for human operators to weed out false information. When dealing with intelligence work, we need to trust the data and the results we get. Experts in several fields keep wondering whether or not artificial intelligence will eventually replace them. 

While generative AI has obvious applications that intelligence and military personnel can see, it poses significant risks. It will be up to the respective authorities to determine whether or not the issues are solved and whether or not the benefits justify the costs.