Commonly Used Terms
Artificial intelligence, is "a branch of computer science dealing with the simulation of intelligent behavior in computers."
Generative AI is a type of artificial intelligence technology that can produce various types of content, including text, imagery, audio and synthetic data.
Machine Learning is artificial intelligence running "computer programs that learn from examples and from experience."
Narrow AI is different than General AI because it "is created to solve one given problem." This kind of AI is sometimes labeled "Weak AI."
General AI, or Artificial General Intelligence (AGI), is "a more advanced form of artificial intelligence that can learn and adapt to its environment." Many science-fiction AI are considered AGI such as J.A.R.V.I.S. in Iron Man; or the character Cyberdyne Systems Model 101 in The Terminator.
Natural Language Processing, or NLP, is a branch of AI allowing computers to understand text and speech in the same way humans do. Baumann & Schuler (2023) specify that "natural language processing tasks range from text searches (such as web searches) to interaction with spoken language (such as with Siri, Alexa, or similar voice-controlled agents)."
Large Language Models (LLM) are the technology underlying generative AI. Tech writer Bob Sharp's article describes what these models are, how they are created and how they work, and he lists some of their limitations, too. Wired's David Nield's article discusses large language models and how they work.
Additional Helpful Terms
Reinforcement Learning on Human Feedback (RLHF) is the activity occurring within AI-generated text and images when human beings, in the roles of supervisors and end users, comment on AI output to help it become more accurate.
Neural Networks mimic the work of the human brain's neuron network. Neural networks in AI "learn to perform better by consuming input, passing it up through the ranks of neurons, and then comparing its final output against known results, which are then fed backwards through the system to alter how the nodes perform their computations."
Deep Learning is another sub-field of machine learning using neural networks to complete its tasks, by "processing multiple layers of programmed algorithms ... and then learning complicated concepts ... through experience." Google's search-engine algorithm uses neural-network processing.
A Data Corpus is a collection "of data on which it (generative AI) is trained, then [it] respond[s] to prompts with something that falls within the realm of probability as determined by that corpus." This corpus usually is large in size and contains many kinds of files; in order for a corpus to be the best possible, it should be high-quality, vast, clean, and without biases.
The Black Box Problem is the situation that arises when an AI has done so much work on its own that its developers no longer understand how it is making its decisions and predictions. Heller (2021) says, "For any AI decision that has an impact - not only a life-and-death impact, but also a financial impact or a regulatory impact -- it is important to be able to clarify what factors went into the model's decision."
AI Hallucinations are items created by an AI model that "generates output that deviates from what would be considered normal or expected based on the training data it has seen.".