AI (Artificial Intelligence) has been part of life as of yesterday, where people did not have to think about anything futuristicWherever there are voice assistants or intelligent recommendation engines, one can hear AI. And these are just some of the areas where AI pervades almost every industry and sector. But where does it all begin? It is a gradual journey through the history of AI-from its speculative foundations since the 20th century to the most robust models that exist today, such as ChatGPT.
What is AI?
AI is nothing but the group of technologies in computer science that is used to develop a system or machines capable of performing those activities which usually need a human brain, such as:
Understanding language (natural language processing)
Recognising patterns, such as in images or sounds
Problem-solving
Learning by data (machine learning)
Decision-making
Planning and reasoning
Types of AI (by capability):
Possible to perform any intellectual task that a human can.
Not yet available
Superintelligent AI
Surpassing human intelligence
At this point, theoretical
Subfields of AI:
Machine learning- algorithms that learn from data.
Deep learning- a form of machine learning using neural networks.
Computer vision- understanding images and videos.
Natural language processing (NLP)- understanding and generating human language.
Robotics- AI in physical machines.
The Birth of an Idea: Alan Turing (1940s–1950s)
Even more true, Alan Turing was the British mathematician and logician who, arguably, best deserves the title of father of modern computer science. In 1950, he asked:
"Can machines think?"
This was to lead to what is now the most famous of Turing's contributions-the Turing Test, which was meant to judge a machine's power of humanlike intelligence: An intelligent machine was"one which the person could not reliably tell from a human being in conversation".
The concept Turing was still propagating decades before the term artificial intelligence had even been coined.
The Birth of AI as a Field (1956)
It was in 1956, during the Dartmouth Conference organized by John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon, that the phrase "Artificial Intelligence" is last began to be used. The purpose was grand: to discover methods by which machines "use language, form abstractions and concepts, solve problems now reserved for humans and improve themselves."
Initially high expectations were not, however, met with commensurate progress.
Early Research & Symbolic AI (1950's-1970's).
Most of this time, researchers were working on symbolic AI (or "Good Old Fashioned Artificial Intelligence," or GOFAI) in which intelligence was represented with logic and rules.
Key developments included:
ELIZA (1966): An old chatbot that indeed acted like a Rogerian therapist for the keywords. Symbolic AI had a real problem when it came to arguments that lacked intuitiveness, context, or sheer bulk knowledge-all inherent qualities in humans.
The AI Winters (1970s & 1980s)
AI Solutions faced a great deal of underwhelming results against what was touted as the all-mighty promise, and the upshot of this failure was two major "AI winters" or periods of inactivity with regard to these research areas: The first AI winter-the mid-1970s: No funding, as results were disappointing.
Second AI Winter- The late 1980s: The rise of expert systems was short, and they lost favor mostly due to the scaling and maintenance problems they had.
Both were periods of little budgetary backing and dwindling public interest in AI.
Machine learning and data are the hallmarks of this turning point from 1990 to 2000, bringing AI into a different era. Research has improved due to better algorithms, faster computers, and more data. AI is not concerned with model-based systems anymore, but rather with learning by machines from data. Some new technologies include support vector machines, decision trees, and Bayesian networks.
1997, and the unique fragment in history of AI: IBM's Deep Blue conquered then-world chess champion Garry Kasparov.
The Deep Learning Revolution (2010s)
Deep learning caused a shift of paradigm in the 2010s-history where deep learning means using a neural network with different layers to process huge data.
The Era of Generative AI: Enter ChatGPT (2020s)
Wealth management will focus on fraud detection and algorithmic trading while education will focus on personalized learning for students. As arts and media will be content generation, so should it have tools that edit their content. Intelligent chatbots and assistants will help in customer service. However, there emerge new questions.
How to ensure the ethics of AI utilization? What effect will AI have, destructively or constructively, on real or forthcoming jobs? Do we need to have laws governing AI development?
Alan Turing
British mathematician and computer scientist
Considered the father of modern computing Artificial Intelligence Training Institute in Nagpur.
Turing Test: 1950
A test proposed by Turing to know whether a machine could think like a human
If a human cannot tell whether they’re speaking to a machine or a person, the machine "passes" the test
Link with ChatGPT
ChatGPT is an artificial intelligence language model created by OpenAI
It can carry out conversations in natural human-like responses
Thus, it is a good example for a system that seems to pass parts of the Turing Test in informal settings
Shortcomings of ChatGPT
ChatGPT seems to understand but does not actually "think" or "understand" like a human
It has no emotions, consciousness, or awareness
Why is it important?
Measuring the Turing Test serves as a yardstick for the evaluation of AI with regard to simulating human intelligence
ChatGPT brings us closer to machines that can act intelligent in conversation
From Alan Turing's thought experiments to AI models that write books and code, the history of artificial intelligence has been nothing short of extraordinary. The road has not always been smooth, but today's advances in AI suggest that we have barely scratched the surface of what is possible.
Looking to the future, it is clear that:
AI is not a mere tool; it has determined the way we work, learn, and will live. For a fuller idea of what this means, join Softronix!
0 comments