DATA SCIENCE AND MACHINE LEARNING
DATA SCIENCE AND MACHINE LEARNING

what is intelligence ?/ what is artificial intelligence ? / artificial intelligence/ Machine learning / colzlife/ Artificial Intelligence Notes

Scientists have proposed two major “consensus” definitions of intelligence:

  • from Mainstream Science on Intelligence (1994);

A very general mental capability that, among other things, involves the ability to reason, plan, solve problems, think abstractly, comprehend complex ideas, learn quickly and learn from experience. It is not merely book learning, a narrow academic skill, or test-taking smarts. Rather, it reflects a broader and deeper capability for comprehending our surroundings- making sense” of things, or “figuring out” what to do.

  • from Intelligence: Knowns and Unknowns (1995);

Individuals differ from one another in their ability to understand complex ideas, to adapt effectively to the environment, to learn from experience, to engage in various forms of reasoning, [and] to overcome obstacles by taking thought. Although these individual differences can be substantial, they are never entirely consistent: a given person’s intellectual performance will vary on different occasions, in different domains, as judged by different criteria. Concepts of “intelligence” are attempts to clarify and organize this complex set of phenomena.

Thus, intelligence is:

  • the capability to reason
  • to understand
  • to create
  • to Learn from experience
  • to plan and execute complex tasks

What is Artificial Intelligence?

“Giving machines ability to perform tasks normally associated with human

intelligence.”

AI is intelligence of machines and branch of computer science that aims to create it. Also AI consists of design of intelligent agents, which is a program that perceives its environment and takes action that maximizes its chance of success. In addition AI it comes issues like deduction, reasoning, problem solving, knowledge representation, planning, learning, natural language processing, perceptron, etc.

“Artificial Intelligence is the part of computer science concerned with designing intelligence computer systems, that is, systems that exhibit the characteristics we associate with intelligence in human behavior.”

Different definitions of AI are given by different books/writers. These definitions can be divided into two dimensions.

Systems that think like humans Systems that think rationally
“The exciting new effort to make computers think…..machine with minds, in the full and literal sense.” (Haugeland, 1985)   “[The automaton of] activities that we associate with human thinking, activities such as decision-making, problem solving, learning…..” (Bellman, 1978) “The study of mental faculties through the use of computational models.” (Charniak and McDermott, 1985)   “The study of the computations that make it possible to perceive, reason, and act.” (Winston, 1992)
Systems that act like humans Systems that act rationally
“ The art of creating machines that perform functions that require intelligence when performed by people.” (Kurzweil, 1990)   “The study of how to make computer do things at which, at the moment, people are better.” (Rich and Knight, 1991) “Computational Intelligence is the study of the design of intelligent agents.” (Poole et al., 1998)     “AI… is concerned with intelligent behavior in artifacts.” (Nilsson, 1998)

Top dimension is concerned with thought processes and reasoning, where as bottom dimension addresses the behavior.

The definition on the left measures the success in terms of fidelity of human performance, whereas definitions on the right measure an ideal concept of intelligence, which is called rationality.

Human-centered approaches must be an empirical science, involving hypothesis and experimental confirmation. A rationalist approach involves a combination of mathematics and engineering.

Acting Humanly: The Turing Test Approach

The Turing test, proposed by Alan Turing (1950) was designed to convince the people that whether a particular machine can think or not. He suggested a test based on indistinguishably from undeniably intelligent entities- human beings. The test involves an interrogator who interacts with one human and one machine. Within a given time the interrogator has to find out which of the two the human is, and which one the machine.

The computer passes the test if a human interrogator after posing some written questions, cannot tell whether the written response come from human or not.

To pass a Turing test, a computer must have following capabilities:

  • Natural Language Processing: Must be able to communicate successfully in English
  • Knowledge representation: To store what it knows and hears.
  • Automated reasoning: Answer the Questions based on the stored information.
  • Machine learning: Must be able to adapt in new circumstances.

Turing test avoid the physical interaction with human interrogator. Physical simulation of human beings is not necessary for testing the intelligence.

The total Turing test includes video signals and manipulation capability so that the interrogator can test the subject’s perceptual abilities and object manipulation ability. To pass the total Turing test computer must have following additional capabilities:

  • Computer Vision: To perceive objects
  • Robotics: To manipulate objects and move

Thinking Humanly: Cognitive modeling approach

If we are going to say that a given program thinks like a human, we must have some way of determining how humans think. We need to get inside the actual workings of human minds. There are two ways to do this:

  • through introspection: catch our thoughts while they go by
  • through psychological experiments.

Once we have precise theory of mind, it is possible to express the theory as a computer program.

The field of cognitive science brings together computer models from AI and experimental techniques from psychology to try to construct precise and testable theories of the workings of the human mind.

Think rationally: The laws of thought approach

Aristotal was one of the first who attempt to codify the right thinking that is irrefutable reasoning process. He gave Syllogisms that always yielded correct conclusion when correct premises are given.

For example:

Ram is a man

All men are mortal

=> Ram is mortal

These law of thought were supposed to govern the operation of mind: This study initiated the field of logic. The logicist tradition in AI hopes to create intelligent systems using logic programming. However there are two obstacles to this approach. First, It is not easy to take informal knowledge and state in the formal terms required by logical notation, particularly when knowledge is not 100% certain. Second, solving problem principally is different from doing it in practice. Even problems with certain dozens of fact may exhaust the computational resources of any computer unless it has some guidance as which reasoning step to try first.

Acting Rationally: The rational Agent approach:

Agent is something that acts.

Computer agent is expected to have following attributes:

  • Autonomous control
  • Perceiving their environment
  • Persisting over a prolonged period of time
  • Adapting to change
  • And capable of taking on another’s goal

Rational behavior: doing the right thing.

The right thing: that which is expected to maximize goal achievement, given the available information.

Rational Agent is one that acts so as to achieve the best outcome or, when there is uncertainty, the best expected outcome.

In the “laws of thought” approach to AI, the emphasis was given to correct inferences. Making correct inferences is sometimes part of being a rational agent, because one way to act rationally is to reason logically to the conclusion and act on that conclusion. On the other hand, there are also some ways of acting rationally that cannot be said to involve inference. For Example, recoiling from a hot stove is a reflex action that usually more successful than a slower action taken after careful deliberation.

Advantages:

  • It is more general than laws of thought approach, because correct inference is just one of several mechanisms for achieving rationality.
  • It is more amenable to scientific development than are approaches based on human behavior or human thought because the standard of rationality is clearly defined and completely general.

Characteristics of A.I. Programs

  • Symbolic Reasoning: reasoning about objects represented by symbols, and their properties and relationships, not just numerical calculations.
  • Knowledge: General principles are stored in the program and used for reasoning about novel situations.
  • Search: a “weak method” for finding a solution to a problem when no direct method exists. Problem: combinatoric explosion of possibilities.
  • Flexible Control: Direction of processing can be changed by changing facts in the environment.

Foundations of AI:

Philosophy:

Logic, reasoning, mind as a physical system, foundations of learning, language and rationality.

  • Where does knowledge come from?
  • How does knowledge lead to action?
  • How does mental mind arise from physical brain?
  • Can formal rules be used to draw valid conclusions?

Mathematics:

Formal representation and proof algorithms, computation, undecidability, intractability, probability.

  • What are the formal rules to draw the valid conclusions?
  • What can be computed?
  • How do we reason with uncertain information?

Psychology:

Adaptation, phenomena of perception and motor control.

  • How humans and animals think and act?

Economics:

Formal theory of rational decisions, game theory, operation research.

  • How should we make decisions so as to maximize payoff?
  • How should we do this when others may not go along?
  • How should we do this when the payoff may be far in future?

Linguistics:

Knowledge representation, grammar

  • How does language relate to thought?

Neuroscience:

Physical substrate for mental activities

  • How do brains process information?

Control theory:

Homeostatic systems, stability, optimal agent design

  • How can artifacts operate under their own control?

Brief history of AI

– 1943: Warren Mc Culloch and Walter Pitts: a model of artificial boolean neurons to perform computations.

  • First steps toward connectionist computation and learning (Hebbian learning).
  • Marvin Minsky and Dann Edmonds (1951) constructed the first neural network computer
  • 1950: Alan Turing’s “Computing Machinery and Intelligence”
    • First complete vision of AI.

The birth of AI (1956):

  • Dartmouth Workshop bringing together top minds on automata theory, neural nets and the study of intelligence.
    • Allen Newell and Herbert Simon: The logic theorist (first nonnumeric thinking program used for theorem proving)
    • For the next 20 years the field was dominated by these participants.

Great expectations (1952-1969):

– Newell and Simon introduced the General Problem Solver.

  • Imitation of human problem-solving

– Arthur Samuel (1952-) investigated game playing (checkers ) with great success.

– John McCarthy(1958-) :

  • Inventor of Lisp (second-oldest high-level language)
    • Logic oriented, Advice Taker (separation between knowledge and reasoning)

– Marvin Minsky (1958 -)

  • Introduction of microworlds that appear to require intelligence to solve: e.g. blocks- world.
    • Anti-logic orientation, society of the mind.

Collapse in AI research (1966 – 1973):

  • Progress was slower than expected.
    • Unrealistic predictions.
  • Some systems lacked scalability.
    • Combinatorial explosion in search.
  • Fundamental limitations on techniques and representations.
    • Minsky and Papert (1969) Perceptrons.

AI revival through knowledge-based systems (1969-1970):

  • General-purpose vs. domain specific

– E.g. the DENDRAL project (Buchanan et al. 1969) First successful knowledge intensive system.

  • Expert systems

– MYCIN to diagnose blood infections (Feigenbaum et al.)

– Introduction of uncertainty in reasoning.

  • Increase in knowledge representation research.

– Logic, frames, semantic nets, …

AI becomes an industry (1980 – present):

  • R1 at DEC (McDermott, 1982)
  • Fifth generation project in Japan (1981)
  • American response …

Puts an end to the AI winter.

Connectionist revival (1986 – present): (Return of Neural Network):

  • Parallel distributed processing (RumelHart and McClelland, 1986); backprop.

AI becomes a science (1987 – present):

  • In speech recognition: hidden markov models
  • In neural networks
  • In uncertain reasoning and expert systems: Bayesian network formalism

The emergence of intelligent agents (1995 – present):

  • The whole agent problem:

“How does an agent act/behave embedded in real environments with continuous sensory inputs”

Applications of AI: (Describe these application areas yourself)

  • Autonomous planning and scheduling
  • Game playing
  • Autonomous Control
  • Expert Systems
  • Logistics Planning
  • Robotics
  • Language understanding and problem solving
  • Speech Recognition
  • Computer Vision

Knowledge:

Knowledge is a theoretical or practical understanding of a subject or a domain. Knowledge is also the sum of what is currently known.

Knowledge is “the sum of what is known: the body of truth, information, and principles acquired by mankind.” Or, “Knowledge is what I know, Information is what we know.”

There are many other definitions such as:

  • Knowledge is “information combined with experience, context, interpretation, and reflection. It is a high-value form of information that is ready to apply to decisions and actions.” (T. Davenport et al., 1998)
  • Knowledge is “human expertise stored in a person’s mind, gained through experience, and interaction with the person’s environment.” (Sunasee and Sewery, 2002)
  • Knowledge is “information evaluated and organized by the human mind so that it can be used purposefully, e.g., conclusions or explanations.” (Rousa, 2002)

Knowledge consists of information that has been:

  • interpreted,
  • categorised,
  • applied, experienced and revised.

In general, knowledge is more than just data, it consist of: facts, ideas, beliefs, heuristics, associations, rules, abstractions, relationships, customs.

Research literature classifies knowledge as follows:

Classification-based Knowledge » Ability to classify information
Decision-oriented Knowledge » Choosing the best option
Descriptive knowledge » State of some world (heuristic)
Procedural knowledge » How to do something
Reasoning knowledge » What conclusion is valid in what situation?
Assimilative knowledge » What its impact is?

Knowledge is important in AI for making intelligent machines. Key issues confronting the designer of AI system are:

Knowledge acquisition: Gathering the knowledge from the problem domain to solve the AI problem.

Knowledge representation: Expressing the identified knowledge into some knowledge representation language such as propositional logic, predicate logic etc.

Knowledge manipulation: Large volume of knowledge has no meaning until up to it is processed to deduce the hidden aspects of it. Knowledge is manipulated to draw conclusions from knowledgebase.

Importance of Knowledge:

Learning:

It is concerned with design and development of algorithms that allow computers to evolve behaviors based on empirical data such as from sensor data. A major focus of learning is to automatically learn to recognize complex patterns and make intelligent decision based on data.

A complete program is said to learn from experience E with respect to some class of tasks T and performance measure P, if its performance at tasks in T, as measured by P, improves with experience E.

Facebook Comments