AI Unit 3
AI Unit 3
2. What are the three reasons why FOL fails in medical diagnosis?
Ans: The three reasons why the FOL fails in medical diagnosis are:
(i) Laziness: Too much work to lilst the complete set of antecedents and consequents
needed.
(ii) Theoretical ignorance: Medical science has no complete theory for domain.
(iii) Practical ignorance: Even if we know all the rules uncertainity arises because
some tests cannot be run on the patient's body.
3. What is the tool that is used to deal with degree of belief?
Ans: The tool that is used to deal with the degree of belief is the probability theory which
assigns a numeric degree of belief between 0 and 1.
4. For what the utility theory is useful and in what way it is related to decision
theory? Ans: The utility theory is useful to represent and reason with preferences.
The decision theory is related to utility theory as addition of the probability theory and
the utility theory.
5. What is the fundamental idea of the decision theory?
Ans: The fundamental idea of the decision theory is that an agent is rational if and only if it
chooses the action that yields the highest expected utility.
6. How are the probability over the simple and complex propositions are classified?
Ans: The probability over the simple and complex propositions are classified as
(i) Unconditional or Prior probabilities.
(ii)Conditional or Posterior probabilities.
7. State the axioms of probability.
Ans: The axioms are:
14. What are the two ways in which one can understand the semantics of belief
networks?
Ans: The two ways are
1. Network as a representation of the Joint Probability distribution
-used to know how to construct networks.
2. Encoding of a collection of conditional independence statements.
-designing inference procedures.
18. What makes the nodes of the Bayesian network and how are they connected?
Ans: A set of random variables makes up the nodes of the network. Variables may be discrete
or continuous and a set of directed links or arrows connect pair of nodes.
19. What is Conditional Probability Table (CPT)?
Ans: The table representation of the Bayesian network is called the Conditional probability
table.
20. What is Conditioning Case?
Ans: A Conditioning case is just a possible combination of values for the parent nodes.
21. What are the Semantics of Bayesian Network?
Ans: The Semantics can be represented as the
1. Global Semantics (Full joint probability distribution)
2. Local Semantics (Conditional independent statements)
22. State the way to representing the Full joint Distribution.
Ans: A generic entry in the joint distribution is the probability of a conjunction of particular
assignments to each variables, such as
p(x1=x1^----^xn)xn)=p(x1,---- xn)
23. When will a Bayesian Network is said as Compact?
Ans: A Bayesian network is compact, when the information is complete and non-redundant.
The Compactness of Bayesian Network is an example of a very general property of
locally structured systems.
24. Explain Node Ordering.
Ans: The correct order in which to add nodes is to add the root causes (parents first and then
the variables) and the addition process is continued, until we reach the leaves which have no
direct casual inference on other variables.
31. What are the three basic classes of algorithms for evaluating multiply connected
networks?
Ans: The three basic classes of algorithms for evaluating multiply connected networks are
clustering, conditioning methods and stochastic simulation methods.
Ans: An optimal policy is a policy that yields the highest expected utility. It is denoted by .
58. Define proper policy.
Ans: A policy that is guaranteed to reach a terminal state is called a proper policy.
Ans: A finite horizon means that there is fixed time after which the game is over. With a finite
horizon, the optimal action in a given state could change over time. Therefore, the optimal
policy for a finite horizon is non stationary.
Ans: A contraction is a function of one argument, when applied to 2 different inputs, produce
2 output values that are closer together by some constant amount than the original values.
76. Write down Bayes’ Rule (Nov’13)(Apr-May’15)(Nov’15)
MYCIN was an early expert system that used artificial intelligence to identify bacteria
causing severe infections, such as bacteremia and meningitis, and to recommend antibiotics, with
the dosage adjusted for patient's body weight — the name derived from the antibiotics themselves,
as many antibiotics have the suffix "-mycin". The Mycin system was also used for the diagnosis of
blood clotting diseases.
Unit III- 11 Marks
Uncertainty
2. What is non-monotonic reasoning? Explain? (Apr-May ’14)(Nov – Dec ’14)(Nov’15)
A monotonic logic cannot handle various reasoning tasks such as reasoning by default
(consequences may be derived only because of lack of evidence of the contrary), adductive
reasoning (consequences are only deduced as most likely explanations), some important
approaches to reasoning about knowledge
Default reasoning
Adductive reasoning
Abductive reasoning is the process of deriving the most likely explanations of the
known facts. An abductive logic should not be monotonic because the most likely
explanations are not necessarily correct.
For example, the most likely explanation for seeing wet grass is that it rained;
however, this explanation has to be retracted when learning that the real cause of the
grass being wet was a sprinkler. Since the old explanation (it rained) is retracted because
of the addition of a piece of knowledge (a sprinkler was active), any logic that models
explanations is non-monotonic.
If a logic includes formulae that mean that something is not known, this logic should
not be monotonic. Indeed, learning something that was previously not known leads to the
removal of the formula specifying that this piece of knowledge is not known. This second
change (a removal caused by an addition) violates the condition of monotonicity. A logic for
reasoning about knowledge is the auto epistemic logic.
Belief revision
Belief revision is the process of changing beliefs to accommodate a new belief that might be
inconsistent with the old ones. In the assumption that the new belief is correct, some of the
old ones have to be retracted in order to maintain consistency. This retraction in response to
an addition of a new belief makes any logic for belief revision to be non-monotonic. The
belief revision approach is alternative to para consistent logics, which tolerate inconsistency
rather than attempting to remove it.
Probability
Given the available evidence, A25 will get me there on time with probability 0:04
(Fuzzy logic handles degree of truth NOT uncertainty e.g., WetGrass is true to degree 0:2)
These are not claims of a \probabilistic tendency" in the current situation (but might be
learned from past experience of similar situations)
There are 28 = 256 events in the sample space. Each event is determined by a joint
instantiation of all of the variables.
Basic insight: observations are conditionally independent of other hidden squares given
neighboring hidden squares
P(b|known,P13,P22,P31)P(P22,P31)
Set 1 =0.2*{(1*0.2*0.2)+(1*0.8*0.2)+(1*0.*0.8)+0*0.8*0.2)}
=0.2*0.36
=0.072
Set 2 =0.8*{1*0.2*0.2)+(0*0.8*0.2)+(1*0.2*0.8)+(0*0.8*0.8)}
=0.8*0.02
=0.16
P(P13|known,b) =α’(0.072,0.16)
0.072 0.16
= 0.072+0.16 + 0.072+0.16
= 0.31 , 0.689
= 0.31 , 0.69 (approximate)
6. Explain Bayesian networks (Nov’13)(Nov’15)
A simple, graphical notation for conditional independence assertions and hence for compact
specification of full joint distributions
Syntax:
Example
Topology of network encodes conditional independence assertions:
Compactness
Example
Deciding conditional independence is hard in noncausal directions (Causal models and
conditional independence seem hardwired for humans!)
Inference by enumeration
Slightly intelligent way to sum out variables from the joint without actually
constructing its explicit representation
7. Explain Dempster - Shafer Theory in AI (Apr-May’15)(Nov’15)
The belief for the “Alive” and “Dead” hypotheses matches their corresponding
masses because they have no subsets; belief for “Either” consists of the sum of
all three masses (Either, Alive, and Dead) because “Alive” and “Dead” are each
subsets of “Either”. The “Alive” plausibility is 1 − m (Dead) and the “Dead”
plausibility is 1 − m (Alive). Finally, the “Either”
plausibility sums m(Alive) + m(Dead) + m(Either). The universal hypothesis
(“Either”) will always have 100% belief and plausibility —it acts as a checksum
of sorts.
Here is a somewhat more elaborate example where the behavior of belief and
plausibility begins to emerge. We're looking through a variety of detector
systems at a single faraway signal light, which can only be coloured in one of
three colours (red, yellow, or green):
Combining beliefs
Overview
Degrees of truth
Fuzzy logic and probabilistic logic are mathematically similar – both have
truth values ranging between 0 and 1 – but conceptually distinct, due to
different interpretations—see interpretations of probability theory. Fuzzy
logic corresponds to "degrees of truth", while probabilistic logic corresponds
to "probability, likelihood"; as these differ, fuzzy logic and probabilistic logic
yield different models of the same real-world situations.Both degrees of truth
and probabilities range between 0 and 1 and hence may seem similar at first.
For example, let a 100 ml glass contain 30 ml of water. Then we may consider
two concepts: Empty and Full. The meaning of each of them can be
represented by a certain fuzzy set. Then one might define the glass as being 0.7
empty and 0.3 full.
In this image, the meanings of the expressions cold, warm, and hot are
represented by functions mapping a temperature scale. A point on that scale
has three "truth values"—one for each of the three functions. The vertical line
in the image represents a particular temperature that the three arrows (truth
values) gauge. Since the red arrow points to zero, this temperature may be
interpreted as "not hot". The orange arrow (pointing at 0.2) may describe it as
"slightly warm" and the blue arrow (pointing at 0.8) "fairly cold".
Linguistic variables
A linguistic variable such as age may have a value such as young or its
antonym old. However, the great utility of linguistic variables is that they can
be modified via linguistic hedges applied to primary terms. The linguistic
hedges can be associated with certain functions.