0% found this document useful (0 votes)
10 views

22AI2101 CIA II QB With Answer Key

The document is a question bank for an introductory course on Artificial Intelligence and Machine Learning, covering topics like Alpha-Beta pruning, machine learning paradigms, classification methods, and decision trees. It includes questions with answers on various concepts, such as supervised vs unsupervised learning, Naïve Bayes classification, and the PAC learning framework. Additionally, it provides detailed explanations and examples for practical applications of these concepts in real-world scenarios.

Uploaded by

yva8219
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views

22AI2101 CIA II QB With Answer Key

The document is a question bank for an introductory course on Artificial Intelligence and Machine Learning, covering topics like Alpha-Beta pruning, machine learning paradigms, classification methods, and decision trees. It includes questions with answers on various concepts, such as supervised vs unsupervised learning, Naïve Bayes classification, and the PAC learning framework. Additionally, it provides detailed explanations and examples for practical applications of these concepts in real-world scenarios.

Uploaded by

yva8219
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 7

22AI2101/ INTRODUCTION TO ARTIFICIAL INTELLIGENCE AND

MACHINE LEARNING

CIA II (QUESTION BANK WITH ANSWER KEY)

PART A
1. Outline the purpose of Alpha-Beta pruning in game tree search.
Alpha-Beta pruning is used to reduce the number of nodes evaluated in the minimax
algorithm by eliminating branches that cannot possibly influence the final decision,
thus improving computational efficiency.
2. Summarize how Alpha-Beta pruning improves the efficiency of the Minimax
algorithm.
It improves Minimax efficiency by pruning away branches that don't need to be
explored, reducing the time complexity from O(b^d) to O(b^(d/2)) in best cases, where
b is branching factor and d is depth.
3. Identify the main components of a machine learning system.
The main components of a machine learning system are:
o Data: The input examples used for training and testing.
o Model: The algorithm or mathematical representation that learns
patterns from the data to make predictions or decisions.
4. Compare supervised, unsupervised, and reinforcement learning.
Supervised learning uses labeled data to train models (e.g., classification),
unsupervised learning identifies patterns in unlabeled data (e.g., clustering), and
reinforcement learning learns through rewards and punishments from interacting with
an environment (e.g., game playing).
5. Summarize the assumptions of PAC Learning.
PAC (Probably Approximately Correct) Learning assumes that:
The learning algorithm receives independent and identically distributed (i.i.d.)
examples from a fixed but unknown distribution. The hypothesis class contains a
function that can approximate the target concept well.
6. Recall the basic probability laws used in ML.

7. Compare Multi-Class vs. Multi-Label classification.

 Multi-Class Classification involves assigning one label from multiple classes to


each instance (e.g., classifying animals as cat, dog, or bird).

 Multi-Label Classification allows multiple labels to be assigned to a single


instance (e.g., tagging a movie as both "action" and "comedy").
8. Summarize the working of a Naïve Bayes model.
Naïve Bayes is a probabilistic classifier based on Bayes’ Theorem with a strong
assumption of feature independence. It calculates the posterior probability for each
class and predicts the class with the highest probability for a given input.
9. Identify overfitting in Decision Trees.
Overfitting in Decision Trees occurs when the model learns not only the general
patterns but also the noise in the training data. This results in high accuracy on training
data but poor performance on unseen or test data.
10. Illustrate the concept of density-based clustering.
Density-based clustering groups data points that are closely packed together, marking
regions with high density as clusters and low-density regions as noise. A common
example is DBSCAN, which forms clusters based on a minimum number of points
within a given radius.
11. Identify the significance of terminal states in a game tree.
Terminal states in a game tree represent the end points of a game where no further
moves are possible. These states are evaluated using a utility or payoff function to
determine the outcome (win, lose, or draw) and guide decision-making for optimal
moves.
12. Summarize the impact of an evaluation function on the performance of
adversarial search.
An evaluation function estimates the desirability of a game state when terminal states
are not yet reached. Its accuracy significantly impacts the performance of adversarial
search by guiding the algorithm toward favorable moves and improving decision-
making under limited search depth.
13. Illustrate the concept of reinforcement learning.
Reinforcement Learning (RL) is a learning paradigm where an agent learns to make
decisions by interacting with an environment. It receives rewards or penalties based on
its actions and uses this feedback to learn optimal strategies through trial and error.
14. Show how version spaces help in hypothesis elimination.
Version spaces help in hypothesis elimination by maintaining a set of all hypotheses
consistent with the observed training examples. As new examples are introduced,
inconsistent hypotheses are removed from the version space, thereby narrowing down
the possible hypotheses until the target concept is identified.
15. Compute the probability of two independent events.
P(A∩B) = P(A) × P(B) for independent events A and B.
16. Give an example of Bayes’ theorem in ML.
In machine learning, Bayes’ theorem is used in Naive Bayes classifiers to predict the
probability of a class given the input features. For example, to classify an email as
spam or not, Bayes’ theorem calculates:
P(Spam | Email) = [P(Email | Spam) × P(Spam)] / P(Email)
This helps in deciding whether an email is spam based on its content.
17. Name two linear and two non-linear machine learning models.
o Linear: Linear Regression, Logistic Regression
o Non-linear: Decision Trees, Neural Networks
18. Summarize the key assumptions made in Linear Regression.
o Key assumptions in Linear Regression are:
o Linearity: The relationship between the independent and dependent variables is
linear.
o Homoscedasticity: The variance of the errors is constant across all levels of the
independent variables.
19. Compare K-Means and K-Modes Clustering.
K-Means Clustering:

 Works with numeric data using means to update cluster centers.


 Uses Euclidean distance for measuring similarity.

K-Modes Clustering:

 Designed for categorical data using modes to update cluster centers.


 Uses matching dissimilarity measure instead of Euclidean distance.

20. Identify a real-world application of clustering.


Market segmentation, image compression, anomaly detection, document organization.

PART B
1. Explain how Alpha-Beta pruning is applied to a game tree and describe how it improves
the efficiency of the Minimax algorithm. Support your explanation with a suitable example
and diagram.
Alpha-Beta pruning application:
o Explain with game tree example
o Show nodes that would be pruned
o Compare with full minimax evaluation
o Diagram showing pruning process
2. Describe step-by-step how Alpha-Beta pruning works and identify which nodes are pruned
in the game tree. Use a clear example and diagram to support your explanation.
Alpha-Beta pruning steps:
o Initialize alpha (-∞) and beta (+∞)
o Depth-first search
o Update alpha at MAX nodes
o Update beta at MIN nodes
o Prune when alpha ≥ beta
o Example with values showing pruned branches
3. Demonstrate the different learning paradigms in Machine Learning with suitable examples.
Learning paradigms:
o Supervised: Classification/Regression (e.g., spam detection)
o Unsupervised: Clustering/Dimensionality reduction (e.g., customer
segmentation)
o Reinforcement: Reward-based learning (e.g., game playing AI)
o Semi-supervised: Mixed labeled/unlabeled data
Illustrate with an example how the Version Space algorithm can be used to update a
hypothesis space during learning.
Version Space algorithm:
4. o General and specific boundary sets
o Candidate elimination algorithm steps
o Example with hypothesis space updating
o Convergence properties
5. Describe in detail the concept of Probably Approximately Correct (PAC) learning. How
does it ensure model reliability?
PAC Learning:
o Formal definition
o Sample complexity
o Computational complexity
o Relationship to VC dimension
o Examples of PAC-learnable classes
6. Explain, with an example, how conditional probability and Bayes’ theorem are used in
model prediction.
Conditional probability & Bayes:
o Bayes' theorem formula
o Prior, likelihood, posterior
o Example: Medical diagnosis
o Naïve Bayes classifier application
7. Explain the differences between linear and non-linear models with appropriate examples.
Compare their advantages and limitations.
Linear vs Non-linear models:
o Linear: Simpler, interpretable (e.g., Linear Regression)
o Non-linear: Complex patterns (e.g., Neural Networks)
o Examples of each type
o Tradeoffs in bias-variance
8. Explain the steps of performing K-Means clustering on a dataset. Discuss how to choose
the number of clusters and interpret the results.
K-Means steps:
o Initialize centroids
o Assign points to nearest centroid
o Recalculate centroids
o Repeat until convergence
o Methods for choosing k (elbow, silhouette)
o Limitations and variants
9. Apply linear regression to a given dataset and interpret the resulting model. Include
derivation of the equation, assumptions, and error analysis.
Linear regression application:
o Model equation derivation
o Assumptions (linearity, independence, etc.)
o Cost function (MSE)
o Gradient descent
o Evaluation metrics (R², RMSE)
10. Use multiple linear regression on a given dataset with two or more independent variables
to derive coefficients and evaluate the model’s performance.
o Model Setup
o Coefficient Derivation
o Performance Metrics
o Critical Assumptions
o Example
11. Apply Minimax and Alpha-Beta pruning to a game tree and compare their efficiency with
examples.
Minimax vs Alpha-Beta:
o Minimax full tree evaluation
o Alpha-Beta pruning process
o Efficiency comparison
o Game tree example showing pruned nodes
12. Demonstrate how adversarial search techniques are used in real-world AI game-playing
systems like chess engines and strategic decision-making, with examples.
Adversarial search applications:
o Chess engine minimax with evaluation
o Real-time strategy game AI
o Negotiation systems
o Security applications
13. Explain supervised, unsupervised, and reinforcement learning, highlighting the advantages
and limitations of each.
Learning paradigms comparison:
o Supervised: Needs labels, predictive accuracy
o Unsupervised: Finds hidden patterns
o Reinforcement: Long-term reward optimization
o Strengths/weaknesses of each
14. Describe the use of probability theory in machine learning and explain the roles of prior,
likelihood, and posterior in learning models.
Probability in ML:
o Bayesian learning framework
o Prior knowledge incorporation
o Likelihood function
o Posterior updating
o Applications in classification
15. Analyze the PAC learning framework and describe its significance in machine learning
model evaluation. Include examples.
PAC Learning analysis:
o Formal definition
o Sample complexity bounds
o Computational tractability
o Relationship to generalization
o Example applications
16. Apply the basic laws of probability (addition, multiplication, conditional) to solve a real-
world ML classification problem. Provide suitable data.
Probability laws application:
o Problem setup (e.g., medical test)
o Addition rule for unions
o Multiplication for intersections
o Bayes' for conditional
o Complete probabilistic model
17. Differentiate between multi-class and multi-label classification problems. Provide real-
world examples and algorithms used.
Multi-class vs Multi-label:
o Definition and examples
o Evaluation metrics differences
o Algorithm adaptations
o Use cases for each
18. Explain how the Naïve Bayes Classifier works and describe how conditional probability
and Bayes’ theorem are used in classifying text documents.
Naïve Bayes classifier:
o Mathematical formulation
o Feature independence assumption
o Training process
o Text classification example
o Smoothing techniques
19. Construct a decision tree manually using a small dataset. Use entropy and information gain
to justify each split.
Decision tree construction:
o Entropy/Information Gain calculation
o Recursive splitting
o Stopping criteria
o Example with small dataset
o Tree visualization
20. Apply the key assumptions and limitations of decision trees to a problem and demonstrate
how overfitting can be reduced using pruning or ensemble methods.
1. Key Assumptions:
 Features are treated as conditionally independent
 Splits are axis-aligned (perpendicular to feature axes)
 Uses greedy, locally optimal splitting (maximizes immediate information gain)
2. Main Limitations:
 High tendency to overfit (creates overly complex trees)
 Poor generalization with noisy data
 Instability (small data changes cause large tree structure changes)
 Bias toward features with more levels/higher variance
3. Pruning Solution:
 Post-pruning: Grows full tree then removes insignificant branches
 Uses validation set to determine optimal tree size
 Reduces complexity while maintaining accuracy
 Example: Cost-complexity pruning with α parameter tuning
4. Ensemble Methods:
 Random Forest: Builds multiple de-correlated trees via bagging
 Boosting (XGBoost): Sequentially corrects errors with weighted trees
 Both reduce variance and improve generalization
 Example: Single tree 75% accuracy → Random Forest 85% accuracy

PART C
1. Analyze the effectiveness of supervised, unsupervised, and reinforcement learning
paradigms in solving a real-world task like fraud detection or customer segmentation,
supporting your analysis with examples.
Learning paradigms analysis:
o Supervised for fraud detection (labeled fraud cases)
o Unsupervised for customer segmentation (no labels)
o Reinforcement for adaptive systems (sequential decisions)
o Compare effectiveness for each task
2. Examine and formulate an approach to determine whether a concept class is PAC-learnable,
discussing the factors that influence sample complexity in the model.
PAC-learnability approach:
o Define concept class
o Determine VC dimension
o Calculate sample complexity
o Discuss approximation parameters (ε, δ)
o Example analysis
3. Analyze a dataset containing student test scores and demographic data to determine which
supervised learning model (e.g., Decision Tree, Naïve Bayes, or Linear Regression) is most
suitable for predicting final grades. Justify your choice.
Model selection analysis:
o Dataset exploration
o Feature analysis
o Model suitability comparison
o Evaluation metric selection
o Final recommendation with justification
4. Design and analyze a predictive model using multilinear regression to forecast house prices
based on features like size, location, and age. Explain each step from data preparation to
model validation.
House price prediction:
o Data cleaning/preprocessing
o Feature selection/engineering
o Model formulation
o Coefficient interpretation
o Validation approach
o Error analysis

You might also like