澳洲 bilogy Assignment 代写

  • 100%原创包过,高质代写&免费提供Turnitin报告--24小时客服QQ&微信:120591129
  • Data Mining Process
    Project/Business Understanding:
    Identify potential benefits, risks and
    efforts of successful project.
    澳洲 bilogy Assignment 代写
    Data Understanding: Sufficient
    relevant data
    Visual assessment of basic
    relationships and properties
    Data quality (missing values)
    Abnormal cases (outliers)
    Data Preparation: Selection,
    correction and modification of data
    Modeling: Extract knowledge out of
    data in the form of a model
    Predictive – Explanatory
    Evaluation
    Deployment
    Data Mining Cycle
    Data Understanding
    澳洲 bilogy Assignment 代写
    Main Goal
    Gain general insights about the data that will potentially be
    helpful for the further steps in the data analysis process
    Not driven exclusively by goals and methods of later steps
    Approach data from neutral viewpoint
    Never trust data before carrying out simple plausibility
    checks
    At the end of Data Understanding we know much better
    whether the assumptions we made during the Project
    Understanding phase concerning: representativeness,
    informativeness, and data quality are justified
    Visualisation: Overview of basic characteristics of data and
    check plausibility
    Simple statistics
    Outliers, missing values, data quality
    Data Visualisation
    Bar chart: Frequency distribution for categorical attribute
    Histogram: Frequency distribution for numerical attribute
    澳洲 bilogy Assignment 代写
    Divide values into bins and show a bar plot of the number of
    objects in each bin
    Height of each bar indicates the number of objects in bin
    Shape of histogram depends on number of bins
    Boxplots
    Very compact method to visualise distribution of one
    attribute
    Many boxplots can fit in single plot: Useful for comparing
    distributions
    Scatterplots
    Relationship between two attributes (linear/ non-linear)
    Axes represent two considered attributes
    Each instance in the dataset is represented by a point
    Correlation between attributes
    Outliers
    With class label info: Separability of classes
    Correlation Analysis
    Scatterplots can give us an idea about correlations
    between pairs of variables
    Pearson’s correlation coefficient: Measure of linear
    association between 2 numerical attributes. Always
    between -1 and 1
    Even if a functional dependency exists between two
    attributes and the function is monotone, if it is non-linear
    then Pearson’s correlation coefficient can be far away from
    -1 and 1
    Rank correlation coefficients overcome this by relying on
    the ordering of the values of the attributes: Spearman’s rho
    Outliers
    Outlier
    A value or a data object that is far away or very different from
    most or all of the other data
    Intuitive but imprecise definition
    It might be worthwhile to exclude outliers from analysis
    Different methods more robust to outliers than others
    Categorical Attribute: value that occurs with very low
    frequency
    Numerical Attribute: Detection much more difficult.
    Boxplot, Scatterplot
    For multidimensional data much more complicated
    approaches need to be used
    Missing Values
    Missing values: One of the most important problems in real
    applications
    Not one best way of handling missing values
    Missing Completely at Random: No special
    circumstances or special values of the variable in question
    lead to higher or lower chances for values to be missing
    Missing at Random: Probability of a missing value
    depends on some other variable(s) Y but conditionally
    on Y it is independent of the value of X
    Nonignorable missing: Occurrence of missing values
    directly depends on the true value of the attribute
    Distinguish Between Types of Missing Values
    Distinction between MCAR and MAR: In case of MAR
    other attributes can be used to predict whether value is
    missing
    Turn considered attribute into binary variable: 1 if value
    exists, 0 if it is missing
    Build a classifier to predict binary variable using as inputs
    other variables
    Determine error rate
    MCAR: Error rate is approx. equal to proportion of missing
    values
    MAR: Error rate is significantly lower (it could also be
    non-ignorable missing)
    In general not possible to distinguish non-ignorable
    missing from the other two cases using only available data
    Treating Missing Values
    Explicit Value: Replace with new value for attribute
    MISSING (nominal attributes)
    If the fact that the value is missing carries information
    about the value itself (non-ignorable missing) introduction
    of new value can help because it can express an intention
    not captured by other attributes
    Better Approach Introduce new binary variable indicating
    that value was missing in original dataset and then
    substitute missing value
    If neither other attributes or imputed value help but the fact
    that the value was missing is important, binary variable
    captures this
    If no such missing value pattern is present the imputed
    value can be used without introducing MISSING value
    Relevance of Attribute
    More realistic problem
    Information available: X = H: P(G) high, X = L: P(G) low
    General Decision Rule
    Given the risk forecast of an applicant, X = {H,L}:
    o(G|X = x) =P(G|X = x)
    P(B|X = x)>‘
    g
    Relevance of Attribute
    More realistic problem
     
    TN + FP
    Sensitivity: Minimise misclassification of Class 1 records
    (also called Recall)
    Specificity: Minimise misclassification of Class 0 records
    ROC Curve
    Critical points on ROC curve
    (TPR,FPR)
    (0,0): All records classified 0
    (1,1): All records classified 1
    (0,1): Ideal model
    Random Classifier: Diagonal Line
    Below diagonal line: Prediction is opposite of true class
    Good classifier: As close as possible to upper left corner
    Area Under ROC (AUC): Summarises ROC curve into a
    single number
    Cost-Sensitive Learning
    Cost of Misclassification
    C(i,j): Cost of misclassifying a pattern from class i to class j
    Cost Matrix:
    Predicted Class
    C(i,j)
    1
    Increasing variable xjby 1:
    Increases log(o(1|xi)) by βj
    Increases o(1|xi) by factor of eβj
    If xjis binary then xj= 1 increases o(1|xi) by eβj
    Synopsis Logistic Regression
    Linear predictor:
    Accommodates quantitative and qualitative variables
    (dummy)
    Enables transformations and combinations (interactions)
    while retaining interpretability. Logistic regressions extends
    this idea to binomial data
    Explanatory model:
    Contribution of individual variables
    Model comparison – Model selection
    Confidence interval (not covered)
    Linear relationship between attribute values and probability
    of success
    Non-linearities can be overcome using discretisation
    Decision Trees
    Decision Tree Approach
    Ask series of questions about attributes to determine class
    Build decision tree from top to bottom (from root to leaves)
    Greedy selection of a test attribute
    Compute an evaluation measure for all attributes
    Select the attribute with the best evaluation
    Greedy Strategy
    Grows a decision tree by making a series of locally optimal
    decisions about how to partition the data
    Divide and conquer / recursive descent
    Divide examples according to the values of the test attribute
    Apply the procedure recursively to the subsets (Hunt’s
    algorithm)
    Characteristics of Decision Tree Induction
    Non-parametric: No assumptions about the type of
    probability distributions satisfied by the data
    Finding optimal decision tree is computationally infeasible:
    Greedy heuristic approaches
    Decision tree induction algorithms construct trees quickly
    even for very large train sets
    Easy to interpret: Especially for small trees
    Robust to presence of noise: Especially when methods to
    avoid overfitting are employed
    Redundant attributes do not adversely affect accuracy
    If dataset contains many irrelevant attributes then some
    could be accidentally chosen by tree-growing algorithm.
    Feature selection
    Characteristics of Decision Tree Induction
    Data fragmentation: Number of records at leaf nodes can
    become too small to make statistically significant decision
    – Impose threshold on minimum number of records per
    node
    Subtree can be replicated many times within a decision
    tree making the model more complex and harder to
    interpret
    Robust performance w.r.t. choice of impurity measure
    Treatment of missing values
    Small changes in train set can yield entirely different tree
    Performance is robust
    Performance adversely affected by too many interval
    scaled variables (Discretisation)
    Artificial Neural Networks (ANN)
    ANNs inspired by attempts to model biological neural
    systems
    Brain consists of a large number of interconnected simple
    processing units (neurons)
    Learning in human brain takes place by changing the
    strength of the synaptic connection between neurons
    through repeated stimulation by the same impulse
    Perceptron
    Perceptron: Simple Model of a Neuron
    Each input node is connected via a weighted link to the
    summing junction
    Weights emulate strength of synaptic connection between
    neurons
    Training adapts weights to reduce error
    Can solve linearly separable problems
    Artificial Neural Networks
    Number of simple processing
    units (nodes)
    Organised in Layers
    Output layer: Returns prediction
    Input layer: Receives inputs
    Hidden layers: Layers between
    input and output layers
    Topology: 5 × 3 × 1
    Multilayer Perceptrons: Only
    Feed-forward connections
    More complicated decision boundaries can be
    approximated using more nodes and more layers
    Design Issues in ANNs
    Systems that combine automatic feature extraction with
    classification process
    Increasing the number of hidden nodes and the number of
    hidden layers ANNs can become very flexible classifiers
    Flexibility can easily result to Overfitting
    Selecting appropriate topology is Critical
    No general rule for how to choose the number of hidden
    layers and the size of the hidden layers
    Small neural networks might not be flexible enough to fit the
    data. Large neural networks tend to overfitting
    Cannot handle missing values
    Black box models: Explaining what an ANN has learned is
    not straightforward
    Very sensitive to chosen feature vector: Variable selection
    and preprocessing necessary
    Ensemble Methods
    Central Idea
    Improve accuracy by combining predictions of multiple
    classifiers
    Conditions for performance improvement
    1 Base classifiers (close to) independent
    2 Base classifiers better than random guessing
    Constructing Ensemble Classifiers: Bagging
    Bagging – Bootstrap Aggregating
    Create many training sets
    through Bootstrapping
    (resampling with replacement)
    Build classifier for each train set
    Use majority vote to predict
    Reduces variance of base classifiers
    Unstable classifier: Sensitive to minor perturbations in
    train data
    Bagging reduces generalisation error of unstable classifiers
    (Decision trees, Neural networks, k–nearest neighbours)
    Can be detrimental for stable/ robust classifiers because
    the size of the train set is reduced
    Does not focus on particular instances of training data
    Boosting
    Example: Weights determine sampling distribution
    Initially all weights are equal 1/N
    At each round i = 1,2,...
    Draw bootstrap sample Dibased on weights
    Base classifier built on Diand used to classify all examples
    from original dataset D
    Increase weights of misclassified examples
    Misclassified examples more likely to be chosen in
    subsequent rounds
    Attention focused on difficult to classify examples