
Data mining. Data analysis.
Tags | AlgorithmsComputer scienceMathProbabilityScienceStatistics |
---|---|
Created | |
Updated |
Resources
Foundations
Complement
Requisites
Introduction
What is Data mining?
Data mining is the process of finding insight patterns from data. Its goal is those patterns are information unexpected, non-trivial, useful, understandable, exact, and true.
Another point of view is in the DIKW pyramid context, where data mining converts data to information.
What is not Data Mining? | What is Data Mining? |
Look up how you can center a div on Stack Overflow | 60% of questions on StackOverflow is about web |
Create a relational database in 3NF | List what is the best clients in Mexico and answer why they buy us. |
List clients from SQL command. | |
Data mining tasks
The process is as follows:

Where data mining can be focused to generate descriptive or predictive information by applying algorithms, scalability concepts, computing architectures, and automation for handling large data.
Descriptive information means the process is on presenting insight into current variables from the data. Some descriptive techniques are clustering and association rules. Sometimes it’s called analytic processing.
Predictive information means the process is on presenting insight into future variables from the data. Some prediction techniques are models for the class attributes (machine learning) and anomaly detection. Sometimes it’s called inference of models.
Great opportunities but big challenges
Today, we collect, store, prepare, represent, model, reason, and visualize data to operate different data modalities such as ontologies, structured, networks, text, multimedia, and signals in order to give insights into Society’s major problems such as health care, climate change, etc. And from a market’s view, some applications are credit card fraud detection, network intrusion detection, and sales promotion.
However, some challenges are scalability, context, quality, streaming, high dimensionality, heterogeneous and complex data, data ownership and distribution, and non-traditional analysis.
Warning. Buzzwords. Marketing.
Research
Special Interest Group (SIG) on Knowledge Discovery and Data Mining (SIGKDD)
Operations Research Center
https://dl.acm.org/conference/kdd
Ecosystem
Economics Applications
Story
http://infolab.stanford.edu/~ullman/
Ethics
Concepts
Data Mining
Statistics
Probability
Machine Learning
Operations Research
IA
Data Science
Big Data
Data Analytics
Business Analysts
Domain Experts (Subject matter experts)
Domain Mining expert.
FAQ
Notes and references
Data mining project management
What is the scope of the project? Which questions have to answer?
Workers:
- Project leader.
- Specialist in data analysis/data mining.
- Expert with knowledge of the domain.
- Information technology (IT) support.
- Consumer.
Objectives
Deliverables
Roles and responsibilities
Current situation
Timeline
Costs and benefits
Input: Concepts, Instances, and Attributes. Data preparation. Data preprocessing.
graph TD
DataPreparation[Data preparation] --> DataCleaning["Data Cleaning"]
DataPreparation[Data preparation] --> FeatureExtraction["Feature extraction"]
DataPreparation[Data preparation] --> DataTrasnformation["Data transformation"]
DataPreparation[Data preparation] --> DataNormalization["Data normalization"]
DataPreparation[Data preparation] --> MissingDataImputation["Missing Data imputation"]
DataPreparation[Data preparation] --> NoiseIdentification["Noise identification"]
This is the most time-consuming step.
We suppose data exists, but sometimes is harder to use. So, Data mining doesn’t concern data collection or generation. But, quality data is necessary in order to make a good analysis. What is quality data? It’s a statistics question.
DataWarehouse and OLAP technology.
Data sources
The most important factor to analyze data is the quality of the data, which means data is reliable and represents the defined target population.
You must get data from studies such as
- Surveys or polls.
- Experiments.
- Observational and other studies.
You must get data from the daily operations of an organization such as
- Operational Databases.
- Data warehouses.
- Historical databases.
- Purchased data.
Data preparation steps
The following are steps to prepare data:
1. Create a data table from different sources.
You must query data sources such as operational databases, surveys, experiments, data warehouses, and historical databases, and integrate them as a data table.
We suppose a table but for other purposes XML or JSON.
Data Wrangling
https://missing.csail.mit.edu/2020/data-wrangling/
2. Characterize variables. Data understanding.
You must characterize your data table in standard ways.
Observations, Variables, data value.
graph TD
Object --> Observations
Variable X_1 | Variable X_2 | Variable X_3 | Variable X_4 | |
O1 | ||||
O2 | ||||
O3 | ||||
Observations are entity registers where variables describe them. A class instance.
Independent variables are sometimes called features.
Dependent variables are what you look up for. They are sometimes called targets.
Variables are observation attributes, so variables are sometimes called attributes.
Class attributes???
attributes?
Variable roles can be target or feature.
Your measurement has got to have units. Please, use International standards.
Random variables?
https://www.statisticshowto.com/attribute-variable/
Continuous and Discrete variables and their data type.
If variable X is 1 value where , then it is constant.
If variable X is 2 values, then it is dichotomous.
If variable values are , then it is discrete.
If variable values are , then it is continuous.
Identify data type by Variable.
Scales of Measurement
Scales of measurement are the relationship restriction between variable values and sensitivity. In general, it is a better high than low, where the highest is the Ratio.
Is it a partially ordered set? Is it a meaningful order? | Is it a meaningful distance? | Does it have absolute zero? It means that the lowest element in the set has got to be 0. It is a linear unit? It is proportional? | Example | |
Nominal. It names things. | No. | No. | No. | Eye colors whose values are “blue” and “brown”. |
Ordinal. It shows the position of something in a list, and nothing else. So, it is about order rather than amount. | Yes. | No. | No. | Level of satisfaction whose values are “low”, “medium”, and “high”. |
Interval. It represents amounts rather than orders. | Yes. | Yes. | No. | Celsius and Fahrenheit. |
Ratio. It represents proportions. | Yes. | Yes. | Yes. | Kelvin and Rankine. |
Cardinal vs Ordinal??
Roles in analysis
ID. These variables identify observations. For example, name.
Descriptors. These variables whose only purpose is describing data.
Response (classes or labels). These variables are predictive model outputs.
Frequency distribution
Variables.
Case study
Check out how the Iris flower data set or Fisher's Iris data set is understood.
References
Guyon & Elisseeff - An Introduction to Variable and Feature Selection - Journal of Machine Learning Research 3 (2003) 1157-1182
3. Clean data and deal with missing values
Consolidate observations
References
Computerphile. (2015, October 09). The Trouble with Missing Data - Computerphile. Youtube. Retrieved from https://www.youtube.com/watch?v=oCQbC818KKU&ab_channel=Computerphile
4. Data reduction.
Data reductions ways

We can manually remove variables that will not contribute to any analysis in obvious variables and small datasets.
graph TD
Reduction["Reduction techniques"] --> DR["Data reduction"]
Reduction["Reduction techniques"] --> SR["Sample numerosity reduction"]
Reduction["Reduction techniques"] --> CR["Cardinality reduction"]
DR --> FeatureExtraction["Feature extraction/construction"]
DR --> PCA["Principal component analysis"]
DR --> FactorAnalysis["Factor analysis"]
DR --> MDS["Multidimensional Scaling"]
DR --> LLE["Locally linear embedding"]
Other positive effects on the data reduction are:
- Speed up the processing of the DM algorithm.
- Improve data quality.
- Make the results easier to understand.
But how do you search for the correct data? A brute force search has an average performance of where is the feature cardinality and is the desired number of features in the subset. So, it’s not a good idea.
The following are the most relevant techniques that solve it.
Instance selection or generation
NEAREST NEIGHBOR RULE
https://www.cs.rit.edu/~rlaz/PatternRecognition/slides/kNearestNeighbor.pdf


Dimension reducer. Features selection. The curse of dimensional
Principal Component Analysis (PCA)
Factor analysis
Feature extraction
Textual
Latent Semantic Analysis (LSA)
Probabilistic Latent Semantic Analysis (pLSA)
LDA
A Bayesian version of pLSA
LSI
lda
5. Transform variables. Simplification.
Normalization.
Min-Max
Z-score
Decimal Scaling Normalization
Value mapping.
Distributed file system and MapReduce
Discretization.
Formally, it means giving examples and a target classes.
Some algorithms are ChiMerge, CAIM,
Output: Decision tree (C4.5 algorithm), Aprori, KNN, and Naive Bayes.
Tools. A
Meta-learning.
Some properties are static (versus dynamic), univariate (versus multivariate), supervised (versus unsupervised), splitting (versus merging), global (versus local), direct (versus incremental), evaluation measures (they are provided by domain or algorithm), parametric (versus nonparametric), disjoint (versus nondisjoint) —fuzzy logic, ordinal (versus nominal).
We compare methods using a number of intervals, inconsistent, predictive, …
Taxonomy.
Wu, X., Kumar, V., Ross Quinlan, J., Ghosh, J., Yang, Q., Motoda, H., ... & Steinberg, D. (2008). Top 10 algorithms in data mining. Knowledge and information systems, 14(1), 1-37.
Data discretization: taxonomy and big data challenge (2015). Sergio Ram ́ırez-Gallego, Salvador Garc ́ıa, H ́ector Mouri ̃no-Tal ́ın, David Mart ́ınez-Rego, Ver ́onica Bol ́on-Canedo, Amparo Alonso-Betanzos, Manuel Ben ́ıtez, and Francisco Herrera. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 6(1):5-21. DOI: 10.1002/widm.1173
Linear Transformations
Quadratic transformations
Non-polynomial Approximations of Transformations
Polynomial Approximations of Transformations
Rank Transformations
Box-Cox transformations
Dimension reduction methods
Transformations via Data Reduction
Algorithms
Aggregation.
6. Segment table
Create subsets of data to facilitate analysis and answer specific questions.
Random Samples.
Concept description: Characterization and comparison
References
García Salvador Luengo Julián & Herrera F. (2015). Data preprocessing in data mining. Springer International Publishing : Imprint : Springer. https://doi.org/10.1007/978-3-319-10247-4
FAQ
How do I prepare images?
Output: Knowledge Representation. Implementation of the analysis.
Descriptive modeling and predictive modeling.
Data Analytics
The category outputs that are determined by Untitled are
- Summarizing the data. Visualizing and Exploring data. Exploratory Data Analysis (EDA)..
- Finding hidden relationships.
- Making predictions.
They interact among themselves.
The methods are
- Summary tables
- Graphs
- Descriptive statistics
- Inferential statistics
- Correlation statistics
- Searching
- Grouping
- Mathematical models
Implementation
Worked example
The next table shows a series of retail transactions monitored by the main office of a computer game store.
Generate a contingency table summarizing the variables Store and Product category.
Generate the following summary tables:
- Grouping by Customer and the showing a count of the number of observations and the sum of Sale price ($) for each row.
- Grouping by Store and showing a count of the number of observations and the mean Sale price ($) for each row.
- Grouping by Product category and showing a count of the number of observations and the sum of the Profit ($) for each row.
Create a histogram of Sales price ($) using the following intervals 0 to less than 20, 20 to less than 30, 30 to less than 40, 40 to less than 50.
Create a scatterplot showing Sales price ($) against Profit ($)
Solution.
https://gist.github.com/sanchezcarlosjr/10fbc3d4ec8e5118d4a2960c0e67f0e7
Measurement and Data
Data Mining Primitives, Languages, and System Architectures
Visualizing and Exploring data. Exploratory Data Analysis (EDA).
Descriptive modeling
Finding patterns and rules
Data Analysis and Uncertainty
Mining Association rules in large databases
Cluster Analysis
Mining Complex Types of Data
Tables and Graphs
Diagrams from a Domain-specific language (DSL)
https://mermaid-js.github.io/mermaid/#/
Graphviz
https://avdi.codes/tools-for-turning-descriptions-into-diagrams/
https://www.augmentedmind.de/2020/12/20/plain-text-diagrams-as-code/
https://modeling-languages.com/text-uml-tools-complete-list/
What's the difference between a graph, a chart, and a plot?
Images
Statistics
Grouping
Supervised vs unsupervised
Grouping methods are methods whose goal is segmenting, finding hidden relationships, and understanding data.
Cluster
Types of variables
Data set limit
Interpretable and actionable
Overlapping groups
Identify outliers
endograma
Hierarchical Agglomerative Clustering
K-Means
Preconditions.
- Data measurement should be Interval.
- Don’t mix units.
- Data should be normalized.
k_means(data, k groups)
allocate an observation randomly
Association rules
The Association rules method is an unsupervised groping method.
Preconditions.
- Data measurement should be Categorial (Nominal or Ordinal).
- Instances have to be 0 or 1.
Apriori algorithm
Decision trees
Preconditions.
- Categorical and continuous variables.
FAQ
Prediction
Predictive
Predictive Modeling for Classification
Predictive Modeling for Regression
Models and Patterns. Classification and Prediction
Real Machine Learning
- Unsupervised learning. No classes or Labels.
- Clustering.
- Dimensionality.
- Supervised learning. Classes or Label.
- Classification.
- Regression.
- Deep learning.
- Reinforcement learning.
Paradigms of Machine Learning.
Symbolysts.
Connectionists
Bayesians.
Analogizers.
Evolutionaries.
The task of Machine Learning.
Classification. If you have discrete labels, then you need classification algorithms.
Regression. If you have continuous labels, then you need regression algorithms.
Contingency table and confusion matrix.
Holdout (1/3 testing, 2/3 training).
K-fold cross-validation K=10.
Linear Regression.
Nonlinear regression. If you can, transform nonlinear regression to a linear relationship using a mathematical transformation.
Data transformations
Ensemble Learning
Moving on
Classification
Classification trees
k-Nearest Neighbors
Logistic regression
Naive Bayes classifiers
Neural networks
Rule-based classifiers
Support vector machines
Regression
Regression trees
k-Nearest Neighbors
Linear regression
Neural networks
Nonlinear regression
Partial least squares
Support vector machines
Associate rules
Regression modeling
Multiple regression and model building
Logistic regression
Naive Bayes estimation and Bayesian networks
Tools
Weka
Data Mining: Practical Machine Learning Tools and Techniques
https://www.cs.waikato.ac.nz/ml/weka/
Search and Optimization Methods
Score Functions for Data Mining Algorithms. Credibility: Evaluating What’s been learned
Data Organization and Databases
Evaluation & Deployment
Some kinds of deliverables are
- Static report to management.
- Software integration.
Then, measure and monitor performance and review the project.
Report examples
https://raw.githubusercontent.com/mrsac7/Data-Mining-Project/master/Report.pdf
Storytelling
Summary
Data mining process
• Kurgan L. A. and Musilek P. (2006). A survey of Knowledge discovery and Data Mining process models. The Knowledge Engineering Review, Vol. 21:1, 1-24. doi:10.1017/S0269888906000737. Recuperado de: http://biomine.cs.vcu.edu/papers/KER-KDDM2006.pdf
Software Ecosystem
SAS SEMMA???
- The WEKA data mining workbench
- Microsoft’s OLE DB for Data Mining
- DBMiner
- Clementine data mining suite
- SPSS statistical software
- SAS Enterprise Miner
Applications and Trends in Data Mining
Project ideas to learn data mining
Email spam detector
Modeling response to direct mail marketing
Wordnet
https://archive-beta.ics.uci.edu/
Intelligence
https://www.fbiic.gov/public/2008/feb/ODNI_Data_Mining_Report.pdf
References
Case of study
https://arxiv.org/pdf/0803.2969.pdf