Follow+Star Public Account, don’t miss the wonderful content
Source | Renesas Embedded Encyclopedia
What is Artificial Intelligence?
The rapid development of human technology continuously brings exciting new technologies. Currently, Artificial Intelligence (AI) is undoubtedly one of the hottest technologies.
Artificial Intelligence, abbreviated as AI, is a new scientific discipline that studies and develops theories, methods, technologies, and application systems used to simulate, extend, and enhance human intelligence. It is a scientific field that constructs computers and machines capable of reasoning, learning, and acting, which typically requires human intelligence or involves data scales beyond human analytical capabilities.
In the diagram below, terms such as artificial intelligence, machine learning, deep learning, data mining, and pattern recognition are often seen, along with many articles and discussions about their interrelations. Generally, artificial intelligence is a broad research field formed by the intersection of many disciplines; machine learning is a branch of artificial intelligence that provides many algorithms, while deep learning is an important branch of machine learning based on artificial neural networks; data mining and pattern recognition focus on algorithm applications. They complement each other and require support from knowledge in other fields.
The goals of artificial intelligence include: reasoning, knowledge representation, automated planning, machine learning, natural language understanding, computer vision, robotics, and strong artificial intelligence, encompassing eight aspects.
• Knowledge representation and reasoning include: propositional calculus and resolution, predicate calculus and resolution, which can derive some formulas or theorems.
• Automated planning includes the planning, actions, and learning of robots, state space search, adversarial search, and planning.
• Machine learning is a research area developed from a sub-goal of AI, aimed at helping machines and software to learn from experience to solve encountered problems.
• Natural language processing is another research area developed from a sub-goal of AI, aimed at facilitating communication between machines and humans.
• Computer vision is a field that arose from the goals of AI, aimed at identifying and recognizing objects that machines can see.
• Robotics is also derived from the goals of AI, aimed at giving a machine a physical form to perform actual actions.
• Strong artificial intelligence is one of the main goals of AI research, also referred to as artificial general intelligence (AGI), or the capability to perform general intelligent behaviors. Strong AI typically connects artificial intelligence with human characteristics such as consciousness, perception, knowledge, and self-awareness.
The three elements of artificial intelligence: data, algorithms, and models.
• Data is fundamental, discovering patterns in data based on data.
• Algorithms are key, determining how to solve models using data and turning them into executable programs.
• Models are core, explaining data and discovering data patterns.
Machine Learning in Artificial Intelligence
Machine learning is an important branch of artificial intelligence that improves system performance through computational methods and learning from experience. It includes: supervised learning, unsupervised learning, and reinforcement learning. The algorithms include: regression algorithms (least squares, LR, etc.), instance-based algorithms (KNN, LVQ, etc.), regularization methods (LASSO, etc.), decision tree algorithms (CART, C4.5, RF, etc.), Bayesian methods (naive Bayes, BBN, etc.), kernel-based algorithms (SVM, LDA, etc.), clustering algorithms (K-Means, DBSCAN, EM, etc.), association rules (Apriori, FP-Growth), genetic algorithms, artificial neural networks (PNN, BP, etc.), deep learning (RBN, DBN, CNN, DNN, LSTM, GAN, etc.), dimensionality reduction methods (PCA, PLS, etc.), and ensemble methods (Boosting, Bagging, AdaBoost, RF, GBDT, etc.).
Deep Learning in Artificial Intelligence
Deep learning is an important branch of machine learning, based on artificial neural networks. The learning process of deep learning is termed deep because the structure of artificial neural networks consists of multiple input, output, and hidden layers. Each layer contains units that convert input data into information for the next layer for specific prediction tasks. Thanks to this structure, machines can learn through their data processing. For example, in a two-layer network, where a is the value of a “unit,” w represents the “connection” weight, and g is the activation function, typically using the sigmoid function for convenience in differentiation. Matrix operations simplify the formulas in the diagram: a(2) = g( a(1) * w(1) ), z = g( a(2) * w(2) ). Let the true value of the training sample be y, the predicted value be z, and define the loss function loss = (z – y)2. The goal of optimizing all parameters w is to minimize the loss sum for all training data, transforming this problem into an optimization problem commonly solved using gradient descent algorithms. Generally, the backpropagation algorithm is used to calculate gradients layer by layer from back to front and ultimately solve each parameter matrix.
Traditional Machine Learning in Artificial Intelligence
• Linear Regression
o Linear regression is about finding a straight line, using data points to find the best-fit line. It attempts to represent the independent variable (x value) and the numerical outcome (y value) by fitting the straight line equation to the data. As per the formula, y=kx+b, where y is the dependent variable, x is the independent variable, and k and c are values obtained from the given dataset.
• Logistic Regression
o Logistic regression is similar to linear regression, but it is used for binary classification outputs (i.e., the output has only two possible values), predicting the final output as a nonlinear S-shaped function.
This logic function maps intermediate results to the outcome variable y, with values ranging between 0 and 1. These values represent the probability of y occurring. The properties of the S-shaped logistic function make logistic regression more suitable for classification problems.
• Decision Tree
o Decision trees are used for regression and classification problems. The trained model learns to predict the target variable’s value by learning the decision rules represented by the tree. The tree consists of nodes with corresponding attributes. At each node, questions about data are asked based on available features, with left and right branches representing possible answers, and the final nodes (leaf nodes) corresponding to predicted values. The importance of each feature is determined through a top-down approach, with higher nodes indicating more important attributes. As illustrated in the example of who in the crowd prefers using credit cards, if a person is married and over 30 years old, they are more likely to have a credit card (100% preference). Test data is used to generate the decision tree.
• Naïve Bayes
o Naïve Bayes is based on Bayes’ theorem. It measures the probability of each class, with the conditional probability of each class given the x value. This algorithm is used for classification problems to yield a binary “yes/no” result.
• Support Vector Machine (SVM)
o Support Vector Machine is a machine learning algorithm used to solve binary classification problems, finding a separating hyperplane in the sample space that separates samples of different categories while maximizing the minimum distance from two point sets to this hyperplane. As shown in the diagram, there are black and white point samples, and the goal of the support vector machine is to find a line (H3) that separates the black and white points while maximizing the sum of distances of all black and white points to this line (H3).
• K-Nearest Neighbors (KNN)
o KNN is an instance-based learning algorithm, or lazy learning that postpones all calculations until classification. It uses the nearest neighbors (k) to predict unknown data points. The value of k is a key factor in prediction accuracy, whether for classification or regression, and measuring the weight of neighbors is very useful, with closer neighbors weighted more heavily than distant ones. The disadvantage of KNN is its sensitivity to the local structure of the data. It has high computational demands, requiring normalization of data to ensure each data point is within the same range.
• K-Means
o K-Means is an unsupervised learning algorithm that provides a solution for clustering problems. The K-Means algorithm partitions n points (which can be a sample observation or an instance) into k clusters, ensuring that each point belongs to the cluster corresponding to the nearest average (centroid). This process continues until the centroids do not change.
• Random Forest
o Random Forest is a very popular ensemble machine learning algorithm. The basic idea of this algorithm is that many people’s opinions are more accurate than an individual’s opinion. In random forests, we use decision tree ensembles (see Decision Trees). To classify new objects, we take votes from each decision tree, combine the results, and make a final decision based on the majority vote.
• Dimensional Reduction
o Dimensional reduction refers to the process of reducing the number of random variables under certain conditions to obtain a set of “uncorrelated” main variables, which can be further divided into feature selection and feature extraction methods.
The Vision Board development board for machine vision in the embedded field has arrived.
What is the performance of the industry’s first Cortex-M85 MCU?
MCU million times read and write flash memory test.
Leave a Comment
Your email address will not be published. Required fields are marked *