30 Cards
AI
Take your first step towards learning about artificial intelligence with all the definitions of important AI concepts and terms with simple flashcards.
Back to listLearn AI concepts
Active Learning
Active learning is an important subdomain of machine learning. It involves a learning algorithm with the capability of making interactive queries to users in order to label data with the required outputs. Active learning algorithms could proactively choose the subset of examples for labeling from a collection of unlabeled data. One of the primary principles underlying active learning is the ability of the ML algorithm to reach higher accuracy with a limited number of training labels. Active learning relies on offering flexibility for the algorithm to choose training data.
Artificial Neural Network
Artificial Neural Networks are algorithms that simulate the brain function with training to model complicated patterns alongside forecasting problems. Artificial Neural Networks or ANNs include artificial neurons, also known as units. The units are organized in a collection of layers that build the complete Artificial Neural Network. Layers in the ANN could harbor different numbers of units, depending on the complexity of the neural networks. You can find an input layer and an output layer in ANNs alongside other hidden layers.
Association Rule Learning
It refers to an unsupervised learning approach that verifies the dependency of a data item on another data object. It would map the data items according to the dependencies to ensure more profitability. Association rule learning focuses on the identification of exclusive relationships between variables in the dataset. Association rule learning uses multiple rules to discover the relations between variables. It is one of the most promising concepts in machine learning that finds applications in web usage mining, continuous production, and others.
Autoencoder
Autoencoder is a deep learning algorithm that has been created in order to receive inputs and transform them into different outputs. The algorithms are capable of image construction, and they can serve useful applications in the domain of unsupervised machine learning. You could utilize an autoencoder for data compression and to reduce its dimensionality. It includes three layers in its architecture, such as the encoder, code, and decoder layer. Autoencoder is capable of reconstructing the original input even from a compressed version.
Bayes’ Theorem
Bayes’ theorem is an important rule that helps determine the probability of an event by utilizing uncertain knowledge. In the case of probability theory, the theorem establishes the relationship between conditional probability and the marginal possibilities of two random events. The theorem plays a crucial role in updating the prediction of probabilities of events through the observation of new information from the real world. It has many practical applications in the domain of AI, including machine learning, robotics, and drug testing.
Bias
Bias in machine learning or AI is the tendency of algorithms to exhibit biases like humans. Machine learning bias occurs when algorithms deliver biased results systematically due to flawed assumptions in the machine learning process. You can come across two common types of bias in machine learning algorithms, such as inductive bias and confirmation bias. Inductive bias refers to assumptions used for predicting outputs based on inputs that the model has never encountered before.
Classification
Classification is an important supervised learning technique. Classification models can ensure the prediction of a class label for example. The class label could also indicate whether a user will return or a certain transaction is fraudulent. Classification models could also predict whether a specific image includes a car. The classification technique in AI is productive for businesses with large volumes of historical training data, alongside labels, to specify the presence of an asset in the concerned group.
Clustering
Clustering is the process of arranging similar objects into different groups within a machine-learning algorithm. AI models could capitalize on the benefits of assigning related objects across different clusters. Clustering offers multiple use cases in data science. Clustering involves scanning unlabeled datasets in the ML model alongside establishing measurements corresponding to the particular data point features. Subsequently, the cluster analysis method classifies and places data points in a group that has similar features.
Confidence Interval
It is a tool used for the quantification of uncertainty in the estimates of ML algorithms. You can use confidence intervals to add limits or possibilities on a population parameter. Confidence intervals are different from tolerance intervals, which describe the limits of data obtained from distribution. The application of the confidence interval in applied ML revolves around ensuring the effective presentation of skills for a predictive model. It can also help in presenting the skill of a classification model.
Convolutional Neural Network
The Convolutional Neural Network, or CNN, is a subdomain of machine learning and one of the variants of artificial neural networks. The networks could serve different applications with support for a broad range of data types. Convolutional Neural Network is a type of network architecture suited for deep learning algorithms. The most common applications of a convolutional neural network are image recognition and tasks that focus on processing pixel data. CNNs are an ideal choice of neural networks for computer vision tasks.
Decision Tree
It refers to a supervised learning technique that applies to classification as well as regression problems. However, it serves as the best choice for solving classification problems. The tree-structured classifier features internal nodes that represent features of a dataset, while the branches serve as representations of decision rules. The leaf nodes provide a representation of the output. Decision node and leaf node are the two critical components of a decision tree and serve as a graphical representation of all solutions to one problem.
Deep Learning
Deep learning is one of the most noticeable subdomains of machine learning. It is a neural network featuring three or more layers that work on simulating human behavior. The deep learning approach focuses on allowing the neural network to learn from massive volumes of data. Additional hidden layers could help in optimizing and refining the approximate predictions by a single layer. Deep learning is the driving power behind different AI applications and services to improve automation.
Ensemble Methods
It’s one of the innovative techniques in ML, which involves a combination of different base models to generate one predictive model with optimal performance. It assures better performance in prediction and helps in resolving statistical problems, representational problems, and computational problems in machine learning. The common methods for independent development of ensemble methods include majority vote, randomness injection, random forest, error-correcting output coding, and feature-selection ensembles.
Feature Learning
Feature learning refers to a collection of techniques that allow systems to automatically discover the representations required for the detection of features or classification in raw data. The applications of feature learning help in removing the manual feature from the engineering feature. It helps a machine in learning the features as well as utilizing the features for addressing a particular task. The primary motivation underlying the applications of feature learning is the requirement of computationally convenient inputs for machine learning tasks.
F-Score
The F-Score is an important metric for the accuracy of an AI model on a specific dataset. It is also known as F1-score and serves as an evaluation of binary classification systems for distributing examples into negative and positive categories. The F-score serves as an effective tool thanks to a combination of precision and recall associated with this model. It is also represented as the harmonic mean of the model’s recall and precision. The F-score generally helps in evaluating information of retrieval systems, such as search engines.
Human-in-the-Loop
Human-in-the-Loop or HITL systems in AI are the systems that enable humans to offer direct feedback to a model. It helps in drawing predictions below a specific level of confidence. The technical explanation for HITL suggests that it is a combination of active learning and supervised machine learning with the involvement of humans. Human users are actively involved in the ML training procedure for HITL systems, especially in the training and testing stages of the algorithm development.
Hyperparameter Tuning
All the datasets and models in training for ML models require different collections of hyperparameters. Hyperparameters are a type of variable, and the only way to determine the hyperparameters involves multiple experiments. The experiments involve the selection of a set of hyperparameters, alongside running them through the model, with a process called hyperparameter tuning. The hyperparameter tuning process actually involves sequential training of the ML model with different collections of hyperparameters.
Composite AI
Composite AI or multidisciplinary AI is the combined utilization of multiple AI techniques in order to improve learning efficiency. It could also offer a broader level of knowledge representations. Composite AI emphasizes the fact that you cannot have a single AI technique for all purposes. Users can obtain the best results from AI in solving complex business issues by combining different types of AI. Composite AI shifts the attention away from the ‘one size fits all’ approach for ML models.
Knowledge Graph
The knowledge graph is a representation of the network of real-world objects, concepts, situations, or events, alongside highlighting the relationships between them. The information about their relationship goes into the graph database, where it is visualized in the form of a graph. The knowledge graph includes three important components, such as labels, edges, and nodes. Knowledge graphs can serve different use cases, such as accessibility and integration of data sources or adding context to data-driven AI techniques.
Machine Learning
Machine learning is a popular subdomain of artificial intelligence and computer science. It utilizes data and algorithms to simulate the learning approaches of humans to achieve better accuracy. Machine learning is a crucial component in the growing domain of data analytics and AI. The algorithms use statistical methods to make classifications and predictions, alongside uncovering major insights from data mining projects. The insights would then play a major role in driving decision-making, with an impact on growth metrics.
Multi-modal Learning
Multi-modal learning or multi-modal AI refers to the domain of artificial intelligence that could combine different types or modes of data. It can help in generating more accurate predictions alongside insightful conclusions. Multi-modal AI also helps in making more precise predictions regarding real-world issues. The multi-modal AI systems rely on training with video, text, images, speech, and audio, along with a broad collection of traditional numerical data sets. Multi-modal AI also ensures the use of multiple data types in combination.
Pattern Recognition
It refers to the process of using a machine learning algorithm to recognize patterns. You can also understand pattern recognition as the classification of data according to existing knowledge. It also relies on statistical information obtained from patterns alongside their representation. The most promising highlight of pattern recognition is the variety of applications possible for this concept. For example, pattern recognition is useful for multimedia document recognition, speech recognition, and speaker identification.
Principal Component Analysis
PCA refers to an unsupervised learning algorithm that can help in dimensionality reduction for machine learning use cases. It involves a statistical process for converting observations regarding correlated features into a collection of linearly uncorrelated features by using orthogonal transformation. The new transformed features would be the Principal Components, and PCA serves as a renowned tool for predictive modeling and exploratory data analysis. PCA also provides a reliable technique for drawing strong patterns from a specific dataset.
Random Forest
Random Forest is a machine learning algorithm associated with the supervised learning technique category. You can use the Random Forest algorithm for classification as well as regression tasks in machine learning. Interestingly, the Random Forest algorithm leverages the ensemble learning concept and combines multiple classifiers to obtain solutions to a complex problem. It serves as a classifier with different decision trees on subsets of the concerned dataset. It uses the average of all subsets in the training data to improve prediction accuracy.
Recurrent Neural Networks
Recurrent Neural Networks are a variant of artificial neural networks that utilize time series data or sequential data. Deep learning algorithms generally serve valuable applications in temporal or ordinal problems, like speech recognition, language translation, image captioning, and NLP. Recurrent neural networks use training data like convolutional neural networks for learning. The distinct highlight of recurrent neural networks is the ‘memory’ of information obtained from previous inputs that can influence existing inputs and outputs.
Regression
Regression is one of the most popular concepts in machine learning within the supervised learning approach. It refers to the process of training an algorithm with input features as well as output labels. As a result, regression analysis could provide a clear impression of the relationship between variables through estimates of the impact of one variable on the other. The primary goal of the regression algorithm focuses on plotting a best-fit curve between the data and offering predictions for continuous values.
Reinforcement Learning
It refers to the machine learning method which focuses on providing rewards for favorable behaviors and penalties for undesired behavior. The reinforcement learning agent perceives and understands its environment, and then takes action and learns through a trial-and-error approach. Developers have to create a method for offering rewards for desired behaviors alongside implementing punishment for negative behaviors. Reinforcement learning helps in training agents to seek long-term rewards to offer the optimal solution.
Semi-supervised Learning
Semi-supervised learning is the variant of the machine learning model that provides representation for the intermediate place between unsupervised and supervised learning algorithms. Semi-supervised learning is a crucial solution for the setbacks in supervised and unsupervised learning. For example, the primary setback with supervised learning is the need for hand labeling. On the other hand, unsupervised learning has limited applications, thereby pointing toward the importance of semi-supervised learning.
Turing Test
The Turing Test is an important method for inquiry in the domain of AI that determines the capability of a computer to think like humans. Alan Turing suggested that a computer could have artificial intelligence if it could respond to questions like humans under certain conditions. The test has also been criticized as it restricts the nature of questions used in testing the potential of computers to exhibit human intelligence. Some of the notable variations of the Turing test include the Reverse Turing test, and the Total Turing test.
Underfitting
Underfitting refers to a situation in data science where the data model cannot capture the relationship between input and output variables with accuracy. As a result, the model can generate a high error rate on the training set as well as the input data. Underfitting is evident in scenarios where the model is very simple or needs more training time and input features or lesser regularization. Under Fitted models could not identify the dominant trends in the data, thereby leading to training errors alongside poor performance in the model.