New pages
12 June 2025
- 11:5011:50, 12 June 2025 Chemical Thermodynamics (hist | edit) [2,653 bytes] Thakshashila (talk | contribs) (Created page with "== Chemical Thermodynamics == '''Chemical Thermodynamics''' is the branch of thermodynamics that studies the interrelation of heat and work with chemical reactions or physical changes of state within chemical systems. It provides the framework to predict whether a reaction will occur spontaneously and to what extent it proceeds. === Basic Concepts === Chemical thermodynamics deals with the energy changes and equilibrium conditions in chemical reactions, focusing on va...")
- 11:4911:49, 12 June 2025 Entropy (hist | edit) [2,526 bytes] Thakshashila (talk | contribs) (Created page with "== Entropy == '''Entropy''' (symbol <math>S</math>) is a fundamental thermodynamic property that measures the degree of disorder or randomness in a system. It quantifies the number of microscopic configurations that correspond to a thermodynamic system's macroscopic state. === Definition === Entropy is related to the number of possible microstates (<math>\Omega</math>) by the Boltzmann equation: <math> S = k_B \ln \Omega </math> where: * <math>S</math> = entropy...")
- 11:4911:49, 12 June 2025 Phase Equilibrium (hist | edit) [2,990 bytes] Thakshashila (talk | contribs) (Created page with "== Phase Equilibrium == '''Phase equilibrium''' refers to the condition where multiple phases of a substance coexist in equilibrium without any net change in their amounts over time. It occurs when the chemical potential of each component is the same in all coexisting phases, ensuring no driving force for phase change. === Basics === In a system involving different phases (solid, liquid, gas), phase equilibrium is established when the rates of phase transitions (such a...")
- 11:4811:48, 12 June 2025 Thermodynamic Potential (hist | edit) [2,941 bytes] Thakshashila (talk | contribs) (Created page with "== Thermodynamic Potential == '''Thermodynamic potentials''' are scalar quantities used in thermodynamics to describe the equilibrium and spontaneous behavior of physical systems. They are functions of state variables such as temperature, pressure, volume, and entropy, and provide criteria for spontaneous processes and equilibrium under different constraints. === Overview === Thermodynamic potentials combine the system's internal energy with other thermodynamic paramet...")
- 11:4711:47, 12 June 2025 Gibbs Free Energy (hist | edit) [2,466 bytes] Thakshashila (talk | contribs) (Created page with "== Gibbs Free Energy == '''Gibbs Free Energy''' (denoted as <math>G</math>) is a thermodynamic potential that measures the maximum reversible work a thermodynamic system can perform at constant temperature and pressure. It is an important concept in chemistry and physics, used to predict the spontaneity of chemical reactions and phase changes. === Definition === Gibbs Free Energy is defined as: <math>G = H - TS</math> where: * <math>G</math> = Gibbs free energy *...")
11 June 2025
- 11:4411:44, 11 June 2025 Convolutional Neural Network (hist | edit) [3,251 bytes] Thakshashila (talk | contribs) (Created page with "== Convolutional Neural Networks (CNNs) == A '''Convolutional Neural Network (CNN)''' is a type of deep learning model specially designed for working with '''image data''' ๐ท. CNNs are widely used in computer vision tasks like image classification, object detection, and face recognition. === ๐ง Why CNNs for Images? === Images are large (millions of pixels), and fully connected neural networks don't scale well with size. CNNs solve this by using convolution operati...")
- 11:1211:12, 11 June 2025 Backpropagation (hist | edit) [3,012 bytes] Thakshashila (talk | contribs) (Created page with "== Backpropagation == '''Backpropagation''' (short for "backward propagation of errors") is a fundamental algorithm used to train neural networks. It calculates how much each weight in the network contributed to the total error and updates them to reduce this error. === ๐ง Purpose === The main goal of backpropagation is to: * Minimize the '''loss function''' (error) ๐ * Improve model accuracy over time by adjusting weights ๐ง === ๐ How It Works (Step-by-Ste...")
- 10:0910:09, 11 June 2025 Exploding Gradient Problem (hist | edit) [2,974 bytes] Thakshashila (talk | contribs) (Created page with "== Exploding Gradient Problem == The '''Exploding Gradient Problem''' is a common issue in training deep neural networks where the gradients grow too large during backpropagation. This leads to very large weight updates, making the model unstable or completely unusable. === ๐ What Are Gradients? === Gradients are computed during the backpropagation step of training. They help the model understand how to change its weights to reduce error. :<math> \text{Gradient} =...")
- 10:0610:06, 11 June 2025 Vanishing gradient problem (hist | edit) [2,942 bytes] Thakshashila (talk | contribs) (Created page with "== Vanishing Gradient Problem == The '''Vanishing Gradient Problem''' is a common issue encountered during the training of deep neural networks. It occurs when the gradients (used to update weights) become extremely small, effectively preventing the network from learning. === ๐ง What is a Gradient? === In neural networks, gradients are values calculated during '''backpropagation'''. They show how much the model's weights should change to reduce the loss (error). The...")
- 09:0609:06, 11 June 2025 Example of ReLU Activation Function (hist | edit) [519 bytes] Thakshashila (talk | contribs) (Created page with "== ReLU (Rectified Linear Unit) Example == The ReLU function is defined as: :<math>f(x) = \max(0, x)</math> This means: * If ''x'' is '''positive''', it stays the same. * If ''x'' is '''negative''', it becomes ''0''. === Real Number Examples === {| class="wikitable" ! Input (x) ! ReLU Output f(x) |- | -3 | 0 |- | -1 | 0 |- | 0 | 0 |- | 2 | 2 |- | 5 | 5 |} In this table: * Negative numbers become 0 ๐ซ * Positive numbers pass through โ This makes ReLU very fast...")
10 June 2025
- 06:3506:35, 10 June 2025 Gradient Descent (hist | edit) [2,685 bytes] Thakshashila (talk | contribs) (Created page with "= Gradient Descent = '''Gradient Descent''' is an optimization algorithm used in machine learning and deep learning to minimize the cost (loss) function by iteratively updating model parameters in the direction of steepest descent, i.e., the negative gradient. == What is Gradient Descent? == Gradient Descent helps find the best-fit parameters (like weights in a neural network or coefficients in regression) that minimize the error between predicted and actual values. I...")
- 06:3406:34, 10 June 2025 Normalization (Machine Learning) (hist | edit) [2,564 bytes] Thakshashila (talk | contribs) (Created page with "= Normalization (Machine Learning) = '''Normalization''' in machine learning is a data preprocessing technique used to scale input features so they fall within a similar range, typically between 0 and 1. This helps improve model performance, especially for algorithms sensitive to the scale of data. == Why Normalize Data? == Some machine learning algorithms (e.g., K-Nearest Neighbors, Gradient Descent-based models, Neural Networks) perform better when input features ar...")
- 06:1106:11, 10 June 2025 Dimensionality Reduction (hist | edit) [2,413 bytes] Thakshashila (talk | contribs) (Created page with "= Dimensionality Reduction = '''Dimensionality Reduction''' is a technique in machine learning and data analysis used to reduce the number of input variables (features) while preserving as much relevant information as possible. == Why Use Dimensionality Reduction? == High-dimensional data can lead to problems such as: * '''Overfitting:''' Too many features can cause the model to learn noise. * '''Increased Computation:''' More features = more time and resources....")
- 06:1006:10, 10 June 2025 Unsupervised Learning (hist | edit) [1,970 bytes] Thakshashila (talk | contribs) (Created page with "= Unsupervised Learning = '''Unsupervised Learning''' is a type of machine learning where the model learns patterns and structures from unlabeled data without predefined outputs. == What is Unsupervised Learning? == In unsupervised learning, the input data has no associated labels. The goal is to explore the dataโs inherent structure, group similar data points, or reduce the dataโs dimensionality. == Common Types of Unsupervised Learning == * '''Clustering:''' G...")
- 06:0806:08, 10 June 2025 Clustering (hist | edit) [2,211 bytes] Thakshashila (talk | contribs) (Created page with "= Clustering = '''Clustering''' is an unsupervised machine learning technique that groups data points into clusters such that points in the same cluster are more similar to each other than to those in other clusters. == What is Clustering? == Unlike supervised learning, clustering does not use labeled data. The goal is to find natural groupings or patterns within the data based on similarity or distance measures. == Types of Clustering == * '''Partitioning Methods:'...")
- 06:0506:05, 10 June 2025 Regression (hist | edit) [2,202 bytes] Thakshashila (talk | contribs) (Created page with "= Regression = '''Regression''' is a type of supervised learning used to predict a continuous output variable based on one or more input features. == What is Regression? == In regression tasks, the goal is to model the relationship between input variables (features) and a continuous target variable. The model learns to estimate the output value for new inputs. == Types of Regression == * '''Simple Linear Regression:''' Models the relationship between a single input...")
- 06:0406:04, 10 June 2025 Supervised Learning (hist | edit) [2,167 bytes] Thakshashila (talk | contribs) (Created page with "= Supervised Learning = '''Supervised Learning''' is a type of machine learning where the model learns to map input data to output labels using a labeled dataset. == What is Supervised Learning? == In supervised learning, each training example includes both the input features and the corresponding correct output (label). The goal is for the model to learn the relationship between inputs and outputs so it can predict the labels for new, unseen data. == Types of Superv...")
- 06:0006:00, 10 June 2025 Hyperparameter Tuning (hist | edit) [2,480 bytes] Thakshashila (talk | contribs) (Created page with "= Hyperparameter Tuning = '''Hyperparameter Tuning''' is the process of optimizing the hyperparameters of a machine learning model to improve its performance on a specific task. == What are Hyperparameters? == Hyperparameters are settings or configurations external to the model that control the learning process. They are not learned from the data but set before training. Examples of hyperparameters include: * Learning rate in neural networks * Number of trees in a...")
- 05:5905:59, 10 June 2025 Bias-Variance Tradeoff (hist | edit) [2,658 bytes] Thakshashila (talk | contribs) (Created page with "= Bias-Variance Tradeoff = '''Bias-Variance Tradeoff''' is a fundamental concept in machine learning that describes the balance between two sources of error that affect model performance: bias and variance. == What is Bias? == Bias refers to the error introduced by approximating a real-world problem, which may be complex, by a simpler model. A model with high bias pays little attention to the training data and oversimplifies the problem. * High bias can cause '''unde...")
- 05:5805:58, 10 June 2025 Regularization (hist | edit) [2,440 bytes] Thakshashila (talk | contribs) (Created page with "= Regularization = '''Regularization''' is a technique in machine learning used to prevent '''overfitting''' by adding extra constraints or penalties to a model during training. == Why Regularization is Important == Overfitting happens when a model learns noise and details from the training data, harming its ability to generalize on new data. Regularization discourages overly complex models by penalizing large or unnecessary model parameters. == Common Types of Regul...")
- 05:5605:56, 10 June 2025 Train-Test Split (hist | edit) [2,316 bytes] Thakshashila (talk | contribs) (Created page with "= Train-Test Split = '''Train-Test Split''' is a fundamental technique in machine learning used to evaluate the performance of a model by dividing the dataset into two parts: a training set and a testing set. == What is Train-Test Split? == The dataset is split into: * '''Training Set:''' Used to train the machine learning model. * '''Testing Set:''' Used to evaluate how well the trained model performs on unseen data. This helps measure the modelโs ability to ge...")
- 05:5305:53, 10 June 2025 Underfitting (hist | edit) [1,937 bytes] Thakshashila (talk | contribs) (Created page with "= Underfitting = '''Underfitting''' occurs when a machine learning model is too simple to capture the underlying pattern in the data, resulting in poor performance on both training and unseen data. == What is Underfitting? == Underfitting means the model fails to learn enough from the training data. It shows high errors during training and testing because it cannot capture important trends. == Causes of Underfitting == * '''Model Too Simple:''' Using a linear model...")
- 05:5105:51, 10 June 2025 Overfitting (hist | edit) [2,515 bytes] Thakshashila (talk | contribs) (Created page with "= Overfitting = '''Overfitting''' is a common problem in machine learning where a model learns the training data too well, including its noise and outliers, resulting in poor performance on new, unseen data. == What is Overfitting? == When a model is overfitted, it captures not only the underlying pattern but also the random fluctuations or noise in the training dataset. This causes the model to perform excellently on training data but badly on test or real-world data...")
- 05:4605:46, 10 June 2025 Evaluation Metrics (hist | edit) [3,123 bytes] Thakshashila (talk | contribs) (Created page with "= Evaluation Metrics = '''Evaluation Metrics''' are quantitative measures used to assess the performance of machine learning models. Choosing the right metric is essential for understanding how well a model performs, especially in classification and regression problems. == Why Are Evaluation Metrics Important? == * Provide objective criteria to compare different models. * Help detect issues like overfitting or underfitting. * Guide model improvement and selection. * R...")
- 05:4505:45, 10 June 2025 Cost-Sensitive Learning (hist | edit) [2,729 bytes] Thakshashila (talk | contribs) (Created page with "= Cost-Sensitive Learning = '''Cost-Sensitive Learning''' is a machine learning approach that incorporates different costs for different types of classification errors, helping models make better decisions in situations where misclassification errors have unequal consequences. == Why Cost-Sensitive Learning? == In many real-world problems, different mistakes have different costs. For example: * In medical diagnosis, a false negative (missing a disease) may be more co...")
- 05:4405:44, 10 June 2025 Area Under Precision-Recall Curve (AUPRC) (hist | edit) [2,331 bytes] Thakshashila (talk | contribs) (Created page with "= Area Under Precision-Recall Curve (AUPRC) = The '''Area Under the Precision-Recall Curve''' ('''AUPRC''') is a single scalar value that summarizes the performance of a binary classification model by measuring the area under its Precision-Recall (PR) curve. == What is the Precision-Recall Curve? == The Precision-Recall Curve plots: * '''Precision''' (y-axis): the proportion of true positive predictions among all positive predictions. * '''Recall''' (x-axis): the pro...")
- 05:4005:40, 10 June 2025 Imbalanced Data (hist | edit) [2,634 bytes] Thakshashila (talk | contribs) (Created page with "= Imbalanced Data = '''Imbalanced Data''' refers to datasets where the classes are not represented equally. In classification problems, one class (usually the positive or minority class) has far fewer examples than the other class (negative or majority class). == Why is Imbalanced Data a Problem? == Machine learning models often assume that classes are balanced and try to maximize overall accuracy. When data is imbalanced, models tend to be biased toward the majority...")
- 05:3605:36, 10 June 2025 Cross Validation (hist | edit) [2,746 bytes] Thakshashila (talk | contribs) (Created page with "= Cross-Validation = '''Cross-Validation''' is a statistical method used to estimate the performance of machine learning models on unseen data. It helps ensure that the model generalizes well and reduces the risk of overfitting. == Why Cross-Validation? == When training a model, it is important to test how well it performs on data it has never seen before. Simply evaluating a model on the same data it was trained on can lead to overly optimistic results. Cross-validat...")
- 05:3505:35, 10 June 2025 Model Selection (hist | edit) [3,186 bytes] Thakshashila (talk | contribs) (Created page with "= Model Selection = '''Model Selection''' is the process of choosing the best machine learning model from a set of candidate models based on their performance on a given task. It is a critical step to ensure the selected model generalizes well to new, unseen data. == Why Model Selection is Important == Different algorithms and model configurations may perform differently depending on the dataset and problem. Selecting the right model helps: * Improve prediction accur...")
- 05:3405:34, 10 June 2025 Threshold Tuning (hist | edit) [2,652 bytes] Thakshashila (talk | contribs) (Created page with "= Threshold Tuning = '''Threshold Tuning''' is the process of selecting the best decision threshold in a classification model to optimize performance metrics such as Precision, Recall, F1 Score, or Accuracy. It is crucial in models that output '''probabilities''' rather than direct class labels. == Why Threshold Tuning Matters == Many classifiers (e.g., Logistic Regression, Neural Networks) output a probability score indicating how likely an instance b...")
- 05:3305:33, 10 June 2025 AUC Score (hist | edit) [2,689 bytes] Thakshashila (talk | contribs) (Created page with "= AUC Score (Area Under the Curve) = The '''AUC Score''' refers to the '''Area Under the Curve''' and is a popular metric used to evaluate the performance of classification models, especially in binary classification tasks. Most commonly, AUC represents the area under the ROC Curve (Receiver Operating Characteristic Curve) or under the Precision-Recall Curve (PR Curve). == What is AUC? == AUC measures the ability of a model to distinguish between positive and...")
- 05:3205:32, 10 June 2025 Precision-Recall Curve (hist | edit) [3,086 bytes] Thakshashila (talk | contribs) (Created page with "= Precision-Recall Curve = The '''Precision-Recall Curve''' (PR Curve) is a graphical representation used to evaluate the performance of binary classification models, especially on '''imbalanced datasets''' where the positive class is rare. It plots '''Precision''' (y-axis) against '''Recall''' (x-axis) for different classification thresholds. == Why Use Precision-Recall Curve? == In many real-world problems like fraud detection, disease diagnosis, or spam filtering,...")
- 05:3005:30, 10 June 2025 Model Evaluation Metrics (hist | edit) [3,410 bytes] Thakshashila (talk | contribs) (Created page with "= Model Evaluation Metrics = '''Model Evaluation Metrics''' are quantitative measures used to assess how well a machine learning model performs. They help determine the accuracy, reliability, and usefulness of models in solving real-world problems. == Importance of Evaluation Metrics == Without evaluation metrics, it's impossible to know whether a model is effective or not. Metrics guide model selection, tuning, and deployment by measuring: * Accuracy of predictions...")
- 05:2905:29, 10 June 2025 Classification (hist | edit) [3,130 bytes] Thakshashila (talk | contribs) (Created page with "= Classification = '''Classification''' is a fundamental task in '''machine learning''' and '''data science''' where the goal is to predict discrete labels (categories) for given input data. It is a type of '''supervised learning''' since the model learns from labeled examples. == What is Classification? == In classification, a model is trained on a dataset with input features and known target classes. Once trained, the model can assign class labels to new, unseen dat...")
- 05:2705:27, 10 June 2025 ROC Curve (hist | edit) [2,678 bytes] Thakshashila (talk | contribs) (Created page with "= ROC Curve = The '''ROC Curve''' ('''Receiver Operating Characteristic Curve''') is a graphical tool used to evaluate the performance of binary classification models. It plots the '''True Positive Rate (TPR)''' against the '''False Positive Rate (FPR)''' at various threshold settings. == Purpose == The ROC Curve shows the trade-off between sensitivity (recall) and specificity. It helps assess how well a classifier can distinguish between two classes. == Definitions...")
- 05:2605:26, 10 June 2025 Micro F1 Score (hist | edit) [2,700 bytes] Thakshashila (talk | contribs) (Created page with "= Micro F1 Score = The '''Micro F1 Score''' is an evaluation metric used primarily in '''multi-class''' and '''multi-label classification''' tasks. Unlike Macro F1 Score, it calculates global counts of true positives, false positives, and false negatives across all classes, then uses these to compute a single Precision, Recall, and F1 Score. It is most useful when the dataset is '''imbalanced''' and you care more about overall performance than per-class fai...")
- 05:2505:25, 10 June 2025 Weighted F1 (hist | edit) [2,730 bytes] Thakshashila (talk | contribs) (Created page with "= Weighted F1 Score = The '''Weighted F1 Score''' is a metric used in multi-class classification to evaluate model performance by computing the F1 Score for each class and taking the average, weighted by the number of true instances for each class (i.e., the class "support"). It is especially useful when working with '''imbalanced datasets''', where some classes are more frequent than others. == Definition == :<math> \text{Weighted F1} = \sum_{i=1}^{C} w_i \cdot F1_i...")
- 05:2405:24, 10 June 2025 Macro F1 (hist | edit) [2,412 bytes] Thakshashila (talk | contribs) (Created page with "= Macro F1 Score = The '''Macro F1 Score''' is an evaluation metric used in multi-class classification tasks. It calculates the F1 Score independently for each class and then takes the average (unweighted) across all classes. Unlike the regular F1 Score, which is typically applied to binary classification, the Macro F1 is designed for problems involving more than two classes. == Definition == 1. Compute Precision and Recall for each class individually 2. Compute...")
- 05:2305:23, 10 June 2025 Complementary metrics (hist | edit) [3,031 bytes] Thakshashila (talk | contribs) (Created page with "= Complementary Metrics in Machine Learning = '''Complementary Metrics''' refer to pairs or groups of evaluation metrics that together provide a more complete and balanced understanding of a classification modelโs performance. Because no single metric is perfect, especially in real-world and imbalanced datasets, these metrics are used together to highlight different strengths and weaknesses of a model. == Why Use Complementary Metrics? == Using only one metric like...")
- 05:2205:22, 10 June 2025 F1 Score (hist | edit) [2,460 bytes] Thakshashila (talk | contribs) (Created page with "= F1 Score = The '''F1 Score''' is a performance metric used in classification problems that balances the trade-off between Precision and Recall (also known as Sensitivity). It is especially useful when the dataset is imbalanced, and both false positives and false negatives are important. == Definition == The F1 Score is the '''harmonic mean''' of Precision and Recall. :<math> F1 = 2 \cdot \frac{\text{Precision} \cdot \text{Recall}}{\text{Precision} + \te...")
- 05:2105:21, 10 June 2025 Specificity (hist | edit) [2,494 bytes] Thakshashila (talk | contribs) (Created page with "= Specificity = '''Specificity''', also known as the '''True Negative Rate (TNR)''', is a performance metric in binary classification tasks. It measures the proportion of actual negative instances that are correctly identified by the model. == Definition == :<math> \text{Specificity} = \frac{TN}{TN + FP} </math> Where: * '''TN''' = True Negatives โ actual negatives correctly predicted * '''FP''' = False Positives โ actual negatives incorrectly predicted as positi...")
- 05:2005:20, 10 June 2025 Sensitivity (hist | edit) [2,293 bytes] Thakshashila (talk | contribs) (Created page with "= Sensitivity = '''Sensitivity''', also known as '''Recall''' or the '''True Positive Rate (TPR)''', is a performance metric used in classification problems. It measures how well a model can identify actual positive instances. == Definition == :<math> \text{Sensitivity} = \frac{TP}{TP + FN} </math> Where: * '''TP''' = True Positives โ actual positives correctly predicted * '''FN''' = False Negatives โ actual positives incorrectly predicted as negative Sensitivit...")
- 05:2005:20, 10 June 2025 Accuracy (hist | edit) [2,110 bytes] Thakshashila (talk | contribs) (Created page with "= Accuracy = '''Accuracy''' is one of the most commonly used metrics to evaluate the performance of a classification model in machine learning. It tells us the proportion of total predictions that were correct. == Definition == :<math> \text{Accuracy} = \frac{TP + TN}{TP + TN + FP + FN} </math> Where: * '''TP''' = True Positives * '''TN''' = True Negatives * '''FP''' = False Positives * '''FN''' = False Negatives Accuracy answers the question: '''"Out of all predict...")
- 05:1805:18, 10 June 2025 Recall (hist | edit) [1,738 bytes] Thakshashila (talk | contribs) (Created page with "= Recall = '''Recall''' is a metric used in classification to measure how many of the actual positive instances were correctly identified by the model. It is also known as '''sensitivity''' or the '''true positive rate'''. == Definition == :<math> \text{Recall} = \frac{TP}{TP + FN} </math> Where: * '''TP''' = True Positives โ correctly predicted positive instances * '''FN''' = False Negatives โ actual positives incorrectly predicted as negative Recall answers th...")
- 05:1805:18, 10 June 2025 Precision (hist | edit) [1,617 bytes] Thakshashila (talk | contribs) (Created page with "= Precision = '''Precision''' is a metric used in classification tasks to measure how many of the predicted positive results are actually correct. It is also known as the '''positive predictive value'''. == Definition == :<math> \text{Precision} = \frac{TP}{TP + FP} </math> Where: * '''TP''' = True Positives โ correct positive predictions * '''FP''' = False Positives โ incorrect positive predictions Precision helps to answer the question: '''"Of all the items la...")
- 05:1305:13, 10 June 2025 Confusion Matrix (hist | edit) [2,740 bytes] Thakshashila (talk | contribs) (Created page with "= Confusion Matrix = '''Confusion Matrix''' is a performance measurement tool used in machine learning, particularly for classification problems. It provides a summary of prediction results on a classification problem by comparing the actual labels with those predicted by the model. == What is a Confusion Matrix? == A confusion matrix is a table that describes the performance of a classification model. It shows how many instances were correctly or incorrectly predicte...")
5 June 2025
- 04:2204:22, 5 June 2025 Neural Network (hist | edit) [3,999 bytes] Thakshashila (talk | contribs) (Created page with "= Neural Network = '''Neural Networks''' are a class of algorithms within Machine Learning and Deep Learning that are designed to recognize patterns. They are inspired by the structure and function of the biological brain and are used to approximate complex functions by learning from data. == Overview == A neural network consists of interconnected units (called '''neurons''' or '''nodes''') organized in layers. These layers process input data through weighted c...")
- 04:2104:21, 5 June 2025 Data Science (hist | edit) [3,648 bytes] Thakshashila (talk | contribs) (Created page with "= Data Science = '''Data Science''' is an interdisciplinary field that uses scientific methods, algorithms, and systems to extract knowledge and insights from structured and unstructured data. It integrates techniques from statistics, computer science, and domain-specific knowledge to turn raw data into actionable intelligence. == Overview == Data Science combines aspects of data analysis, machine learning, data engineering, and software development to address complex...")
- 04:2004:20, 5 June 2025 Deep Learning (hist | edit) [3,701 bytes] Thakshashila (talk | contribs) (Created page with "= Deep Learning = '''Deep Learning''' is a subfield of Machine Learning concerned with algorithms inspired by the structure and function of the brain, known as artificial neural networks. It is at the heart of many recent advances in Artificial Intelligence. == Overview == Deep learning models automatically learn representations of data through multiple layers of abstraction. These models excel at recognizing patterns in unstructured data such as images, audio,...")
- 04:2004:20, 5 June 2025 Artificial Intelligence (hist | edit) [3,871 bytes] Thakshashila (talk | contribs) (Created page with "= Artificial Intelligence = '''Artificial Intelligence (AI)''' is a branch of computer science that aims to create systems or machines that exhibit behavior typically requiring human intelligence. These behaviors include learning, reasoning, problem-solving, perception, language understanding, and decision-making. == Overview == Artificial Intelligence involves the design and development of algorithms that allow computers and software to perform tasks that would normal...")