Viqus Logo Viqus Logo
Home
Categories
Language Models Generative Imagery Hardware & Chips Business & Funding Ethics & Society Science & Robotics
Resources
AI Glossary Academy CLI Tool Labs
About Contact
Back to Glossary
Machine Learning Intermediate Also: Feature Extraction, Feature Construction

Feature Engineering

Definition

The process of selecting, transforming, or creating input variables (features) from raw data to make machine learning algorithms work more effectively on a given problem.

In Depth

Feature Engineering is the art and science of shaping raw data into a form that machine learning algorithms can learn from effectively. Raw data rarely arrives in a machine-learnable format: dates need to be decomposed, categories need to be encoded numerically, outliers need to be handled, and domain-specific combinations of variables often carry more signal than the raw variables themselves.

Common feature engineering techniques include normalization and standardization (scaling numerical features to comparable ranges); one-hot encoding (converting categorical variables to binary columns); creating interaction terms (multiplying two variables together to capture their joint effect); log transformations (compressing skewed distributions); and temporal features (extracting day-of-week, time-since-event, or rolling averages from time series).

The impact of Feature Engineering on model performance can be dramatic — often larger than switching from one algorithm to another. In many real-world ML projects, 70-80% of time is spent on data collection and Feature Engineering rather than model development. The rise of deep learning has partially automated feature learning for unstructured data (images, text), but Feature Engineering remains essential for structured/tabular data, which dominates enterprise ML applications.

Key Takeaway

Feature Engineering is what separates a model that works from one that works well — the quality of input features often matters more than the sophistication of the algorithm applied to them.

Real-World Applications

01 E-commerce recommendation: creating 'days since last purchase', 'average order value', and 'preferred category' features from raw transaction logs.
02 Credit risk modeling: combining income, debt-to-income ratio, and payment history velocity into higher-signal features for loan default prediction.
03 Sensor data analytics: extracting statistical features (mean, variance, peak frequency) from raw time-series sensor readings for anomaly detection.
04 Text classification: creating TF-IDF features, character n-grams, or named entity counts from raw text before feeding to classical ML models.
05 Sports analytics: engineering possession efficiency metrics, shot quality scores, and momentum indicators from raw game event data.

Frequently Asked Questions

What is the difference between feature engineering and feature selection?

Feature engineering creates new features from raw data (e.g., extracting 'day of week' from a timestamp). Feature selection chooses the most informative subset from existing features (e.g., removing redundant or low-value columns). Both aim to improve model performance, but engineering adds information while selection removes noise.

Is feature engineering still important with Deep Learning?

Deep Learning reduces the need for manual feature engineering because neural networks can learn useful representations from raw data. However, feature engineering remains valuable — even for deep models, well-crafted input features can speed training, improve accuracy, and reduce data requirements. For tabular data, classical ML with good feature engineering often outperforms deep learning.

What are common feature engineering techniques?

Common techniques include: one-hot encoding (converting categories to binary vectors), normalization/standardization (scaling numerical features), polynomial features (capturing non-linear relationships), temporal features (extracting hour, day, season from timestamps), text vectorization (TF-IDF, word embeddings), and domain-specific transformations (log transforms, interaction terms).