skip to content
luminary.blog
by Oz Akan
bird

What are Features in Machine Learning?

Choosing the right features is crucial for building an accurate and efficient model.

/ 4 min read

Table of Contents

Introduction

In machine learning, features are the measurable properties or characteristics of the data that are used as input for a model. Choosing the right features is crucial for building an accurate and efficient model. This guide explores features from basic to advanced levels with examples to help you understand their significance in different contexts.

1. What Are Features in Machine Learning?

A feature is an individual measurable property of a dataset. Features are used as inputs to train a machine learning model. The performance of a model heavily depends on the quality and relevance of features.

Example of Features in a Dataset

Consider a dataset predicting house prices. The dataset might have the following features:

FeatureDescription
Size (sq. ft)The area of the house
Number of BedroomsTotal bedrooms in the house
Number of BathroomsTotal bathrooms in the house
LocationNeighborhood or zip code
Year BuiltThe year the house was constructed
Price (Target)The house price (Output Variable)

In this example, all the columns except Price are features that help predict the price of a house.

2. Types of Features

a. Numerical Features (Continuous & Discrete)

Numerical features are represented by numbers and can be either:

  • Continuous: Values that can take any range (e.g., height, weight, temperature).
  • Discrete: Values that take only specific numbers (e.g., number of rooms in a house).

Example:

  • Continuous: Age = 25.3 years
  • Discrete: Number of children = 2

b. Categorical Features

Categorical features represent distinct groups or categories. They can be:

  • Nominal (No natural order) – e.g., Gender (Male, Female, Other), Car Brand (Toyota, Ford, Tesla).
  • Ordinal (Has a natural order) – e.g., Education Level (High School, Bachelor’s, Master’s, PhD).

Example:

  • Nominal: Color = Red
  • Ordinal: Customer Satisfaction = High

c. Boolean (Binary) Features

Binary features take only two values, usually 0 or 1.

Example:

  • Has Pets? (Yes=1, No=0)
  • Customer Churned? (Yes=1, No=0)

d. Text Features

Text data can be used as a feature in Natural Language Processing (NLP).

Example:

  • Product Review: “This phone has excellent battery life.”

Text features require conversion into numerical form using techniques like TF-IDF or word embeddings.

e. Temporal Features These features capture time-related information.

Example:

  • Timestamp: 2024-02-28 14:30:00
  • Day of the Week: Monday
  • Elapsed Time Since Last Purchase: 30 days

f. Image and Audio Features

For computer vision and audio processing, features can be:

  • Pixel values (images)
  • Spectrograms or MFCCs (audio)

3. Feature Engineering

Feature engineering is the process of creating new features or modifying existing ones to improve model performance.

  • Feature Creation: Generating new features from raw data (e.g., calculating the ratio of two variables).
  • Feature Transformation: Modifying existing features to improve their usability (e.g., normalizing numerical values).
  • Feature Selection: Choosing the most relevant features for a task to reduce dimensionality and computational cost.

a. Feature Scaling

Scaling numerical features ensures they are within the same range.

  • Min-Max Scaling: Rescales values between 0 and 1.
  • Standardization (Z-score Normalization): Converts values to a distribution with mean 0 and standard deviation 1.

b. One-Hot Encoding (OHE)

Used for categorical variables where each category is converted into a separate binary column.

Example:

ColorRedBlueGreen
Red100
Blue010
Green001

c. Feature Extraction

Creating new meaningful features from raw data.

  • From dates: Extracting Year, Month, Day of the Week from a timestamp.
  • From text: Counting the number of words or sentiment analysis scores.
  • From images: Extracting edges or colors using Convolutional Neural Networks (CNNs).

d. Feature Selection

Feature selection reduces the number of irrelevant or redundant features.

  • Filter Methods: Using correlation or statistical tests.
  • Wrapper Methods: Using model performance to select features.
  • Embedded Methods: Feature selection happens during training (e.g., Lasso Regression, Decision Trees).

4. Advanced Topics

a. Feature Engineering for Time Series Data

  • Lag Features: Using past values as features (Sales at t-1 to predict Sales at t).
  • Rolling Statistics: Creating moving averages (e.g., 7-day rolling mean).

b. Feature Representation for Deep Learning

  • Embeddings: Representing categorical variables in lower-dimensional continuous space (e.g., word embeddings in NLP).
  • Autoencoders: Unsupervised feature extraction for deep learning models.

c. Synthetic Features

Creating new features that don’t exist in raw data.

Example:

  • Polynomial Features: x², x³ for non-linear models.
  • Interaction Features: Height × Weight to improve predictions.

5. Conclusion

Features are the foundation of machine learning models. Selecting and engineering the right features can significantly improve performance. From basic numerical and categorical features to advanced feature engineering techniques, understanding how to transform raw data into meaningful inputs is a crucial skill for any machine learning practitioner.