Python/machine_learning/ridge_regression/model.py

83 lines
2.6 KiB
Python
Raw Normal View History

2024-10-23 13:57:43 +00:00
import numpy as np
import pandas as pd
2024-10-23 14:07:10 +00:00
2024-10-23 13:57:43 +00:00
class RidgeRegression:
2024-10-23 15:31:03 +00:00
def __init__(self,
alpha: float = 0.001,
regularization_param: float = 0.1,
num_iterations: int = 1000,
) -> None:
self.alpha: float = alpha
self.regularization_param: float = regularization_param
self.num_iterations: int = num_iterations
self.theta: np.ndarray = None
def feature_scaling(
2024-10-23 15:31:03 +00:00
self, x: np.ndarray
) -> tuple[np.ndarray, np.ndarray, np.ndarray]:
2024-10-23 15:31:03 +00:00
mean = np.mean(x, axis=0)
std = np.std(x, axis=0)
2024-10-23 13:57:43 +00:00
# avoid division by zero for constant features (std = 0)
std[std == 0] = 1 # set std=1 for constant features to avoid NaN
2024-10-23 15:21:58 +00:00
x_scaled = (x - mean) / std
return x_scaled, mean, std
2024-10-23 14:07:10 +00:00
2024-10-23 15:25:20 +00:00
def fit(self, x: np.ndarray, y: np.ndarray) -> None:
2024-10-23 15:21:58 +00:00
x_scaled, mean, std = self.feature_scaling(x)
m, n = x_scaled.shape
2024-10-23 13:57:43 +00:00
self.theta = np.zeros(n) # initializing weights to zeros
2024-10-23 15:31:03 +00:00
for _ in range(self.num_iterations):
2024-10-23 15:21:58 +00:00
predictions = x_scaled.dot(self.theta)
2024-10-23 13:57:43 +00:00
error = predictions - y
2024-10-23 13:57:43 +00:00
# computing gradient with L2 regularization
gradient = (
2024-10-23 15:21:58 +00:00
x_scaled.T.dot(error) + self.regularization_param * self.theta
) / m
2024-10-23 13:57:43 +00:00
self.theta -= self.alpha * gradient # updating weights
2024-10-23 15:31:03 +00:00
def predict(self, x: np.ndarray) -> np.ndarray:
x_scaled, _, _ = self.feature_scaling(x)
return x_scaled.dot(self.theta)
2024-10-23 15:14:04 +00:00
2024-10-23 15:25:20 +00:00
def compute_cost(self, x: np.ndarray, y: np.ndarray) -> float:
2024-10-23 15:21:58 +00:00
x_scaled, _, _ = self.feature_scaling(x)
2024-10-23 13:57:43 +00:00
m = len(y)
2024-10-23 15:21:58 +00:00
predictions = x_scaled.dot(self.theta)
2024-10-23 15:31:03 +00:00
cost = (
1 / (2 * m)) * np.sum((predictions - y) ** 2) + (
self.regularization_param / (2 * m)
) * np.sum(self.theta**2)
2024-10-23 13:57:43 +00:00
return cost
2024-10-23 14:07:10 +00:00
def mean_absolute_error(self, y_true: np.ndarray, y_pred: np.ndarray) -> float:
2024-10-23 13:57:43 +00:00
return np.mean(np.abs(y_true - y_pred))
2024-10-23 14:07:10 +00:00
2024-10-23 13:57:43 +00:00
# Example usage
if __name__ == "__main__":
2024-10-23 15:31:03 +00:00
data = pd.read_csv("ADRvsRating.csv")
x = data[["Rating"]].to_numpy()
y = data["ADR"].to_numpy()
2024-10-23 13:57:43 +00:00
y = (y - np.mean(y)) / np.std(y)
2024-10-23 14:17:34 +00:00
# added bias term to the feature matrix
2024-10-23 15:25:20 +00:00
x = np.c_[np.ones(x.shape[0]), x]
2024-10-23 13:57:43 +00:00
2024-10-23 14:17:34 +00:00
# initialize and train the ridge regression model
2024-10-23 14:07:10 +00:00
model = RidgeRegression(alpha=0.01, regularization_param=0.1, num_iterations=1000)
2024-10-23 15:21:58 +00:00
model.fit(x, y)
2024-10-23 13:57:43 +00:00
# predictions
2024-10-23 15:21:58 +00:00
predictions = model.predict(x)
2024-10-23 13:57:43 +00:00
# results
print("Optimized Weights:", model.theta)
2024-10-23 15:21:58 +00:00
print("Cost:", model.compute_cost(x, y))
print("Mean Absolute Error:", model.mean_absolute_error(y, predictions))