Portfolio Report:
Sensor-Based Failure Analysis and Predictive Modeling
Paulo Protachevicz
protachevicz@gmail.com
github.com/Protachevicz

1. Exploratory Analysis of Sensor Variables

This project illustrates how I applied statistical and machine learning techniques to analyze sensor data and anticipate system failures in operational cycles. I started by analyzing key sensor distributions such as Temperature, Pressure, VibrationX, VibrationY, VibrationZ, and Frequency.

image

Distributions were compared between normal operation cycles and those preceding failure events. This type of exploratory data analysis is essential to identify signals that are potentially predictive of anomalies or equipment degradation.

Highlights:

Takeaway: This stage provides the foundation for effective feature selection and modeling by identifying features that correlate with future failure events.

2. Failure Rate by Operational Configuration

Sensor cycles were grouped by configuration variables (Preset_1 and Preset_2), and their associated failure rates were computed. This approach helps identify configurations more prone to instabilities.

image

The heatmap highlights critical regions in the operational space. These findings support strategic intervention such as reconfiguring default parameters or adding redundancy in high-risk zones.

3. Classification of Failure Causes

Using threshold-based rules, failures were categorized into:

image

image

Insight: Different physical stressors contribute to failures, reinforcing the need for tailored diagnostics and multi-sensor fusion.

4. Predictive Modeling of Failure Events

The goal was to anticipate failures using classification models trained on normal cycles. I evaluated:

Models were validated via cross-validation and hold-out test sets (15 positive, 15 negative).

Logistic Regression:

Gradient Boosting:

Conclusion: Logistic Regression demonstrated superior generalization and reliability in high-recall scenarios — critical in predictive maintenance.

5. Feature Importance and Explainability (SHAP)

Feature importance was derived using:

  1. Gradient Boosting’s feature_importances_

  2. Absolute coefficients from Logistic Regression

  3. SHAP values for model interpretability

image

image

image

image

Takeaways:

Final Remarks

This project highlights my ability to:

This workflow is applicable across multiple domains: industrial monitoring, IoT systems, anomaly detection, healthcare devices, and beyond.