Project 002 // Computer Vision + Haptics // Provisional Patent
Remote Identification and Detection of Genital Skin Cancer — a multi-modal diagnostic system combining CNN-based visual classification with a custom non-Newtonian haptic sensor array to replicate the tactile and visual information of a clinical skin lesion examination, remotely.
// 01 — Problem Statement
Skin cancer is the most common malignancy globally, with melanoma responsible for the majority of skin cancer deaths despite representing only ~1% of cases. The critical factor in survival is early detection — 5-year survival for localized melanoma exceeds 98%; for metastatic disease it drops below 30%. Routine screening appointments are therefore high-stakes interventions, not optional wellness visits.
Dermatological diagnosis depends on two parallel information channels: visual appearance (color, border irregularity, asymmetry, diameter — the clinical ABCDE criteria) and tactile feedback (lesion elevation, firmness, texture variability, raised margins). An experienced dermatologist palpates suspicious lesions — the mechanical properties of a melanoma feel meaningfully different from a benign nevus, particularly at the margins.
COVID-19 eliminated in-person dermatology appointments in 2020. Telehealth platforms offered visual consultation via photograph — but photographs are fundamentally incapable of transmitting tactile information. The haptic diagnostic channel was severed entirely. Dermatologists were making triage decisions with half the clinical data they normally use, and patients with suspicious lesions had no viable alternative.
The Core Hypothesis
If we can capture the mechanical signature of a skin lesion — its topographic pressure distribution, firmness profile, and edge compliance — and transmit it alongside visual data, we can give remote dermatologists a clinically equivalent information set to an in-person exam.
Diagnostic Information Channels
// 02 — System Design
The core hardware innovation in RIDGE is a custom 40×40mm pressure sensor array designed to capture the continuous topographic pressure profile of a skin lesion when the array is pressed against the skin surface. The critical material choice — a non-Newtonian pressure-sensitive medium — enables continuous spatial pressure distribution rather than discrete point measurements.
Component Breakdown
Visual System
Haptic System
// 03 — Visual Classification Pipeline
The visual classification pipeline processes a photograph of the lesion through a sequence of preprocessing, feature extraction, and classification stages. Each stage was designed to handle the specific degradation modes of clinical photography (inconsistent lighting, varying resolution, skin tone variation) rather than the controlled conditions of benchmark datasets.
Public dermatology datasets have two critical failure modes for clinical use: severe class imbalance (benign cases outnumber malignant by 20:1 or more) and inconsistent labeling standards across source institutions. Training on uncurated public data produces models with high accuracy on benchmark tests that fail on clinical populations. RIDGE used a manually curated dataset with dermatologist-reviewed labels.
Class Distribution — Curated Dataset
Curation Methodology
Source aggregation: ISIC Archive, HAM10000, and DermNet — reviewed for labeling consistency before inclusion.
Dermatologist review: Ambiguous labels flagged and reclassified by a practicing dermatologist. ~12% of initially included images were rejected or reclassified.
Targeted augmentation: Rotation, color jitter, zoom applied exclusively to underrepresented malignant classes — not applied to benign classes to avoid artificially inflating apparent class balance.
Skin tone stratification: ITA angle distribution checked to ensure no demographic bias. Dataset spans Fitzpatrick scale I–VI.
// 04 — Multi-Modal Fusion
The central architectural challenge in RIDGE is combining two fundamentally different data representations — spatial image tensors from the visual system and pressure matrix time-series from the haptic array — into a single classification decision. Early fusion (concatenating raw inputs) and mid fusion (combining at an intermediate feature layer) both produced worse results than the final late fusion architecture.
Late Fusion Architecture
Key Finding
Haptic features added 4–6% accuracy improvement on borderline cases (dysplastic nevus vs. early melanoma) compared to visual-only classification. These are precisely the cases where an in-person dermatologist would palpate the lesion — the system learned the same diagnostic heuristic from data.
// 05 — Technical Challenges
Problem
Skin lesions have continuous topographic texture variation — a melanoma has raised, irregular margins with variable firmness that grades into normal tissue over millimeters. Discrete FSR point sensors measure pressure at fixed locations and cannot capture this continuous gradient directly. Interpolation between discrete points produces smooth pressure maps that erase exactly the edge irregularity that is diagnostically significant.
Solution
Non-Newtonian pressure-sensitive material as the intermediate sensing layer. When pressed against a lesion, the material deforms in direct proportion to the local surface topography — the continuous skin surface topography is encoded in the continuous deformation of the material, which the sensor array then reads. The material acts as an analog spatial integrator between the continuous skin surface and the discrete sensor array. This remains the most imperfect element of RIDGE — material consistency between applications and temperature sensitivity are unsolved challenges that I want to address in future work.
Problem
Standard public dermatology datasets (ISIC 2019, HAM10000) produce high benchmark accuracy but poor clinical performance — because they contain images taken under controlled photographic conditions by trained medical photographers, not under the variable conditions of telehealth photography. A model trained on these datasets learns features that are valid for benchmark images but brittle to real-world photo quality variation.
Solution
Two-stage curation: first, source aggregation with cross-dataset label consistency checking; second, deliberate addition of degraded-condition images (motion blur, variable lighting, low resolution, partial occlusion) to the training set to improve robustness. Also applied color normalization via Macenko stain normalization (adapted from histopathology) to reduce skin tone bias. The result was a 3% drop in benchmark accuracy but a measurably more robust model on held-out real-world telehealth images reviewed by the clinical advisor.
Problem
A softmax output of 0.87 for "melanoma" from a neural network does not mean 87% probability of melanoma in any calibrated sense — it means the model's output neuron fired at 0.87, which is only meaningful relative to the training distribution. Presenting this as a probability to a dermatologist making a triage decision is potentially dangerous overconfidence, particularly on out-of-distribution inputs.
Solution
Temperature scaling post-hoc calibration on a validation set, which rescales softmax outputs to better match empirical frequencies. Additionally, the output report distinguishes between high-confidence classifications (top class probability >0.85, clear visual features) and low-confidence classifications (top probability <0.6, borderline cases) with explicit recommendations: high-confidence benign = routine monitoring, low-confidence any class = in-person biopsy referral regardless of the visual prediction. The system's most useful function in ambiguous cases is to route to a human expert, not to replace one.
// 06 — Results
Validation Note
Accuracy validated on a held-out test set of 1,200 images not seen during training. Clinical review by 2 dermatologists and 1 radiologist confirmed that the system's borderline case handling — routing ambiguous predictions to in-person referral — represented a clinically sound triage behavior. The false negative rate on malignant lesions in the high-confidence classification tier was 2.1%.
"Visual classification was a tractable ML problem once I had good data. Replicating the physical texture of a skin lesion — that required physics, materials science, and signal processing simultaneously, and I still don't think I fully solved it."
// 07 — Key Learnings
Seeking full-time roles in robotics, computer vision, or ML engineering — preferably medical technology. Graduating May 2026.
Get in Touch ← All Projects