Fetal Compromise Detection Using CTG Data: Can Deep Learning Save the Day?

Okay, folks, buckle up because we’re diving deep—no, like REALLY deep—into the world of fetal monitoring during labor. And no, I’m not talking about those handheld Doppler devices your grandma used to freak out about. We’re talking cutting-edge tech here: deep learning algorithms analyzing Cardiotocography (CTG) data. Think of it as giving doctors superhuman hearing for babies in the womb.

Why is this a big deal, you ask? Well, imagine being able to predict potential problems before they become serious, potentially saving tiny lives. That’s what this whole shebang is about—using fancy computer models to make childbirth safer for everyone involved.

Data, Data Everywhere: Where Do We Even Begin?

First things first, you can’t train a computer model without feeding it a whole lotta data. That’s where the CTU-UHB Intrapartum Cardiotocography Database comes in, a goldmine of info collected between two thousand ten and two thousand twelve. This database has a whopping five hundred and fifty-two CTG recordings, each tracking two crucial things: Fetal Heart Rate (FHR) and Uterine Contractions (UC).

But wait, there’s more! We’re not just throwing raw data at the computer and hoping for the best. Oh no, we gotta be way more strategic than that.

Cleaning Up the Data: It’s Like Organizing Your Sock Drawer, But for Science

Imagine trying to bake a cake with a recipe written in a foreign language. That’s kinda what raw CTG data is like—a jumbled mess that needs some serious translation before it makes any sense. That’s where preprocessing comes in, the less glamorous but oh-so-crucial step in our deep learning journey.

  • Artifact Removal: First, we gotta get rid of all the junk—think of it like filtering out spam emails. Outliers, those weird blips in the data that make no physiological sense, get the boot and are replaced with zeros.
  • Signal Interpolation: Next up, we gotta fill in the blanks. Small gaps in the data, less than fifteen seconds, are stitched together using linear interpolation. It’s like using Photoshop to fix a tiny tear in a precious photo.
  • Downsampling: Finally, we gotta simplify things a bit. The recordings are downsampled to .25 Hz, making the data more manageable for our hungry algorithms to digest. It’s like compressing a giant image file without losing too much resolution.

Introducing FHR-LINet: The Superhero of Fetal Monitoring (Well, Almost)

Now for the main event—the deep learning model itself! Drumroll please… Introducing FHR-LINet, a convolutional neural network (CNN) with a very important job: analyzing FHR signals to predict fetal compromise. This ain’t your grandma’s CNN either, folks. This model is “input length invariant,” meaning it can handle FHR signals of any length. Talk about talent!

The Architecture: It’s Like Building a High-Tech Baby Monitor

Imagine FHR-LINet as a sophisticated machine with multiple layers, each performing a specific task.

  • Multi-scale convolutional layers: These layers are like the eyes of the model, scanning the FHR signal for patterns and anomalies.
  • Max pooling: This layer acts like a filter, picking out the most important features from the convolutional layers.
  • Global average pooling: This layer helps to prevent overfitting, ensuring the model can generalize well to new data.
  • Fully connected layers: These layers connect all the previous layers together and make the final prediction about fetal compromise.

And to make sure our model doesn’t get too big for its britches (aka overfitting), we throw in some batch normalization and dropout layers. It’s all about keeping things balanced and preventing the model from memorizing the training data instead of actually learning from it.

Training the Model: It’s Like Teaching a Baby to Ride a Bike, But With More Math

Now that we’ve built our fancy FHR-LINet, it’s time to train it. Think of this as the model’s bootcamp, where it learns to distinguish between normal and compromised FHR signals. We don’t just throw it in the deep end, though. We gotta be patient teachers and use some clever techniques.

  • Epochs, Mini-Batches, and Class Weights, Oh My!: We train the model for sixty-five epochs, which is like making it read the entire CTG dataset sixty-five times. But instead of feeding it all the data at once, we break it down into mini-batches of thirty-two recordings. And because there are more normal cases than compromised cases in our dataset, we use class weights to give the model a little extra nudge when it encounters a compromised case.
  • Cross-Validation: It’s Like Taking Multiple Choice Tests: To make sure our model isn’t just memorizing the training data, we use a five-fold stratified cross-validation approach. It’s like dividing the class into five groups and giving each group a slightly different version of the test. This helps us evaluate how well the model generalizes to unseen data.
  • Data Augmentation: Because More Data is Always Better: Remember those sixty-minute FHR recordings? Well, we get even more mileage out of them by creating overlapping thirty-minute windows. It’s like cutting a pizza into more slices—same amount of pizza, but more opportunities to learn!
  • Excluding the Gray Area: To make things a little easier for our model, we exclude intermediate cases (those with a pH between seven point zero five and seven point one five) from the training data. This helps to reduce label errors and makes the distinction between normal and compromised cases crystal clear.

Putting FHR-LINet to the Test: Does It Pass With Flying Colors?

Training is done, the model is prepped—it’s time for the final exam! But we’re not talking about your typical multiple-choice test here. We’re evaluating FHR-LINet’s performance using three different approaches, each more challenging than the last:

Approach 1: The Sliding Window Challenge

Imagine a fifteen-minute window sliding across the sixty-minute FHR recording in five-minute increments. If the model flags any of these windows as potentially compromised, BAM—the entire recording is classified as positive. It’s a tough test, but hey, nobody said saving babies was easy.

Approach 2: The Cumulative Challenge

This time, the window starts at fifteen minutes and keeps growing in five-minute increments. The model has to analyze increasingly longer segments of the recording, making it harder to miss subtle signs of compromise. It’s like starting with a single puzzle piece and gradually adding more until you see the whole picture.

Approach 3: The Whole Enchilada

No pressure, but in this approach, the model has to analyze the entire sixty-minute FHR recording in one go. It’s the ultimate test of FHR-LINet’s ability to detect compromise even in the most challenging cases.

Grading the Model: More Than Just a Letter Grade

We’re not just looking for a simple “pass” or “fail” here. We need to know exactly how well FHR-LINet performs, and for that, we use some fancy metrics:

  • True Positive Rate (TPR): This tells us how good the model is at correctly identifying compromised cases. Think of it as the model’s “sensitivity” to danger.
  • False Positive Rate (FPR): This measures how often the model cries wolf when there’s no real danger. We want this number to be low, as too many false alarms can lead to unnecessary interventions.
  • Time to Predict (TTP): This is where things get really interesting. TTP tells us how long it takes the model to accurately predict compromise. In a time-sensitive situation like childbirth, every minute counts, so a faster TTP could be the difference between a happy outcome and a tragic one.

And to make sure we’re comparing apples to apples, we optimize the model’s threshold values for each approach, targeting specific False Positive Rates (five percent, ten percent, fifteen percent, and twenty percent). It’s all about finding the sweet spot between sensitivity and specificity.

FHR-LINet vs. MCNN: The Ultimate Showdown

What good is a new model if it’s not better than the existing ones, right? That’s why we pit FHR-LINet against the current state-of-the-art model, MCNN, in a head-to-head competition. We’re talking same data, same evaluation approaches, same everything—it’s a fair fight to the finish.

Leveling the Playing Field:

To make sure the comparison is fair, we implement and evaluate MCNN using the same rigorous methodology as FHR-LINet. We even train MCNN on both FHR and UC data, just like our model, to see if adding more data makes a difference.

The Moment of Truth: Can FHR-LINet Outperform the Champ?

We put both models through their paces, testing their performance on varying input lengths (thirty, forty-five, and sixty minutes). It’s like asking a runner to complete a sprint, a mid-distance run, and a marathon—we want to know which model has the stamina to go the distance.

And the results are in! But you’ll have to keep reading to find out who comes out on top… (cliffhanger alert!)