Peer Review History

Original SubmissionMarch 14, 2022
Decision Letter - Chim C. Lang, Editor

PONE-D-22-07622Multicenter validation of a machine learning phase space electro-mechanical pulse wave analysis to predict elevated left ventricular end diastolic pressure at the point-of-carePLOS ONE

Dear Dr. Burton

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Kindly consider the comments made by the comments of the reviewers and respond accordingly.  Please ensure that your decision is justified on PLOS ONE’s publication criteria and not, for example, on novelty or perceived impact.

Please submit your revised manuscript by the Aug 05 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Chim C. Lang

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf  and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please include your tables as part of your main manuscript and remove the individual files. Please note that supplementary tables (should remain/ be uploaded) as separate "supporting information" files.

3. Thank you for stating the following in the Competing Interests section:

“I have read the journal's policy and authors of this manuscript have the following competing interests.

Sanjeev Bhavnani MD is a scientific advisor to Corvista Health and Blumio; consultant to Bristol Meyers Squibb, Pfizer, and Infineon Semiconductor; data safety monitoring board chair at Proteus Digital; has received research support from Scripps Clinic and the Qualcomm Foundation and is member of the healthcare innovation advisory boards at the American College of Cardiology, American Society of Echocardiography, and BIOCOM (all non-profit institutions with all positions voluntary).

Jeremiah P. Depta MD reports consulting fees from Edwards Lifesciences LLC, Boston Scientific, V wave Medical Ltd and Abbot.

Brett Bennett MD reports payment or honoraria for lecture from Philips.

Horace R. Gillins BS, Ian Shadforth EngD, Emmanuel Lange, Abhinav Doomra MScAC, Mohammad Firouzi MSc, Farhad Fathieh PhD, Timothy Burton BComp, Ali Khosousi PhD, Shyam Ramchandami PhD and William E. Sanders Jr. MD report employment by CorVista Health, and stock options in the same.

Frank Smart MD reports grants or contracts from Abbot (GUIDE HF clinical trial), NIH / Ohio State (DCM genetic study), Duke Clinical Research (Transform HF), CorVista Health (Pulmonary Hypertension clinical trial), and participation on a Data Safety Monitoring Board or Advisory Board (Abbott Medical; GUIDE-HF Steering committee).

All other authors report no disclosures.”

Please confirm that this does not alter your adherence to all PLOS ONE policies on sharing data and materials, by including the following statement: "This does not alter our adherence to  PLOS ONE policies on sharing data and materials.” (as detailed online in our guide for authors http://journals.plos.org/plosone/s/competing-interests).  If there are restrictions on sharing of data and/or materials, please state these. Please note that we cannot proceed with consideration of your article until this information has been declared.

Please include your updated Competing Interests statement in your cover letter; we will change the online submission form on your behalf.

4. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions.

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

5. We note that Supplementary Appendix Section 9 includes an image of a patient in the study.

As per the PLOS ONE policy (http://journals.plos.org/plosone/s/submission-guidelines#loc-human-subjects-research) on papers that include identifying, or potentially identifying, information, the individual(s) or parent(s)/guardian(s) must be informed of the terms of the PLOS open-access (CC-BY) license and provide specific permission for publication of these details under the terms of this license. Please download the Consent Form for Publication in a PLOS Journal (http://journals.plos.org/plosone/s/file?id=8ce6/plos-consent-form-english.pdf). The signed consent form should not be submitted with the manuscript, but should be securely filed in the individual's case notes. Please amend the methods section and ethics statement of the manuscript to explicitly state that the patient/participant has provided consent for publication: “The individual in this manuscript has given written informed consent (as outlined in PLOS consent form) to publish these case details”.

If you are unable to obtain consent from the subject of the photograph, you will need to remove the figure and any other textual identifying information or case descriptions for this individual.

6. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

Additional Editor Comments:

This paper has been reviewed. A number of comments have been raised that require responses and revision.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: In this manuscript, the authors aimed to verify if phase space ML analysis with orthogonal voltage gradient and photoplethysmography as features can be used to predict elevated left ventricular end diastolic pressure. Overall, the manuscript is well-structured and clearly written. Below I list some comments for this manuscript.

1. The authors can provide more clear description for the definition of the indices used as features (perhaps provide some figure illustrations).

2. Minor comment: since this is a multi-center study, the authors can provide descriptions about the chosen centers. For example, why choosing these centers? Any special subject characteristic for each center?

Reviewer #2: Introduction:

1. It would be nice to include some literature reviews on any previous studies predicting LVEDP using non-invasive measurement, if any.

Results:

1. Are all the results shown in the results section using the ensembled model?

Discussion:

1. The discussion should focus on the results presented in this study, and the literature reviews should be moved to the introduction.

2. In the limitation, the authors mentioned that study subjects taking diuretics might affect the specificity. Have the authors tried to remove those subjects and rerun the model to test that hypothesis?

3. The authors should add the study populations are patients likely to have CAD to the limitation.

4. The authors should also add the possibility of overfitting to the limitation unless they can justify the large number of features they used in the study.

Methods:

1. The study population are patients that are likely to have CAD, which seems to be a biased dataset to predict LVEDP. If the aim is being able to predict the LVEDP values for just patients who are likely to have CAD, then this would be acceptable, but if the aim is to predict LVEDP for a wider range of populations, then it would make sense to include other patients irrelevant to having CAD.

2. Line 336: Is it without CVD or CAD? If it is CVD, then which diseases are considered here? As CVD is a very wide category.

3. Is the validation group also patients likely to have CAD?

4. What type of catheter was used to measure LVEDP, specs?

5. Line 353: it says the following 4 categories, but there are 5 listed

6. Line 368: what are the symptoms are considered here for HF

7. From Line 414 – 420, please give some specific numbers for the cut-offs, e.g. high-frequency noise above XXX Hz?? Maximum measurable value XXX?

8. Could you give a bit more details on how the features are calculated? I imagine you have both OVG and PPG signals with multiple cardiac cycles. Do you calculate the features cycle by cycle? If so, how do you get the final subject level features based on the cycle features? Do you calculate the features over the whole recorded signal? If so, what is the time window you use for each recording? If the time window is constant, then how many seconds? If not, please justify.

9. Could you provide a list of features used in the appendix? And how they are calculated?

10. Is the outlier detection for the cycle level features or in-between subjects as well?

11. The rationale for choosing the 13 machine-learning models is unclear and how the hyper-parameters were chosen is also unclear. The authors claim that the ensemble outputs are intended to outperform any single model, but would they be able to provide some results showing that?

12. In terms of training the model, what type of training was used? Cross-validation?

13. Can the author provide some information about what they did to prevent overfitting the models?

14. Can the author provide some details on how the 95% CIs were calculated?

General question:

1. How do you avoid overfitting the models?

2. Have you investigated the correlations between features?

Figures:

Figure 8: could you add what each colour and lines represent?

Figure 9: This figure is not very clear. Personally, I don’t think it is necessary to have a figure to describe the random forest, as it is a very well known method. I think it is best to describe what setup was used in the Random Forest, such as bootstrapping=True, how many estimators, any limits for the number of trees, leaves etc. Same for the figure of XGBoost in the appendix.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 1

We kindly request that you review the cover letter and response to reviewers documents (editor and reviewers, respectively), as they contain images and text formatting intended to facilitate your review, which are not possible to enter into this plain text field.

Attachments
Attachment
Submitted filename: Response to Reviewers.pdf
Decision Letter - Chim C. Lang, Editor

PONE-D-22-07622R1Multicenter validation of a machine learning phase space electro-mechanical pulse wave analysis to predict elevated left ventricular end diastolic pressure at the point-of-carePLOS ONE

Dear Sir,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

==============================

Kindly note the comments of the reviewer to assist in the further revision needed for your manuscript, Please ensure that your decision is justified on PLOS ONE’s publication criteria and not, for example, on novelty or perceived impact.

For Lab, Study and Registered Report Protocols: These article types are not expected to include results but may include pilot data. 

==============================

Please submit your revised manuscript by Oct 27 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Chim C. Lang

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Additional Editor Comments (if provided):

Kindly note the comments of the Reviewer to assist in your resubmission

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Partly

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: 1. About overfitting

a. Generally, a good ratio between sample size and parameters to avoid overfitting is somewhere above 10 to 1. This means for an average of 149 features, you would need about 1490 subjects. If I am not mistaken, the maximum number included in this study is 684 post exclusions. It is true that the tree-based methods (Random Forest and XGB) and the elastic net can help reduce the risk of overfitting, but it does not mean you will not overfit because you are using these models. I have seen many overfitting cases when people use those models. I do see the authors trying to manage overfitting by limiting the maximum depth, but there are other parameters as well that need to be looked into.

b. One of my additional questions here is how did you choose the hyperparameters for your tree-based models? What type of tunning was performed?

c. Similarly, what type of tunning was conducted for the elastic net model? What makes you choose 0.003 or 0.01 for alpha?

d. Also, an AUC of 0.81 from the validation does not prove that overfitting did not occur since it is just one data point. One can easily argue if there is no overfitting, you could have an AUC of 0.95.

e. In summary, it is still not very clear how the hyperparameters were chosen for each model. And it is unfair for the authors to conclude that the overfitting did not occur based on an AUC of 0.81 from validation. Also, please remove the word “high-performing”, as a word like “high” or “low” is subjective and not scientific.

2. Regarding the type of catheter, the authors have replied to Comment #10 that the choice of the catheter was left to the discretion of each cardiologist, which means the device for measuring the ground truth can be different between time and the medical centre. Have the authors analysed the potential impact of this?

3. Is there a reason that the authors can only provide a list of the families of features used in the study instead of a list containing the actual features?

4. What type of cross-validation did the authors use? Can they specify it in the paper?

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 2

Dear Reviewer #2,

*Please find these responses in a separate submission file that has the tables formatted for better readability than when copied here*

Thank you for your thorough review and valuable comments. We appreciate and agree with your concern on overfitting and the corresponding risks. We view the recognition of this problem as one of the most significant in the field of machine learning. While we cannot definitively prove that overfitting has not occurred, we believe that the application of our various methods acts to sufficiently mitigate the of likelihood of its occurrence.

Specifically, our most critical control for overfitting was the single evaluation on a test dataset separated from the development set by enrollment date. Further, the probability of our model achieving the performance of 0.81 by chance on a dataset of that size (which could be the case if our development set was overfit) is low, as shown by the comparison between the curve and the random predictor that lies along the diagonal of the ROC space (p<2E-16 by DeLong’s test).

We also implemented a series of additional controls to increase the chance that our algorithm would generalize to an unseen dataset:

Use of modelling methodologies with regularization and penalty terms that are well-suited to high-dimensional spaces.

Cross-validation model development with careful monitoring of fold-to-fold standard deviations

Careful choice of models to ensemble based on hyperparameter stability

Below we provided detailed explanations of our methodologies to address your specific concerns, and we hope that we have addressed them to your satisfaction. Our responses appear below in bold, provided in italics where applicable in the response, and modified manuscript text in track changes.

Sincerely,

Sanjeev Bhavnani MD

Specific Comment #1: In order to clarify our imposed safeguards in reference to overfitting, we have revised the text on lines 350-355.

Author Response: The use of an ensemble, as is the case, does not increase the likelihood of overfitting but rather mitigates it by reducing the dependence on a single model. The methods employed to avoid overfitting include the use of cross-validation within the development data, using simple models with regularization and penalty terms, and testing the performance of the model on unseen data (only doing so only once). With respect to the models in particular; first, each model is exposed to an average of 149 features (with a range of 89-194, supplement section 10).

Generally, a good ratio between sample size and parameters to avoid overfitting is somewhere above 10 to 1. This means for an average of 149 features, you would need about 1490 subjects. If I am not mistaken, the maximum number included in this study is 684 post exclusions. It is true that the tree-based methods (Random Forest and XGB) and the elastic net can help reduce the risk of overfitting, but it does not mean you will not overfit because you are using these models. I have seen many overfitting cases when people use those models. I do see the authors trying to manage overfitting by limiting the maximum depth, but there are other parameters as well that need to be looked into.

Author Response: To clarify the number of subjects – the N=684 that you referenced is the validation cohort, while the models were developed on the N=1272 development cohort (and specifically, each model using a subset of the available cohort, as shown in supplement section 10). With respect to your other comments in this section, we hope that they will be addressed in the responses to your other questions.

One of my additional questions here is how did you choose the hyperparameters for your tree-based models? What type of tunning was performed?

Author Response: The possible hyperparameter values are shown below for XGB and Random Forest and were chosen to provide a range over the conservative values of each hyperparameter, with some exponential changes where appropriate to ensure coverage. Each combination of hyperparameters settings was trained within the development dataset only using cross-validation (recall that the ensemble was locked, and the validation data evaluated only once).

XGB Hyperparameters

Hyperparameter Possible Values

Learning Rate 0.1, 0.3, 0.5

Maximum Tree Depth 2, 3, 5, 7

Minimum Child Weight 1, 3, 5

Number of Trees 5, 10, 50, 100

Regularization Alpha 0.1, 0.3, 0.5

Random Forest Hyperparameters

Hyperparameter Possible Values

Maximum Tree Depth 2, 3, 5, 7

Minimum Samples Per Leaf 1, 25, 50, 100

Number of Trees 5, 10, 50, 100, 500

A heuristic stability assessment was performed during the selection of each model. Specifically, for a given model, the cross-validation performance of each “neighboring” hyperparameter specification was compared with that of the model under consideration, and the difference evaluated, with the goal of determining whether that region of the hyperparameter space generated similar cross-validation performances as a measure of stability. To illustrate the method, consider the toy example below using XGB, which assumes that XBG only requires the learning and minimum child weight hyperparameters. Varying these parameters each generates an AUC value, as described in the following table.

Minimum Child Weight

Learning Rate 1 3 5

0.1 AUC 0.1,1 AUC 0.1,3 AUC 0.1,5

0.3 AUC 0.3,1 AUC 0.3,3 AUC 0.3,5

0.5 AUC 0.5,1 AUC 0.5,3 AUC 0.5,5

The resultant AUC values can be used to assess the stability of learning rate = 0.3 and minimum child weight = 3 in the following manner. Should the stability value be close to zero, then that indicates that the performance at that specific learning rate and minimum child weight is relatively insensitive to changes to adjacent hyperparameter values.

Stability= 〖AUC〗_0.3,3- 1/8 (〖AUC〗_0.1,1+〖AUC〗_0.3,1+ 〖AUC〗_0.5,1+ 〖AUC〗_0.1,3+ 〖AUC〗_0.5,3 〖+ AUC〗_0.1,5+ 〖AUC〗_0.3,5+ 〖AUC〗_0.5,5 )

The extension of the methodology is straightforward to higher dimensions of hyperparameters, where the configuration of interest is compared to each of its neighbors in that high-dimensional space, which are found based on changing a single hyperparameter by one step in the assessed range of that hyperparameter.

Finally, please note that this is a heuristic that we used to guide our selection of the hyperparameters settings, and we do not claim that it is an optimal strategy. For instance, a limitation of this heuristic is the uneven spacing between hyperparameter settings – though we viewed this as necessary to explore the hyperparameter space, it does result in nonuniform distance between neighbors. A possible future extension of this method could correct for that nonuniform distance.

Similarly, what type of tunning was conducted for the elastic net model? What makes you choose 0.003 or 0.01 for alpha?

Author Response: Please see the explored hyperparameter space for Elastic Net in the table below. The same hyperparameter stability assessment performed for Random Forest and XBG (as described above in b.) was also performed for Elastic Net.

Elastic Net Hyperparameters

Hyperparameter Possible Values

Alpha 0, 0.003, 0.01, 0.1, 0.5

Fit Intercept True, False

L1 Ratio 0, 0.01, 0.1, 0.5

Normalize True, False

Also, an AUC of 0.81 from the validation does not prove that overfitting did not occur since it is just one data point. One can easily argue if there is no overfitting, you could have an AUC of 0.95.

Author Response: The point is well taken. Please see discuss on page one regarding AUC.

In summary, it is still not very clear how the hyperparameters were chosen for each model. And it is unfair for the authors to conclude that the overfitting did not occur based on an AUC of 0.81 from validation. Also, please remove the word “high-performing”, as a word like “high” or “low” is subjective and not scientific.

Author Response: We agree that the terms inferring “high/low” are subjective and the text has been altered accordingly on lines 368-369, as shown below. With respect to the hyperparameter selection, we hope that it is now clear from the response to question 1b.

Finally, the ultimate test of overfitting is the performance on unseen blinded data, which yielded an AUC of 0.81.

Specific Comment #2: Regarding the type of catheter, the authors have replied to Comment #10 that the choice of the catheter was left to the discretion of each cardiologist, which means the device for measuring the ground truth can be different between time and the medical centre. Have the authors analysed the potential impact of this?

Author Response: The reviewer makes an excellent point that variations in catheter technique could affect the consistency of measurements. Such sources of variability are an important concern. However, since 6-French to 8-French catheters are the primary options for left ventricular catheterization, recorded pressure should vary minimally as a function of catheter selection. Moreover, since a variety of catheter types and sizes are used in clinical practice, specifying a particular catheter type may impair the generalizability of the resultant algorithm.

Specific Comment #3: Is there a reason that the authors can only provide a list of the families of features used in the study instead of a list containing the actual features?

Author Response: Describing all of the features was out of scope for this paper, and therefore we provided feature description at the level of the family in the interest of brevity. However, we did provide detailed information on several of the features (Fig 3 and Supplement Section 3) to allow the reader to gain an appreciation for the prediction mechanism. In addition, many of the features are proprietary in nature.

Specific Comment #4: What type of cross-validation did the authors use? Can they specify it in the paper?

Author Response: The type of cross-validation used was stratified 5-fold cross-validation repeated for 100 iterations (to vary the train/test folds). Please see the modified text from lines 522-523 below:

Each of the 13 models were individually performant based on stratified 5-fold cross-validation repeated for 100 iterations (to vary the train/test folds) within the development data (S11) but represent unique analyses of LVEDP assessment.

Attachments
Attachment
Submitted filename: Author Response To Reviewer_SEPT 24.docx
Decision Letter - Chim C. Lang, Editor

Multicenter validation of a machine learning phase space electro-mechanical pulse wave analysis to predict elevated left ventricular end diastolic pressure at the point-of-care

PONE-D-22-07622R2

Dear Dr

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Chim C. Lang

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Formally Accepted
Acceptance Letter - Chim C. Lang, Editor

PONE-D-22-07622R2

Multicenter validation of a machine learning phase space electro-mechanical pulse wave analysis to predict elevated left ventricular end diastolic pressure at the point-of-care

Dear Dr. Bhavnani:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Chim C. Lang

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .