Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2020 Mar;11(2):242-252.
doi: 10.1055/s-0040-1708049. Epub 2020 Apr 1.

Natural Language Processing to Extract Meaningful Information from Patient Experience Feedback

Affiliations

Natural Language Processing to Extract Meaningful Information from Patient Experience Feedback

Khalid Nawab et al. Appl Clin Inform. 2020 Mar.

Abstract

Background: Due to reimbursement tied in part to patients' perception of their care, hospitals continue to stress obtaining patient feedback and understanding it to plan interventions to improve patients' experience. We demonstrate the use of natural language processing (NLP) to extract meaningful information from patient feedback obtained through Press Ganey surveys.

Methods: The first step was to standardize textual data programmatically using NLP libraries. This included correcting spelling mistakes, converting text to lowercase, and removing words that most likely did not carry useful information. Next, we converted numeric data pertaining to each category based on sentiment and care aspect into charts. We selected care aspect categories where there were more negative comments for more in-depth study. Using NLP, we made tables of most frequently appearing words, adjectives, and bigrams. Comments with frequent words/combinations underwent further study manually to understand factors contributing to negative patient feedback. We then used the positive and negative comments as the training dataset for a neural network to perform sentiment analysis on sentences obtained by splitting mixed reviews.

Results: We found that most of the comments were about doctors and nurses, confirming the important role patients ascribed to these two in patient care. "Room," "discharge" and "tests and treatments" were the three categories that had more negative than positive comments. We then tabulated commonly appearing words, adjectives, and two-word combinations. We found that climate control, housekeeping and noise levels in the room, time delays in discharge paperwork, conflicting information about discharge plan, frequent blood draws, and needle sticks were major contributors to negative patient feedback. None of this information was available from numeric data alone.

Conclusion: NLP is an effective tool to gain insight from raw textual patient feedback to extract meaningful information, making it a powerful tool in processing large amounts of patient feedback efficiently.

PubMed Disclaimer

Conflict of interest statement

None declared.

Figures

Fig. 1
Fig. 1
Preprocessing of the raw text. All text converted to lowercase. Numbers, special characters, and extra spaces removed. Certain words were similar in meaning, for example doc, dr, and doctors, all three were replaced with physician. Words were also reduced to base form, for example explained, explaining, and explain mean the same thing, therefore reduced to one word.
Fig. 2
Fig. 2
Natural language processing methods: ( A ) part of speech tagging, using representation of words and their tags; ( B ) output of the parts of speech tagger. Each word is coupled with its tag; ( C ) summary of neural network based model; and ( D ) evaluation metrics of the model.
Fig. 3
Fig. 3
( A , B ) Graphical representation of the training process of the model. Accuracy did not improve significantly after 15 epochs, but loss continued to decrease and started going up after 35 epochs.
Fig. 4
Fig. 4
Classification of mixed sentiment comments. All mixed comments were split into sentences. The deep learning model trained with the preclassified positive and negative comments was then used to classify the sentences into positive, negative, or neutral.
Fig. 5
Fig. 5
Word Cloud representations. The larger the word, the more frequent it appeared in the comments. ( A ) “Nurse” and “doctor” are the most frequent words in both positive and negative comments; ( B ) “room” also appears more frequently in negative comments.
Fig. 6
Fig. 6
Comments in each category: ( A ) based on sentiment and ( B ) based on care aspect.
Fig. 7
Fig. 7
Comments in each category based on care aspect: ( A ) positive comments and ( B ) negative comments.
Fig. 8
Fig. 8
Positive and negative comments in each category based on care aspect. Most care aspect categories had more positive comments than negative except for comments about room, tests and treatments, and discharge.
Fig. 9
Fig. 9
Classification of mixed sentiment feedback by deep learning model. A mixed comment may have many sentences; this figure shows how different sentences in one mixed feedback may be classified based on sentiment by the classification model.

References

    1. Press I. Concern for the patient's experience comes of age. Pat Exper J. 2014;1(01):4–6.
    1. Press Ganey.History & MissionAvailable at:https://www.pressganey.com/about/history-mission. Accessed December 11, 2019
    1. CMS.gov.HCAHPS: Patients' Perspectives of Care SurveyAvailable at:https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Inst.... Accessed December 11, 2019
    1. CMS.gov.The HCAHPS survey–Frequently asked questionsAvailable at:https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Inst.... Accessed December 11, 2019
    1. Office of the Legislative Counsel.Patient Protection and Affordable Care Act; Health-related portions fo the Heatlh Care and Education Reconciliation Act of 2010Available at:http://housedocs.house.gov/energycommerce/ppacacon.pdf. Accessed December 11, 2019