NLP

CRIS Natural Language Processing

CRIS NLP Applications Catalogue

Natural Language Processing (NLP) is a type of Artificial Intelligence, or AI, for extracting structured information from the free text of electronic health records. We have set up the Clinical Record Interactive Search (CRIS) NLP Service to facilitate the extraction of anonymised information from the free text of the clinical record at the South London and Maudsley NHS Foundation Trust. Research using data from electronic health records (EHRs) is rapidly increasing and most of the valuable information is sought in the free text. However, manually reviewing the free text is very time consuming. To overcome the burden of manual work and extract the information needed, NLP methods are utilised and in high demand throughout the research world.

CRIS Natural Language Processing

Welcome to the NLP Applications Catalogue. This provides detailed information about various apps developed within the CRIS service.

The webpage published and regularly updated here contains the details and performance of over 80 NLP applications relating to the automatic extraction of mental health data from the EHR that we have developed and currently routinely deployed through our NLP service.

This webpage provides details of NLP resources which have been developed since around 2009 for use at the NIHR Maudsley Biomedical Research Centre and its mental healthcare data platform, CRIS. We have set up the CRIS NLP Service to facilitate the extraction of anonymised information from the free text of the clinical record. Research using data from electronic health records (EHRs) is rapidly increasing and the most valuable information is sometimes only contained in the free text. This is particularly the case in mental healthcare, although not limited to that sector.

The CRIS system was developed for use within NIHR Maudsley BRC. It provides authorised researchers with regulated, secure access to anonymised information extracted from South London and Maudsley’s EHR.

The South London and Maudsley provides mental healthcare to a defined geographic catchment of four south London boroughs (Croydon, Lambeth, Lewisham, Southwark) with around 1.3 million residents, in addition to a range of national specialist services.

General Points of Use

All applications currently in production at the CRIS NLP Service are described here.

Our aim is to update this webpage at least twice yearly so please check you are using the version that pertains to the data extraction you are using.

Guidance for use: Every application report comprises four parts:

1) Description– the name of application and short explanation of what construct(s) the application seeks to capture.

2) Definition - an account of how the application was developed (e.g. machine-learning/rule-based, the terms searched for and guidelines for annotators), annotation classes produced and interrater reliability results (Cohen’s Kappa).

3) Performance – precision and recall are used to evaluate application performance in pre-annotated documents identified by the app as well as un-annotated documents retrieved by keyword searching the free text of the events and correspondence sections of CRIS. a) Precision is the ratio of the number of relevant (true positive) entities retrieved to the total number of entities (irrelevant -false positive- and relevant -true positive)) retrieved. b) Recall is the ratio of the number of relevant (true positive) entities retrieved to the number of relevant (true positive and false negative) entities available in the database. Performance testing is outlined in chronological order for either pre-annotated documents, unannotated documents retrieved through specific keyword searches or both. The latest performance testing on the list corresponds to results produced by the version of the application currently in use by the NLP Service. Search terms used for recall testing are presented, where necessary. Similarly, details are provided for any post-processing rules that have been implemented. Notes relating to observations by annotators and performance testers are described, where applicable.

4) Production – information is provided on the version of the application currently in use by the NLP Service and the corresponding deployment schedule.

Symptom scales (see proposed allocations)

As the number of symptom applications is increasing, we regularly evaluate how to make these available to researchers in a flexible and meaningful manner.

To this end, and in order to reduce the risk of too many and/or highly correlated variables in analyses, we are currently utilising symptom scales that group positive schizophreniform, negative schizophreniform, depressive, manic, disorganized and catatonic symptoms respectively.

The group of ‘other’ symptoms represent symptoms that have been developed separately for different purposes and that are intended to be used individually rather than in scales.

Each symptom receives a score of 1 if it’s scored as positive within a given surveillance period.

Individual symptoms are then summed to generate a total score of:

• 0 – 16 for positive schizophreniform

• 0 – 12 for negative schizophreniform

• 0 – 21 for depressive

• 0 – 8 for manic

• 0 – 8 for disorganized

• 0 – 4 for catatonic

We are encouraging researchers, unless there is a particular reason to be discussed with the NLP team, to use the scales for extracting and analysing data relating to symptom applications.

Version

V3.3

Contents

Symptoms

Aggression
Agitation
Anergia
Anhedonia
Anosmia
Anxiety
Apathy
Arousal
Bad Dreams
Blunted Affect
Circumstantiality
Cognitive Impairment
Concrete Thinking
Delusions
Derailment
Disturbed Sleep
Diurnal Variation
Drowsiness
Early Morning wakening
Echolalia
Elation
Emotional Withdrawal
Eye Contact (Catergorisation)
Fatigue
Flight of Ideas
Fluctuation
Formal Thoughts Disorder
Grandiosity
Guilt
Hallucinations (All)
Hallucinations - Auditory
Hallucinations - Olfactory Tactile Gustatory (OTG)
Hallucinations - Visual
Helplessness
Hopelessness
Hostility
Insomnia
Irritability
Loss of Coherence
Low energy
Mood instability
Mutism
Negative Symptoms
Nightmares
Obsessive Compulsive Symptoms
Paranoia
Passivity
Persecutory Ideation
Poor Appetite
Poor Concentration
Poor Eye Contact
Poor Insight
Poor Motivation
Poverty Of Speech
Poverty Of Thought
Psychomotor Activity (Catergorisation)
Self Injurious Action
Smell
Social Withdrawal
Stupor
Suicidal Ideation
Tangentiality
Taste
Tearfulness
Thought Block
Thought Broadcast
Thought Insertion
Thought Withdrawal
Waxy Flexibility
Weight Loss
Worthlessness

Physical Health Conditions

Asthma
Bronchitis
Cough
Crohn's Disease
Falls
Fever
Hypertension
Multimorbidity - 21 Long Term Conditions (Medcat)
Pain
Rheumatoid Arthritis
HIV
HIV Treatment

Contextual Factors

Amphetamine
Cannabis
Chronic Alcohol Abuse
Cocaine or Crack Cocaine
MDMA
Smoking
Education
Occupation
Lives Alone
Loneliness
Violence

Interventions

CAMHS - Creative Therapy
CAMHS - Dialectical Behaviour Therapy (DBT)
CAMHS - Psychotherapy/Psychosocial Intervention
Cognitive Behavioural Therapy (CBT)
Depot Medication
Family Intervention
Medication
Social Care - Care Package
Social Care - Home Care
Social Care - Meals on Wheels

Outcome and Clinical Status

Blood Pressure (BP)
Body Mass Index (BMI)
Brain MRI report volumetric Assessment for dementia
Cholesterol
EGFR
HBA1C
Lithium
Mini-Mental State Examination (MMSE)
Neutrophils
Non-Adherence
Diagnosis
Treatment- Resistant Depression
Bradykinesia (Dementia)
Trajectory
Tremor (Dementia)
QT
White Cells

Miscellaneous

Family Contact
Forms
Quoted Speech

Symptom Scales (see notes)

Symptoms

Aggression

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of aggressive behaviour in patients, including verbal, physical and sexual aggression.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“reported to be quite aggressive towards…”,

“violence and aggression, requires continued management and continues to reduce in terms of incidents etc”.

 

Examples of negative / irrelevant mentions (not included in the output):

“no aggression”,

“no evidence of aggression”

“aggression won’t be tolerated”.

 

Search term(case insensitive): *aggress*

Evaluated Performance

Cohen's k = 85% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 75%

Additional Notes

Run schedule– Monthly

Other Specifications

Version 1.0, Last updated: xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Agitation

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of agitation

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Very agitated at present, he was agitated”,

“He was initially calm but then became agitated and started staring and pointing at me towards”,

“Should also include no longer agitated. “

 

Examples of negative / irrelevant mentions (not included in the output):

“He did not seem distracted or agitated”,

“Not agitated”,

“No evidence of agitation”,

“A common symptom of psychomotor agitation”.

 

Search term(case insensitive): *agitat*

Evaluated Performance

Cohen's k = 85% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 79%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 30 random documents):

Precision (specificity / accuracy) = 82%

Additional Notes

Run schedule– Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Anergia

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of anergia

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“feelings of anergia…”

 

Examples of negative / irrelevant mentions (not included in the output):

“no anergia”,

“no evidence of anergia”,

“no feeling of anergia”.

 

Search term(case insensitive): *anergia*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 89%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 30 random documents):

Precision (specificity / accuracy) = 93%

Additional Notes

Run schedule– Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2015-007619

10.1136/bmjopen-2021-056541

Anhedonia

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of anhedonia (inability to experience pleasure from activities usually found enjoyable).

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ X had been anhedonic”,

“ X has anhedonia”.

 

Examples of negative / irrelevant mentions (not included in the output):

“no anhedonia”,

“not anhedonic”,

Used in a list, not applying to patient (e.g. typical symptoms include …);

Uncertain (might have anhedonia, ?anhedonia, possible anhedonia);

 

Search term(s): *anhedon*

Evaluated Performance

Cohen's k = 85% (testing done on 50 random documents).

 

Instance level, (testing done on 100 random documents):

Precision (specificity / accuracy) = 93%

Recall (sensitivity / coverage) = 86%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 30 random documents):

Precision (specificity / accuracy) = 87%

Additional Notes

Run schedule– Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2015-007619

10.1136/bmjopen-2021-056541

Anosmia

Return to contents

CRIS NLP Service

Brief Description

Application to extract and classify mentions related to anosmia.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Loss of enjoyment of food due to anosmia”,

“COVID symptoms such as anosmia”

 

Examples of negative / irrelevant mentions (not included in the output):

“Nil anosmia”,

“Anosmia related to people other than the patient”,

“Mentions of medications for it”,

“Don’t come to the practice if you have any covid symptoms such as anosmia”

 

Search term(case insensitive): Anosmia*

Evaluated Performance

Cohen's k = 83% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 93%

Additional Notes

Run schedule– On demand

Other Specifications

Version 1.0, Last updated:xx

Anxiety

Return to contents

CRIS NLP Service

Brief Description

Application to extract and classify mentions related to (any kind of) anxiety.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Affirmed, Negated, and Irrelevant.

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZZ shows anxiety problems”

 

Examples of negative / irrelevant mentions (not included in the output):

“ZZZZ does not show anxiety problems”

“If ZZZZ was anxious he would not take his medication”

 

Search Terms (case insensitive): Available on request

Evaluated Performance

Cohen’s k = 94% (testing done on 3000 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 87%

Recall (sensitivity / coverage) = 97%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Apathy

Return to contents

CRIS NLP Service

Brief Description

Application to extract the presence of apathy

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ continues to demonstrate apathy”

“ some degree of apathy noted”

 

Examples of negative / irrelevant mentions (not included in the output):

“denied apathy”

“no evidence of apathy”

“may develop apathy or as a possible side effect of medication”

“*apathy* found in quite a few names”

 

Search term(s): *apath*

Evaluated Performance

Cohen's k=86% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 93%

Recall (sensitivity / coverage) = 86%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 30 random documents):

Precision (specificity / accuracy) = 73%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Arousal

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of arousal excluding sexual arousal.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“...the decisions she makes when emotionally aroused”,

“...during hyperaroused state”,

“following an incidence of physiological arousal”

 

Examples of negative / irrelevant mentions (not included in the output):

“mentions of sexual arousal”,

“not aroused”,

“annotations include unclear statements and hypotheticals”

 

Search term(s): *arous*

Evaluated Performance

Cohen's k = 95% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 91%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Bad Dreams

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of experiencing a bad dream

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZZ had a bad dream last night”,

“she frequently has bad dreams”,

“ZZZZZ has suffered from bad dreams in the past”,

“ZZZZZ had a bad dream that she was underwater”,

“ he’s been having fewer bad dreams”

 

Examples of negative / irrelevant mentions (not included in the output):

“she denied any bad dreams”,

“does not suffer from bad dreams”,

“she said it might have been a bad dream”,

“he woke up in a start, as if waking from a bad dream”,

 

Search term(s): bad dream*

Evaluated Performance

Cohen's k = 100% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Blunted Affect

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of blunted affect

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“his affect remains very blunted”,

“objectively flattened affect”,

“states that ZZZZZ continues to appear flat in affect”

 

Examples of negative / irrelevant mentions (not included in the output):

“incongruent affect”,

“stable affect”,

“typical symptoms include blunted affect”,

“slightly flat affect”,

 

Search term(s): *affect*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 80%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Circumstantiality

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of circumstantiality

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“loose associations and circumstantiality”,

“circumstantial in nature”,

“some circumstantiality at points”,

“speech is less circumstantial”

 

Examples of negative / irrelevant mentions (not included in the output):

“no signs of circumstantiality”,

“no evidence of circumstantial”

“Such as a hypothetical cause of something else”

 

Search term(s): *circumstan*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 94%

Recall (sensitivity / coverage) = 92%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Cognitive Impairment

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of cognitive impairment. The application allows to detect cognitive impairments related to attention, memory, executive functions, and emotion, as well as a generic cognition domain. This application has been developed for patients diagnosed with schizophrenia.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Affirmed and relating to the patient and Negated/Irrelevant.

Output and Definitions

The output includes-

 

Examples of positive mentions:

“patient shows attention problems (positive) “

“ZZZ does not show good concentration “

“ZZZ shows poor concentration “

“patient scored 10/18 for attention “

“ZZZ seems hyper-vigilant “

 

Examples of negative / irrelevant mentions (not included in the output):

“patient uses distraction technique to ignore hallucinations “

“attention seeking”

“patient needs (medical) attention”

“draw your attention to…”

 

Search term(s): attention, concentration, focus, distracted, adhd, hypervigilance, attend to

Evaluated Performance

Cohen’s k:

Cognition – 66% (testing done on 3000 random documents)

Emotion – 84% (testing done on 3000 random documents)

Executive function – 40% (testing done on 3000 random documents)

Memory – 68% (testing done on 3000 random documents)

Attention – 99% (testing done on 2616 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 96%

Recall (sensitivity / coverage) = 92%

 

Patient level testing done on all patients with an F20 diagnosis (testing done on 100 random documents):

Precision (specificity / accuracy) = 78%

Recall (sensitivity / coverage) = 70%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

DOI

10.3389/fdgth.2021.711941

Concrete Thinking

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of concrete thinking.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“text referring to ‘concrete thinking’”,

“speech or answers to questions being ‘concrete’”,

“the patient being described as ‘concrete’ without elaboration”,

“answers being described as concrete in cognitive assessments”,

“‘understanding’ or ‘manner’ or ‘interpretations’ of circumstances being described as concrete”

 

Examples of negative / irrelevant mentions (not included in the output):

“no evidence of concrete thinking”

“reference to concrete as a material (concrete floor, concrete house etc.)”

“no concrete plans”,

“delusions being concrete”

 

Search term(s): Concret*, "concret* think*"

Evaluated Performance

Cohen's k = 83% (testing done on 50 random documents)

 

Instance level (testing done on 146 random documents):

Precision (specificity / accuracy) = 84%

Recall (sensitivity / coverage) = 41%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Delusions

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of delusions

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“paranoid delusions”,

“continued to express delusional ideas of the nature”

“no longer delusional- indicates past”

 

Examples of negative / irrelevant mentions (not included in the output):

“no delusions”,

“denied delusions”

“delusions are common”

 

Search term(s): *delusion*

Evaluated Performance

Cohen's k = 92% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 93%

Recall (sensitivity / coverage) = 85%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Derailment

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of derailment.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Very agitated at present, he was agitated”,

“He was initially calm but then became agitated and started staring and pointing at me towards”,

“Should also

“he derailed frequently”,

“there was evidence of flight of ideas”,

“thought derailment in his language” ‘speech no longer derailed’.

 

Examples of negative / irrelevant mentions (not included in the output):

“no derailment”,

“erratic compliance can further derail her stability”

“train was derailed”

 

Search term(s): *derail*

 

 

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 88%

Recall (sensitivity / coverage) = 95%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Disturbed Sleep

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of disturbed sleep.

Development Approach

Development approach: Rule-Based.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“complains of poor sleep”,

“sleep disturbed”,

“sleep difficulty”,

“sleeping poorly”

 

Search term(s): "disturbed sleep*", "difficult* sleep*", "poor sleep*"

Evaluated Performance

Cohen's k = 75% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 88%

Recall (sensitivity / coverage) = 68%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 2.0, Last updated:xx

Diurnal Variation

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of diurnal variation of mood

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"patient complaints of diurnal variation"

"he reported diurnal variation in his mood"

"Diurnal variation present"

"some diurnal variation of mood present"

 

Examples of negative / irrelevant mentions (not included in the output):

"no diurnal variation"

"diurnal variation absent"

"diurnal variation could be a symptom of more severe depression"

"we spoke about possible diurnal variation in his mood"

 

Search term(s): diurnal variation

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 Random Documents):

Precision (specificity / accuracy) = 94%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Drowsiness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of drowsiness

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZZ appeared to be drowsy”,

“She has complained of feeling drowsy”

 

Examples of negative / irrelevant mentions (not included in the output):

“He is not drowsy in the mornings”,

“She was quite happy and did not appear drowsy”,

“In reading the label (of the medication), ZZZZZ expressed concern in the indication that it might make him drowsy”,

“Monitor for increased drowsiness and inform for change in presentation”,

 

Search term(s): drows*

Evaluated Performance

Cohen's k = 83% (testing done on 1000 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 80%

Recall (sensitivity / coverage) =100%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Early Morning wakening

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of early morning wakening.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ patient complaints of early morning awakening”,

“he reported early morning wakening”,

“Early morning awakening present”,

“there is still some early morning wakening”

 

Examples of negative / irrelevant mentions (not included in the output):

“no early morning wakening”,

“early morning wakening absent”,

“early morning awakening could be a symptom of more severe depression”,

“we spoke about how to deal possible early morning wakening”

 

Search term(s): early morning wakening

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Echolalia

Return to contents

CRIS NLP Service

Brief Description

Application to extract occurrences where echolalia is present.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“no neologisms, but repeated what I said almost like echolalia”,

“intermittent echolalia”,

“some or less echolalia”

 

Examples of negative / irrelevant mentions (not included in the output):

“no echolalia”,

“no evidence of echolalia”,

“Echolalia is not a common symptom”,

“Include hypotheticals such as he may have some echolalia, evidence of possible echolalia”

 

Search term(s): *echola*

Evaluated Performance

Cohen's k = 88% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1017/S0033291721004402

10.1136/bmjopen-2016-012012

Elation

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of elation.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“mildly elated in mood”,

“elated in mood on return from leave”,

“she appeared elated and aroused”

 

Examples of negative / irrelevant mentions (not included in the output):

“ZZZZZ was coherent and more optimistic/aspirational than elated throughout the conversation”,

“no elated behaviour" etc.

“In his elated state there is a risk of accidental harm”,

“monitor for elation”,

 

Search term(s): *elat*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 94%

Recall (sensitivity / coverage) = 97%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Emotional Withdrawal

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of emotional withdrawal

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

To any description of the patient being described as withdrawn or showing withdrawal but with the following exceptions (which are annotated as unknown):

• Alcohol, substance, medication withdrawal

• Withdrawal symptoms, fits, seizures etc.

• Social withdrawal (i.e. a patient described as becoming withdrawn would be positive but a patient described as showing ‘social withdrawal’ would be unknown – because social withdrawal is covered in another application).

• Thought withdrawal (e.g. ‘no thought insertion, withdrawal or broadcast’)

• Withdrawing money, benefits being withdrawn etc.

 

 

Examples of negative / irrelevant mentions (not included in the output):

Restricted to instances where the patient is being described as not withdrawn and categorised as unknown.

 

Search term(s): withdrawn

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2015-007619

Eye Contact (Catergorisation)

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of eye contact and determine the type of eye contact.

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

the application successfully identifies the type of eye contact (as denoted by the keyword) in the context (as denoted by the contextstring)

e.g., keyword: ‘good’; contextstring: ‘There was good eye contact’

 

Negative mentions: the application does not successfully identifies the type of contact (as denoted by the keyword) in the context (as denoted by the contextstring). The keyword does not related to the eye contact

e.g., keyword: ‘showed’; contextstring: ‘showed little eye-contact’.

Keyword: the term describing the type of eye contact

ContextString: the context containing the keyword in its relation to eye-contact

 

Search term(s): Eye *contact*

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 80%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Fatigue

Return to contents

CRIS NLP Service

Brief Description

Application to identify symptoms of fatigue.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZ has been experiencing fatigue”,

“fatigue interfering with daily activities”

 

Examples of negative / irrelevant mentions (not included in the output):

“No mentions of fatigue”,

“her high levels of anxiety impact on fatigue”,

“main symptoms of dissociation leading to fatigue”

“ZZZZ is undertaking CBT for fatigue”.

 

Search term(s): Fatigue, exclude ‘chronic fatigue syndrome’.

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 78%

Recall (sensitivity / coverage) = 95%

Additional Notes

Run schedule – Monthly

Other Specifications

Version: xx, Last updated:xx

Flight of Ideas

Return to contents

CRIS NLP Service

Brief Description

Application to extract instances of flight of ideas.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Mrs ZZZZZ was very elated with by marked flights of ideas”,

“marked pressure of speech associated with flights of ideas”,

“Some flight of ideas”.

 

Examples of negative / irrelevant mentions (not included in the output):

“no evidence of flight of ideas”,

“no flight of ideas”

“bordering on flight of ideas”

“relative shows FOI”

 

Search term(s):*flight* *of* *idea*

Evaluated Performance

Cohen's k = 96% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2021-056541

Fluctuation

Return to contents

CRIS NLP Service

Brief Description

The purpose of this application is to determine if a mention of fluctuation within the text is relevant

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Mrs ZZZZZ’s mood has been fluctuating a lot”,

“suicidal thoughts appear to fluctuate”

 

Examples of negative / irrelevant mentions (not included in the output):

“no evidence of mood fluctuation”,

“does not appear to have significant fluctuations in mental state”

“unsure whether fluctuation has a mood component”,

“monitoring to see if fluctuations deteriorate”

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 87%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1017/S0033291724001922

Formal Thoughts Disorder

Return to contents

CRIS NLP Service

Brief Description

Application to extract occurrences where formal thought disorder is present.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“deteriorating into a more thought disordered state with outbursts of aggression”,

“there was always a degree thought disorder”,

“Include some formal thought disorder”

 

Examples of negative / irrelevant mentions (not included in the output):

“no signs of FTD”,

“NFTD”

“?FTD”,

“relative shows FTD”,

 

Search term(s): *ftd*,*formal* *thought* *disorder*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 83%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Grandiosity

Return to contents

CRIS NLP Service

Brief Description

Application to extract occurrences where grandiosity is apparent.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZZ was wearing slippers and was animated elated and grandiose”,

“reduction in grandiosity”,

”No longer grandiose”

 

Examples of negative / irrelevant mentions (not included in the output):

“No evidence of grandiose of delusions in the content of his speech”,

“no evidence of grandiose ideas”

“his experience could lead to grandiose ideas”

 

Search term(s): *grandios*

Evaluated Performance

Cohen's k = 89% (testing done on 50 random documents).

 

Instance level: (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 91%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Guilt

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of guilt.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she then feels guilty/angry towards mum”,

“Being angry is easier to deal with than feeling guilty”,

“Include feelings of guilt with a reasonable cause and mentions stating”,

“no longer feels guilty”

 

Examples of negative / irrelevant mentions (not included in the output):

“No feeling of guilt”,

“denies feeling hopeless or guilty”

“he might be feeling guilty”,

“some guilt”,

“sometimes feeling guilty”

 

Search term(s): *guil*

Evaluated Performance

Cohen's k = 92% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 83%

 

Patient level testing done on all patients with primary diagnosis code of F32* and F33* (testing done on 90 random documents):

Precision (specificity / accuracy) = 93%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Hallucinations (All)

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of hallucinations.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“her husband was minimising her hallucinations”,

“continues to experience auditory hallucinations”,

“doesn’t appear distressed by his hallucinations”,

“he reported auditory and visual hallucinations”,

“this will likely worsen her hallucinations”,

“his hallucinations subsided”,

“Neuroleptics were prescribed for her hallucinations”.

 

Examples of negative / irrelevant mentions (not included in the output):

“denied any hallucinations”,

“no evidence of auditory hallucinations”,

“pseudo(-) hallucinations”,

“hallucinations present?”,

 

Search term(s): hallucinat*

Evaluated Performance

Cohen's k = 83% (testing done on 100 random documents).

 

Instance level (testing done on 100 dandom documents):

 

Precision (specificity / accuracy) = 84%

Recall (sensitivity / coverage) = 98%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 2.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Hallucinations - Auditory

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of auditory hallucinations non-specific to diagnosis.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Seems to be having olfactory hallucination”,

“in relation to her tactile hallucinations”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies auditory, visual, gustatory, olfactory and tactile hallucinations at the time of the assessment”,

“denied tactile/olfactory hallucination”

“possibly olfactory hallucinations”

 

Search term(s): auditory hallucinat*

Evaluated Performance

Cohen's k = 96% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 80%

Recall (sensitivity / coverage) = 84%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Hallucinations - Olfactory Tactile Gustatory (OTG)

Return to contents

CRIS NLP Service

Brief Description

Application to extract occurrences where auditory hallucination is present. Auditory hallucinations may be due to a diagnosis of psychosis/schizophrenia or may be due to other causes, e.g. due to substance abuse.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“seems to be having olfactory hallucinations”,

“in relation to her tactile hallucinations”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies auditory, visual, gustatory, olfactory and tactile hallucinations at the time of the assessment”,

“denied tactile/olfactory hallucinations”

“possibly olfactory hallucinations”

 

Search term(s): *olfactory*, *hallucin*, *gustat* , *tactile*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

 

Precision (specificity / accuracy) = 78%

Recall (sensitivity / coverage) = 68%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Hallucinations - Visual

Return to contents

CRIS NLP Service

Brief Description

Application to extract occurrences where visual hallucination is present. Visual hallucinations may be due to a diagnosis of psychosis/schizophrenia or may be due to other causes, e.g. due to substance abuse.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of negative / irrelevant mentions (not included in the output):

“denied any visual hallucination”,

“not responding to visual hallucination”,

“no visual hallucination”,

“if/may/possible/possibly/might have visual hallucinations”,

“monitor for possible visual hallucination”

 

Search term(s): visual hallucinat*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Helplessness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of helplessness.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Ideas of helplessness secondary to her physical symptoms present”,

“ideation compounded by anxiety and a sense of helplessness”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies uselessness or helplessness”,

“no thoughts of hopelessness or helplessness”.

“there a sense of helplessness”,

“helplessness is a common symptom”

 

Search term(s): *helpless*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 93%

Recall (sensitivity / coverage) = 86%

 

Patient level testing done on all patients with primary diagnosis of F32* or F33* (testing done on 30 random documents):

Precision (specificity / accuracy) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2021-056541

Hopelessness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of hopelessness

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“feeling very low and hopeless”,

“says feels hopeless”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies hopelessness”,

“no thoughts of hopelessness or helplessness”

“there a sense of hopelessness”,

“hopelessness is a common symptom”

 

Search term(s):*hopeles*

Evaluated Performance

Cohen's k = 90% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 90%

Recall (sensitivity / coverage) = 95%

 

Patient level testing done on all patients with a primary diagnosis of F32* or F33*:

Precision (specificity / accuracy) = 87%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2021-056541

Hostility

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of hostility.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“increased hostility and paranoia”,

“she presented as hostile to the nurses”

 

Examples of negative / irrelevant mentions (not included in the output):

“not hostile”,

“denied any feelings of hostility”

“he may become hostile”,

“hostility is something to look out for”

 

Search term(s): *hostil*

Evaluated Performance

Cohen's k = 94% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Insomnia

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of insomnia.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“initial insomnia”,

“contributes to her insomnia”,

“problems with insomnia”,

“this has resulted in insomnia”,

“this will address his insomnia”

 

Examples of negative / irrelevant mentions (not included in the output):

“no insomnia”,

“not insomniac"

“Typical symptoms include insomnia”,

“might have insomnia”,

 

Search term(s): *insom*

Evaluated Performance

Cohen's k = 94% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 94%

 

Patient level testing done on all patients with primary diagnosis of F32* or F33* (testing done on 50 random documents):

Precision (specificity / accuracy) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2021-056541

10.1002/gps.6097

Irritability

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of irritability.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“can be irritable”,

“became irritable”,

“appeared irritable”,

“complained of feeling irritable”

 

Examples of negative / irrelevant mentions (not included in the output):

“no evidence of irritability”,

“no longer irritable”,

“irritable bowel syndrome”,

“becomes irritable when unwell”,

 

Search term(s): *irritabl*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 83%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Loss of Coherence

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of incoherence or loss of coherence in speech or thinking.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“patient was incoherent”,

“his speech is characterised by a loss of coherence”

 

Examples of negative / irrelevant mentions (not included in the output):

“patient is coherent”,

“coherence in his thinking”

“coherent discharge plan”,

“could not give me a coherent account”,

 

Search term(s): coheren*, incoheren*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 98%

Recall (sensitivity / coverage) = 95%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 50 random documents):

Precision (specificity / accuracy) = 93%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Low energy

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of low energy.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“low energy”,

“decreased energy”,

“not much energy”,

“no energy”

 

Examples of negative / irrelevant mentions (not included in the output):

“no indications of low energy”,

“increased energy”

“..., might be caused by low energy”,

“monitor for low energy”,

 

Search term(s): *energy*

Evaluated Performance

Cohen's k = 95% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 82%

Recall (sensitivity / coverage) = 85%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 50 random documents).

Precision (specificity / accuracy) = 76%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Mood instability

Return to contents

CRIS NLP Service

Brief Description

This application identifies instances of mood instability.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she continues to have frequent mood swings”,

“expressed fluctuating mood”

 

Examples of negative / irrelevant mentions (not included in the output):

“no mood fluctuation”

“no mood unpredictability”,

“denied diurnal mood variations”

“she had harmed others in the past when her mood changed”,

“tried antidepressants in the past but they led to fluctuations in mood”,

 

Search term(s): "*mood*" and "chang*", "extremes", "fluctuat*", "instability", "*labil*", "*swings*", "*unpredictable*", "unsettled", "unstable", "variable", "*variation*", or "volatile".

Evaluated Performance

Cohen's k = 91% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 70%#

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Mutism

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of mutism.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she has periods of 'mutism”,

“he did not respond any further and remained mute”

 

Examples of negative / irrelevant mentions (not included in the output):

“her mother is mute”,

“muted body language”

 

Search term(s): *mute* , *mutism*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 75%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Negative Symptoms

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of negative symptoms.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she was having negative symptoms”,

“diagnosis of schizophrenia with prominent negative symptoms”

 

Examples of negative / irrelevant mentions (not included in the output):

“no negative symptom”,

“no evidence of negative symptoms”

“symptoms present?”,

“negative symptoms can be debilitating” Definitions:

 

Search term(s): *negative* *symptom*

Evaluated Performance

Cohen's k = 85% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 86%

Recall (sensitivity / coverage) = 95%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Nightmares

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of nightmares.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she was having nightmares”,

“unsettled sleep with vivid nightmares”

 

Examples of negative / irrelevant mentions (not included in the output):

“no nightmares”,

“no complains of having nightmares”

“it’s been a nightmare to get this arranged”,

“a nightmare scenario would be….”

 

Search term(s): nightmare*

Evaluated Performance

Cohen's k = 95% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Obsessive Compulsive Symptoms

Return to contents

CRIS NLP Service

Brief Description

Application to identify obsessive-compulsive symptoms (OCS) in patients with schizophrenia, schizoaffective disorder or bipolar disorder

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

• Text states that patient has OCD features/symptoms

• Text states that patient has OCS

• Text including hoarding, which is considered part of OCS, regardless of presence or absence of specific examples

• Text states that patient has either obsessive or compulsive or rituals or Yale-Brown Obsessive Compulsive Scale (YBOCS) [see keywords below] and one of the following:

o Obsessions or compulsions are described as egodystonic

o Intrusive, cause patient distress or excessive worrying/anxiety

o Patient feels unable to stop obsessions or compulsions

o Patient recognises symptoms are irrational or senseless

• Clinician provides specific YBOCS symptoms

• Text reports that patient has been diagnosed with OCD by clinician

 

Negative annotations of OCS include

• Text makes no mention of OCS

• Text states that patient does not have OCS • Text states that patient has either compulsions or obsessions, not both, and there is no information about any of the following:

o Patient distress

o Obsessive or compulsive symptoms described as egodystonic

o Inability to stop obsessions or compulsions

o Description of specific compulsions or specific obsessions

o Patient insight

• Text states that non-clinician observers (e.g., patient or family/friends) believe patient has obsessions or compulsions without describing YBOCS symptoms.

• Text includes hedge words (i.e., possibly, apparently, seems) that specifically refers to OCS keywords

• Text includes risky, risk-taking or self-harming behaviours

• Text includes romantic or weight-related (food-related) words that modify OCS keywords

Evaluated Performance

Cohen's k = 80% (testing done on 600 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 72%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Paranoia

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of paranoia. Paranoia may be due to a diagnosis of paranoid schizophrenia or may be due to other causes, e.g. substance abuse.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“vague paranoid ideation”,

“caused him to feel paranoid”

 

Examples of negative / irrelevant mentions (not included in the output):

“denied any paranoia”,

“no paranoid feelings”

“relative is paranoid about me”,

“paranoia can cause distress”

 

Search term(s): *paranoi*

Evaluated Performance

Cohen's k = 92% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 86%

Recall (sensitivity / coverage) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

Passivity

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of passivity.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“patient describes experiencing passivity”,

“patient has experienced passivity in the past but not on current admission”

 

Examples of negative / irrelevant mentions (not included in the output):

"denies passivity",

"no passivity".

“passivity could not be discussed”,

“possible passivity requiring further exploration”,

“unclear whether this is passivity or another symptom”

 

Search term(s):passivity

Evaluated Performance

Cohen's k = 83% (testing done on 438 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 89%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Persecutory Ideation

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of ideas of persecution.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she was having delusions of persecution”,

“she suffered persecutory delusions”,

“marked persecutory delusions”,

“paranoid persecutory ideations”,

“persecutory ideas present”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies persecutory delusions”,

“he denied any worries of persecution”,

“this might not be a persecutory belief”,

“no longer experiencing persecutory delusions”

 

Search term(s): [Pp]ersecu*

Evaluated Performance

Cohen's k = 91% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 80%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Poor Appetite

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of poor appetite (negative annotations).

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes- Positive annotations( applied to adjectives implying a good or normal appetite):

“Appetite fine”

“Appetite and sleep OK”,

“Appetite reasonable”,

“appetite alright”,

“sleep and appetite both preserved”

 

Exclude Negative annotations:

“loss of appetite”,

“reduced appetite”,s

“decrease in appetite”,

“not so good appetite”,

“diminished appetite”,

“lack of appetite”

 

Exclude Unknown annotations:

“Loss of appetite as a potential side effect”,

“as an early warning sign, as a description of a diagnosis (rather than patient experience)”, “describing a relative rather than the patient, ‘appetite suppressants’” Definitions: Search term(s): *appetite* within the same sentence of *eat* *well*, *alright*, excellent*, fine*, fair*, good*, healthy, intact*, not too bad*, no problem, not a concern*.

Evaluated Performance

Cohen’s k = 91% (Done on 50 random documents). Instance level, Random sample of 100 random documents: Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 71% Patient level – All patients with primary diagnosis code F32* or F33* in a structured field, random sample of 30 (one document per patient), Precision (specificity / accuracy) = 97%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Poor Concentration

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of poor concentration.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“my concentration is still poor”,

“she found it difficult to concentrate”,

“he finds it hard to concentrate”

 

Examples of negative / irrelevant mentions (not included in the output):

“good attention and concentration”,

“participating well and able to concentrate on activities”

“‘gave her a concentration solution”,

“talk concentrated on her difficulties”,

 

Search term(s): *concentrat*

Evaluated Performance

Cohen's k = 95% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 84%

Recall (sensitivity / coverage) = 60%

 

Patient level testing done on all patients with primary diagnosis code F32* or F33* (testing done on 50 random documents):

Precision (specificity / accuracy) = 76%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Poor Eye Contact

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of poor eye contact.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“looked unkempt, quiet voice, poor eye contact”,

“eye contact was poo”,

“she refused eye contact”,

“throughout the conversation she failed to maintain eye contact”,

“unable to engage in eye contact”,

“eye contact was very limited”,

“no eye contact and constantly looking at floor”

 

Examples of negative / irrelevant mentions (not included in the output):

“good eye contact”,

“he was comfortable with eye contact”,

“she showed increased eye contact”,

“I noticed reduced eye contact today”

 

Search term(s): Available on request

Evaluated Performance

Cohen’s k = 92% (testing done on 100 random documents).

 

Patient level testing done on all patients (testing done on 100 Random Documents):

Precision (specificity / accuracy) = 81%

Recall (sensitivity / coverage) = 65%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Poor Insight

Return to contents

CRIS NLP Service

Brief Description

Applications to identify instances of poor insight.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Lacking/ Lack of insight”

“Doesn’t have insight”

“No/ None insight”

“Poor insight”

“Limited insight”

“Insightless”

“Little insight”

 

Examples of negative / irrelevant mentions (not included in the output):

“Clear insight”

“Had/ Has insight”

“There is a lengthy and unclear description of the patient’s insight, without a final, specific verdict”

“Insight was not assessed”

 

Search term(s): insight

Evaluated Performance

Cohen’s k = 88% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 87%

Recall (sensitivity / coverage) = 70%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1136/bmjopen-2019-028929

Poor Motivation

Return to contents

CRIS NLP Service

Brief Description

This application aims to identify instances of poor motivation.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“poor motivation”,

“unable to motivate’ self”,

“difficult to motivate’ self”,

“struggling with motivation”

 

Examples of negative / irrelevant mentions (not included in the output):

“patient has good general motivation”,

“participate in alcohol rehabilitation”,

“tasks/groups designed for motivation”,

“comments about motivation but not clearly indicating whether this was high or low”,

 

Search term(s): "Motivat*" and "lack*", "poor", "struggle*", or "no"

Evaluated Performance

Cohen's k = 88% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 45%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2015-007619

10.1136/bmjopen-2021-056541

Poverty Of Speech

Return to contents

CRIS NLP Service

Brief Description

Application to identify poverty of speech.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“he continues to display negative symptoms including blunting of affect, poverty of speech”,

“he does have negative symptoms in the form of poverty of speech”

“less poverty of speech”

 

Examples of negative / irrelevant mentions (not included in the output):

“no poverty of speech”,

“poverty of speech not observed”

“poverty of speech is a common symptom of…, “

“?poverty of speech”

 

Search term(s): speech within the same sentence of poverty, impoverish.

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 87%

Recall (sensitivity / coverage) = 85%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2015-007619

Poverty Of Thought

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of poverty of thought.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“poverty of thought was very striking”,

“evidence of poverty of thought”,

“some poverty of thought”

 

Examples of negative / irrelevant mentions (not included in the output):

“no poverty of thought”,

“no evidence of poverty of thought”

“poverty of thought needs to be assessed”,

“poverty of thought among other symptoms”

 

Search term(s): *poverty* *of* *thought*

Evaluated Performance

Cohen's k = 90% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 93%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Psychomotor Activity (Catergorisation)

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of psychomotor activity and determine the level of activity

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes- Positive annotations:

Positive/correct mentions identifies the level of psychomotor activity (as denoted by the keyword) in the context (as denoted by the contextstring). In addition, psychomotor_activity column correctly states whether the reference to abnormal levels of psychomotor activity in the contextstring.

For example: Keyword: ‘psychomotor agitation’; Contextstring: ‘patient showed psychomotor agitation’;

Negativity: ‘No’; psychomotor_activity: ‘psychomotor agitation’

 

Negative/incorrect/irrelevant mentions do not successfully identify the level of activity (as denoted by the keyword) in the context (as denoted by the contextstring). Or an instance of psychomotor activity is noted as negated.

 

For example: Keyword: ‘psychomotor activity’; Contextstring: ‘normal psychomotor activity’; Negativity: ‘yes’;psychomotor_activity’: ‘psychomotor activity’

Keyword: ‘psychomotor activity’; Contextstring: ‘change in psychomotor activity’; Negativity: ‘yes’; psychomotor_activity: ‘psychomotor activity’

Evaluated Performance

Cohen's k = xx Instance level, Random sample of 100 Random Documents: Precision (specificity / accuracy) = 92% Recall (sensitivity / coverage) = 92%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Self Injurious Action

Return to contents

CRIS NLP Service

Brief Description

This application will identify descriptions of deliberate self injurious action taken by the subject of the record

Development Approach

Development approach: BERT Model

Classes produced: Annotations are positive where the text in the text field refers to deliberate self injurious actions, or where texts that contain segments of forms that positively note self-injurious behaviour.

Search term(s): 'self-har*', 'Self-injur*', 'Deliberate self harm'.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"The Patient harms himself regularly"

"The Patient has a history of deliberate self harm"

(in a Form): DSH (Y) N

 

Examples of negative / irrelevant mentions (not included in the output):

"Has had thoughts about self-harm"

"High risk of self-injurious behaviour"

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 84%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Smell

Return to contents

CRIS NLP Service

Brief Description

Application to identify symptoms of loss of smell.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ she has not recovered her sense of smell since she contracted COVID-19 in May 2021”,

“ Complains of loss of smell and loss of tastes”

 

Examples of negative / irrelevant mentions (not included in the output):

“Negative annotations include denies any symptoms of loss of smell”,

” Her mother could not smell the food she made”.

“ no one else could smell it either”,

“ she was unsure whether her smell had been affected”

 

Search term(s): Loss of smell/ Lack of smell

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Social Withdrawal

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of social withdrawal.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes- Positive annotations:

“she is withdrawn socially from friends and family”,

“Mr ZZZZZ became very isolated and socially withdrawn”,

“ some social withdrawal”

 

Exclude Negative annotations:

“not being socially withdrawn”,

“no evidence of being socially withdrawn”

 

Exclude ‘Unknown’ annotations:

“social withdrawal is common in depression”,

“need to ask about social withdrawal”. Definitions: Search term(s): Social within the same sentence of withdraw.

Evaluated Performance

Cohen's k = 100% (50 un-annotated documents - 25 events/25 attachments, search term ‘withdraw*’). Instance level, Random sample of 100 Random Documents: Precision (specificity / accuracy) = 60% Recall (sensitivity / coverage) = 86% Patient level – Random sample of 30 (one document per patient)

Precision (specificity / accuracy) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2016-012012

10.1136/bmjopen-2015-007619

Stupor

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of stupor. This includes depressive stupor, psychotic stupor, catatonic stupor, dissociative stupor and manic stupor.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZ presented in a psychotic stupor”,

“man with stuporous catatonia”,

“he is in a depressive stupor”,

“his presentation being a schizoaffective stupor”,

“periods of being less responsive/stuporous”

 

Examples of negative / irrelevant mentions (not included in the output):

“not in the state of stupor”,

“presentation not suggestive of depressive stupor”,

“?manic stupor”,

“drink himself to stupor”,

 

Search term(s): Stupor*

Evaluated Performance

Cohen's k = 96% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 88%

Recall (sensitivity / coverage) = 87%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1017/S0033291721004402

10.1136/bmjopen-2016-012012

Suicidal Ideation

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of suicidal ideation - thinking about, considering, or planning suicide.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Her main concerns were his low mood QQQQQ suicidal ideation”,

“He has recently sent a letter to mom describing suicidal ideation”,

“QQQQQ then advised of suicidal ideation”

 

Examples of negative / irrelevant mentions (not included in the output):

“There was no immediate risk in relation to self-harm or current suicidal ideation”,

“There has been no self-harm and no suicidal ideation disclosed to QQQQQ”,

“Suicidal ideation is a common symptom in depression”,

“It wasn’t certain if she was experiencing suicidal ideation”

 

Search term(s): *suicide* ideat*

Evaluated Performance

Cohen's k = 92% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 81%

Recall (sensitivity / coverage) = 87%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Tangentiality

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of tangentiality.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ he was very tangential lacked goal directed thinking”,

“ there was evidence of tangential speech”

 

Examples of negative / irrelevant mentions (not included in the output):

“ no evidence of formal thought disorder or tangentiality of thoughts”,

“there was no overt tangentiality or loosening of associations”

“there can be tangentiality”,

“FTD is characterised by tangentiality”,

“ go off on a tangent”

 

Search term(s): *tangent*

Evaluated Performance

Cohen's k = 81% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 99%

Recall (sensitivity / coverage) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Taste

Return to contents

CRIS NLP Service

Brief Description

Application to identify symptoms of loss of taste within community populations

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“the patient reported loss of enjoyment of food due to loss of taste”,

“COVID symptoms present such as loss of taste”

 

Examples of negative / irrelevant mentions (not included in the output):

“the patient denied loss of taste”,

“patients’ mother reported loss of taste due to COVID”

 

Examples of negative / irrelevant mentions (not included in the output):

“the patient is not sure if he has lost his taste”,

“don’t come to the practice if you have any COVID symptoms such as loss of taste etc"

 

Search term(s): Loss of taste*, lack of taste*

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Tearfulness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of tearfulness

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“appeared tearful”,

“was tearful (including was XX and tearful; was tearful and YY)”,

“became tearful”,

“moments of tearfulness”,

“a bit tearful”

 

Examples of negative / irrelevant mentions (not included in the output):

“not tearful”,

“no tearfulness”,

“less tearful”,

“couldn’t remember being tearful”

 

Search term(s): *tearful*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2021-056541

Thought Block

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of thought block.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“showed some thought block”,

“thought block”,

“paucity of thought”

 

Examples of negative / irrelevant mentions (not included in the output):

“denies problems with thought block”,

“no thought block elicited”

“thought block can be difficult to assess”,

“ …among thought block and other symptoms”

 

Search term(s): *thought* *block*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents)

 

Instance level (testing done on 100 Random Documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 75%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Thought Broadcast

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of thought broadcasting.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"patient describes experiencing thought broadcasting",

"patient has experienced thought broadcasting in the past but not on current admission".

 

Examples of negative / irrelevant mentions (not included in the output):

"denies thought broadcasting",

"no thought broadcasting".

" thought broadcasting could not be discussed",

"possible thought broadcasting requiring further exploration"

 

Search term(s): Though* within the same sentence of broadcast*

Evaluated Performance

Cohen's k = 94% (testing done on 95 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 86%

Recall (sensitivity / coverage) = 92%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Thought Insertion

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of thought insertion.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"patient describes experiencing thought insertion" or "patient has experienced thought insertion in the past but not on current admission".

 

Examples of negative / irrelevant mentions (not included in the output):

"denies thought insertion",

"No thought insertion".

"thought insertion could not be discussed",

"possible thought insertion requiring further exploration"

 

Search term(s): "Though*" and "insert*"

Evaluated Performance

Cohen's k = 97% (testing done on 96 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 81%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – On Request

Other Specifications

Version 1.0, Last updated:xx

Thought Withdrawal

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of thought withdrawal.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"patient describes experiencing thought withdrawal" ,

"patient has experienced thought withdrawal in the past but not on current admission".

 

Examples of negative / irrelevant mentions (not included in the output):

"denies thought withdrawal",

"no thought withdrawal".

"thought withdrawal could not be discussed",

"possible thought withdrawal requiring further exploration",

 

Search term(s): "Though*" and "withdraw*"

Evaluated Performance

Cohen's k = 95% (testing done on 76 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 90%

Recall (sensitivity / coverage) = 88%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1136/bmjopen-2021-057433

Waxy Flexibility

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of waxy flexibility. Waxy flexibility is a psychomotor symptom of catatonia as associated with schizophrenia, bipolar disorder, or other mental disorders which leads to a decreased response to stimuli and a tendency to remain in an immobile posture.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“she presents as catatonic with waxy flexibility”,

“exhibiting waxy flexibility”

 

Examples of negative / irrelevant mentions (not included in the output):

“no waxy flexibility”,

“no evidence of waxy flexibility”

“his right pre-tibial region was swollen and waxy and slightly pink”,

“waxy flexibility is a very uncommon symptom”

 

Search term(s): *waxy*

Evaluated Performance

Cohen's k = 96% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 80%

Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1017/S0033291721004402

10.1136/bmjopen-2016-012012

Weight Loss

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of weight loss.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“significant weight loss”,

“pleased with his weight loss”

 

Examples of negative / irrelevant mentions (not included in the output):

“no weight loss”,

“denies weight loss”.

“maintain adequate dietary intake and avoid weight loss”,

“the latter reduced in line with weight loss”

 

Search term(s): "weight*" and "loss" or "lost"

Evaluated Performance

Cohen’s k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 90%

Recall (sensitivity / coverage) = 88%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2021-056541

10.1002/gps.6097

10.1002/gps.5659

Worthlessness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of worthlessness

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“feeling worthless”,

“feels hopeless and worthless”

 

Examples of negative / irrelevant mentions (not included in the output):

“no worthlessness”,

“denies feelings of worthlessness”

“his father had told him that he was worthless”,

“would call them worthless”

 

Search term(s): *worthless*

Evaluated Performance

Cohen's k = 82% (50 un-annotated documents - 25 events/25 attachments, search term ‘worthless*’). Instance level, Random sample of 100 Random Documents: Precision (specificity / accuracy) = 88% Recall (sensitivity / coverage) = 86% Patient level – Random sample of 30 (one document per patient)

Precision (specificity / accuracy) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1192/j.eurpsy.2021.18

10.1136/bmjopen-2021-056541

Physical Health Conditions

Asthma

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnosis of asthma.

Development Approach

Development approach: Sem-EHR

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

‘past medical history: eczema, asthma’,

‘diagnosed with asthma during childhood’,

‘uses inhaler to manage asthma symptoms’,

‘suffered from an asthma attack’,

‘ZZZZZ suffers from severe asthma’,

‘Mrs ZZZZZ has mild asthma’.

 

Search term(s): Ontology available on request

Evaluated Performance

Cohen’s k = 98% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 95%

Recall (sensitivity / coverage) = 84%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Bronchitis

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnosis of bronchitis

Development Approach

Development approach: Sem-EHR

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

‘Recently had COPD (chronic obstructive pulmonary disease’,

‘ZZZZ had chronic bronchitis,

‘Past diagnosis: chronic obstructive airway disease’,

‘physical health history: asthma, bronchitis’,

‘centrilobular emphysema’.

 

Search term(s): Ontology available on request

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 48%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Cough

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of coughing.

Development Approach

Development approach: Machine- learning

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“She has been experiencing a cough for the last week and is going to call her GP”,

“ZZZ called ahead of today’s session reporting a cough so we agreed to move the session to over the phone due to current COVID guidance”,

“He has been to the GP due to coughing up sputum”

 

Examples of negative / irrelevant mentions (not included in the output):

“she denied any coughing or shortness of breath”,

“He stated he was unwell with a cold last week, no cough or cough reported”,

“She is feeling very distresses because people were coughing near her on the bus”,

“Her son is currently off school with bad cough”

 

Search term(s): Cough*

Evaluated Performance

Cohen's k = 79% (testing done on 150 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 80%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Crohn's Disease

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnosis of Crohn’s disease.

Development Approach

Development approach: Sem-EHR

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“recently been diagnosed with crohn’s disease”,

“ZZZZ has crohn’s disease”,

“she has a history of crohn’s disease”,

“has been hospitalised due to severe crohn’s disease”,

“physical health history: asthma, diabetes, hypertension, crohn’s disease”

 

Search term(s): Ontology available on request

Evaluated Performance

Cohen’s k = 98% (testing done on 50 random documents).

 

Patient level test done on all patients (testing done on 50 random documents):

Precision (specificity / accuracy) = 94%

Recall (sensitivity / coverage) = 78%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Falls

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of falls or falling.

Development Approach

Development approach: Rules-Based

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

Output:

• Fall_single_episode, any reference to a single fall (regardless of when it happened) e.g., ‘he fell last night’, ‘he had one fall 10 years ago’).

• Fall_recurrent: any reference to more than one fall (regardless of when they happened), e.g. ‘he reported recurrent falls’, ‘she had a couple of falls.

• Not_relevant: to capture irrelevant mentions or false positives, e.g. ‘in the fall’, ‘falling in love’ or any other fall mention such as risk of falling, side effect of this medication is risk of falling.

 

Note 1: positive annotations must refer to the patient and not someone else.

His mother had one fall > NOT_RELEVANT

 

Note 2: hypothetical statements should not be counted

If she took this medication, she might be at risk of falling > NOT_RELEVANT

 

Note 3: classes should be chosen on an annotation level: “She had a fall 10 months ago and then had another fall yesterday” should end up as two single-episode annotations, but “she had a couple of falls: 10 months ago and yesterday” would end up as a FALL_RECURRENT

 

Note 4: accidental falls are to be considered relevant

He fell from the bed > FALL_SINGLE_EPISODE

 

Note 5: mentions where a fall is "suggested" but not explicitly written (e.g. 'Fall pendant', 'Falls clinic', 'Falls referral', 'Falls prevention advice') should be considered as NOT_RELEVANT Definitions: Search term(s): Fall*, fell

Evaluated Performance

Cohen's k = xx

 

Patient level testing done on all patients (testing done on 50 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 58%

Additional Notes

Run schedule – On Request

Other Specifications

Version 1.0, Last updated:xx

Fever

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with any symptom of fever developed within the last month.

Development Approach

Development approach: Machine-learning

Classification of past or present symptom: Past.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“She informed me on the phone she has had a fever all week”,

“ZZZ has been taking paracetamol for a fever”,

“Attended A&E reporting fever”,

“She felt feverish”

 

Examples of negative / irrelevant mentions (not included in the output):

“I asked if she had any symptoms, such as fever, which she denied”,

“Temperature was checked for signs of fever, none observed”

“Her son had a fever last night and she can’t make it to today’s session”

“She reported worrying over what to do if the baby developed a fever”

 

Search term(s): fever*

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – On Request

Other Specifications

Version 1.0, Last updated:xx

Hypertension

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnosis of hypertension or high blood pressure

Development Approach

Development approach: Sem-EHR

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Recently been diagnosed with hypertension”,

“ZZZZ has high blood pressure”,

“she has a history of hypertension”,

“physical health history: asthma, diabetes, high blood pressure”

 

Search term(s): Ontology available on request

Evaluated Performance

Cohen’s k = 91% (testing done on 50 random documents).

 

Instance level (testing done on 200 random documents):

Precision (specificity / accuracy) = 94%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Multimorbidity - 21 Long Term Conditions (Medcat)

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnosis of physical health conditions (21 conditions in total, including arthritis, asthma, atrial fibrillation, cerebrovascular accident, chronic kidney disease, chronic liver disease, chronic obstructive lung disease, chronic sinusitis, coronary arteriosclerosis, diabetes mellitus, eczema, epilepsy, heart failure, systemic arterial hypertensive disorder, inflammatory bowel disease, ischemic heart disease, migraine, multiple sclerosis, myocardial infarction, parkinson's disease, psoriasis, transient ischemic attack).

Development Approach

Development approach: Machine-learning

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“He reported that he suffers from diabetes and hypertension”,

“Ms ZZZZ has a history of atopy including asthma”,

“Physical health history: asthma, diabetes, high blood pressure: Nil” ,

“Physical health: lung disease confirmed”

Evaluated Performance

Cohen's k = 91% (testing done on 50 random documents)

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Pain

Return to contents

CRIS NLP Service

Brief Description

Application to determine if a mention of pain (or related words, such as sore, ache, *algia, *dynia etc.) within the text is relevant i.e. associated with the patient and refers to physical pain.

Development Approach

Development approach: Machine-learning

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

‘she is in constant pain’;

‘he suffers from severe headaches’;

he is taking pain killers due to a pulled muscle’ Definitions:

 

Search term(s): %dynia%, '%algia%, %burn%', % headache%, % backache%, % toothache%, % earache%, % ache%, %sore%, %spasm%, % colic%, % cramp%, % hurt%, % sciatic%, % tender%, % pain %, % pains%, % painful%

Evaluated Performance

Cohen’s k = 86% for Attachment (testing done on 865 random documents)

Cohen’s k = 91% for Event (testing done on 458 random documents).

 

Patient level testing done on all patients (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 78%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Rheumatoid Arthritis

Return to contents

CRIS NLP Service

Brief Description

Application to identify patients with diagnoses of rheumatoid arthritis.

Development Approach

Development approach: Sem-EHR

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“ZZZZZ has been in pain due to her rheumatoid arthritis”,

“she has been bedbound with rheumatoid arthritis this week”,

“medication for her rheumatoid arthritis”,

“physical health comorbidities: hypertension, rheumatoid arthritis”,

“diagnosed with rheumatoid arthritis is 1988”

 

Search term(s): Ontology available on request

Evaluated Performance

Cohen’s k = 98% (testing done on 50 random documents).

 

Patient level testing done all patients (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

HIV

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of HIV diagnosis.

Development Approach

Development approach: Machine-learning

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

ZZZZZ was diagnosed with HIV (only include cases where a definite HIV diagnosis is present in the text)

 

Search term(s): hiv

Evaluated Performance

Cohen's k = 98% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 70%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

HIV Treatment

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of HIV treatment.

Development Approach

Development approach: Machine-learning

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

Include any positive references to the search terms below.

 

Search term(s): Anti-retroviral, antiretroviral, ARV, HAART, cART , ART, CD4, Undetectable, Abacavir, Lamivudine, Zidovudine, Aptivus, Atazanavir, Atripla, Celsentri, Cobicistat, Combivir, Darunavir, Didanosine, Dolutegravir, Edurant, Efavirenz, Elvitegravir, Emtricitabine, Emtricitabine, Emtriva, Enfuvirtide, Epivir, Etravirine, Eviplera,

Fosamprenavir.

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 98%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Contextual Factors

Amphetamine

Return to contents

CRIS NLP Service

Brief Description

To identify instances of amphetamine use.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“denies current use of amphetamine, however last reported using 3 months ago”,

“first took amphetamines at the age of 15”,

“UDS: +ve amphetamine”,

“ZZZZZ has been trying to give up amphetamine for the last 2 months”,

“ZZZZZ was found in possession of large quantities of amphetamines”,

“She admitted to having bought amphetamine 2 days ago” ,

“amphetamine-psychosis”

 

Examples of negative / irrelevant mentions (not included in the output):

“ZZZZZ denies use of alcohol and amphetamine”,

“ZZZZZ has not used amphetamine for the last week”,

“ZZZZZZ’s mother has a history of amphetamine abuse” – subject other than patient,

“ZZZZZ is planning on taking amphetamine this weekend” – future or conditional event,

 

Search term(s): Amphetamin*

Evaluated Performance

Cohen's k = 84% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 80%

Recall (sensitivity / coverage) = 84%

Additional Notes

Run schedule – Monthly

Other Specifications

Version xx, Last updated:xx

Cannabis

Return to contents

CRIS NLP Service

Brief Description

To identify instances of cannabis use.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“He is a cannabis smoker”,

“she smoked cannabis when at uni”

“she stopped using cannabis 3 years ago”

 

Examples of negative / irrelevant mentions (not included in the output):

“denied taking any drugs including cannabis”,

“no cannabis use”

“she stated in hash voice”,

“pot of yoghurt”

 

Search term(s): cannabis ,skunk, weed, Pot, marijuana, grass ,THC, hash, cannabinoids, resin, hashish, weeds, Cannabis- ,spices, Spice, ganja, CBD, cannabis-induced, Cannabinoid, cannabies, grasses, Cannaboids, marijuana, cannabbase, cannabis-free, skunk- cannabis, Hashis, cannabis-related, cannabi, cannabise, cannabinoids, cannabis-use, marijuana, cannabus, cannabiss, weed- skunks

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 93%

 

Current instance level (testing done on 30 random documents)

Precision (specificity / accuracy) = 72%

Additional Notes

Run schedule – Monthly

Other Specifications

Version xx, Last updated:xx

Chronic Alcohol Abuse

Return to contents

CRIS NLP Service

Brief Description

This application identifies instances of Chronic Alcohol Abuse within the text of CRIS clinical texts where Chronic Alcohol Abuse in the subject of the text is mentioned.

Development Approach

Development approach: Machine-learning.

Search terms(s): ‘Alcoholism’, ‘an alcoholic’, ‘problem* drinker’, ‘drinking problem’, ‘problem with drink*’, ‘problem with alcohol’, ‘alcohol problem’, ‘excessive drinker’, ‘drink* in excess’, ‘consumes alcohol excessively’, ‘consumes alcohol in excess’, ‘heavy drinking’, ‘drink* heavily’, ‘drink* excessively’, ‘alcohol related disorder’, ‘alcohol use disorder’, ‘consumes excessive amounts of alcohol’, ‘regularly gets drunk’.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"zzzz has a drinking problem"; or "zzz regularly drinks too much"; or "zzz is an alcoholic".

"zzz had a historic drinking problem"; or "zzz used to drink too much"

 

Examples of negative / irrelevant mentions (not included in the output):

"zzz drank too much last night"

"used to drink excessively but does not anymore"

Evaluated Performance

Cohen’s k = 92% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Cocaine or Crack Cocaine

Return to contents

CRIS NLP Service

Brief Description

To identify instances of cocaine or crack cocaine use.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“denies current use of cocaine, however last reported using 3 months ago”,

“first smoked cocaine at the age of 15”,

“UDS: +ve cocaine”,

“ZZZZZ has been trying to give up cocaine for the last 2 months”,

“ZZZZZ was found in possession of large quantities of cocaine”,

“She admitted to having bought cocaine 2 days ago” ,

“He has stopped taking cocaine”.

 

Examples of negative / irrelevant mentions (not included in the output):

“ZZZZZ denies use of street drugs such as cocaine”,

“ZZZZZ has not used cocaine for the last week”,

“ZZZZZZ’s mother has a history of crack abuse” – another subject other than the patient,

“ZZZZ is planning on taking cocaine this weekend” – future or conditional events,

 

Search term(s): Cocaine*, crack

Evaluated Performance

Cohen's k = 95% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 84%

Recall (sensitivity / coverage) = 97%

Additional Notes

Run schedule – Monthly

Other Specifications

Version xx, Last updated:xx

MDMA

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of MDMA use

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“denies current use of MDMA, however last reported using 3 months ago”,

“first took MDMA at the age of 15”,

“UDS: +ve MDMA”,

“ZZZZZ has been trying to give up MDMA for the last 2 months”,

“ZZZZZ was found in possession of large quantities of MDMA”,

“She admitted to having bought MDMA 2 days ago”,

“He has stopped taking MDMA”

 

Examples of negative / irrelevant mentions (not included in the output):

“ZZZZZ denies use of street drugs such as MDMA”,

“ZZZZZ has not used MDMA for the last week”,

“ZZZZZZ’s mother has a history of MDMA abuse” – another subject other than the patient,

“ZZZZ is planning on taking MDMA this weekend” – future or conditional events,

 

Search term(s): mdma

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 99%

Additional Notes

Run schedule – Monthly

Other Specifications

Version xx, Last updated:xx

Smoking

Return to contents

CRIS NLP Service

Brief Description

This application distinguishes between people who are a) current smokers, b) current non-smokers (ever smoked) and c) non-smokers. This application may at times bring back contradictory information on the same patient since patient may start smoking and stop smoking and because of the varied level of information available to the clinician.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

 

Output and Definitions

The output includes-

 

“…is a non-smoker”, “… was/is not a smoker”, “… doesn’t smoke”, “ZZZZZ denies ever smoking”, or “… is currently not smoking”

 

“…smokes 20 cigarettes a day”, “… has been smoking for 10 years”, “…is a smoker”, “ZZZZZ smokes in the ward”,

“…went to garden for a smoke”, “ZZZZZ is stable when smoking”, “…has a history of heavy smoking”, “Consider stopping smoking”, “ZZZZZ found smoking in her room” or “… is a tobacco user”

 

“… used to smoke”, “… has quitted smoking”, “… stopped smoking”, “ZZZZZ is an ex-smoker” or “…was a smoker”.

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 85%

Recall (sensitivity / coverage) = 89%

Additional Notes

Run schedule – Weekly

Other Specifications

Version xx, Last updated:xx

Education

Return to contents

CRIS NLP Service

Brief Description

Application to identify the highest level of education at patient level.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Examples of positive mentions:

Group 1: A level group

 

Rule Stage of course

Accepted: Accepted for A-level course or equivalent (course or institution)

Ongoing: Started course but not (yet) completed (including evidence of attending relevant institution)

Dropped Out: Started course but not completed - dropped out

Expelled: Started course but not completed - expelled

Failed: Completed course – failed all exams

Completed: Completed course

Passes: Passed at least one exam

Applied_undergrad: Applied for university / course

 

Note: aspirations, plans, application only are not accepted.

Group 3: University

 

Rule Stage of Course

Accepted: Accepted for course / institution

Ongoing: Started course but not (yet) completed

Dropped out: Started course but not completed - dropped out

Expelled: Started course but not completed - expelled

Failed: Completed course – failed

Completed: Completed course

Passed: Passed / graduated

Applied_University: Applied for University

 

 

Group 4: unqualified group

 

Rule Definition

Unqualified: A specific reference in notes describing as having left school without any qualifications.

GSCE_Dropped_out: Started GCSE course but not completed - dropped out

GSCE_Expelled: Started GCSE course but not completed - expelled

GSCE_Failed: Completed GCSE course – failed all exams

 

School leaving age

 

Examples

He left school at the age of 16 years

Was 19 years old when she left school

Mrs ZZZZZ left school at 15 without any qualifications

 

 

Group 2: GCSE group

 

Rule Stage of Course

Ongoing: Started GCSE course (or equivalent) but not (yet) completed

Completed: Completed GCSE course or equivalent

Passed: Passed at least one exam (GSCE or equivalent)

Applied_A-level: Applied for 6th form (college) / A-level

Evaluated Performance

GCSE:

Cohen's k = 90% (testing done on50 random documents).

 

No qualifications:

Cohen’s k = 100% (testing done on 50 random documents).

Additional Notes

Run schedule – On request

Other Specifications

Version xx, Last updated:xx

Occupation

Return to contents

CRIS NLP Service

Brief Description

Application to identify occupations/work descriptions and who these refer to.

Development Approach

Development approach: Machine- learning and Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Examples of positive mentions:

There are two parts to each annotation: Firstly, the occupation feature is annotated - this could be a job title, for example a ‘builder’; or a job description, for example ‘working in construction’. Secondly, the occupation relation is annotated: who the occupation belongs to, for example the patient or their family member.

 

Unpaid occupational categories were included (e.g. student, unemployed, homemaker, volunteer). Depending on the text available, extractions can state a specific job title (e.g. head-teacher) or a general occupational category (e.g. self-employed).

 

Work aspirations were excluded from annotations. Frequently extracted health/social care occupations (e.g. psychiatrist) are not annotated as belonging to the patient, in order to maximise precision.

Occupation feature (text) – the job title (e.g. ‘hairdresser’)

Occupation relation (text) – who the occupation belongs to (e.g. ‘patient’)

 

The full annotation guideline document is available on request.

 

Search term(s): Gazetteer available on request.

Evaluated Performance

Cohen’s k = 77% for occupation feature (testing done on 200 random documents).

 

Cohen’s k = 72% for occupation relation (testing done on 200 random documents).

 

Instance level (testing done on 200 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 79%

 

Patient level testing done on all patietns aged >= 16 years (testing done on 82 random documents):

Precision (specificity / accuracy) = 96%

Additional Notes

Run schedule – On request

Other Specifications

Version xx, Last updated:xx

Lives Alone

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of living alone.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Lives on her own”, Who- none ,

“She lives alone”, Who- She

“He presently lives alone on 7th floor”, Subject – He

“His father lives alone”, Subject – Father

 

Search term(s): "Lives alone", "Lives by himself", "Lives by herself", "Lives on his own", "Lives on her own".

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 83%

Precision (Subject) = 61%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Loneliness

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of loneliness.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“the patient is lonely”,

“the patient confirms they have a sense/feeling of loneliness”,

“preventing further loneliness”

 

Examples of negative / irrelevant mentions (not included in the output):

“Patient is not lonely”,

“denies being lonely”,

“the patient’s family member is lonely”;

“they are participating in an activity on a ward to prevent boredom/loneliness”.

 

Search term(s): lonely, loneliness.

Evaluated Performance

Cohen’s k= 81% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 87%

Recall (sensitivity / coverage) = 100%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1002/gps.6097

10.1007/s00127-024-02663-9

10.1002/gps.5630

10.1007/s00127-021-02079-9

Violence

Return to contents

CRIS NLP Service

Brief Description

Application to identify and classify different types of violence.

Development Approach

Development approach: Bert model

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Violence_type – It will describe difference violence types, values are Emotional, Financial, Irrelevant, Physical(non-sexual), Sexual and Unspecified.

 

Physical (non-sexual): e.g., punching, hitting, slapping, attach with a weapon, that is not sexual in nature.

‘She was beaten by her boyfriend’ or ‘someone spat at him’

 

Search term(s): “abus”, “assault”, “attack”, “beat”, “fight”, “hit”, “punch, “push”, “threw”, “violenc”.

 

Sexual: e.g., sexual assault, rape.

‘Patient showed sexually inappropriate behaviour’ or ‘zzz was raped when she was 12 years old’

 

Search term(s): “abus”, “assault”, “rape”.

 

Emotional: e.g., gaslighting psychological abuse

‘He was psychological abused from the age of 10’ or ‘She is subjected to emotional abuse by her boyfriend’

 

Search term(s): “abus”, “emotional abus”, “emotional manipulat”, “emotionally abus”, “gaslight”, “psychological abus”.

 

Financial: financial and economic abuse e.g., withholding of money

‘zzz has suffered from financial abuse’ or ‘She has had problems with stress, money problem, financial abuse and relationship problems’

 

Search term(s): “abus”, “assault”, “economic abus”, “financial abus”, “financially abus”, “struck”, “violenc”.

 

Unspecified: violence is being discussed but the type of violence has not otherwise been specified.

‘He was abuse as a child by his father’ or ‘he has a history of being abused’

 

Irrelevant: not a mention of violence, the keyword is ambiguous and, in the sense used, it does not refer to violence. In this case, no other attributes should be added

‘Patient mentioned that her child was exposed to an abuse relationship’ or ‘He often feels this as a stabbing pain’

 

Temporality – it will describe whether event is past or not past. Values are Past and Not_Past. Details of values are:

Past: The violence occurred more than a year ago.

E.g., ‘zzz was sexually abuse several years ago in a past relationship’ or ‘’zzz was subjected to emotional abuse throughout childhood (when discussing an adult)

Recent: Violence is occurring now, in the past year.

e.g., ‘zzz was hit last week’

Unclear: It is not clear if the violence being described is past or recent.

e.g., ‘zzz was coerced in to taking part against his will’

 

Presence- Values are actual and unclear.

Threat: e.g., ‘He made a gesture as if to attack staff’ or ‘The patient threatened to hit me’

Actual: e.g., ‘he hit the nurse’

255

 

Kw_text – This will pick up the keyword from the text, based on that it will decide violence type.

 

Polarity- Values are Abstract, Affirmed and Negated. Details for values are given below:

Affirmed- The mention of violence is discussing something has happened, including threats.

Negated - The mention of violence is discussing something that has not happened, such as the absence of violence.

e.g., ‘No violence or aggression noted’

Abstract - Violence is mentioned, but is being conjectured, speculated, or hypothesised about. For example, possible violence or risks of violence.

‘Clinician wondered whether there was emotional abuse’

 

Patient_role – values are Perpetrator, Unclear, Victim and Witness. Details are:

Victim: The patient was the victim of the violence.

e.g., ‘zzz mentioned that she was gaslighted by ex-partner’

Perpetrator: The patient perpetrated the violence

e.g., ‘Patient made threats to kill and attack staff’

Witness: The patient witnessed the violence, rather than being a victim or perpetrator.

e.g., ‘zzz saw his sister being punch by dad’

 

Setting – values are Domestic, not domestic and Not known. Details are:

Domestic: Violence occurred in a domestic setting, including intimate partner violence (family members, intimate partners, ex-intimate partners and household members).

e.g., ‘Patient stabbed his roommate’ or ‘She was hit by her boyfriend’

Not domestic: Violence did not happen in a domestic setting.

e.g., ‘She was abused whilst in care’

Not known: Whether or not the setting was domestic is not known.

Evaluated Performance

Physical Violence (testing done on 50 random documents):

Precision (specificity / accuracy): Type=76%, Setting=78%, Presence=78%, Patient_role=64%, Polarity=74%.

 

Financial Violence (testing done on 50 random documents):

Precision (specificity / accuracy): Type=98%, Setting=86%, Presence=96%, Patient_role=92%, Polarity=96% .

 

Emotional Violence (testing done on 50 random documents):

Precision (specificity / accuracy): Type=100%, Settings=98%, Presence=100%, Patient_role=96%, Polarity=100%.

 

Sexual Violence (testing done on 50 random documents):

Precision (specificity / accuracy): Type=82%, Settings=90%, Presence=84%, Patient_role=78%

Additional Notes

Run schedule – Quarterly

Other Specifications

Version 2.0, Last updated:xx

Interventions

CAMHS - Creative Therapy

Return to contents

CRIS NLP Service

Brief Description

This application is to extract creative therapy intervention that covers art, play, music and drama interventions provided by CAMHS for children and young people from the free text.

Development Approach

Development approach: Text hunter

Classification of past or present instance: Both.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

"engaging with ART therapy",

"she joined others in music and dancing session. Staff engages her with musical instruments. ZZZZZ appears pleasant in the music session",

"had an art session with the O.T this afternoon and did colouring".

 

Examples of negative / irrelevant mentions (not included in the output):

"mum also thought that art therapy separately for ZZZZZ could be helpful",

"mum keen to access therapy for children, either art/play therapy or counselling to help process feelings around the accident".

 

Search term(s): "art", "session", "play", "music", or "drama" and “assessment”, “need*”, “intake”, “appointment”, “appt”, “support”, “intervention”, “session”, “saw”, “therapy”, “follow up”, “refresher”, or “top-up”.

 

“art: seen”, “play: seen”, “music: seen”, “drama: seen”, “art: reviewed”, “play: reviewed”, “music: reviewed”, “drama: reviewed”.

 

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 84%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

CAMHS - Dialectical Behaviour Therapy (DBT)

Return to contents

CRIS NLP Service

Brief Description

This application is to extract DBT intervention provided by CAMHS for children and young people from the free text.

Development Approach

Development approach: Text hunter

Classification of past or present instance: Both.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Attended a DBT group yesterday and needed some encouragement",

"She frequently attends groups help by the therapy such as DBT and Movement therapy.",

"ZZZZZ came to his first DBT group",

 

Examples of negative / irrelevant mentions (not included in the output):

"It discuss about mum's DBT therapy",

"parents/carers to be invited to DBT assessment:",

"A number of options for further support were discussed including DBT skills sessions".

 

Search term(s): “Dialectical Behaviour Therapy” or "DBT" and “assessment”, “need*”, “intake”, “appointment”, “Appt”, “support”, “intervention”, “session”, “therapy”, “saw”, “attended”, “continued”, “follow up”, “refresher”, or “top-up”.

 

“Dialectical Behaviour Therapy: seen”, “DBT: seen”, “Dialectical Behaviour Therapy: reviewed”, “DBT: reviewed”, “Dialectical Behaviour Therapy: reviewed”.

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 92%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

CAMHS - Psychotherapy/Psychosocial Intervention

Return to contents

CRIS NLP Service

Brief Description

This application is to extract the psychosocial intervention/psychotherapy provided by CAMHS for children and young people from the free text.

Development Approach

Development approach: Text hunter

Classification of past or present instance: Both.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Psychotherapy assessment - 2nd parent meeting",

"Psychotherapy assessment review meeting with dad, and myself",

"Systemic Psychotherapist conducted the current assessment of ZZZZZ and her family"

 

Examples of negative / irrelevant mentions (not included in the output):

"Hence could not attend their regular psychotherapy appointment.",

"ZZZZZ is attending 1:1 counselling for his emotional needs with a trained counsellor."

 

Search term(s):

“psychotherap*” within a few words of “assessment”, “need*”, “intake”, “appointment”, “support”, “intervention”, “session”, “attended”, “continued”,

“psychological therap*” within a few words of “assessment”, “need*”, “intake”, “appointment”, “support”, “intervention”, “session”, “saw”, “continued”, “follow up”, “ refresher”, or “top-up”.

“counsell*” within a few words of “assessment”, “need*”, “intake”, “appointment”, “support”, “intervention”, “session”, “saw”, “follow up”, “refresher”, or “top-up”.

“counsel*” within a few words of “attended”, “session”, or “continued”.

“psychodynamic” within a few words of “assessment”, “need*”, “intake”, “appointment”, “support”, “intervention”, “session”, “attended”, “saw”, “continued”, “follow up”, “refresher”, or “top-up”.

“psychotherapy*” within a few words of “saw”, “follow up”, “refresher”, or “top-up”.

“attended for psychological therapy”, “psychotherapy*: seen”, “psychological therap*: seen”, “counsell*: seen”, “psychodynamic: seen”, “psychotherapy*: reviewed”, “psychological therap*: reviewed”, “counsell*: reviewed”, or “psychodynamic: reviewed”.

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 80%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Cognitive Behavioural Therapy (CBT)

Return to contents

CRIS NLP Service

Brief Description

An application to identify instances of delivered sessions of Cognitive Behavioural Therapy (CBT).

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes- Definitions:

Search term(s):

1.1 Inclusions:

A session of CBT is defined as an event (excluding ward progress notes) having “CBT” or

“Cognitive Behavioural Therapy” or

“Cognitive Therapy” followed by “session”, “assessment” or “follow up” plus the following variations specified below:

1.2 Assessment session:

“Attended for CBT”

“LICBT” & “session”

“CBT appointment”

“CBT appt”

“saw ZZZZZ for CBT”

“CBT: Seen”

“CBT: Reviewed”

“Session X of CBT”

“X CBT”

“Xst CBT”

“CBT #X”

“CBT #X”

“SX CBT”

“session of CBT”

“continued with CBT” “CBT psychology session”

“session X of CBT”

“Met with ZZZZZ to continue the CBT work.”

“MIND WORKOUT (CBT GROUP)“

 

1.3 Follow Up:

“CBT follow up appointment”,

“CBT 12-month follow-up”

 

1.4 Alternative terms for CBT

“SX HICBT”

“SX LICBT”

Evaluated Performance

Cohen's k = xx 1) Instance level (Overall) , Random sample of 100 Random Documents: Precision (specificity / accuracy) = 85% Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1136/bmjopen-2016-015297

10.1136/bmjopen-2019-034913

Depot Medication

Return to contents

CRIS NLP Service

Brief Description

Depot antipsychotic medications are used to treat psychotic disorders via an injection that can be delivered 2-weekly to 6-monthly. The evidence suggests that missing 2 depots/year results in a significantly higher relapse rate, and yet there is not a reliable method within the trust for identifying when depots are due. We aim to develop an NLP app that can identify when a depot has been given so that it can be established when the next one is due.

Development Approach

Development approach: Rule-based.

Classification of past or present instance: Both.

Details of the rules used in the app:

Words ‘given, administered, gave’ is used in various combination with the list of medication.

Words ‘Depot, accepted, complied, attend’ is used in various combination with the list of medication.

In some cases dose (in value) is used in various combination with the list of medication.

Output and Definitions

The output includes-

 

Examples of positive mentions:

x attended for his depot-25mg given into left buttock”;

“flupenthixol 40mg IM given”;

“Paliperidone depot 100mg given x had his 2nd loading dose of paliperidone 100mg Deltoid on (L) side”;

“zuclopenthixol decanoate 300mg was given on the left upper gluteal muscle”;

“fluphenazine administered as per prescription chart”.

 

Examples of negative / irrelevant mentions (not included in the output):

“Since the start of this admission he has been switched to an aripirazole depot”;

“paliperidone depot injection 100mg STAT on dd/mm/yy”;

“two options discussed with him were clozapine and clopixol depot”;

“flupenthixol 100mg every 2 weeks, last given dd/mm/yy”;

“no need for olanzapine oral, as we plan to start depot today”;

Evaluated Performance

Testing done on 250 random documents:

Precision (specificity / accuracy) = 90.8%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Family Intervention

Return to contents

CRIS NLP Service

Brief Description

The application identifies instances of family intervention delivery.

Development Approach

Development approach: Rule-Based.

Classes produced:

The application will produce the following 6 features for each annotation: -

FI Session: Y/N

 

Session n: Session number

 

Stage: Assessment, first session, last, treatment, follow-up,

 

Subject: Both patient and carer/Carer/Patient but patient only relevant FI intervention for Behavioural Family Therapy (BFT). – Note if a single subject + patient then annotate as both (“ZZZZZ and carer”) and if more than one other attendee then annotated as family (“ZZZZ, mum and sister”).

 

Delivery: Individual Family/Multi Family – note Multi family groups are not generally practiced in the psychosis services but will be in the eating disorders service

 

Outcome: Attended, DNA, cancelled

Output and Definitions

The output includes- Definitions:

FI Session

Inclusions

A session of FI is defined as an event having “FI” or equivalent terms ("family intervention", “FI”, "family therapy", "family work", "family workshop","systemic work", "systemic therapy", "family session", “FTWS”, “Behavioural Family Therapy”, “BFT”, ”BFI”, “FIP”) followed by “session” or equivalent terms (“appt”, “Appointment”, “Ass”, “Assessment”, “Reviewed”, “Seen”) and additional terms specified below.

 

Exclude “family meeting” and “carer” from NLP app but include in the heading section – exclude at the combined_view stage.

 

Note - FIP refers to Family Intervention in Psychosis

 

Assessment session:

Other terms that should be included

“FI Assessment” Assessment

“FI: Ax” Assessment

“Assessment and FI in the same sentence” Assessment

 

Treatment session Other terms that should be included:

 

“Attended for FI”

“FI appointment”

“FI appt”

“saw ZZZZZ for FI”

“FI: Seen”

“FI: Reviewed”

“Session X of FI”

“X FI”

“Xst FI”

“FI #X”

“FI #X”

“SX FI”

“session of FI”

“continued with FI”

“session X of FI”

“Met with ZZZZZ to continue the FI work.”

 

Follow up

“FI follow up appointment”

“FI 12-month follow-up” Exclusions:

The following combinations below with FI in the same sentence are considered as exclusions. Note if the above inclusion criteria are met then this would be considered a positive hit independently of below but if only “next session” and FI were present in the same sentence this wouldn’t be annotated as a positive hit: -

“next session -/-” (day/month)

“next session 2nd”

“next session _._._” (day/month/year)

“Next session _._” (day/month)

“next appointment -/-” (day/month)

“next appointment 2nd”

“next appointment _._._” (day/month/year)

“Next appointment _._” (day/month)

“next appt -/-” (day/month)

“next appt 2nd”

“next appt _._._” (day/month/year)

“Next appt _._” (day/month)

 

Session n

Where a FI session has been indicated record the session number where specified. Note include first and last. Think about proximity – usually “Session x” but also examples of 1st session of FI, etc…

 

Other terms

“Final FI session” “last FI Session”

“Final session of FI”

“last session of FI”

 

Stage

Assessment terms:

“FI Assessment”

“FI: Ax”

“Assessment” and “FI” in the same sentence

Some services e.g picup service has mid therapy assessment

 

Follow-up terms

“FI Follow up appointment”

“FI Follow up appt”

 

Subject

Inclusions

Both patient and carer

Carer/

Patient but patient only relevant for Behavioural Family Therapy (BFT) (only in psychosis services)

 

Delivery

Inclusions

Group or individual therapy 253

 

Outcome

Attended, DNA, cancelled by carer, cancelled by patient, cancelled by staff

Evaluated Performance

Cohen's k = 88% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 87%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

Medication

Return to contents

CRIS NLP Service

Brief Description

The Medication Application distinguishes between medications that are currently prescribed (i.e. at the time of the document was written) and medications prescribed to the patient in the past. The application ignores medications that might be prescribed in the future. This is because a clinician may write that a patient should be prescribed a certain drug if their condition worsens but that may never happen to the patient. The Medication application does not calculate daily dose of a drug, just the dose given at a single point in time.

The application output is linked to BNF codes to enable researchers to filter by drug class. N.B. Some drugs with antidepressant BNF codes appear more frequently as antipsychotics (e.g., flupentixol). Care should be taken when extracting patients who have ever used an antidepressant to ensure antipsychotic usage is not erroneously included. Corresponding dosage information is informative in determining whether a patient used a drug as an antipsychotic or as an antidepressant.

Development Approach

Development approach: Rule-based.

Output and Definitions

The application appears to preferentially detect medications:

(a) With corresponding dosage information

(b) Written in this format: ‘Medication:’ ‘Current medications:’

Evaluated Performance

Antipsychotic (testing done on 50 random documents):

For current medication: Precision (specificity / accuracy) = 94% when the dose is mentioned

For current medication: Precision (specificity / accuracy) = 72% when the dose is not mentioned.

 

Antidepressants (tested done on 50 random documents):

For current medication: Precision (specificity / accuracy) = 90% when the dose is mentioned

For current medication: Precision (specificity / accuracy) = 72% when the dose is not mentioned.

 

Mood Stabilisers (testing done on 50 random documents):

For current medication: Precision (specificity / accuracy) = 98% when the dose is mentioned

For current medication: Precision (specificity / accuracy) = 76% when the dose is not mentioned.

 

Dementia (testing done on 50 random documents):

For current medication: Precision (specificity / accuracy) = 88% when the dose is mentioned

For current medication: Precision (specificity / accuracy) = 66% when the dose is not mentioned.

 

Amlodipine (testing done on 50 Random Documents):

Precision (specificity / accuracy) = 99% when the drug is mentioned; Recall (sensitivity / coverage) = 88%.

Precision (specificity / accuracy) = 99% when the dose is mentioned.

 

Diabetic drugs with BNF code ‘060101*’ or ‘060102*’ - Random sample of 50 Random Documents:

Precision (specificity / accuracy) = 94% when the dose is mentioned.

Recall (sensitivity / coverage) = 88% when the dose is mentioned.

Precision (specificity / accuracy) = 94% when the dose is not mentioned.

Additional Notes

Run schedule – Weekly

Other Specifications

Version 2.0, Last updated:xx

Social Care - Care Package

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of receiving current, recommended or planned general care package. This is a generic term relating to any social care intervention. This could be a specific type of social care (e.g. Meals on wheels, regular home care) or general mention of a package. ‘Status’ states whether patient currently has a care package, will get one in the future or that there is potential to receive it. ‘Subject’ states who the receiver of the care package is.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"carer came twice that day as per the care package",

"should receive a package of care from next week",

"we have recommended the care package to be increased".

 

Examples of negative / irrelevant mentions (not included in the output):

"refused the care package",

"tried to discuss having a package of care after discharge, but refused to converse".

 

Search term(s): "care package", "package of care"

Evaluated Performance

Cohen’s k = 95% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 88%

Recall (sensitivity / coverage) = 72.8%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Social Care - Home Care

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of home care/help. This is help by someone who comes to assist the patient with activities of daily living.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"patient gets home help 3x daily",

"carers admit he can be agitated during the day at home",

"home care plan to start next week",

"meeting to discuss potential home care",

"suggested home help which the family are considering".

 

Examples of negative / irrelevant mentions (not included in the output):

"home care FORMCHECBOX",

"carer support to be given to sister of ZZZ".

 

Search Term: "home care" , "carer visits" , "carer support" , "home carer" , "‘home help" , "home helper", "home carer had visited".

Evaluated Performance

Cohen’s k = 95% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 79%

Recall (sensitivity / coverage) = 65%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Social Care - Meals on Wheels

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of receiving current or recommended meals on wheels (food delivery, usually from a private firm).

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

"receiving MOW",

"recommended for MOW’,

"planning to have meals on wheels from Monday",

"will arrange for ZZZ to have MOW",

"isn’t eating his MOW",

"prefers the Wiltshire farms to the original MOW",

"discussed MOW",

"would be happy to have MOW".

 

Examples of negative / irrelevant mentions (not included in the output):

"mow the law",

"he is mow in mood".

 

Search term(s): "MOW", "meals on wheel*".

Evaluated Performance

Cohen’s k = 95% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 71%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – On Request

Other Specifications

Version 1.0, Last updated:xx

Outcome and Clinical Status

Blood Pressure (BP)

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of blood pressure scores in the format of overall score, systolic blood pressure

score and diastolic blood pressure score.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"BP: 126/85 systolic:126, diastolic:85",

"blood pressure was 126/73 systolic:126, diastolic:73".

 

Search term(s): "blood pressure", "bp"

Evaluated Performance

Cohen's k = 98% (testing done on 100 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) (Overall) = 98%

Precision (Systolic) = 98%

Precision (Diastolic) = 98%

Precision (Full Score) = 98%

Precision (Same day) = 92%

Precision (One Week) = 98%

Precision (One Month) = 98%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Body Mass Index (BMI)

Return to contents

CRIS NLP Service

Brief Description

Application to identify body mass index (BMI) scores.

Development Approach

Development approach: Rule-Based.

Classification of past or present symptom: Both.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Bmi 45, bmi:46, Body Mass Index is 22.9, 16 BMI"

"Bmi 45 kg/m2, BMI 47 Kg/m2 , BMI 22.8 kg/m 2"

 

Examples of negative / irrelevant mentions (not included in the output):

"Bmi centile 46, Bmi centile 77, He is on the 34th centile for BMI, BMI above 96th centile"

"Her BMI is 48 kg, BMI: 22 kg, BMI/Weight : 103.2 kg"

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) (Overall) = 89%

Precision (Same Day) = 66%

Precision (One Week) = 71%

Precision (One Month) = 72%

Precision (Three Months) = 73%

Recall (sensitivity / coverage) = 78%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1136/gpsych-2022-100819

Brain MRI report volumetric Assessment for dementia

Return to contents

CRIS NLP Service

Brief Description

Application for automated extraction of mentions of dementia-related volumetric assessments from plain text brain MRI reports

• This model has only been trained on brain MRI reports, where the clinical indication was for the investigation of dementia

• This had not been validated for use on other imaging reports, and for other clinical indications would be unlikely to return many results as the terms of interest are unlikely to be mentioned

• The terms of interest are unlikely to appear outside of the radiologists report, other than if they are copied into the patient notes and/or a clinical letter

Development Approach

Development: Machine-learning, based on spaCy python library

Output and Definitions

The output includes- Output:

• GVL = Global Volume Loss – Present

• NO_GVL = Global Volume Loss – Absent

• RVL = Regional Volume Loss – Present

• NO_RVL = Regional Volume Loss – Absent

• HVL = Hippocampal Volume Loss – Present

• NO_HVL = Hippocampal Volume Loss – Absent Definitions: Search term(s): Patient referred for an MRI from memory clinic. Takes the plain text MRI report as input and identifies and returns labelled spans or token that are predicted as belonging to any of the below 6 classes which is given in Output section.

For each span:

Text (string), label (String), score (float)

E.g., ‘brain volume is normal’ = NO_GVL

‘Severe bilateral hippocampal atrophy’ = HVL

 

The underlying span categorization approach allows multiple labels to be applied to the same span, e.g., ‘no regional predominant or hippocampal atrophy’ = NO_RVL and NO_HVL

 

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

GVL: F-score = 0.95

Precision (specificity / accuracy) = 95%,

Recall (sensitivity / coverage) = 95%

 

NO_GVL: F-score = 0.85

RVL: F-score = 0.68

NO_RVL: F-score = 0.92

HVL: F-score = 0.89

NO_HVL: F-score = 0.93

F1 score averaged over all 6 categories on holdout test set was 0.89

Additional Notes

Run schedule – Yearly

Other Specifications

Version 1.0, Last updated:xx

Cholesterol

Return to contents

CRIS NLP Service

Brief Description

To identify total cholesterol score of a patient, in clinical notes. Total cholesterol referred to as total cholesterol or serum cholesterol

Development Approach

Development: Rule-based.

Classification of past or present symptoms: Both.

Classes produced: In clinical notes, cholesterol level is referred as totals cholesterol score and serum cholesterol.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Total cholesterol 06-Sep-2022 4.8 mmol/L",

"Serum cholesterol level 08-March-2001 3.8 mmol/L",

"Serum total cholesterol level 14-June-1998 6.3 mmol/L",

"Total cholesterol level 19-April-2010 5.3",

"19-March-2020 6 mmol/L serum cholesterol level".

 

Examples of negative / irrelevant mentions (not included in the output):

"Serum cholesterol > 4.0 mmol/L".

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

 

Precision (specificity / accuracy) = 82%

Recall (sensitivity / coverage) = 95%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 2.0, Last updated:xx

EGFR

Return to contents

CRIS NLP Service

Brief Description

To extract results from the eGFR blood test that allows monitoring of kidney function.

Development Approach

Development approach: Rule-based (Jape Rules).

Classes produced: Annotations were positive in entries with a test descriptor of eGFR followed by a numerical result that was in keeping with an eGFR and was in proximity to the eGFR term, with appropriate units.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"eGFR result (EPI) >90" (Score given as 90)

"eGFR result (EPI) 80 ml/min/1.73m*2" (Score given as 80)

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 95%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

HBA1C

Return to contents

CRIS NLP Service

Brief Description

The application will use a structured code to identify instances where HbA1c* and its results are found within CRIS from non-structured fields (i.e. case notes). This will help provide a clearer indication of how HbA1c is being recorded within CRIS.

 

*HbA1c can be obtained from a routine blood test and refers to glycated haemoglobin. It develops when haemoglobin, a protein within red blood cells that carries oxygen throughout your body joins with glucose in the blood, becoming 'glycated'. By measuring glycated haemoglobin (HbA1c), clinicians are able to get an overall picture of what our average blood sugar levels have been over a period of weeks/months. For people with diabetes, this is important as the higher the HbA1c, the greater the risk of developing diabetes-related complications. Therefore, it is important to ensure that this is being recorded and monitored effectively within South London and Maudsley as we know that those with psychosis are at a greater risk of diabetes.

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

"HbA1c was 40", "HbA1c 40", "HbA1c was 40mmol/mol", "HbA1c was 40mmol",

"HbA1c was 15%".

 

Examples of negative / irrelevant mentions (not included in the output):

"HbA1c was measured and found to be within normal range",

"HbA1c was measured on 11/11/19",

"HbA1c 10/10/18".

 

N.B: The application was not developed with upper or lower score limits. However, during testing anything lower than 3% or 9mmol and anything higher than 24% or 238mmol was excluded:

 

HbA1c mmol/mol %:

Normal - Below 42 mmol/mol; Below 6.0%,

Prediabetes - 42 to 47 mmol/mol; 6.0% to 6.4%,

Diabetes - 48 mmol/mol or over; 6.5% or over

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) (Overall) = 92%

Recall (sensitivity / coverage) = 76%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 2.0, Last updated:xx

DOI

10.1136/bmjopen-2022-069635

Lithium

Return to contents

CRIS NLP Service

Brief Description

The application is designed to detect the results of the blood test looking for lithium levels in the blood. This test is used for finding the appropriate dosing of lithium

Development Approach

Development approach: Rule-based (Jape rules).

Classes produced: Annotations are positive when there is text around the word lithium and a number. The syntax should relate the number to the lithium and indicate that it is a level, either through this being mentioned explicitly in the text or the number relating to appropriate units for a lithium level.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Lithium 0.86 mmol/L 0.4-1.0" (level 0.86)

"Lithium 0.54 mmol/L 0.4-1.0" (level 0.54)

"Lithium 0.5 from blood test on 20/04"

 

Examples of negative / irrelevant mentions (not included in the output):

"Plan: 1. Care plan for at least 1.5-2L fluid intake per QQQ, there is essential while on lithium; 2. 1:1 with intentional rounding" (returns 2.1),

"Lithium 600mmg Nocte" (returns 600).

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 86%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

Mini-Mental State Examination (MMSE)

Return to contents

CRIS NLP Service

Brief Description

This app identifies MMSE scores and returns information on:

- MMSE score (overall and subdivided into numerator and denominator)

- Associated date

Development Approach

Development approach: Machine-learning.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

Numerator should be a number from 0 to 30 and denominator should always be 30. Date is identified in the format of DD/MM/YYYY.

Evaluated Performance

Cohen's k = 90% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents:

Precision (specificity / accuracy) (Numerator) = 97%

Precision (Denominator) = 98%

Precision (Date- Same Day) = 68%

Precision (Date- One Week) = 76%

Precision (Date- Two Weeks) = 81%

Precision (Date- One Month) = 84%

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 93%

Recall (sensitivity / coverage) = 94%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1093/ije/dyac185

10.1016/j.jamda.2022.04.045

10.1136/bmjopen-2021-055447

10.1016/j.jamda.2021.08.011

10.1080/13607863.2021.1947967

10.1136/bmjopen-2019-035779

10.1002/gps.5420

10.1002/gps.5330

10.1007/s10654-020-00643-2

10.1016/j.exger.2019.02.019

10.1016/j.jamda.2018.03.009

10.1371/journal.pone.0178562

Neutrophils

Return to contents

CRIS NLP Service

Brief Description

The application is designed to pick up the result of blood tests looking at the white cell count, which can be raised or lowered as a result of being physically unwell or as a side-effect of medications. It is designed to function to pick up results that are inputted from laboratories as well as those mentioned in the notes

Development Approach

Development approach: Rule-based (Jape rules).

Classes produced: Examples are positively annotated where the number identified is clearly related to the description white cell count or WCC. There was no distinction made between those that were felt to refer to a previous result or a current result.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Neutrophils 2.51 10^9/L 1.50-6.30" (algorithm returns 2.51),

"Neutrophils 2.2 10^9/L 1.70-6.10 10^9/L" (algorithm returns 2.2),

"Neutrophils 7.4" (algorithm returns 7.4).

 

Examples of negative / irrelevant mentions (not included in the output):

"Percentage neutrophils 48.6%"

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 97%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

Non-Adherence

Return to contents

CRIS NLP Service

Brief Description

An application to identify instances of clinician-reported non-adherence/non-compliance to psychiatric treatment (including medication) from patients’ mental health records.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive, Negative and Unknown.

Search term(s): 'Adheren*', 'complian*', 'non-adheren*', 'nonadheren*', 'noncomplian*'.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"We have to accept non-compliance for the time being";

"Mum reports non-compliance with medication that *patient* denies";

“her disengagement with Home Treatment Team and non-compliance with medication”.

 

Exclude Negative mentions of non-adherence include:

“Compliant with medication”;

“Isolates himself, disengages from services, non-compliance, alcohol abuse, depression.”

 

Exclude Unknown mentions of non-adherence:

"Partially adherent”

“fluctuating adherence”

"having problems with compliance".

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 86%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Diagnosis

Return to contents

CRIS NLP Service

Brief Description

Application to extract instances of diagnosis.

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

 

Output and Definitions

The output includes-

 

The main aim is to look for a standard or as close as possible to a definitive standard diagnosis:

1.) When reading through document, if you come across phrase(s) similar to the examples below:

……Diagnosis: Fxx.x diagnosis name……(this could be with or without the colon, or could even have several colons and/or other punctuation marks before they diagnosis name, following each

……Diagnosis Fxx.x diagnosis name……

……Diagnosis: diagnosis name……

……Diagnosis: Fxx.x……

Highlight this as ‘Diagnosis’ – please label the annotation just as I have specified it (i.e. with a CAPITAL D).

2.) The following features have been added under the Diagnosis annotation:

ICD10: if there is a name of a diagnosis, but no ICD10 code, find the ICD10 code and fill in under the feature ICD10

Diagname: if there is a diagnosis name then please copy this in the annotation feature. Please copy the exact diagnosis name even if it varies from the official ICD10 name.

Diffdiag – add this only if there is a differential diagnosis. This kind of diagnosis is often mentioned because usually most documents are trying to find out what the diagnosis is and in the process give a possible diagnosis which is vague or will not be the correct one eventually.

Nonpsychdiag – any definite diagnosis where the annotation does not come under the F group diagnosis. For example, COPD. Definitions: Search term(s):

Gazetteer of diagnoses and ICD10 codes.

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 50 random documents):

F20/ Schizophrenia - Precision (specificity / accuracy) = 96%

F20/ Schizophrenia - Recall (sensitivity / coverage) = 63%

 

F20 - Precision (specificity / accuracy) = 100%

F20 - Recall (sensitivity / coverage) = 65%

 

SMI - Precision (specificity / accuracy) = 95%

SMI - Recall (sensitivity / coverage) = 43%

 

Schizoaffective - Precision (specificity / accuracy) = 80%

Schizoaffective - Recall (sensitivity / coverage) = 29%

 

Depression - Precision (specificity / accuracy) = 100%

Depression - Recall (sensitivity / coverage) = 40%

 

Patient level testing done on all patients with primary diagnosis of learning disability F7 or "learning dis*" (testing done on 50 random documents):

Precision (specificity / accuracy) = 93%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

Treatment- Resistant Depression

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of treatment-resistant depression.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“X year history of treatment resistant depression”,

“problems with low mood (resistant depression)”,

“diagnosis: treatment resistant depression”,

“resistant endogenous depression, suffers from chronic treatment resistant depression”,

“referred for management of treatment resistant recurrent depression”

 

Examples of negative / irrelevant mentions (not included in the output):

“talked about ways in which they might resist allowing each other’s depression to …”,

“has a diagnosis of treatment resistant schizophrenia and depression”,

 

Search term(s): "depression" and "*resist*" , "*resist* and "depression"

Evaluated Performance

Cohen's k = 85% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 77%

Recall (sensitivity / coverage) = 95%

 

Patient level testing done on all patients with primary diagnosis code of F32* or F33* testing done on 50 random documents):

Precision (specificity / accuracy) = 90%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 1.0, Last updated:xx

Bradykinesia (Dementia)

Return to contents

CRIS NLP Service

Brief Description

To identify instances of bradykinesia in the context of dementia.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“presence of bradykinesia”,

“motor symptoms – moderate bradykinesia L>R”.

 

Examples of negative / irrelevant mentions (not included in the output):

“absence of bradykinesia”,

“he was moving easily in bed and transferring independently with no bradykinesia or tremor”

“bradykinesia is a symptom of dementia”,

“difficult to assess if it has caused any bradykinesia”,

“SHO to look out for bradykinesia” Definitions:

 

Search term(s): Bradykine*

Evaluated Performance

Cohen's k = 100% (testing done on 50 random documents).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 91%

Recall (sensitivity / coverage) = 84%

Additional Notes

Run schedule – Monthly

Other Specifications

Version 2.0, Last updated:xx

DOI

10.1016/j.exger.2020.111223

Trajectory

Return to contents

CRIS NLP Service

Brief Description

Application to identify evidence from clinical text that something is getting better or getting worse.

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

 

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Hallucinations have improved”,

“living conditions are worse than they were",

“Her piano playing has improved’,

“His mother’s eye-sight is better”.

 

Examples of negative / irrelevant mentions (not included in the output):

“His drinking is better when attends the session”,

“she may have improved”,

“his knee could deteriorate”,

“The relationship with her mother deteriorated…”.

 

Evaluated Performance

Cohen's k = xx

 

Insance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 97%

Recall (sensitivity / coverage) = 76%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Tremor (Dementia)

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of tremor in patients with dementia.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“There was evidence of a tremor when writing…”,

“…with a degree of resting tremor…”

 

Examples of negative / irrelevant mentions (not included in the output):

“There are no reports of any noticeable motor symptoms such as tremor…”,

“No dystonic movement or tremor”.

“ZZZZ will be reviewed with regards to side effects and if there is no tremor then can have another 75mg of Paliperidone”,

“there is a family history of tremor”.

 

Search term(s): *Tremor*

Evaluated Performance

Cohen's k = 100% (50 un-annotated documents - 25 events/25 attachments, search term ‘tremor*’).

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 63%

Recall (sensitivity / coverage) = 96%

 

Patient level testing done on all patients with a dementia diagnosis (testing done on 100 random documents):

Precision (specificity / accuracy) = 83%

Recall (sensitivity / coverage) = 92%

 

Patient level testing done on all patients (testing done on 30 random documents):

Precision (specificity / accuracy) = 83%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

DOI

10.1016/j.exger.2020.111223

QT

Return to contents

CRIS NLP Service

Brief Description

Application to identify instances of QT interval or Corrected QT interval, QTc.

Development Approach

Development approach: Machine-learning.

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

The output includes-

 

Examples of positive mentions:

“QTc now 497”,

“QTc 450”,

“Uncorrected QT 384ms”

“QTc – 404”

“QTc: 421”

“QTC interval of 430ms”

“QTc= 442ms”

 

Examples of negative / irrelevant mentions (not included in the output):

“increase the QTc by 1.3ms" (excluded since the patient’s actual QTc not stated),

“QTc interval was less than 440ms" (excluded since the patient’s actual QTc not stated),

“QT was given 7 days" (excluded since QT is referring to the patient.),

“Date of birth: - QT- 1R "i '^n^^jSTtv^eN" (excluded since a random mix of letters not clinically relevant),

“recommended QTC interval less than 440ms in men and less than 470ms in women" (excluded since it is stating the recommended ranges).

 

Search term(s): "QT" or "QTc" followed by "=, <, >, or -", or followed by number

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 94%

Recall (sensitivity / coverage) = 96%

Additional Notes

Run schedule – On Request

Other Specifications

Version 1.0, Last updated:xx

White Cells

Return to contents

CRIS NLP Service

Brief Description

This app is designed to pick up the result of blood tests looking at the white cell count, which can be raised or lowered as a result of being physically unwell or as a side effect of medications. It is designed to function to pick up results that are inputted from laboratories as well as those mentioned in the notes.

Development Approach

Development approach: Rule-based (Jape Rules)

Classes produced: Examples are positively annotated where the number identified is clearly related to the description white cell counts or WCC. There was no distinction made between those that were felt to refer to a previous result or a current result.

Output and Definitions

The output includes-

 

Examples of positive mentions:

"Total white cell count 4.16 10^9/L 4.00-11.00 10^9/L" (algorithm returns 4.16)

"Total white cell count 5.4 10*9/L 4.00-11.00 10*9/L" (algorithm returns 5.4)

"White cell count 11.5 x 10^9/L HI" (algorithm returns 11.5)

Evaluated Performance

Testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Additional Notes

Run schedule – Weekly

Other Specifications

Version 1.0, Last updated:xx

Miscellaneous

Family Contact

Return to contents

CRIS NLP Service

Brief Description

This application is designed to positively identify instances where the subject of a clinical record has had contact with a family member, friend or a carer.

Development Approach

Development approach: BERT Model

Classes produced: Annotations are positive where the subject has had present or recent contact with a family member or carer.

Output and Definitions

The output includes-

 

Examples of positive mentions:

“Met with father.”

“Spoke with mother last week”

“Was recently visited by carer.”

 

Examples of negative / irrelevant mentions (not included in the output):

“Has not met with family”

“Met with doctor”

“Spoke to sister last year”

Evaluated Performance

Testing done on 100 random documents:

Precision (specificity / accuracy) = 84%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Forms

Return to contents

CRIS NLP Service

Brief Description

Application to identify documents that include form structures such as yes/no questions and other checklists.

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

Manual annotators look for forms as stand-alone documents or as part of the wider document text and annotate either positive (that text corresponds to a form and therefore the document contains a form or is a form), or negative (that text doesn’t correspond to a form and therefore the document neither contains a form nor is a form).

A form is identified based on:

1) Presence of the term ‘form’ within document heading

2) Presence of yes/no questions

3) Presence of checkboxes

 

Forms are identified as such even if they are not filled in, are part of a letter or email or correspond to symptom checklists.

The following rules were applied by the app to determine the presence of a form.

-Parse the HTML Text, identifying the text within tags.

-Identify the term “Form” within heading, paragraph of emboldened tags with lengths of less than 80

- Identify the presence of yes/no questions

-Identify the presence of check boxes

-Identify unique text to particular forms (this is an evolving part of the app that is updated as more information becomes available).

-Identify terms to exclude the document from being a form. These are terms that indicate the document may be a letter or an email (such as a greeting to open or close a letter, or terminology to indicate an email reply, such as “re:”.

Evaluated Performance

IRR = 92% - Cohen’s k could not be computed.

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 100%

Recall (sensitivity / coverage) = 76%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx

Quoted Speech

Return to contents

CRIS NLP Service

Brief Description

Application to extract text within quotation marks

Development Approach

Development approach: Rule-Based

Classification of past or present symptom: Both.

Classes produced: Positive

Output and Definitions

Rules:

Regular expression matching is used to identify text occurring within matching quotation mark pairs [(" & "), (' & '), (“ & ”), (‘ & ’)] in the EHR. To avoid mistaking apostrophes used in contractions for the start of quoted phrase, a quote followed by a sequence (‘c’, ‘d’, ‘e’, ‘m’, ‘n’, ‘s’, ‘t’ , ‘ve’, ‘re’, ‘s’, ‘ll’, ‘all’ ) was treated as an apostrophe not a quote. A similar check was performed for end quotes. Once a quoted phrase was identified, any sub-quotations occurring within that quote were assumed to be part of the larger quotation. The length of quoted phrases was allowed to vary from one word to more than a paragraph; however, a maximum length of 1,500 characters was applied to avoid extracting the entire text where a quote was not properly closed. Phrases that consisted only of emails or starting with “https” were removed using standard regular expression pattern matching and substitution procedures.

 

Annotation guidelines:

 

Manual annotation involves identifying the full scope of human identifiable quoted speech.

Manual annotations that are deemed to be quoted speech if they include:

• Matched pairs of quotation marks as specified by the quotation algorithm [(" & "), (' & '), (“ & ”), (‘ & ’)].

• Mismatched quotation mark pairs e.g. “hello’

• Quotation mark pairs that start with the end quote e.g. ”goodbye“

• Incorrect/unusual quotation mark types that are not specified in the algorithm e.g. backticks in `x`

• Identifiable quotations that have no end quotation mark e.g. “I am always sad. The patient’s general mood was… Manual annotations that are not deemed to be quoted speech if they include:

• Emails or URLs (https)

Evaluated Performance

Cohen's k = xx

 

Instance level (testing done on 100 random documents):

Precision (specificity / accuracy) = 82%

Additional Notes

Run schedule – On request

Other Specifications

Version 1.0, Last updated:xx