Automated Generation of Accurate & Fluent Medical X-ray Reports Paper Review

murlocKing·2022년 3월 15일
0

Paper

목록 보기
15/15

JRC 2022 Paper Review, 22/03/18 Seongyun, Lee

Prior works

image-based captioining

perform reasonably well in addressing the language fluency aspect

but, less satisfactory in terms of clinical accuracy

  1. closely tied to the textual characteristic of medical reports, which typically consists of many long sequences describing various disease related symptoms and related topics in precise and domain-specific terms
  2. lack of full use of rich contextual information that encodes prior knowledge. these information include for example clinical document of the patient describing key clinical history and indication from doctors, and multiple scans from different 3D views - information that are typically existed in abundance in practical scenarios, as in the Open-I, MIMIC-CXR

Framework

  • classifier
    • read multiple CXR images and extracts the global visual feature representation via a multi-view image encoder
    • they are disentangled into multiple low-dimensional visual embedding
    • text encoder reads clinical documents (doctor indication), summarize
    • visual, text-summarized embeddings are entangled via add&layer norm → contextualized embedding in terms of disease-related topic
  • generator
    • enriched disease embedding as input → generates word by word
  • interpreter
    • generated text → fine-tuning to align to the checklist of disease-related topic from the classifier

Classifier

  • Multi-view Image Encoder
    • m CXR images {Xi}i=1m\{{X_i}\}_{i=1}^{m}{xi}i=1m\{x_i\}_{i=1}^{m} (multi-view latent features). x have c-dimension, c is # of features via a shared DenseNet-121 image encoder.
    • multi-view latent features x can be obtained by max-pooling, m=1 ⇒ multi-view encoder → single-image encoder
    • 여기서는 CXR image가 여러 각도에서 촬영된 것을 multi-view로 표현. 각각의 각도에서 촬영된 CXR images에서 각각 latent vector를 추출하고 이를 encoding
  • Text Encoder
    • text document TT has length ll{w1,w2,...,wl}\{w_1, w_2, ... , w_l\}, wiRe,TRlew_i \in R^e, \therefore T \in R^{l*e}
    • transformer encoder가 TTHH로 encoding
    • HRleH\in R^{l*e}, hi=Encoder(wiw1,w2,...,wl)h_i = Encoder(w_i|w_1, w_2, ...,w_l)
    • Q={q1,q2,...,qn}Q=\{q_1,q_2, ... , q_n\}, representing n disease related topics (pneumonia), to be queried from document
    • Dtxt=Softmax(QHT)HD_{txt}=Softmax(QH^T)H, DtxtD_{txt} is text-summarized embedding
  • Contextualized Disease Embedding
    • ϕj(x)=AiT+bj\phi_j(x)=A_i^T+b_j, image encoding, jnj\le n, nn is # of diseases
    • Dimg=ϕj(x)nRneD_{img} = \phi_j(x) * n \in R^{n*e}
    • Dfused=LayerNorm(Dimg+Dtxt)\therefore D_{fused} = LayerNorm(D_{img} + D_{txt})
    • entanglement of visual and texttual information allows our model to mimic the hospital workflow, to screen the disease’s visual representations conditioned on the patients’ clinical history of doctors’ indication
    • but, this embedding by itself is insufficient for generating accurate medical reports, this lead us to conceive a follow-up enriched representation below
  • Enriched Disease Embedding
    • DfusedD_{fused}를 가지고 각각의 n개의 질병에 대해 k개의 disease state중 하나로 classification한다.
    • disease state = SRkeS \in R^{k*e}
    • p=Softmax(DfusedST)p=Softmax(D_{fused}S^T)
    • Dstates=ySD_{states} = yS if traning phase else pSpS, y=labely=label
    • finally eenriched disease embedding Denriched=Dstates+Dtopics+DfusedD_{enriched}=D_{states}+D_{topics}+D_{fused}
    • DtopicsD_{topics} is disease names

Generator

hidden state for each word = hih_{i}, masked multi-head attention, feed-forward layer를 N번 반복

hi=Encoder(wiw1,w2,...,wi1,d1,d2,...,dn)h_i = Encoder (w_i|w_1, w_2, ...,w_{i-1}, d_1,d_2,...,d_n)

여기서 wiw_i는 medical report를 의미함, did_iDenrichedD_{enriched}.

pword=Softmax(HWT)p_{word}=Softmax(HW^T)

weighted word embedding, generated report = W^=pwordW\hat{W} = p_{word}W

Interpreter

generated reports are often distorted in the process, such that they become inconsistent with the original output of the classification module - the enriched disease embedding that encodes the disease and symptom related topics

fully differentiable network module to estimate the checklist of disease-related topics based on the generator’s output and to compare with the original output of the classifcation module

this provides a meaningful feedback loop to regulate the generated reports which is used to fine-tune the generated report

D^text=Softmax(QH^T)H^\hat{D}_{text} = Softmax(Q\hat{H}^T)\hat{H}, H^\hat{H}is W^\hat{W}’s hidden state, QQ is quried diseases

pint=Softmax(D^txtST)p_{int}=Softmax(\hat{D}_{txt}S^T) is predicted disease label of interpreter

LI=L_I= -1N1\over Ni=1nj=1kyijlog(pint,ij)\sum_{i=1}^n\sum_{j=1}^ky_{ij}log(p_{int},ij)

yijy_{ij} is groud truth disease label

guide to force the word representations, force the generator to move toward producing a correct word representations

Experiments

  • Datasets
    • MIMIC-CXR
      • 227,385 medical reports of 65,379 patients, associated with 377,100 images from multiple view → AP(anterior-posterior), PA(posterior-anterior), LA(lateral), including comparison, clinical history, indication, reasons for examination, impression, and findings
      • this paper utilizes multi-view image, adopt as contextual information the concatenation of the clinical history, reason for examination, and indication sections
      • for consistency, this paper only focus on generating findings
    • Open-I
      • 3,955 radiology studies that correspond to 7,470 frontal and lateral chest X-rays
      • some radiology studies are associated with more than one chest X-ray image
      • Each study typically consists of impression, findings, comparison, and indication section
      • this paper utilize multi-view images and indication section
      • for generating medical reports, concatenating the impression and the findings sections as the target output
  • Results
    • Language Generation Performance
      • 모든 metric에서 SOTA를 차지하였다. 특히 MIMIC-CXR의 경우 single view image만 사용한 모델 조차 기존 SOTA를 압도하는 것이 인상적. 또한 interpreter module의 사용이 많은 이득을 가져다 줌을 확인할 수 있었다.
    • Clinical Accuracy Performance
      • to evaluate the clinical accuracy of the generated reports, they use the LSTM CheXper labeler as a universal measurement. F1, precision, recall metrics on 14 common diseases. and they also use macro, micro scores. high macro score means the detection of all 14 diseases is improved. high micro score implies the dominant diseases are improved
      • 모든 metric에서 SOTA를 차지 하였다.

Ablation study

  • Enriched disease embedding
    • enriched disease embedding 쓰는게 좋다
  • Contextualized embedding
    • contextualized embedding 쓰는게 좋다

Conclusion

본 논문은 three module approach로 X-ray image와 환자의 증상정보가 담긴 clinical document를 동시에 사용하여 기존 SOTA 모델에 비해 높은 성능을 얻을 수 있었고, interpreter를 사용해서 생성된 medical report를 다시 한번 fine-tune하면서 language fluecy뿐만 아니라 clinical accuray도 향상시킬 수 있었다.

profile
재미있는 딥러닝

2개의 댓글

comment-user-thumbnail
2024년 3월 1일

Unlock your body's full potential with Profunction.ca physiotherapy https://maps.google.com/maps/dir//Pro+Function+Health+Care+Team+140+Ann+St+Suite+103+London,+ON+N6A+1R3+Canada/@42.9930194,-81.2560469,14z/data=!4m5!4m4!1m0!1m2!1m1!1s0x882ef19d3e70de8d:0x37d7394910fe724e services in 2024. Our expert team is dedicated to restoring your mobility, relieving pain, and enhancing your overall well-being. Whether you're recovering from an injury, managing a chronic condition, or seeking to optimize your athletic performance, we tailor our treatments to meet your unique needs.

At Profunction.ca, we blend cutting-edge techniques with compassionate care to ensure you receive the highest quality of treatment. From manual therapy and therapeutic exercises to state-of-the-art modalities, we employ a comprehensive approach to help you achieve your goals.

Experience personalized care and unparalleled expertise at Profunction.ca. Let us guide you on your journey to a healthier, more active life. Take the first step towards optimal health and vitality – schedule your appointment with Profunction.ca physiotherapists today.

답글 달기
comment-user-thumbnail
2024년 6월 19일

The automated generation of accurate and fluent medical X-ray reports represents a significant advancement in healthcare technology. This paper review delves into the efficacy of such systems in improving diagnostic accuracy and streamlining report generation processes. By leveraging machine learning algorithms and natural language processing techniques, these systems ensure precise interpretation of X-ray images while facilitating the creation of comprehensive reports. Furthermore, the integration of features like refresh eye drops ustomer service enhances user experience and fosters efficiency in clinical workflows.

답글 달기