“A lesão 1 no QIE da mama esquerda corresponde a um nódulo redondo com 21 mm, margens bem circunscritas, sem agrupamento de microcalcificações associado.”
Figure 1. Different modalities may hold complementary information.
In many areas where information related to a certain entity is available from more than one source (e.g. visual, audio or textual information) it may be desirable to use multiple sources to find the closest match. Put in other words, we may desire an application where queries are from one domain (text) and retrieved objects are from another domain (audio). An everyday example of this is music search; we can perform music search on a database by using a “microphone-captured song snippet” (audio), or alternatively use “part of the song lyrics” (text). Both queries should return the (same) music we are looking for.
The same applies to breast cancer screening in which information is derived from images but also medical reports.
This work aims to use a set of annotated medical exams containing mammograms, an expert report and other relevant meta-information to develop models that can perform: cross-modal, uni-modal and multi-modal classification systems for Computer-Aided-Diagnosis (CAD) for breast cancer. These can include: 1) automatic medical report generation, 2) diagnosed findings localization on the mammogram, and others. Ultimately, this work contributes to improve state-of-the-art in CAD for mammography. The work is to be developed at INESCTEC, where research in this area is conducted daily.
This is an ambitious thesis project in the sense that it includes three areas of research:
[1] - Gao H, Aiello Bowles EJ, Carrell D, Buist DS. Using natural language processing to extract mammographic findings. J Biomed Inform. 2015;54:77-84. https://doi.org/10.1016/j.jbi.2015.01.010