In healthcare it is frequently challenging to collect individuals’ health data for purposes other than the patients’ treatment. This is particularly the case in relation to the health data of minority groups, whether, for example, based on gender, ethnicity or disability. The lack of data and lack of representative data means that the results stemming from the AI may be biased, lacking in integrity and not fully reliable.
It is important for the collection of representative health data that patients have trust in the organisations which collect their data and understand the purposes for which their data is collected to unlock the availability of representative data. To gain patient trust it is essential that there is sufficient transparency as to the purpose and objectives of the use to which the personal data will be put, as well as assurances concerning its confidentiality, security and the period of its retention.
Bias can also occur from the selection of the AI training data that is input by AI developers and/or in their further development of the AI algorithms. A diverse AI developer team, informed as to the risks of insufficient or unrepresentative data and of bias, should be deployed. Additionally, AI itself may “learn” to confirm that it is processing sufficient and representative data.
As well as data collection, good data management will also be essential, whether handling personal data, anonymised or aggregated data. Monitoring and auditing data usage from collection and use of training data and other input data through to AI model usage and outputs will support accountability and trust.
Addressing cyber risk
Because the life sciences sector is at the forefront of scientific innovation it is already an attractive target for cyber criminals, including state-sponsored attackers. This means the cyber risks all businesses face – business interruption, costs, potential data loss, regulatory action and reputational harm, to name a few – are heightened in life sciences.
Some attacks are intended to delay, disrupt or undermine trust in critical research projects and cause economic and social harm. Other attacks tend to be focused on stealing personal data, valuable research data and other intellectual properly.