Part three: 15 pragmatic guidelines to handle data quality issues
Posted: 13 September 2024 | Dr Raminderpal Singh (Hitchhikers AI and 20/15 Visioneers) | No comments yet
In this four-part series, Dr Raminderpal Singh will discuss the challenges surrounding limited data quality, and some pragmatic solutions. In this third article, he discusses pragmatic guidelines to help support better data quality.
In the first article in this series, published Wednesday 14 August, we discussed the importance that data quality plays in the effectiveness of data analyses by Machine Learning (ML) and AI. In part two, we looked at the impact that poor data quality can have on ML models. Here, we outline 15 pragmatic guidelines to ensure better data quality in early drug discovery and to identify potential issues with compromised data.
- Data collection and entry
Standardisation, by implementing standardised operating procedures (SOPs) for data collection and entry, is a key consideration. This includes consistent use of units, naming conventions, and data formats. The team involved in data collection should be trained on the importance of data quality and the specific protocols to follow in order to minimise errors.
- Data validation
Automated checks should be carried out with automated validation scripts, which check for common issues such as missing values, duplicates, outliers and inconsistencies in units or formats. Furthermore, manual reviews of a subset of the data should be performed periodically to identify any issues that automated checks might miss.
- Data cleaning
A strategy should be developed for handling missing data, such as deciding when to use imputation, exclude data points, or flag datasets for further investigation. Also, methods should be implemented to identify and investigate outliers, determining whether they represent true variability or errors.
- Data integration
It is essential to ensure that data from different sources or experiments are harmonised before integration. This includes reconciling different naming conventions, units and formats. Consistency and correctness can be checked by cross validation, in which cross-referencing methods validate integrated datasets.
- Data documentation
Researchers should maintain detailed metadata for each dataset, including information about the origin, collection method, and any preprocessing steps. This helps in tracking data provenance and understanding the context. Also, version control systems for datasets should be used to track changes and ensure that any modifications are well documented and reversible.
- Data monitoring
Data should be monitored continuously, recording quality metrics like completeness, accuracy, and consistency, throughout the data lifecycle. Moreover, automated alerts can be set up to notify relevant personnel if data quality metrics fall below predefined thresholds.
- Data auditing
Conduct regular data audits to assess the overall quality of your datasets. This involves checking for adherence to data quality standards and identifying any systemic issues. Also, maintain audit trails that log all data processing steps, transformations, and any changes made to the data to ensure traceability and accountability.
- Bias and variability checks
Regularly assess your datasets for potential biases, such as over-representation of certain chemical scaffolds or biological targets. Statistical techniques should be used to quantify bias and take corrective actions. Analyse the variability in your data, particularly in biological assays, to understand the level of noise and its impact on model performance.
- Data redundancy and duplication checks
Implement robust mechanisms for detecting and removing duplicate records, to prevent skewing of the data. Then, techniques such as correlation analysis can be used to identify and eliminate redundant features that do not contribute new information.
- Data imbalance handling
Continuously monitor the balance of different classes in your data (eg, active versus inactive compounds). Address imbalances through methods such as over-sampling, under-sampling, or synthetic data generation. If data imbalance is unavoidable, consider using model algorithms that are better suited to handle imbalanced data.
- Data security and access control
Security measures can be implemented to protect data from corruption, loss or unauthorised access, ensuring that data integrity is maintained. With this, access to data can be restricted based on roles and responsibilities to prevent unauthorised modifications or data entry errors.
- Communication and collaboration
Foster collaboration between data scientists, domain experts and IT professionals to ensure that data quality requirements are clearly understood and addressed. A feedback loop should be established, whereby issues identified by data scientists or model results are communicated back to the experimental team to refine data collection processes.
- Use of quality control samples
Include quality control samples (eg, known standards or replicates) in experimental runs to monitor and ensure consistency in assay performance. Regularly analyse the results of quality control samples to identify any drifts or deviations in experimental conditions that could impact data quality.
- Data quality metrics and reporting
Define specific data quality metrics such as accuracy, completeness, consistency, timeliness, and uniqueness. Use these metrics to evaluate and report on the quality of your data regularly.
When reporting data quality metrics to stakeholders, ensure transparency to facilitate continuous improvement.
- Continuous improvement
When data quality issues are identified, perform a root cause analysis to understand the underlying reasons and implement corrective actions. You should treat data quality improvement as an iterative process, continuously refining and enhancing your strategies as new challenges and technologies emerge.
In the next article, which will be published Tuesday 24 September, we will present views from an industry veteran on the topic of data quality.
About the author
Dr Raminderpal Singh
Dr Raminderpal Singh is a recognised visionary in the implementation of AI across technology and science-focused industries. He has over 30 years of global experience leading and advising teams, helping early to mid-stage companies achieve breakthroughs through the effective use of computational modelling.
Raminderpal is currently the Global Head of AI and GenAI Practice at 20/15 Visioneers. He also founded and leads the HitchhikersAI.org open-source community. He is also a co-founder of Incubate Bio – a techbio providing a service to life sciences companies who are looking to accelerate their research and lower their wet lab costs through in silico modelling.
Raminderpal has extensive experience building businesses in both Europe and the US. As a business executive at IBM Research in New York, Dr Singh led the go-to-market for IBM Watson Genomics Analytics. He was also Vice President and Head of the Microbiome Division at Eagle Genomics Ltd, in Cambridge. Raminderpal earned his PhD in semiconductor modelling in 1997. He has published several papers and two books and has twelve issued patents. In 2003, he was selected by EE Times as one of the top 13 most influential people in the semiconductor industry.
For more: http://raminderpalsingh.com; http://20visioneers15.com; http://hitchhikersAI.org; http://incubate.bio
Related topics
Artificial Intelligence, Targets
Related organisations
20/15 Visioneers, HitchhikersAI