top of page

Critical Reflection on Data Preprocessing Experiences

  • Writer: Vusi Kubheka
    Vusi Kubheka
  • Nov 19, 2024
  • 3 min read

Data preprocessing is both a meticulous and indispensable part of working with data. My experiences in this area have demonstrated how labour-intensive the process can be, yet it is impossible to overstate its importance. Skipping or rushing through this stage risks mismanaging the dataset, potentially resulting in errors in analysis or invalid conclusions.


My exposure to datasets such as those related to heart attack prediction and air pollution has underscored how critical it is to understand the dataset’s context, structure, and nuances.

A key insight from my experiences is the necessity of grasping the terminology and distinguishing between the numerical and categorical features of a dataset. While working with the heart attack dataset, for instance, I spent a significant amount of time learning what features like "serum cholesterol" or "thalassemia test results" meant. This was not just an academic exercise but a practical necessity to decide which preprocessing techniques to apply. Without this understanding, I risked dropping relevant features or misinterpreting the relationships between variables.


The air pollution dataset presented a similar challenge. Variables such as PM2.5 levels, AQI categories, and temperature required a nuanced understanding of environmental science concepts to ensure accurate manipulation. This process highlighted how understanding the dataset’s domain is just as crucial as knowing the technical functions available in data science libraries like Pandas or Seaborn.


As Sihem (2023) aptly observes in their thesis Domain Knowledge and Functions in Data Science, relying solely on data science techniques is insufficient. Domain knowledge remains essential for extracting meaningful insights from data. The thesis builds on prior research (Kitchin, 2014), emphasising the integration of domain expertise with data science to unlock advanced insights. This concept resonates strongly with my experiences.


In the absence of domain knowledge, data preprocessing risks becoming a mechanical task rather than an informed, strategic process. For example, dropping columns or imputing missing values without understanding their significance in the dataset’s context could lead to a loss of critical information or the introduction of bias. Similarly, the choice of normalisation or standardisation methods for numerical features depends heavily on understanding the dataset’s specific use case.


The insight from Viaene (2013) that data scientists are not domain experts aligns with my perspective on the value of cross-functional collaboration. While working on projects, I often found myself relying on external sources such as academic articles, domain-specific glossaries, or guidance from peers to contextualise the data. This reliance highlights the importance of teams comprising diverse expertise, where data scientists work alongside domain experts. These collaborative teams not only fill knowledge gaps but also enhance the overall quality of the analysis by ensuring a holistic approach to data handling.


Another critical reflection from my experience is that preprocessing is iterative rather than linear. Often, initial steps like detecting missing values or outliers revealed deeper inconsistencies in the dataset, requiring me to revisit earlier stages. For instance, in one project, I initially dropped rows with missing data, only to realise later that certain missing values had a specific meaning and needed to be imputed differently. Similarly, handling outliers was not just about removing extreme values but understanding their potential relevance to the dataset’s objectives.


One of the more challenging yet rewarding aspects of data preprocessing is creating dummy variables for categorical features. While this task seems straightforward, it requires careful thought to avoid issues like the dummy variable trap, where multicollinearity can distort analysis. These considerations further underscore the balance between technical skill and contextual understanding in preprocessing.


Reflecting on these experiences, I am struck by how data preprocessing is not merely a technical exercise but a cognitive and collaborative one. It requires the humility to acknowledge gaps in one’s domain knowledge and the willingness to engage deeply with the dataset. It also calls for a disciplined approach to ensure that the cleaned data not only meets technical standards but also retains its analytical integrity.



 


References


Sihem, A. Y. (2023). Domain Knowledge and Functions in Data Science (Doctoral dissertation, INSA Lyon).


Stijn Viaene. Data scientists aren’t domain experts. IT Professional,

15(6):12–17, 2013.


Rob Kitchin. Big data, new epistemologies and paradigm shifts. Big data

&society, 1(1):2053951714528481, 2014

Comments


  • Linkedin
  • Kaggle_logo_edited
  • Twitter
bottom of page