Guys, did you use any high level methods to fill nan values or just leave it as it is? There are many nans in the data and filling it using basics e.g. mean/median/mode or even some imputers as knn or mice is not really good idea as it only adds some noise and change the real statistical distribution of values.
Since the data is a hourly average of weather indicators for 5 days,(121 hours), i think a good way to deal with such NaN values is by looping through the list and averaging it on hour basis, maybe 8 hours interval or so. And also I think there are probably some missing values in the temp column as there are 0.0 values for temperature which is'nt possible in Uganda.
It is better to not fill null values than absurdly filling with mean, median, mode in this particular problem statement. Although Some variables have a certain pattern of values. Analyze each Var separately and fill nulls accordingly if you see a pattern.
@Krishna_Priya. In fact, you are very right. I just finished my EDA on these datasets. I could see many features with very different patterns like you said. Just to be taking them one by one is my concern right now considering the fact that we have 770 numerical variables(3 discrete e.g location,min_precip,median_precip and the rest continuos) after using label encoding for variable 'location' . Now only categorical variable is the ID. Truly, stressful though, but we are getting there gradually.
@Adeyinka_Michael You are on the correct track. Yeah, this data certainly requires a bit extensive research :)