Академический Документы
Профессиональный Документы
Культура Документы
Sayar Karmakar, Tamal Kumar De, Deborshee Sen, Arka Bhattacharjee, Riddhiman Bhattacharya, Abhirup Mondal
Introduction
In this project we have done some internal consistency checking on a survey data. As the size of the data is huge, one can surely ask the question how reliable the responses are. To address this question we carry out some tests on the dataset and draw some conclusion.
There might be inconsistency between the response about the general price increase and commodity wise price increase. Suppose a respondent says that commodity prices would increase in general more than the current rate in next 3 months. But when asked about commodity wise price change he replies that price would decline for every such commodity. We can do tests for this kind of inconsistency for both the 3 month data and the 1 year data. These are the inconsistencies between the responses of a particular respondent. We might also be interested at whether the responses of the people surveyed by a particular investigator in a particular city and time are close or not. If in some cases we see too much variation or too much consistency, then we might suspect that something has gone wrong. We can also assign a consistency percentage to every investigator depending on the proportion of inconsistent response. Then we might look how this proportion has varied over investigators in the same city and time. We might also look for a pattern between the consistency percentage and the workload. As the time for collecting the data were spaced by 3 months we could tally the data collected for expected inflation rate after 3 months with the data on current inflation rate obtained after three months. But this was not done as the persons interviewed or the investigators would not remain same over the time period. Also inflation rate changes are highly unpredictable and while doing consistency checks we are not interested in determining how well the respondents predicted the future inflation rates. So this would not give us anything useful.
Rules
We first replace the responses A,B,C,D and E for indirect questions (good wise) with 10,30,50,70 and 90 respectively. This is our response (3). Then we replaced the responses of the direct question (inflation rate) A,B..J by 5,15 95 respectively. The 3 month or 1 year inflation rate is response (2) and the current inflation rate is response (1). We call a response consistent if i. ii. iii. iv. v. Response(2) Response(1) > 35 and Response(3) =10 or 15 < Response(2) Response(1) < 35 and Response(3) <= 30 or -15 < Response(2) Response(1) < 15 and Response(3) <= 50 or -15 < Response(2) Response(1) < 15 and 30 <= Response(3) <= 70 or Response(2) Response(1) < -35 and Response(3) >= 50
These rules were chosen carefully. To simplify, we illustrate the first rule. If response(3) is 10 that means price has increased more than current rate and so difference between response(2) and response(1) should be high. Similarly the others rule can be verified. We carry out the same tests for both three months and 1 year. These are two types of consistency checks.
Consistency of the general price increase rate and commodity-wise price increase rates
Here, we want to see whether the responses in the price change in general are consistent with the responses for separate goods. If for example someone responds that price would decrease for every commodity but he says that price would increase in general then this is an inconsistency. Here, we want to compare a single categorical response with a vector of categorical response. Here we do not know how much weight these different commodities attribute to the general economic structure. So, we cannot compare the general response with some typical weighted mean of the responses in the categories. What we did instead is we look at the range of the responses in the good-wise categorization. If the general response lie outside this then we say that it is inconsistent. It seems too much conservative, but actually this is the best we can do. Any weighted mean would always lie in the range of the responses. As we do not know the actual weights, our measure of inconsistency would remain invariant under any kind of weighing schema. We do this good-wise consistency check both for 3 months and 1 year over various cities over time. Once we define these 4 types of inconsistency, we say a response is consistent if it is consistent in all the 4 cases and if the direct-indirect consistency cannot be checked due to unavailability of data we say the response is consistent if it is consistent in good-wise check.
Goodwise consistency for march and june, All for septmber & december
Then we concentrated on different time points and see consistency percentage for different cities. In the following plots the cities are plotted in the following order: Ahmedabad, Bangalore, Bhopal, Chennai, Guwahati, Hyderabad, Jaipur, Kolkata, Lucknow, Mumbai, New Delhi, Patna.
March
June
September
December
We can easily notice here that there is a sharp decline in proportion of consistent responses in Ahmedabad in September, and again in Patna in December. Also Hyderabad in September and Bangalore in December also have markedly low consistency percentage. We observe the percentage of consistent responses is quite less in quite a few cases leading us to the conclusion the overall consistency of responses is not quite high. Next we plotted the consistency proportions for the cities for the 4 months taken together. In the following plot we can easily see most cities have about average percentage of consistent responses. Only exception is Patna for which overall percentage of consistent responses falls even below 70%.
Bhopal
Chennai
Guwahati
Hyderabad
Jaipur
Kolkata
Mumbai
Lucknow
New Delhi
Patna
We also plotted all the workload-consistency proportions for the all the investigators together .
So we see that there is no specific pattern in the consistency proportions. They are more or less comparable for every such categorization.
If a particular investigator has measures for at least 2 questions inconsistent then we say that hes suspicious or unreliable. If someone has measures for 4 or more questions inconsistent then we identify that investigator as highly unreliable.
Results
We first removed the investigators who have surveyed less than 20 persons as calculating a measure of spread based on paired samples taken from a very small population size would not be very meaningful. Out of the rest 180 investigators 32 are found were found to unreliable or suspicious and 10 were highly suspicious. Among the cities the highest number of unreliable investigators are from Chennai and New Delhi. From Chennai there were 4 investigators who were unreliable or suspicious and 4 were highly suspicious among a total of 20 investigators. And from New Delhi 2 were highly suspicious and 3 more were suspicious or unreliable. Ahmedabad scored a clean sheet as no investigator from this city was identified as unreliable.
Summary
1. First of all we decide upon what we mean by consistency. We mainly address the respondentwise consistency here. 2. For a particular respondent, we look at their responses at current inflation and 3 month after inflation. If their difference and the response in general category are contradictory we call it an inconsistency. Same was done for 1 year also. 3. We checked whether the response in general price change is consistent with the response in price changes over different commodities. This was done for both 3 month and 1 year response. 4. Once we defined these inconsistencies we call a respondent inconsistent if he has inconsistency in any one of these 4 types. Then we looked at the consistency percentage of different investigators and whether this percentage has any effect on the workload or not. 5. We also checked whether sex and occupation of the respondent have a special effect on the proportion of inconsistency. 6. After dealing with this respondent-wise inconsistency we checked whether the responses over a city are close or not. We use a particular measure for this purpose. We try to look at the consistency of variation from city to city and also for different investigators.
Future works
1. The measure we used to judge a particular respondent is consistent in his direct and indirect questions is somewhat subjective to our intuition. We can look for some measure which would be invariant under subjective judgments. 2. In literature there are some tests for checking consistency of real-valued responses such as C ronbachs alpha, McDonalds omega, etc. These measures can be used if we can replace these categories by suitable representative values. Likert scaling is one of the methods we can use. 3. In order to replace these categorical data by suitable real numbers, we were thinking to use the cumulative frequency table. If suppose from the histogram we found that the responses are following a normal pattern then we could replace i-th category by the expectation of the normal truncated at phi-inv (Pi) and phi-inv (Pi-1). If not normal then we try to guess the distribution and use corresponding quartiles. 4. We were saying that a respondent is inconsistent if we are getting an inconsistency in any of the 4 types we discussed. Instead we can look for some combination of these 4 inconsistencies to get a consistency measure for each respondent. We call a particular respondent inconsistent if he or she has consistency measure less than a particular cut-off.
Conclusion
This is mainly a comparative study. We could not carry out the consistency of the first two types in all
responses due to unavailability of data. These type of systemic errors should be avoided in future surveys. Among the rest 7999 responses, 6273 (78.422%) were consistent for the first two types (direct-indirect). 14099 (88.124%) responses were consistent for the second two type of consistency (good-wise) among all the 15999 responses. We see that only 12755 (79.724%) among the 15999 responses are consistent considering all these four types. So we see that consistency rate is falling when we are considering all the four measures together and the consistency between direct and indirect question is affecting the consistency measure a lot. There might be two reasons possible behind this. One is inflation rate is a complicated concept that not many common men are aware of. So that type of inconsistency is more prominent. Another reason might be that we were a bit conservative in deciding the rules for checking good-wise consistency. Overall we see that the general understanding of inflation seems to be quite poor among the people, hence an overall consistency of only 79%. September. Patna is also quite low at consistency percentage in the month of December. The other cities
Among the cities, Ahmedabad showed very high consistency in March but very low consistency in
did not show any such pattern. For other cities there were variations in different time points but not drastic changes like this two cities. So we might suspect that something has went wrong in these two cities in those specific time points. Our suspicion is mainly bad choice of sample or some failure on the part of the investigators who carried out the survey.
Among the months, the overall consistency has decreased from March to December. But no drastic change
is there. So we could not conclude something very specific about consistency in these months.
We tried to see whether the workload has affected a particular investigator. We found no specific pattern
for that. Linear regression has shown the slope coefficient term is insignificant. This shows that whatever inconsistency these investigators had shown was not due to their workload, maybe perception of the common man played a big role in these consistency measures. there is no specific dependence of consistency on age, sex and occupation of the respondents. Although consistency increases slightly over different age groups, its not a significant difference. It maybe shows that older and middle aged people have a slightly better idea about inflation. variation present in the population distribution. We used this measure of variation to check the reliability of investigators and found out that among the 180 investigators who surveyed 20 persons or more, 32 were suspicious and unreliable and 10 were highly suspicious. As there was no dependence of consistency percentage of an investigator on workload and since many of the investigators were identified as unreliable, we would recommend not using these investigators in future surveys. are having markedly less variation and Patna has more variation than other cities. This may be due to different degrees of understanding of the concept of inflation in different cities or due to different actual values of inflation in different cities.
Then we wished to see whether categorizing these common men shows us some specific pattern. We see
Next, we look at the variation of each responses. We used the sample dispersion as an estimate of the
When we divided the responses city wise, although the difference is not large, we see Guwahati and Jaipur