B.S.Chandrasekhar
Since the announcement of elections to the Lok Sabha and four State Assemblies in January 2004, the media is flooded with a large number of opinion polls. By the time voting started there were at least twenty pre-poll surveys and on the very first day of voting we had half-a-dozen exit polls. As many more pre-poll surveys and exit polls have been promised for the next three weeks, this article tries to examine the theoretical basis of election forecasting and its actual practice in our country.
Election surveys are not new to
There is yet another resemblance to soaps: the viewer has plenty to choose from. We have estimates to satisfy everyone. If you are a supporter of NDA you have the estimate of 340 seats for NDA and 105 for Congress and if you are a Congress supporter you can quote another agency, which has given 200 seats to Congress and only 230 to NDA. Similarly there has been wide variety in the results of exit polls of the first phase of the elections to satisfy all shades of political beliefs. Even at state level there is choice- in Karnataka Congress supporters are happy that NDTV exit polls say that
Election forecasting is a risky business everywhere. Even in a country like
Election forecast studies need expertise from various disciplines like Statistics, Sociology, Psychology, Political Science etc. There are many stages in such forecasting and the most important part is perhaps that of estimating the vote shares of different parties. The vote shares as we all know are estimated through large-scale sample surveys and the success of such surveys depends on at least three factors - the selection of a representative sample, developing a good questionnaire and properly conducting the fieldwork. Another crucial stage in forecasting is converting vote percentages to estimates of seats in Lok Sabha or the Assembly.
Sampling - Theory
The basic principle of sampling is that we can draw reasonably valid conclusions about the whole (population or universe) by examining only a small part (sample). Sampling is based on probability theories and is applied in almost all fields of knowledge. According to statistical theory from a representative and adequate sample we can estimate a population value within a certain interval with a certain degree of confidence. The range of the interval depends on the extent of sampling error (standard error), which in turn depends on the size of the sample. Without going to the details we can say that while estimating a percentage from a sample of 10,000 the maximum error will be 0.50 percent; and in a survey of 10000 voters if 30 percent have said that they will vote for Congress then we can say with a confidence of 95 percent that in India as whole 29 to 31 percent will be voting for Congress. In such surveys there is also another type of error called non-sampling error and controlling this error is equally important for forecasting.
The important requisite of a good sample is that it should be representative. Representation can be achieved if we select the sample ‘at random’, which means by ensuring that each unit in the population has an equal probability to be included in the sample. The second requisite is that the size of the sample should be adequate- it should not be too small.
A few other points about sampling could be mentioned here. The sample error is not dependent on the population size- whether the sample is 0.01 percent or it is 0.0001 percent of the total voters does not affect the error. Increasing the sample size does not reduce the error in the same proportion. For a sample of 10000 if the error is 0.50 percent for a sample of 400000 the error will not be reduced by a quarter (to 0.125) but only by half (to 0.25). A larger sample invariably reduces the quality of fieldwork and increases the non-sampling error and as such a larger sample is not always a better sample. The standard error for the state level estimates will be very much more than the error for the national estimates.
Let us see how this theory is put into practice in election surveys. Selecting randomly 10000 from 680 million voters is just impossible and this selection is done in multi-stages, the State, Region, Constituency and Polling booth etc. In this process the standard error increases but as long as the principle of ‘randomness’ is applied at each stage it is possible to estimate the error using different formulae. However more often than not convenience dictates the selection of the constituencies and the polling booths (more about this later) compromising the ‘scientific’ nature of such studies.
In all election studies at the final stage for selecting the voters a sampling system with set quotas for different demographic and social groups is adopted. Strictly speaking, the theories of sampling will not be applicable for quota sampling as there is purposive selection. However as long as there is no deliberate selection based on political leanings of the respondents a sample based on quotas is considered as good as a random sample.
As stated earlier with a sample of 10000 the standard error for the national estimates will be 0.50 percent but in the same survey the error for a particular State estimate could be as high as 2.5 percent and this is the reason why a survey, which has given correct estimates at the national level, would have gone wrong in some states. Most of the present surveys are not sensitive enough to track even 2-3 percent changes at state level and if some researcher talks about one percent shift against one party in one state or 2 percent shift in favour of another party in another state such statements have to be taken with not just a pinch but a large spoon of salt.
Questionnaire
In election surveys information is collected through a questionnaire and it is very important that this questionnaire has been prepared with care. "If you ask a silly question, you get a silly answer" and in many surveys such silly questions do creep in. The questions should elicit answers, which are reliable or consistent and are valid or relevant to the context. Textbooks on Social Research methods list some of the important questions, which the designer of a questionnaire should ask himself before framing a question like the following:
Just to give a simple example most people can give reliable answer to the question ‘Which soap you are using?’ But answering the question ‘Why are you using this soap’ may not be that easy (if it was so there would not have been so many brands of soaps in the market). When such a question is asked the respondent could arbitrarily tick one of the choices listed and if the same question is asked at a different time he could tick a different answer altogether. The answers to such questions will not be consistent or reliable. The framers of the questionnaires often forget this elementary point. One recent election forecast programme was discussing the findings about ‘the fear factor’ forgetting that a question on such a subject in a survey may well be self-defeating.
In most of the surveys though proper care is taken in the preparation of the original questionnaire in English often the translation to regional languages is left to interviewers who may not have a good knowledge of the language and idioms of the area where the survey is to be conducted. A badly prepared or translated questionnaire can spoil an otherwise well organized survey.
Field Work
The quality of survey research mainly depends on how well the fieldwork is conducted and this in turn depends on the professional capabilities of the interviewers who collect the data. The interviewer has to establish proper rapport with the respondents. His or her personality, the manners, tone of asking questions and even the "body language" contribute to the collection of accurate information. There has to be a proper selection of such interviewers and they should also be adequately trained. In addition there has to be proper supervision during the fieldwork.
The current elections surveys are being conducted by different agencies. In our country there are only two or three market research agencies with a network of field units with professionally trained researchers. Even these agencies have to recruit inexperienced local field workers for large size surveys. The other agencies presently conducting election surveys surface only during election times and there will always be a big question mark about the capabilities of their field staff and to what extent their work is supervised.
The agencies conducting the fieldwork have to work to stringent dead lines and more often than not convenience dictates their selection of constituencies, polling booths etc. When an agency has to complete 25000 interviews in the course of 6-8 hours in hundreds of towns and villages, one can only imagine how ‘representative’ will be the sample and to what extent quality of fieldwork is ensured. Laloo Prasad Yadav is quite justified in complaining that these surveys do not cover his constituency of voters. Even in normal times market researchers seldom reach the far off places and below-the- poverty line section of the population. (See the same author’s comparison of Census and NRS figures of TV ownership in rural areas at www.thehoot.org/mediaresearch ).
In large-scale surveys things can go wrong in a number of ways and they do go wrong. Nevertheless in such surveys the universal law of errors operates in an interesting way. If a set of bad interviewers pulls the data in one direction another set will balance by pulling it in the opposite direction.
Converting Vote Shares to Seats
Sample surveys provide estimates of vote shares of different parties and the next crucial part in forecasting is converting vote shares into seats. There are no ‘scientific’ formulae and forecasters have their own methods based on their accumulated experience. Recently Congress was justifiably peeved when the same data provided by a market research agency for Maharashtra was interpreted by two forecasters differently, one giving a majority of seats to that party and the other giving a big chunk of these seats to its rival. Forecasting the number of seats is comparatively easy when there is direct contest between two parties and it is quite hazardous when there are three or more contestants. Some years back in Uttar Pradesh BJP got 2-3 percent more votes than the Socialist Party (SP) but the latter got ten seats more than BJP and no forecaster could have anticipated this possibility. For a similar reason in the last parliamentary elections all the forecasters underestimated the strengths of SP and BSP.
Survey on Surveys
How surveys may go wrong could be illustrated with an example. CSDS has reported the results of ‘a poll on polls’ it conducted sometime back in
Q1. ‘Have you read or heard about polls?’ Yes-41%; No 59%
Q2. ‘Were you influenced by these surveys? Yes 10.2%; No 30.8%
Q3. ‘How were you influenced?’ Shifted from BJP to Congress 5.3% Bandwagon effect); Shifted from Congress to BJP to 4.9% (Sympathy factor).
Q1 is simple which could be easily answered and 41 percent is a reasonable level of awareness for a literate community. But Q2 is not as simple a question as it looks. Political analysts have always been saying that voting decisions are influenced by many factors and isolating one factor from the others is not easy. If at all such a question were to be asked there should have been a third option- ‘Do not know/Cannot say’ and a majority of the respondents might well have ticked that answer. If we assume that the respondents have answered sincerely 10 percent of voters and a quarter of those exposed to the results being influenced makes a strong case for banning such surveys. Q3 and the answers to this raise more questions- sympathy for whom? Candidate? Party? Its leaders? What benefit in joining the bandwagon? etc.
Sample surveys are not panacea for everything and in this case it was the wrong method to be adopted. There are more suitable methods like depth interviews, projective techniques, longitudinal studies etc. to research such subjects.
Track Record
Public memory is short. The morning’s newspaper becomes raddi by the afternoon and the evening’s television programmes are forgotten by the next morning and public memory on media is shorter. Media memory on media failures is perhaps the shortest. Let us recollect how correct were our election forecasters in the state assembly elections held in 2002:
Gujarat: Early forecasts- BJP may win two-third majority; mid-term corrections - Congress gaining grounds (‘Patel discontent’) but still BJP will win, Congress winning according to one magazine; Final Forecasts - comfortable majority to BJP. Result- Sweep by BJP- 25-30 percent more seats than the final forecasts.
Madhya Pradesh; Early forecasts- Congress will lose; mid-term corrections - Congress gaining grounds (‘Hindu card of CM working) Final forecasts- comfortable majority to BJP. Result- Sweep by BJP- 20-25 percent more seats than the final forecast.
Rajasthan; Early forecasts- Congress will win; mid-term corrections Congress doing well (‘Record of CM’); Final Forecasts- Tough fight- Congress can scrape through. Result two-thirds majority to BJP.
Chattisgarh; All stages Congress winning; Result comfortable majority to BJP.
Forecasters were right about
As mentioned in the beginning forecasting in
Impact on Voters
There has been extensive research on the role of media in elections, as a matter of fact serious research on media started with the study of Presidential elections in
General elections in
(Chandrasekhar was Director, Audience Research in Doordarshan.