George,
To answer your first question, no I do not agree. The peer-reviewed research does prove polygraph to be better then chance at detecting deception. I purposefully used a rather dated research for my first post because even it provides a better then chance accuracy in the review method used. If we look at Bersh's study alone, "....70.6 percent guilty correct and 80 percent innocent correct.", the overall accuracy rate is 75.3%. This is one of the first field polygraph studies. There have been changes made to polygraph which have improved the overall accuracy, some of which I discussed in my previous post. Bersh?s study also uses non-polygraph evaluators to confirm results. Thus, the 75.3% accuracy is achieved in part through information independent of the polygraph data. Lykken?s argument is that the results are independent of the polygraph charts. I have stated previously that this is not completely true. Some of the evaluators? based their decisions on information independent of the polygraph charts. However, the examiners? original decisions were based on the polygraph chart data. Regardless of all that, it shows better then chance in detecting both truth and deception in a field setting.
The combined studies of Bersh and Barland and Raskin do not illustrate polygraph to be 67.87% accurate per se. The studies illustrate that polygraph was accurate to this degree for the particular confirmation method used. Your argument is that polygraph is "not better then chance". Although the previous studies do not reflect the current accuracy rate of polygraph, using the given confirmation method the combine studies do support a better then chance accuracy rate.
On a related but separate note, just because Barland and Raskins study did not appear in a professional journal upon its release does not mean it was not peer-reviewed and accepted. Just because Lykken does not like the results and the method is not his, does not make the study invalid. You may get some of the people to agree some of the time but you can?t get all of the people to agree all of the time.
Since I have stated that there were improvements made to polygraph which have increased its? accuracy, I will quote some more recent studies to support the increased accuracy of Polygraph. All of these studies appear in professional journals and provide better then chance results for CQT polygraph detecting deception.
Quote: From
http://www.polygraph.org/research.htm
Patrick, C. J., & Iacono, W. G. (1991). Validity of the control question polygraph test: The problem of sampling bias. Journal of applied Psychology, 76(2), 229-238.
Sampling bias is a potential problem in polygraph validity studies in which posttest confessions are used to establish ground truth because this criterion is not independent of the polygraph test. In the present study, criterion evidence was sought from polygraph office records and from independent police files for all 402 control question test (CQTs) conducted during a 5-year period by a federal police examiners in a major Canadian city. Based on blind scoring of the charts, the hit rate for criterion innocent subjects (65% of whom were verified by independent sources) was 55%; for guilty subjects (of whom only 2% were verified independently), the hit rate was 98%. Although the estimate for innocent subjects is tenable given the characteristics of the sample on which it is based, the estimate for the guilty subsample is not. Some alternatives to confession studies for evaluating the accuracy of the CQT with guilty subjects are discussed
Podlesny, J. A., & Truslow, C. M. (1993). Validity of an expanded-issue (Modified General Question) polygraph technique in a simulated distributed-crime-roles context. Journal of Applied Psychology, 78(5), 788-797.
The validity of an expanded-issue control-question technique that is commonly used in investigations was tested with simulations of thief, accomplice, confidant, and innocent crime roles. Field numerical scores and objective measures discriminated between the guilty and innocent groups. Excluding inconclusives (guilty =
-18.1%, innocent = 20.8%), decisions based on total numerical scores were 84.7% correct for the guilty group and 94.7% correct for the innocent group. There was relatively weaker, but significant, discrimination between the thief group and the other guilty groups and no significant discrimination between the accomplice group and confident group. Skin conductance, respiration, heart rate, and cardiograph measures contributed most strongly to discrimination
Honts, C. R. (1996). Criterion development and validity of the CQT in field application. The Journal of General Psychology, 123(4), 309-324.
A field study of the control question test (CQT) for the detection of deception was conducted. Data from the files of 41 criminal cases were examined for confirming information and were rated by two evaluators on the strength of the confirming information. Those ratings were found to be highly reliable, r = .94. Thirty-two of the cases were found to have some independent confirmation. Numerical scores and decisions from the original examiners and an independent evaluation were analyzed. The results indicated that the CQT was a highly valid discriminator. Excluding inconclusives, the decision of the original examiners were correct 96% of the time, and the independent evaluations were 93% correct. These results suggest that criteria other than confessions can be developed and used reliably. In addition, the validity of the CQT in real-world settings was supported.
You ask how valid it shows polygraph to be. It is you who have purported that the CQT is not better then chance at detecting between truth and deception. Since you have in the past set the terms for rational discourse, I see it to be your burden to prove what is chance and what peer-reviewed studies have proven polygraph to be below the chance level. You would also have to establish what scientifically the chance level is. In a separate message thread you wrote:
Quote: Re: How Countermeasures are Detected on the Charts
« Reply #52 on: 12/11/01 at 15:51:49 »
In addition, the chance level of accuracy is not necessarily 50/50. It is governed by the base rate of guilt. For example, in screening for espionage, where the base rate of guilt is quite small (less than 1%), an accuracy rate of over 99% could be obtained by ignoring the polygraph charts and arbitrarily declaring all "tested" to be truthful. However, such a methodology would not work better than chance.
This is an estimation of base rate on your part. One cannot establish an ultimate true base rate in a field setting for truthful and deceptive because it will vary. One cannot control for the number of cases that will produce one or the other result within a field setting. A toxicologist cannot say that 50% of his cases will detect a presents of XYZ in the field because it will vary. A toxicologist can say that if XYZ is present then it will be detected, if the test works. Since polygraph measures for deception, polygraph can only produce one of two results, if the test works. Thus, the exam/test will have a 50% chance of producing deception or no deception, if the test works. If the exam/test does not work in either discipline, there is an outside contaminant that is thwarting the ability of the exam to produce an acceptable result.
As for your second question, no I do not agree that polygraph lacks both standardization and control.
Standardization:
The instrumentation must meet a standardized criterion. The examiner must meet a specified standard criterion. There is a very standardized process followed in a specific issue polygraph that is discussed in you book. The examiner must follow the process from the beginning to the end. This is standardization. The given question formats must contain a standardized number of a given type of question. This is another standardization. The given question format must follow a standardized sequence. The chart tracings must be of a certain standard of quality for acceptable scoring purposes. The scoring must be done in an acceptable standardized scoring method and must meet a standardized scoring result to make a decision. The fact is that there are numerous standardized methods within polygraph that prove it to be standardized.
Control:
The examiner is required to conduct the polygraph in a sterol environment that is free of visual and audio distraction. The examiner is required to assess the examinee?s medical background to control for outside contaminants that may hinder the ability of the instrument to obtain suitable tracings. The examiner must attempt to control for movement by the examinee to control for outside contaminants that may hinder the ability of the instrument to obtain suitable tracings. The examiner must conduct an acquaintance exam to control for the possibility of undisclosed medical of physical variants that may contaminate or hinder the ability of the instrument to obtain suitable tracings. Again, polygraph controls for a number of variables and thus does not lack control.
You quiet frequently use Furdey and Lykken for references. It should be noted that both of these individuals do have motive to be bias in their opinions toward CQT but support polygraph. Dr. Furedy has repeatedly condemned the use of polygraph but only the CQT method. Furedy is a proponent for polygraph when it uses the Guilty Knowledge Test (GKT). The GKT lacks the extensive research, reviews, critical debate, and sheer numbers of its use in the field setting that the CQT has endured and produced over time. Lykken is under the same ideology as Furdey. Their bias may be genuinely for scientific purposes but I think not. Their motive more likely is synonymous with the old cliché that plagues polygraph, "My question format is better then yours." Why argue so intensely over this issue of the question format to use? Polygraph is attempting to even further standardize an already standardized method. Searching for further standardized format, polygraph looked to the academic community because of its? wealth of resources, ability to formulate experimental design, and ability to conduct extensive controlled laboratory research on the methods. In the academic community, those who posses the acceptable methods are the ones who get the research grant money to perfect and substantiate their methods. This may be somewhat of an off issue for this discussion. However, if one looks at why polygraph has not been overwhelmingly accepted in the scientific community regardless of its? high validity marks, one can see that lack of agreement is a major issue that hold up is overall acceptance. I have spoken with scientists, psychologists, and many other scientific disciplines. These people say that polygraph is valid, not flawless. Again, the reoccurring theme that hinders polygraph is the lack of agreement amongst the ranks. The irony is that those who are causing such confusion and thwarting the acceptance of polygraph as a standardized scientific method are the ones who were sought out to aid in doing just the opposite. Further, these individuals are not even polygraph examiners. "Those who can do. Those who can?t teach."
Further more, Lykken?s argument against presence of standardization and control in polygraph is elusive babble. He say?s things like, "There are virtually an infinite number of dimensions along which the R [relevant] and the so-called "C" ["control"] items of the CQT could differ. These differences include such dimensions as time (immediate versus distant past), potential penalties (imprisonment and a criminal record versus a bad conscience), and amount of time and attention paid to "developing" the questions (limited versus extensive). Accordingly, no logical inference is possible based on the R versus "C" comparison. For those concerned with the more applied issue of evaluating the accuracy of the CQT procedure, it is the procedure?s in-principle lack of standardization that is more critical." He has haphazardly taken terms used in polygraph, thrown them about into a paragraph, imposed his own opinion of there meanings and uses without reference to support, and finally drawn a conclusion that has nothing to do with the previous statements made. The fact is, none of Lykken?s gibberish has a thing to do with the standardized methods of polygraph and/or even the physiological data for which its? findings are based on.
You state:
Quote:
Other uncontrolled (and uncontrollable) variables that may reasonably be expected to affect the outcome of a polygraph interrogation include the subject's level of knowledge about CQT polygraphy (that is, whether he/she understands that it's a fraud) and whether the subject has employed countermeasures.
I can with quite certainty say that you have no research or data to support your opinion on this issue. There are no published research studies that have measured specifically for this variance you speak of and/or one that has concluded with data that will concur with your ideology.
You then use Dr. Richardson?s explanation in attempts to illustrate a scientific procedure that has been accepted and its? basis for acceptance.
Quote: You will notice in Drew?s explanation that he states, "if the test works". To be fair, maybe all forensic sciences should be put to the standards of validity measurement that is held against polygraph, inconclusive results included. It is a known fact that even controlled testing for proficiency purposes in accepted scientific practices can often go a rye. When this happens, the results can be inconclusive and/or even false. For example I will give a hypothetical scenario; A standardized control sample of urine leaves an accredited proficiency testing company. That sample is known to contain benzoylecgonine. The test is weather the sample contains benzoylecgonine or not. During the shipping process the cabin of the airplane that contains the samples loses atmospheric pressure. The loss of pressure causes the cabin temperature to plummet to -80 degrees F. When the airplane descends, the cabin pressure and temperature returns to normal atmospheric conditions for the region, let us say 70 degrees F for this hypothetical scenario. This change can happen quite quickly considering the sometimes rapid descent of airplanes. The sample arrives at the lab and is tested. It is found to contain no presence of benzoylecgonine and is reported as such.
Drew also speaks about the test for known to verify the test and the instrumentation works. A polygraph examiner should be conducting an acquaintance exam/test, as stated in APA polygraph procedures outline. This exam/test checks for the ability of the test to work. If the subject has an autonomic response to the known lie, the test works. If the subject does not have an autonomic response to the known lie, the test does not work. The subject is instructed not to move and to follow specific directions. If the subject does not cooperate and attempts to augment his responses in any way on this non-intrusive exam/test, the subject is intentionally attempting to hide his natural responses. I know of only one reason for someone to augment his or her response. That is, they are going to attempt to deceive. I believe any reasonable person would come to the same conclusion. Now if the exam/test works, I have a true physiological response created by a known lie and a true homeostasis or tonic level measurement contained in the known truth. This data can be used to confirm the remainder of the exam/test data collected.
You referenced, for comparison purposes, phrenology and graphology. Phrenology and graphology measures no known and/or research proven methods. These once experimental methods do not even remotely compare to polygraphs extensive research, documentation of known and proven physiological responses, and proven accuracy. A closer but distant comparison you may have used would be questioned documents, since it is a forensic science. From your same source of information,
http://www.skepdic.com/graphol.html , the following appears. "Real handwriting experts are known as forensic document examiners, not as graphologists. Forensic (or questioned) document examiners consider loops, dotted "i's" and crossed "t's," letter spacing, slants, heights, ending strokes, etc. They examine handwriting to detect authenticity or forgery." I believe the author of this site accepts questioned documents as a scientific discipline. Polygraph measures known physiological responses of the subject to detect deception. Polygraph has provided more favorable research, standardization and validity then questioned documents. I know you have read the research study in which polygraph was putt head to head against questioned documents and latent fingerprints.
Although it is not my burden to prove the validity, I will give an example of how polygraphs tested validity stands up against other accepted science:
Quote: From
http://www.iivs.org/news/3t3.html
STATEMENT ON THE SCIENTIFIC VALIDITY OF THE 3T3 NRU PT TEST (AN in vitro TEST FOR PHOTOTOXIC POTENTIAL)
At its 9th meeting, held on 1-2 October 1997 at the European Centre for the Validation of Alternative Methods (ECVAM), Ispra, Italy, the ECVAM Scientific Advisory Committee (ESAC) 1 unanimously endorsed the following statement:
The results obtained with the 3T3 NRU PT test in the blind trial phase of the EU/COLIPA2 international validation study on in vitro tests for phototoxic potential were highly reproducible in all the nine laboratories that performed the test, and the correlations between the in vitro data and the in vivo data were very good. The Committee therefore agrees with the conclusion from this formal validation study that the 3T3 NRU PT is a scientifically validated test which is ready to be considered for regulatory acceptance.
???
General information about the study:
A. The study was managed by a Management Team consisting of representatives of the European Commission and COLIPA, under the chairmanship of Professor Horst Spielmann (ZEBET, BgVV, Berlin, Germany). The following laboratories participated in the blind trial on the 3T3 NRU PT test: ZEBET (the lead laboratory), Beiersdorf (Hamburg, Germany), University of Nottingham (Nottingham, UK), Henkel (Dusseldorf, German), Hoffman-La Roche (Basel, Switzerland), L'Oréal (Aulnay-sous-Bois, France), Procter & Gamble (Cincinnati, USA), Unilever (Sharnbrook, UK), and Warsaw Medical School (Warsaw, Poland).
B. This study began in 1991, as a joint initiative of the European Commission and COLIPA. Phase I of the study (1992-93) was designed as a prevalidation phase, for test selection and test protocol optimisation. Phase II (1994-95) involved a formal validation trial, conducted under blind conditions on 30 test materials which were independently selected, coded and distributed to nine laboratories. The results obtained were submitted to an indepedent statistician for analysis. Data analysis and preparation of the final report took place during 1996-97.
C. A number of tests at different stages of development were included in the study, but the 3T3 NRU PT test was found to be the one most ready for validation. It is a cytotoxicity test, in which Balb/c mouse embryo-derived cells of the 3T3 cell line are exposed to test chemicals with and without exposure to UVA under carefully defined conditions. Cytotoxicity is measured as inhibition of the capacity of the cell cultures to take up a vital dye, neutral red. The prediction model requites a sufficient increase in toxicity in the presence of UVA for a chemical to be labelled as having phototoxic potential.
D. Two versions of the prediction model were applied by the independent statistician. The phototoxicity factor (PTF) version compared two equi-effective concentrations (the IC50 value, defined as the concentration of test chemical which reduces neutral red uptake by 50%) with and without UV light. However, since no 1C50 value was obtained for some chemicals in the absence of UVA, another version was devised, based on the Mean Phototoxic Effect (MPE), whereby all parts of the dose-response curves could be compared.
The two versions of the prediction model were applied to classify the phototoxic potentials of the 30 test chemicals on the basis of the in vitro data obtained in the nine laboratories. Comparing these in vitro classifications with the in vivo classifications independently assigned to the chemicals before the blind trial began, the following overall contingency statistics were obtained for the 3T3 NRU PT test:
PIF version MPE version
Specificity: 90% 93%
Sensitivity: 82% 84%
Positive predictivity: 96% 96%
Negative predictivity: 64% 73%
Accuracy: 88% 92%
E. Other methods in the study included the human keratinocyte NRU PT test, the red blood cell PT test, the SOLATEX PT test, the histidine oxidation test, a protein binding test, the Skin2 ZK1350 PT test, and a complement PT test. The other methods showed varying degrees of promise, e.g. as potential mechanistic tests for certain kinds of phototoxicity, and this will be the subject of further reports.
Considering the above, I would conclude that polygraph has provided more then sufficient overall accuracy data and done so over a greater test period of both laboratory and field settings to prove scientifically valid.
Comparison example:
Quote: From:
http://www.polygraphplace.com/docs/acr.htm In their recent review, Raskin and his colleagues (12) also examined the available field studies of the CQT. They were able to find four field studies (13) that met the above criteria for meaningful field studies of psychophysiological detection of deception tests. The results of the independent evaluations for those studies are illustrated in Table 2. Overall, the independent evaluations of the field studies produce results that are quite similar to the results of the high quality laboratory studies. The average accuracy of field decisions for the CQT was 90.5 percent. (14) However, with the field studies nearly all of the errors made by the CQT were false positive errors. (15)
http://www.polygraphplace.com/docs/AMICUS%20CURIAE%20RE%20THE%20POLYGRAPH%20Draf... aSub-group of subjects confirmed by confession and evidence.
bDecision based only on comparisons to traditional control questions.
cResults from the mean blind rescoring of the cases "verified with maximum certainty" (p.235)
dThese results are from an independent evaluation of the "pure verification" cases.
_______________
Although the high quality field studies indicate a high accuracy rate for the CQT, all of the data represented in Table 2 were derived from independent evaluations of the physiological data. This is a desirable practice from a scientific viewpoint, because it eliminates possible contamination (e.g. knowledge of the case facts, and the overt behaviors of the subject during the examination) in the decisions of the original examiners. However, independent evaluators rarely offer testimony in legal proceedings. It is usually the original examiner who gives testimony. Thus, accuracy rates based on the decisions of independent evaluators may not be the true figure of merit for legal proceedings. Raskin and his colleagues have summarized the data from the original examiners in the studies reported in Table 2, and for two additional studies that are often cited by critics of the CQT. (16) The data for the original examiners are presented in Table 3. These data clearly indicate that the original examiners are even more accurate than the independent evaluators.
<
http://www.polygraphplace.com/docs/AMICUS%20CURIAE%20RE%20THE%20POLYGRAPH%20Draf...;
aCases where all questions were confirmed.
bIncludes all cases with some confirmation.
The above comparison supports the assertion that polygraph when using the CQT meets standard validity test requirements to be considered a scientific test method.