Premium
The Application of the Central Limit Theorem and the Law of Large Numbers to Facial Soft Tissue Depths: T‐Table Robustness and Trends since 2008 ,
Author(s) -
Stephan Carl N.
Publication year - 2014
Publication title -
journal of forensic sciences
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.715
H-Index - 96
eISSN - 1556-4029
pISSN - 0022-1198
DOI - 10.1111/1556-4029.12328
Subject(s) - pooling , central limit theorem , population , statistics , table (database) , robustness (evolution) , sample (material) , proxy (statistics) , mathematics , soft tissue , computer science , medicine , data mining , artificial intelligence , biology , surgery , chemistry , biochemistry , environmental health , chromatography , gene
By pooling independent study means ( x ¯ ), the T‐Tables use the central limit theorem and law of large numbers to average out study‐specific sampling bias and instrument errors and, in turn, triangulate upon human population means (μ). Since their first publication in 2008, new data from >2660 adults have been collected ( c .30% of the original sample) making a review of the T‐Table's robustness timely. Updated grand means show that the new data have negligible impact on the previously published statistics: maximum change = 1.7 mm at gonion; and ≤1 mm at 93% of all landmarks measured. This confirms the utility of the 2008 T‐Table as a proxy to soft tissue depth population means and, together with updated sample sizes (8851 individuals at pogonion), earmarks the 2013 T‐Table as the premier mean facial soft tissue depth standard for craniofacial identification casework. The utility of the T‐Table, in comparison with shorths and 75‐shormaxes, is also discussed.