Additional comparisons of randomization-test procedures for single-case multiple-baseline designs: Alternative effect types
| dc.contributor.author | Levin, Joel R. | |
| dc.contributor.author | Ferron, John M. | |
| dc.contributor.author | Gafurov, Boris S. | |
| dc.date.accessioned | 2017-11-02T23:51:28Z | |
| dc.date.available | 2017-11-02T23:51:28Z | |
| dc.date.issued | 2017-08 | |
| dc.identifier.citation | Additional comparisons of randomization-test procedures for single-case multiple-baseline designs: Alternative effect types 2017, 63:13 Journal of School Psychology | en |
| dc.identifier.issn | 00224405 | |
| dc.identifier.doi | 10.1016/j.jsp.2017.02.003 | |
| dc.identifier.uri | http://hdl.handle.net/10150/625957 | |
| dc.description.abstract | A number of randomization statistical procedures have been developed to analyze the results from single-case multiple-baseline intervention investigations. In a previous simulation study, comparisons of the various procedures revealed distinct differences among them in their ability to detect immediate abrupt intervention effects of moderate size, with some procedures (typically those with randomized intervention start points) exhibiting power that was both respectable and superior to other procedures (typically those with single fixed intervention start points). In Investigation 1 of the present follow-up simulation study, we found that when the same randomization-test procedures were applied to either delayed abrupt or immediate gradual intervention effects: (1) the powers of all of the procedures were severely diminished; and (2) in contrast to the previous study's results, the single fixed intervention start-point procedures generally outperformed those with randomized intervention start points. In Investigation 2 we additionally demonstrated that if researchers are able to successfully anticipate the specific alternative effect types, it is possible for them to formulate adjusted versions of the original randomization-test procedures that can recapture substantial proportions of the lost powers. | |
| dc.language.iso | en | en |
| dc.publisher | PERGAMON-ELSEVIER SCIENCE LTD | en |
| dc.relation.url | http://linkinghub.elsevier.com/retrieve/pii/S0022440517300171 | en |
| dc.rights | © 2017 Society for the Study of School Psychology. Published by Elsevier Ltd. All rights reserved. | en |
| dc.rights.uri | http://rightsstatements.org/vocab/InC/1.0/ | |
| dc.subject | Single-case intervention research | en |
| dc.subject | Multiple-baseline design | en |
| dc.subject | Randomization statistical tests | en |
| dc.subject | Alternative effect types | en |
| dc.title | Additional comparisons of randomization-test procedures for single-case multiple-baseline designs: Alternative effect types | en |
| dc.type | Article | en |
| dc.contributor.department | University of Arizona | en |
| dc.identifier.journal | Journal of School Psychology | en |
| dc.description.note | 24 month embargo; Available online 9 March 2017. | en |
| dc.description.collectioninformation | This item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu. | en |
| dc.eprint.version | Final accepted manuscript | en |
| html.description.abstract | A number of randomization statistical procedures have been developed to analyze the results from single-case multiple-baseline intervention investigations. In a previous simulation study, comparisons of the various procedures revealed distinct differences among them in their ability to detect immediate abrupt intervention effects of moderate size, with some procedures (typically those with randomized intervention start points) exhibiting power that was both respectable and superior to other procedures (typically those with single fixed intervention start points). In Investigation 1 of the present follow-up simulation study, we found that when the same randomization-test procedures were applied to either delayed abrupt or immediate gradual intervention effects: (1) the powers of all of the procedures were severely diminished; and (2) in contrast to the previous study's results, the single fixed intervention start-point procedures generally outperformed those with randomized intervention start points. In Investigation 2 we additionally demonstrated that if researchers are able to successfully anticipate the specific alternative effect types, it is possible for them to formulate adjusted versions of the original randomization-test procedures that can recapture substantial proportions of the lost powers. |
