Inhoud blog
  • Waarom leerlingen steeds slechter presteren op Nederlandse scholen; en grotendeels ook toepasselijk op Vlaams onderwijs!?
  • Waarom leerlingen steeds slechter presteren op Nederlandse scholen; en grotendeels ook toepasselijk op Vlaams onderwijs!?
  • Inspectie in Engeland kiest ander spoor dan in VlaanderenI Klemtoon op kernopdracht i.p.v. 1001 wollige ROK-criteria!
  • Meer lln met ernstige gedragsproblemen in l.o. -Verraste en verontwaardigde beleidsmakers Crevits (CD&V) & Steve Vandenberghe (So.a) ... wassen handen in onschuld en pakken uit met ingrepen die geen oplossing bieden!
  • Schorsing probleemleerlingen in lager onderwijs: verraste en verontwaardigde beleidsmakers wassen handen in onschuld en pakken uit met niet-effective maatregelen
    Zoeken in blog

    Beoordeel dit blog
      Zeer goed
      Goed
      Voldoende
      Nog wat bijwerken
      Nog veel werk aan
     
    Onderwijskrant Vlaanderen
    Vernieuwen: ja, maar in continuïteit!
    27-11-2013
    Klik hier om een link te hebben waarmee u dit artikel later terug kunt lezen.Veel kritiek op PISA en PISA-topman Schleicher
    Veel kritiek op PISA en topman Andreas Schleicher: statistische blunders, misbruik van PISA door OESO e.d.

    1.Er verschijnen steeds meer bijdragen over foute statistische berekeningen in de PISA-studies.

    In vorige bijdragen in dit facebook verwezen we al naar een recente studie van de univ. van Helsinki waaruit bleek dat Finland allesbehalve een onderwijsparadijs is inzake qualit...
    y en equity zoals het voorgesteld wordt in PISA-publicaties. De Finse onderzoekers stelden met eigen tests vast dat de Finse 15-jarigen vrij zwak presteerden voor wiskunde en andere basisvakken - en dit in tegenstelling met de PISA-voorstelling. Ze stelden de voorbije 11 jaar ook een grote achteruitgang vast en slaan alarm. De Finse beleidsmakers werden blijkbaar misleid door de hoge PISA-score en verwaarloosden de niveaubewaking.

    2. Hadley schrijft over PISA en Schleicher: It is absurd that someone with no academic status as a statistician can have so much influence on the educational research. The world's greatest experts on statistics have heavily criticised Schleicher's methodology in peer reviewed papers and in articles in the press. Anyone who doubts this should Google any one of Professor Spiegelhalter from Cambridge, Professor Svend Kreiner and Professor Karl Bang Christiensen from Copenhagen, Professor Harvey Goldstein from Bristol and Oxford’s Professor Jenny Ozga, look at their academic record and see what they say about PISA. They are not shrills in the pay of the teachers' unions or left wing zealots trying to harm our children's education. They are gifted independent people who simply know what they are talking about with no axe to grind, who can recognise an abuse of statistics when they see it. Mr Schleicher simply does not really understand what he is talking about and it is embarrassing that so many people take so much notice of him.

    3.In The Guardian lazen we gisteren in een bijdrage van Peter Wilby volgende kritieken.

    *But does Pisa have sufficiently robust data to justify its growing weight? Many critics think not. For one thing, the tests don't work as people think they work. You would expect all pupils to answer the same questions. In fact, according to an analysis by Copenhagen University in Denmark, only 10% of those who took part in Pisa 2006 were tested on all 28 reading questions, and about half weren't tested on reading at all. The OECD feeds real scores into a statistical device called the Rasch model so that it can work out "plausible values" for children who weren't tested. Schleicher says: "We want to test lots of different things but we have limited time. So we give the students different tests with overlapping content." He says it's a long-established statistical technique, enhanced by modern technology. But some statisticians insist it can't work for Pisa, because different test items work differently in different countries. Pisa's league tables, they say, are almost meaningless. In 2006, the UK could have finished anywhere between 14th and 30th on reading, Canada anywhere between second and 25th, Japan anywhere between eighth and 40th.

    *A second objection to Pisa is that it can't take account of social, economic and cultural differences. Can it make sense to compare, say, Peru, which has high levels of child labour and limited internet access, with western European countries? Can UK schools really learn anything from east Asian countries, with their deeply ingrained respect for authority?

    *Pisa critic: Schleicher's work threatens a global standardisation of education, wiping out school systems that were embedded in diverse local cultures, values and traditions. "The very meaning of public education is being recast," write the American academics Heinz-Dieter Meyer and Aaron Benavot, editors of Pisa, Power and Policy, a collection of learned papers published this year, "from a project aimed at forming national citizens and nurturing social solidarity to a project driven by economic demands." Schools, they argue, are increasingly "subject to the imperatives of efficiency, calculability, predictability and control".

    Until recently, almost nobody questioned the merits of Pisa and its league tables, only their interpretation. Now, as the world's schoolmaster announces the latest planetary exam results, we can expect the political arguments to intensify.

     4. We citeren ook nog even uit davidts' blog 

    The PISA methodology is complex and rather opaque, in spite of the substantial amount of material published in the technical reports. Briefly:

    Individual students only answer a minority of questions. Multiple ‘plausible values’ are then generated for all students assuming a particular statistical model, essentially estimating what might have happened if the student had answered all the questions. These ‘plausible values’ are then treated as if they are the results of complete surveys, and form the basis of national scores (and their uncertainties) and hence rankings in league tables. But the statistical model used to generate the ‘plausible scores’ is demonstrably inadequate – it does not fit the observed data. This means the variability in the plausible scores is underestimated, which in turn means the uncertainty in the national scores is underestimated, and hence the rankings are even less reliable than claimed. Here's a little more detail on these steps.

    (1). Individual students only answer a minority of questions.

    Svend Kreiner has calculated that in 2006, about half did not answer any reading questions at all, while "another 40 per cent of participating students were tested on just 14 of the 28 reading questions used in the assessment. So only approximately 10 per cent of the students who took part in Pisa were tested on all 28 reading questions."

    (2). Multiple ‘plausible values’ are then generated for all students assuming a particular statistical model

    A simple Rasch model (PISA Technical Report , Chapter 9) is assumed, and five values for each student are generated at random from the 'posterior' distribution given the information available on that student. So for the half of students in 2006 who did not answer any reading questions, five 'plausible' reading scores are generated on the basis of their responses on other subjects.

    (3). These ‘plausible values’ are then treated as if they are the results of surveys with complete data on all students

    The Technical Report is not clear about how the final country scores are derived, but the Data Analysis manual makes clear that these are based on the five plausible values generated for each student: they then use standard methods to inflate the sampling error to allow for the use of 'imputed' data.

    “Secondly, PISA uses imputation methods, denoted plausible values, for reporting student performance. From a theoretical point of view, any analysis that involves student performance estimates should be analysed five times and results should be aggregated to obtain: (i) the final estimate; and (ii) the imputation error that will be combined with the sampling error in order to reflect the test unreliability on the standard error.

    All results published in the OECD initial and thematic reports have been computed accordingly to these methodologies, which means that the reporting of a country mean estimate and its respective standard error requires the computation of 405 means as described in detail in the next sections.” There does seem to be some confusion in the PISA team about this - in my interview with Andreas Schleicher, I explicitly asked whether the country scores were based on the 'plausible values', and he appeared to deny that this was the case.

    (4). The statistical model used to generate the ‘plausible scores’ is demonstrably inadequate.

    Analysis using imputed ('plausible') data is not inherently unsound, provided (as PISA do) the extra sampling error is taken into account. But the vital issue is that the adjustment for imputation is only valid if the model used to generate the plausible values can be considered 'true', in the sense that the generated values are reasonably 'plausible' assessments of what that student would have scored had they answered the questions.

    A simple Rasch model is assumed by PISA, in which questions are assumed to have a common level of difficulty across all countries - questions with clear differences are weeded out as “dodgy”. But in a paper in Psychometrika, Kreiner has shown the existence of substantial Differential Item Functioning” (DIF) - i.e. questions have different difficulty in different countries, and concludes that the “The evidence against the Rasch model is overwhelming.”

    The existence of DIF is acknowledged by Adams (who heads the OECD analysis team), who says “The sample sizes in PISA are such that the fit of any scaling model, particularly a simple model like the Rasch model, will be rejected. PISA has taken the view that it is unreasonable to adopt a slavish devotion to tests of statistical significance concerning fit to a scaling model.”. Kreiner disagrees, and argues that the effects are both statistically significant and practically important.

     (5) This means the variability in the plausible scores is underestimated

    The crucial issue, in my view, is that since these 'plausible values' are generated from an over-simplified model, they will not represent plausible values as if the student really had answered all the questions. Kreiner says “The effect of using plausible values generated by a flawed model is unknown”.

    I would be more confident than this, and would expect that the 'plausible values' will be ‘under-dispersed’, ie not show a reasonable variability. Hence the uncertainty about all the derived statistics, such as mean country scores, will be under-estimated, although the extent of this under-estimation is unknown. It is notable that PISA acknowledge the uncertainty about their rankings (although this is not very prominent in their main communications), but the extra variability due to the use of potentally-inappropriate plausible values will inevitably mean that the rankings would be even less reliable than claimed. That is the reason for my scepticism about PISA's detailed rankings.


    Geef hier uw reactie door
    Uw naam *
    Uw e-mail *
    URL
    Titel *
    Reactie * Very Happy Smile Sad Surprised Shocked Confused Cool Laughing Mad Razz Embarassed Crying or Very sad Evil or Very Mad Twisted Evil Rolling Eyes Wink Exclamation Question Idea Arrow
      Persoonlijke gegevens onthouden?
    (* = verplicht!)
    Reacties op bericht (0)



    Archief per week
  • 30/04-06/05 2018
  • 23/04-29/04 2018
  • 16/04-22/04 2018
  • 09/04-15/04 2018
  • 02/04-08/04 2018
  • 26/03-01/04 2018
  • 19/03-25/03 2018
  • 12/03-18/03 2018
  • 05/03-11/03 2018
  • 26/02-04/03 2018
  • 19/02-25/02 2018
  • 12/02-18/02 2018
  • 05/02-11/02 2018
  • 29/01-04/02 2018
  • 22/01-28/01 2018
  • 15/01-21/01 2018
  • 08/01-14/01 2018
  • 01/01-07/01 2018
  • 25/12-31/12 2017
  • 18/12-24/12 2017
  • 11/12-17/12 2017
  • 04/12-10/12 2017
  • 27/11-03/12 2017
  • 20/11-26/11 2017
  • 13/11-19/11 2017
  • 06/11-12/11 2017
  • 30/10-05/11 2017
  • 23/10-29/10 2017
  • 16/10-22/10 2017
  • 09/10-15/10 2017
  • 02/10-08/10 2017
  • 25/09-01/10 2017
  • 18/09-24/09 2017
  • 11/09-17/09 2017
  • 04/09-10/09 2017
  • 28/08-03/09 2017
  • 21/08-27/08 2017
  • 14/08-20/08 2017
  • 07/08-13/08 2017
  • 31/07-06/08 2017
  • 24/07-30/07 2017
  • 17/07-23/07 2017
  • 10/07-16/07 2017
  • 03/07-09/07 2017
  • 26/06-02/07 2017
  • 19/06-25/06 2017
  • 05/06-11/06 2017
  • 29/05-04/06 2017
  • 22/05-28/05 2017
  • 15/05-21/05 2017
  • 08/05-14/05 2017
  • 01/05-07/05 2017
  • 24/04-30/04 2017
  • 17/04-23/04 2017
  • 10/04-16/04 2017
  • 03/04-09/04 2017
  • 27/03-02/04 2017
  • 20/03-26/03 2017
  • 13/03-19/03 2017
  • 06/03-12/03 2017
  • 27/02-05/03 2017
  • 20/02-26/02 2017
  • 13/02-19/02 2017
  • 06/02-12/02 2017
  • 30/01-05/02 2017
  • 23/01-29/01 2017
  • 16/01-22/01 2017
  • 09/01-15/01 2017
  • 02/01-08/01 2017
  • 26/12-01/01 2017
  • 19/12-25/12 2016
  • 12/12-18/12 2016
  • 05/12-11/12 2016
  • 28/11-04/12 2016
  • 21/11-27/11 2016
  • 14/11-20/11 2016
  • 07/11-13/11 2016
  • 31/10-06/11 2016
  • 24/10-30/10 2016
  • 17/10-23/10 2016
  • 10/10-16/10 2016
  • 03/10-09/10 2016
  • 26/09-02/10 2016
  • 19/09-25/09 2016
  • 12/09-18/09 2016
  • 05/09-11/09 2016
  • 29/08-04/09 2016
  • 22/08-28/08 2016
  • 15/08-21/08 2016
  • 25/07-31/07 2016
  • 18/07-24/07 2016
  • 11/07-17/07 2016
  • 04/07-10/07 2016
  • 27/06-03/07 2016
  • 20/06-26/06 2016
  • 13/06-19/06 2016
  • 06/06-12/06 2016
  • 30/05-05/06 2016
  • 23/05-29/05 2016
  • 16/05-22/05 2016
  • 09/05-15/05 2016
  • 02/05-08/05 2016
  • 25/04-01/05 2016
  • 18/04-24/04 2016
  • 11/04-17/04 2016
  • 04/04-10/04 2016
  • 28/03-03/04 2016
  • 21/03-27/03 2016
  • 14/03-20/03 2016
  • 07/03-13/03 2016
  • 29/02-06/03 2016
  • 22/02-28/02 2016
  • 15/02-21/02 2016
  • 08/02-14/02 2016
  • 01/02-07/02 2016
  • 25/01-31/01 2016
  • 18/01-24/01 2016
  • 11/01-17/01 2016
  • 04/01-10/01 2016
  • 28/12-03/01 2016
  • 21/12-27/12 2015
  • 14/12-20/12 2015
  • 07/12-13/12 2015
  • 30/11-06/12 2015
  • 23/11-29/11 2015
  • 16/11-22/11 2015
  • 09/11-15/11 2015
  • 02/11-08/11 2015
  • 26/10-01/11 2015
  • 19/10-25/10 2015
  • 12/10-18/10 2015
  • 05/10-11/10 2015
  • 28/09-04/10 2015
  • 21/09-27/09 2015
  • 14/09-20/09 2015
  • 07/09-13/09 2015
  • 31/08-06/09 2015
  • 24/08-30/08 2015
  • 17/08-23/08 2015
  • 10/08-16/08 2015
  • 03/08-09/08 2015
  • 27/07-02/08 2015
  • 20/07-26/07 2015
  • 13/07-19/07 2015
  • 06/07-12/07 2015
  • 29/06-05/07 2015
  • 22/06-28/06 2015
  • 15/06-21/06 2015
  • 08/06-14/06 2015
  • 01/06-07/06 2015
  • 25/05-31/05 2015
  • 18/05-24/05 2015
  • 11/05-17/05 2015
  • 04/05-10/05 2015
  • 27/04-03/05 2015
  • 20/04-26/04 2015
  • 13/04-19/04 2015
  • 06/04-12/04 2015
  • 30/03-05/04 2015
  • 23/03-29/03 2015
  • 16/03-22/03 2015
  • 09/03-15/03 2015
  • 02/03-08/03 2015
  • 23/02-01/03 2015
  • 16/02-22/02 2015
  • 09/02-15/02 2015
  • 02/02-08/02 2015
  • 26/01-01/02 2015
  • 19/01-25/01 2015
  • 12/01-18/01 2015
  • 05/01-11/01 2015
  • 29/12-04/01 2015
  • 22/12-28/12 2014
  • 15/12-21/12 2014
  • 08/12-14/12 2014
  • 01/12-07/12 2014
  • 24/11-30/11 2014
  • 17/11-23/11 2014
  • 10/11-16/11 2014
  • 03/11-09/11 2014
  • 27/10-02/11 2014
  • 20/10-26/10 2014
  • 13/10-19/10 2014
  • 06/10-12/10 2014
  • 29/09-05/10 2014
  • 22/09-28/09 2014
  • 15/09-21/09 2014
  • 08/09-14/09 2014
  • 01/09-07/09 2014
  • 25/08-31/08 2014
  • 18/08-24/08 2014
  • 11/08-17/08 2014
  • 04/08-10/08 2014
  • 28/07-03/08 2014
  • 21/07-27/07 2014
  • 14/07-20/07 2014
  • 07/07-13/07 2014
  • 30/06-06/07 2014
  • 23/06-29/06 2014
  • 16/06-22/06 2014
  • 09/06-15/06 2014
  • 02/06-08/06 2014
  • 26/05-01/06 2014
  • 19/05-25/05 2014
  • 12/05-18/05 2014
  • 05/05-11/05 2014
  • 28/04-04/05 2014
  • 14/04-20/04 2014
  • 07/04-13/04 2014
  • 31/03-06/04 2014
  • 24/03-30/03 2014
  • 17/03-23/03 2014
  • 10/03-16/03 2014
  • 03/03-09/03 2014
  • 24/02-02/03 2014
  • 17/02-23/02 2014
  • 10/02-16/02 2014
  • 03/02-09/02 2014
  • 27/01-02/02 2014
  • 20/01-26/01 2014
  • 13/01-19/01 2014
  • 06/01-12/01 2014
  • 30/12-05/01 2014
  • 23/12-29/12 2013
  • 16/12-22/12 2013
  • 09/12-15/12 2013
  • 02/12-08/12 2013
  • 25/11-01/12 2013
  • 18/11-24/11 2013
  • 11/11-17/11 2013
  • 04/11-10/11 2013
  • 28/10-03/11 2013
  • 21/10-27/10 2013

    E-mail mij

    Druk op onderstaande knop om mij te e-mailen.


    Gastenboek

    Druk op onderstaande knop om een berichtje achter te laten in mijn gastenboek


    Blog als favoriet !

    Klik hier
    om dit blog bij uw favorieten te plaatsen!


    Blog tegen de wet? Klik hier.
    Gratis blog op https://www.bloggen.be - Meer blogs