AustLII Home | Databases | WorldLII | Search | Feedback

University of Technology Sydney Law Research Series

You are here:  AustLII >> Databases >> University of Technology Sydney Law Research Series >> 2018 >> [2018] UTSLRS 10

Database Search | Name Search | Recent Articles | Noteup | LawCite | Author Info | Download | Help

Vijeyarasa, Ramona and Bello y Villarino, Jose-Miguel --- "The Indicator Fad: How Quantifiable Measurement Can Work Hand-in-hand with Human Rights - A Response to Sally Engle Merry's The Seductions of Quantification" [2018] UTSLRS 10; (2018) 50 International Law and Politics 985

Last Updated: 21 September 2020

The Indicator Fad: How Quantifiable Measurement Can Work Hand-in-Hand with Human Rights—A Response to Sally Engle Merry’s The Seductions of Quantification

José-Miguel Bello y Villarino & Ramona Vijeyarasa, Ph.D[∗]

I. Introduction

Indicators, understood broadly to mean any type of organized categories that can be individually assessed—quantitatively or qualitatively—have received escalated attention in the twenty-first century, both as a way of understanding legal and social environments and as a method of presenting information for further evaluation. This heightened public interest in what are most commonly institutionally-created numerical rankings or scores has been paralleled by academic attention, with leading scholars, research centers, and even international organizations, showing great concern for—and devoting significant resources to—the question of what can, or should, be measured and how.[1]
Among the social aspects which can be both measured and compared, human rights—understood broadly—figure very highly. If there are some inalienable and global rights that are linked to the human condition and dignity, measuring the enjoyment of those rights in different parts of the world should allow us to compare the extent to which different States are fulfilling their human rights obligations and assess both improvements and declines over time.
Some may argue that the core human rights standards were drafted at a time when effective monitoring was neither intended nor achievable.[2] Yet, monitoring of global progress towards more equal, more just, and more sustainable societies is a core component of the human rights and development sectors today. As a result, methods for counting progress have proliferated, triggering a substantial debate about what constitutes a comparable, somewhat objective, universal standard against which all countries can be measured.
Developing such a universal standard is not a simple endeavor. While global human rights treaties set out agreed principles for what are meant to be universally accepted human rights norms, interpretation is an evolutionary process, responding to changes in societies. This is reflected in the issuing of General Recommendations and General Comments, which the nine core human rights treaty bodies have the power to do.[3]
Despite this global norm setting, determining what should be measured in each case—as well as how such measurement should occur—was always likely to stimulate substantial debate. The goal of this process may also be contentious. Do we undertake this exercise for mere comparison purposes, or are we serving the broader goal of monitoring human rights towards their universal enjoyment? Should the outcomes of numerical ranking shape policy or, adopting a more extreme approach, be used as the basis for a system of rewards and punishments in response to a State’s performance?
Regardless of the answers to these questions, the span of this debate proves that we live in an “indicator-prone culture,”[4] which, perhaps, has been accepted without a thorough validation of the need for and purpose of measuring human rights as a whole. In this context, Sally Engle Merry’s book, The Seductions of Quantification, raises valid questions about how and why we have become surrounded by multiple systems of country rankings. For example, there are indices that seek to show us which countries are more corrupt (e.g., the Corruption Perception Index), or less protective of the rights of women (e.g., the Social Institutions and Gender Index, SIGI; or the Gender Inequality Index, GII), or better at guaranteeing our civil liberties (e.g., the Human Freedom Index or the Democracy Index), or overall, simply a better place to live (e.g., the Human Development Index). Such indices have proliferated without the time to step back and adequately assess what these figures really mean, either individually or when placed side by side.
Merry’s book offers an anthropological and sociological take on the issue of indicators in relation to the field of human rights. She is not the first, and will certainly not be the last, to discuss the intractable problems with human rights compliance, application, and enforcement—in theory and in practice—of which, we believe, indicators are a part. Yet, we approach Merry’s arguments from a legal and practical point of view, as scholars and practitioners in the field of international law.
A reply to her book requires, first, an overview of her work, which we concede is limited for its reductionist character. Merry offers an overarching critique of the indicator culture, both in terms of the process by which indicators are developed and the ways in which they are applied. The three dominant concerns articulated throughout her book are (1) the roles played by power and knowledge in the creation, establishment, affirmation, and reiteration of a limited set of indicators for any one topic;[5] (2) the assumed universal nature of indicators and, in turn, their inability to reflect important contextual and cultural differences; and (3) that the choice and design of indicators define the very problem that is being measured,[6] which alone is problematic, but even more so if done in a way that is “decontextualized, homogenized, and remote from local systems of meaning.”[7] Moreover, Merry argues these indicators are often based on previous indicator projects and hence reinforce the same understanding, blocking out new notions or interpretations of the concept at hand.[8]
In The Seductions of Quantification, Merry highlights why indicators have become so popular. As articulated in her earlier critical pieces, Merry writes that it is often considered easier to assess compliance with a treaty through numbers.[9] In concrete terms, indicators simplify the process of evaluating progress and monitoring it overtime,[10] and they provide comparability, both for individual countries over the years and between countries. Hence, what has resulted from this “seduction” is numerous indices that rank and compare countries and show gains and losses and winners and losers in each area.
In response to this embrace of indicators, Merry offers three indicator projects as examples to lay out her criticisms. First, Merry dedicates two chapters of her book to the issue of violence against women, discussing the process by which the U.N. Statistical Commission (UNSC) set out to develop a set of indicators and guidelines to collect data on violence against women.[11] To make her point, Merry discusses four approaches to the topic of violence: a gender equality approach, a human rights approach, a criminal justice approach, and an approach based on national statistical capacity. By comparing the four approaches, she highlights how while all purport to measure the same idea (i.e., violence against women) in practice, they are “not measuring the same thing at all, even though they are calling it by the same name.”[12]
While Merry concedes that the effort to quantify violence against women has helped to advance women’s rights by bringing greater visibility to the problem, raising awareness, and displacing the belief that violence is a problem for only the “unruly few,”[13] her biggest gripe with the attempt to quantify the issue is the suggestion that any of these approaches can reveal an “objective truth” about violence.[14] To the contrary, any approach—whether a feminist approach or a human rights one—constructs our understanding of violence in the first place. Merry is concerned that the data may be viewed as the result of “an objective scientific process of data collection and analysis.”[15]
Merry’s second indicator project is the annual-produced United States Trafficking in Persons report (“TIP report”). The centerpiece of the TIP reports is a system for ranking countries in terms of their efforts to prevent and respond to human trafficking. Tier 1 countries are considered the most compliant, Tier 3 the least.[16] A fourth tier, the Tier 2 Watch List was added in 2004.[17] For those countries ranked Tier 3, there is a risk of “negative consequences” in their relations with the United States.[18] Merry identifies numerous problems with the TIP report indicators, including how they have influenced what we understand to be trafficking and lack of objectivity, given the system is heavily intertwined with politics—and the political agenda of the U.S. government in particular.[19] As will be discussed in Part V, Merry is not the first author to highlight the clear agendas behind how trafficking is framed and understood. In our view, the TIP report—heavily critiqued by feminist academics, policy analysts, and other governments—is a parochial evaluation, lacking a strong methodology and with results that are too heavily influenced by the politics of the day. As such, the TIP report is perhaps a too convenient choice for Merry to demonstrate the problems with quantification. Moreover, we hardly consider the report to be a quantitative indicator, having a more qualitative nature that is subsequently, in a non-transparent manner, transformed into a ranking with tiers.
The third indicator project that Merry discusses is the development, between 2005 and 2012, by the Office of the High Commissioner for Human Rights (OHCHR) of a set of indicators to measure core human rights, such as the right to education and the right to life.[20] With this example, Merry’s three neat categories become blurred. It is doubtful that Merry intends to suggest that freedom from violence against women is not a human rights concern. A prohibition of gender-based violence has been established as a key principle of customary international law;[21] and one of ten human rights treaty bodies,[22] the Committee on the Elimination of Discrimination against Women (CEDAW), just released a general recommendation on gender-based violence.[23] The first and third indicators projects are thus naturally overlapping.
Similar reasoning could apply to the second and third projects. The set of indicators developed by the OHCHR, for instance, include the “elimination of forced labour,” with human trafficking offered as an example.[24] “Violence at work, forced labour and trafficking” are specifically addressed as part of the violence against women indicators.[25] In the end, all three projects are about measuring compliance with human rights standards, looking into certain rights in more or less detail. In relation to the OHCHR project, Merry’s main concern seems to be a bias towards the global North in the development of those indicators. We will look into the underlying logic of that argument in section II.A.
This article analyzes Merry’s main critiques of the growing interest—Merry might say obsession—with quantifying progress on human rights and development. It assesses whether Merry adequately defends and sustains her critiques through her three examples, and it questions whether these critiques are any different when it comes to other mechanisms for accountability and tracking change within the human rights system at large. It also analyzes Merry’s choice of projects presented in the book as examples: for instance, are these particularly weak examples or are the critiques directed at them equally sustained when it comes to other examples of accountability sought through quantification and rankings? We conclude by arguing that Merry’s view, while raising a valid point (i.e. the enthusiastic use of indicators hides the significant shortcomings of quantification) does not offer a balanced evaluation of how to monitor human rights through quantitative approaches, especially in terms of a cost-benefit analysis and the best means to ensure access to the relevant information presented in a particular data set.

A. Defining the Concepts: What Are Indicators?

The term indicator is used for many different kinds of numerical representations with no real consistency. Its meaning ranges from something that indicates or points to a fact, such as that a person has been trafficked, to an elaborate combination of data merged into a single rank or score.[26] Merry defines quantification as “the use of numbers to describe social phenomena in countable and commensurable terms.”[27] Yet, in her book—as in her previous works—Merry focuses on social phenomena largely related to the enjoyment of human rights rather than other measureable phenomena. Unsurprisingly, an expansive number of social phenomena are currently tracked and quantified, many of which have only limited or indirect connections with human rights. Some examples include patterns of wealth, the use of social media, penetration of smart mobile use, and demographic data related to sexual orientation.
Given that a large body of existing indices, and particularly the better known ones, tend to measure the enjoyment or protection of different aspects of human rights, we begin on the basis that that those indices are often proxies for human rights accountability. In other words, if we measure a population’s access to drinkable water, we would be looking at compliance with the right to water and sanitation. If we rank countries on the basis of access to contraceptives for their citizens, we would be trying to assess the enjoyment of sexual and reproductive rights. If we quantify how much corruption is perceived in a country, we would be doing so because we believe people should live free of corruption and under the rule of law. In the end, we are quantifying human rights. And, yet, this says very little about what we mean when we talk about quantification, because this concept covers many different interpretations of complex processes that are analyzed with very different lenses and at different levels.
In general, instead of discussing “indicators” at large, this article, and most of Merry’s related work, could be more properly described as scrutinizing “human rights indicators,” which have been defined as “information that is presented quantitatively and that can be related to human rights norms and standards; that addresses and reflects the human rights concerns and principles; and that are used to assess and monitor promotion and protection of human rights.”[28] To help explain what is meant by “quantitatively,” in the following section we present five different levels of analysis for how we understand indicators and measurement.

B. Distinguishing Five Levels of Analysis

Indicators can be analyzed from at least five different approaches, or “levels of analysis.”[29] First is the “ontological” level, which examines their nature of being and becoming a reality. Put simply, this normative analysis looks into whether and why indicators should exist. Further, assuming that indicators add value, at this level, we can also interrogate how indicators can be grouped in order to achieve a specific aim, especially if the goal is to establish some kind of hierarchy or taxonomy for the indicators developed. We can also ask whether indicators and their ranks can be subdivided into smaller groups according to similarities and differences.
If we agree that indicators are useful or necessary, as well as a valid way of assessing compliance with human rights, the second level focuses on the construction of those indicators. As a human construction, the choices made in the process of developing an indicator—or a system of indicators—are naturally influenced by cultural backgrounds and related biases. Given that we claim and defend that human rights are by their definition universal, we should be able to measure compliance with them in a globally agreed upon way. At the same time, we should recognize how cultural biases affect those choices. Unlike Merry, who sees indicators as failing to adequately respond to cultural divergences,[30] this second level of analysis enables us to see culture embedded in the choice of what gets measured and how.
Third, assuming that an indicator is developed in a globally accepted manner and that it is feasible to measure it (e.g., quantitatively or qualitatively), we can move into the procedural and technical level, where we need to understand how the necessary information (i.e., inputs) to determine a performance for that indicator (i.e., output) can be collected, processed, validated and, finally, presented. Again, all such steps towards collecting inputs to develop an output are understood as a way of assessing compliance with human rights in a comparable manner, with whatever disclaimers may be necessary.
At the next level, the evaluation level, we can assess what those indicators say about compliance with human rights in a given jurisdiction. We must assess the reliability of the information we have obtained from the process. This is a prerequisite to undertaking a normative analysis of those results, such as is “performance x of country A, better than performance y of country B?” or “what is the desirable/ideal performance in relation to indicator N measuring—for example—compliance with the right to a fair trial?”
Finally, the fifth level is the consequential level. At this stage, we consider whether it is fair and reasonable for performance of a given indicator to result in practical repercussions. One can think of punishments and rewards, or “carrots and sticks” as possible legitimate consequences.
Having set out these five stages—the ontological stage, the construction stage, the procedural and technical stage, the evaluation stage, and finally, the consequential stage—the next sections will assess Merry’s arguments against the stages above. At each of these points, we analyze how Merry defends that indicators are failing to meet their expected effect, how she sustains these criticisms, and what alternatives exist for human rights accountability.

II. The Ontology of Indicators: What They Are and Why We Need Them

Merry’s book opens with a harsh critique of indicators: we have been fooled as to their appropriateness, believing them to be objective when they are not, relying on them in our fight to address global problems, when in fact, they are not universally applicable.[31] Her critique is extended not only to the human rights and the development sector communities but also to the media and civil society at large:

Once indicators are established and settled, they are typically portrayed in the media as accurate descriptions of the world. They offer forms of information that satisfy the unease and anxiety of living in a complex and ultimately unknowable world. They address a desire for unambiguous knowledge, free of political bias. Statistical information can be used to legitimate political decisions as being scientific and evidence-based in a time when politics is questioned. They are buoyed up by the rise in bureaucracy and faith in solutions to problems that rely on statistical expertise. Such technocratic knowledge seems more reliable than political perspectives in generating solutions to problems, since it appears pragmatic and instrumental rather than ideological. These are the seductions of quantification.[32]
While her criticisms are largely substantiated, Merry lacks the background analysis that helps explain the increased interest in quantification, particularly during the past two decades. Moreover, readers are left questioning whether some of Merry’s legitimate critiques are equally directed at other systems of accountability. In other words, is the issue the indicators themselves or the reality of how States Parties respond when being held accountable for their observance of human rights? Finally, does Merry merely present a negative appraisal of the status quo or offer a potential and feasible alternative? In the next sections, we seek to answer some of these questions and offer a more balanced and nuanced view of the use of indicators for the monitoring of compliance with human rights standards and its current shortcomings.

A. Indicators and Their Western Origins

It is hard to disagree with Merry that the interest in and use of indicators is, to a significant extent, a non-inclusive process, or, if collaborative, often driven by representatives of States Parties from Europe and the Americas.[33] The real question here is if this is something particular to the development of indicators created to measure rights and welfare or part of a wider picture in which, as Merry describes it, the “global North”[34] countries—generally rich and developed States—are at the origin of the existing international system of human rights and global governance. We could go even further and claim that, in the end, the whole idea of human rights is a Western (or European) driven idea, shaped through the global and regional international institutions created during the twentieth century. These were, to a significant extent, the result of a promotion of an understanding of human rights supported by the global North.[35]
Based on this reasoning, to offer one example, we can measure political freedom according to certain proxies such as access to information, freedom to manifest political ideas or freedom to vote. In doing so, we adopt a particular way of understanding political freedom. This particular understanding corresponds to the ideas promoted by the bodies entrusted with their monitoring, such as the Human Rights Committee, which monitors the International Covenant on Civil and Political Rights.[36] It cannot be surprising that the output of this Committee is rooted in the ways in which global North countries understood these rights in the first place and the role they played in influencing the drafting of the treaties themselves.
In this respect, there is an intrinsic Western bias with the use of indicators because they try to measure understandings of concepts with a heavy Western influence. Yet, does this problem—if it is indeed problematic—justify a shift away from indicators altogether? For Merry, her concerns with the Western origins of indicators appear to be part of a broader critique of the power dynamics inherent in the development of indicators. We therefore turn to the construction stage—the identification of what gets measured and how.

III. How Are the Indicators Chosen and Developed? Constructing Indicators

Merry draws on Foucault to establish her critique of the tight knit relationship between indicators and power. Power determines what gets measured and the practices of measurement and quantification, which in turn produce knowledge. In the words of Foucault, “power and knowledge directly imply one another.”[37] Moreover, once indicators are established, normalized, and readily used, we tend not to return to the initial concern of who played the most central role in their development. Merry argues, “[O]nce decisions about categorization and aggregation are made, the categories may come to seem objective and natural, while the power exerted in creating them disappears.”[38] In turn, we stop questioning why a certain measure is being taken —or, probably more relevant here, on whose determination—and the outcome is considered a new norm.

A. The Voice of the North and Measuring Violence Against Women

When discussing the first of the three projects—the development of indicators related to violence against women—Merry critiques how the process of developing indicators to measure the gravity of and responses to violence shapes our understanding of the issue: “As the UNSC decided what to measure, how to categorize it, and how to gather data, it defined the phenomenon.”[39] Merry in effect is arguing that the indicators for violence set the boundaries for what society understands to mean by the term violence.
However, for experts from the field of women’s rights or, specifically, violence suffered by women, Merry’s claim is a stretch. Both rape and intimate-partner violence were unquestionably recognized as a form of violence against women long before this exercise. Her comment clearly speaks to the issues of child marriage, female genital mutilation (i.e., FGM, although Merry uses the term “cutting”), sexual slavery during wartime, and violence while in police custody, issues that tend to be included in broader conceptions of violence, rather than narrower approaches that focus on rape and intimate partner violence.[40]. However, again, it is a stretch to suggest that until this moment in time violence was not associated with these rights violations. We must recall that the development of the indicators that Merry analyzes started in the mid-2000s, long after the Convention on the Elimination of Discrimination against Women (CEDAW) was drafted (1989),[41] the Beijing Declaration and Platform for Action issued (1995),[42] or the CEDAW Committee’s General Recommendations No. 12 (1989)[43] and No. 19 (1992)[44] on violence against women adopted.
Take for instance the case of FGM. FGM has long been recognized as not only a form of physical violence but also emotional violence,[45] the violence itself closely intertwined with the perverse control over women’s sexuality that is exercised as a result of the practice. Merry herself makes this point in the pages that follow:

As the national and international focus on the issue developed, it gradually expanded from a narrow concern with rape and domestic violence to a broader analysis of the way patriarchal social structures, cultural understandings of gender difference, and social and economic inequality contributed to a wide range of forms of violence.[46]
In many respects, for readers, the UNSC’s decision to develop a set of indicators on violence seems an odd choice to sustain Merry’s critique. In fact, that process evidences how an extensive, well-debated, and open-minded effort to develop a set of indicators can result in a fairly strong outcome. Merry explains how the initial list of indicators to measure violence against women—a set that were subsequently approved without further discussion—was in fact very narrow.[47] While it included FGM,[48] it excluded a host of other forms of violence, including state violence by police or the military and harassment in the workplace.
However, after much negotiation, the final guidelines, while recognizing the difficulty of measuring the severity of violence, stretched the categories of acts and consequences. The final list of consequences of violence to be measured under that set of indicators includes injuries, mental health repercussions, miscarriage, violence during pregnancy, and fear of the perpetrator,[49] despite discussions during their development about the difficulty of measuring fear.[50] The guidelines also emphasize ethical and safety concerns during interviews, a major concern of the World Health Organization’s (WHO) study and feminist activists involved in the project.[51] Overall, there was an expansion in scope in terms of how violence was framed and experienced as a controlling behavior.
As Merry herself states, “[G]ender equality advocates worked hard to bring their perspective to the UNSC indicator project. . . [and] succeeded in expanding the UNSC’s initial narrow view of violence against women.”[52] Their more expansive questionnaire and detailed ethical and safety recommendations were incorporated into the guidelines, facilitating greater disclosure of the scope of the problem.[53] Recognizing the “triumph” of the feminists at the negotiating table, it is confusing when Merry critiques the fact that “[t]hrough the process of deciding how to measure something like severe physical violence, the concept is established and defined.”[54]
To the contrary, Merry’s extensive discussion suggests that the process of developing an agreed set of what can be measured is a negotiation, creates a dialogue about women’s experiences—in this case, of violence and what matters for women who experience violence—and considers the feasibility of undertaking the measurement that is being proposed. Merry’s direct exposure to the process, as described in the book, highlights that there are many players involved in the development of indicators. Perhaps, therefore, it is less a question of critiquing the indicators themselves but more about influencing the expertise of the people at the table—ensuring for women’s rights issues, for instance, that there are feminist perspectives rooted in international human rights standards.
A more interesting approach regarding bias in the development of indicators would be to analyze the possible underlying reasons for the assumptions that serve as the basis of indicators. Merry, for example, shows her own interpretation of the drivers of trafficking when she discusses her second indicator project, the U.S. State Department’s Trafficking in Persons Reports (TIP reports).[55] When referring to the TIP reports, she states that the reports “expose deficiencies in state efforts to tackle rural poverty or oppressive marriage practices, even though both of these also fuel trafficking.”[56] This assumption contradicts existing research, which challenges the necessary causal relationship between human trafficking and poverty or trafficking[57] and specific manifestations of gender inequality that are often presumed to be relevant without substantiation.[58]
Merry concedes that “experience in measuring violence against women trumped the skepticism of the statisticians.”[59] In fact, the preferred model of the feminists—even if a northern-dominated perspective—was the only one annexed to the UNSC guidelines published in 2013, [60] thereby giving the suggestion that it should be preferred too globally when it comes to national data collection. That a feminist expertise “won” in the exercise of developing indicators on violence against women reflects positively on the potential of indicators to garner interest in measuring change in pertinent women’s rights issues.

B. Expertise Inertia or Learning from Experience?

A final question regarding indicator construction is the appropriateness of Merry's use of the phrase “expertise inertia.”[61] Merry highlights how expert papers from past meetings often form the basis of future discussions and new lists of indicators. She uses the example of a 2007 discussion on indicators related to violence against women, the substance of which was based on formerly used indicators from the U.N. High Commissioner for Refugees, U.N. Statistical Commission for Latin America and the Caribbean (UNECLAC), the U.N. Economic Commission for Europe (UNECE), the European Union, Canada, the United States, United Kingdom, Italy, and Spain.[62] Similarly, a 2009 “Friends of the Chair” of the U.N. Statistical Commission (UNSC) meeting in Mexico relied on a preexisting national survey to finalize its list of indicators.[63] In short, to Merry, these examples demonstrate the tendency to rely on previously applied indicators rather than generate new ones.
However, it is questionable that a reliance on past lessons is necessarily a bad thing. Moreover, Merry notes that not all indicator projects are equally flawed[64] and offers several alternative approaches to measuring violence against women. For example, a project initiated by the UNECE and UNECLAC adopted a “feminist vision of violence against women.”[65] The final survey model included controlling behaviors, emotional abuse, fear, and their effects on everyday activities in addition to physical and sexual violence.[66] The WHO’s study of domestic violence and women’s health from 2005 also helped to influence the UNECE survey instrument, which incorporated a public health perspective in its focus on psychological harm and family relationships.[67]
Merry also notes the work of Yakin Ertürk, the second Special Rapporteur on Violence against Women—successor to Radhika Coomaraswamy—who had a broad vision for what should be included when we measure violence against women. As Merry notes, Erturk proposed three broad indicators: “grave violence,” covering issues like rape, intimate partner violence, sexual harassment; “femicide” (i.e., the murder of women); and third, “social tolerance” (i.e., social and cultural contexts that allow violence to continue).[68] Merry labels Ertuk’s approach as a feminist one, defining violence in its broadest sense, and the battlefield broadly too, to include, for example, violence in the work place.[69]
Readers are left with a strong sense, therefore, that at least in the examples offered on measuring violence against women, we have—particularly with strong feminist expertise at the table—come a long way to a more expansive and gender-sensitive means of measuring experiences of and responses to violence against women and that cumulative processes do not necessarily mean “expertise inertia.”

IV. How to Transform Information into Quantitative Measurements: Universal Indicators and the Differences Lost in the Cracks

Merry shows great concern with how local systems of knowledge “must be squeezed” into global ones.[70] For example, to measure violence against women, throwing acid in the face of one’s wife in Bangladesh is equated to shooting a domestic partner in the United States.[71] There is little scope for nuanced and local difference. This is what Merry calls the problem of “producing commensurability.”[72]
Merry’s concerns echo the debates over the universalism of human rights which in part explain the existence of universal indicators. The strength of the normative framework established by the Universal Declaration of Human Rights (UDHR) in 1948 is that it calls on all states to accept that all human beings are universally entitled to the basic rights and freedoms contained in it.[73] Human rights are considered universal. If we accept this, we should accept that we can create universal systems to measure these global principles.
However, some may still argue that an indicator designed to measure a particular human right will only be as universal as the right itself. For instance, we have seen an extensive number of reservations by Muslim States to the Convention on the Elimination of Discrimination against Women (CEDAW), namely in relation to Articles 2, 9, 10 and, particularly, to 16 on family life.[74] This is despite the fact that CEDAW is one of the most ratified and globally accepted treaties.[75] One could argue that any indicator trying to assess compliance with those articles would not respect the cultural differences of certain countries. Yet, the problem would then lie not in the indicator but in the (contended) “cultural insensitivity” of the right itself.
A related but different issue appears when the right itself is not under debate, but instead the proxy used to measure the right does not adequately capture cultural or social differences.[76] This valid point—which Merry covers succinctly through examples—has been explored in the context of indicators used in the development-sector to measure well-being and enjoyment of fundamental human rights. For example, former Millennium Development Goal (MDG) Three—on the promotion of gender equality and the empowerment of women[77]—has been criticized for its lack of global applicability.[78] One of its indicators was gender parity in education (i.e., Target 3A) at the primary and secondary level.[79] If we apply it to the Latin American context, this region would be assessed as a champion of gender equality, as disparity in school enrollment and completion between boys and girls has been almost eliminated.[80] Yet, gender inequality remains rife, to varying degrees and across a range of issues, across the region.[81] This would suggest the need for a Latin America-specific target on gender equality that does not focus on girls in schools.
A third problem is developing an indicator that ignores local manifestations of broader categories. When discussing her violence against women indicator project, Merry details the difficulty in developing a global measure of violence against women based on severity. In this context, she points out that there are “national and local differences in what is interpreted as a severe act,”[82] giving the example of Bangladesh. At a “Friends of the Chair” meeting, Bangladesh suggested adding to the list of physical violence actions like acid throwing, dropping from a higher place, and a kick in the abdomen; and to the list of psychological violence, adding pressure for dowry, not paying attention to the children, and rebuking the giving of birth to a female child.[83] All examples, Merry points out, are particularly relevant to Bangladesh and South Asia in general,[84] and not including them in an indicator means not measuring an egregious violation of the human rights of women in the region. This is a valid criticism—although it seems that the final version of the indicator covered these situations—but the real question is if non-quantitative existing human rights mechanisms would result in more encompassing, culturally-sensitive measurements.
At the same time, accommodating nuance and local difference, instead of relying on broader categories, can be a problem. Merry does not offer this critique, but a fourth aspect of the issue of commensurability—and an equally problematic one—occurs when an indicator is developed for a country or region-specific problem and then assessed on a global scale. For example, in 2009 the Social Institutions and Gender Index—introduced in 2007 by the Organisation for Economic Cooperation and Development (OECD) and only applicable to countries in the global South—contained only two measures for barriers to women’s enjoyment of their civil liberties, one of which was an obligation to wear a veil or burqa to cover parts of the body in public.[85] Such an indicator reflects a bias from the outset targeting the Muslim religion and is an indicator that makes no sense when applied on a global scale. Similarly, the choice to include a measure on the prevalence of Female Genital Mutilation (FGM)—unquestionably an egregious violation of the rights of women and girls—is also open to critique as this too is not a globally prevalent phenomenon. Is measuring the prevalence of FGM more justified than measuring rates of correctional rape of lesbian women or acid attacks on women?[86]
In practice, indicators designed primarily for the purpose of human rights accountability are no more flawed in their attempt to present a universal measure than the human rights system itself. Having accepted the cons and many pros of a universal human rights system—and we are of the view that human rights are truly universal and must be accepted as such if progress on their enjoyment is to succeed—we must accept that no global measure will be able to account for the many micro nuances among countries. This is the space for country-based indicators, which should be fully able to measure human rights accountability but within a national context.[87]
Merry’s critique is both passionate and understandable: “the complexity of relationships—their histories, the interplay of love and fear, the role of kinship, and residence patterns—disappear.”[88] Problematically, however, the only way to collect this type of data is qualitatively, and it is hard to imagine how countries can collect national data on the scale intended by the UNSC, or the data necessary to allow for any kind of commensurability, if all subjective, qualitative considerations are to be included. It is here that we turn to the next step of collecting and processing data and related practicalities.

V. The Practicalities of Collecting and Processing the Data—The Example of Human Trafficking

As discussed in Part I, it is not enough to accept that an indicator is a valid way to acquire and present information—at the ontological level—and agree on what we are measuring and how—at the construction level. Before creating an indicator, it is also important to consider how the necessary information (i.e., inputs) to determine a performance for that indicator (i.e., output) can be collected, processed, validated, and, finally, presented. At this procedural and technical level, the purpose is to produce a measure that can be compared horizontally—among jurisdictions—or sequentially—in the same jurisdiction, but over a particular number of years.
Here, there are two separate issues worth examining. First, there is the possible existence of an intrinsic bias, where the process of quantification is “adjusted” to potentially achieve a more “palatable” outcome for those who initiated—or controlled or even paid for—the quantification. Second, there is a more technical aspect, which looks into the methodology and transparency of the process itself. To consider these two aspects, we will use Merry’s project on the TIP reports produced by the U.S. Government.
It is important to recall our previous comments on the particular choice of this indicator and its reliability.[89] Despite this, the TIP report is a frequently cited source, precisely because the U.S. Government is the only one that produces an annual ranking of other governments’ actions, something that other reports and organizations operating in the field of trafficking choose not to do.[90] However, considered by some a moral crusade,[91] significant questions have been raised as to whether the U.S. government has access to sufficient data or the political independence to make an accurate assessment.[92] In fact, we suggest that Merry is wrong when she lists other systems that measure trafficking, including that of the ILO, UNODC, and IOM but nonetheless concludes that, “[a]t least for now, trafficking is what the TIP Reports measure.”[93] Many of these other organizations are more respected for their contributions to trafficking and more readily turned to than the U.S. government.[94]
Moreover, it is problematic to discuss a ranking process—calling it an indicator is probably excessive—produced by one actor about the performance on trafficking in persons of its allies and enemies. Merry’s view is that overtime the United States TIP reports’ method of data collection has improved and “it has acquired a more settled and legitimate status.”[95] This is, at best, questionable. As a result, Merry’s decision to dedicate an extensive number of pages to critiquing what is already a heavily critiqued system of rankings is surprising.[96] We therefore respond to Merry’s arguments concerning the development of indicators to measure human trafficking but with reservation: the TIP reports are a poor example, an odd choice because it is somewhat self-serving to prove the shortcomings of indicators at large.
The agenda of the U.S. government necessarily permeates the whole exercise. Yet, this is not particular to the TIP reports or other indicators used to measure human trafficking. There are clear agendas behind how trafficking is framed and understood,[97] and as such, there will also be agendas behind any process designed to measure trafficking. If we undertake a qualitative analysis of trafficking across a number of countries, the reports of different State Parties, the accounts provided by different people consulted for the exercise, and the interpretation of a particular country context—again, based on a set of consulted experts—will always have an element of subjectivity. The best way to address the impact of such subjectivity in the final assessment is to promote transparency by different actors about their potential interests in a particular outcome.
Yet in the case of the TIP reports, this subjectivity engulfs their utility. The politics behind the tier system of ranking countries is very clear and has been critiqued elsewhere.[98] In 2012, for example, Kosovo, Moldova, and Mexico were ranked Tier 2, China was ranked Tier 2 Watch List, whereas Cuba was ranked Tier 3.[99] Some authors note that the report is “fused with the US’ foreign policy agenda” and flawed by “messy slippages with the use of numbers and statistics.”[100] Moreover, as Merry points out, the Tier system is used to determine the U.S. government’s spending in this area, which is not insignificant. Merry notes that by 2005, the U.S. government had spent $295 million combatting trafficking in 120 countries.[101]
Therefore, when Merry critiques the TIP report as being “an apparently objective way of understanding the world,”[102] two responses seem evident. First, the TIP reports are not actually believed to be objective. This is suggested by the fact that quite extensive critiques of the report continue after the noted methodological improvements, even by several of the authors that Merry herself draws upon for her analysis. It is hard to believe that any of the experts working in the field of trafficking would see these reports as objective. Perhaps it is only the U.S. government that presents the report as being objective, but many readers know better.
It is also not new to suggest, as Merry does, that there is a tendency within trafficking circles to repeat the same stories and statistics.[103] There is in fact very limited space for researchers and practitioners working on trafficking to debate outside of an “echo chamber” of repeated ideas.[104] There have also been challenges to the key players in the trafficking movement for defining the very phenomenon under discussion: the key stakeholders involved in trafficking “not only analyse the phenomenon of human trafficking and debate among themselves about the causes and consequences, but in fact establish the lens through which society understands human trafficking.”[105] Moreover, “[s]uch assumptions easily become self-fulfilling: where Siddharth Kara searches for ‘slaves,’ he finds slaves; if we searched for trafficked women who fit the victim archetype, we would indeed find them.”[106]
We are therefore left unsure about how to respond to Merry’s discussion of the TIP reports. This project is a poor choice, particularly compared to the much richer discussion Merry offers on the violence against women indicator projects—and here we emphasize “projects”—or in her discussion of the initiatives to develop human rights indicators presented in chapters seven and eight. Obviously, if all indicators—or even a significant number of them—were developed with the poor methodology and blatant political bias present in the TIP reports, indicators would have never acquired the success[107] or the myth of objectivity[108] that Merry grants them.
However, this is no impediment to acknowledging the value of Merry’s critique of indicators at the procedural and technical level. Even some of the most known (i.e., “successful”) and methodologically robust (i.e., “objective”) indicators, when analyzed carefully, show obvious shortcomings. One recent example has been the publication of a compelling criticism of the Corruption Perception Index (CPI),[109] challenging the CPI for its lack of objective grounding, despite a thorough review of its methodology conducted in 2012.[110] With this in mind, it is essential to always demand a transparent methodology for any kind of quantification or measurement process, so it can be scrutinized and evaluated against the vested interests of those involved, and for the release of any potential conflict of interest in terms of funding, staff interests, organizational values, and objectives.

VI. What Indicators Say About Compliance with Human Rights: Where the Real Problem Lies

At the evaluation level, we assess what a given indicator says about compliance with human rights in a given jurisdiction and ask how indicators have been used to measure human rights compliance and to heighten accountability. Merry argues that a “[p]roliferation of indicators . . . springs . . . from the desire for accountability.”[111] Accepting that accountability is desirable, we must therefore ask whether the non-quantitative human rights mechanisms are any better as monitoring systems.
The first question to consider is whether it is possible to monitor or track human rights through other means. The answer is obviously yes, as it was traditionally done by the different human rights international mechanisms, such as using envoys, commissioners, general reporting, fact finding missions, voluntary transmission of comprehensive reports, sampling, and media tracking. We argue that the issue is not the lack of alternatives for measuring and quantifying a situation, but a cost-benefit analysis of these options in terms of the adequacy of the portrait presented, fairness, costs, accessibility of data, and comparability. This analysis is missing in Merry’s book.
The situation of the human rights mechanisms at the turn of the century described by James Crawford is a perfect illustration of previous attempts to assess human rights compliance. In a well-known text from 2000, Crawford offered a detailed list of problems with the U.N. human rights system, posing to readers the question of whether the system was in crisis. The problems Crawford listed include “corrosive effects” of the backlog in state reporting, linked to delays in present-day reporting, such as a delay between the submission of a report and when it is finally considered;[112] resource constraints in terms of personnel, and financial and technological capacities faced by Committees;[113] procedural issues, such as lack of follow-up mechanisms for periodic reports and communications;[114] substantive concerns about the composition of committees;[115] and limited political support from States.[116] Many argue that the introduction of the Universal Periodic Review (UPR) did not adequately address these concerns,[117] although UPR has somewhat heightened accountability and recognition of the universality of human rights.[118]
The first and more essential dilemma Merry raises is the problematic nature of judging compliance based on information provided by the organization being judged. She goes on to highlight the problem of “gaming” the indicator: the ways in which those who are being evaluated present data or change their approach purely to get a better ranking, with no substantial impact or change to the issue at hand. Merry argues, “Rather than revealing truth, indicators create it. However, the result is not simply a fiction but a particular way of dividing up and making known one reality among many possibilities.”[119] Data acts to suggest the existence of one particular truth, despite the shortcomings of the data collection itself; it is, therefore, a misleading “truth.” In addition, it then influences development and policy planning.[120]
Merry’s concerns are valid. Data shows only one perspective on a situation, and the concern is heightened because the data collector can present the data in a way that shows it in the best light. When it comes to human rights, however, surely we can shed some of our skepticism. We must recognize those States that genuinely wish to improve the well-being of their residents and who have a real interest in a system of data collection that would enable them to better understand the situations they are grappling with and how they can better invest their own resources to improve enjoyment of human rights. However, no one wants to be seen as a human rights violator, and it is true that some States may abuse quantitative indicators to diminish the problems reflected in their country statistics.
However, in these instances, qualitative data proves similarly problematic when it comes to unsupportive States. Those data collectors who are more willing to honestly participate in a qualitative process to monitor their performance are more likely to produce more reliable quantifiable data as well. The difference can probably be seen with those States who are reluctant participants in qualitative evaluation processes of their human rights situations. In a given year, one can mislead others in qualitative and quantitative terms, but when the exercise is repeated year after year, it becomes more difficult to coherently produce false quantitative data.
A separate issue Merry raises—and perhaps a less explored one in the literature on indicators—is the potential dependency of the United Nations on funding by the States subject to evaluation.[121] Yet, one could argue that there are many examples of countries threatening to withdraw funding for the publication of facts or decisions by U.N. bodies that have nothing to do with rankings or quantifications. One can think for example of the finally withdrawn report accusing Israel of imposing an apartheid regime in the Occupied Territories,[122] or the never-published report on the killing and maiming of children in warfare in Yemen by Saudi Arabia.[123] In fact, a more accurate view could be the exact opposite situation. In general, countries are more willing to be subjected to global exercises where there is some naming and shaming for everybody than particular investigations into any one pinpointed domestic context. The very existence of the Universal Periodic Review (UPR)—and the debates leading to its creation—would support this view.[124]

VII. Attaching Consequences to Indicators

Many of the criticisms Merry highlights are not particular to indicators but relate to the international system we have created to protect human rights. Perhaps with the exception of some regional systems—namely ECHR and IACHR—treaty bodies using indicators lack penalties or rewards for good or bad performance beyond issuing public comments of praise or critique.[125] The same could be said for those that do not use indicators. The U.N. Trafficking Protocol for instance was enacted in Palermo in 2000,[126] and having defined the concept of trafficking,[127] was arguably the origin of the U.S. TIP report.[128] The Protocol itself suffers from a weak implementation mechanism, operating under a “loose” oversight of a working group of States Parties attached to the broader Conference on Parties to the U.N. Convention against Transnational Organized Crime.[129] In other words, the lack of consequences for non-compliance is a problem stemming from the broader system of international human rights law, or international law, and cannot be solely attributed to indicators.
It is true as well that efforts to promote indicators of human rights compliance face resistance from countries concerned about mechanisms that compare and rank them on the basis of their compliance.[130] And yet, this is no different from those reporting mechanisms that do not use indicators or collect aggregated data. For example, country-specific reports on their fulfilment of conventions against torture often remained unpublished due to the reluctance of those Governments to accept them and make them public.[131]
To the contrary, indicators can offer individual states a baseline. Current performance can be checked against previous performance, based on general agreed-upon standards. This would allow countries to be judged only against themselves, eliminating the common claim that developing countries are judged against developed countries’ standards. Rewarding improvements and punishing deteriorations would then seem fairer than attaching those consequences to reporting on concrete incidents that may not be comparable. This could also be a way of implementing the evolving and improving nature of human rights,[132] making it obvious if a country goes backwards in its protection of human rights.
All of this is a different scenario from the one Merry describes on the consequences that the U.S. government attaches to its TIP rankings. In fact, the U.S. government—like all governments—is entitled to spend its donor aid as it considers appropriate.[133] If the government feels more confident to do this based on a report that it has produced itself, there is no legitimate reason to stand in the way of such an approach. Any country could do the same, basing its decisions on quantitative or qualitative appraisals conducted by them or on their behalf. All this proves that this issue is not linked to quantification mechanisms or rankings, but rather the gathering of data in order to make certain decisions which are political in nature. It is the politics behind the decision that should be the subject of the critique.
It is also important to note that governments and fellow States Parties are far from the only bodies that use data as the basis for establishing consequences for other entities. The media is also a massive consumer of data. In this respect, it is important to distinguish between how governments or the United Nations, for instance, may attach consequences to a particular system of ranking or performance based on quantitative data and the way in which the media may consume and report on such data. Merry fails to adequately recognize such a distinction.
For example, when she discusses the quantification of corruption, she refers to the use of “expert opinion surveys,” correctly arguing that collecting data is expensive and expert opinion surveys are a cheaper alternative.[134] In this case, she questions the idea of sending questionnaires to local experts to gather their experiences of corruption, arguing that while this is clearly less expensive, it is “also less comprehensive and accurate.”[135] However, Merry does not accurately represent this example. The index she refers to measures “perception of corruption” and not “corruption.”[136] The methodology is designed to survey people on their perception of corruption and Transparency International (TI), the entity which created that index, presents the results as such. In fact, TI has an alternative index—the Global Corruption Barometer—which surveys people about their corruption experiences in the last twelve months,[137] trying to directly assess the prevalence of corruption in a given jurisdiction. This allows for data on corruption prevalence to exist alongside data on perceptions of the levels of corruption.
The real issue here is how the media—and particularly non-expert media—reports on data developed from an index versus what the actual index is purporting to measure. This problem would be magnified if the reporting of those indices and quantification formed the basis of policy decisions or was used to apply the “evidence-based governance” that Merry describes at the beginning of her book.[138] Yet, Merry fails to provide an example of such a problem in practice.

VIII. CONCLUSION

As a general observation, quantification is not intrinsically better or worse than other systems of accountability in assessing compliance with human rights, but it does require a more robust methodology than other mechanisms to be useful. Here, the term “useful” denotes both a process and an end result that is a fair representation of a body of data and where that data can be presented in a manageable format, allowing consumers of that data to assess a situation at a particular moment in time but also changes over time. Such rigor may not be as demanding for some non-quantitative approaches.
Yet, there are some obvious advantages of quantitative measurement. Merry is correct that it cannot all be quantitative, but this does not justify a widespread questioning of the substantial interest in the quantification of human rights indicators that exists today. There are bad apples and good apples, including within Merry’s examples. In the same way that the TIP report is an example of a bad apple, the outputs of the UNECE Project,[139] the UNSC Project,[140] and the work of Yakin Ertürk[141] in the context of violence against women are examples of good apples in terms of how they have defined the issue to be measured and against which governments and other actors will be held to account. The very process of developing such indicators involved an open and honest debate about what would get measured as part of a reasonably global effort towards heightened accountability in relation to violence suffered by women.[142]
As a second conclusion, Merry is also correct in stating that, “some human rights are more amenable to quantification than others,”[143] but this does not mean that quantification should be excluded from monitoring compliance for certain rights. In those cases, the methodology should be defined to align as closely as possible with the specific right. When proxies are used, these choices should be transparent and any necessary simplification clearly understood by those collecting the data and later by those using it.
According to Merry, proxy measures, giving the example, among others, of using investment in prison and police reform to assess compliance with civil and political rights, “provide far greater specificity to treaty obligations but only focus on a few measures while neglecting many other dimensions of a given right.”[144] This is true, but needs to be compared with the alternative systems available to monitor those “given rights,” especially in terms of costs and benefits. A thorough and systematic reporting on the use of unnecessary force and torture by police officials would be a much better way to track compliance with the relevant articles of the Convention Against Torture, but it is unlikely to be feasible in the majority of developing countries, nor possible in the majority of countries globally on a systematic and regular basis, which is what is required for ongoing accountability.
In connection to this point, we disagree with Merry when she argues that “treaty bodies see their task as evaluating compliance with legal standards based on the data presented, not on checking the quality of the data themselves.”[145] This may be the case in some treaty monitoring bodies, but it is definitely not the case for some others. In fact, in some instances, treaty body concluding comments address very specifically the process of collecting information and contrast the data presented in States Parties’ reports with submissions and reports from NGOs, local and international, which allow members to challenge the data with concrete examples.[146] It is true that “[b]y using apparently objective, scientifically produced statistical data, treaty bodies can enhance their authority in rendering judgments about human rights,”[147] but as far as there has been an adequate check of the underlying data, we do not perceive this as a problem but rather as a legitimate option in light of the alternative—governments showing unwillingness to follow concluding comments or denying their legitimacy because such data is lacking.
Finally, Merry does not sufficiently explore the difference between quantification and how the media reports on organized data. At some point, it is necessary to distinguish between the use of quantification by experts in the field of international human rights law and its use by the general public. Therefore, concerns with what and how something is measured must be separated from concerns about how such data is interpreted and consumed by end users.
Indicators may appear objective from a distance, but when we look at them closely, it is easy to see their flaws. Merry’s book can be praised as an effort to make this clear to non-experts who come across her work. Regretfully, this point is well-known to specialized practitioners who value indicators for their practicality in helping to measure similarities and differences, advancement, and gaps in our progress towards development and well-being. Merry has evidently invested an enormous amount of work in analyzing the three indicator projects that are the subject of The Seductions of Quantification. As academics and practitioners in international law, we would have hoped for a deeper study of the trade-offs between the different alternatives (i.e., the non-quantitative ones) to assessing compliance with human rights and, potentially, a practical insight on more robust and useful ways to select and develop indicators.


[∗]José-Miguel Bello y Villarino is a member of the Diplomatic Corps of Spain (on leave as of this writing) and a PhD candidate at the University of Sydney, researching corruption and international law. Ramona Vijeyarasa is a Chancellor’s Post-Doctoral Research Fellow at the University of Technology Sydney. Her scholarship focuses on international women’s rights law, migration, trafficking, sexual and reproductive rights, and the role of women leaders in promoting gender-responsive legislation. The views expressed in this article are the authors’ views alone and may not represent any of the entities they are associated with. The authors are grateful to the editorial staff at the N.Y.U. Journal of International Law and Politics, especially Yekaterina Fomitcheva for her useful substantive suggestions. . As always, all errors are our own.
[1] See generally, e.g., Gauthier de Beco, Human Rights Indicators for Assessing State Compliance with International Human Rights, 77 NORDIC J. INT. LAW 23 (2008); Maria Green, What We Talk About When We Talk About Indicators: Current Approaches to Human Rights Measurement, 23 HUM. RTS. Q. 1062 (2001); Angela Hawken & Gerardo L. Munck, Cross-National Indices with Gender-Differentiated Data: What Do They Measure? How Valid Are They?, 111 SOC. INDICATORS RES. 801 (2013); Johannes P. Jütting et al., Measuring Gender (In)Equality: The OECD Gender, Institutions and Development Data Base, 9 J. HUM. DEV. 65 (2008); Debra J. Liebowitz & Susanne Zwingel, Gender Equality Oversimplified: Using CEDAW to Counter the Measurement Obsession, 16 INT'L STUD. REV. 362 (2014); AnnJanette Rosga & Margaret L. Satterthwaite, The Trust in Indicators: Measuring Human Rights, 27 BERKELEY J. INT'L L. 253 (2009); Irene van Staveren, To Measure is to Know? A Comparative Analysis of Gender Indices, 71 REV. SOC. ECON. 339 (2013); Boris Branisa et al., New Measures of Gender Inequality: The Social Institutions and Gender Index (SIGI) and Its Subindices (Courant Res. Ctr., Discussion Paper No. 10, 2009), http://EconPapers.repec.org/RePEc:got:gotcrc:010;
[2] ANNE F. BAYEFSKY, THE U.N. HUMAN RIGHTS TREATY SYSTEM: UNIVERSALITY AT THE CROSSROADS 2 (1st ed. 2001).
[3] There are nine treaties but ten treaty bodies. The Convention against Torture and Other Cruel, Inhuman or Degrading Treatment or Punishment (Dec. 10, 1984, 1465 U.N.T.S. 85) has its own treaty body , as well as the Subcommittee on Prevention of Torture and other Cruel, Inhuman or Degrading Treatment or Punishment (SPT) established pursuant to the Optional Protocol to the Convention against Torture and Other Cruel, Inhuman or Degrading Treatment or Punishment (Dec. 18, 2010, 42 ILM 26).
[4] We prefer this term to “indicator culture,” used by Merry, as we believe that there is an inclination to favor information supported or presented as indicators or rankings, but not necessarily to conduct human rights work through indicators. Cf. SALLY ENGLE MERRY, THE SEDUCTIONS OF QUANTIFICATION: MEASURING HUMAN RIGHTS, GENDER VIOLENCE, AND SEX TRAFFICKING 9, 220 (2016).
[5] Id. at 20.
[6] Id. at 21.
[7] Id. at 3.
[8] Id. at 31.
[9] See Sally Engle Merry, Measuring the World: Indicators, Human Rights, and Global Governance, 52 CURRENT ANTHROPOLOGY SUPPLEMENT 3, S83, S84(2011); see also THE QUIET POWER OF INDICATORS: MEASURING GOVERNANCE, CORRUPTION, AND RULE OF LAW20 (Sally Engle Merry, Kevin E. Davis, & Benedict Kingsbury eds., 2015) discussing the issue in wider terms beyond human rights.
[10] See MERRY, supra note 4, at 16–19 (offering a few examples of “successful indicators” as a simplified overview).
[11] A brief overview of the contents of the discussions in this process can be found in UN STATISTICAL COMMISSION, REPORT ON THE MEETING OF THE FRIENDS OF THE CHAIR OF THE UNITED NATIONS STATISTICAL COMMISSION ON STATISTICAL INDICATORS ON VIOLENCE AGAINST WOMEN 26 (2010), https://unstats.un.org/unsd/demographic/meetings/vaw/docs/finalreport.pdf.
[12] MERRY, supra note 4, at 107.
[13] Id. at 107.
[14] See id. at 5 (“Rather than revealing truth, indicators create it.”).
[15] Id. at 111.
[16] DEPARTMENT OF STATE, TRAFFICKING IN PERSONS REPORT (2017), https://www.state.gov/documents/organization/271339.pdf.
[17] DEPARTMENT OF STATE, TRAFFICKING IN PERSONS REPORT 25 (2004), https://www.state.gov/documents/organization/34158.pdf.
[18] DEPARTMENT OF STATE, supra note 16 at 29.
[19] Id. at 114. For a detailed discussion on the issue of politicization see generally Kirstin Roster, Politicization of the Trafficking in Persons Report; Is Political Proximity to the US Associated with Better Rankings? (McCourt Sch. Pub. Pol'y, 2016), https://repository.library.georgetown.edu/handle/10822/1040854.
[20] MERRY, supra note 4 at 194–206.
[21] Comm. on the Elimination of Discrimination Against Women, General Recommendation No. 35 on Gender-Based Violence Against Women, Updating General Recommendation No. 19, U.N. Doc. CEDAW /C/GC/35 2 (2017).
[22] The ten human rights treaty bodies are the Committee on Civil and Political Rights; the Committee on Economic, Social and Cultural Rights; the Committee on the Elimination of Discrimination against Women; the Committee on the Elimination of Racial Discrimination; the Committee on the Rights of the Child; the Committee on the Rights of Persons with Disabilities; the Committee against Torture; the Committee on Migrant Workers; and the Sub-committee on the Prevention of Torture and the Committee on Enforced Disappearance. Cf. supra note 3.
[23] See Comm. on the Elimination of Discrimination Against Women, supra note 21; see also Ramona Vijeyarasa, A Quarter of a Century Later: Violence Against Women and CEDAW’s General Recommendation (forthcoming 2018) (manuscript at 35) (on file with authors).
[24] U.N. High Comm'r for Hum. Rts., Human Rights Indicators: A Guide to Measurement and Implementation, at 95, U.N. Doc. HR/Pub/12/5 (2012).
[25] Id. at 99.
[26] MERRY, supra note 4, at 13.
[27] Id. at 1.
[28] Malhotra & Fasel, supra note 1, at 3.
[29] Merry et al. describe four phases in the life of an indicator: conceptualization, production, uses and effects, and impacts, supra note 9 at 10.
[30] MERRY, supra note 4, at 168 (citing Russel Lawrence Barsh, Measuring Human Rights: Problems of Methodology and Purpose, 15 HUM. RTS. Q. 87 (1993), and Todd Landman, Measuring Human Rights: Principle, Practice, and Policy, 26 HUM. RTS. Q. 906 (2004)). Merry’s point is better explained in her own words. See MERRY, supra note 4, at 215.
[31] MERRY, supra note 4 at 1, 19–21.
[32] MERRY, supra note 4, at 4.
[33] Id. at 180.
[34] Id. at 204.
[35] See generally Andrew Moravcsik, The Origins of Human Rights Regimes: Democratic Delegation in Postwar Europe, 54 INT. ORG. 217 (2000) (offering an account of the origins of the institutionalization of human rights regimes in the European context).
[36] For example, the link between access to information, freedom to manifest political ideas and the right to take part in the conduct of public affairs is a recurrent reasoning of the Human Rights Committee, often referring to these concepts as closely connected. See UN HUMAN RIGHTS COMMITTEE (HRC), CCPR GENERAL COMMENT NO. 25: ARTICLE 25 (PARTICIPATION IN PUBLIC AFFAIRS AND THE RIGHT TO VOTE), THE RIGHT TO PARTICIPATE IN PUBLIC AFFAIRS, VOTING RIGHTS AND THE RIGHT OF EQUAL ACCESS TO PUBLIC SERVICE 8 (1996); UN HUMAN RIGHTS COMMITTEE (HRC), CCPR GENERAL COMMENT NO. 34 ARTICLE 19: FREEDOMS OF OPINION AND EXPRESSION 13 (2011), http://www2.ohchr.org/english/bodies/hrc/docs/gc34.pdf.
[37] MERRY, supra note 4 at 29 (quoting MICHEL FOUCAULT, DISCIPLINE AND PUNISH: THE BIRTH OF THE PRISON 27 (1979)).

[38] Id. at 77.
[39] Id. at 44.
[40] Id.
[41] Convention on the Elimination of All Forms of Discrimination Against Woman, Dec. 18, 1979, 1249 U.NT.S. 13 (entered into force Sept. 3, 1981).
[42] See U.N. Entity for Gender Equality and the Empowerment of Women, Rep. of the Fourth World Conf. on Women, U.N. Doc. A/CONF. 177/20/Add.1 (1995); see also U.N. Entity for Gender Equality and the Empowerment of Women, Beijing Declaration and Platform of Action, Fourth World Conf. on Women (1995), http://www.un.org/womenwatch/daw/beijing/platform/.
[43] Comm. on the Elimination of Discrimination Against Women [CEDAW], General Recommendation 12 (1989), http://www.un.org/womenwatch/daw/cedaw/recommendations/recomm.htm#recom12.
[44] Comm. on the Elimination of Discrimination Against Women [CEDAW], General Recommendation 19 (1992), http://www.un.org/womenwatch/daw/cedaw/recommendations/recomm.htm#recom19.
[45] Ruth F. Lax, Socially sanctioned violence against women: Female genital mutilation is its most brutal form, 28 CLINICAL SOCIAL WORK JOURNAL; NEW YORK 403–412, 502 (2000) .
[46] MERRY, supra note 4, at 46.
[47] Id. at 62.
[48] Id.
[49] Id. at 82.
[50] Id. at 82.
[51] Id. at 73.
[52] Id. at 90.
[53] Id.
[54] Id. at 85.
[55] See,U.S. DEP’T OF STATE, supra note 16.
[56] Id. at 20.
[57] RAMONA VIJEYARASA, SEX, SLAVERY AND THE TRAFFICKED WOMAN: MYTHS AND MISCONCEPTIONS ABOUT TRAFFICKING AND ITS VICTIMS (2015).
[58]
Id. at 141-144.
[59] MERRY, supra note 4, at 74.
[60] Id. at 73.
[61] Id. at 6.
[62] Id. at 71.
[63] Id.
[64] Id. at 25, 216.
[65] Id. at 66.
[66] Id. at 88.
[67] Id.
[68] Id. at 64.
[69] Id.
[70] Id. at 21.
[71] Id. at 77-78.
[72] Id. at 212.
[73] G.A. Res. 217 (III) A, pmbl. & art. 2, Universal Declaration of Human Rights (Dec. 10, 1948).
[74] Declarations, Reservations and Objections to CEDAW, U.N. WOMEN, http://www.un.org/womenwatch/daw/cedaw/reservations-country.htm (last visited Mar. 17, 2018); see Judith Resnik, Comparative (In)equalities: CEDAW, the Jurisdiction of Gender, and the Heterogeneity of Transnational Law Production, 10 INT'L J. CONST. L. 531, Footnote 78 (2012).
[75] See Convention on the Elimination of All Forms of Discrimination Against Women: Status, UNITED NATIONS TREATY COLLECTION, https://treaties.un.org/Pages/ViewDetails.aspx?src=IND&mtdsg_no=IV-8&chapter=4&lang=en (last updated Mar. 31, 2018) (detailing ratification status).
[76] MERRY, supra note 4, at 77..
[77] G.A. Res. 55/2, 20, United Nations Millennium Declaration (Sept. 18, 2000); see also Promote Gender Equality and Empower Women, UNITED NATIONS DEV. PROGRAMME, http://www.undp.org/content/undp/en/home/sdgoverview/mdg_goals/mdg3 (last visited Mar. 17, 2018).
[78] Ceri Hayes, Out of the Margins: The MDGs Through a CEDAW Lens, 13 GEND. & DEV. 67, 68 (2005).
[79] UNITED NATIONS & 2015 TIME FOR GLOB. ACTION, THE MILLENNIUM GOALS DEVELOPMENT REPORT 2015, 28 (2015), http://www.undp.org/content/dam/undp/library/MDG/english/UNDP_MDG_Report_2015.pdf.
[80] See generally SUZANNE DURYEA ET AL., THE EDUCATIONAL GENDER GAP IN LATIN AMERICA AND THE CARIBBEAN (Social Science Research Network) (2007), https://papers.ssrn.com/abstract=1820870 (last visited Apr 10, 2018); see also HAYES, supra note 78, at 68.
[81] HAYES, supra note 78, at 68..
[82] MERRY, supra note 4, at 77.
[83] Id. at 77–78.
[84] Id. at 78.
[85] Liebowitz & Zwingel, supra note 1, at 375. The other measure of civil liberties under the Social Institutions and Gender Index was the freedom of women to move outside of the home. See generally The OECD Social Institutions and Gender Index, ORG. FOR ECON. CO-OPERATION & DEV. [OECD], http://www.oecd.org/dev/development-gender/theoecdsocialinstitutionsandgenderindex.htm#results (last visited Mar. 17, 2018).
[86] See Ramona Vijeyarasa, In Defence of CEDAW: The Need to Return to the Foundations of Women’s Rights Law to Measure Progress on Gender Equality (forthcoming 2018) (manuscript at 19) (on file with authors).
[87] Id. Ramona Vijeyarasa proposes a quantitative accountability system, grounded in CEDAW, where each article is translated into a measurable indicator with direct domestic relevance.
[88] MERRY, supra note 43, at 62.
[89] See supra, Introduction.
[90] VIJEYARASA, supra note 58, at 27.

[91] See Laura Agustín, TIP: Trafficking in Persons, the No Methodology Report, NAKED ANTHROPOLOGIST (June 26, 2009), https://www.lauraagustin.com/tip-trafficking-in-persons-the-no-methodology-report.

[92] VIJEYARASA, supra note 58, at 8.
[93] MERRY, supra note 4, at 138.
[94] See for example Liz Kelly, “You Can Find Anything You Want”: A Critical Reflection on Research on Trafficking in Persons within and into Europe, 43 INTERNATIONAL MIGRATION 235–265, 254 (2005),
[95] MERRY, supra note 4 114.
[96] To her credit, Merry makes a valid argument that the measurement of human trafficking has proven particularly elusive. Trafficking is indeed poorly understood and not well defined. These factors make it harder to successfully develop a system of measurement. MERRY, supra note 3, at 129.
[97] VIJEYARASA, supra note 58, at 19–39. While many stakeholders have a legitimate interest in countering trafficking, we have also seen how human trafficking has been presented in a sensationalized or biased light to obtain particular gains. This spans a host of agendas, from a desire for abolition of prostitution to the financial and readership goals of profit-driven media to government goals around regulated or closed borders.
[98] VIJEYARASA, supra note 58, at 8.
[99] U.S. DEP’T OF STATE, TRAFFICKING IN PERSONS REPORT: JUNE 2012, 53 (2012), https://www.state.gov/documents/organization/192587.pdf.
[100] JUST THE “TIP” OF THE ICEBERG: THE 2011 TRAFFICKING IN PERSONS REPORT (TIP) FALLS SHORT OF EXPECTATIONS HUFFINGTON POST, https://www.huffingtonpost.com/pardis-mahdavi/just-the-tip-of-the-icebe_1_b_888618.html (last visited Apr 10, 2018); see generally Roster, supra note 19 on the politicization of global rankings).
[101] MERRY, supra note 4, at 149. Ramona Vijeyarasa has elsewhere noted the role that the U.S. government has played as a donor in distorting global priorities when it comes to aid and anti-trafficking programs. VIJEYARASA, supra note 58, at 27–28.
[102] MERRY, supra note 4, at 138.
[103] Id. at 125.
[104] VIJEYARASA, supra note 58, at xvii.
[105] Id. at 19.
[106] Id. at 20.
[107] MERRY, supra note 4, at 16.
[108] Id. at 19.
[109] Alina Mungiu-Pippidi & Ramin Dadašov, Measuring Control of Corruption by a New Index of Public Integrity, 22 EUR. J. CRIM. POL'Y & RES. 369, 416, 417 (2016).
[110] Transparency International, CPI 2017 TECHNICAL METHODOLOGY NOTE 1 (2017), http://files.transparency.org/content/download/2183/13748/file/CPI_2017_Technical%20Methodology%20Note_EN.pdf.
[111] MERRY, supra note 4, at 3.
[112] James Crawford, The UN Human Right Treaty System: A System in Crisis?, in THE FUTURE OF UN HUMAN RIGHTS TREATY MONITORING , 4–5 (2000).
[113] Id. at 6–7.
[114] Id. at 7–8.
[115] Id. at 9.
[116] Id. at 10.
[117] See generally Allehone Mulugeta Abebe, Of Shaming and Bargaining: African States and the Universal Periodic Review of the United Nations Human Rights Council, 9 HUM. RTS. L. REV. 1, 19-30 (2009); Felice D. Gaer, A Voice Not an Echo: Universal Periodic Review and the UN Treaty Body System, 7 HUM. RTS. L. REV. 109, 135-139 (2007).
[118] Elvira Domínguez-Redondo, The Universal Periodic Review—Is There Life Beyond Naming and Shaming in Human Rights Implementation?, 4 N. Z. L. REV. 67, 694-705 (2012); Rhona Smith, “To See Themselves as Others See Them”: The Five Permanent Members of the Security Council and the Human Rights Council’s Universal Periodic Review, 35 HUM. RTS. Q. 1, 22-26 (2013).
[119] MERRY, supra note 4, at 5.
[120] Id. at 5.
[121] Id. at 168.
[122] See Raoul Wootliff, End Bias or Else, Israel Threatens UN Ahead of Chief’s Visit, TIMES ISR. (Aug. 27, 2017), https://www.timesofisrael.com/end-bias-or-else-israel-threatens-un-ahead-of-chiefs-visit/.
[123] Colum Lynch, Saudi Arabia Threatened to Break Relations With U.N. over Human Rights Criticism in Yemen, FOREIGN POL'Y (June 7, 2015), https://foreignpolicy.com/2016/06/07/saudi-arabia-threatened-to-break-relations-with-un-over-human-rights-criticism-in-yemen/.
[124] This point is expressly made in the Preamble of the Resolution adopted by the General Assembly which created the Human Rights and the UPR as “universality” and “non-selectivity”, as a recognition of the focus on a few situations and countries in the work of the Commission on Human Rights, the predecessor of the Human Rights Council. G.A. Res. 60/251, Preamble, U.N. Doc. A/RES/60/251 (Apr. 3, 2006), https://undocs.org/en/A/RES/60/251
[125] MERRY, supra note 4, at 167.
[126] Protocol to Prevent, Suppress and Punish Trafficking in Persons, Especially Women and Children, Supplementing the United Nations Convention Against Transnational Organized Crime, opened for signature Nov. 15, 2000, 2237 U.N.T.S. 319.
[127] Id. art. 3(a).
[128] DEPARTMENT OF STATE, TRAFFICKING IN PERSONS REPORT 8 (2001), https://www.state.gov/documents/organization/4107.pdf.
[129] Anne T. Gallagher, Two Cheers for the Trafficking Protocol, 4 ANTI-TRAFFICKING REV. 14, 22 (2015).
[130] MERRY, supra note 4, at 167.
[131] For example, the Assembly of the Council of Europe has recently called on the European Committee for the Prevention of Torture and Inhuman or Degrading Treatment or Punishment (CPT) to make public any unpublished reports concerning its member States. Eur. Consult. Ass., Recommendation 2123 (2018) Provisional Version: Strengthening International Regulations against Trade in Goods used for Torture and the Death Penalty, Doc. No. 14454, art. 10(6) (2018).
[132] See generally GENDER, NUTRITION, AND THE HUMAN RIGHT TO ADEQUATE FOOD: TOWARD AN INCLUSIVE FRAMEWORK (Anne C. Bellows et al. eds., 2016) (arguing for a more expansive notion of human rights encompassing an affirmative right to adequate food).
[133] For example the U.S. funds abolitionists groups, but not NGOs favouring the legalization of sex work. MERRY, supra note 4, at 153.
[134] MERRY, supra note 4, at 7.
[135] Id.
[136] Corruption Perceptions Index 2016, TRANSPARENCY INT'L (Jan. 2017), https://www.transparency.org/news/feature/corruption_perceptions_index_2016.
[137] People and Corruption: Citizens’ Voices from Around the World, TRANSPARENCY INT'L (Nov. 14, 2017), https://www.transparency.org/_view/publication/8064.
[138] MERRY, supra note 4, at 10.
[139] Id. at 65.
[140] Id. at 54.
[141] Id. at 63, 64.
[142] Id. at 63-66. Merry’s analysis sheds positive light on the process of indicator development under the UN bodies for European (UNECE) initiative, one that “relief on social science scholars with extensive knowledge of gender-based violence and survey techniques developed for measuring it,” was clearly “concerned with women who experience violence” and adopted “a more feminist political stance and an interest in maximizing disclosure” at 67.
[143] Id. at 172.
[144] Id. at 171.
[145] Id. at 170.
[146] Cossette D. Creamer & Beth A. Simmons, Ratification, Reporting, and Rights: Quality of Participation in the Convention Against Torture, 37 HUM. RTS. Q. 579, 585 (2015).
[147] MERRY, supra note 4, at 170.


AustLII: Copyright Policy | Disclaimers | Privacy Policy | Feedback
URL: http://www.austlii.edu.au/au/journals/UTSLRS/2018/10.html