Non-Existent Humans: How Stateless Persons are Forced to Work Informally

 . 8 MIN READ

Opening a bank account, obtaining a driver’s license, or even getting married—while many of us may take these activities for granted, they are impossible for people without a formal nationality, or so-called stateless persons. Yet, the lack of a bank account is just a minor detail compared to the more severe consequences of being stateless, particularly a lack of access to healthcare, education, or employment. As Isa, a stateless person from the former Yugoslavia, explains, “To be without documents and a nationality is as if you never existed in this world.” Together with at least 10 million stateless people across the globe, Isa belongs to the most vulnerable group of humans, who are often even described as “legal ghosts.” Well-known examples of such demographic groups not having the nationality of any country include the Rohingya in Myanmar or Eastern Europeans after the collapse of the Soviet Union. In recent years, public awareness about the issue of statelessness has increased due to increased efforts by the UN and many international non-governmental organizations. In fact, the UN has even launched a campaign to end statelessness by 2024, identifying flawed administrative procedures, which can be fixed rather quickly in theory, as the main cause of statelessness.

However, this campaign is unlikely to be successful because the public discourse so far has forgotten about the connection between statelessness and informal employment. As part of globalization, what Professor Nina Glick Schiller calls a new “global labor regime” has emerged, in which transnational companies looking for inexpensive labor will incentivize workers to move from their less-developed home countries to more prosperous nations. While the workers, which include both stateless persons and their documented counterparts, undoubtedly hope for a better future, they soon encounter restrictive and complicated administrative procedures in their new country that effectively prevent them from applying for citizenship. With few legal protections, stateless labor migrants in particular then become a pawn sacrifice of the “global labor regime.” This strategy of capitalizing on human defenselessness renders informal labor possible at the expense of stateless persons, particularly in newly industrialized countries in Southeast Asia like Malaysia. Therefore, merely blaming flawed administrative procedures will be insufficient to properly address the issue of statelessness. More creative and comprehensive solutions such as a “transnational labor citizenship,” a type of proxy identity provided through membership in a transnational civil society organization, are needed so that the seeming contradiction of non-existent humans does not become reality for even more people.

As UN High Commissioner for Refugees, António Guterres was leading in the start of the UN campaign to end statelessness by 2024. Photo by Eric Bridiers, CC BY-ND 2.0, accessed via Flickr.

How Statelessness and Informal Employment Connect

To understand why the “global labor regime” is even effective at incentivizing low-skilled workers to migrate, it is important to discuss the precarious living situations that many migrants face at home. On the one hand, such insecure conditions are caused by the subsistence-level living standards found in many parts of the world, where low-skilled workers typically receive low wages and have few opportunities to enhance their skills. On the other hand, people’s desire to accumulate previously unimaginable amounts of wealth has crushed the labor and social rights that had been won throughout the 19th and 20th centuries. This capitalist centralization of power and wealth has removed the governmental frameworks and social institutions that once prevented labor exploitation and has served to reinstitute the kind of dispossession that most people have only heard of in the context of the Middle Ages. As a result, many low-skilled workers can only dream about the kind of financial security, labor protection, and long-term security that citizens in industrialized countries have come to take for granted. This phenomenon is well summarized by three scholars from the International Institute of Social Studies, who recently showed how “foreign employment demands to be understood as precarious work undertaken to mitigate existing conditions of precarity at home.” The desire to migrate appears quite natural given the perilous living situations many low-skilled workers find themselves in.

It is exactly these push factors, which are driving people with low levels of education out of their home country, that the new “global labor regime” profits off of in order to meet the increased demand of low-wage labor caused by globalization. As the global market increasingly demands faster production of consumer goods at lower prices, large companies can only stay competitive internationally by reducing workers’ wages. Additionally, the significant demographic changes in many industrialized nations have shrunk these countries’ populations and domestic labor supplies, forcing them to increasingly depend on low-wage migrant workers. Yet, national governments and large companies have to actively cooperate on an informal level to ensure that these migrant workers can consistently be employed at a comparably low price. This is where the idea of a “global labor regime” comes in. Scholars have coined this term to describe the mutual effort of governments and companies to exclude particularly stateless labor migrants from access to citizenship. Built on nationalism and seemingly justified by a defense of the modern welfare state, this “global labor regime” systemically keeps migrants without a nationality in statelessness. For instance, governments in East Asia intentionally refuse to adopt more inclusive administrative procedures that would more easily grant stateless migrant workers from Southeast Asian countries citizenship rights as doing so would endanger economic growth. In short, national governments purposely accept the statelessness of migrants so that their economic sector can more easily exploit these people via low-wage labor. This practice is deeply troubling from a human rights perspective because it effectively dismisses the humanity of a nation’s stateless migrant workforce.

Stateless migrant workers are frequently exploited via low-wage labor, such as in the agricultural sector. Photo by Lance Cheung, Public Domain, accessed via Flickr. 

How Stateless Workers Suffer

This “global labor regime” fundamentally intertwines statelessness and informal employment, creating devastating consequences for the stateless workers who have become the regime’s pawn sacrifice. Without some type of government-issued identification, which thereby systematically excludes them from the formal labor market, stateless people are forced into chronic economic instability. For instance, they are denied access to almost all public-sector jobs within almost all countries since employment at federal or state agencies is generally restricted to citizens. As these public-sector jobs often present a key avenue to getting paid employment, especially for women in Southeast Asia, victims of the “global labor regime” are compelled to carry out cheaper, more dangerous under-the-table work even after many years in their new country. Yet, while ordinary citizens might be able to escape the low-wage sector, stateless persons are rarely able to find paths toward upward mobility. Given that the main mode of acquiring citizenship by birth in almost all Asian countries is by ius sanguinis (based on the nationality of one or both parents), generation after generation is ultimately trapped in a hopeless combination of poverty and obscurity. And, as if this is not enough already, most labor laws do not apply to workers without an official nationality, opening the door for employers to exploit, sexually harass, or even abuse their stateless employees. This is not to mention excessive working hours without overtime pay or the substandard working conditions that have almost become the new normal to many labor migrants. Unfortunately, these are the heavy burdens stateless workers have to bear at their workplace. Within their private life, stateless people similarly have to suffer from countless other examples of marginalization and discrimination.

A particularly heartbreaking example of the type of marginalization that stateless workers are confronted with can be found in Sabah, a Malaysian state in the northern part of the island Borneo. For decades, Sabah’s most crucial economic branches, such as rural oil palm plantations, have massively depended on migrant labor. Yet, as a professor at Hasanuddin University in Indonesia has described, a system of “mutual illegal collaboration” between local politicians, the police, and employers has prevented any provisions that would have granted stateless migrant workers the right to permanent employment. Instead, Sabah’s labor regime has ensured that foreign stateless workers are chained to one particular employer. The victims are people such as Rozita, whose father once came to Sabah as a refugee from the Philippines. Due to a series of family moves, Rozita lost her birth certificate and a copy of her father’s passport, which effectively made her stateless as she has been unable to prove her nationality ever since. By now, Rozita has four children who all lack birth certificates because of their mother’s statelessness. The family is effectively trapped in a vicious cycle of migration, statelessness, and informal employment. As Rozita says herself: “I don’t know how to get documents for my children. I want to go to Immigration or the Police to ask them, but I am scared.” Rozita, like so many others around the world, has become a victim of the new “global labor regime.”

Stateless workers, and particularly their children, face desperate living conditions. Photo by Transformer18, CC BY 2.0, accessed via Flickr. 

How Solutions Do Exist

Although it might at first seem hard to solve this issue effectively, there is still hope for people like Rozita and her children. Most obviously, stateless workers could become naturalized citizens of the country they currently reside in. This measure would not only boost the worker’s income, but it would also indirectly benefit the host country’s economy through increased spending and demand, which would undermine many nations’ demand for exploitable, low-wage workers. In this sense, improving statelessness determination procedures or granting stateless workers unimpeded access to a nation’s labor market would also contradict the interests of many transnational companies. These are but a few of the reasons for why the current UN initiative, which sees flawed administrative procedures as the predominant cause of statelessness, will most likely not succeed. Instead, a proposal for a “transnational labor citizenship” put forth by Fordham University Professor Jennifer Gordon seems much more promising. Such citizenship would link the permission to enter a country in search of work to the membership in a “cross-border worker association [that would] allow migrants to carry benefits and services with them as they move.” Gordon’s idea has already gained support among numerous scholars who agree that this kind of “proxy citizenship” through membership within a civil society organization would be quite effective if labor organizations are empowered to develop a transnational capacity. For instance, labor movements in Indonesia as well as the Philippines have developed greater organizational capacity in recent years. A combination of political activism from below by civil society and continuing global initiatives from above by the UN could intensify this trend, possibly also spreading it to other Southeast Asian countries like Malaysia or Singapore. If continued efforts by the international community or domestic pressures empower these labor organizations to a sufficient degree, a “transnational labor citizenship” could actually defeat the “global labor regime.” In the end, Rozita and her children might still have a chance.

Although the right to a nationality has theoretically been enshrined within Article 15 of the Universal Declaration of Human Rights since 1948, stories like those of Rozita have shown us how statelessness and informal employment are still possible, particularly within newly industrialized countries in Asia. A new “global labor regime” driven by international competition for the lowest prices purposely denies stateless persons the right to citizenship to ensure that these stateless migrant workers can consistently be employed at low wages. The result could hardly be more catastrophic as humans become a pawn sacrifice that is continuously exploited, sexually harassed, and physically abused. People like Isa and Rozita become legal ghosts, non-existent, without any rights, privileges, or happiness. As the current world fails them, their sole remedy might be a transnational labor citizenship. The stakes could not be higher.

Cover photo by Jan Arendtsz, CC BY-ND 2.0, accessed via Flickr.

Ending Poverty in Sub-Saharan Africa with Online Education

 . 5 MIN READ

Education plays a critical role in ending poverty. For example, the winners of the Nobel Memorial Prize in Economic Sciences in 2019 Abhijit Banerjee and Esther Duflo found that schooling increases earnings relatively linearly, in that each additional year of primary education adds about eight percent to earnings for a child: over the standard six years of primary schooling, income would increase for a child by about 50 percent. These gains are especially important in sub-Saharan Africa, which has the highest out-of-school rates for all ages; 33 million of the 61 million global out-of-school children live in the region.

Despite the large benefits of widespread increased education, the logistics of education can be difficult. In an analysis of seven African countries, United Nations Children's Fund (UNICEF) found that on the average school day, a fourth to a half of the teachers are absent, which results in a wasted school day for the students who do attend and lowers the actual effects of going to school. Moreover, barriers in attending school, such as fees, traveling time, and child morbidity, prevent out-of-school rates from decreasing.

Online Education

However, these roadblocks are not insurmountable. Online education has been on the rise in sub-Saharan Africa due to it being a cost-effective alternative to formal education as well as a way to widen educational opportunity. Already, this transition is being felt in the higher education category. Africa’s public universities are wrestling with expanding capacity without sacrificing quality which has allowed online-based institutions such as Unicaf, a pan-African university, to assert itself in the education system. Unicaf provides every student with a tablet and downloadable, offline course materials. A similar program in India, the Indira Gandhi National Open University, educates over three million students. In light of these case studies, online education could serve as a powerful tool for poverty reduction.

Tackling Barriers to Education

The primary reason that online education is so popular is for the convenience advantages for students. Specifically for students in poverty, online education tackles a number of barriers including travel time, school cost, and labor schedule.

The most important determinant of primary school enrollment is proximity.

Travel time
In their book, Improving Primary Education in Developing Countries, the World Bank describes how the most important determinant of primary school enrollment is proximity.The most important determinant of primary school enrollment is proximity. This insight has also been corroborated by the Jameel Poverty Action Lab’s (JPAL) studies, which have shown that creating local schools resulted in very large gains in enrollment as many areas with low enrollment rates are remote and affected by conflict. At the moment, 27 percent of school-age children in sub-Saharan Africa live more than two kilometers away from the nearest school, and the World Bank estimates that if these distances were reduced to less than two kilometers, their enrollment rates would rise by 15 percentage points. Clearly, the virtual aspect of online education would remove this issue of travel time and locality.

School fees and other costs
As an alternative to tax-based funding of education, a number of sub-Saharan African countries implemented school fees. However, they paid insufficient attention to the effect of these fees on access to education for the poor, which led to decreased primary school enrollment, especially among the poorest and most vulnerable children. In fact, a randomized controlled trial by Innovations for Poverty Action showed that students whose school fees were removed were 55 percent more likely to complete secondary school. Similarly, JPAL revealed that even reducing other schooling costs such as by providing free school uniforms or school meals also had positive impacts on attendance. Comparing the costs of universities in sub-Saharan Africa, it is clear that online institutions clearly win this competition; Unicaf costs US$4,000 for a degree compared to the average of US$3,000 per year of traditional universities. This suggests that, at least at the college level, online education is a very viable way to minimize costs.

Labor Schedule
The UN Educational, Scientific, and Cultural Organization (UNESCO) reported that about a quarter of five to 14-year-olds in sub-Saharan Africa were engaged in child labor in 2004. These working children faced attendance disadvantages of about 30 percent to 67 percent. Moreover, because a majority of the child laborers are working on subsistence agricultural farming for the family, online classes would also facilitate the education of those who have to farm during the school day so they can take classes later in the evening. The UN Girls’ Education Initiative also describes how for young girls, sibling care is a major barrier to education. Considering the gender gaps between school completion—36 percent for girls and 42 percent for boys—in sub-Saharan Africa for lower secondary school, online education offers hope of leveling the playing field. By having the ability to enroll in online courses, young girls would be able to learn at their own pace and on their own hours rather than on a rigid school day schedule.

Concerns

Despite these benefits, there are a number of logistical concerns with the feasibility of extensive online education in sub-Saharan Africa.

Internet Infrastructure
One of the primary concerns is Internet access: a fundamental part of online education is the online part of it. Sub-Saharan Africa has the lowest Internet penetration of any region in the world at just 40 percent. Moreover, the gaps between Internet usage varied with income, up to a 29 percentage point difference between high- and low-income earners in certain countries. Considering these lower Internet penetration rates, online education would be less suited to tackle the poorest of the poor, but could benefit those on the higher end of the poverty spectrum. However, these rates could change soon. A number of companies are working to bring universal Internet access across the globe. For instance, the US Federal Communications Commission recently approved aerospace company SpaceX to launch more than 4,000 satellites into space as part of their universal satellite Internet program Starlink. A number of other companies such as Amazon and Facebook are also pursuing these types of satellite Internet plans.

Quality of Online Education
Moreover, the quality of online education is another concern. In a report last year about online education, researchers at George Mason University revealed that, though general learning outcomes are similar for online and in-person courses, at-risk populations such as those of lower socioeconomic background underperform in the online setting as compared to the traditional setting. They suggest that those who are the most disadvantaged may end up experiencing a lower quality of education, because traditional classes offer valuable instructor-student interaction whereas online education does not. Because schooling quality has significant impacts on later earnings, decreased quality of online education may diminish the effects that receiving an online education would have regarding social mobility.

Overlooked Primary and Secondary Education
Unfortunately, most online education in sub-Saharan Africa is at the tertiary education level, but only about 9 percent of the population even enrolls in tertiary education. Truly tackling the barriers of education would involve increasing digital education opportunities for lower levels of education because that is where the most impact is needed.

Key Takeaways

Ultimately, online education is an increasingly-utilized, but still under-used tool for tackling a lack of educational attendance in sub-Saharan Africa for its benefits of no travel time, cheaper costs, and flexible hours. However, concerns such as the existing digital infrastructure, quality of schooling, and targeted student population suggest that it alone is not enough to drive a welfare revolution in sub-Saharan Africa.

A critical tool for tackling inequality in the 21st century.

Online education may not end poverty, but it is a leap towards the right direction, a critical tool for tackling inequality in the 21st century.

Cutthroat Academia: Invisible Innovators

 . 7 MIN READ

For immigrants chasing the ever elusive American Dream, the phrase, “education is the greatest equalizer,” is repeated, then recited, until it becomes a promise. In academia and scientific research, fulfilled dreams abound. On the tip of the iceberg, the immigrant professors, just awarded tenure, or the immigrant scientists, perhaps now working on a COVID-19 vaccine in big pharma, are living proof of the American Dream. They are the lucky ones—sometimes the exceptional ones. Others become invisible workers driving scientific innovation and research; their stories are rarely told. This is too often the plight of the immigrant postdoctoral (postdoc) researcher. Placed into a system that tends to exploit immigrant labor, the unlucky ones go from blue-collar work to, in extreme cases, 100-hour work weeks inside clandestine labs. The American Dream is put on indefinite hold.

Indentured to Research

Academia is not a true meritocracy. At the golden tip of the hierarchy are the tenured professors, who hold the job for life. Choosing these professors, a tiny sliver of the doctorate-holding population, is an inherently political process. The demand for that coveted slot is what has fueled the relevance of the postdoc.

Officially, a postdoc position is marketed towards candidates who have completed graduate school. It allows them to receive invaluable training, additional research experience, and strong mentorship—everything that, on paper, seems to prepare graduates for tenure-track professorships. A vast majority of students believe they need to do at least one postdoc in order to be considered for tenure-track positions. In reality, it’s a job built on empty promises.

What the job really entails is far less rosy than advertised. Postdocs are often hired through scientific grants awarded to professors, so they are responsible for one thing: creating valuable output. There is a simple equation that determines job security—publications equals grants equals employment. The 40 hour work week becomes a myth, as output is measured by number of publications, not the hours spent on research. Salary is neither measured by time nor skill. The average annual salary hovers around US $40,000, with minimal employee benefits. Because most postdocs are in their 30s or 40s, it’s a salary that goes towards supporting a family. In contrast, full time researchers hired to do the same work are paid up to twice that amount, and the income gap becomes much larger for researchers in industry. It is clear that postdocs are often valued as a source of cheap labor. The justification for this is that a postdoctoral fellowship is not a job, but rather a temporary training period that serves as a down payment on a stellar academic career to come. So, in spite of what often looks like a blatantly exploitative job, junior researchers still compete for postdoc opportunities  at research facilities working under well-connected faculty.

For immigrants, however, a postdoctoral fellowship is often neither temporary nor freely chosen. A postdoc is meant to be a temporary placeholder until something better comes along, but what if nothing better is up for grabs? It’s especially telling that, as of 2018, 51 percent of postdocs were foreign nationals. Immigrants more often than not realize that a postdoc is rarely a short-term affair; being a postdoc becomes a career, and an unstable, stagnated one at that. When grants dry up, postdocs simply move on to the next lab, the next professor. It’s common and almost expected to spend five or more years as a postdoc before moving on to the next job, and even then there’s no guarantee that the next job will be any better.

Because the US job market tends to be less forgiving for foreign nationals, the postdoc is a vital opportunity for immigrants to make their mark through important publications in well-known journals. Only by building up a substantial portfolio can they further a career in academia or even find a position in industry. Unfortunately, individual potential for innovative research often stagnates after too many years of postdoctoral research. Recall that most postdocs are in their thirties or forties, and more often than not, the years spent in school and then research make them overqualified for most entry level positions in industry, but under-qualified for other senior roles.

Condemned by the Visa

As much as the postdoc can become a semi-permanent position, it is also a volatile one for immigrants. The stamp of foreign status can be used as leverage to guarantee output. In some cases, this is used to take advantage of workers afraid of being sent back to their home country. Right after signing off on a postdoc’s visa renewal, one professor decided that he would be withholding 30 percent of the previously agreed upon salary. In another lab, one student remembers seeing foreign postdocs who were waiting for experiments to run sneaking in naps on the floor of the lab during 100 hour work weeks. Extreme situations like these are unfortunately more common than most would think.

The lack of quantitative surveys on postdocs, much less foreign postdocs, creates room for  marginalization. However, there is a wealth of informal anecdotal evidence suggesting that professors abuse their power in these kinds of situations. The reason for this is that professors are most often the ones who approve visa renewals; the research institution rarely has any oversight in this matter. Visas in exchange for compliance, the right to stay in exchange for labor. All of these stories, of course, are whispered. As one university administrator claims, these events are frighteningly normal. The abuse or otherwise tough conditions are often normalized, and most postdocs refuse to file formal complaints because they fear retribution. In small scientific communities, reputation and connections mean everything, and the goodwill of a professor is critical. Filing a report could very well be a career-ending move.

This exploitation then comes to occur systematically due to the simple fact that immigrants have much more to lose than a job or a promotion to a tenure-track position. Career aside, to stay in the United States, immigrants must have a visa. This sacred document is always at stake, and ironically is what often traps them in the postdoc position.

For postdocs, the most common visa is the J-1 visa, and less frequently, the H1-B visa. Both of these visas are contingent on employment. However, a postdoc position is often synonymous with instability. The job is tied to availability of funds: if a professor’s grant is spent, a postdoc will very likely be let go. Having a job and a visa sponsor will always be the top priority for foreign postdocs; why reach for  competitive positions in industries or tenure-track professorships when postdoc positions are available and frankly more realistic? Immigrants have the most to lose, and the unpredictable politics of visas are often to blame for that.

With the Trump Administration’s recent visa suspensions that placed further quotas and restrictions on H-1B and J-1 visas, the competition for limited positions open to foreigners, especially in industry, will only intensify. For foreigners, the only way to escape the postdoc trap seems to be through luck, brilliance, or obtaining resident or citizenship status. Unfortunately, on all three counts, the odds seem to be stacked against the average postdoc.

So You Think You Can Get a Green Card?

For the majority of workers that don’t catch a lucky break, have a brilliant streak, or snatch up one of the limited spots in industry, what options remain? Because the job market invariably favors US citizens and permanent residents, the only way to get a leg up is to apply for a green card. The green card is the golden ticket for immigrants. It holds the promise of security and employment opportunities; it is the key that unlocks the American Dream. Not surprisingly, the paths to winning that plastic card are incredibly limited for all immigrants, including postdocs. Assuming that the average foreign postdoc is neither married to a US citizen nor seeking asylum, one of two options is a test of luck: the green card lottery. Citizens of certain countries are prohibited from applying; even if eligible, the odds are astronomical. In 2017, 50,000 cards were awarded out of a pool of 19 million applicants. The second option, and most viable for postdocs, is to apply for an employment-based green card.

Within this broad category, there are several classes of visas with different levels of preference. The lowest preference levels, EB-5 and EB-4, are for wealthy business investors and religious workers, respectively. This clearly excludes the typical postdoc, who is generally paid less than US$15 an hour. Two categories up are the EB-3, for professionals and other skilled workers with at least two years of experience, and the EB-2, for individuals with advanced degrees or exceptional abilities. At first glance, this sounds feasible. Unfortunately, the EB-3 and EB-2 both require employer sponsorship to prove that the applicant has or will have a permanent, full-time job—and a postdoctoral fellowship is officially deemed temporary. To apply for these two categories, an immigrant postdoc would need a full-time faculty position or a position in industry.

The last and arguably only resort for any postdoc without a full-time employment offer is the holy grail of green cards: the EB-1. Often referred to as the “genius visa,” it is as daunting as it sounds. It is reserved for individuals with “extraordinary abilities” such as “outstanding professors or researchers; and multinational executives and managers.” To be eligible, an applicant has to be at the very top percentiles of their field, and proof—such as publications or positions held—is required. The only reason it is the only feasible option for many foreign postdocs is that they do not need a permanent position to apply. Nonetheless, it is incredibly daunting and difficult for junior researchers to apply for and be awarded this visa. The immigration system is set up so that a select few are successful. The question is, what happens to the rest?

Exposing a Broken System

The fundamental problem in combating the exploitation of foreign postdocs, and immigrants in general, is the lack of visibility. The majority of the time, these postdocs simply go unseen even though it is clear that they and other junior researchers are responsible for much of the output in research institutions. The current system in academia and in domestic politics is designed to lure and almost coercively retain highly skilled foreign researchers. It uses them, but refuses to acknowledge them through protective regulations. After postdocs leave one institution, most universities and research institutions do not track career outcomes for immigrant postdocs. They effectively become transient workers. Some continue building successful careers in academia, or find lucrative positions in industries. As for  the rest, do they bounce between low-level, low-paying academic positions or simply relocate back home, dreams unfulfilled?

All Mine: The United States and the Ottawa Treaty

 . 5 MIN READ

Landmines, as Major General Michael Beary of the United Nations Interim Force in Lebanon aptly describes them, are an “insidious menace” that disproportionately kill civilians. In recent decades, one of the most successful international agreements in recent decades has resulted in significant steps towards their elimination in warfare. The Convention on the Prohibition of the Use, Stockpiling, Production and Transfer of Anti-Personnel Mines and on their Destruction, better known as the Ottawa Treaty, was signed in 1997. The treaty aims for the complete elimination of anti-personnel landmines in warfare and has to a significant extent accomplished that aim. To date, 162 countries have signed onto the treaty, signaling their actionable intention to never develop, produce, use, stockpile, transfer, or retain landmines. The United States, notably, has not acceded to the treaty. Through an examination of the history of landmines, the campaign against them, and the process towards developing the Ottawa Treaty, the successes and failures of the Treaty can be fully understood. Furthermore, the United States’ past relationship with the Treaty can be adequately explored along with a path towards future ascension.

The history of landmines goes back to the American Civil War. The inventor of the first landmine, Gabriel Rains, assisted the Confederacy in their deployment on the battlefield—at the time, land mines consisted of a piece of sheet iron with a brass—covered fuse activated by movement of an object attached to an explosive primer. Despite concerns over their use including a temporary ban by Confederate General James Longstreet, thousands of land mines were deployed during the Civil War, resulting in an unknown number of what Union General George McLellan described as “murderous and barbarous” casualties. Rains himself declared that “Each new invention of war has been assailed and denounced as barbarous… yet each in its own turn notwithstanding has taken its position by the universal consent of nations according to its efficiency in human slaughter.” His words proved prophetic: land mines became a weapon of war used by nearly every nation in the world throughout the World Wars and after. Outside of their devastating effect on civilian populations during wartime, landmines can remain buried for decades after their intended use. For example, Afghanistan saw landmines planted indiscriminately across the country during conflicts in the late 20th century, which continue to kill thousands and serve as a major obstacle to development; Egypt cites a figure of some 23 million land mines within its borders, causing locals to refer to some tracts of the country’s deserts as “The Devil’s Gardens.” By the mid-1990’s, landmines senselessly killed around 26,000 people annually.

In the early 1990’s, an international movement against the use of these devastating weapons began in earnest. Spurred by an international campaign to ban landmines made up of nonprofits including Handicap International, Medico International, Human Rights Watch, the Mines Advisory Group, the Vietnam Veterans of America Foundation and Physicians for Human Rights, in December of 1993 the United Nations considered a moratorium on the exportation of anti-personnel mines. However, the ultimate amended Protocol that emerged from three years of negotiations on the subject fell short of total prohibition. Independently, some countries—including Belgium, Canada, Switzerland, and France—destroyed their stockpiles and stopped exporting land mines. In late 1996, countries frustrated by the lack of international progress convened in Ottawa for a summit entitled the “International Strategy Conference: Towards a Global Ban on Anti-Personnel Mines.” After establishing a framework for negotiations going forward, 97 nations signed the Brussels Declaration of 1997, which called for a comprehensive ban on the production, stockpiling, and transfer of landmines, the destruction of stockpiles of landmines, and international assistance in the clearance of land mines from conflict zones. This declaration became the basis of the Ottawa Treaty, which was officially signed in Canada in December of 1997.

The Treaty has largely been successfully implemented, though some areas of enforcement are lacking. For one, the codification of an absolute ban on landmines has stigmatized their use and contributed to a new international norm against their deployment in warfare. More concretely, the treaty has directly caused the elimination of 87 countries’ landmine stockpiles; more than 40 million landmines have been destroyed since 1997, saving tens of thousands of lives. Additionally, the Treaty has established meaningful mechanisms for compensation and rehabilitation for victims of landmines and their families, including drastic increases in funding for organizations providing medical, social, and economic support to affected populations. However, some nations are actively trying to leave the Treaty, such as Finland: Finnish Defense Minister Jussi Niinisto called the Ottawa Treaty a “blunder by a peacetime fool,” and urges Finland to remove itself from the Treaty in the interest of its defenses against Russia. Even worse, some states openly defy the Treaty’s ban. A notable recent example is Myanmar. The Burmese military has laid mines in Rohingya villages and along their border with Bangladesh, directly impeding the return of Rohingya Muslims from their place of refuge. Unrelenting international pressure is the best recourse and solution to force noncompliant states to accede to the Ottawa Treaty and stop their use of landmines. Thirty-two nations are not a party to the Ottawa Treaty; however, not all are openly using landmines in warfare. Though the Treaty has seen great success with its signatories, progress towards the complete elimination of landmines in the world will require the ascension of noncompliant states.

More than 40 million landmines have been destroyed since 1997, saving tens of thousands of lives.

The United States—along with the likes of Iran, China, Russia, Pakistan, and Syria—is not a party to the Ottawa Treaty. Although America supported the development process of the treaty, it did not sign it in 1997. The Clinton administration declined to accede to the Treaty under pressure from the Pentagon, which was concerned with the strategic importance of landmines along the Demilitarized Zone (or DMZ) between North and South Korea. The United States, in conjunction with South Korea, has deployed thousands of landmines across the DMZ to serve as a deterrent against North Korean invasion. Robert Beecroft, a State Department official under President Clinton, said that the United States would have signed the treaty if not for the issue of Korea. Little movement was made on the issue until 2014, when the Obama administration announced that it would accede to the Ottawa Treaty in every way except on the Korean peninsula. President Obama announced at the time that “we will begin destroying our stockpiles not required for the defense of South Korea. And we’re going to continue to work to find ways that would allow us to ultimately comply fully and accede to the Ottawa Convention.” The United States pledged to destroy its stockpile of nearly 10 million landmines, not produce anymore, and totally restrict their use in warfare. Advocates of the land mine ban applauded this move, but questioned the logic of requiring an exception for Korea. Steven Goose, arms director at the Human Rights Watch, said that “a geographic exception to the ban is no more acceptable today than when the treaty was negotiated.”

A new opportunity presented by recent diplomatic efforts in Korea could be the beginning of the United States’ ascension in full to the Ottawa Treaty. The recent thawing of diplomatic tension on the Korean peninsula included a 2018 summit between South Korean President Moon Jae-In and North Korean leader Kim Jong Un; both South and North Korea agreed to begin dismantling landmines in the DMZ. Engineers began removing landmines from the region within days of the summit, with the stated goal of total removal. If this policy is successful between North and South Korea, and landmines are eliminated from the DMZ, then the United States would have no reason not to fully sign on to the Ottawa Treaty.

A new opportunity presented by recent diplomatic efforts in Korea could be the beginning of the United States' ascension in full to the Ottawa Treaty.

The Ottawa Treaty is the result of the complicated, ugly history of landmines and the subsequent late-20th century movement towards their elimination. It is a success in progress: though it has undoubtedly saved lives and made the world safer, landmines are still in use in some countries around the world. The United States has so far failed to accede to the Treaty, largely due to the outlier of the Korean peninsula and the necessity of landmines in the DMZ as a deterrent. New developments towards the removal of those mines, though, may finally mean that America can add its full, unconditional support to a global landmine ban.

Europe's Old Beacon of New Hope: The Overdue Creation of a European Army

 . 6 MIN READ

For more than seven decades, NATO has been an “anchor of stability and a beacon of hope.” North America and Europe have even been called “two sides of the same coin” with regard to international security interests. Since the end of World War II, the close ties between the two continents have guaranteed the security of the Western world. Yet, in early June 2020, US President Donald Trump announced plans to withdraw approximately 9,700 US troops from Germany, thereby reducing the United States’ military presence in the region by 30 percent. Former US Ambassador to Germany Richard Grenell explained the motives behind this momentous decision, criticizing Germany’s comparably low defense spending and describing how “[US] taxpayers are getting a little bit tired of paying too much for the defense of other countries.” From the US point of view, Germany, like most of Europe, has become an unreliable defense policy partner because of their insufficient financial support for NATO.

However, from a European perspective, Trump’s attacks constitute a threat to the alliance’s cohesion. After Trump suggested that the United States completely withdraw from NATO,  a mere 11 percent of Germans trust him to respond appropriately to global developments. With Russian military activities in Ukraine endangering Europe’s security from the east and Trump questioning the future of NATO from the west, the political leaders of the European Union now feel pressured to find new ways of ensuring the continent’s safety. German Chancellor Angela Merkel remarked how “the times, in which we could rely on others without reservations, are over. This means that we Europeans must take our fate into our own hands, if we want to survive as a community.” While Trump aims at strengthening NATO by having its members contribute an equitable amount of money, his decision to withdraw US soldiers from Germany may actually push the European Union toward developing its own military strategy. Trump’s pressure might actually promote the type of European military integration that has long been proposed but never fully realized.

Ramstein Air Base, European headquarter of the U.S. Air Force, in Germany. Photo by Kenny Holston, public domain, accessed via Wikimedia Commons.

A Slow Start

In the 1950s, the leaders of six major European countries including Germany and France agreed to form the European Defence Community. Implementation, however, ultimately failed after a veto by the French National Assembly, which feared German rearmament so soon after World War II. Instead, Germany, along with the other major European powers, joined NATO to unite their various military forces under the US umbrella. As a result, the United States became Europe’s de facto military protector during the Cold War; it was not until the fall of the Berlin Wall for the idea of European military integration resurfaced again. Since the collapse of the Soviet Union, and the subsequently decreased military presence of the United States in Europe, numerous politicians—such as German Chancellor Helmut Kohl in 1991, French Prime Minister Alain Juppé in 1996, and British Prime Minister Tony Blair in 1999—have advocated for the implementation of a common European defense union. Yet, all of this grand rhetoric only turned into minor actions.

It took until 2004 and the establishment of the European Defence Agency (EDA) for the first political step toward European military integration, providing an informal forum for European ministries of defense to collaborate without any kind of legal framework. Eventually, the 2009 Treaty of Lisbon built a statutory foundation for European military cooperation and even included a Mutual Assistance Clause in case one member state would become the victim of “armed aggression on its territory.” For the first time, military integration had become a formal item on the European agenda. Today, initiatives such as the Permanent Structured Cooperation (PESCO), the European Defense Fund, and an annual Review of (European) Defense try to deepen the continent’s military cooperation by financing collaborative defense industrial projects and synchronizing the member state’s individual defense plans with one another. This institutional groundwork has enabled the various European Union member states to jointly carry out military and civilian missions across Europe (UkraineBosnia, and Herzegovina), Africa (Somalia and Mali), and Asia (Georgia and Afghanistan).

Training activity as part of the EU Training Mission in Mali. Photo by DPAO EUTMMali, CC BY-SA 4.0, accessed via Wikimedia Commons.

Insurmountable Hurdles?

Despite these promising developments, European military integration still faces many of the same challenges that existed in the 1950s. Some countries freely deploy their military personnel across the globe: France was the driving force behind the ouster of Libyan dictator Muammar Gaddafi in 2011. Other countries, like Germany, have considerable reservations about assigning their soldiers overseas. These asymmetric security preferences combined with the consensus-based institutional organization of the European Union make it difficult for member states to agree upon joint military operations. The addition of growing Euroskepticism among numerous member states makes it very unlikely that a full supranational military force could ever replace national armies. Not only do most countries still value their own military sovereignty over cooperative efforts, but the complicated institutional organization of the European Union would make an efficient military crisis management of a European army almost impossible. Cooperation between countries’ individual armies, then, appears to be the only viable option to further European military integration.

However, even joint military missions are strained in the context of COVID-19. As the EU faces GDP declining by more than 12 percent, the continent’s defense expenditure will likely be severely restricted. While a similar phenomenon occurred after the global economic crisis of 2008, the impact on Europe’s defense budget will be even greater this time around. Experts have already warned that this looming decrease in military spending could have massive negative effects on European military integration. If European leaders do not act decisively, the progress made toward a European defense community might be lost—possibly forever.

Moreover, external pressures from international actors could seal the fate of Europe’s joint defense aspirations. The United States in particular has strongly resisted EU attempts to develop its military capabilities, worried these efforts could limit US control over regional military decisions. The United States also fears European military integration for economic reasons: a united European defense community could exclude US defense contractors. US ambassador to NATO, Kay Bailey Hutchison, acknowledges how “we do not want this to be a protectionist vehicle for the EU…. We want the Europeans to have capabilities and strength, but not to fence off [US] products.” US opposition to extensive European military integration, whether through indirect political pressure or aggressive lobbying within the European Parliament, is as old as the idea of integration itself.

European leaders at a European Council Meeting in 2017. Photo by Tauno Tõhk, CC BY 2.0, accessed via Wikimedia Commons.

Now or Never

The removal of US troops from Germany could make European leaders realize how the United States’ current economic priorities have dominated their traditional strategic political alliances. Security experts like diplomat Wolfgang Ischinger have warned for almost five years that “[while] a strong transatlantic partnership will remain the best security guarantee for Europe, this preferred option may not be available in the future.” Current polls indicate that 75 percent of Europeans advocate for common defense and security policies, and that 68 percent would like the European Union to cooperate more closely on defense. At any rate, European citizens seem to have understood these warnings. It remains to be seen if the politicians have. If former US Vice President Joe Biden, an outspoken advocate of refreshing transatlantic relations and strengthening NATO, wins the US presidential elections in November, European leaders may easily fall back into the habit of relying on their American ally. Attempts to increase European military integration would once again amount to little.

Alternatively, the European Union could take the numerous warnings of security experts seriously and could satisfy the populace by furthering a holistic European defense strategy. A unified military would finally allow Europe to properly defend itself without having to rely on the assistance of outside forces. How? Simple coordination and an expanded mandate of the European Defence Agency would be sufficient. Importantly, none of that would necessitate the surrender of national military sovereignty, and coherent individual defense spending plans and joint investment into European industrial defense projects would actually allow member states to save money. There is no lack of arguments for military integration and no lack of ideas for how to achieve such integration; there is only a lack of political determination.

For more than seven decades, EU member states have claimed that they do not want to lose their military sovereignty to justify implementing halfhearted integration policies. US military assistance lulled Europe’s leaders into a sense of security and made the hurdle of establishing a European army appear insurmountable. Now, Trump’s decision to withdraw soldiers from Germany could finally push the European Union toward developing its own military strategy. While North American and European international security interests might have once been two sides of the same coin, this perception has masked what is possible under a united European front. It is time to put the topic of further military integration on top of the continent’s political agenda; the future of Europe depends on it.

Too Much Information: Ineffective Intelligence Collection

 . 9 MIN READ

Alex Young. Originally published in the HIR Summer 2013 Issue.

In this age of digitalization and technology, intelligence agencies across the globe process massive amounts of information about individuals, sub-state actors, and governments every day. Intelligence experts ana military leaders often assume that the goal of intelligence work is to gather as much information as possible in order to formulate a more comprehensive picture of the world. The United States, in particular, has become a global epicenter of intelligence work—4.2 million US citizens, more than 10% of the country's population, have some form of security clearance. However, this aggressive intelligence gathering does not make for better-informed government agencies or higher quality security policy. Instead, excessive information collection leads to information overload on both the individual and institutional levels, impairing the US intelligence community's ability to do its job. What’s more, US government agencies do not use this information effectively, due to overclassification problems. These inefficiencies in intelligence ultimately sow instability in the international system and increase the likelihood of conflict between states.

Too Much Information

The US intelligence community is currently inundated with information. This poses a serious challenge to effective intelligence work. Overwhelmed by data, analysts lose the ability to pick out what is important and fail to make good judgments. In a 1970 book, futurist Alvin Toffler of the International Institute for Strategic Studies coined the term "information overload" to describe situations in which an excess of information results in poorer decision making. Today, this phenomenon holds true on both an individual and an institutional level. Modern psychology teaches that the human brain can only focus effectively on so much information at a time. As a person tries to complete more tasks simultaneously, his or her efficacy in dealing with each individual task diminishes in a phenomenon called "cognitive overload." Psychologist Lucy Jo Palladino writes that information overload leads to added stress, indecisiveness, and less effective analysis of decisions. This feature of human attention has clear implications for security policy: attempting to collect more and more information makes a nation less secure when it overloads intelligence analysts.

Information overload carries over to the institutional level in three ways. First, institutional skill is in some ways nothing more than an aggregation of individual talent. If every member of a group is overwhelmed by an excess of information, then the organization as a whole is unable to operate effectively. Secondly, institutions may face the challenge of circular reporting. In the process of collecting massive amounts of information, agencies may collect the same information twice from different sources. When faced with high volumes of incoming reports, intelligence agencies cannot easily prevent this duplication of data. It is particularly difficult to detect circular reporting when the shared provenance of the information is obscured—for example, when the information is delivered to intelligence officers through secondary sources. Not only does circular reporting add to inefficiency, it also can lead analysts to place too much importance on the twice-reported information. This is because analysts measure the credibility of intelligence reporting in part based on how many independent sources confirm the report. That standard becomes problematic, though, if one source appears to multiply through circular reporting. Third, organizations fall prey to the sheer complexity of their own intelligence gathering frameworks. These three factors make information overload a serious problem at the institutional level.

To make things worse, institutions are notoriously ineffective at evaluating themselves. Leaders of organizations like those inside the US intelligence community are more likely to keep their jobs or to be promoted if the organization they manage appears to be successfully and efficiently living up to expectations. The head of an agency therefore has a vested interest in appearing competent, leading to strong efforts to avoid unfavorable evaluations, especially from the inside. This means that leaders face powerful incentives to suppress negative evaluations, marginalize the internal evaluators, and overlook their own shortcomings. Intelligence institutions thus cannot easily identify or resolve information overload.

Unfortunately, it may prove impossible to solve both individual cognitive overload and institutional information overload: the solution to one problem exacerbates the other. Individuals should take breaks and limit their number of concurrent projects in order to minimize cognitive overload. This suggests that organizations should hire more analysts to deal with the flow of intelligence. However, as organizations grow larger and more unwieldy, the problems of circular reporting and complexity become only more serious.

These cognitive and institutional problems are degrading the capability of US intelligence agencies to keep track of world events and to use information to guard against security threats. Since the terrorist attacks on September 1 1th, 2001, the US intelligence community has ballooned: since the attack, 263 separate organizations have been created or reformulated. Those offices represent 20% of the government organizations that do intelligence work. The analysts among the 854,000 people with top-secret security clearances produce an overwhelming number of intelligence reports—so many that the Office of the Director of National Intelligence cannot keep track of exactly how many reports are completed each year. Perhaps most strikingly, the National Security Agency intercepts and stores 1.7 billion emails, phone calls, and other communications every day, a small portion of which are organized into 70 databases. Paradoxically, then, by trying to do more, US intelligence agencies are accomplishing less.

Too Much Secrecy

The US intelligence community does a poor job sharing information internally, between agencies and between analysts, because vast portions of that information are overclassified. Overclassification occurs either when information is classified but should not be, or when information that is classified should be classified at a lower level.

This is by no means a new problem. Almost six decades ago, a Department of Defense report argued that overclassification had "reached serious proportions." More recently, public figures from John Kerry to Donald Rumsfeld have expressed strong concerns about how many documents are classified in the United States. Excessive classification has persisted because it arises from a fundamentally perverse incentive structure facing decision makers. Officials who decide whether to classify documents and how strictly to limit their circulation face virtually no consequences if they classify a document whose contents did not warrant such a designation. On the other hand, those officials are punished severely for failures to classify sensitive information. This leads decision makers to err on the side of caution, choosing to classify documents at higher levels in uncertain cases. The result is massive overclassification and institutional failure to make information available where and when it is needed.

The numbers are striking. In 2010, officials in the US intelligence community made 22 million more classification decisions than they had in 2009, reaching an annual total of 76.8 million classification activity events. These estimates include both original decisions made about the classification status of new incoming information as well as derivative decisions reviewing previously classified information. Spending levels also act as an illuminating illustration of the scale of overclassification. The Information Security Oversight Office, the US government agency that keeps track of trends in classification and intelligence management, publishes annual reports on the state of the intelligence community. In 2011, the US government spent almost US$11.31 billion on all classification efforts. This number does not include the costs of classification by the Central Intelligence Agency, the Defense Intelligence Agency, the Office of the Director of National Security, the National Geospatial-Intelligence Agency, the National Reconnaissance Office, or the National Security Agency. Those expenditures are themselves classified.

By contrast, the total cost of all declassification by the US government was only US$52.76 million—more than 200 times less than the amount spent on classification. Perhaps most tellingly, Elizabeth Goitein, Co-Director of the Liberty and National Security Program at the Brennan Center for Justice, reports that in 92 percent of cases in which a member of the public appeals for the declassification of a record, the agency in question determines that at least part of the relevant document did not need to remain classified. Secrecy is out of control in the US intelligence community.

Overclassification has become an obstacle to intelligence sharing across agencies, potentially leaving analysts in the CIA without easy access to necessary information gathered by the NSA (or other agencies), and therefore, with diminished ability to formulate an accurate picture of the world around us. The lack of transparency that necessarily results from such large-scale classification also decreases accountability, thereby reducing the incentive for analysts to carry out accurate intelligence reporting. Analysts cannot easily be reprimanded or commended for their work unless their superiors can gauge the accuracy of the information they produce and use. Excessive secrecy also precludes open discussion of security policy questions, fueling public ignorance on issues of national security and eliminating the government's ability to take into account the voice of the people. Essentially, the US intelligence community has limited itself by placing too much emphasis on secrecy and not enough on efficiency.

Intelligence and International Insecurity

The failures of US intelligence do more than just erode US security. Given that the United States shares intelligence with many of it allies and coordinates with militaries across the globe, especially since 9/11, lapses in judgment and inconsistencies in intelligence on the part of US analysts cause ripple effects throughout the military and intelligence communities across the world. Since 1946, the United States has upheld a signals intelligence sharing agreement—often called "Five Eyes"—with the United Kingdom, Canada, Australia, and New Zealand. Washington also cooperates closely with newer allies in the Middle East and South Asia, including Israel, Saudi Arabia, Yemen, and Pakistan. US intelligence agencies have even reached out to governments that have traditionally not been their greatest partners—nations such as China and, before the Arab Spring, Libya and Syria.

This means that any failures of US intelligence are multiplied and spread across the international community. These shortcomings in US intelligence collection have serious security implications for the world as a whole. States make decisions about entering, exiting, and preparing for war based on their perceptions of the international system, their views of how power is distributed, and their understandings of what capabilities other states have. All of these assessments are shaped by the United States' and other nations' abilities to collect accurate, relevant information and distribute that intelligence to allies.

Furthermore, less effective intelligence work heightens the chance of war between states. One classic problem in political science deals with why war occurs: conflict is costly for both winners and losers, which seems to suggest that it is irrational for two states to wage war. One prominent explanation for the existence of war, then, is that states act rationally but make mistakes due to imperfect information. That is, if the world were in fact exactly the way a state perceived it, then that state would be acting rationally. But, because the world differs in some significant way from that state's view of it, the state makes irrational choices. Specifically, governments make miscalculations surrounding relative military capabilities, strategies, the intentions of allies to provide support, or the resolve of military and civilian leadership to pursue a drawn out conflict. A state's misunderstanding of one or more of these factors could lead it to overestimate its ability to win wars, leading it to enter more conflicts.

Historically, there have been many wars founded on misinformation or incomplete intelligence—conflicts which could have been avoided by better intelligence work. Overclassification prevented US intelligence analysts from making the right connections in the months and days leading up to the terrorist attacks of September 11th, 2001; the 9/11 Commission later blamed those intelligence gaps on "overclassification and excessive compartmentalization of information among agencies." Moreover, the US war in Iraq from 2003 to 2011 began because of a widely held belief that Iraqi President Saddam Hussein possessed and was perhaps willing to use weapons of mass destruction. This claim proved false, but the war nonetheless claimed more than 50,000 US and Iraqi lives, left more than 100,000 people wounded, and condemned Iraq to years of instability. With better intelligence work, the US intelligence community could have seen this miscalculation before it was too late.

Today, faulty intelligence gathering still poses a threat to peace. For example, without accurate information about Iran's nuclear capabilities and intelligence about the nature and locations of its nuclear plants, a risk-averse Israel could overestimate the need to take drastic, preemptive measures against Iran. Israeli Prime Minister Benjamin Netanyahu has already called for international action to stop Iranian development of functional nuclear weapons, calling that prospect… the main [problem] facing not only myself and Israel, but the entire world." Israel's decision to strike or not to strike Iran depends in large part on the Israeli government and military's perceptions about Iran's strength, capabilities, and intentions. If faulty intelligence leads Israel to believe that Iran is putting the finishing touches on a nuclear arsenal or that Iran intends to use a nuclear weapon against Israel, Israel will likely carry out an airstrike on suspected Iranian nuclear facilities. Such action may be acceptable or even responsible if Iran is indeed in possession of a nuclear weapon, but if that intelligence turns out to be false, an Israeli strike would destabilize the region without achieving much. Inadequate information therefore continues to have the potential to create unnecessary conflict.

Conclusion

US military strategy and security policy depend heavily on the ability of intelligence analysts to piece together a clear, coherent, and accurate picture of the world. Unfortunately, the agencies tasked with gathering intelligence have overreached their capacities and grown far too unwieldy, while at the same time placing too much importance on secrecy. This hinders their ability to do their job: due to individual cognitive overload, institutional information overload, and the overclassification of intelligence, US intelligence offices can no longer accomplish what they are intended to do. By trying to do too much, the United States has harmed its own ability to collect intelligence. This problem grows even more serious when one considers that the United States shares intelligence with allies across the globe. Unless the United States rapidly improves its ability to gather and sort through intelligence, the international system will remain less stable and more prone to conflict for the foreseeable future.

Consent by Default: When Civil Society Uses Surveillance

 . 9 MIN READ

For centuries, people have protested against perceived injustices, often facing punishment as the consequence of their participation. While authoritarian regimes tend to suppress such protests, which are potential threats to their power, liberal democracies are generally more tolerant, allowing participants to continue their activities and remain anonymous to both the state and civilian third parties. Over the last decade, however, technological advancements have increasingly eroded this expectation of anonymity, both during protests and on a daily basis.

Today’s ubiquitous mobile devices offer constant connectivity, a seemingly benevolent feature that also enables third-party entities to easily collect vast amounts of data. This technology generates much more data with much greater precision, across a much larger number of users than ever before—all of which make the collected information very powerful. Many firms have created analytical tools for these data, providing precise information about the locations visited by individuals and putting it on sale. Civil society groups in the United States have already taken notice of this trend, hoping to take advantage of this emerging resource to further their political interests; peer organizations around the world will undoubtedly follow suit. While the various concerns about this technology have yet to be addressed, one thing is clear: personal data will play an increasingly important role in global politics moving forward.

Local SEO Google map image I created for a blog post on local marketing. An everyday interaction with technology.
A map application is open on a smartphone. Photo by henry perks / Unsplash

From Communication to Data Generation

As all of those who are familiar with landlines know, phones were not always mechanisms for semi-covert, third-party surveillance. Today’s cell phones were originally intended to improve connectivity by making telecommunication more portable. This purpose, however, has shifted significantly over time as masses of people—first in the developed world, and more recently in the global south—began to be able to afford this technology, thereby creating a highly lucrative market and putting pressure on firms to constantly improve their products. As a result, a tool originally meant for voice communication gradually evolved into a tool that could also transmit text and images to people around the world.

The spread of this technology has been truly global, extending well beyond Western industrialized societies. For example, China, the world’s most populous country, is now a leader in rolling out 5G networks, while India, the second-most populous, has more cell phones than toilets, in addition to the world’s least expensive cellular data. Most of sub-Saharan Africa has also adopted cell phones, although only a minority of people use smartphones, which are more concerning from a privacy perspective. While the usage environments are undoubtedly different in each of the many countries experiencing a boom in cellular communication, it is clear that there is an increasing number of people around the world who are making themselves vulnerable to external tracking. This trend coincides with the relatively recent emergence of civilian cellular tracking, making the present moment an important turning point for the role of mobile data in politics. Accordingly, potential regulatory decisions about data sharing are also more influential than ever before.

#Save Our Internet
A protest in Stuttgart, Germany against the European Union's Article 13, a law that requires online platforms to filter uploaded content to protect copyrighted material. Photo by Christian Wiediger / Unsplash

Civil Society — And Governments

Without regulation, the data from mobile technology will soon have profound effects on political systems around the world. The example of a hypothetical protest in a liberal democracy shows the many uses of this data. First, third parties can track the approximate number of people in attendance, providing valuable information to protest organizers, opposing groups, and interested political parties. Second, they can perform additional data analyses to understand the protesters’ demographics as a group, which is yet another valuable piece of information. Finally, at the individual level, third parties can identify protesters’ destinations following the event, tracking them to their homes and workplaces. This creates more opportunities for political entities to pursue their interests through tactics such as direct-mail campaigning and individually tailored fundraising efforts.

In the hands of mainstream civil society groups, this type of information is not necessarily harmful, particularly considering that personal information has long been printed in telephone and address directories without any major issues. However, people have moved closer to either end of the political spectrum in recent years, making civility in confronting political disagreement all but impossible. This means that political groups are no longer necessarily restricted by social norms of propriety, or even legality, when selecting methods of using and disseminating information. These methods include using social media campaigns to defame individuals, doxxing opponents to discourage continued political participation, and even threats of physical harm. These strategies have effectively weaponized personal information, making it unsafe for individuals to allow their data to be collected in the first place.

Furthermore, the methods that make mobile data attractive to civil society groups are the same ones that make the data a potent tool for governments to suppress the political elements of civil society. For instance, the address data used in direct-mail campaigning is the very same data needed to make an arrest, while event attendance data can easily be used to determine the amount of military or police force needed to end a protest. Unfortunately, this is not just a hypothetical situation: governments around the world are already using these forms of surveillance and have been actively collecting the information necessary to suppress protests and eliminate dissent.

More alarmingly, the use of digital surveillance is not limited to authoritarian regimes such as Russia, China, and Iran. One prominent example is India, a country often lauded for its relatively stable democracy, where the government has developed a cryptic intelligence gathering system to spy on its citizens. Similarly, South Korea—a democratic model of governance upheld by the liberal world order as the ultimate goal for its northern neighbor—collects significant amounts of digital information from cell phones, including financial transactions and contact records, whenever they connect to a cellular tower. The COVID-19 pandemic has already demonstrated the potency of mobile data: South Korea was able to launch an aggressive contact tracing campaign without creating a new dedicated app, showing that the types of data that are normally collected are more than sufficient to pinpoint individual citizens’ whereabouts with great accuracy.

Big data is watching you
A sticker on a light pole warns about the privacy concerns associated with data collection practices. Photo by ev / Unsplash

Regulation: A No-Win Situation

On the surface, it may appear that the third parties collecting the information should be at fault for the potentially devastating consequences linked to this surveillance. They, along with technology manufacturers, have created the physical and digital infrastructure needed to track individuals’ every activity. They have also developed the tools used to convert bits of information constantly sent from phones into large datasets filled with rows of coherent, personally identifiable data. Nonetheless, governments actually hold the key to the issue of misused information.

From a historical perspective, this new form of information-gathering is only the latest form of the digital control exercised by governments around the world. The previously preferred method of suppressing dissent was to simply shut down cellular networks or access to the Internet, preventing protesters and other aggrieved parties from communicating with each other during critical times. Countries as diverse as the United StatesRussiaUkraine, and India have used this technique to quell civil unrest, not to mention other parts of the world where such shutdowns simply go unreported. This common experience of digital control indicates that governments, regardless of their regime type, have a major stake in tools that help mitigate the impacts of social unrest and are unlikely to deprive themselves of such powerful technologies.

Moreover, whether or not governments regulate the practice of third-party data collection, it is unlikely that user privacy will change substantively. As previously mentioned, governments benefit from the information gathered, and because it would be impractical to create double standards prohibiting the practice for civilian entities, they are likely to continue allowing the data collection. Alternatively, incumbent data firms are likely to oppose any restrictive regulatory changes in data collection, particularly those that have built their business models around this capacity. Many other businesses would also be opposed: for example, in the United States, a sizable portion of advertising is done by targeting a particular audience or demographic, which requires the collection of individual consumer data. Finally, in addition to the inevitable opposition that a move to ban personal digital surveillance would bring, today’s average smartphone user does not have a clear understanding of the information being collected and transmitted because of the technological and legal complexity involved. This means that the people most affected by regulatory changes, that is, individual users, are unlikely to see any meaningful change in their daily interactions with personal digital devices.

The ignorance of today’s technology users has already proven detrimental to existing regulation. In 2018, the General Data Protection Regulation (GDPR) went into effect across the European Union, providing protection to individuals by requiring data firms to be more transparent about their processes. However, the EU reported in a survey that two years after the GDPR was implemented, only 22 percent of respondents actually read the terms and conditions of digital services, and only 41 percent knew the privacy settings for all of the apps on their smartphones. These low rates of understanding, particularly among a coalition of advanced industrial societies, clearly indicate that change is difficult at the individual level. Furthermore, the US state of California implemented a similar Consumer Privacy Act (CCPA) in January 2020, with analysts and law professors predicting that users are unlikely to take the proactive steps necessary to prevent their data from being gathered and used. Collectively, these relatively recent developments make one thing clear: even when well-intentioned and presumably well-designed regulation is implemented, it does not necessarily mitigate the issues surrounding civilian digital surveillance.

Several individuals using computers and other electronic devices. Photo by Marvin Meyer / Unsplash

When Democracy Breaks Down

Given the substantial barriers to regulation, it is imperative to confront the alternative of an unregulated world for individual digital surveillance technology. Such a world would not necessarily be an unlivable one; cellular phones have arguably improved the lives of many, and most people who do not already own one are eager to obtain one and experience the convenience of portable telecommunication. However, a livable world does not mean that its conditions are desirable. In particular, an unregulated world means that data can be constantly transmitted to unknown third parties, creating a lingering threat to individual privacy.

Unfortunately for the developed world, much of which is composed of liberal democracies, this threat to privacy can also be a threat to the viability of a democratic regime. A lack of privacy reduces individuals’ sense of security and poses a major barrier to participation in all forms of political activity. The loss of privacy becomes particularly damaging to democracies when personal data is sold to individuals and groups on both sides of the debate, allowing an individual to be hurt twice, whether physically, emotionally, socially, or financially. Without protections and adequate enforcement mechanisms in place, individuals from all walks of life will be discouraged from participating in protests and other forms of public self-expression, stifling not only the possibility of positive change but also threatening the very foundation of democracy.

This digital menace presents the ultimate challenge for democracies around the world. In the near future, they will have to balance the needs of citizens, government, and private enterprise in crafting appropriate responses to the emerging trend of mobile data collection. A daunting challenge in itself, the process promises to be especially difficult in countries such as the United States where corporations are legally considered people under certain circumstances and thus deserving of certain civil protections. Any restrictive regulation on data, even if intended to protect democracy, would threaten the ability of legal persons in the form of corporations to exist freely, thus undermining democratic values. Even if this issue is resolved, a new one will inevitably emerge. When this happens, most users are unlikely to notice because they are unaware of how the technology and the law surrounding it work, making regulation unlikely to become an important issue for the average citizen. The second challenge becomes to increase public awareness of ways to avoid third-party data collection, a task that will surely prove most difficult of all. Unless resolved, ongoing public ignorance about the dangers of surveillance will allow digital interference in politics to continue unhindered and slowly erode the foundations of democracy.

Soil Degradation and Crippling Debt: The Agrarian Crisis in India

 . 5 MIN READ

Every day, four farmers in India take their own lives. According to data released from the National Crime Record Bureau, 10,655 people working in agriculture committed suicide in India in 2017. According to the same report, the largest causes of these suicides include stresses caused by failed crops and crippling debts. This is the result of a deeply complex crisis occurring in the agriculture industry.

In India, over 70 percent of the population depends on agriculture for income; yet farming there is an incredibly risky and potentially dangerous business. Farmers are unable to reliably produce profitable crops due to outdated farming practices and fluctuating prices. Most suicides in India are committed by farmers who produce cash crops like cotton and sugarcane, which are high risk and “not based on [the] principle of sustained and resilient high yield." In addition, farmers are often forced to take on crippling debt. The costs of being a farmer in India have been increasing since the 1990s, but farmer income has plateaued or even decreased in that time period. The government continues to attempt to solve this agrarian crisis, but no solution has yet been successful in creating sustainable profitability for the agriculture sector.

Arable land in India has become gradually less productive, due to soil degradation and unsustainable farming techniques. BKP Sinha, a journalist for The Pioneerwrites, “Besides a host of other factors, the main factor being attributed to a large number of such suicides is sickness of our soil, steep decline in groundwater table, deteriorating discharge of rivers and flooding due to mismanagement of land in the catchment and riparian areas.” In India, 53 percent of land is estimated to be affected by soil erosion and land degradation and 83 percent is affected by water and wind erosion. Unless there is something done to fix the environmental effects of current farming practices, it is estimated that within 20 years, one third of arable land will be lost. This, along with the economic difficulties facing farmers, creates a crippling agrarian crisis.

India's Response

In order to stimulate the economy in the long run, the Indian government plans on moving farmers into manufacturing and other industries in residential and urban areas, overtime replacing the farming industry with larger, international corporations. However, this solution fails in many ways. Firstly, there are not enough jobs available in urban areas to provide income for displaced farmers moved into the city. Secondly, handing farming over to large, international organizations does not guarantee any solution to the negative environmental impact and unsustainable techniques currently taking place. Sinha notes, “Increasing quantity of outputs with continuously decreasing uncertain output and risk due to climate change in the rained areas makes agriculture less resilient and more risk prone.”

To solve the problem of fluctuating prices in the market, the government has been working to moderate production surpluses. The Cotton Corporation of India (CCI) says they are ready to buy more than five million bales of cotton from farmers in the 2019-2020 season, which is the highest amount produced in five years. This has the potential to keep the price of raw cotton and a profitable price for farmers, but is incredibly costly to the government and a difficult long-term solution.

Currently, the government has been offering loan waivers in an attempt to decrease the debilitating debt facing many farmers. This has been ineffective as the waivers unintentionally ruin farmers’ credit and weaken the banking system in India. Additionally, they decrease the incentives for farmers to pay off their loans because they often choose to just wait for the next waiver. Worst of all, despite this costly program, indebtedness remains high because farmers often rent farmland from money lenders, who could be accurately described as loan sharks. These money lenders do not abide by loan waivers, so there is no benefit in this program for farmers that owe debt to money lenders.

The government has also implemented the Prime Minister Kisan Samman Nidhi (PM-Kisan) scheme which gives 6,000 rupees to every land-owning farm family. However, the majority of families facing the worst levels of indebtedness do not own the land they farm and thus this program is limited in its ability to help the most poverty-stricken farmers. Furthermore, the central government provides a long list of initiatives to assist farmers, including PM-Kisan, fertilizer subsidy, crop insurance, farmer pension, food subsidy, and more. Nevertheless, all of these plans have been unsuccessful at curbing the crisis.

Other Solutions

On March 11, 2018, thousands of farmers arrived in Mumbai to protest the government policies and inaction that has negatively affected their community. Among other pleas, such as debt waivers, protesters called for the implementation of India’s Forest Rights Act of 2006, which gives land to peasants of indigenous tribal communities who traditionally cultivate forest acres and makes that land legally available to them. This policy, if implemented as it was designed, could provide land to many farmers that currently are in debt from renting land to cultivate. Providing farmers with their traditional farmland could ease the long-term financial burden on families and secure lasting profit for local farmers.

Currently, one reason farming is so risky is because 52 percent of India’s cultivable farmland is reliant on rainwater, which can be unpredictable. The government could invest in infrastructure to support farming, specifically irrigation canals. Another reason farmers are losing profits is because they are unable to save their produce before selling it. This forces them to sell at the market price, without being able to wait for a better price and time their sales. If the government in India built or subsidized warehouse facilities farmers would be able to store their produce and sell it when they could make the greatest profit.

As previously discussed, the current governmental programs of farm-loan waiver has been ineffective and costly. For example, the central government spends the equivalent of over US$11 billion on fertilizer subsidies alone. Not only have these policies been ineffective, but fertilizer subsidies, and other subsidies like it, have depleted the soil and the farm output, which increases demand for sustainable agriculture.

One of the reasons the Indian government has been so interested in the potential of large agricultural corporations is because they have the capital to implement modern farming methods, which are far more efficient and can decrease costs and increase profits in India’s agriculture sector. However, it is possible to implement these methods while allowing current farmers to continue their current agriculture jobs with increased profitability. Independent farmers and small farming practices would be able to hire precision farming equipment, “even sensors, drones, satellite support apart from tractors, harvesters and planters.” Instead of investing in the capital to own these technologies, farmers could rent them for a season.  Instead of subsidizing fertilizer and other farming methods that are taxing on the environment and less effective than modern technology, the government could subsidize precision farming technology and equipment. This method of farming would result in a greater quantity and quality of output and more sustainable farming practices.

Regardless of which solutions are tried and implemented, this issue will be difficult to address. It has created strong welfare, economic, and environmental concerns for the region to weather. Without some sort of changes, the agrarian crisis, and more importantly the needless loss of life through suicides, may remain unchanged.

Green Gold: Making Money (and Fighting Deforestation) with Yerba Mate

 . 7 MIN READ

Argentina’s Misiones province has a bit of a problem. As several local newspapers have reported, thefts of the province’s main agricultural commodity, often referred to simply as “the green leaf” or “green gold,” have risen, and authorities, despite making several arrests, are at a loss.

Misiones is not a hotbed of the coca trade, but rather a center of production for one of Argentina, Uruguay, and Paraguay’s most beloved commodities: yerba mate, the leaves used to prepare a herbal infusion called mate (pronounced mah-tay, with two syllables). As a Paraguayan villager told Harvard’s ReVista, drinking mate “is such an intrinsic part of us that you really don’t think about it until somebody else points it out.” However, unlike other caffeinated drinks like Italian coffees and Chinese teas, mate has not largely made it past the Southern Cone and into the United States’ globalized drink palate, likely because it is an acquired taste: one self-described “reluctant” mate drinker described it as a mix of “green tea and coffee, with hints of tobacco and oak.”

Yet mate’s global popularity has finally started to increase in the last decade or so. Entrepreneurs have started to tap into the growing American market for alternative teas—and into Gen Z’s preference for more ecologically and socially conscious products—with shade-grown yerba mate. Since mate production provides an economic incentive to reforest the Atlantic rainforest in Brazil, Argentina, and Paraguay, these emerging companies demonstrate how ecological restoration and profit can go hand in hand.

The Ecology of Mate Production

Yerba mate has been around a long time. Sixteenth-century Spanish colonizers in Paraguay marvelled at how the indigenous Guaraní people who imbibed the infusion could work longer and harder than them. After the Guaraní ignored a 1616 ban of the green leaf, Jesuit missionaries noticed its value to their indigenous laborers and planted it alongside other agricultural commodities at their missions; it soon became a valuable export commodity for the missionaries, who could claim a tax exemption to the chagrin of profit-seeking Paraguayan merchants. After Spanish monarch Carlos III, who wanted to reduce the church’s influence in the Spanish realms, expelled the Jesuits, they abandoned their plantations, although Paraguayans still harvested the natural stands that grew in the Atlantic rainforest.

Large-scale plantation agriculture did not return until the 1890s, when a massive influx of European immigrants—German, Ukrainian, Polish, and others—arrived in Argentina and Brazil’s yerba mate belt. German colonists in the area astutely noticed that they would have to adopt crops already adapted to the soil and climatic conditions in order to survive, and mate fit the bill perfectly. By 1915, the plantation system had become widespread in Misiones, and Argentina overtook Brazil and Paraguay as the largest producer of mate. After all, plantation agriculture had double the yields of solely exploiting the native forests.

Yerba mate ready to be drunk. Photo by Edgar Carlos Franetovic, public domain, accessed via Wikimedia Commons. 

While plantation agriculture increases production, it also has an ecological cost. As with other deforested ecosystems, yerba mate plantations are more exposed to soil erosion and degradation, which poor soil management practices exacerbate. Unsurprisingly, abandoned mate plantations have lower levels of nitrogen and phosphorus than the virgin ecosystem.

Such an extractive model, though, is unsustainable. As a result, yerba mate planters have gradually started adopting an agroforestry model, which involves growing other trees  alongside the mate trees. While this model decreases the productivity of plantations when viewed alone, the ecosystem as a whole benefits because the other trees provide shade for the mate trees that helps the leaves to retain more nutrients and flavor, additional organic material which improve soil moisture, and another root system to prevent erosion. Intercropping with Paraná pine, for instance, reduced soil erosion and moisture loss while improving soil fertility. When managed sustainably, such intercropping can also help farmers diversify their plots with another source of income—primarily timber and fruit—and contribute to the reforestation of the Atlantic rainforest.

Putting Trees in “Paper Parks”

That rainforest, largely coexistent with the mate belt, has seen better days: across Brazil, Argentina, and Paraguay, it has lost more than 93 percent of its land cover. While mate plantations have certainly contributed to deforestation, other culprits have emerged in the past ten years: soybean production and cattle ranching. Argentina and Paraguay both rank within the top five soybean exporters by value, and rising demand for both products has driven rapid expansion into the Atlantic rainforest. In Paraguay alone, soybeans and beef exports comprise 90 percent of combined exports, meaning that the country is dependent on just a few commodities—commodities that also drive deforestation.

To combat this deforestation, international NGOs have seized upon agroforestry as a sustainable and income-generating solution, and shade-grown yerba mate is a primary example. For instance, the World Wildlife Fund started a program in 2001 to create forest corridors that linked previously fragmented forest areas by using yerba mate to secure the buy-in of local communities and funding local female-owned cooperatives such as La Hoja Completa. Around the same time, local cooperative Titrayju sprung up to benefit the indigenous community with organic and eco-friendly production practices. Guyra Paraguay, meanwhile, bought 7,000 hectares of land within San Rafael National Park, regarded as a “paper park” within the country, to develop shade-grown mate agriculture alongside several indigenous communities. Finally, Fundación Moisés Bertoni seeded 856 hectares with around 1.3 million mate seedlings as part of a broader World Bank project to reforest the Paraná rainforest.

These programs have largely succeeded in preserving existing forest habitat because they reduce the economic incentives for deforestation; destroying a forest that already has an income-generating activity makes little economic sense. At the same time, agroforestry projects that plant trees can slowly yet surely reforest areas that were once exclusively yerba mate plantations and add new habitat for local wildlife. Whether farmers plant mate plants in previously existing forests or new trees to augment previous mate plantations, both options create new carbon sinks as well, since mate and shade trees both take carbon out of the atmosphere. And of course, the projects generated income for the local community, especially women and indigenous people.

A New Way to Caffeinate

Even while yerba mate is increasingly becoming the solution to an ecological crisis in the mate belt, its demand abroad has started to increase. That ecological role, in turn, plays a central role in an increasingly savvy marketing campaign by companies espousing principles of “market-driven regeneration.” Profit is definitely on these companies’ minds, but ultimately, the story of yerba mate shows how profit and environmentalism are not mutually incompatible.

A group of five Californian friends founded Guayaki, widely recognized as the pioneer of this model, in 1996. Having founded the company right out of college, Guayaki’s founders quickly seized on college students, with their insatiable desire for caffeine, as a key market by using the tried-and-true technique of collegiate brand ambassadors. Beyond marketing its triple bottom line of social, environmental, and financial performance, the company touts mate’s benefits, especially when compared to coffee or other energy drinks: a more balanced stimulant effect with more natural antioxidants than coffee and fewer artificial chemicals than energy drinks. It also helps that its products are not straight yerba mate, but sparkling beverages with mate as a key ingredient instead.

That marketing campaign has paid off, especially as the markets for organic products, healthy beverages, and “exotic” flavors have started to expand. Campus papers at the University of California-San Diego and Occidental College reported that mate had become a staple at dining halls and off-campus convenience stores. As one Occidental student said, “Every day I will see someone in class with a drink.” Guayaki now makes more than US$40 million a year in sales, and the company reported that it had sold 390,000 cans across 171 universities in 2017 and 2018.

The company’s two other bottom lines have prospered as well. It has stewarded more than 81,000 acres of rainforest, planted 8,800 native trees, and created 670 jobs that paid a living wage of around 2.3 million guaranís (US$357) per month since organic, shade-grown mate can fetch two to three times the yield per pound than plantation-grown mate. These increased wages are critical in a region where the average farmer makes far less than that.

Mate for sale in a market in Barcelona. Photo by Bob Warrick, CC-SA-3.0, accessed via Wikimedia Commons. 

Other companies have emulated Guayaki’s sustainable business model and have started to compete with it. Mi Mate, founded by students at UCSD, follows it almost to the letter, purchasing its mate from an independent farm that pays its workers a living wage and has contributed to the construction of local schools. Eco Teas also specializes in organic mate, and Clean Cause uses 50 percent of its profits to support teens recovering from drug addiction. Others have seized upon the product but not necessarily the mission: teaRIOT has incorporated a mate flavor into one of its drinks, German brand Club-Mate has developed a mate-based soft drink, and even Pepsi has its own mate brand. Yerba mate is “green gold” for more than just Argentines.

That status has attracted some criticism, which highlights the neoliberalism inherent in “exotic” marketing and the dubious ethics of white men profiting off an indigenous commodity. These arguments certainly make a valid point, given the long history of U.S. companies profiting off Latin American commodities. However, they also ignore the lengths companies like Guayaki and Mi Mate have gone to consult the indigenous communities involved, fund projects of actual interest to the community, and pay far more than market price for mate. Indeed, both Argentine and indigenous growers alike appreciate the company’s focus on environmental consciousness and the extra income the mate provides.

The Future of Yerba Mate

With international companies getting on the mate bandwagon, yerba mate looks destined for a future more oriented towards the Global North than the Global South. Overall, it seems that this commodity—previously limited by its small geographic range only in the Southern Cone—will become yet another local taste adapted for a global audience. Two trends will likely continue as mate continues its path of globalization. First, young people will lead the craze as they search for new and exciting beverages that still have social and environmental benefits. Second, straight yerba mate will not be the bestseller; rather, mixed beverages with mate as a component will succeed, since consumers generally dislike its otherwise strong taste.

This increasing global demand for mate will have local benefits, but only if more companies adopt the agroforestry model that Guayakí has adopted and pay attention to local and indigenous concerns. It will be a slow battle—after all, neither shade trees nor mate trees spring up overnight. But with the Atlantic rainforest and local livelihoods at stake, it is a battle worth fighting.

Cover photo by Miguel Mendez, CC-SA-3.0, accessed via Wikimedia Commons.

非洲日益壮大的科学界:一场新的复兴

技术变革一直是发达国家与发展中国家之间最大的鸿沟之一。科学领域的重量级人物通常都包括经济实力最强的国家。据世界银行统计,美国、德国和日本每年在研发 上的投入均占GDP的 2.8% 以上。然而,21 世纪的全球经济发展刺激了全球对研发的投资增加,因为科学技术在全球经济中的重要性日益提升。非洲大陆也开始参与到这个不断发展的领域,与此同时,新的教育基础设施造就了一支有前途的年轻科学劳动力队伍。学术界和工业界的领袖们现在试图动员这些聪明的年轻人,共同发出一个口号:“用非洲解决方案解决非洲问题。”

众所周知,非洲的经济和教育体系面临着历史性的劣势。殖民主义在政治和经济上的衰落使大多数非洲新独立国家陷入剧烈的政治动荡和严重的资源匮乏,几乎无法实现稳定与发展。在 21 世纪,整个非洲大陆的非洲人正在改变这一现状。当代南非艺术家玛丽·西班德、周游世界的尼日利亚演说家兼小说家奇玛曼达·恩戈齐·阿迪奇和百万富翁、苏丹裔英国慈善家穆·易卜拉欣都通过成为众人瞩目的焦点为非洲参与者开辟了空间。这种新的“自己动手”战略无疑在经济和政治上统一了非洲大陆。“非洲问题,非洲方案”最初是与非洲联盟密切相关的口号,现在也适用于科学,在职业发展和教育投资的浪潮中将产业、学术界和政治思想家联系起来。

许多学术界和工业界的领导人采取了国际和集体的方式,力求动员新一代决策者、科学家和工程师解决非洲的问题,从失业和教育不足到迫在眉睫的气候变化和可持续发展危机。几十年来,许多机构和倡议一直致力于非洲大陆科学家和研究人员的培训和专业发展,其中包括撒哈拉以南非洲的非洲科学技术机构网络。非洲大陆的其他此类网络,包括非洲数学科学研究所和非洲基础物理及其应用学院,在非洲各地选拔和培训小班非洲学生,进一步推动了非洲科学资产的集体化。

然而,由于主要期刊读者集中在美国和类似国家,这些新机构的非洲研究人员的工作很少得到全球关注。在这个行业中,专利和出版物不仅使研究领域合法化,而且可以成就或毁掉职业生涯,因此不为人知是职业发展的一大障碍。经济和政治机会最终迫使许多受过培训的人才移民,导致大规模人才流失。

为了阻止训练有素的人才外流,科学界的领导人正在通过行动和言论来推广非洲解决方案。《科学非洲》是一本泛非洲同行评议期刊,于 2018 年推出,旨在传播、推广和突出非洲大陆和世界各地的非洲研究。该杂志由非洲数学科学研究所 (AIMS) 发起的下一个爱因斯坦论坛 (NEF) 出版,该论坛致力于将非洲科学和政策与全球科学界联系起来,特别是通过赋予年轻人权力。最近,该组织协调了非洲科学周,以团结非洲大陆追求科学发展。这些努力继续代表着非洲大陆科学复兴日益增长的能量。用 NEF 的话来说:“我们相信下一个爱因斯坦将是非洲人。”

除了倡导组织,各个领域的专家也谈到了尚未开发的潜力。在 2019 年关于 IPCC 最新全球变暖预测报告的一次采访中,法国国家发展研究所高级研究主任 Arona Diedhiou 指出:“长期以来,国际层面提出的可持续发展方案与非洲当地的现实之间存在着差距……现在是时候进行范式转变,以便提出由非洲人制定的、针对非洲的解决方案。”在描述未来几年非洲大陆面临的气候挑战时,Diedhiou 还强调了教育和激励年轻人的可能性。

虽然这些机构、组织和领导者在专业发展方面发挥了重要作用,但必须考虑它们的缺点。通过延续“最优秀人才”的选拔和提升,这些系统有可能加剧地方人才流失。如果没有同等重要的支持团队、实验室设施和金融投资,训练有素的科学家就无法取得进步;他们可以轻松地搬迁到有这些资源的地方。但并非一切都失去了。在加强研发基础设施方面,非洲科学界有几个意想不到的优势:渴望教育和就业的年轻人群体迅速增长,气候变化的压力需要创新解决方案。

不管世界其他国家怎么看,随着 21 世纪的到来,非洲的科学抱负只会越来越大。尽管非洲大陆的历史中存在着政治和经济上的困难,但非洲领导人正在翻开新的一页。非洲复兴就在眼前。它将带来一个多么奇妙的世界。


Africa’s Growing Scientific Communities: A New Renaissance

Technological change has always been one of the largest dividers between developed and developing countries. The scientific heavy-hitters have traditionally included the most economically powerful nations. According to the World Bank, every year, the United States, Germany, and Japan all spend upwards of 2.8 percent of GDP on research and development. However, global economic development in the 21st century has spurred increased investment in research and development worldwide, as science and technology have become increasingly critical in the global economy. The African continent has begun to take part in this growing sector as well, at a time when new infrastructure in education has created a promising young scientific workforce. Leaders in academia and industry now seek to mobilize these bright young minds towards a unifying rallying cry: “African solutions to African problems.”

The historical disadvantages faced by Africa’s economic and educational systems are well-known. The political and economic withering of colonialism left most of Africa’s new independent states struggling with violent political upheavals and devastating resource scarcities, allowing little stability or development. In the 21st century, Africans across the continent are changing that narrative. Contemporary South African artist Mary Sibande, globetrotting Nigerian speaker and novelist Chimamanda Ngozi Adichie, and millionaire Sudanese-British philanthropist Mo Ibrahim are carving out spaces for African players by claiming the spotlight. This new do-it-yourself strategy has certainly unified the continent economically and politically. Originally a mantra strongly associated with the African Union, “African solutions to African problems” has come to apply to science as well, bridging industries, academics, and political thinkers in a wave of professional development and educational investment.

Many leaders in academia and industry have taken an international and collective approach that seeks to mobilize a new generation of policymakers, scientists, and engineers towards solving Africa’s problems, from unemployment and limited education to the looming crisis of climate change and sustainable development. For decades, numerous institutions and initiatives have been working on the training and professional development of scientists and researchers on the continent, including the network of African Institutions of Science and Technology in Sub-Saharan Africa. Other such networks across the continent, including the African Institute for Mathematical Sciences and the African School of Fundamental Physics and its Applications, select and train small classes of African students at various locations across the continent, furthering the collectivization of Africa’s scientific assets.

However, due to the centralization of major journal readerships in the United States and similar countries, African researchers at these newer institutions receive little global exposure for their work. In an industry where patents and publications not only legitimize research areas but also make or break careers, invisibility is a large obstacle to professional growth. Economic and political opportunity ultimately drive many trained individuals to emigrate, contributing to brain drain on a massive scale.

In order to stem the outflow of trained talent, leaders in the scientific community are promoting the African solutions approach through action as well as rhetoric. The Scientific African, a pan-African peer-reviewed journal, was launched in 2018 to circulate, promote, and highlight African research within the continent and around the world. It is being published by the Next Einstein Forum (NEF), an initiative of the African Institute for Mathematical Sciences (AIMS), which has worked to connect African science and policy with the global scientific community, particularly through the empowerment of young people. Recently, the organization coordinated an Africa Science Week to unify the continent in the pursuit of scientific development. These efforts continue to represent the energy growing on the continent for a scientific renaissance. In the words of the NEF: “We believe the next Einstein will be African.”

In addition to advocacy organizations, experts in various fields speak about untapped potential. In a 2019 interview about the latest IPCC global warming prediction report, Arona Diedhiou, Senior Research Director at the French National Research Institute for Development, argued that “For a long time, there has been a gulf between the sustainable development options suggested at the international level and the local African realities…. It is time for a paradigm shift in order to propose solutions for Africa, developed by Africans.” And describing the climate challenges that face the continent in the coming years, Diedhiou also stressed the possibility of educating and energizing young people.

While these institutions, organizations, and leaders have been instrumental in the growth of professional development, it is important to consider their drawbacks. By perpetuating the selection and elevation of “the best of the best,” these systems have the potential to exacerbate brain drain on a local level. Well-trained scientists cannot achieve progress without comparably valued support teams, laboratory facilities, and financial investments; they can easily relocate to where those resources are available. But all is not lost. To bolster R&D infrastructure, the African scientific community has a couple of unexpected advantages: a rapidly growing population of young people hungry for education and employment, and the pressure of climate change necessitating innovative solutions.

Despite what the rest of the world may believe, Africa’s scientific aspirations are only growing as the 21st century goes on. In spite of the political and economic hardships embedded in the fabric of the continent’s history, leaders are turning the page. The African Renaissance is upon us. What a wondrous world it will bring.

在线客服
在线咨询