The primary causative organism, Plasmodium falciparum, requires both the Anopheles mosquito and humans to complete its life cycle: sexual reproduction of the protozoan occurs in the mosquito; an immature form is then transmitted to the human via the bite of the mosquito. In a person the parasite goes to the liver, replicates, and moves into the bloodstream, where it attacks red blood cells for their hemoglobin. Some of the plasmodia become sexually mature and are transmitted back to another biting mosquito. Three other Plasmodium species also infect humans.
At the onset of malaria, bouts of chills (ague) and fever lasting several hours and occurring every three or four days are the usual symptoms. If the disease is not treated, the spleen and the liver become enlarged, anemia develops, and jaundice appears. Death may occur from general debility, anemia, or clogging of the vessels of cerebral tissues by affected red blood cells. Cerebral malaria is most commonly seen in infants, pregnant women, and nonimmune travelers to endemic areas.
P. falciparum creates protein knobs on the surfaces of the red blood cells it attacks. These knobs attach the cell to the lining of the blood vessel, preventing its removal to the spleen for destruction. The parasite slows detection by the immune system by changing the makeup of the knobs periodically, substituting or rearranging its 150 "var" (variability) genes, a strategy unique to malaria. A pattern of remission and relapse results as the immune system learns each new "code" only to have it again changed. Patients with malaria gradually do develop immunity that modifies the course of the disease, but this immunity has a degree of strain specificity.
The bark of the cinchona and its product, quinine, have been used in the treatment of malaria for centuries. After World War II, they were largely replaced by the synthetic analog chloroquine. The use of chloroquine, in addition to the use of DDT for mosquito control, was expected to eradicate the disease, but a World Health Organization campaign (1955-69) to eradicate the disease globally (by controlling mosquitoes long enough to allow the human population to become disease free) proved unsuccessful. Despite that, spraying successfully eradicated the disease in some areas (Sardinia, Japan, and Taiwan).
In the 1960s several strains of the malarial parasite developed resistance to chloroquine. This, plus the growing immunity of mosquitoes to insecticides, has caused malaria to become one the of world's leading re-emerging infectious diseases, infecting an estimated 300 million people a year and killing more than a million. Mefloquine may be used in areas where the disease has become highly resistant to chloroquine, but some strains are now resistant to it and other drugs. Artemisinin (derived from sweet wormwood) in combination with other drugs is now in many cases the preferred treat for resistant strains. Amodiaquine in combination with sulfadoxine and pyrimethamine has also been shown to be effective, and malarone (atovaquone and proguanil) also is used for resistant strains. Vaccines against malaria are still experimental. Spraying is still used to control malaria-transmitting mosquitoes, but fish that feed on mosquito larva also have been employed.
Malaria is one of the most common infectious diseases and an enormous public health problem. The disease is caused by protozoan parasites of the genus Plasmodium. Only four types of the plasmodium parasite can infect humans; the most serious forms of the disease are caused by Plasmodium falciparum and Plasmodium vivax, but other related species (Plasmodium ovale, Plasmodium malariae) can also affect humans. This group of human-pathogenic Plasmodium species is usually referred to as malaria parasites.
Malaria parasites are transmitted by female Anopheles mosquitoes. The parasites multiply within red blood cells, causing symptoms that include symptoms of anemia (light headedness, shortness of breath, tachycardia etc.), as well as other general symptoms such as fever, chills, nausea, flu-like illness, and in severe cases, coma and death. Malaria transmission can be reduced by preventing mosquito bites with mosquito nets and insect repellents, or by mosquito control measures such as spraying insecticides inside houses and draining standing water where mosquitoes lay their eggs.
Although some are under development, no vaccine is currently available for malaria; preventive drugs must be taken continuously to reduce the risk of infection. These prophylactic drug treatments are often too expensive for most people living in endemic areas. Most adults from endemic areas have a degree of long-term infection, which tends to recur and also possess partial immunity (resistance); the resistance reduces with time and such adults may become susceptible to severe malaria if they have spent a significant amount of time in non-endemic areas. They are strongly recommended to take full precautions if they return to an endemic area. Malaria infections are treated through the use of antimalarial drugs, such as quinine or artemisinin derivatives, although drug resistance is increasingly common.
Malaria has infected humans for over 50,000 years, and may have been a human pathogen for the entire history of our species. Indeed, close relatives of the human malaria parasites remain common in chimpanzees, our closest relatives. References to the unique periodic fevers of malaria are found throughout recorded history, beginning in 2700 BC in China. The term malaria originates from Medieval Italian: mala aria — "bad air"; and the disease was formerly called ague or marsh fever due to its association with swamps and marsh land.
Scientific studies on malaria made their first significant advance in 1880, when a French army doctor working in the military hospital of Constantine in Algeria named Charles Louis Alphonse Laveran observed parasites for the first time, inside the red blood cells of people suffering from malaria. He therefore proposed that malaria was caused by this protozoan, the first time protozoa were identified as causing disease. For this and later discoveries, he was awarded the 1907 Nobel Prize for Physiology or Medicine. The protozoan was called Plasmodium by the Italian scientists Ettore Marchiafava and Angelo Celli. A year later, Carlos Finlay, a Cuban doctor treating patients with yellow fever in Havana, provided strong evidence that mosquitoes were transmitting disease to and from humans. This work followed earlier suggestions by Josiah C. Nott, and work by Patrick Manson on the transmission of filariasis.
However, it was Britain's Sir Ronald Ross working in the Presidency General Hospital in Calcutta who finally proved in 1898 that malaria is transmitted by mosquitoes. He did this by showing that certain mosquito species transmit malaria to birds and isolating malaria parasites from the salivary glands of mosquitoes that had fed on infected birds. For this work Ross received the 1902 Nobel Prize in Medicine. After resigning from the Indian Medical Service, Ross worked at the newly-established Liverpool School of Tropical Medicine and directed malaria-control efforts in Egypt, Panama, Greece and Mauritius. The findings of Finlay and Ross were later confirmed by a medical board headed by Walter Reed in 1900, and its recommendations implemented by William C. Gorgas in the health measures undertaken during construction of the Panama Canal. This public-health work saved the lives of thousands of workers and helped develop the methods used in future public-health campaigns against this disease.
The first effective treatment for malaria was the bark of cinchona tree, which contains quinine. This tree grows on the slopes of the Andes, mainly in Peru. This natural product was used by the inhabitants of Peru to control malaria, and the Jesuits introduced this practice to Europe during the 1640s where it was rapidly accepted. However, it was not until 1820 that the active ingredient, quinine, was extracted from the bark, isolated and named by the French chemists Pierre Joseph Pelletier and Joseph Bienaimé Caventou.
In the early twentieth century, before antibiotics, patients with syphilis were intentionally infected with malaria to create a fever, following the work of Julius Wagner-Jauregg. By accurately controlling the fever with quinine, the effects of both syphilis and malaria could be minimized. Although some patients died from malaria, this was preferable to the almost-certain death from syphilis.
Although the blood stage and mosquito stages of the malaria life cycle were identified in the 19th and early 20th centuries, it was not until the 1980s that the latent liver form of the parasite was observed. The discovery of this latent form of the parasite finally explained why people could appear to be cured of malaria but still relapse years after the parasite had disappeared from their bloodstreams.
Malaria causes about 250 million cases of fever and approximately one million deaths annually. The vast majority of cases occur in children under the age of 5 years; pregnant women are also especially vulnerable. Despite efforts to reduce transmission and increase treatment, there has been little change in which areas are at risk of this disease since 1992. Indeed, if the prevalence of malaria stays on its present upwards course, the death rate could double in the next twenty years. Precise statistics are unknown because many cases occur in rural areas where people do not have access to hospitals or the means to afford health care. Consequently, the majority of cases are undocumented.
Although co-infection with HIV and malaria does cause increased mortality, this is less of a problem than with HIV/tuberculosis co-infection, due to the two diseases usually attacking different age-ranges, with malaria being most common in the young and active tuberculosis most common in the old. Although HIV/malaria co-infection produces less severe symptoms than the interaction between HIV and TB, HIV and malaria do contribute to each other's spread. This effect comes from malaria increasing viral load and HIV infection increasing a person's susceptibility to malaria infection.
Malaria is presently endemic in a broad band around the equator, in areas of the Americas, many parts of Asia, and much of Africa; however, it is in sub-Saharan Africa where 85– 90% of malaria fatalities occur. The geographic distribution of malaria within large regions is complex, and malaria-afflicted and malaria-free areas are often found close to each other. In drier areas, outbreaks of malaria can be predicted with reasonable accuracy by mapping rainfall. Malaria is more common in rural areas than in cities; this is in contrast to dengue fever where urban areas present the greater risk. For example, the cities of Vietnam, Laos and Cambodia are essentially malaria-free, but the disease is present in many rural regions. By contrast, in Africa malaria is present in both rural and urban areas, though the risk is lower in the larger cities. The global endemic levels of malaria have not been mapped since the 1960s. However, the Wellcome Trust, UK, has funded the Malaria Atlas Project to rectify this, providing a more contemporary and robust means with which to assess current and future malaria disease burden.
Malaria is not just a disease commonly associated with poverty, but is also a cause of poverty and a major hindrance to economic development. The disease has been associated with major negative economic effects on regions where it is widespread. A comparison of average per capita GDP in 1995, adjusted to give parity of purchasing power, between malarious and non-malarious countries demonstrates a fivefold difference ($1,526 USD versus $8,268 USD). Moreover, in countries where malaria is common, average per capita GDP has risen (between 1965 and 1990) only 0.4% per year, compared to 2.4% per year in other countries. However, correlation does not demonstrate causation, and the prevalence is at least partly because these regions do not have the financial capacities to prevent malaria. In its entirety, the economic impact of malaria has been estimated to cost Africa $12 billion USD every year. The economic impact includes costs of health care, working days lost due to sickness, days lost in education, decreased productivity due to brain damage from cerebral malaria, and loss of investment and tourism. In some countries with a heavy malaria burden, the disease may account for as much as 40% of public health expenditure, 30-50% of inpatient admissions, and up to 50% of outpatient visits.
Severe malaria is almost exclusively caused by P. falciparum infection and usually arises 6-14 days after infection. Consequences of severe malaria include coma and death if untreated—young children and pregnant women are especially vulnerable. Splenomegaly (enlarged spleen), severe headache, cerebral ischemia, hepatomegaly (enlarged liver), hypoglycemia, and hemoglobinuria with renal failure may occur. Renal failure may cause blackwater fever, where hemoglobin from lysed red blood cells leaks into the urine. Severe malaria can progress extremely rapidly and cause death within hours or days. In the most severe cases of the disease fatality rates can exceed 20%, even with intensive care and treatment. In endemic areas, treatment is often less satisfactory and the overall fatality rate for all cases of malaria can be as high as one in ten. Over the longer term, developmental impairments have been documented in children who have suffered episodes of severe malaria.
Chronic malaria is seen in both P. vivax and P. ovale, but not in P. falciparum. Here, the disease can relapse months or years after exposure, due to the presence of latent parasites in the liver. Describing a case of malaria as cured by observing the disappearance of parasites from the bloodstream can therefore be deceptive. The longest incubation period reported for a P. vivax infection is 30 years. Approximately one in five of P. vivax malaria cases in temperate areas involve overwintering by hypnozoites (i.e., relapses begin the year after the mosquito bite).
Malaria is caused by protozoan parasites of the genus Plasmodium (phylum Apicomplexa). In humans malaria is caused by P. falciparum, P. malariae, P. ovale, P. vivax and P. knowlesi. P. falciparum is the most common cause of infection and is responsible for about 80% of all malaria cases, and is also responsible for about 90% of the deaths from malaria. Parasitic Plasmodium species also infect birds, reptiles, monkeys, chimpanzees and rodents. There have been documented human infections with several simian species of malaria, namely P. knowlesi, P. inui, P. cynomolgi, P. simiovale, P. brazilianum, P. schwetzi and P. simium; however, with the exception of P. knowlesi, these are mostly of limited public health importance. Although avian malaria can kill chickens and turkeys, this disease does not cause serious economic losses to poultry farmers. However, since being accidentally introduced by humans it has decimated the endemic birds of Hawaii, which evolved in its absence and lack any resistance to it.
The parasite's primary (definitive) hosts and transmission vectors are female mosquitoes of the Anopheles genus. Young mosquitoes first ingest the malaria parasite by feeding on an infected human carrier and the infected Anopheles mosquitoes carry Plasmodium sporozoites in their salivary glands. A mosquito becomes infected when it takes a blood meal from an infected human. Once ingested, the parasite gametocytes taken up in the blood will further differentiate into male or female gametes and then fuse in the mosquito gut. This produces an ookinete that penetrates the gut lining and produces an oocyst in the gut wall. When the oocyst ruptures, it releases sporozoites that migrate through the mosquito's body to the salivary glands, where they are then ready to infect a new human host. This type of transmission is occasionally referred to as anterior station transfer. The sporozoites are injected into the skin, alongside saliva, when the mosquito takes a subsequent blood meal.
Only female mosquitoes feed on blood, thus males do not transmit the disease. The females of the Anopheles genus of mosquito prefer to feed at night. They usually start searching for a meal at dusk, and will continue throughout the night until taking a meal. Malaria parasites can also be transmitted by blood transfusions, although this is rare.
Malaria in humans develops via two phases: an exoerythrocytic and an erythrocytic phase. The exoerythrocytic phase involves infection of the hepatic system, or liver, while the erythrocytic phase involves infection of the erythrocytes, or red blood cells. When an infected mosquito pierces a person's skin to take a blood meal, sporozoites in the mosquito's saliva enter the bloodstream and migrate to the liver. Within 30 minutes of being introduced into the human host, they infect hepatocytes, multiplying asexually and asymptomatically for a period of 6–15 days. Once in the liver these organisms differentiate to yield thousands of merozoites which, following rupture of their host cells, escape into the blood and infect red blood cells, thus beginning the erythrocytic stage of the life cycle. The parasite escapes from the liver undetected by wrapping itself in the cell membrane of the infected host liver cell.
Within the red blood cells the parasites multiply further, again asexually, periodically breaking out of their hosts to invade fresh red blood cells. Several such amplification cycles occur. Thus, classical descriptions of waves of fever arise from simultaneous waves of merozoites escaping and infecting red blood cells.
Some P. vivax and P. ovale sporozoites do not immediately develop into exoerythrocytic-phase merozoites, but instead produce hypnozoites that remain dormant for periods ranging from several months (6–12 months is typical) to as long as three years. After a period of dormancy, they reactivate and produce merozoites. Hypnozoites are responsible for long incubation and late relapses in these two species of malaria.
The parasite is relatively protected from attack by the body's immune system because for most of its human life cycle it resides within the liver and blood cells and is relatively invisible to immune surveillance. However, circulating infected blood cells are destroyed in the spleen. To avoid this fate, the P. falciparum parasite displays adhesive proteins on the surface of the infected blood cells, causing the blood cells to stick to the walls of small blood vessels, thereby sequestering the parasite from passage through the general circulation and the spleen. This "stickiness" is the main factor giving rise to hemorrhagic complications of malaria. High endothelial venules (the smallest branches of the circulatory system) can be blocked by the attachment of masses of these infected red blood cells. The blockage of these vessels causes symptoms such as in placental and cerebral malaria. In cerebral malaria the sequestrated red blood cells can breach the blood brain barrier possibly leading to coma.
Although the red blood cell surface adhesive proteins (called PfEMP1, for Plasmodium falciparum erythrocyte membrane protein 1) are exposed to the immune system they do not serve as good immune targets because of their extreme diversity; there are at least 60 variations of the protein within a single parasite and perhaps limitless versions within parasite populations. Like a thief changing disguises or a spy with multiple passports, the parasite switches between a broad repertoire of PfEMP1 surface proteins, thus staying one step ahead of the pursuing immune system.
Some merozoites turn into male and female gametocytes. If a mosquito pierces the skin of an infected person, it potentially picks up gametocytes within the blood. Fertilization and sexual recombination of the parasite occurs in the mosquito's gut, thereby defining the mosquito as the definitive host of the disease. New sporozoites develop and travel to the mosquito's salivary gland, completing the cycle. Pregnant women are especially attractive to the mosquitoes, and malaria in pregnant women is an important cause of stillbirths, infant mortality and low birth weight, particularly in P. falciparum infection, but also in other species infection, such as P. vivax.
The best-studied influence of the malaria parasite upon the human genome is the blood disease, sickle-cell disease. In sickle-cell disease, there is a mutation in the HBB gene, which encodes the beta globin subunit of haemoglobin. The normal allele encodes a glutamate at position six of the beta globin protein, while the sickle-cell allele encodes a valine. This change from a hydrophilic to a hydrophobic amino acid encourages binding between haemoglobin molecules, with polymerization of haemoglobin deforming red blood cells into a "sickle" shape. Such deformed cells are cleared rapidly from the blood, mainly in the spleen, for destruction and recycling.
In the merozoite stage of its life cycle the malaria parasite lives inside red blood cells, and its metabolism changes the internal chemistry of the red blood cell. Infected cells normally survive until the parasite reproduces, but if the red cell contains a mixture of sickle and normal haemoglobin, it is likely to become deformed and be destroyed before the daughter parasites emerge. Thus, individuals heterozygous for the mutated allele, known as sickle-cell trait, may have a low and usually unimportant level of anaemia, but also have a greatly reduced chance of serious malaria infection. This is a classic example of heterozygote advantage.
Individuals homozygous for the mutation have full sickle-cell disease and in traditional societies rarely live beyond adolescence. However, in populations where malaria is endemic, the frequency of sickle-cell genes is around 10%. The existence of four haplotypes of sickle-type hemoglobin suggests that this mutation has emerged independently at least four times in malaria-endemic areas, further demonstrating its evolutionary advantage in such affected regions. There are also other mutations of the HBB gene that produce haemoglobin molecules capable of conferring similar resistance to malaria infection. These mutations produce haemoglobin types HbE and HbC which are common in Southeast Asia and Western Africa, respectively.
Severe malaria is commonly misdiagnosed in Africa, leading to a failure to treat other life-threatening illnesses. In malaria-endemic areas, parasitemia does not ensure a diagnosis of severe malaria because parasitemia can be incidental to other concurrent disease. Recent investigations suggest that malarial retinopathy is better (collective sensitivity of 95% and specificity of 90%) than any other clinical or laboratory feature in distinguishing malarial from non-malarial coma.
From the thick film, an experienced microscopist can detect parasite levels (or parasitemia) down to as low as 0.0000001% of red blood cells. Diagnosis of species can be difficult because the early trophozoites ("ring form") of all four species look identical and it is never possible to diagnose species on the basis of a single ring form; species identification is always based on several trophozoites.
There are several families of drugs used to treat malaria. Chloroquine is very cheap and, until recently, was very effective, which made it the antimalarial drug of choice for many years in most parts of the world. However, resistance of Plasmodium falciparum to chloroquine has spread recently from Asia to Africa, making the drug ineffective against the most dangerous Plasmodium strain in many affected regions of the world. In those areas where chloroquine is still effective it remains the first choice. Unfortunately, chloroquine-resistance is associated with reduced sensitivity to other drugs such as quinine and amodiaquine.
There are several other substances which are used for treatment and, partially, for prevention (prophylaxis). Many drugs may be used for both purposes; larger doses are used to treat cases of malaria. Their deployment depends mainly on the frequency of resistant parasites in the area where the drug is used. One drug currently being investigated for possible use as an anti-malarial, especially for treatment of drug-resistant strains, is the beta blocker propranolol. Propranolol has been shown to block both Plasmodium's ability to enter red blood cell and establish an infection, as well as parasite replication. A December 2006 study by Northwestern University researchers suggested that propranolol may reduce the dosages required for existing drugs to be effective against P. falciparum by 5- to 10-fold, suggesting a role in combination therapies.
Currently available anti-malarial drugs include:
The development of drugs was facilitated when Plasmodium falciparum was successfully cultured. This allowed in vitro testing of new drug candidates.
Extracts of the plant Artemisia annua, containing the compound artemisinin or semi-synthetic derivatives (a substance unrelated to quinine), offer over 90% efficacy rates, but their supply is not meeting demand. In 2007, the Bill & Melinda Gates Foundation contributed $13.6m to support research at the University of York to develop fast and high-yield strains of artemisia, with researchers predicting an increase in yield of up to 1000% compared to current varieties. One study in Rwanda showed that children with uncomplicated P. falciparum malaria demonstrated fewer clinical and parasitological failures on post-treatment day 28 when amodiaquine was combined with artesunate, rather than administered alone (OR = 0.34). However, increased resistance to amodiaquine during this study period was also noted.
Since 2001 the World Health Organization has recommended using artemisinin-based combination therapy (ACT) as first-line treatment for uncomplicated malaria in areas experiencing resistance to older medications. The most recent WHO treatment guidelines for malaria recommend four different ACTs. While numerous countries, including most African nations, have adopted the change in their official malaria treatment policies, cost remains a major barrier to ACT implementation. Because ACTs cost up to twenty times as much as older medications, they remain unaffordable in many malaria-endemic countries. The molecular target of artemisinin is controversial, although recent studies suggest that SERCA, a calcium pump in the endoplasmic reticulum may be associated with artemisinin resistance. Malaria parasites can develop resistance to artemisinin and resistance can be produced by mutation of SERCA. However, other studies suggest the mitochondrion is the major target for artemisinin and its analogs.
In February 2002, the journal Science and other press outlets announced progress on a new treatment for infected individuals. A team of French and South African researchers had identified a new drug they were calling "G25". It cured malaria in test primates by blocking the ability of the parasite to copy itself within the red blood cells of its victims. In 2005 the same team of researchers published their research on achieving an oral form, which they refer to as "TE3" or "te3". As of early 2006, there is no information in the mainstream press as to when this family of drugs will become commercially available.
In 1996, Professor Geoff McFadden stumbled upon the work of British biologist Ian Wilson, who had discovered that the plasmodia responsible for causing malaria retained parts of chloroplasts, an organelle usually found in plants, complete with their own functioning genomes. This led Professor McFadden to the realisation that any number of herbicides may in fact be successful in the fight against malaria, and so he set about trialing large numbers of them, and enjoyed a 75% success rate.
These "apicoplasts" are thought to have originated through the endosymbiosis of algae and play a crucial role in fatty acid bio-synthesis in plasmodia. To date, 466 proteins have been found to be produced by apicoplasts and these are now being looked at as possible targets for novel anti-malarial drugs.
Although effective anti-malarial drugs are on the market, the disease remains a threat to people living in endemic areas who have no proper and prompt access to effective drugs. Access to pharmacies and health facilities, as well as drug costs, are major obstacles. Médecins Sans Frontières estimates that the cost of treating a malaria-infected person in an endemic country was between US$0.25 and $2.40 per dose in 2002.
Methods used to prevent the spread of disease, or to protect individuals in areas where malaria is endemic, include prophylactic drugs, mosquito eradication, and the prevention of mosquito bites. The continued existence of malaria in an area requires a combination of high human population density, high mosquito population density, and high rates of transmission from humans to mosquitoes and from mosquitoes to humans. If any of these is lowered sufficiently, the parasite will sooner or later disappear from that area, as happened in North America, Europe and much of Middle East. However, unless the parasite is eliminated from the whole world, it could become re-established if conditions revert to a combination that favors the parasite's reproduction. Many countries are seeing an increasing number of imported malaria cases due to extensive travel and migration. (See Anopheles.)
There is currently no vaccine that will prevent malaria, but this is an active field of research.
Many researchers argue that prevention of malaria may be more cost-effective than treatment of the disease in the long run, but the capital costs required are out of reach of many of the world's poorest people. Economic adviser Jeffrey Sachs estimates that malaria can be controlled for US$3 billion in aid per year. It has been argued that, in order to meet the Millennium Development Goals, money should be redirected from HIV/AIDS treatment to malaria prevention, which for the same amount of money would provide greater benefit to African economies.
Brazil, Eritrea, India, and Vietnam have, unlike many other developing nations, successfully reduced the malaria burden. Common success factors included conducive country conditions, a targeted technical approach using a package of effective tools, data-driven decision-making, active leadership at all levels of government, involvement of communities, decentralized implementation and control of finances, skilled technical and managerial capacity at national and sub-national levels, hands-on technical and programmatic support from partner agencies, and sufficient and flexible financing.
Before DDT, malaria was successfully eradicated or controlled also in several tropical areas by removing or poisoning the breeding grounds of the mosquitoes or the aquatic habitats of the larva stages, for example by filling or applying oil to places with standing water. These methods have seen little application in Africa for more than half a century.
Efforts to eradicate malaria by eliminating mosquitoes have been successful in some areas. Malaria was once common in the United States and southern Europe, but the draining of wetland breeding grounds and better sanitation, in conjunction with the monitoring and treatment of infected humans, eliminated it from affluent regions. In 2002, there were 1,059 cases of malaria reported in the US, including eight deaths. In five of those cases, the disease was contracted in the United States. Malaria was eliminated from the northern parts of the USA in the early twentieth century, and the use of the pesticide DDT eliminated it from the South by 1951. In the 1950s and 1960s, there was a major public health effort to eradicate malaria worldwide by selectively targeting mosquitoes in areas where malaria was rampant. However, these efforts have so far failed to eradicate malaria in many parts of the developing world - the problem is most prevalent in Africa.
Sterile insect technique is emerging as a potential mosquito control method. Progress towards transgenic, or genetically modified, insects suggest that wild mosquito populations could be made malaria-resistant. Researchers at Imperial College London created the world's first transgenic malaria mosquito, with the first plasmodium-resistant species announced by a team at Case Western Reserve University in Ohio in 2002. Successful replacement of existent populations with genetically modified populations, relies upon a drive mechanism, such as transposable elements to allow for non-Mendelian inheritance of the gene of interest.
On December 21, 2007, a study published in PLoS Pathogens found that the hemolytic C-type lectin CEL-III from Cucumaria echinata, a sea cucumber found in the Bay of Bengal, impaired the development of the malaria parasite when produced by transgenic mosquitoes. This could potentially be used one day to control malaria by using genetically modified mosquitoes refractory to the parasites, although the authors of the study recognize that there are numerous scientific and ethical problems to be overcome before such a control strategy could be implemented.
Several drugs, most of which are also used for treatment of malaria, can be taken preventively. Generally, these drugs are taken daily or weekly, at a lower dose than would be used for treatment of a person who had actually contracted the disease. Use of prophylactic drugs is seldom practical for full-time residents of malaria-endemic areas, and their use is usually restricted to short-term visitors and travelers to malarial regions. This is due to the cost of purchasing the drugs, negative side effects from long-term use, and because some effective anti-malarial drugs are difficult to obtain outside of wealthy nations.
Quinine was used starting in the seventeenth century as a prophylactic against malaria. The development of more effective alternatives such as quinacrine, chloroquine, and primaquine in the twentieth century reduced the reliance on quinine. Today, quinine is still used to treat chloroquine resistant Plasmodium falciparum, as well as severe and cerebral stages of malaria, but is not generally used for prophylaxis. Of interesting historical note is the observation by Samuel Hahnemann in the late 18th century that over-dosing of quinine leads to a symptomatic state very similar to that of malaria itself. This lead Hahnemann to develop the medical Law of Similars, and the subsequent medical system of Homeopathy.
Modern drugs used preventively include mefloquine (Lariam), doxycycline (available generically), and the combination of atovaquone and proguanil hydrochloride (Malarone). The choice of which drug to use depends on which drugs the parasites in the area are resistant to, as well as side-effects and other considerations. The prophylactic effect does not begin immediately upon starting taking the drugs, so people temporarily visiting malaria-endemic areas usually begin taking the drugs one to two weeks before arriving and must continue taking them for 4 weeks after leaving (with the exception of atovaquone proguanil that only needs be started 2 days prior and continued for 7 days afterwards).
The first and historically the most popular insecticide used for IRS is DDT. While it was initially used to exclusively to combat malaria, its use quickly spread to agriculture. In time, pest-control, rather than disease-control, came to dominate DDT use, and this large-scale agricultural use led to the evolution of resistant mosquitoes in many regions. During the 1960s, awareness of the negative consequences of its indiscriminate use increased ultimately leading to bans on agricultural applications of DDT in many countries in the 1970s.
Though DDT has never been banned for use in malaria control and there are several other insecticides suitable for IRS, some advocates have claimed that bans are responsible for tens of millions of deaths in tropical countries where DDT had once been effective in controlling malaria. Furthermore, most of the problems associated with DDT use stem specifically from its industrial-scale application in agriculture, rather than its use in public health.
The World Health Organization (WHO) currently advises the use of 12 different insecticides in IRS operations. These include DDT and a series of alternative insecticides (such as the pyrethroids permethrin and deltamethrin) to both combat malaria in areas where mosquitoes are DDT-resistant, and to slow the evolution of resistance. This public health use of small amounts of DDT is permitted under the Stockholm Convention on Persistent Organic Pollutants (POPs), which prohibits the agricultural use of DDT. However, because of its legacy, many developed countries discourage DDT use even in small quantities.
The distribution of mosquito nets impregnated with insecticide (often permethrin or deltamethrin) has been shown to be an extremely effective method of malaria prevention, and it is also one of the most cost-effective methods of prevention. These nets can often be obtained for around US$2.50 - $3.50 (2-3 euro) from the United Nations, the World Health Organization, and others.
For maximum effectiveness, the nets should be re-impregnated with insecticide every six months. This process poses a significant logistical problem in rural areas. New technologies like Olyset or DawaPlus allow for production of long-lasting insecticidal mosquito nets (LLINs), which release insecticide for approximately 5 years, and cost about US$5.50. ITNs have the advantage of protecting people sleeping under the net and simultaneously killing mosquitoes that contact the net. This has the effect of killing the most dangerous mosquitoes. Some protection is also provided to others, including people sleeping in the same room but not under the net.
Unfortunately, the cost of treating malaria is high relative to income, and the illness results in lost wages. Consequently, the financial burden means that the cost of a mosquito net is often unaffordable to people in developing countries, especially for those most at risk. Only 1 out of 20 people in Africa own a bed net. Although shipped into Africa mainly from Europe as free development help, the nets quickly become expensive trade goods. They are mainly used for fishing, and by combining hundreds of donated mosquito nets, whole river sections can be completely shut off, catching even the smallest fish.
A study among Afghan refugees in Pakistan found that treating top-sheets and chaddars (head coverings) with permethrin has similar effectiveness to using a treated net, but is much cheaper.
A new approach, announced in Science on June 10, 2005, uses spores of the fungus Beauveria bassiana, sprayed on walls and bed nets, to kill mosquitoes. While some mosquitoes have developed resistance to chemicals, they have not been found to develop a resistance to fungal infections.
Vaccines for malaria are under development, with no completely effective vaccine yet available. The first promising studies demonstrating the potential for a malaria vaccine were performed in 1967 by immunizing mice with live, radiation-attenuated sporozoites, providing protection to about 60% of the mice upon subsequent injection with normal, viable sporozoites. Since the 1970s, there has been a considerable effort to develop similar vaccination strategies within humans. It was determined that an individual can be protected from a P. falciparum infection if they receive over 1000 bites from infected, irradiated mosquitoes.
It has been generally accepted that it is impractical to provide at-risk individuals with this vaccination strategy, but that has been recently challenged with work being done by Dr. Stephen Hoffman of Sanaria, one of the key researchers who originally sequenced the genome of Plasmodium falciparum. His work most recently has revolved around solving the logistical problem of isolating and preparing the parasites equivalent to a 1000 irradiated mosquitoes for mass storage and inoculation of human beings. The company has recently received several multi-million dollar grants from the Bill & Melinda Gates Foundation and the U.S. government to begin early clinical studies in 2007 and 2008. The Seattle Biomedical Research Institute (SBRI), funded by the Malaria Vaccine Initiative, assures potential volunteers that "the  clinical trials won't be a life-threatening experience. While many volunteers [in Seattle] will actually contract malaria, the cloned strain used in the experiments can be quickly cured, and does not cause a recurring form of the disease." "Some participants will get experimental drugs or vaccines, while others will get placebo.
Instead, much work has been performed to try and understand the immunological processes that provide protection after immunization with irradiated sporozoites. After the mouse vaccination study in 1967, it was hypothesized that the injected sporozoites themselves were being recognized by the immune system, which was in turn creating antibodies against the parasite. It was determined that the immune system was creating antibodies against the circumsporozoite protein (CSP) which coated the sporozoite. Moreover, antibodies against CSP prevented the sporozoite from invading hepatocytes. CSP was therefore chosen as the most promising protein on which to develop a vaccine against the malaria sporozoite. It is for these historical reasons that vaccines based on CSP are the most numerous of all malaria vaccines.
Presently, there is a huge variety of vaccine candidates on the table. Pre-erythrocytic vaccines (vaccines that target the parasite before it reaches the blood), in particular vaccines based on CSP, make up the largest group of research for the malaria vaccine. Other vaccine candidates include: those that seek to induce immunity to the blood stages of the infection; those that seek to avoid more severe pathologies of malaria by preventing adherence of the parasite to blood venules and placenta; and transmission-blocking vaccines that would stop the development of the parasite in the mosquito right after the mosquito has taken a bloodmeal from an infected person. It is hoped that the sequencing of the P. falciparum genome will provide targets for new drugs or vaccines.
The first vaccine developed that has undergone field trials, is the SPf66, developed by Manuel Elkin Patarroyo in 1987. It presents a combination of antigens from the sporozoite (using CS repeats) and merozoite parasites. During phase I trials a 75% efficacy rate was demonstrated and the vaccine appeared to be well tolerated by subjects and immunogenic. The phase IIb and III trials were less promising, with the efficacy falling to between 38.8% and 60.2%. A trial was carried out in Tanzania in 1993 demonstrating the efficacy to be 31% after a years follow up, however the most recent (though controversial) study in the Gambia did not show any effect. Despite the relatively long trial periods and the number of studies carried out, it is still not known how the SPf66 vaccine confers immunity; it therefore remains an unlikely solution to malaria. The CSP was the next vaccine developed that initially appeared promising enough to undergo trials. It is also based on the circumsporoziote protein, but additionally has the recombinant (Asn-Ala-Pro15Asn-Val-Asp-Pro)2-Leu-Arg(R32LR) protein covalently bound to a purified Pseudomonas aeruginosa toxin (A9). However at an early stage a complete lack of protective immunity was demonstrated in those inoculated. The study group used in Kenya had an 82% incidence of parasitaemia whilst the control group only had an 89% incidence. The vaccine intended to cause an increased T-lymphocyte response in those exposed, this was also not observed.
The efficacy of Patarroyo's vaccine has been disputed with some US scientists concluding in The Lancet (1997) that "the vaccine was not effective and should be dropped" while the Colombian accused them of "arrogance" putting down their assertions to the fact that he came from a developing country.
The RTS,S/AS02A vaccine is the candidate furthest along in vaccine trials. It is being developed by a partnership between the PATH Malaria Vaccine Initiative (a grantee of the Gates Foundation), the pharmaceutical company, GlaxoSmithKline, and the Walter Reed Army Institute of Research In the vaccine, a portion of CSP has been fused to the immunogenic "S antigen" of the hepatitis B virus; this recombinant protein is injected alongside the potent AS02A adjuvant. In October 2004, the RTS,S/AS02A researchers announced results of a Phase IIb trial, indicating the vaccine reduced infection risk by approximately 30% and severity of infection by over 50%. The study looked at over 2,000 Mozambican children. More recent testing of the RTS,S/AS02A vaccine has focused on the safety and efficacy of administering it earlier in infancy: In October 2007, the researchers announced results of a phase I/IIb trial conducted on 214 Mozambican infants between the ages of 10 and 18 months in which the full three-dose course of the vaccine led to a 62% reduction of infection with no serious side-effects save some pain at the point of injection. Further research will delay this vaccine from commercial release until around 2011.
The Malaria Control Project is currently using downtime computing power donated by individual volunteers around the world (see Volunteer computing and BOINC) to simulate models of the health effects and transmission dynamics in order to find the best method or combination of methods for malaria control. This modeling is extremely computer intensive due to the simulations of large human populations with a vast range of parameters related to biological and social factors that influence the spread of the disease. It is expected to take a few months using volunteered computing power compared to the 40 years it would have taken with the current resources available to the scientists who developed the program.
An example of the importance of computer modeling in planning malaria eradication programs is shown in the paper by Águas and others. They showed that eradication of malaria is crucially dependent on finding and treating the large number of people in endemic areas with asymptomatic malaria, who act as a reservoir for infection. The malaria parasites do not affect animal species and therefore eradication of the disease from the human population would be expected to be effective.