Maverick Citizen


The Sputnik Vaccine case study (Part Two): How fast is too fast?

Covid gave researchers a sense of urgency. ‘We all wanted a vaccine and so we want to go as fast as we can, but that speed must never come at the expense of the scientific process.’ (Image: Adobe Stock)

In June, Bhekisisa started doing research for an article that would explain how well Russia’s Sputnik V vaccine works and to gauge whether it would be suitable for use in South Africa’s Covid-19 vaccine roll-out. Soon enough, it was realised Sputnik V had flouted the rules of every part of the system that produces scientific research.

This is the second of a four-part series published by Daily Maverick. Part One reveals that Sputnik V is plagued by red flags and question marks surrounding its clinical trials and results. Part Two explains how checkpoints in the scientific system are put in place to ensure that research can be sped up in a safe way. But during the Covid pandemic, the accelerated timeline to push out research in an effort to help has exposed gaps in the process.

“It must be an emergency.”

That’s what most motorists assume when an ambulance zooms by on the highway — and why they’re not that surprised when it happens.

Just the same, it was predictable that research on Covid-19 would stream into the academic publishing sphere when the pandemic started in 2020. People’s lives were at stake.

On the road, speed limits set the parameters of how fast you should drive, but there is a certain speed above the limit that is slightly faster while still being relatively safe.

Excessive speed, however, remains reckless, even for ambulances.

Daredevil drivers are kept in check by traffic cops who help to avoid accidents. Likewise, the peer-review and regulatory review processes in the research world help vet studies that are presented to the community.

The risks of speeding up research without peer-review cops

Covid gave researchers a sense of urgency, says Mitchell Warren, executive director of the US-based advocacy organisation AVAC.

“We all wanted a vaccine and so we want to go as fast as we can, but that speed must never come at the expense of the scientific process.”

You can put pedal to the metal on vaccine research and still ensure the safety of the jabs, Warren says.

But only if the peer-review traffic cops are on duty.

Warren explains: “There is a way to accelerate the process without sacrificing safety, ethics, community engagement and rigorous reviews, both peer-review and regulatory reviews. As long as those things still happen, we should be able to go fast.”

So what happens if the traffic cops are speeding, too?

peer review sputnik v
For the more ‘serious research’, the peer review was sped up drastically with people given between 36 to 48 hours to look through the submission. Given the circumstances, there may have been situations where ‘you find that little bits and pieces slipped through the cracks’. (Photo: Alexandra Beier / Getty Images)

Enter the Sputnik V vaccine.

The jab was approved and used in Russia’s roll-out before its clinical trials were completed. So there were, in fact, no traffic cops on patrol at all.

“They were short-circuiting a process in an incredibly risky way,” Warren says.

But then, The Lancet, one of the world’s most prestigious medical journals, published a state-funded paper on Sputnik’s phase-three trial, granting the jab a certain degree of credibility.

In the warped and accelerated sphere forged by Covid-19, trust in scientists’ integrity (both as producers and reviewers of research) is paramount.

In this second part of our Sputnik V series, we look at the risks of speeding up research without the safeguards of a peer-review cop on duty. It turns out, the traffic cops aren’t that great at stopping accidents, and the situation has worsened during Covid.

What is peer review?

Not all journals carry the same weight, which is why some are seen as more credible than others. In the science world, the biggest and most influential include the New England Journal of Medicine, Science, Nature and The Lancet.

One thing these publications have in common is that they subject paper submissions to a process called peer review. That means that the research is sent to a group of experts in a particular field (their peers) to assess (review) the quality and validity of the work.

Roganie Govender, a speech and language therapist at University College London Hospital, writes in Elsevier that the scrutiny of peers also helps share the ethical responsibility of research which could affect people’s lives. Moreover, she argues, engaging with science as a reviewer helps strengthen your own work and writing.

“It acts as a filter to ensure that only high-quality research is published, especially in reputable journals, by determining the validity, significance and originality of the study,” explain a group of researchers in the Journal of the International Federation of Clinical Chemistry and Laboratory Medicine.

But peer review is far from perfect.

A 2019 study by Elsevier and Sense about Science surveyed more than 3,000 researchers across various disciplines to understand how they felt about peer review. It found 75% of scientists were satisfied with the peer review system, with 90% believing the process helped improve the quality of the papers produced.

But although the researchers surveyed did not want to completely replace the existing system, they did believe it had areas that could be improved.

One aspect that allows for failings is that the whole process of peer review is entirely reliant on people, their expertise and their time.

As deputy editor of the South African Medical Journal, Bridget Farham explains: “Peer review is one of the bottlenecks in publishing because it takes time to get papers that have been submitted, peer-reviewed.”

Typically, Farham says, people get two weeks to go through the paper, but even with that time scientists often need several reminders to complete the review. She also points out that often “people just don’t respond to our requests for peer review”.

The range of experts available to a journal depends on its size and reputation. A smaller journal will have a smaller pool of researchers to call on to assess potential papers for publication, whereas researchers are more eager to do reviews for more “famous” journals and for high-impact publications (where the papers are likely to be shared and cited more often).

How does peer review work?

The formal peer-review process has its flaws, even when there’s no pandemic putting extra pressure on researchers, argues Richard Lessells, an infectious diseases specialist at the KwaZulu-Natal Research Innovation and Sequencing Platform (Krisp).

When scientists submit their work to a journal they will sometimes be asked to suggest a few of their peers to act as reviewers. Lessells explains: “Of course, [researchers] might pick people that they think might look favourably upon [their] science.”

But this is where journal editors come in. Their role is to evaluate the list of suggested researchers and decide whether to approach others, Lessells says. Not all journals allow researchers to suggest reviewers, though.

Peer review takes place for every article published in medical journals around the world — and there are tens of thousands of them. That means researchers’ inboxes are often bursting with requests to review research — for free.

“The reality is that journals sometimes have to send requests to 10 researchers to get one person to accept,” Lessells explains.

Researchers who do accept the offer are doing it for “the good of science” and are signing up to a process that could take hours.

Once a scientist has agreed to review somebody’s research, there are two main things they would look at.

  1. How well is the study designed? 

For this, researchers look at the way the study was set up, ranging from the broader aspects of the paper, such as its structure and arguments, to the finer details of the research methodology, how the participants were chosen and how well the intervention (say, a vaccine) is delivered.

Randomised control trials are considered the gold standard of clinical research. In these types of studies, participants are divided into a group that gets the real intervention and a group that gets a placebo, or dummy drug. This process is “randomised” so that characteristics such as age or sex are equally represented in each group.

When evaluating a randomised control trial, reviewers will assess how well the “blinding” of the study was conducted. That means they’re evaluating how well the researchers ensured that nobody in the study knew which participants received the real intervention and which ones got a placebo.

This is also called the “internal validity” of a study.

  1. How generalisable are the results? 

Researchers also look at “external validity” when assessing a study’s veracity, to help them decide how confidently the research results can be applied to the general population.

Reviewers will look at the population studied in terms of age, race, gender and health status.

A study conducted only in a group of army personnel, for example, will include quite a narrow population group, Lessells explains. They are likely to all be of a specific age and health status that won’t represent the general population because there will be very few older people included in the study.

What are the limitations of peer review?

Peer review assumes that in the realm of science, people will check their personal bias at the door, which may not always be the case. This is particularly true in smaller research communities.

Papers go out blind to reviewers, which means they don’t know who the authors are or which institutions the research is coming from. This is to prevent someone from negatively reviewing a paper based on their feelings about the person behind the research.

But in reality, it’s a bit harder to control.

“Because of the tiny academic community in South Africa, people do actually have a pretty good idea of who they’re reviewing,” says Farham.

“The best thing to do is to try and find an independent commentator, which is obviously going to be more difficult in a small environment like South Africa.”

Ideally, journals would hope that reviewers would decline if they knew there was a conflict with the research or person submitting the paper — for example, if a reviewer had previously published a paper with one of the authors or worked closely with them.

According to Lessells: “There’s only so many people who have the skills and the expertise to review certain science. But if you all know each other, it can become difficult because the journal wants to find people who can be objective and don’t have a conflict of interest.”

Even in these cases, though, Lessells believes it is possible to objectively review a paper in tight-knit scientific communities — but it requires reviewers to thoroughly interrogate papers and focus purely on the science.

“It all comes down to the integrity of the science, the integrity of the authors and the integrity of the reviewers and the editors.”

Things ‘slip through the cracks’ during Covid

In the fast-paced and dynamic space of Covid research, there is often a competitive element among journals themselves.

A health consumer advocate, Hilda Bastian says: “People believe that they should trust those journals, that peer review and editorial control are great and the better and more famous the journal, the higher the quality control.”

Bastian was a founding member of the Cochrane Collaboration, which specialises in systematic reviews of medical research — studies that evaluate all the evidence previous research has found when investigating the same topic.

Journals “also want to rush to print with really high-profile research that is good for their name and gets them a lot of press attention”, which means that they sometimes “cut corners”.

For instance, they may want to be the first to share a breakthrough study or high-profile work, such as a vaccine’s trial results.

Locally, the South African Medical Journal adopted a rapid publication model for Covid submissions, says Farham.

This rush, coupled with constrained resources, sometimes came at the sacrifice of peer review.

Farham explains: “A lot of the trust element came from the fact that the people who were submitting [papers] are extremely well-known clinicians and researchers and academics within the country. They’re the people I would normally be asking to peer-review articles, so in some cases I would waive peer review.”

Farham, however, says this mainly happened with correspondence pieces that contained “snapshots of clinical data” and not with research articles (although the South African Medical Journal does not publish clinical trial data).

For more “serious research”, she says, the peer review was sped up drastically, with people given between 36 and 48 hours to look through the submission. Given the circumstances, there may have been situations where “you find that little bits and pieces slipped through the cracks”. In those scenarios people could submit responding queries in correspondence that the journal would publish.

The issue of things “slipping through the cracks” isn’t limited to smaller journals. Even higher-profile scientific publications have made mistakes in the past two years, with some printing retractions.

One example is The Lancet, the second-ranked general and internal medicine journal in the world.

chloroquine research
An employee checks the production of chloroquine phosphate, the production of which resumed after a 15-year break, at a pharmaceutical company in Nantong City, Jiangsu province, China in 2020. (Photo: EPA-EFE / Xu Congjun)

One of the world’s biggest controversies surrounding Covid treatment in 2020 concerned the antimalarial drug hydroxychloroquine.

On 22 May 2020, The Lancet published a study claiming that hydroxychloroquine or chloroquine could lead to heart damage and endanger patients in hospital. This paper led to the World Health Organization pressing pause on use of the drug in its clinical trial assessing potential Covid-19 treatments.

But two weeks later, the journal retracted the paper.

Bastian says: “When The Lancet publishes these papers [for instance, on Sputnik V and hydroxychloroquine] it appears to have the high-quality stamp on it like it’s gone through the ultimate process. It absolutely hasn’t.”

One of the biggest controversies in the world of Covid treatment last year was the antimalarial drug hydroxychloroquine. (Photo: Christopher Occhicone / Bloomberg)

The Lancet tightened its peer-review process after the retraction was published. The journal now has stricter rules for authors and peer-reviewers. For example, study authors are expected to share a statement that outlines which data will be shared and when. It has to include an assurance that more than one author has verified datasets used in the research.

When a study is based on real-world datasets, peer-review protocols at The Lancet now also require that at least one reviewer sees the full dataset being studied. In addition, very large datasets will be reviewed by a data scientist and reviewers are explicitly asked whether they have doubts about the integrity of the research paper.

The Lancet is not the only prestigious journal that has had to withdraw a Covid-19 study.

Research was also retracted by what is considered the world’s leading medical journal, the New England Journal of Medicine.

On 1 May, the journal published a study that claimed the use of heart medication called ACE inhibitors — which lower blood pressure — could reduce the risk of death among Covid patients. The paper was withdrawn by the authors on 4 June.

In both cases the authors retracted their papers because the underlying data used in the studies were provided by what turned out to be an untrustworthy source and could not be verified.

Reckless driving: The consequences of speeding up Covid science

These two papers were not alone in their mistakes. As amazing as it is to hit warp speed on the production of scientific information, this unprecedented pace also places strain on an already struggling system.

A June paper in BMC Medical Research Methodology identified common reasons behind retractions as errors in the data analysis or methodology of the studies. But there is also an issue with the role of peer review in these times.

The authors point out that in their analysis of eight Covid papers, “one was peer-reviewed in less than three days and the second displayed editorial conflict of interest”.

One of the issues they identified was: “While faster peer-reviewing does not necessarily equate with poorer review quality and faster peer-reviewed time is encouraged at a time of crisis, it remains unclear how thorough the peer reviewing is and how potential conflicts of interest are handled.”

A year before, in June 2020, Nature Human Behaviour published a piece analysing the first four months of Covid publications (from January to April 2020). During this time, an average of 367 Covid articles were published a week, most of which were accepted for publication in under a week.

The study compares this with the rate of publication during the Ebola outbreak, where an average of four articles were published per week, and took just more than two weeks to be accepted. In pre-Covid times, the same journals took just over three months to accept papers.

Moreover, a June 2020 paper published in Scientometrics found that Covid had slashed the publication time for papers tenfold. Worryingly, 11% of the papers assessed had been published in less than 24 hours.

“It is hard to believe that an article can be read by an editor and peer-reviewed properly and published in less than 24 hours,” the paper argues. “It could be said that this time of crisis is revealing a dark side of some journals.”

In their oversight of Covid research, Retraction Watch, an accountability mechanism for journals that highlights flawed science and misconduct, identified 132 retracted papers as of 10 August.

Although 17 of the 132 papers were shared as preprints or in other formats (such as presentations) that had not undergone peer review, 87% of the studies were published in journals that supposedly vetted the work.

Retraction Watch found 13 of the 132 retracted papers (one out of every 10) were published in journals that are regarded among the top 10% of global academic journals by the research database Scopus. Among these journals, five of the retractions were from The Lancet. In 2019, the journal retracted two pieces, one of which was a book review.

Retractions of Covid-19 research in journals were more than three times as high as retractions of research about the H1N1 (swine flu) outbreak in 2009, found a June study published in Accountability in Research.

Where does Sputnik V come in?

The Sputnik V vaccine trials have been afforded credibility by having results published in The Lancet.

Barry Schoub, who heads the ministerial advisory committee on Covid-19 vaccines, argues that there’s only so much journal editors can do with the data provided.

“I would more or less trust the high citation journals such as The Lancet, it’s one of the oldest medical journals. Their scientific integrity is pretty good.”

But even so, the trial is clouded by a litany of concerns and red flags:

  1. The study design wasn’t public;
  2. There wasn’t proper reporting of adverse events or side effects; and
  3. There was a lack of diversity in the trial’s participants.

For more detail on these issues, see Part One of our series. 

So, even without a retraction entering the equation, there are serious questions about the underlying data used in the reported results of the Sputnik study, some of which may have passed through peer review.

The fact that most research papers are not retracted doesn’t mean they’re without error, a 2019 study published in the journal Trials found.

The researchers, led by Bad Science author Ben Goldacre, monitored all the trials reported in five prestigious journals over six weeks. They included the New England Journal of Medicine, The Lancet, Journal of the American Medical Association, British Medical Journal, and Annals of Internal Medicine.

All the publications endorse guidelines designed to combat bad reporting on randomised clinical trials, called Consort (Consolidated Standards of Reporting Trials). 

Using these guidelines, the researchers assessed the published journal reports against the study protocols and entries on trial registries.

They were looking for misreported trials, and that’s just what they found.

Not only did all five journals flout Consort’s rules, they mostly rejected correction letters sent by Goldacre and his colleagues which pointed out the errors.

Nearly 90% of trials assessed required a correction letter, but fewer than half of the letters sent were published by journals — and the corrections that were published only saw the light of day about 99 days later.

An increasing reliance on preprints

At the forefront of all of this is the reminder that peer review isn’t perfect. As with the people upon whom the process relies, it has its flaws.

It’s the simple truth of the matter: people miss things. It doesn’t matter how much time you have or how carefully you pore over the study, sometimes mistakes still slip through.

That’s why more eyes always help.

The more people who are able to look at a paper, assess it and give input, the better.

A May paper in Science Progress notes that the scrutiny of other scientists provides a fail-safe for the review process. The authors write: “The retracted Lancet and [New England Journal of Medicine] papers failed at the peer review level; however, scientific community attention captured the errors in record time.” This is what led to the short time it took for the papers to be retracted.

Under Covid, the publication of preprints has — counterintuitively — allowed for more scrutiny.

Preprints are when scientists choose to share their research, in the form of a draft paper, with the public without it needing to undergo peer review. Often with Covid papers, research is first shared in this way and then submitted to a journal to avoid delays in getting the information out.

Through social media platforms such as Twitter, preprints about Covid treatments and vaccines are put in front of far more expert eyeballs than they otherwise would have been, which can help researchers improve the paper they eventually submit for peer review.

Krisp’s Lessells explains: “It’s clear to me that preprints are a good thing.”

Of course, bad science may be published this way, but Lessells cautions that bad science slips through the formal peer-review process, too.

“The more people see your work, the higher the chances are that bad science will be flagged as such.” DM/MC

This story was produced by the Bhekisisa Centre for Health Journalism. Sign up for the newsletter.


"Information pertaining to Covid-19, vaccines, how to control the spread of the virus and potential treatments is ever-changing. Under the South African Disaster Management Act Regulation 11(5)(c) it is prohibited to publish information through any medium with the intention to deceive people on government measures to address COVID-19. We are therefore disabling the comment section on this article in order to protect both the commenting member and ourselves from potential liability. Should you have additional information that you think we should know, please email [email protected]"