Researchers are using statistics and data science tools to accurately and efficiently estimate the number of identified victims killed in the Syrian civil war.

In a new paper, which will appear in the *Annals of Applied Statistics*, scientists report on a four-year effort to combine a data-indexing method called “hashing with statistical estimation.” The new method produces real-time estimates of documented, identified victims with a far lower margin of error than existing statistical methods for finding duplicate records in databases.

“Throwing out duplicate records is easy if all the data are clean—names are complete, spellings are correct, dates are exact, etc.,” says coauthor Beidi Chen, a computer science graduate student at Rice University.

“The war casualty data isn’t like that. People use nicknames. Dates are sometimes included in one database but missing from another. It’s a classic example of what we refer to as a ‘noisy’ dataset. The challenge is finding a way to accurately estimate the number of unique records in spite of this noise.”

Using records from four databases of people killed in the Syrian war, Chen, Duke statistician and machine learning expert Rebecca Steorts and Rice computer scientist Anshumali Shrivastava estimated there were 191,874 unique individuals documented from March 2011 to April 2014.

That’s very close to the estimate of 191,369 compiled in 2014 by HRDAG, a nonprofit organization that helps build scientifically defensible, evidence-based arguments of human rights violations.

But while HRDAG’s estimate relied on the painstaking efforts of human workers to carefully weed out potential duplicate records, hashing with statistical estimation proved to be faster, easier, and less expensive. The researchers say hashing also has the important advantage of a sharp confidence interval: The range of error is plus or minus 1,772, or less than 1 percent of the total number of victims.

“The big win from this method is that we can quickly calculate the probable number of unique elements in a dataset with many duplicates,” says Patrick Ball, HRDAG’s director of research. “We can do a lot with this estimate.”

Shrivastava says the sharpness of the hashing estimate is due to the technique used to index the casualty records. Hashing involves converting a complete data record—a name, date, place of death, and gender in the case of each Syrian war casualty—into one number called a hash.

An algorithm that considers the alphanumeric information in a record produces each hash, and a “hash table,” which works much like the index in a book stores them. The more textual similarity there is between two records, the closer together their hashes are in the table.

“Our method—unique entity estimation—could prove to be useful beyond just the Syrian conflict,” says Steorts, assistant professor of statistical science at Duke.

She says the algorithm and methodology could be used for medical records, official statistics, and industry applications.

“As we collect more and more data, duplication is becoming a more timely and socially important problem,” Steorts says. “Entity resolution problems need to scale to millions and billions of records. Of course, the most accurate way to find duplicate records is having an expert check every record. But this is impossible for large data sets, since the number of pairs that needs to be compared grows dramatically as the number of records increase.”

For example, a record-by-record analysis of all four Syrian war databases would entail some 63 billion paired comparisons, she says.

“If you make assumptions, like dates that are close might be duplicates, you can reduce the number of comparisons that are needed, but every assumption comes with a bias, and ultimately you want an unbiased estimate,” says Shrivastava, assistant professor of computer science at Rice.

“One statistical approach that avoids bias is random sampling. So perhaps choose 1 million random pairs out of the 63 billion, see how many are duplicates and then apply that rate to the entire dataset. This produces an unbiased estimate, which is good, but the likelihood of finding duplicates purely by random is quite low, and that gives a high variance.

##### ‘Deep learning’ goes faster with organized data

“In this case, for example, random sampling could also estimate the documented counts at around 191,000,” he says. “But it couldn’t tell us with any certainty whether the count was 176,000 or 216,000 or some number in between.

“In recent work, my lab has shown that hashing algorithms that were originally designed to do search can also be used as adaptive samplers that precisely mitigate the high variance associated with random sampling,” Shrivastava says.

“Resolving every duplicate seems very appealing, but it is the harder way of estimating the number of unique entities. The new theory of adaptive sampling with hashing allows us to directly estimate unique entity counts efficiently, with high confidence, without resolving the duplicates.”

Shrivastava and Steorts say they are planning future research to apply the hashing technique for unique entity approximation to other types of datasets.

The National Science Foundation and the Air Force Office of Scientific Research funded the work.

*Source: **Rice University*