Literature DB >> 30530688

Newcomb-Benford law and the detection of frauds in international trade.

Andrea Cerioli1, Lucio Barabesi2, Andrea Cerasa3, Mario Menegatti4, Domenico Perrotta5.   

Abstract

The contrast of fraud in international trade is a crucial task of modern economic regulations. We develop statistical tools for the detection of frauds in customs declarations that rely on the Newcomb-Benford law for significant digits. Our first contribution is to show the features, in the context of a European Union market, of the traders for which the law should hold in the absence of fraudulent data manipulation. Our results shed light on a relevant and debated question, since no general known theory can exactly predict validity of the law for genuine empirical data. We also provide approximations to the distribution of test statistics when the Newcomb-Benford law does not hold. These approximations open the door to the development of modified goodness-of-fit procedures with wide applicability and good inferential properties.
Copyright © 2019 the Author(s). Published by PNAS.

Entities:  

Keywords:  Newcomb–Benford law; anomaly detection; customs fraud; customs valuation; statistical antifraud analysis

Year:  2018        PMID: 30530688      PMCID: PMC6320519          DOI: 10.1073/pnas.1806617115

Source DB:  PubMed          Journal:  Proc Natl Acad Sci U S A        ISSN: 0027-8424            Impact factor:   11.205


The contrast of fraud in international trade, and the corresponding protection of national budgets, is a crucial task of modern economic regulations. To give an idea of the volumes involved, in 2016 the customs duties flowing into the European Union (EU) budget amounted to more than 20 billion euros and provided about 15% of the total own resources of the EU. Huge losses thus occur when the value of imported goods is underreported (e.g., ref. 1). Most statistical antifraud techniques for international transactions fall in the class of unsupervised methods, with outlier detection and (robust) cluster analysis playing a prominent role (2–5). The rationale is that the bulk of international trade data are made of legitimate transactions and major frauds may stand out as highly suspicious anomalies. Considerable emphasis is also put on procedures that provide stringent control of the number of false positives (6), since substantial investigations like the one reported in ref. 1 are demanding and time consuming. A related crucial requirement is the ability to deal with massive datasets of traders and to provide—as automatically as possible—a ranking of their degree of anomaly. This information is essential for the design of efficient and effective audit plans, a major task for customs offices. In this work we consider fraud detection through the Newcomb–Benford law (NBL). This law defines a probability distribution for patterns of significant digits in real positive numbers. It relies on the intriguing fact that in many natural and human phenomena the leading—that is, the first significant—digits are not uniformly scattered, as one could naively expect, but follow a logarithmic-type distribution. We refer to refs. 7–10 for an historical summary of the NBL, an extensive review of its challenging mathematical properties, and a survey of its more relevant applications. Despite its long history, the mathematical and statistical challenges of the NBL have been recognized only recently. From a mathematical perspective, appropriate versions of the law appear in integer sequences, such as the celebrated Fibonacci sequence (8) or the factorial sequence (11). The law also emerges in the context of floating-point arithmetic (12), while a deep probabilistic study was carried out by Hill (13). A seminal note by Varian (14) suggested the idea that agreement with the NBL could validate the “reasonableness” of data. Since then, it is now rather well known—mainly due to the work of Nigrini (see ref. 7, for a review of such studies)—that the NBL can be used as a forensic accounting and auditing tool for financial data. The law has been shown to be a valuable starting point for forensic accountants and to be applicable in a number of auditing contexts, such as external, internal, and governmental auditing. It has also been found successful for identifying the presence of misconduct in other domains, including the identification of irregularities in electoral data (15, 16), campaign finance (17), and economic data (18). Although the cited advances may suggest applicability of the NBL to international trade, there remain major unanswered questions that we address in our work. The first one concerns the trustworthiness of the NBL for genuine—that is, nonfraudulent—transactions. As shown in ref. 19, no general known theory can exactly predict whether the NBL should hold in any specific application, whose data-generating process cannot be known with certainty, even in the absence of fraud or other data manipulations; see also refs. 20–22 for related concerns. Our first goal is then to provide insight on the suitability of the NBL for modeling the distribution of digits of genuine transaction values arising in international trade. We use the Italian import market as a specimen for our study, but our approach is general and can be replicated for any country for which detailed customs data are available. Knowledge of the conditions under which the NBL should be expected to hold in the absence of data manipulation is an essential ingredient for the implementation of large-scale monitoring processes in which tens (or even hundreds) of thousands of traders are screened in an automatic and fast way with the aim of identifying the most suspicious cases. In we describe a web application that has been developed to assist customs officers and auditors in this screening task, which can be executed in full autonomy on their own datasets. It may instead be very difficult to ascertain whether anomaly should be attributed to fraud or to model failure if the NBL does not provide a suitable model for genuine transactions; see also ref. 23, p. 193, for a similar concern. Our second goal is to deepen our knowledge of the empirical behavior of NBL-conformance tests by investigating their power under different contamination schemes. The adoption of such tests for antifraud screening is based on the assumption that fabrication of data closely following the law is difficult and that fraudsters might be biased toward simpler digit distributions, such as the discrete uniform or the Dirac. We also quantify the corresponding false positive rates, to make explicit the different and possibly conflicting facets that empirical researchers have to balance in practice. The third aim of our work is to provide corrections to test statistics when the NBL does not hold. This is typically the case for traders who operate on a limited number of products, so that there is not enough variability in their transactions. Even if the NBL is not a suitable model for genuine transaction digits, the conformance tests based on our modified statistics have the appropriate empirical size in the absence of data manipulation, while the usual tests turn out to be potentially very liberal. We argue that, having the required size under general trade conditions and being competitive in terms of power, the conformance tests based on our modified statistics are recommended. Therefore, they extend the applicability of large-scale monitoring processes of international trade data to a wider range of practical situations.

The NBL

Statistical Background.

Let be the first, the second, …, significant digit of the positive real number . Let be a positive real random variable defined on the probability space . The NBL implies (and vice versa) that the following joint probability function holds for each ,where and for . A practically important special case is that of the first two significant digits (), for which Eq. reduces toSimilarly, the marginal probability function of iswhile the marginal probability function of isWe refer to ref. 24 for a summary of the mechanisms that give rise to NBL-distributed data in accounting and finance. Among these, there are several statistical motivations for adopting the NBL as a model for the digits appearing in genuine international transactions. A major methodological basis relies on a limit theorem derived by Hill (13), to which we refer for the technical details. A key mathematical concept is that of a random probability measure, which is a function —where is the space of probability measures on —defined on the underlying probability space . For each Borel set the function is a random variable; that is, is a probability measure on for each . Another important related concept is that of a sequence of -random samples, where . It is a sequence of random variables defined on such that, for each , the first random variables are drawn independently from the same random probability distribution , selected according to the random probability measure , the subsequent random variables are drawn independently from the same random probability distribution , in turn selected according to the random probability measure , and so on. Hill’s limit theorem then states that, if satisfies some invariance conditions related to either the scale or the base of measurement, for each the -random -samples sequence converges to the NBL with probability one. That is, for each and for , as A second reason for adopting the NBL is that multiplicative processes—which are at the heart of many financial data—generate NBL-distributed data. More precisely, if is a sequence of independent and identically distributed random variables such that , as the sequence converges to the NBL with probability one (theorem 8.16 in ref. 8). It can be shown that convergence is extremely fast since it is exponential in (25). It is also remarkable that, given two independent random variables and only one of which follows the NBL, the product is distributed according the NBL provided that (theorem 8.12 in ref. 8). Finally, NBL-distributed data may also originate from random variables raised to integer powers. If is an absolutely continuous random variable, as the sequence converges to the NBL with probability one (theorem 8.8 in ref. 8).

Relevance for International Trade.

Our applied focus is on transactions involving EU traders; we refer to for the institutional regulations supporting their analysis. By international trade data we mean the data collected by EU member states for imports and exports that are declared by national traders and shipping agents using the form called the Single Administrative Document (SAD). The value that we analyze for antifraud purposes is the “statistical value” reported in each SAD, which also includes the costs of insurance and freight (CIF) and is given in euros by taking into account the exchange rate (26). Our interest is then on random variables defined on the product spacewhere and are nonnegative random variables representing the (CIF-type) unit price in euros and the traded quantity in transaction . If we rephrase in the context of trade, corresponds to the number of transactions made by the trader of interest, so that is the available sample of transaction values, and the ratio is the corresponding number of traded goods (provided that is an integer). There are different economic reasons suggesting that the distribution of the significant digits contained in may, under some conditions, be well approximated by the NBL. First, markets are hit by specific shocks and show peculiar reactions to common shocks (27). This, coupled with differences in the trader size and product quality, generates different economic processes for prices and quantities determination, which imply in turn that the observed data of prices and quantities may be described by different trader-specific probability distributions, not exactly predictable in advance. In view of , it is then sensible to anticipate good conformance to the NBL when a trader operates by importing or exporting a sufficiently large number of different goods, even if none of the product-specific marginal distributions of digits follows the law. The economic literature also shows that traders have different degrees of market power. Trading operations are affected by market and country features, such as different trade costs and different access to credit (e.g., ref. 28). Therefore, transactions made with different counterparties may be characterized by different economic processes, yielding distributions for transaction values that can be conceived to vary randomly from one product to another for each trader. The significant-digit distribution in international transactions can thus be expected to adhere to the NBL when the trader makes a sufficiently large number of operations, with a sufficiently large number of counterparties, possibly located in different countries.

A Contamination Model for Fraud

The Model.

We phrase our antifraud approach within the framework of a trader-specific contamination model where each fraud corresponds to an outlier. For this purpose, we need a slight change in notation and we write for the number of transactions made by trader , which operates on distinct products and for which the positive random variable now represents a transaction value. We then defineand let denote the total number of traders in the market. For and each , the general form of our contamination model iswhere is the probability of observing in the absence of fraud, is the probability of the same event for a manipulated transaction, and is the probability of fraud for trader . Although it is convenient to work in the digit space through , model has a counterpart in the transaction space defined by . The latter is given in . Model provides a principled framework for antifraud analysis of international trade data. Indeed, trader may be considered a potential fraudster if the null hypothesisis rejected, in favor of the alternative , based on independent copies of , say . A useful tractable version of contamination model assumes that the probability of observing a given -ple of digits in a genuine transaction of trader depends on the trader features only through the values of and ; that is,Therefore, for each , the model becomeswith again stating the absence of fraud. Model implies that the random vector is independent of any other trader-specific random variable, given the values of and . Although this structure is clearly an approximation, it is coherent with the discussion about the economic elements that make the NBL a plausible model for the digit distribution in genuine international transactions. A further bonus of models and is that they make clear the antifraud advantages of our methodology over the often uninformative analysis of aggregated data, as given, for example, in ref. 18. In the latter instance, for each , the underlying contamination model would bewhere the quantities involved are now constant for the whole (product-specific) market. Testing the hypothesis that in this restricted model requires a sample obtained from traders, for which just one replicate is available. However, the inferential conclusion that is much less informative than rejection of for some . In fact, yields no information on the specific traders that are responsible for rejection and identification of the fraudsters must be left to further nonstatistical investigations. Another notable advantage is that models and acknowledge the existence of a trader-specific propensity to fraud.

Testing the Absence of Fraud.

The usual hypothesis of interest in the antifraud literature (7, 10) iswhich corresponds to when is the NBL. Several statistics exist for testing for a given value of , the simplest one being the χ2 statisticwhere is the frequency of the -ple in the sample of transactions for trader . It is a standard result that, as , when is true, with . In practice only NBL marginals of low order are analyzed. The two-digit version of , that is, , tests the fit to the 2D marginal of the NBL given in , while the corresponding 1D marginal hypotheses are tested through and , respectively. In our empirical study we also consider the multiple-stage approach proposed by Barabesi et al. (6) with the aim of introducing a more stringent control on the proportion of false discoveries. This approach tests a decreasing sequence of lower-dimensional marginals of the NBL through their exact conditional distributions. Specifically, in the simple two-step version that we consider here, the method of Barabesi et al. (6) first tests the two-digit marginal of the NBL by comparing to the quantiles of its exact distribution under the null, which are approximated through an efficient Monte Carlo scheme. Then, if the 2D NBL is rejected, the fit to the 1D marginals is tested by and . These lower-dimensional tests use the exact conditional distributions of and , given rejection of the 2D hypothesis, instead of their marginal ones. Type-I error rates are thus controlled at the prescribed level (e.g., 1%) at each step of the procedure, both in the two-digit and in the one-digit tests. Furthermore, the outcome on the one-digit tests reveals which digit is responsible for nonconformance to . Since χ2 tests may also have some shortcomings (ref. 10, chap. 37), additional procedures not based on and less formal methods are considered in . Qualitative findings are similar in all cases. Nevertheless, for our purposes it is instructive to look at the results for χ2 tests, because their distribution (either exact or asymptotic) is known under the NBL. We can thus look at the agreement between the empirical and the nominal distribution of the test statistics to assess whether genuine transactions actually follow the law, that is, if in (or in ) is the NBL.

Adequacy of the NBL for Trade Data

Although the theoretical results sketched in the statistical background and the subsequent economic arguments broadly motivate the adoption of the NBL as a sensible model for genuine transactions in the context of international trade, it is unclear how they may fit to empirical transactions whose generating mechanism cannot be exactly known and obviously involves only a finite number of terms. One goal of our study is then to provide evidence on the quality of the NBL assumption to the digit distribution of transaction values for noncheating traders that operate in real international markets. For this purpose, we assume that our contamination model holds with for each trader. We also take as a sensible and practically workable approximation to this model in the absence of a priori information on the trader. We simulate nonmanipulated statistical values, according to definition , for “idealized” traders in each relevant configuration of trade represented by a pair . For this aim, we sample transactions with replacement from the Cartesian product spaceswhere and denote the sets of unit prices (CIF-type) and traded quantities, respectively, originated in all of the market transactions involving good , is the number of such transactions, and is the total number of goods in the market. The details of the simulation algorithm are reported in . In our experimental setting the values of and are fixed by design, while in empirical analysis we instead condition on the observed values of and for the trader under scrutiny. We replicate genuine international trading behavior in one specific EU market by picking unit price and traded quantity at random from the database of one calendar year Italian customs declarations, after appropriate trader and product anonymization making it impossible to infer the features of individual operators. Two databases of simulated transactions (pseudo-datasets) similar to those analyzed in this work can be accessed through , where their structure is explained. A description of our code is also given in . For each idealized trader and a chosen value of , we compare the observed distribution of digits to the theoretical NBL values through the test statistic . This statistic will be asymptotically distributed as if is indeed the NBL. Furthermore, its exact distribution under the -digit NBL hypothesis can be approximated to an arbitrary degree of accuracy through the Monte Carlo approach of Barabesi et al. (6). We thus take the discrepancy between the estimated distribution of , computed by averaging over the Monte Carlo replicates of , and its reference null distribution, say , as a measure of the adequacy of the NBL assumption in model . Formally, let be the quantile of and let denote the indicator function of a given set . Our Monte Carlo estimate is computed asfor in the usual range of significance levels. Although a value of close to does not imply that the empirical distribution of is well approximated by over all its support, it tells us that the approximation is satisfactory for the purpose for which is computed in antifraud analysis. The insight that we gain from our study is twofold. First, we shed light on the trading configurations—represented in terms of pairs —that ensure close agreement between and the NBL in the market from which all of the sets and are obtained. Second, we explore the effect of sparseness of digit counts on the distribution of when is small or moderate. The bulk of our results deal with the simple first-digit statistic , which is likely to be method of choice by many antifraud practitioners in automated large-scale auditing processes. As a reference, we also provide the estimated test sizes for the two-stage (TS) version of the procedure of Barabesi et al. (6) and for the two-digit statistic . The former is intended to be a reasonable compromise between simplicity of use and strong reduction in the rate of false detections, while the latter is often recommended in applications with not-too-small sample sizes (ref. 7, p. 79). We estimate test sizes using for a wide range of pairs , with . The chosen grid represents the features of some of the most relevant traders in the empirical analysis of customs declarations. In fact, the importers for which cover less than 14% of the recorded transactions in our customs database and an even smaller quota in terms of traded values. Very big traders are not common: To give an idea, for less than 0.1% of the importers in the database, and almost 40% of the recorded transactions refer to traders with . We present only the findings for the case , similar conclusions being valid for other significance levels. Table 1 displays the estimated sizes of the test of the first-digit marginal hypothesis for both (using the quantiles of its asymptotic distribution) and TS. These estimates are computed on idealized noncheating traders, pooled across different scenarios with the same pair . One striking feature of the reported values of in Table 1 is that they vary considerably according to the specific trading configuration. This result clearly supports the conjecture that in a realistic market scenario both and are crucial factors in determining the adequacy of the NBL as a valid model for the empirical digit distribution in the absence of data manipulation. If only one digit is considered, a sample size of transactions can be considered sufficiently large to justify the asymptotic approximation to the distribution of and the adoption of the NBL as a reasonable model for , provided that the number of traded products is large as well (around 20, say). Similar findings hold for all of the pairs taken into account in our experiment and provide an empirical verification of the speed of convergence to the NBL anticipated by the asymptotic framework of Hill’s result . An interesting remark is that for is closer to when , thus suggesting that convergence in is faster when . On the other hand, TS yields a very conservative test when the NBL provides a satisfactory model. This result is hardly surprising, since TS tests the first-digit hypothesis at nominal size in the conditional distribution of , given previous rejection of the two-digit NBL hypothesis. In , we also investigate the fit of the whole empirical distribution of to the nominal distribution.
Table 1.

Estimated test sizes (Eq. 11) for the first-digit statistic , using the asymptotic quantile , and for the TS version of the procedure of Barabesi et al. (6), based on Monte Carlo replicates for each configuration , with

mt
No. of transactionsTest1510204080100200500
nt=50V{1}(t)0.0530.0270.0180.0140.011
TS0.0240.0030.0010.0000.000
nt=100V{1}(t)0.0710.0450.0270.0160.0120.0110.011
TS0.0490.0130.0040.0010.0000.0000.000
nt=200V{1}(t)0.0940.0690.0470.0260.0160.0120.0110.010
TS0.0700.0350.0130.0030.0010.0000.0000.000
nt=500V{1}(t)0.1320.1260.0970.0620.0310.0170.0160.0120.010
TS0.1030.0840.0490.0170.0030.0000.0000.0000.000

Model holds with for each trader. The nominal test size is .

Estimated test sizes (Eq. 11) for the first-digit statistic , using the asymptotic quantile , and for the TS version of the procedure of Barabesi et al. (6), based on Monte Carlo replicates for each configuration , with Model holds with for each trader. The nominal test size is . Our results point to the conclusion that the NBL is not a satisfactory model when is much smaller than . This statement is verified consistently over all market configurations and does not depend on the specific testing methodology. Indeed, also the potentially very conservative TS procedure can become considerably liberal if . The same is true for other adjustments to that control for multiplicity of tests among traders, not reported here. We argue that lack of variability in the transactions made by trader is the main reason for the discrepancy between the NBL and when is small. Whatever the interpretation, our simulation results confirm that the asymptotic framework set by does not hold if , requiring instead . Our results also quantify how much deleterious can be the effect of keeping fixed on the distribution of test statistics. Indeed, they show that in this setting an increase of the sample size worsens the situation, since it points to a “wrong” asymptotic direction. The clear message is then that standard conformance tests, such as , should not be used for antifraud purposes when , because the hypotheses and cannot be taken any longer to be equivalent. We conclude this section with a glimpse of the performance of the two-digit statistic , when either the asymptotic quantile or the exact 0.99 quantile from Barabesi et al. (6) is used. The estimated test sizes, now based on Monte Carlo replicates for each configuration , are reported in Table 2. As expected, convergence to the distribution is slower than convergence to in the one-digit case. The adoption of exact quantiles should thus be preferred with , except in the instance of large values of both and . Our results confirm the relationship between accuracy of the NBL approximation and the ratio , suggesting as a sensible rule of thumb when the exact quantiles are used. They also provide a clue of the strategy to be adopted with more complex large- procedures.
Table 2.

Estimated test sizes (Eq. 11) for the two-digit statistic , using the asymptotic quantile (As) and the exact 0.99 quantile (Ex) from Barabesi et al. (6), based on Monte Carlo replicates for each configuration , with

mt
No. of transactionsTest1510204080100200500
nt=50As0.0640.0390.0350.0290.026
Ex0.0400.0170.0130.0110.010
nt=100As0.0830.0480.0330.0230.0210.0200.019
Ex0.0680.0320.0190.0130.0110.0100.010
nt=200As0.1020.0690.0430.0250.0180.0140.0160.014
Ex0.0950.0590.0340.0180.0120.0100.0110.009
nt=500As0.1410.1250.0870.0520.0270.0160.0140.0120.012
Ex0.1370.1200.0820.0470.0230.0130.0120.0100.009

Model holds with for each trader. The nominal test size is .

Estimated test sizes (Eq. 11) for the two-digit statistic , using the asymptotic quantile (As) and the exact 0.99 quantile (Ex) from Barabesi et al. (6), based on Monte Carlo replicates for each configuration , with Model holds with for each trader. The nominal test size is .

Enemy Brothers: Power and False Positive Rate

When model holds with for one or more traders, we write and for the sets corresponding to noncheating traders and fraudsters, respectively. Power (P) is defined as the proportion of traders in that are correctly identified as potential fraudsters. The false positive rate (FPR) is the proportion of rejections of the null hypothesis that turn out to be wrong, since they refer to traders that belong to . Both performance measures play a crucial role when antifraud analysis is put into practice. In our simulations we compare the results under different contaminant distributions , with . Our first contamination instance assumes that the first two digits of transactions from trader are generated according to the discrete uniform distribution on . Therefore,for and . The uniform distribution provides an unfavorable scenario for fraud detection, since is then close to the NBL marginal probability for most digit pairs . Our second contamination scheme instead concentrates frauds on a specific digit pair, say , randomly selected from the discrete uniform distribution on . The contaminated model thus becomesAlthough this Dirac-type contamination may at first sight appear extreme, our experience with manipulated declarations is that similar patterns may arise rather frequently among the transactions found to be fraudulent, especially when contamination is due to the attempt to circumvent threshold-depending duties, either “ad valorem”—that is, computed as a percentage of the declared value—or fixed. In fact, the attempt to declare quantities below the threshold (or above it, according to the specific regulation) typically produces a bias in the corresponding values similar to that represented by a Dirac-type model. Other instances of contamination are considered in . We consider the simplified case where is the same for each . We take , to represent three increasing levels of individual propensity to fraud. We also define the proportion of fraudsters in the whole market aswhere is the set of all traders. We fix , to investigate the effect of different degrees of fraud diffusion in the market. Our estimates of P and FPR are based on idealized traders, independently generated in each configuration. Nonmanipulated transactions are again simulated with the algorithm described in . We restrict our analysis to the market configurations for which the NBL approximation to is good, and the empirical test sizes closely match the nominal one, to avoid confounding between power and lack of fit. We give results only for the configurations with . Pairs where is of the same order of magnitude as yield qualitatively similar findings and are not reported. Table 3 shows the estimated values of P and FPR under the uniform contamination model for , using the asymptotic quantile , and for the TS version of the procedure of Barabesi et al. (6). Not surprisingly, the detection rates are low in the case of sporadic contamination (). It is apparent that no statistical method can be expected to have high power against “well-masked” frauds, unless the number of contaminated transactions becomes relatively large. Indeed, it is clearly seen that P rapidly grows with both and , leading to almost sure detection of fraudsters even through the potentially very conservative TS procedure (e.g., when and ). Both methods thus prove to be able to identify the traders belonging to if there is enough information on the contaminant distribution in the available data, also in the unfavorable framework provided by . The value of FPR is much higher with , as expected, except in some instances of low contamination, where the number of hypotheses rejected by TS is very small and the estimate of FPR is overwhelmed by its sampling variability. The choice between and TS should then depend on the user’s attitude toward FPR and toward the power reduction implied by TS in situations of intermediate contamination. The value of does not have a major impact on P, thus suggesting that our procedures can be equally effective in detecting isolated fraudsters and more diffuse illegal trading behavior. However, a considerable increase in FPR is to be expected in the former situation, especially for .
Table 3.

Uniform contamination model 12

ς=0.05ς=0.10
τt=0.2τt=0.5τt=0.8τt=0.2τt=0.5τt=0.8
Trade configurationTestPFPRPFPRPFPRPFPRPFPRPFPR
nt=50V{1}(t)0.0340.8650.1960.5460.5860.3020.0300.7790.2000.3460.5740.178
mt=50TS0.0020.0000.0080.0000.1540.0130.00010.0190.0000.1330.007
nt=100V{1}(t)0.0580.7880.4360.2970.9380.1840.0430.7050.4250.1750.9240.097
mt=100TS0.0040.0000.0700.0540.5740.0030.0020.6670.0630.0000.5390.002
nt=200V{1}(t)0.0600.7780.8100.17910.1510.0970.4840.8010.10910.097
mt=200TS0.0060.5000.3560.0000.9640.0020.0050.4440.3450.0030.9590.004
nt=500V{1}(t)0.2720.40110.16010.1540.2810.22610.06910.081
mt=500TS0.0280.2630.9320.00010.0040.0290.0650.9280.00010.000

Shown are estimated power (P) and false positive rate (FPR) for the first-digit statistic , using the asymptotic quantile , and for the TS version of the procedure of Barabesi et al. (6), based on Monte Carlo replicates for each pair . The nominal test size is .

Uniform contamination model 12 Shown are estimated power (P) and false positive rate (FPR) for the first-digit statistic , using the asymptotic quantile , and for the TS version of the procedure of Barabesi et al. (6), based on Monte Carlo replicates for each pair . The nominal test size is . Table 4 repeats the analysis under the Dirac-type scheme . The contaminant distribution is now well separated from and both methods generally have excellent detection properties, with some minor differences only in the problematic case . However, FPR is much higher for . In such contamination frameworks the TS procedure thus comes closer to performing like an “ideal” test, leading to the identification of most potential fraudsters with a very small number of false alarms. The effect of is still minor on P, while it is more noticeable on FPR for .
Table 4.

The same as Table 3, but now for contamination model 13

ς=0.05ς=0.10
τt=0.2τt=0.5τt=0.8τt=0.2τt=0.5τt=0.8
Trade configurationTestPFPRPFPRPFPRPFPRPFPRPFPR
nt=50V{1}(t)0.7120.2180.9980.19910.1840.6960.1210.9960.09210.108
mt=50TS0.5200.76310.00210.0020.5550.00510.00310.001
nt=100V{1}(t)0.8760.18910.18810.1450.8910.09510.08310.081
mt=100TS0.9720.00810.00410.0000.9800.00110.00310.001
nt=200V{1}(t)0.9720.16910.16710.1500.9670.09110.07910.076
mt=200TS10.00410.00610.00010.00210.00210.000
nt=500V{1}(t)10.17110.15810.17610.07810.09410.071
mt=500TS10.00610.00010.00010.00210.00310.003
The same as Table 3, but now for contamination model 13

Corrections to Goodness-of-Fit Statistics

We now focus on the trading configurations for which the NBL does not provide a satisfactory representation of the genuine digit distribution , that is, when . In this case, the reported distributional results are no longer valid for or for the exact Monte Carlo approach of Barabesi et al. (6). The true probability should replace the NBL version of in to obtain valid tests of hypothesis . Since is unknown, we resort to our Monte Carlo algorithm for simulating nonfraudulent transactions and we compute a model-free approximation to the null distribution function of . This approximation is then used to obtain a test of . Similar testing procedures have proved to be useful in other domains, in the case of correlated observations and other distributional misspecifications (e.g., ref. 29 and the references therein). If is the trader of interest, let be an idealized noncheating trader such that , while and . The set of transactions for trader is randomly generated according to the algorithm described in , and the resulting statistical values are collected in vector , say. Correspondingly, let be the test statistic computed for trader . Under model , the significant-digit random variables associated to the elements of can be considered as independent copies of those associated to the elements of , in the absence of data manipulation. We thus estimate the unknown null distribution function as a Monte Carlo average over replicates of . This yieldsfor , andfor the corresponding estimate of the quantile. Therefore, we reject hypothesis at nominal test size , and we consider trader a potential fraudster, ifwhere is the observed value of . Motivated by large-scale applications, Efron (30) describes a related methodology for empirically estimating a null distribution when the standard theoretical model (such as the NBL in the case of digit counts) does not hold. This approach uses the available data to estimate an appropriate version of the distribution of the test statistic under the null hypothesis. However, it is apparent that empirical null estimation is not directly feasible when recast in the framework of models and . One reason is that the method generally requires a known parametric form for the null distribution, whose parameters are then estimated from the available realizations of the test statistic. Even more fundamentally, in our applied context there is no guarantee that the proportion of genuine transactions is large for each trader, that is, that is small for each in models and , thus violating a key assumption for empirical null estimation (ref. 30, p. 98). On the other hand, the proportion of transactions that involve manipulated data and their impact on is arguably small when considering the Cartesian products defined in Eq. . First, both and are not trader specific, since they contain all of the transactions in the market for the corresponding good, and the resulting idealized transactions are further aggregated to obtain the required basket of transactions on products. Second, as already reviewed in the statistical background, an intrinsic robustness property of the NBL specification of our contamination model arises from decomposition , since the product of independent random variables follows the NBL if only one of the factors does, regardless of the other factors (ref. 8, p. 188). We may thus expect a reduction in the contamination effect produced by a manipulated element of (respectively, ), after multiplication by a genuine element of (respectively, ). Third, if the NBL does not hold, the contaminant distribution for a trader may not be too far from the genuine distribution for some other trader , which further reduces the degree of anomaly of the corresponding realizations in the whole market. We thus see our estimate as the outcome of an extended null estimation approach, where is estimated by exploiting all of the potential samples that could have been observed given the realized transactions in the market. Since the cardinality of this sample space is very large, we finally resort to Monte Carlo simulation for approximating the extended empirical null. Table 5 reports the estimated sizes for different values of and for , when test is performed at on the same sets of idealized traders already considered in Table 1, and the Monte Carlo average in is computed on independent replicates for each value of . The analysis for the case is given in . In all instances, comparison with the estimated sizes of the liberal test (copied from Table 1) shows that the improvement provided by our procedure is paramount. The appropriate size is also reached when grows, while is kept fixed. Therefore, our approach provides a valid test of even when the asymptotic framework does not comply with the requirements of Hill’s limit theorem.
Table 5.

Estimates of test size, P, and FPR using modified procedures 15 and 16, with , for different values of and for

Uniform contamination (Eq. 12)Dirac-type contamination (Eq. 13)
τt=0τt=0.5τt=0.8τt=0.5τt=0.8
No. of transactionsTestα^PFPRPFPRPFPRPFPR
nt=100V{1}(t)0.0710.4140.7160.9280.60010.57910.572
Test 150.0100.00010.00010.8500.16710.180
Test 160.0110.3500.3290.8640.1790.9900.16110.144
nt=200V{1}(t)0.0940.8120.68310.63010.64810.634
Test 150.0100.00010.00010.8780.15710.187
Test 160.0120.6780.2130.9340.1820.9980.1530.9920.175
nt=500V{1}(t)0.13210.71910.71410.70810.717
Test 150.0100.0040.9830.00010.7760.17310.154
Test 160.0100.8940.1890.9380.1490.9960.17110.143

The estimated test sizes for are also given as a reference. The nominal test size is . The number of independent idealized traders in each market configuration is for procedure and for procedure , P and FPR. when computing P and FPR

Estimates of test size, P, and FPR using modified procedures 15 and 16, with , for different values of and for The estimated test sizes for are also given as a reference. The nominal test size is . The number of independent idealized traders in each market configuration is for procedure and for procedure , P and FPR. when computing P and FPR We then compute P and FPR for test , under the uniform contamination model and the Dirac-type contamination scheme , using the same sets of idealized traders already considered in Tables 3 and 4. For simplicity, we restrict our analysis to and , similar qualitative conclusions being reached in the other cases. The results are again reported in Table 5 and in , for and , respectively. We see that test can have severe difficulties in discriminating between and , unless and are well separated or is close to one. One reason for the observed loss of power is the large number of goods that are potentially involved in the Monte Carlo estimation process. Indeed, for each idealized trader contributing to , but the specific goods for which the digit distribution is obtained usually vary from trader to trader. This variability inflates the quantile estimate , especially when the ratio increases. We can obtain an improved estimate of the required quantile by adopting a refined version of model . In this specification the genuine digit distribution depends not only on , but also on the specific set of goods, say , dealt with by trader . Consequently, we now generate the behavior of idealized noncheating traders with the constraint that . Let denote the corresponding Monte Carlo estimate of , computed as in . Then,and hypothesis is rejected at nominal test size ifThe number of ways in which a basket of products can be selected out of possible goods will be huge in any real-world scenario. Computation of thus becomes trader specific and cannot be automated before knowing the exact composition of , differently from , which depends only on the pair . Nevertheless, estimation time is still acceptable for routine application of the methodology. For instance, in our experiment computation of using replicates takes on average less than 0.5 s for a trader with and . The performance of the refined test procedure is displayed in Table 5 (for ) and in (for ). All of the estimated sizes are very close to the nominal target and similar to those obtained through . Power values are comparable for the three reported tests when the genuine and the contaminant digit distributions are well separated. However, our proposals are still preferred since their FPR is considerably lower than for . It is in the case of intermediate contamination, as under the uniform model, that the refined estimator shows much higher efficiency than . In this instance rule ensures that the reduction in power with respect to the test is minor, while keeping considerably lower values of FPR. We thus conclude that, having the appropriate size and power properties comparable to those of the liberal standard procedure, our modified tests and are recommended whenever the attained levels of FPR can be tolerated in practice.

Case Studies

To illustrate the use of the proposed procedure and its ability to detect relevant value manipulations, we first discuss the case of a trader extracted from an archive of fraudulent declarations provided by the Italian customs after appropriate data anonymization. The same archive was also used in ref. 6. The trader under scrutiny has import transactions on products from January 2014 to June 2015. The quantities and values appearing in the declarations of the three most traded products (not labeled for confidentiality reasons) are represented as (red) solid circles in the scatter plots of Fig. 1. The information displayed in such scatter plots is the input for some commonly adopted (robust) regression techniques aiming at the automatic detection of value frauds in customs data; see, e.g., ref. 31 and for further details. However, the plots for this trader do not provide clear evidence of substantial undervaluation or of other major anomalies, although two of the declarations displayed in Fig. 1, Center were found to be fraudulent after substantial investigation. Our testing procedure instead produces a strong signal of contamination of the digit distribution. In fact, restricting for simplicity to the first digit, we obtain and , based on simulated traders with the same values of and . By applying rule , we can thus conclude that hypothesis can be safely rejected when the focus is shifted from individual transactions, as in Fig. 1, to the whole trader activity, as in our test.
Fig. 1.

Quantity-value scatter plots for the three most traded products by an Italian operator convicted for two false declarations. The transactions made by this trader are represented as (red) solid circles.

Quantity-value scatter plots for the three most traded products by an Italian operator convicted for two false declarations. The transactions made by this trader are represented as (red) solid circles. The strength of evidence against the null may suggest the existence in the administrative records of this trader of a larger number of manipulated declarations than the two already detected. It also suggests that our method could be helpful in providing authorities with evidence of potential fraud among traders not previously classified as fraudsters or even not considered as suspicious. In view of contamination models and , and of our simulation results, we expect this information gain to be higher in the case of serial misconduct. Additional investigations for this trader are given in . Although all methods point to the same conclusion, we remark that simple graphical tools for conformance checking—such as histograms—require substantial human interpretation and thus cannot be routinely applied on thousands of traders. We now move to (anonymized) data provided by the customs office of another EU member state, not disclosed for its specific confidentiality policy, that we label as MS2. The data were collected in the context of a specific operation on undervaluation, focusing on a limited set of products traded by fraudulent operators that have systematically falsified the import values. The traders classified as nonfraudulent were audited by the customs officers of MS2 and no indications of possible manipulation of import values were found. Although the absence of fraud can never be anticipated with certainty, we can thus place good confidence on these statements of genuine behavior. In we provide empirical investigations of the first-digit distribution of the 15 traders in this small benchmark study for which , as in our simulation experiments. We apply test instead of test , since the available database is limited to a basket of fraud-sensitive products, and we keep and for each observed pair . We give the estimated P value of each test, computed as , and—as a reference—the asymptotic P value from the distribution that assumes validity of the NBL. It can be seen that our approach gives very good results, both when applied to fraudsters—it clearly rejects the hypothesis of no contamination for five traders—and in the case of genuine behavior—none of the supposedly honest traders is flagged by our test at . Therefore, this study supports the claim that our methodology can be an effective aid to the preparation of the audit plans of customs services, given its ability to point to potential serial fraudsters, in agreement with current guidelines for the customs modernization process (32). We finally note the beneficial effect of our correction for one supposedly honest trader shown in , whose small basket of traded products may imply spurious deviation from the NBL when the classic approximation is used. An extreme example of this effect is also shown in .

Discussion

We have developed a principled framework for goodness-of-fit testing of the NBL for antifraud purposes, with a focus on customs data collected in international trade. Our approach relies on a trader-specific contamination model, under which fraud detection has close connections with outlier testing. We have given simulation evidence, in the context of a real EU market, showing the features of the traders for which we can expect the genuine digit distribution to be well approximated by the NBL. Our simulation experiment is an empirical study addressing this issue in detail in the context of international trade, where the contrast of fraud has become a crucial task and substantial investigations are often demanding and time consuming. We have also provided simulation-based approximations to the distribution of test statistics when the conditions ensuring the validity of the NBL do not hold. These approximations open the door to the development of goodness-of-fit procedures with good inferential properties and wide applicability. Our methodology is general and potentially applicable to any country, or year, for which detailed customs data are available. Being mostly automatic, it is suited to be implemented in large-scale monitoring processes in which thousands of traders are screened to find the most suspicious cases. It can also be a valuable aid to the design of efficient and effective audit plans. Although we expect our general guidelines to remain valid in other empirical studies, the specific quantitative findings may clearly vary from one country (year) to another. A bonus of our contamination approach is that it makes clear the setting in which statistical antifraud analysis takes place. Our conformance testing procedures mainly aim at the detection of serial fraudsters, for which information accumulates in the corresponding transaction records. The generation of low-price clusters of anomalous transactions is a typical consequence of this cheating behavior, and robust clustering techniques can also be used for its detection (e.g., ref. 4). However, rejection of our goodness-of-fit null hypotheses often provides more compelling evidence of fraud, also because it may not be easy to identify the low-price clusters that actually correspond to illegal declarations. Testing conformance to the NBL, or to another suitable distribution for genuine digits, thus shifts the detection focus from individual transactions to the full set of data from each trader. A word of caution concerns the fact that not all possible frauds can be detected by our method, even when we restrict to manipulation of transaction values. For instance, we cannot expect any statistical procedure (including our own proposal) to have high power against data fabrication methods that preserve the validity of the NBL, at least approximately, and against occasional frauds for which statistical tests are not powerful enough. Therefore, we do not see our methodology as the ultimate antifraud tool, but as a powerful procedure to be possibly coupled with additional information. We support integration of the signals provided by our method with those obtained through alternative statistical techniques and with less technical model-free analyses—such as those developed in refs. 7 and 10—that can be applied on a restricted number of traders. Indeed, we see our approach as a suitable automatic tool for selecting the most interesting cases for additional qualitative and quantitative investigations, while ensuring control of the statistical properties of the adopted tests.
  2 in total

1.  Testing mutual independence between two discrete-valued spatial processes: a correction to pearson chi-squared.

Authors:  Andrea Cerioli
Journal:  Biometrics       Date:  2002-12       Impact factor: 2.571

2.  Statistical detection of systematic election irregularities.

Authors:  Peter Klimek; Yuri Yegorov; Rudolf Hanel; Stefan Thurner
Journal:  Proc Natl Acad Sci U S A       Date:  2012-09-24       Impact factor: 11.205

  2 in total
  7 in total

1.  Newcomb-Benford law helps customs officers to detect fraud in international trade.

Authors:  Lucas Lacasa
Journal:  Proc Natl Acad Sci U S A       Date:  2018-12-12       Impact factor: 11.205

2.  The role of diversity and ensemble learning in credit card fraud detection.

Authors:  Gian Marco Paldino; Bertrand Lebichot; Yann-Aël Le Borgne; Wissam Siblini; Frédéric Oblé; Giacomo Boracchi; Gianluca Bontempi
Journal:  Adv Data Anal Classif       Date:  2022-09-28

3.  Benchmarking of public health surveillance of COVID-19 in Colombia: First semester

Authors:  Jose William Martínez; Juan Camilo Martínez; Diego Alejandro Rincón; Diego Alejandro Salazar; Juan Daniel Castrillón; María Del Pilar Gómez; Oscar Felipe Suárez; Juan Pablo Vélez; Ángela María Valencia; Sandra Gómez; Ángel María Rincón; Álvaro J Idrovo; José Moreno-Montoya; Franklyn E Prieto-Alvarado; Alexandra Hurtado-Ortiz
Journal:  Biomedica       Date:  2020-10-30       Impact factor: 0.935

4.  Characterizing Human Cell Types and Tissue Origin Using the Benford Law.

Authors:  Sne Morag; Mali Salmon-Divon
Journal:  Cells       Date:  2019-08-29       Impact factor: 6.600

5.  On the authenticity of COVID-19 case figures.

Authors:  Adrian Patrick Kennedy; Sheung Chi Phillip Yam
Journal:  PLoS One       Date:  2020-12-08       Impact factor: 3.240

6.  Detecting fabrication in large-scale molecular omics data.

Authors:  Michael S Bradshaw; Samuel H Payne
Journal:  PLoS One       Date:  2021-11-30       Impact factor: 3.240

7.  Testing for Benford's Law in very small samples: Simulation study and a new test proposal.

Authors:  Andrea Cerasa
Journal:  PLoS One       Date:  2022-07-22       Impact factor: 3.752

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.