Explainable Artificial Intelligence (XAI)** and its importance in Tax Administration

General aspects

Artificial intelligence (AI) has been impacting not only business, but also our daily lives, from facial and image recognition to predictive analytics based on machine learning, conversational applications, autonomous machines and hyper-personalized systems[1]. The kinds of decisions and predictions that AI-enabled systems make are becoming much deeper and they may become critical to life, death, and personal well-being (Schmelzer, 2019).

Although they affect everyone’s life, few of us are able to understand how and why AI systems make their decisions and how they apply in our lives. Nor is it possible to reconstruct the path followed by many algorithms based on the results found. This opacity leads to mistrust in the results. Particularly noteworthy are applications in healthcare, driverless cars and drone operations in warfare. Today, it is increasingly important, for certain applications, to also recognize the fundamental human rights and legal impacts associated with AI system decisions or recommendations. Furthermore, in cases brought to court, judges increasingly confront machine learning algorithms, seeking explanations on which they can base their decisions.

This context strengthens the debate about which decisions based on AI algorithms need explanation, and with what types of clarifications. Thus, the explanation of how they have reached a certain conclusion could be important to allow those affected by it, to challenge or change that result.

The AI algorithms

AI is based on distinct types of algorithms, which act on data sets, in models such as machine learning (ML), deep learning (DL) and neural networks (NN). According to IBM, too many of these models look like “black boxes”, which are impossible to interpret. Neural networks, used in deep learning, are often the most difficult for humans to understand. Bias, often based on race, gender, age, or location, is a risk that has always been present in AI model training. In addition, according to the same source, the performance of the model may vary or degrade, as the production data is different from the training data. This adds an additional decisive argument for an institution to continuously monitor and manage AI models, promote elucidation and evaluate the decisions.

Algorithms and legal consequences

Operational and legal problems with algorithm-based systems are not new, the literature has some examples, such as the case of the fraud detection system in unemployment insurance claims of the state of Michigan (United States), which improperly charged fines and blocked returns of taxes to thousands of beneficiaries, and now responds to legal processes. In 2019, a Dutch court determined that an algorithm used in the detection of social assistance fraud violated human rights and ordered the government not to use it anymore (Wykstra, 2020).

The tax administration, which is pleased with the potential of AI algorithms to improve its performance in different areas, must also consider the effects of algorithms when the issue is subject to discussion in court.

Kuzniacki and Tylinski (2020) indicate that the greatest source of risks for the integration of AI into the legal process in Europe are the General Data Protection Regulations-RGPD[2] and the European Convention on Human Rights-ECHR[3]. The legal articles mentioned by the authors are the following:

  • The right to an explanation (articles 12, 14 and 15 of the GDPR): this right deals with the transparency of an AI model. If the tax authorities take a decision based on an AI model, they must be able to explain to taxpayers how this decision was made, providing sufficiently complete information so that they can act to challenge the decision, correct inaccuracies or request its cancellation.
  • The right to human intervention (Article 22 of the GDPR): The right to human intervention is a legal tool to challenge decisions that are based on data processed through automatic means, such as AI models. The idea is that taxpayers should have the right to challenge an automated AI model decision and have it reviewed by a person (usually a tax auditor). For this to be effective, the taxpayer must first know how the model arrived at the conclusion (previous right).
  • The right to a fair trial (Article 6 of the ECHR): this right includes the following elements:

    • (i) lthe minimum guarantees of equality of means and
    • (ii) the right to defense.

    It means that taxpayers must be able to effectively review the information on which the tax authorities base their decisions. The authors exemplify that taxpayers should have the right to know the relevant legal factors used to decide on a tax law application and the logic behind the AI model that prompted the authorities to make a certain decision, in order to fully understand how that decision was taken.

  • The prohibition of discrimination and the protection of property (ECHR): The prohibition of discrimination together with the protection of property determines that the tax authorities make decisions in a non-discriminatory manner, that is, that they do not treat taxpayers differently without an objective and reasonable justification. This requirement refers to the fact that certain AI algorithms tend to bear biases, as a consequence of the imbalance of the data used to train them.

Kroll, Huey et al. (2017) indicates that the use of advanced and complicated AI models, such as neural networks, is not yet transparent enough to be the only tool that makes tax decisions without human supervision. Therefore, instead of neural networks, they recommend in these cases to use easily explainable AI models, such as decision trees, different versions of the simultaneous use of multiple machine learning methods (ensemble learning), so that it is possible to provide meaningful information about the logic involved in the model. The results must appear in an intelligible way by humans, as well as the factors that led the model to make the decision to apply or not apply the tax law. This allows people to challenge such decisions, even before they are officially issued.

What is XAI?

The context described above led to the need for more advanced and structured studies on the subject, consolidated in a new field, named Explainable Artificial Intelligence or XAI.

We can define XAI as a set of processes and methods, which allows people to understand and trust the results and guidance generated by machine learning algorithms. The Explainable AI can describe an AI model, its expected impact, and potential biases[4]. Anyway, it promotes accountability (algorithm accountability).

In this context, XAI also helps promote end-user trust, model auditability, and productive use of AI, as well as mitigating reputational, security, and legal compliance risks.

Companies such as IBM have identified the importance of XAI and offer resources such as blogs and toolkits, so their customers may start studying this field.

It seems that the legal area will be the main interested party and the functional driver of the XAI studies of a tax administration.

In the current situation, in European countries and in the United States, there is no clear guidance from the courts in cases dealing with decisions based on algorithms. Since there are no consolidated precedents, courts may take decisions based on the criteria of each judge.

Deeks (2021) argues that judges will face a variety of cases in which they must demand explanations for algorithmic decisions, recommendations or predictions. Due to these demands, the judges will play a fundamental role in shaping the nature and form of XAI. Using the tools of common law, courts can construct what the role of the XAI should be in different legal contexts, including criminal, administrative and civil cases. Thus, he concludes that the judges are the main actors in the definition and meaning of XAI.

The same author affirms that as the courts employ and develop the existing case law in the face of predictive algorithms, which arise in a series of litigation, they will create the “common law of XAI”, a law sensitive to the requirements of different audiences (judges, jurors, plaintiffs or defendants) and different uses for the provided explanations (criminal, civil or administrative law settings).

Final remarks

The need to explain AI algorithms is still a new topic for tax administrations in LAC (and in the world). But the progress made in the use of AI makes it available in all areas of taxation, and lawsuits will surely follow, as the legal principles and general data protection laws in force or in preparation are often based on principles of their European and North American models.

According to Doshi-Velez, Kortz et al. (2017), in this new field of legal and technological research, the role of AI algorithm explanation to guarantee its responsible use must be periodically reassessed to adapt to a changing technological world. Access to greater computational resources can reduce the computational burden associated with explanation, but it also allows the use of even more sophisticated algorithms, increasing the challenges associated with an accurate explanation.

On the other hand, not every AI solution needs to be “explainable.” Thus, the first step is to identify this need before the design.

Likewise, it is well-advised to be alert for the oncoming controversies.

Bibliographic References:

Deeks, A. 2021. The Judicial demand for explainable Artificial Intelligence. Columbia Law Review. Vol. 119. No. 7. Available at: https://www.columbialawreview.org/content/the-judicial-demand-for-explainable-artificial-intelligence/

Doshi-Velez, F., Kortz, M. et al. 2017. Accountability of AI under the law: The role of explanation. Working Draft from Berkman Klein Center Working Group on AI Interpretability. Available at: https://arxiv.org/ftp/arxiv/papers/1711/1711.01134.pdf

Kroll, J., Huey, J. et al. 2017. Accountable algorithms. University of Pennsylvania Law Review. Available at: https://scholarship.law.upenn.edu/penn_law_review/vol165/iss3/3/

Kuzniacki, B., Tylinski, K. 2020. Legal risks stemming from the integration of artificial intelligence (AI) to tax law. Kluwer International tax blog. Available at: kluwertaxblog.com/2020/09/04/legal-risks-stemming-from-the-integration-of-artificial-intelligence-ai-to-tax-law/

Schmelzer, R. 2019. Understanding Explainable AI. Available at: https://www.forbes.com/sites/cognitiveworld/2019/07/23/understanding-explainable-ai/

Wykstra, S. 2020. How the government’s use of algorithm serves up false fraud charges. Salon.com. Available at: https://www.salon.com/2020/06/14/how-the-governments-use-of-algorithm-serves-up-false-fraud-charges_partner/

**Explainable AI in English
[1]Hyper-personalization is the most advanced way that brands can tailor their marketing to individual customers. It creates personalized and specific experiences using data, analytics, artificial intelligence and automation (Deloitte).
[2]See https://gdpr-info.eu/
[3]See https://www.echr.coe.int/Documents/Convention_ENG.pdf
[4]See https://www.ibm.com/watson/explainable-ai

1,984 total views, 3 views today

Disclaimer. Readers are informed that the views, thoughts, and opinions expressed in the text belong solely to the author, and not necessarily to the author's employer, organization, committee or other group the author might be associated with, nor to the Executive Secretariat of CIAT. The author is also responsible for the precision and accuracy of data and sources.

Leave a Reply

Your email address will not be published.

CIAT Subscriptions

Browse through the site without restrictions. Consult and download the contents.

Subscribe to our electronic newsletters:

  • Blog
  • Academic offer (Only in spanish)
  • Newsletter
  • Publications
  • News alert

Activate subscription

CIAT Members

Representatives, Correspondent and Authorized staff (TA)