{ "cells": [ { "cell_type": "markdown", "id": "849014ca", "metadata": {}, "source": [ "# Information-Theoretic Approach\n", "\n", "Information theory developed by Shannons and others as a branch of applied mathematics, electrical engineering, and computer science is involved in the quantification of information which is often a probability distribution function. Since the electron density is a continuous probability distribution function, as early as in the year of 1980, information theory has been applied to DFT to study atoms and molecules. We call this particular category of work the information-theoretic approach (ITA). \n", "\n", "There are three different representation using in ITA, they are: \n", "\n", "* Electron density representation\n", "\n", "* Shape function representation\n", "\n", "* Atoms-in-molecules representation" ] }, { "cell_type": "markdown", "id": "4c59a2cb", "metadata": {}, "source": [ "## Electron Density Representation\n", "\n", "Using the electron density $\\rho(\\mathbf{r})$ as the probability function in information theory, one obtains the first ITA representation. A key measure of information is entropy, quantifying the uncertainty involved in predicting the value of the distribution function. Shannon entropy is the first such a measure widely used in the literature, which reads\n", "\n", "$$\n", "S_{\\text{S}} = -\\int \\rho(\\mathbf{r}) \\ln \\rho(\\mathbf{r}) d\\mathbf{r} = \\int s_{\\text{S}}(\\mathbf{r}) d\\mathbf{r} \\tag{1}\n", "$$ \n", "\n", "where $s_{\\text{S}}(\\mathbf{r})$ is the Shannon entropy density and $\\rho(\\mathbf{r})$ is the total electron density, satisfying the following condition in relation to the total number of electrons, N, of the system.\n", "\n", "$$\n", "\\int \\rho(\\mathbf{r}) d\\mathbf{r} = N \\tag{2}\n", "$$\n", "\n", "Shannon entropy measures the spatial delocalization of the electronic density. The second important measure in information theory is the Fisher information $I_{\\text{F}}$, defined as followse\n", "\n", "$$\n", "I_{\\text{F}} \n", "\\equiv \\int i_{\\text{F}}(\\mathbf{r}) d\\mathbf{r} \n", "= \\int \\frac{|\\nabla\\rho(\\mathbf{r})|^2}{\\rho(\\mathbf{r})} d\\mathbf{r} \\tag{3}\n", "$$\n", "\n", "which is a gauge of the sharpness or concentration of the electron density distribution. In Eq.(3), $i^{\\prime}_{\\text{F}}(\\mathbf{r})$ is the Fisher information density and $\\mathbf{\\nabla} \\rho(\\mathbf{r})$ is the density gradient. Earlier, we haveproved that there is an equivalent expression for the Fisher information in terms of the Laplacian of the electron density $\\nabla^2 \\rho(\\mathbf{r})$\n", "\n", "$$\n", "I^{\\prime}_{\\text{F}} \n", "\\equiv \\int i^{\\prime}_{\\text{F}}(\\mathbf{r}) d\\mathbf{r} \n", "= -\\int \\nabla^2\\rho(\\mathbf{r}) \\ln \\rho(\\mathbf{r}) d\\mathbf{r} \n", "\\tag{4}\n", "$$\n", "\n", "Equations (3) and (4) are equal in the sense that they can be derived by partial integration from one to the other, and that the two integrals have the same value. As have been shown, local behaviors of the two integrals, $i_F({\\mathbf{r}})$, and $i^{\\prime}_F(\\mathbf{r})$, are markedly different. More importantly, we have proved the existence of the following rigorous relationship among the three quantities $s_s(\\mathbf{r})$, $i_F({\\mathbf{r}})$, and $i^{\\prime}_F(\\mathbf{r})$,\n", "\n", "$$\n", "S_S = \\int s_S(\\mathbf{r}) d\\mathbf{r} \n", "= -N + \\frac{1}{4\\pi} \n", "\\int\\int \\frac{i_F({\\mathbf{r}}) - i^{\\prime}_F(\\mathbf{r}^{\\prime})}{|\\mathbf{r} - \\mathbf{r}^{\\prime}|}\n", "d \\mathbf{r} d\\mathbf{r}^{\\prime}\n", "\\tag{5}\n", "$$\n", "\n", "whose validity has subsequently been verified by numerical results.\n", "\n", "The third quantity in the same spirit is the Ghosh-Berkowitz-Parr (GBP) entropy\n", "\n", "$$\n", "S_{GBP} = -\\int \\frac{3}{2}k\\rho(\\mathbf{r})\\left[c+ ln\\frac{t(\\mathbf{r};\\rho)}{t_{TF}(\\mathbf{r};\\rho)} \\right] d\\mathbf{r}\n", "\\tag{6} \n", "$$\n", "\n", "where $t(\\mathbf{r}, \\rho)$ is the kinetic energy density, which is related to the total kinetic energy $T_s$ via\n", "\n", "$$\n", "\\int t(\\mathbf{r}, \\rho) d\\mathbf{r} = T_S\n", "\\tag{7}\n", "$$\n", "\n", "and $t_{\\text{TF}}(\\mathbf{r}, \\rho)$ is the Thomas-Fermi kinetic energy density,\n", "\n", "$$\n", "t_{\\text{TF}}(\\mathbf{r}, \\rho) = c_K \\rho^{5/3}(\\mathbf{r}) \n", "\\tag{8}\n", "$$\n", "\n", "with k as the Boltzmann constant, $c = (5/3) + \\ln(4\\pi c_K/3)$, and $c_K = (3/10)(3\\pi^2)^{2/3}$. The GBP entropy originates from the effort to transcribe the ground-state density functional theory into a local thermodynamics through the phase-space distribution function $f(\\mathbf{r}, \\mathbf{p})$ which is a function of both the electron position $\\mathbf{r}$ and momentum $\\mathbf{p}$ as its two basic variables. The conditions of such a recast of DFT into thermodynamics are that the phase-space distribution function is associated with the ground state electron density $\\rho(\\mathbf{r})$ and kinetic energy density $t(\\mathbf{r}, \\rho)$ through the following relationships\n", "\n", "$$\n", "\\rho(\\mathbf{r}) = \\int f(\\mathbf{r}, \\mathbf{p}) d\\mathbf{p} \n", "\\tag{9}\n", "$$\n", "$$\n", "t(\\mathbf{r}, \\rho) = \\frac{1}{2} \\int f(\\mathbf{r}, \\mathbf{p}) p^2 d\\mathbf{p} \n", "\\tag{10}\n", "$$\n", "\n", "The specific form of the local kinetic energy $t(\\mathbf{r}, \\rho)$ used is thefollowing,\n", "\n", "$$\n", "t(\\mathbf{r};\\rho) = \\sum_i \\frac{1}{8} \\frac{\\nabla \\rho_i \\dot \\nabla \\rho_i}{\\nabla\\rho_i}-\\frac{1}{8} \\nabla^2 \\rho\n", "\\tag{11}\n", "$$\n", " \n", "Very recently, three information-theoretic quantities, Renyi entropy, Tsallis entropy, and Onicescu information energy, are introduced as new reactivity descriptors in DFRTs. The Renyi entropy oforder n, where n > 0 and n / 1, is defined as\n", "\n", "$$\n", "R_n^r = \\frac{1}{n-1}ln\\int \\left[ \\frac{\\rho^n(\\mathbf{r})}{\\rho^{n-1}_0(\\mathbf{r})} d\\mathbf{r} \\right]\n", "\\tag{12}\n", "$$\n", "\n", "When n approaches to 1, the Renyi entropy, Eq.(14), reduces tothe Shannon entropy, Eq.(3). The Tsallis entropy of order n isdefined as follows\n", "\n", "$$\n", "T_n = \\frac{1}{n-1} \\int \\left[1 - \\ln \\frac{\\rho^n(\\mathbf{r})}{\\rho^{n-1}_0(\\mathbf{r})} d\\mathbf{r} \\right]\n", "\\tag{13}\n", "$$\n", "\n", "It is a generalization of the standard Boltzmann-Gibbs entropy.The common term in Eqs.(14) and (15) is the integral of the n-th power of the electron density, which is called the Onicescuinformation energy of order n\n", "\n", "$$\n", "E_n = \\frac{1}{n-1}\\int \\rho^n(\\mathbf{r}) d\\mathbf{r} \n", "\\tag{14}\n", "$$\n", "\n", "Onicescu introduced this quantity in an attempt to define a finer measure of dispersion distribution than that of Shannon entropy in information theory\n", "\n", "Closely related to the concept of entropy in information theory is the relative entropy, which is a non-symmetric measureof the entropy difference between two probability distribution functions. Well known examples in the literature are the relative Shannon entropy, also called information gain, Kullback-Leibler divergence, or information divergence, defined byo\n", "\n", "$$\n", "S_r = \\int \\rho(\\mathbf{r}) ln \\frac{\\rho(\\mathbf{r})}{\\rho_0(\\mathbf{r})} d\\mathbf{r}\n", "\\tag{15}\n", "$$\n", "\n", "and the relative Renyi entropy of order n\n", "\n", "$$\n", "R^r_n = \\frac{1}{n-1} \\ln \\left[ \\int \\frac{\\rho^n(\\mathbf{r})}{\\rho^{n-1}_0(\\mathbf{r})} d\\mathbf{r} \\right]\n", "\\tag{16}\n", "$$\n", "\n", "where $\\rho_0(\\mathbf{r})$ is the reference state density satisfying the same normalization condition as $\\rho(\\mathbf{r})$. This reference density can be from the same molecule with different conformation or from the reactant of a chemical reaction when the transition state is investigated.\n", "\n", "