(image) (image) [Previous] [Next]

Charge Trapping and Single-Defect Extraction in Gallium-Nitride Based MIS-HEMTs

Chapter 6 Extraction of the Characteristic Time-Constants

This chapter deals with the extraction of the characteristic time-constants from the stochastic capture and emission events of RTN signals. Section 6.1 will be dedicated to a short description of the basics of HMMs and their relevance for RTN producing single-charge defects. The following sections introduce the most basic case, a simple two-state defect (Section 6.2) followed by more complex cases, namely defects with multiple states (Section 6.3) and systems composed of several arbitrarily shaped defects (Section 6.4). In Section 6.6 different histogram-based methods will be discussed, which allow extracting the time constants of certain defects from their stochastic capture and emission events. In the last part of this chapter, a method to extract the time constants of multiple defects with an arbitrary number of states, an algorithm to train a certain HMM to a set of observations, the Baum-Welch Algorithm will be introduced. After discussing the basics of the Baum-Welch algorithm, an implementation of a HMM library (see Appendix A) will be tested for its robustness against data sampled from a known system of defects.

6.1 Markov Processes and the Hidden Markov Model

Markov processes are widely used to describe stochastic transitions between two or more abstract states across many fields of science (physics, chemistry, speech recognition, robotics, etc.) [110, 153, 154].

In the real world, statistical processes produce observable signals which can be measured by some kind of device. In the case of charge transfer reactions in MIS-HEMTs, the charge cannot be measured directly, but only indirectly due to its electrostatic influence on the (math image). This fact potentially introduces noise in the measurements, which depends on the device itself, the measurement equipment and other systematic errors like the mapping from the drain or gate current to (math image) [89].

Throughout the next sections, the following notation is used:

\( T \)

\( \dots \) the length of the observation sequence

\( N \)

\( \dots \) the number of states in the model

\( M \)

\( \dots \) the number of observable symbols

\( Q = \lbrace q_0,q_1,\dots ,q_{N-1}\rbrace \)

\( \dots \) the set of possible states of the Markov Model

\( X = \lbrace x_0,x_1,\dots ,x_{T-1}\rbrace \)

\( \dots \) the sequence of states from \( Q \)

\( o = \lbrace 0,1,\dots ,{M-1}\rbrace \)

\( \dots \) the set of possible observations

\( \mathcal {O} = \lbrace \mathcal {O}_0,\mathcal {O}_1,\dots ,\mathcal {O}_{T-1}\rbrace   \)

\( \dots \) the sequence of observations from \( o \)

\( \underline {\bm {k}} \)

\( \dots \) the state transition probability matrix

\( \underline {\bm {b}} \)

\( \dots \) the observation probability matrix

\( \pi \)

\( \dots \) the initial state probabilities

The working principle can be seen in Figure 6.1. The grey region denotes the inner state sequence \( X \) of the Markov Model, which can be one of the states \( x_i \) for each item in the observed sequence \( \mathcal {O} \). Note that the each inner state \( x_i \) can only be identified by its corresponding observation \( \mathcal {O}_i \).

(-tikz- diagram)

Figure 6.1: A schematic of the Hidden Markov Model. The hidden state sequence \( x_i \) is unknown and should be reconstucted from the corresponding observations \( \mathcal {O}_i \).

As a subset of all stochastic processes, Markov processes can be described as a series of stochastic events, where each event \( x_t \) from a discrete state space \( Q \) occurs at a certain time \( t \). In general, the set of events \( X \) is described by:

(6.1) \begin{equation} X(t)=\left \lbrace x_t\right \rbrace ,\ t = 0 \dots T-1 \end{equation}

Each of the events \( X(t) \) is determined by its own CDF,

(6.2) \begin{equation} F(x_t,t) = F(x_t) = P\left \lbrace X(t) = x_t \right \rbrace .   \eqlabel {stochasticCDF} \end{equation}

The CDF of the whole set \( X \) can be found by writing down the joint CDF for all \( T \) events:

(6.3) \begin{equation} F(x_0,x_1,...,x_{T-1};t_0,t_1,...,t_{T-1})=P\left \lbrace X(t_0) = x_0, X(t_1) = x_1,...,X(t_{T-1}) = x_{T-1} \right \rbrace \eqlabel {jointCDF} \end{equation}

To actually construct the CDF for a given series of events, conditional probabilities are used to express the probability of the next observation, given a certain history of observations. In general, this can be a very complex task, since the conditional probability depends on all past observations. At that point, the so called Markov property helps to simplify the problem. It states, that for a Markov process, the conditional probability to enter the next state only depends the current state [153, 154]. With other words, Markov processes have no memory and thus the probability to reach a certain state at time \( t+1 \) only depends on the current state:

(6.4) \begin{equation} P\left \lbrace X(t+1) = x_{t+1}|X(t) = x_t \right \rbrace \eqlabel {probabilityMarkov} \end{equation}

In the context of defect capture and emission events we look at continuous-time discrete-space Markov processes, also called Markov chains, which will be used in the following section to calculate the PDF of a simple two-state defect.

The state transition probability matrix \( \underline {\bm {k}} \) is of size \( N\times N \). It contains the conditional probabilities to go from state \( i \) to state \( j \) and each of the rows sum to one because the probability of being in one of the states is one (i.e.,\( \underline {\bm {k}} \) is row stochastic). Note that in this case the braces only mark the instants in time of the state sequence \( X \) as the transition probabilities of Markov chains are time independent.

(6.5) \begin{equation} k_{ij}=P\lbrace X(t+1)=q_j|X(t)=q_i\rbrace \eqlabel {rtnextract:hmm:k} \end{equation}

The observation probability matrix \( \underline {\bm {b}} \) is also row stochastic and time independent. It holds the probabilities to observe the symbol \( o_i \) given a certain state \( q_i \). The size is \( N\times M \), as the number of possible observations not necessarily reflects the number of inner states. One example would be thermal transitions of a defect without charge transfer.

(6.6) \begin{equation} b_{i}=P\lbrace o_i(t)|q_i(t)\rbrace \eqlabel {rtnextract:hmm:b} \end{equation}

The HMM is fully defined by \( \underline {\bm {k}} \), \( \underline {\bm {b}} \) and \( \pi \), and is denoted by \( \lambda ( \underline {\bm {k}}, \underline {\bm {b}}, \pi ) \).